Senior Data Engineer

Cobra Studio is seeking a Senior Data Engineer (Python/Java) with 5+ years of experience. This full-time, remote role (paid in USD) focuses on building cloud-based data lakes, real-time pipelines (Kafka/PubSub), and scalable ETL processes using tools like DBT, Airflow, and GCP technologies. Strong SQL and data warehouse experience required.

Data Engineering
Data Warehouse
DBT
Python
Java
ETL
ELT
Airflow
Kafka
GCP
SQL
* Only applicants living in: Argentina, Chile, Uruguay, Paraguay, Ecuador, Guatemala, El Salvador, Honduras, República Dominicana

JOB DESCRIPTION:

Benefits and conditions

  • 100% remote work – collaborate from anywhere in the world.
  • Flexible working hours to support a healthy work-life balance.
  • All necessary equipment provided to set you up for success.
  • Engaging, dynamic projects with plenty of room for learning and professional growth.
  • Competitive compensation in USD.

Requirements

  • 5+ years of experience as a Data Engineer.
  • Strong programming skills in Python or Java.
  • Proven experience building data lakes in cloud environments (preferably GCP).
  • Expertise in ETL/ELT pipeline development, using tools like Airflow.
  • Solid experience with DBT and data warehouse metric modeling.
  • Hands-on experience with real-time streaming using Kafka or Pub/Sub.
  • Familiarity with data processing tools such as Hadoop, Spark, or their managed equivalents in GCP (e.g. Dataproc, Dataflow).
  • Proficiency in writing complex SQL queries.
  • Minimum B2 English level (verbal and written communication).

Soft skills

  • Strong problem-solving and analytical thinking.
  • Effective communication and collaboration in remote teams.
  • High attention to detail and data quality.
  • Proactive and self-driven work style.
  • Adaptability to changing priorities and technologies.
  • Team-oriented mindset and knowledge sharing.

Responsibilities

  • Design, build, and maintain scalable ETL/ELT pipelines.
  • Develop and optimize data models and metrics using DBT and SQL.
  • Build and manage real-time data streaming solutions (Kafka, Pub/Sub).
  • Develop and maintain cloud-based data lakes and warehouses (preferably in GCP).
  • Collaborate with data analysts, engineers, and product teams to deliver data solutions.
  • Ensure data quality, integrity, and performance across all pipelines.
  • Monitor, debug, and improve data workflows using tools like Airflow and Spark.

----