Senior Data Engineer
Cobra Studio is seeking a Senior Data Engineer (Python/Java) with 5+ years of experience. This full-time, remote role (paid in USD) focuses on building cloud-based data lakes, real-time pipelines (Kafka/PubSub), and scalable ETL processes using tools like DBT, Airflow, and GCP technologies. Strong SQL and data warehouse experience required.
JOB DESCRIPTION:
Benefits and conditions
- 100% remote work – collaborate from anywhere in the world.
- Flexible working hours to support a healthy work-life balance.
- All necessary equipment provided to set you up for success.
- Engaging, dynamic projects with plenty of room for learning and professional growth.
- Competitive compensation in USD.
Requirements
- 5+ years of experience as a Data Engineer.
- Strong programming skills in Python or Java.
- Proven experience building data lakes in cloud environments (preferably GCP).
- Expertise in ETL/ELT pipeline development, using tools like Airflow.
- Solid experience with DBT and data warehouse metric modeling.
- Hands-on experience with real-time streaming using Kafka or Pub/Sub.
- Familiarity with data processing tools such as Hadoop, Spark, or their managed equivalents in GCP (e.g. Dataproc, Dataflow).
- Proficiency in writing complex SQL queries.
- Minimum B2 English level (verbal and written communication).
Soft skills
- Strong problem-solving and analytical thinking.
- Effective communication and collaboration in remote teams.
- High attention to detail and data quality.
- Proactive and self-driven work style.
- Adaptability to changing priorities and technologies.
- Team-oriented mindset and knowledge sharing.
Responsibilities
- Design, build, and maintain scalable ETL/ELT pipelines.
- Develop and optimize data models and metrics using DBT and SQL.
- Build and manage real-time data streaming solutions (Kafka, Pub/Sub).
- Develop and maintain cloud-based data lakes and warehouses (preferably in GCP).
- Collaborate with data analysts, engineers, and product teams to deliver data solutions.
- Ensure data quality, integrity, and performance across all pipelines.
- Monitor, debug, and improve data workflows using tools like Airflow and Spark.
----