Data Engineer – ETL & Streaming

Dubai Tax Free3 days agoContractor External
28.4k - 56.8k / mo
Data Engineer – ETL & Streaming (Kafka, Airflow, Spark) Location: UAE Experience: 4–7 years Department: Data Engineering About the Role We are looking for a Data Engineer to develop robust and scalable data pipelines for both batch and real-time processing using Apache Kafka, Airflow, and Spark. This role focuses on implementing business logic, transformations, and data delivery into data lakes and warehouses. Key Responsibilities • Build ETL and ELT pipelines integrating multiple source systems (RDBMS, APIs, events). • Develop data ingestion and transformation jobs in Spark (PySpark or Scala). • Design and deploy Airflow DAGs to orchestrate end-to-end data workflows. • Implement real-time data ingestion and processing using Kafka (streams, connectors, topics). • Ensure data quality, consistency, and lineage across datasets. • Collaborate with data modelers and BI teams to deliver curated datasets. • Troubleshoot pipeline failures, optimize jobs for performance and cost efficiency. • Work closely with the platform team to integrate pipelines into the enterprise data platform. • Follow DevOps and CI/CD best practices for deployment and monitoring. Skills & Qualifications • Strong programming skills in Python (preferred) or Scala/Java. • Hands-on experience in Kafka (producers, consumers, stream processing, schema management). • Proficient in Airflow DAG development, scheduling, and task dependency handling. • Solid experience with Spark (batch, streaming, dataframes, and optimization). • Working knowledge of SQL and experience with data modeling (Star/Snowflake). • Experience with data lake and warehouse environments (Azure Synapse, Databricks, Snowflake, Redshift, etc.). • Understanding of data versioning, partitioning, and incremental load strategies. • Familiar with Git-based development, CI/CD, and observability practices. Job Type: Contract Contract length: 12 months Pay: AED15,000.00 - AED30,000.00 per month