Data Engineer – Talend to Spark Migration
We are looking for a Data Engineer with strong experience in Talend ETL workflows and Apache Spark to lead the migration of our data pipelines from Talend to Spark. This role involves re-engineering existing ETL processes, optimizing performance, and ensuring seamless integration with modern big data platforms.
• Analyze existing Talend ETL jobs and design equivalent Spark-based pipelines.
• Migrate and optimize data workflows from Talend to Spark (PySpark/Scala).
• Collaborate with data architects to define scalable data lake/lakehouse solutions.
• Implement best practices for data quality, governance, and security.
• Integrate Spark pipelines with cloud platforms (AWS, Azure, GCP).
• Work with orchestration tools (Airflow, Databricks Jobs, etc.) to schedule and monitor workflows.
Required Skills
• Bachelor’s/Master’s degree in Computer Science, Engineering, or related field.
• Hands‑on experience with Talend ETL development (tMap, tJoin, tFilter, etc.).
• Strong expertise in Apache Spark (PySpark, Spark SQL, Spark Streaming).
• Proficiency in Python, Scala, or Java.
• Solid understanding of data warehousing concepts and SQL.
• Experience with big data ecosystems (Hadoop, Hive, Kafka).
• Familiarity with cloud services (AWS EMR, Azure Databricks, GCP DataProc).
Interested candidates, who wish to apply for the above position, please send in your resume to: ansari.abdulrahman@persolapac.com
We regret to inform you that only shortlisted candidates will be contacted.