Data Engineer(Hadoop, Spark)

Singapore 16 days agoFull-time External
Negotiable
**Key Responsibilities** - Design, develop, and maintain robust data pipelines and ETL processes to support analytics and reporting needs. - Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver high-quality solutions. - Implement data integration solutions across structured and unstructured data sources. - Ensure data quality, integrity, and security across all stages of the data lifecycle. - Optimize data workflows for performance and scalability in cloud and on-premise environments. - Support data migration and transformation initiatives for client projects. - Monitor and troubleshoot data pipeline issues and provide timely resolutions. **Required Qualifications** - Bachelor’s degree in Computer Science, Information Systems, Engineering, or related field. - 3+ years of experience in data engineering or related roles. - Proficiency in SQL and Python or Scala. - Experience with data pipeline tools such as Apache Spark, Kafka, Airflow, or similar. - Familiarity with cloud platforms (AWS, Azure, or GCP). - Strong understanding of data warehousing concepts and tools (e.g., Snowflake, Redshift, BigQuery). - Knowledge of data governance, security, and compliance standards.