Devops and Data Engineer

Singapore 1 months agoFull-time External
43.8k - 60.3k / mo
This job posting has expired and is no longer accepting applications.
Role Overview: ETL & Data Engineer to design, build, and maintain robust data pipelines and backend services that power AI-driven operations. The role involves working with high-volume IT and cloud data, optimizing ETL processes, and integrating with AIOps platforms and ML pipelines. Key Responsibilities: • Build and maintain scalable ETL pipelines for batch and real-time data ingestion, transformation, and loading from diverse sources (IT infrastructure, cloud, monitoring systems, APIs). • Implement data validation, cleansing, and normalization for consistent AI model input. • Develop backend services and APIs to support data ingestion, metadata management, and configuration. • Optimize ETL jobs for performance, fault tolerance, and low latency. • Integrate with AIOps platforms and ML pipelines using REST APIs or event-driven architectures. • Schedule and monitor ETL workflows using tools like Airflow, Prefect, or Dagster. • Support CI/CD pipelines for deploying ETL services and full-stack applications. Required Skills & Tools: • Programming & Scripting: Python, Go (Golang), Java, Ruby, JavaScript/TypeScript (Next.js) • ETL & Data Engineering: Apache NiFi, Spark, Airflow, Flink, Kafka, Talend • Orchestration: Airflow, Prefect, Dagster • Data Storage & Lakes: PostgreSQL, MongoDB, Elasticsearch, Snowflake, BigQuery, S3, GCS, Azure Blob • Streaming Platforms: Kafka, Kinesis, Pub/Sub Good to Have: • Experience with AIOps & Observability tools like Splunk, Dynatrace, AppDynamics, New Relic, Elastic Stack • Familiarity with ITSM systems (ServiceNow) and CMDB integrations • Understanding of metrics, logs, and traces for AI-driven operations