Data Engineer – SC Cleared – Databricks
We are seeking a hands-on Data Engineer with deep expertise in building and managing streaming and batch data pipelines. The ideal candidate will have strong experience working with large-scale data systems operating on cloud-based platforms such as AWS and Databricks. This role also involves close collaboration with hyperscalers and data platform vendors to evaluate and document Proofs of Concept (PoCs) for modern data platforms, while effectively engaging with senior stakeholders across the organisation.
Key Responsibilities:
• Design, develop, and maintain streaming and batch data pipelines using modern data engineering tools and frameworks.
• Work with large volumes of structured and unstructured data, ensuring high performance and scalability.
• Collaborate with cloud providers and data platform vendors (e.g., AWS, Microsoft Azure, Databricks) to conduct PoCs for data platform solutions.
• Evaluate PoC outcomes and provide comprehensive documentation including architecture, performance benchmarks, and recommendations.
Required Experience & Skills:
• Proven experience as a Data Engineer with a strong focus on streaming and batch processing.
• Hands-on experience with cloud-based data plaforms such as AWS/ Databricks.
• Strong programming skills in Python, Scala, or Java.
• Experience with data modeling, ETL/ELT processes, and data warehousing.
• Experience conducting and documenting PoCs with hyperscalers or data platform vendors.
Preferred Qualifications:
• Certifications in AWS, Azure, or Databricks.
• Experience with Snowflake, IBM DataStage, or other enterprise data tools.
• Knowledge of CI/CD pipelines and infrastructure as code (e.g., Terraform, CloudFormation).