Principal Dev Ops Engineer- Data Platform

Los Angeles 22 months agoFull-time External
Negotiable
We are building data platforms with on-prem data centers and public clouds in multiple regions. Our stack is fairly modern. We use Kubernetes, Istio, Kong, Kafka, Flink, Spark. All new applications are developed to run in containers Our cloud infrastructure is completely described and managed using infrastructure-as-code (IAC) and configuration management tools. At the data-layer, the stack is built on top of technologies including Spark, Kafka, Hadoop For monitoring we use VictoriaMetrics, Prometheus and Grafana We use ElasticSearch , Kibana and Fluentd for logs. Responsibilities: • You'll work on high impact projects that improve data availability, scalability, and reliability of our data infrastructure. • You will manage several Kubernetes, Kafka and Hadoop cluster • You will manage resources in AWS and Azure • You will assist our system architecture team to improve the on-prem data platform. • You will design, architect, improve and support new and existing tools to help us operate at scale. • You will own and operate cloud infrastructure in AWS/Azure by using IAC tools like Terraform, Packer, Chef and Ansible • And, finally, join us in our oncall rotation. Qualifications: • Have a strong understanding of Linux, networking and production systems. • Have 10+ years experience building and maintaining infrastructure and services • Experience in managing Production grade Kubernetes clusters and have deep understanding of securing workloads • Strong understanding of Kafka, Hadoop and Spark • Proficient at scripting and programming • Experience in using IAC and config management tools like Terraform, Packer, Chef and Ansible • Bachelor's degree in computer engineering or equivalent Nice to have: • Due to the nature of the work, being a US citizen is preferred but not mandatory. • Some experience operating infrastructure servicing federal customers #LI-AZ2