VLA & LLM Engineer (Model Training & Data Validation)

New York 7 days agoFull-time External
1m - 1.7m / yr
About Mecka AI Mecka AI is building the data infrastructure layer for robotics and embodied AI. We partner with leading AI labs and robotics companies to deliver high-quality, real-world datasets used to train, evaluate, and deploy robotic systems. Our work sits directly between research, data, and real-world execution — where model performance is dictated by data quality. The Role We are looking for a VLA & LLM Engineer to serve as the primary engine for our large model training efforts. Your role is twofold: first, you will own the automated labeling system by training and fine-tuning LLMs and VLAs to process our data at scale. Second, you are the ultimate validator—you must prove that Mecka’s data actually works by training high-performing models that demonstrate superior results in real-world tasks. This is not a theoretical research post. You will own the full process: from writing the math for a new architecture to debugging the production code that ships it. We don't hand you a finished system to work on; you are responsible for getting the models working from the start and meeting tight product deadlines. What You’ll Work On Automated Labeling & Pipeline Intelligence (Primary Focus) Training the Pipeline: Train and fine-tune Large Language Models (LLMs) and Vision-Language-Action (VLA) models to automatically label, segment, and describe massive robotics datasets. Knowledge Application: Use your ML expertise to improve our current processing pipelines. You will apply what you learn from training failures to refine how data is ingested, cleaned, and represented. Semantic Reasoning: Develop systems where LLMs reason about high-level task goals while VLAs ground those goals in low-level sensory and action data. Data Validation (Proving the Product) The "Proof" Phase: You are responsible for proving that Mecka’s data is the best in the world. You will do this by training models that achieve state-of-the-art performance in both simulated and physical environments. Full-Cycle Ownership: You are in charge of the full process. You will handle the setup, training, and debugging of the models. If a robot isn't performing, you are responsible for diagnosing if it’s an architectural flaw or a data gap. Performance Benchmarking: Design experiments to show exactly how Mecka's data improves model generalization and failure-mode recovery. Cross-Functional Support Assisting All Teams: Apply your deep understanding of model mechanics to assist all engineering and research functions. Production Debugging: Write, test, and debug production-grade code. You are motivated by shipping a functional product by the deadline, not just generating metrics for a paper. Who You Are Required Background Education: A PhD or a Research-based Master’s degree in Computer Science, Robotics, Machine Learning, Physics, Mathematics, or a related quantitative field. Proven Research History: A strong track record (publications at conferences like NeurIPS, ICRA, CVPR, or significant open-source contributions) demonstrating your ability to train and innovate within LLM or VLA frameworks. Mathematical Rigor: You understand the first principles of transformers, multi-modal embeddings, and scaling laws. You don't just "tune" models; you understand the underlying math. Production Engineer: You are an expert at writing and debugging production-grade code. You are motivated by shipping a functional product to a deadline rather than just publishing a paper. Full-Stack ML Execution: You have experience taking a model from a blank script to a deployed, working system (in sim or on hardware). Strong Signals: Tenacity: You are comfortable being handed a non-functional system and taking total ownership until it works. Architectural Intuition: You can look at a training curve and intuitively understand which architectural or data-level changes are needed. Builder Mentality: You are research-driven but pragmatic. You want to see your research move real robots and solve real-world data problems. Why This Role Work directly with real-world robotics data at scale Influence how leading robotics teams collect and use data Apply your VLA/LLM expertise to real-world datasets that are larger and richer than almost any academic lab. Compensation Range: $150K - $250K