About the Role
We are looking for a driven and detail-oriented
L5B Program Manager to join our
Frontier Labs AI team, focused on building high-quality, multi-modal data pipelines to support advanced model development and foundational research.
In this role, you will lead the
end-to-end execution of AI data labeling workflows across
text, image, audio, video, and instruction-tuned datasets, partnering closely with researchers, data scientists, product managers, and annotation vendors. You will play a critical role in
scaling and operationalising labeling operations, ensuring that the data used to train and evaluate cutting-edge models is accurate, diverse, and aligned with evolving research needs.
This is a hands-on role for someone who thrives in
high-ambiguity, high-velocity environments and can bring structure and discipline to rapidly evolving labeling workflows
- What You Will Do -Program Execution & Delivery- Manage AI data labeling programs from scoping to delivery, ensuring high-quality annotations at scale.
- Translate Frontier Labs research needs into concrete annotation specs, rubrics, and task designs.
Own timelines, throughput plans, and quality controls for critical datasets used in LLM training and evaluation.
Stakeholder Management- Partner with researchers, data scientists, product, and ops to ensure labeling goals are aligned with model objectives.
- Work cross-functionally to drive task clarity, resolve ambiguity, and incorporate feedback into successive batches.
- Act as the single-threaded owner for specific labeling programs, managing internal and external partners.
Operational Infrastructure- Develop and refine batching strategies, smart sampling plans, and audit workflows.
- Drive QA processes, including golden set calibration, rubric refinement, and disagreement adjudication.
- Ensure traceability from raw inputs to final labeled outputs, and track quality regressions over time.
Process Design & Automation- Identify opportunities to apply model-in-the-loop labeling, active learning, or self-checking pipelines.
- Collaborate with tool owners and engineers to integrate annotation workflows with internal tooling systems.
Own feedback loops that enable raters to improve over time and reduce error variance
- What You Will Need -Bachelor's degree in Engineering, Data Science, Linguistics, or related technical/analytical field.
5+ years of program or project management experience in AI/ML, data ops, or labeling infrastructure.
Demonstrated ability to manage
end-to-end data pipelines in AI/ML or research environments.
Strong working knowledge of
Robotics, Physical AI Data labeling tasks, such as:
- Object detection and recognition
- Semantic & Instance Segmentation
- Depth & Pose Estimation
- Grasp Detection
- Action Segmentation
- Trajectory Labeling
- Prompt-response evaluation
- Instruction tuning
- Dialogue evaluation
- Vision-language QA
- Video slot tagging
- Image Tagging
- Documentation Extraction
- Data collection annotation
- HRI
Experience collaborating with research or model teams to scope data collection requirements.
Excellent written and verbal communication skills
- Preferred Qualifications -
- Experience in frontier AI research environments, such as foundation model labs or GenAI startups.
- Familiarity with tools like Label Studio, Scale AI, SuperAnnotate, Snorkel Flow, or in-house annotation platforms.
- Understanding of LLM training and evaluation lifecycles.
- Experience working with human-in-the-loop systems or model-assisted labeling pipelines.
- Familiarity with multilingual or multi-cultural annotation programs