Helix AI Engineer, Generative AI
full-time
mid
Posted 5 days ago
About this role
Figure is an AI robotics company developing autonomous general-purpose humanoid robots. Our goal is to build embodied AI systems that can perceive, reason, and act in the real world. Figure is headquartered in San Jose, CA, and this role requires 5 days/week in-office collaboration.
Our Helix team is responsible for developing the core AI systems that power humanoid autonomy. We are looking for a Helix AI Engineer, Generative AI to build and scale generative models that enable robots to understand, simulate, and interact with the physical world. This role focuses on training and deploying diffusion and generative models across vision, video, and multimodal domains, with applications spanning perception, data generation, and model-based reasoning.
Responsibilities
Design, train, and deploy large-scale generative models, with a focus on diffusion-based approaches for vision, video, and multimodal data
Develop models that improve robot perception, world modeling, and prediction from raw sensory inputs
Build generative systems for synthetic data creation, augmentation, and dataset scaling for robot learning
Explore and implement state-of-the-art techniques in diffusion, generative modeling, and multimodal foundation models
Optimize training pipelines for large-scale generative models across distributed systems
Work closely with data, training infrastructure, and agent teams to integrate generative models into the full autonomy stack
Evaluate model quality, robustness, and generalization across real-world scenarios
Contribute to the design of scalable experimentation frameworks for generative model development
Requirements
Experience training and deploying generative models (diffusion, autoregressive, or related approaches) at scale
Strong understanding of modern deep learning techniques for vision and/or multimodal systems
Proficiency in Python and deep learning frameworks such as PyTorch
Experience working with large-scale datasets and distributed training systems
Strong experimental rigor and ability to iterate quickly on model performance
Solid software engineering skills and ability to build reliable, maintainable systems
Ability to operate independently and own ambiguous, high-impact technical problems
Bonus Qualifications
Experience with diffusion models for image or video generation
Experience with multimodal foundation models (vision-language or vision-language-action)
Background in synthetic data generation or simulation for robotics or embodied AI
Experience optimizing large-scale training (multi-node, GPU clusters, etc.)
Familiarity with 3D, video prediction, or world models
Prior work in robotics, embodied AI, or real-world ML systems
Publication record in machine learning, computer vision, or generative modeling
The pay offered for this position may vary based on several individual factors, including job-related knowledge, skills, and experience. The total compensation package may also include additional components/benefits depending on the specific role. This information will be shared if an employment offer is extended.
Similar Jobs
Related searches:
On-site Jobs
Mid-Level Jobs
On-site Mid-Level Jobs
Mid-Level Machine LearningMid-Level Generative AIMid-Level Backend & SystemsMid-Level Robotics & AutonomyMid-Level Computer VisionMid-Level AI Infrastructure
AI Jobs in San Jose
Machine Learning in San JoseGenerative AI in San JoseBackend & Systems in San JoseRobotics & Autonomy in San JoseComputer Vision in San JoseAI Infrastructure in San Jose
computer-visionroboticsgpupytorchdeep-learningdistributed-systemsdiffusion-modelsgenerative-ai