Founding Machine Learning – World Models

Back to all jobs
  • One Robot
  • San Francisco, CA
  • Full-Time
  • 1 week ago
  • $150k-275k
Published
April 30, 2026
Location
San Francisco, CA
Category
Job Type

Founding Machine Learning – World Models: our view in 3 lines...

  • The Role: Develop generative world models and training infrastructure to simulate robot manipulation scenes for validating and eventually training robot policies.
  • The Person: Design and train action-conditioned video and dynamics world models, develop architectures for long-horizon coherence, run and debug multi-GPU training infrastructure including custom CUDA, and build the world-model data engine.
  • Requirements: Very strong coding in Python and PyTorch, video generation experience, track record operating training runs at cluster scale, and working knowledge of multi-view geometry, scene reconstruction, and physical priors.

Job Description

We build world models that simulate manipulation scenes faithfully enough to validate, and one day, train policies without touching a robot. You'll develop generative models that make this work, with the controllability and physical fidelity to match real-robot behavior.

What you'll do:

  • Train video and dynamics models: Develop world models with action conditioning for manipulation policies.
  • Push long-horizon coherence: Develop architectures and training methods that extend rollout quality on hard physical tasks.
  • Own training infrastructure: Run multi-GPU clusters, write custom CUDA, debug at scale.
  • Build the world-model data engine: Design, implement, and improve a data engine that allows the world model to compound learning across customers and manipulation tasks.

Requirements:

  • Very strong coding in Python and PyTorch (or similar).
  • Video generation experience: Deep experience training image or video generation models end-to-end.
  • Large-scale training: Track record operating training runs at cluster scale.
  • 3D vision: Working knowledge of multi-view geometry, scene reconstruction, and physical priors.

One Robot builds task-specific world models and an evaluation platform for robot manipulation policies.

Training end-to-end policies for robots is vibes-based today. Teams collect data, train, deploy on a real robot, find out what fails, collect more, retry. We replace the trial-and-error with rigorous validation that tells you where your policy will fail and what data to collect to fix it.

Robotics can't industrialize without an evaluation layer. We're building it.

We're solving challenging technical problems around long-horizon autoregressive generation, world model controllability, and closing the sim-to-real gap. We work with real customer data, real failures, and real deployment pressure.

We're based in San Francisco, backed by Accel, YC, several exited founders, and engineering leaders at leading AI companies.

We're small and deliberately so. Everyone is an IC with deep ownership of a wide surface area. The culture is fast iteration and direct responsibility.

Hemanth Sarabu and Elton Shon co-founded One Robot after leading robot learning together at Industrial Next (YC W22), bringing experience from Google, NASA JPL, and Tesla.

Key Skills
? Key Skills in dark blue have been inferred based on similar industry roles
CUDA Video Generation Multi-gpu Cluster Training 3D Vision Multi-view Geometry Scene Reconstruction Python Machine Learning Pytorch

Subscribe to Career Resources

Get the latest career advice, industry insights, and job opportunities delivered to your inbox.