Job Description

As an AI Research Apprentice you'll push the frontiers of generative and multimodal learning that power our autonomous robots. You will prototype diffusion-based vision models, vision–language architectures (VLAs/VLMs) and automated data-annotation pipelines that turn raw site footage into training gold.
Key Responsibilities
* Design and train diffusion-based generative models for realistic, high-resolution synthetic data.
* Build compact Vision–Language Models (VLMs) to caption, query and retrieve job-site scenes for downstream perception tasks.
* Develop Vision–Language Alignment (VLA) objectives that link textual work-orders with pixel-level segmentation masks.
* Architect large-scale auto-annotation pipelines that transform unlabeled images / point-clouds into high-quality labels with minimal human input.
* Benchmark model performance on accuracy, latency and memory for deployment on Jetson-class hardware; compress with distillation or LoRA.
* Collaborate with ...

Apply for this Position

Ready to join Origin? Click the button below to submit your application.

Submit Application