SIGN IN
AI Robotics Research Engineer, Vision-Language-Action (VLA) jobs in United States
cer-icon
Apply on Employer Site
company-logo

Confidential · 15 hours ago

AI Robotics Research Engineer, Vision-Language-Action (VLA)

Confidential company is looking for a Vision-Language-Action (VLA) Research Engineer to help build the next generation of intelligent robotic systems. The role involves developing models that enable robots to perceive the world, reason with language, and take meaningful actions in real-world environments, while collaborating with systems engineers to translate research into practical applications.
Marketing & Advertising

Responsibilities

Design, implement, and evaluate Vision-Language-Action models for embodied agents and robotic systems
Develop multimodal learning pipelines combining visual perception, language understanding, and action/control
Train and fine-tune large-scale models using simulation and real-world robotic data
Explore approaches such as imitation learning, reinforcement learning, foundation models, and policy learning
Integrate perception and decision-making models with robotic hardware and simulators
Conduct experiments, analyze results, and iterate rapidly on model architectures
Collaborate on research publications, internal reports, and technical documentation
Stay current with the latest research in robotics, multimodal learning, and foundation models

Qualification

Machine LearningComputer VisionDeep Learning FrameworksVision-Language ModelsPythonRobotics ConceptsReinforcement LearningC++Research PublicationsTechnical Documentation

Required

Strong background in machine learning, robotics, computer vision, or NLP
Experience with deep learning frameworks (e.g., PyTorch, JAX, TensorFlow)
Hands-on experience with vision-language models, policy learning, or embodied AI
Solid programming skills in Python (C++ a plus)
Familiarity with robotics concepts such as control, kinematics, sensors, or simulation
Ability to read, implement, and extend research papers

Preferred

Experience training or deploying Vision-Language-Action or multimodal foundation models
Experience with robotic simulators (e.g., Isaac Sim, MuJoCo, Habitat, Gazebo)
Background in reinforcement learning, imitation learning, or offline RL
Experience working with real robotic platforms (manipulation, navigation, mobile robots, etc.)
Publications in top-tier conferences or journals (e.g., RSS, ICRA, CoRL, NeurIPS, ICML, CVPR)
Experience scaling training pipelines on distributed or cloud systems

Company

Confidential

twitter
company-logo
This page is owned and operated by JAF HOLDINGS INC. All rights reserved 2025. jafholdingsinc.com

Funding

Current Stage
Growth Stage

Leadership Team

leader-logo
Roshan Syed
Global Talent Partner
linkedin
Company data provided by crunchbase