Principal Software Engineer - Dynamo jobs in United States
cer-icon
Apply on Employer Site
company-logo

NVIDIA · 3 hours ago

Principal Software Engineer - Dynamo

NVIDIA is an innovative company focused on AI technologies, and they are seeking a Principal Software Engineer for their Dynamo project. The role involves addressing complex challenges in distributed inference systems, including developing Kubernetes deployment stacks, optimizing GPU resource management, and enhancing intelligent routing systems.

Artificial Intelligence (AI)Consumer ElectronicsGPUHardwareSoftwareVirtual Reality
check
Growth Opportunities
check
H1B Sponsor Likelynote

Responsibilities

Collaborate on the design and development of the Dynamo Kubernetes stack
Introduce new features to the Dynamo Python SDK and Dynamo Rust Runtime Core Library
Design, implement, and optimize distributed inference components in Rust and Python
Contribute to the development of disaggregated serving for Dynamo-supported inference engines (vLLM, SGLang, TRT-LLM, llama.cpp, mistral.rs)
Improve intelligent routing and KV-cache management subsystems
Contribute to open-source repositories, participate in code reviews, and assist with issue triage on GitHub
Work closely with the community to address issues, capture feedback, and evolve the framework’s APIs and architecture
Write clear documentation and contribute to user and developer guides

Qualification

RustC++PythonKubernetesDistributed systemsCloud-native deploymentContainer orchestrationLarge-scale inferenceMemory managementOpen-source developmentProblem-solvingCommunication skills

Required

BS/MS or higher in computer engineering, computer science or related engineering (or equivalent experience)
15+ years of proven experience in related field
Strong proficiency in systems programming (Rust and/or C++), with experience in Python for workflow and API development
Experience with Go for Kubernetes controllers and operators development
Deep understanding of distributed systems, parallel computing, and GPU architectures
Experience with cloud-native deployment and container orchestration (Kubernetes, Docker)
Experience with large-scale inference serving, LLMs, or similar high-performance AI workloads
Background with memory management, data transfer optimization, and multi-node orchestration
Familiarity with open-source development workflows (GitHub, continuous integration and continuous deployment)
Excellent problem-solving and communication skills

Preferred

Prior contributions to open-source AI inference frameworks (e.g., vLLM, TensorRT-LLM, SGLang)
Experience with GPU resource scheduling, cache management, or high-performance networking
Understanding of LLM-specific inference challenges, such as context window scaling and multi-model agentic workflows

Benefits

Equity
Benefits

Company

NVIDIA is a computing platform company operating at the intersection of graphics, HPC, and AI.

H1B Sponsorship

NVIDIA has a track record of offering H1B sponsorships. Please note that this does not guarantee sponsorship for this specific role. Below presents additional info for your reference. (Data Powered by US Department of Labor)
Distribution of Different Job Fields Receiving Sponsorship
Represents job field similar to this job
Trends of Total Sponsorships
2025 (1418)
2024 (1356)
2023 (976)
2022 (835)
2021 (601)
2020 (529)

Funding

Current Stage
Public Company
Total Funding
$4.09B
Key Investors
ARPA-EARK Investment ManagementSoftBank Vision Fund
2023-05-09Grant· $5M
2022-08-09Post Ipo Equity· $65M
2021-02-18Post Ipo Equity

Leadership Team

leader-logo
Jensen Huang
Founder and CEO
linkedin
leader-logo
Michael Kagan
Chief Technology Officer
linkedin
Company data provided by crunchbase