Machine Learning Research Intern jobs in United States
cer-icon
Apply on Employer Site
company-logo

d-Matrix · 3 hours ago

Machine Learning Research Intern

d-Matrix is focused on unleashing the potential of generative AI to transform technology. They are seeking a Machine Learning Research Intern to join their team, responsible for investigating algorithm-hardware co-design and developing efficient deep learning techniques for their AI compute engine.

AI InfrastructureArtificial Intelligence (AI)Cloud InfrastructureData CenterSemiconductor
check
H1B Sponsor Likelynote

Responsibilities

Design, implement and evaluate efficient deep neural network architectures and algorithms for d-Matrix's AI compute engine
Engage and collaborate with internal and external ML researchers to meet R&D goals
Engage and collaborate with Software team to meet stack development milestones
Conduct research to guide hardware design
Develop and maintain tools for high-level simulation and research
Port customer workloads, optimize them for deployment, generate reference implementations and evaluate performance
Report and present progress timely and effectively
Contribute to publications of papers and intellectual property

Qualification

PyTorchAlgorithm analysisDeep learningPython programmingC/C++ programmingGPU CUDA programmingNeural network architecturesAutoMLMeta learningNumerical analysisHW accelerator systems

Required

Pursuing Masters/PhD degree in Computer Science, Electrical and Computer Engineering, or a related scientific discipline
High proficiency with PyTorch is a must
High proficiency in algorithm analysis, data structure, and Python programming is a must
Current knowledge in machine learning and modern deep learning
Hands-on experience with modern neural network architectures such as MoEs and Diffusion models

Preferred

Knowledge and experience with efficient deep learning is preferred: quantization, sparsity, distillation
Strong publication records in top machine learning conferences or journals is preferred
Proficiency with C/C++ programming is preferred
Proficiency with GPU CUDA programming is preferred
Experience with AutoML and meta learning is preferred
Experience with numerical analysis preferred
Experience with specialized HW accelerator systems for deep neural network is preferred

Company

d-Matrix

twittertwittertwitter
company-logo
D-Matrix is a platform that enables data centers to handle large-scale generative AI inference with high throughput and low latency.

H1B Sponsorship

d-Matrix has a track record of offering H1B sponsorships. Please note that this does not guarantee sponsorship for this specific role. Below presents additional info for your reference. (Data Powered by US Department of Labor)
Distribution of Different Job Fields Receiving Sponsorship
Represents job field similar to this job
Trends of Total Sponsorships
2025 (20)
2024 (15)
2023 (8)
2022 (7)

Funding

Current Stage
Growth Stage
Total Funding
$429M
Key Investors
Temasek HoldingsTSVC
2025-11-12Series C· $275M
2023-09-06Series B· $110M
2022-04-20Series A· $44M

Leadership Team

leader-logo
Peter Buckingham
Senior Vice President, Software Engineering
linkedin
Company data provided by crunchbase