Argonne National Laboratory · 4 days ago
Software Engineer - AI Inference for Science
Argonne National Laboratory is dedicated to accelerating scientific discoveries and engineering breakthroughs through world-leading computing facilities. They are seeking a Software Engineer to enable AI for science, focusing on scalable inference using HPC systems and AI accelerators, while collaborating with various teams to solve complex scientific problems.
Responsibilities
Explore and engineer solutions for AI inference integrated within scientific workflows
Utilize programmatic access using standard programming interfaces (e.g. OpenAI API)
Submit large batches of prompts for parallel processing
Ensure efficient execution on underlying resources, including ALCF’s HPC systems and AI testbed machines
Adapt and deploy new AI models and functionality as they evolve
Qualification
Required
Experience with at least one AI framework is required, such as PyTorch or TensorFlow
Comprehensive experience programming in one or more programming languages such as Python, C/C++
Ability to create, maintain, and support high-quality software is essential
Work with and contribute to domain-specific software and models
Experience with version control software such as git
Ability to work collaboratively in a fast-paced environment
Effective written and oral communications skills
Ability to model Argonne's core values of impact, safety, respect, integrity and teamwork
RD2: Bachelor's degree and 5+ years of experience, Master's degree and 3+ years of experience, or PhD, or equivalent
RD3: Bachelor's degree and 8+ years of experience, Master's degree and 5+ years of experience, or PhD and 4+ years of experience, or equivalent
Preferred
Experience designing or operating distributed inference or data services, including request routing, asynchronous execution, queueing, fault tolerance, and performance monitoring
Experience integrating services with HPC schedulers (e.g., Slurm, PBS), including resource provisioning, job lifecycle management, and balancing latency-sensitive and throughput-oriented workloads
Experience optimizing AI inference performance (e.g., batching, memory management, model parallelism, quantization, accelerator utilization) on GPU- or accelerator-based systems
Familiarity with secure, multi-user services, including authentication/authorization, API security, and operating within institutional or regulated environments
Experience with running simulations or AI workflows on supercomputers
Company
Argonne National Laboratory
Argonne National Laboratory conducts researches in basic science, energy resources, and environmental management.
H1B Sponsorship
Argonne National Laboratory has a track record of offering H1B sponsorships. Please note that this does not
guarantee sponsorship for this specific role. Below presents additional info for your
reference. (Data Powered by US Department of Labor)
Distribution of Different Job Fields Receiving Sponsorship
Represents job field similar to this job
Trends of Total Sponsorships
2022 (6)
2021 (2)
Funding
Current Stage
Late StageTotal Funding
$41.4MKey Investors
Advanced Research Projects Agency for HealthUS Department of EnergyU.S. Department of Homeland Security
2024-11-14Grant· $21.7M
2023-09-27Grant
2023-01-17Grant
Recent News
Inside HPC & AI News | High-Performance Computing & Artificial Intelligence
2026-01-09
2026-01-03
Inside HPC & AI News | High-Performance Computing & Artificial Intelligence
2025-12-24
Company data provided by crunchbase