Senior Software Engineer, AI Inference Platform jobs in United States
cer-icon
Apply on Employer Site
company-logo

Cerebras · 3 months ago

Senior Software Engineer, AI Inference Platform

Cerebras Systems builds the world's largest AI chip and is seeking a talented Platform Software Engineer to join the team building the Cerebras Inference Platform. In this role, you will design, develop, and operate core backend services and APIs that enable customers to deploy and manage inference workloads on dedicated Cerebras hardware.

AI InfrastructureArtificial Intelligence (AI)ComputerHardwareSemiconductorSoftware
check
Growth Opportunities
check
H1B Sponsor Likelynote

Responsibilities

Design, build, and maintain the core APIs for the Inference Platform, handling model catalog management, deployment of ML workloads, scaling, and status monitoring
Focus on building platform capabilities that optimize for ease-of-use, robustness, and self-service access to inference models and serving
Collaborate with infrastructure and ML engineering teams to ensure high reliability, uptime, and smooth user interactions with the inference service
Design and implement features like multi-tenant support, deployment automation, priority queuing, and caching strategies for user requests
Build robust observability features by integrating with monitoring and telemetry tools (e.g., Prometheus, Grafana) to track system health, performance metrics, and request analytics

Qualification

Backend software developmentPythonRESTful APIsDistributed systemsContainerization (Docker)Orchestration (Kubernetes)Databases (Postgres)Caching systems (Redis)TelemetryObservabilityProblem-solvingCommunication skillsCollaboration skills

Required

Bachelor's or Master's degree in computer science or related field, or equivalent practical experience
5+ years of experience in backend software development, with a focus on service APIs, orchestration platforms, or user-facing infrastructure
Strong proficiency in Python (C++ is good to have)
Experience designing, building, and integrating with RESTful APIs and gRPC services
Solid understanding of distributed systems concepts such as concurrency, scalability, and fault tolerance
Hands-on experience with containerization (Docker) and orchestration frameworks (Kubernetes)
Experience with databases and caching systems (e.g., Postgres, Redis)
Experience with observability, telemetry pipelines, and system monitoring best practices
Strong problem-solving and debugging abilities
Excellent communication and cross-functional collaboration skills

Company

Cerebras

twittertwittertwitter
company-logo
Cerebras Systems is the world's fastest AI inference. We are powering the future of generative AI.

H1B Sponsorship

Cerebras has a track record of offering H1B sponsorships. Please note that this does not guarantee sponsorship for this specific role. Below presents additional info for your reference. (Data Powered by US Department of Labor)
Distribution of Different Job Fields Receiving Sponsorship
Represents job field similar to this job
Trends of Total Sponsorships
2025 (31)
2024 (16)
2023 (18)
2022 (17)
2021 (34)
2020 (23)

Funding

Current Stage
Late Stage
Total Funding
$1.82B
Key Investors
Alpha Wave VenturesVy CapitalCoatue
2025-12-03Secondary Market
2025-09-30Series G· $1.1B
2024-09-27Series Unknown

Leadership Team

leader-logo
Andrew Feldman
CEO & Founder
linkedin
leader-logo
Bob Komin
Chief Financial Officer
linkedin
Company data provided by crunchbase