Yuma AI (YC W23) · 1 month ago
Senior DevOps / Infrastructure & AI LLM Systems Engineer (Hybrid)
Yuma AI is building a next-generation orchestration platform that deploys autonomous AI agents dedicated to customer support in e-commerce. The Senior DevOps / Infrastructure & AI LLM Systems Engineer will take full ownership of cloud infrastructure, deployments, and reliability, while also contributing to AI systems and LLM deployments.
Artificial Intelligence (AI)Customer ServiceE-CommerceGenerative AISoftware
Responsibilities
All cloud infrastructure across AWS, GCP, and Azure
Kubernetes cluster management, scaling, upgrades, and security
CI/CD pipelines (GitHub Actions) and deployment systems
Observability, monitoring, logging, alerting, and reliability practices
Incident response, on-call rotation, and uptime improvements
Cost optimization and infra-level performance tuning
Security best practices, IAM, secrets, policies, and overall infra hygiene
High-scale PostgreSQL (large DB, indexes, performance tuning)
Redis and Sidekiq pipelines, queue scaling, job parallelization
API performance and throughput
Manage and optimize LLM deployments across cloud providers
Improve latency, reliability, and cost through routing and system architecture
Help build and maintain eval pipelines and A/B tests
Contribute directly at the app level (prompts, agents, routing)
Support or prototype self-hosted model experiments (optional but valuable)
Qualification
Required
8+ years of experience in DevOps / infrastructure roles, ideally in fast-paced SaaS or startup environments
Experience scaling production systems and understanding how systems behave under real load
Comfortable deep in Kubernetes or writing Ruby/Python for quick scripts, tools, or LLM evaluations
Experience with Kubernetes, Docker
Strong in at least 2 of AWS, Azure, GCP
Experience with GitHub Actions CI/CD
Experience with PostgreSQL, Redis, Sidekiq
Experience with LLM APIs (OpenAI, Azure, Anthropic; self-hosted a plus)
Experience with Terraform or similar IaC
Strong coding ability to contribute across the stack
Preferred
Experience managing and optimizing LLM deployments across cloud providers
Experience improving latency, reliability, and cost through routing and system architecture
Experience building and maintaining eval pipelines and A/B tests
Experience contributing directly at the app level (prompts, agents, routing)
Experience supporting or prototyping self-hosted model experiments
Benefits
Competitive compensation based on experience and stock options
Flexibility : fully remote in Europe with preference for Barcelona office (Boston office is also an option)
Company
Yuma AI (YC W23)
Yuma is the most advanced AI platform built for e-commerce.
Funding
Current Stage
Early StageTotal Funding
$9.3MKey Investors
Y Combinator
2024-10-30Seed· $5M
2023-12-21Seed
2023-04-05Pre Seed· $0.5M
Recent News
2024-11-27
Pioneer Fund
2023-11-20
2023-04-12
Company data provided by crunchbase