Compiler Engineer - Distributed ML Training @ Gensyn | Jobright.ai
JOBSarrow
RecommendedLiked
0
Applied
0
External
0
Compiler Engineer - Distributed ML Training jobs in San Francisco Bay Area
Be an early applicantLess than 25 applicants
company-logo

Gensyn · 3 hours ago

Compiler Engineer - Distributed ML Training

ftfMaximize your interview chances
BlockchainCryptocurrency

Insider Connection @Gensyn

Discover valuable connections within the company who might provide insights and potential referrals.
Get 3x more responses when you reach out via email instead of LinkedIn.

Responsibilities

Lowering deep learning graphs - from common frameworks (PyTorch, Tensorflow, Keras, etc) down to an IR for training and inference - with particular focus on ensuring reproducibility
Writing novel algorithms - for transforming intermediate representations of compute graphs between different operator representations
Owning two of the following compiler areas:
Front-end - deal with the handshaking of common Deep Learning Frameworks with Gensyn's IR for internal IR usage. Write Transformation passes in ONNX to alter IR for middle-end consumption
Middle-end - write compiler passes for training-based compute graphs, integrate reproducible Deep Learning kernels into the code generation stage, and debug compilation passes and transformations as you go
Back-end : lower IR from middle-end to GPU target machine code

Qualification

Find out how your skills align with this job's requirements. If anything seems off, you can easily click on the tags to select or unselect skills to reflect your actual expertise.

Compiler knowledgeParallel programmingHigh-Level IR/Clang/LLVMLow Level IR/LLVMComputer architectureCompilation understandingDeep Learning understandingGPU kernels optimizationDeep Learning Compiler frameworksRust experienceOpen-source contributions

Required

Compiler knowledge - base-level understanding of a traditional compiler (LLVM, GCC) and graph traversals required for writing code for such a compiler
Solid software engineering skills - practicing software engineer, having significantly contributed to/shipped production code
Understanding of parallel programming - specifically as it pertains to GPUs
Ability to operate on: High-Level IR/Clang/LLVM up to middle-end optimisation; and/or Low Level IR/LLVM targets/target-specific optimisations - particularly GPU specific optimisations
Highly self-motivated with excellent verbal and written communication skills
Comfortable working in an applied research environment with extremely high autonomy

Preferred

Architecture understanding - full understanding of a computer architecture specialised for training NN graphs (Intel Xeon CPU, GPUs, TPUs, custom accelerators)
Compilation understanding - strong understanding of compilation in regards to one or more High-Performance Computer architectures (CPU, GPU, custom accelerator, or a heterogenous system of all such components)
Proven technical foundation - in CPU and GPU architectures, numeric libraries, and modular software design
Deep Learning understanding - both in terms of recent architecture trends + fundamentals of how training works, and experience with machine learning frameworks and their internals (e.g. PyTorch, TensorFlow, scikit-learn, etc.)
Exposure to a Deep Learning Compiler frameworks - e.g. TVM, MLIR, TensorComprehensions, Triton, JAX
Kernel Experience - Experience writing and optimizing highly-performant GPU kernels
Rust experience - systems level programming experience in Rust
Open-source contributions to existing compilers/frameworks with a strong preference for ML compilers/frameworks.

Benefits

Competitive salary + share of equity and token pool
Fully remote work - we hire between the West Coast (PT) and Central Europe (CET) time zones
Relocation Assistance - available for those that would like to relocate after being hired (anywhere from PST through CET time zones)
4x all expenses paid company retreats around the world, per year
Whatever equipment you need
Paid sick leave
Private health, vision, and dental insurance- including spouse/dependents [🇺🇸 only]

Company

Gensyn

twittertwittertwitter
company-logo
Gensyn is a machine learning compute protocol for the world's deep learning models.

Funding

Current Stage
Early Stage
Total Funding
$50.6M
Key Investors
a16z cryptoEden Block7percent Ventures
2023-06-12Series A· $43M
2022-03-21Seed· $6.5M
2021-01-01Pre Seed· $1.1M

Leadership Team

leader-logo
Ben Fielding
Co-founder
linkedin
leader-logo
Harry Grieve
Co-founder
linkedin
Company data provided by crunchbase
logo

Orion

Your AI Copilot