Advanced Technology: Compiler Engineer
full-time
mid
Posted 2 weeks ago
About this role
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras , to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
About The Team
Cerebras builds wafer-scale AI processors—single chips delivering tens of PB/s of memory bandwidth and a dataflow architecture that accelerates at a granularity no multi-device system can match. The Advanced Technology Group (ATG) is Cerebras ’ pathfinding organization. We work ahead of product to explore new architectures, demonstrate breakthrough performance on scientific and AI workloads, and shape the technical roadmap for future Cerebras hardware and software. Our work regularly appears at top-tier venues (Supercomputing, SIAM, IEEE, and NeurIPS ) and directly influences the design of next-generation wafer-scale systems.
About The Role
We are seeking Compiler Engineers to join a small team of specialists working on our emerging Tungsten language compiler. Tungsten is Cerebras’ dataflow programming language, purpose-built for wafer-scale hardware. You will work on the Tungsten compiler from language design through code generation, building the toolchain that translates high-level intent into efficient execution across hundreds of thousands of cores with a memory and interconnect model unlike anything in conventional computing.
This is not incremental work on an existing backend. The architecture is new, the programming model is new, and the compiler is where those two things meet. You will collaborate closely with Cerebras’ ASIC, kernel, and AI teams, and your design decisions will directly shape both the language and the hardware it targets. Beyond the compiler itself, the broader toolchain—runtime, debugger, simulator—is still being built, and we are equally interested in engineers who want to own those pieces of the developer experience on novel hardware.
What You Will Do
Design and implement compiler passes across the Tungsten toolchain: mid-end optimization, backend code generation, instruction scheduling, register allocation, assembler, and linker.
Co-design language constructs that improve expressiveness and performance for dataflow execution on wafer-scale hardware.
Develop and iterate on code generation strategies for complex scientific and AI workloads, analyzing performance bottlenecks and closing the gap between peak and achieved throughput.
Extend the compiler to support future hardware architectures as they move from design to silicon.
Work directly with ASIC architects and application researchers to inform hardware-software co-design decisions.
What We Are Looking For
PhD in Computer Science or Computer Engineering preferred ; exceptional candidates without a graduate degree who demonstrate equivalent depth through published research, significant open-source contributions, or a strong industry track record are encouraged to apply.
Substantial experience in compiler development: IR design, optimization passes, code generation, or backend implementation for novel or non-standard architectures.
Strong grasp of computer architecture: instruction sets, memory models, dataflow execution, and how hardware constraints shape compilation strategy.
Systems-level programming ability in C; comfort reasoning about performance at the instruction and memory-access level.
Ability to think about compilation as a design problem, not just an implementation task: you should have opinions about how language semantics, compiler IR, and hardware capabilities interact.
Excellent communication and interpersonal skills: able to work effectively in a small, fast-moving team where compiler, architecture, and application concerns are deeply intertwined.
Valuable Assets
Experience with compilers for spatial, dataflow, or CGRA architectures where the compilation model diverges sig
Similar Jobs
Related searches:
On-site Jobs
Mid-Level Jobs
On-site Mid-Level Jobs
Mid-Level NLP & Language AIMid-Level Backend & SystemsMid-Level AI Agents & RAGMid-Level AI InfrastructureMid-Level Generative AI
AI Jobs in Sunnyvale
NLP & Language AI in SunnyvaleBackend & Systems in SunnyvaleAI Agents & RAG in SunnyvaleAI Infrastructure in SunnyvaleGenerative AI in Sunnyvale
distributed-systemscode-generationagentsgenerative-ai