Job Description

Overview

Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This enables industry-leading training and inference speeds and allows machine learning users to run large-scale ML applications without managing hundreds of GPUs or TPUs.

Cerebras' current customers include top model labs, global enterprises, and AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras to deploy scale and high-speed inference.

Thanks to the wafer-scale architecture, Cerebras Inference offers fast Generative AI inference, significantly faster than GPU-based hyperscale cloud inference services. This accelerates the user experience of AI applications and enables real-time iteration and deeper model intelligence.

About The Role

We are building ...

Apply for this Position

Ready to join Cerebras? Click the button below to submit your application.

Submit Application