Job Description

Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach enables Cerebras to deliver industry-leading training and inference speeds and allows machine learning users to run large-scale ML applications with less hardware management.

Cerebras' customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras to deploy scale and transform key workloads with ultra-high-speed inference.

Thanks to the wafer-scale architecture, Cerebras Inference offers fast Generative AI inference, significantly faster than GPU-based hyperscale cloud inference services. This speed improvement enhances the user experience of AI applications, enabling real-time iteration and more capable computation.

About The Rol...

Apply for this Position

Ready to join Cerebras? Click the button below to submit your application.

Submit Application