About the role
As an applied machine learning engineer, you will work on adapting state of the art deep learning (DL) models to run on our wafer scale system. This includes both functional validation and performance tuning of a variety of core models for applications like Natural Language Processing (NLP), Large Language Models (LLMs), Computer Vision (CV) and Graph Neural Networks (GNN).
As a member of the Cerebras engineering team you will be implementing models in popular DL frameworks like PyTorch and using insights into our hardware architecture to unlock to full potential of our chip. You will work on all aspects of the DL model pipeline including:
- Dataloader implementation and performance optimization
- Reference model implementation and functional validation
- Model convergence and hyper-parameters tuning
- Model customization to meet customer needs.
- Model architecture pathfinding.
This role will allow you to work closely with partner companies at the forefront of their fields across many industries. You will get to see how deep learning is being applied to some of the world’s most difficult problems today and help ML researchers in these fields to innovate more rapidly and in ways that are not currently possible on other hardware systems.
Responsibilities
- Analyze, implement, and optimize DL models for the WSE
- Functional and convergence of models on the WSE
- Work with engineering teams to optimize models for the Cerebras stack
- Support engineering teams in functional and performance scoping new models and layers
- Work with customers to optimize their models for the Cerebras stack.
- Develop new approaches for solving real world AI problems on various domains.
Requirements
- Bachelor's degree in engineering, science, or related field with 8+ years of experience
- Experience programming in modern language like Python or C++
- In-depth understanding of DL learning methods and model architectures
- Experience with DL frameworks like PyTorch, TensorFlow and JAX
- Familiar with state-of-the-art transformer architectures for language and vision model.
- Experience in model training and hyper-parameter tuning techniques.
- Familiar with different LLM downstream tasks and datasets.
Preferred Skills
- A deep passion for cutting edge artificial intelligence techniques
- Master's or PhD in engineering, science or related field
- Understanding of hardware architecture
- Experience programming accelerators like GPUs and FPGAs
Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Cerebras Systems has pioneered a groundbreaking chip and system that revolutionizes deep learning applications. Our system empowers ML researchers to achieve unprecedented speeds in training and inference workloads, propelling AI innovation to new horizons.
The Condor Galaxy 1 (CG-1), unveiled in a recent announcement, stands as a testament to Cerebras' commitment to pushing the boundaries of AI computing. With a staggering 4 ExaFLOP processing power, 54 million cores, and 64-node architecture, the CG-1 is the first of nine powerful supercomputers to be built and operated through an exclusive partnership between Cerebras and G42. This strategic collaboration aims to redefine the possibilities of AI by creating a network of interconnected supercomputers that will collectively deliver a mind-boggling 36 ExaFLOPS of AI compute power upon completion in 2024.
Cerebras is building a team of exceptional people to work together on big problems. Join us!