ML Research Engineer (Inference)
Cerebras Systemsabout 4 hours ago
Bengaluru, India
Entry Level / Mid Level
H1B Sponsor
Responsibilities
- Implement and adapt transformer-based models (NLP and/or vision) to run on Cerebras hardware.
- Assist in optimizing models for inference performance (latency, throughput).
- Run experiments, analyze results, and support model improvements.
- Help bring up and validate models on the Cerebras system.
- Debug and troubleshoot model or system issues with guidance from senior team members.
- Support profiling and performance analysis using internal tools.
- Collaborate with cross-functional teams (ML, software, hardware) on model integration.
Requirements
- Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
- 1–3 years of experience in software engineering or machine learning in a similar capacity.
- Experience with Python and at least one ML framework (e.g., PyTorch, Transformers).
- Understanding of deep learning concepts (e.g., neural networks, transformers).
- Experience with Generative AI and Machine Learning systems.
- Strong programming skills in Python and/or C++.
Benefits
- Opportunity to build a breakthrough AI platform beyond the constraints of the GPU.
- Ability to publish and open source cutting-edge AI research.
- Work on one of the fastest AI supercomputers in the world.
- Enjoy job stability with startup vitality.
- Experience a simple, non-corporate work culture that respects individual beliefs.
Tech Stack
C++LinuxPythonPyTorch
Categories
AI & MLData Science