about 7 hours ago
Base Salary
$320k - $485k/yr
Responsibilities
- Lead inference for new model architectures and ensure their deployment on cloud platforms.
- Collaborate with the core inference team to integrate new features into cloud services.
- Identify and resolve discrepancies in inference behavior across platforms.
- Design and manage CI/CD infrastructure for inference servers and load balancers.
- Enhance validation processes to improve production cycle times.
- Analyze performance data to identify and address bottlenecks and anomalies.
Requirements
- Significant software engineering experience in high-performance distributed systems.
- Experience with cloud platforms like AWS, GCP, or Azure.
- Proficiency in building automation or test infrastructure.
- Ability to collaborate effectively with cross-functional teams.
- Strong problem-solving skills and autonomy in managing projects.
- Interest in LLM serving, with a willingness to learn.
Benefits
- Competitive compensation and benefits package.
- Optional equity donation matching.
- Generous vacation and parental leave.
- Flexible working hours.
- Collaborative office environment.