about 18 hours ago
Responsibilities
- Design, build, and operate a high-quality data lake and supporting software systems.
- Write production-grade software for services, APIs, tooling, and automation.
- Innovate solutions using Trino and Starburst for complex data management challenges.
- Collaborate with technical leads and product managers to develop data products.
- Leverage AI to enhance dataset access for users across Starburst.
- Enable dataset preparation and model evaluation for AI projects.
- Define and evolve engineering processes and best practices.
- Iterate on data architecture and software systems with a focus on quality.
- Identify emerging patterns in data management and software engineering.
Requirements
- At least 7 years of experience in software and/or data engineering.
- Strong fundamentals in software engineering with proficiency in Java, Python, or Scala.
- Experience building and optimizing data pipelines using Trino, Spark, and dbt.
- Experience designing and building backend services and APIs.
- Familiarity with managing data infrastructure in public clouds, especially AWS.
- Experience with orchestration frameworks like Apache Airflow or Dagster.
- Knowledge of AI application design patterns.
- Fluency in SQL and ability to switch between SQL and programming languages.
- Experience with API integrations for data extraction.
- Knowledge of modern data lake modeling techniques.
- Proficiency with Infrastructure-as-Code tools like Terraform or Ansible.
- Strong communication skills and ability to coordinate across teams.
- Willingness to travel 25% for various company events.
Benefits
- Competitive pay and attractive stock grants.
- Flexible paid time off.
- Supportive and inclusive work environment.
