✨ About The Role
- The role involves designing and building the inference infrastructure that powers OpenAI's products.
- Ensuring the infrastructure can scale to the next order of magnitude is a key responsibility.
- The position includes participation in an on-call rotation to respond to critical incidents as needed.
- The candidate will help create a diverse, equitable, and inclusive culture within the team.
- The job is based in San Francisco, CA, with a hybrid work model of 3 days in the office per week.
âš¡ Requirements
- The ideal candidate will have over 10 years of experience in building core infrastructure.
- Experience in running GPU clusters at scale is essential for success in this role.
- A strong background in operating orchestration systems such as Kubernetes at scale is required.
- The candidate should take pride in building and operating scalable, reliable, and secure systems.
- Comfort with ambiguity and rapid change is crucial for thriving in this dynamic environment.