
Inference Technical Lead, Sora

Inference Technical Lead, Sora
OpenAI
The Inference Technical Lead at OpenAI's Sora team will focus on optimizing model serving efficiency and inference performance for multimodal AI products. This hybrid research and product role requires expertise in model performance optimization and kernel-level systems, contributing to the development of reliable AI systems that benefit society.
Qualification
- Deep expertise in model performance optimization, particularly at the inference layer
- Strong background in kernel-level systems, data movement, and low-level performance tuning
- Excitement about scaling high-performing AI systems for real-world, multimodal workloads
- Ability to navigate ambiguity and set technical direction
- Experience driving complex initiatives to completion
Responsibility
- Perform engineering efforts focused on improving model serving, inference performance, and system efficiency
- Drive optimizations from a kernel and data movement perspective to improve system throughput and reliability
- Partner closely with research and product teams to ensure models perform effectively at scale
- Design, build, and improve critical serving infrastructure to support Sora’s growth and reliability needs
- Assist researchers in developing inference-friendly models




