Senior Software Engineer-Distributed Inference
NVIDIA- Full Time
- Senior (5 to 8 years)
Candidates should have at least 5 years of professional software engineering experience. A strong understanding of modern machine learning architectures and the ability to optimize their performance for inference is essential. Familiarity with PyTorch, NVidia GPUs, and software stacks that optimize them, as well as HPC technologies such as InfiniBand, MPI, and NVLink, is required. Experience in architecting, building, observing, and debugging production distributed systems is necessary, with bonus points for experience in performance-critical systems. Candidates should also be self-directed and willing to learn whatever is needed to solve problems effectively.
The Software Engineer will work alongside machine learning researchers, engineers, and product managers to bring the latest technologies into production. They will enable advanced research through engineering, introduce new techniques and tools to improve the performance and efficiency of the model inference stack, and build tools to identify and address bottlenecks. Additionally, they will optimize code and Azure VMs to maximize hardware utilization.
Develops safe and beneficial AI technologies
OpenAI develops and deploys artificial intelligence technologies aimed at benefiting humanity. The company creates advanced AI models capable of performing various tasks, such as automating processes and enhancing creativity. OpenAI's products, like Sora, allow users to generate videos from text descriptions, showcasing the versatility of its AI applications. Unlike many competitors, OpenAI operates under a capped profit model, which limits the profits it can make and ensures that excess earnings are redistributed to maximize the social benefits of AI. This commitment to safety and ethical considerations is central to its mission of ensuring that artificial general intelligence (AGI) serves all of humanity.