Machine Learning Engineer
SweedFull Time
Mid-level (3 to 4 years)
Candidates should possess a solid background in algorithms, data structures, and system design, with at least 5 years of relevant work experience. Experience working with modern machine learning tooling such as PyTorch and MLFlow is also required. Bonus points are awarded for experience in building and maintaining open-source projects, operating machine learning infrastructure in production, and building highly available serving systems.
The ML Developer Experience team is responsible for creating tools and services that enable users to build production-quality applications using Ray. This includes developing a next-generation ML Ops platform and development tooling centered around Ray, as well as building tools and frameworks for managing the AI development lifecycle from data preparation to training and production serving. Engineers will also contribute to critical infrastructure and architecture pieces that power the platform at scale, and may work on Anyscale workspaces, debugging, dependency management, ML Ops tooling, SDKs, and authentication.
Platform for scaling AI workloads
Anyscale provides a platform designed to scale and productionize artificial intelligence (AI) and machine learning (ML) workloads. Its main product, Ray, is an open-source framework that helps developers manage and scale AI applications across various fields, including Generative AI, Large Language Models (LLMs), and computer vision. Ray allows companies to enhance the performance, fault tolerance, and scalability of their AI systems, with some users reporting over 90% improvements in efficiency, latency, and cost-effectiveness. Anyscale primarily serves clients in the AI and ML sectors, including major companies like OpenAI and Ant Group, who rely on Ray for training large models. The company operates on a software-as-a-service (SaaS) model, charging clients a subscription fee for access to the Ray platform. Anyscale's goal is to empower organizations to effectively scale their AI workloads and optimize their operations.