AI Engineer & Researcher, Inference
SpeechifyFull Time
Junior (1 to 2 years)
Santa Clara, California, United States
Key technologies and capabilities for this role
Common questions about this position
The salary range is $180K - $300K.
This is a hybrid position.
Candidates need 10+ years of experience in customer engineering and field support for enterprise-level AI and datacenter products focused on AI/ML software and generative AI inference, in-depth knowledge of generative AI inference at scale including production deployment, and experience with automation tools and scripting.
The culture emphasizes respect, collaboration, humility, direct communication, and inclusivity, with a team that values differing perspectives for better solutions.
Strong candidates have 10+ years in customer engineering for AI/datacenter products, hands-on expertise in scaling generative AI inference and production deployments, experience with automation scripting, and a passion for tackling challenges with a drive for execution.
AI compute platform for datacenters
d-Matrix focuses on improving the efficiency of AI computing for large datacenter customers. Its main product is the digital in-memory compute (DIMC) engine, which combines computing capabilities directly within programmable memory. This design helps reduce power consumption and enhances data processing speed while ensuring accuracy. d-Matrix differentiates itself from competitors by offering a modular and scalable approach, utilizing low-power chiplets that can be tailored for different applications. The company's goal is to provide high-performance, energy-efficient AI inference solutions to large-scale datacenter operators.