Staff AI Research Scientist - Evaluation, Handshake AI
HandshakeFull Time
Expert & Leadership (9+ years)
Key technologies and capabilities for this role
Common questions about this position
The position is hybrid.
Responsibilities include creating new evaluation benchmarks, working on cross-functional teams for trustworthy evaluations, conducting research on LLM evaluation methods like training LLM judges, and building scalable tools for model performance analysis.
Ideal candidates enjoy building prototypes to test LLM boundaries, have extensive experience reviewing complex data and LLM outputs for quality, are obsessive about rigorous AI measurements that align with real capabilities, and possess strong software engineering skills.
Cohere has an open and inclusive culture, with a team of top researchers, engineers, and designers who obsess over their work, move fast for customers, and value diverse perspectives.
Full-time employees enjoy an open and inclusive culture and work environment, and work closely with a team on the cutting edge.
Provides NLP tools and LLMs via API
Cohere provides advanced Natural Language Processing (NLP) tools and Large Language Models (LLMs) through a user-friendly API. Their services cater to a wide range of clients, including businesses that want to improve their content generation, summarization, and search functions. Cohere's business model focuses on offering scalable and affordable generative AI tools, generating revenue by granting API access to pre-trained models that can handle tasks like text classification, sentiment analysis, and semantic search in multiple languages. The platform is customizable, enabling businesses to create smarter and faster solutions. With multilingual support, Cohere effectively addresses language barriers, making it suitable for international use.