Research Staff, LLMs
DeepgramFull Time
Expert & Leadership (9+ years)
Candidates should have a strong background in pushing the limits of LLMs and experience building high-quality evaluation resources such as datasets, simulators, or environments. A proven track record of developing new methods or data for LLM evaluation, demonstrated through publications at top-tier conferences or popular benchmarks, is required. Deep experience building with and around LLMs, coupled with strong software engineering skills for analyzing and understanding LLM performance, is also necessary.
The Senior Research Engineer, Model Evaluation will develop evaluation benchmarks, datasets, and environments to measure the cutting edge of model capabilities. This role involves conducting research to advance LLM evaluation methods, including training LLM judges, improving evaluation efficiency, and scalably building high-quality datasets. Additionally, the engineer will build scalable tools for investigating and understanding evaluation results, which will be utilized by technical staff, leadership, and the CEO.
Provides NLP tools and LLMs via API
Cohere provides advanced Natural Language Processing (NLP) tools and Large Language Models (LLMs) through a user-friendly API. Their services cater to a wide range of clients, including businesses that want to improve their content generation, summarization, and search functions. Cohere's business model focuses on offering scalable and affordable generative AI tools, generating revenue by granting API access to pre-trained models that can handle tasks like text classification, sentiment analysis, and semantic search in multiple languages. The platform is customizable, enabling businesses to create smarter and faster solutions. With multilingual support, Cohere effectively addresses language barriers, making it suitable for international use.