AI Researcher (Multimodal Perception Models)
TavusFull Time
Mid-level (3 to 4 years)
Key technologies and capabilities for this role
Common questions about this position
The role is based in Tokyo, where you will join the team focused on multimodal companion agents.
Key skills include expertise in large language models (LLMs) particularly in the multimodal domain (vision, audio, text), LLM post-training and evaluation, and the ability to implement and optimize multimodal research concepts.
You will collaborate with a world-class, cross-functional team of researchers and software engineers in a dynamic and collaborative environment, working on cutting-edge research with a focus on safety, ethics, and public benefit.
The role focuses on developing state-of-the-art multimodal companion agents powered by Gemini, including implementation and optimization of research concepts across domains like education, health, and gaming.
This information is not specified in the job description.
Develops artificial general intelligence systems
This company leads in the field of artificial general intelligence (AGI), with notable applications across healthcare, energy management, and biotechnology. Their work in early diagnostic tools for eye diseases, optimizing energy usage in major data centers, and groundbreaking contributions to protein structure prediction underlines their commitment to harnessing AI for diverse practical applications. The company's dedication to pushing the boundaries of AI technology not only propels the industry forward but also creates a dynamic and impactful working environment for its employees.