In a move that could accelerate the enterprise adoption of generative AI, startup Galileo has unveiled a novel suite of AI models specifically designed for evaluating the outputs of large language models like GPT-3.
The new offering, dubbed Galileo Luna, represents a first-of-its-kind approach to GenAI evaluation using what the company calls Evaluation Foundation Models (EFMs). These specialized models are finely tuned for tasks like detecting hallucinations, toxic language, data leaks, and malicious prompts in the responses from AI systems.
"For gen AI to achieve mass adoption, it's crucial that enterprises can evaluate hundreds of thousands of AI responses for hallucinations, toxicity, security risk, and more, in real time," said Vikram Chatterji, Co-Founder and CEO of Galile
Galileo Releases ‘Luna’ to Light Up Enterprise Gen AI Evaluation
- By Abhijeet Adhikari
- Published on
The key innovation in Luna is the use of right-sized, purpose-built EFMs rather than massively over parameterized models.
