Galileo Evaluate® : Rapid Evaluation of Prompts, Chains and RAG systems
Galileo Evaluate is a powerful bench for rapid, collaborative experimentation across LLMs, Prompts, RAG parameters, Vector Context, Retrievers, Chains/Agents, and more. Over time, Galileo becomes your team's centralized Prompt Store with automatic version control.
Core features
Prompt Templating and Versioning - One place to create, manage, and track all versions of your templates
Evaluation - Evaluate your prompts and mitigate your hallucinations using Galileo's Guardrail Metrics.
Experiment tracking - compare different runs and choose the best as per metrics
Getting Started
pageQuickstartLast updated