Galileo Evaluate® : Rapid Evaluation of Prompts, Chains and RAG systems

Galileo Evaluate is a powerful bench for rapid, collaborative experimentation across LLMs, Prompts, RAG parameters, Vector Context, Retrievers, Chains/Agents, and more. Over time, Galileo becomes your team's centralized Prompt Store with automatic version control.

Core features

  • Prompt Templating and Versioning - One place to create, manage, and track all versions of your templates

  • Evaluation - Evaluate your prompts and mitigate your hallucinations using Galileo's Guardrail Metrics.

  • Experiment tracking - compare different runs and choose the best as per metrics

Getting Started

pageQuickstart

Last updated