PyCon India 2025

Ragas: Supercharge Your LLM Application Evaluations
2025-09-15 , Track 1

Ragas is a comprehensive evaluation toolkit designed to supercharge Large Language Model (LLM) application evaluations. Ragas provides objective metrics, intelligent test generation capabilities, and data-driven insights to help developers move away from time-consuming, subjective assessments toward efficient, quantifiable evaluation workflows.


Setup

  1. Environment Setup
git clone https://github.com/explodinggradients/ragas.git
cd ragas

Install uv (recommended package manager)

curl -LsSf https://astral.sh/uv/install.sh | sh
  1. Installation
make install
  1. Run tests to ensure everything works
make test
  1. Explore Available Commands
make help

Important Links

Development Workflow

  • Daily Development: make format, make type, make test
  • Before PR Submission: make run-ci
  • Code Quality: Automated formatting with ruff, type checking with pyright make check
  • Testing: Comprehensive unit tests, e2e tests, and benchmarks make test

Key Development Areas to focus on:

  • Core Metrics: Evaluation metrics for LLM applications
  • Test Generation: Automated test data creation
  • Integrations: Framework connectors and observability tools
  • Documentation: Tutorials, guides, and examples

Prerequisites
  • Python 3.9+ installed on your system
  • Git for version control
  • Basic Python knowledge and familiarity with virtual environments
  • LLM/ML concepts understanding (helpful but not required)
  • Optional: Experience with evaluation frameworks, testing, or LLM applications
Additional Resources

https://github.com/explodinggradients/ragas

Target Audience

Intermediate

maintainer from Ragas