Prerequisites
Step 1: Your First Evaluation
Useevaluate_rubric() to evaluate text with natural language criteria:
With Ground Truth
Compare against a reference answer:Get Detailed Results
Include explanations with your scores:Step 2: Create a Reward Function
For deterministic, local evaluation, use the@osmosis_reward decorator:
Advanced Example: Numeric Tolerance
Step 3: Create a Rubric Evaluator
Use@osmosis_rubric for LLM-based evaluation functions:
Compare Multiple Providers
Evaluate with different LLM providers:Error Handling
Handle common errors gracefully:Next Steps
CLI Quick Start
Batch evaluate datasets with the CLI
API Reference
Complete API documentation
CLI Reference
Full CLI command reference
Decorators & API
Advanced patterns and usage