Athina AI¶
Ragas Metrics on your Production Logs¶
Athina is a production monitoring and evaluation platform. Try the sandbox here.
You can use Athina with Ragas metrics to run evals on production logs, and get granular model performance metrics on your production data.
For example, you can get insights like this visually:
What is my
AnswerRelevancy
score for queries related torefunds
for customer idnike-usa
What is my
Faithfulness
score forproduct catalog
queries using promptcatalog_answerer/v3
with modelgpt-3.5-turbo
▷ Running Athina Programmatically¶
When you use Athina to run Ragas evals programmatically, you will be able to view the results on Athina’s UI like this 👇
Install Athina’s Python SDK:
pip install athina
Create an account at app.athina.ai. After signing up, you will receive an API key.
Here’s a sample notebook you can follow: https://github.com/athina-ai/athina-evals/blob/main/examples/ragas.ipynb
Run the code
import os
from athina.evals import (
RagasAnswerCorrectness,
RagasAnswerRelevancy,
RagasContextRelevancy,
RagasFaithfulness,
)
from athina.loaders import RagasLoader
from athina.keys import AthinaApiKey, OpenAiApiKey
from athina.runner.run import EvalRunner
import pandas as pd
# Set your API keys
OpenAiApiKey.set_key(os.getenv("OPENAI_API_KEY"))
AthinaApiKey.set_key(os.getenv("ATHINA_API_KEY"))
# Load your dataset from a dictionary, json, or csv: https://docs.athina.ai/evals/loading_data
dataset = RagasLoader().load_json("raw_data.json")
# Configure the eval suite
eval_model = "gpt-3.5-turbo"
eval_suite = [
RagasAnswerCorrectness(),
RagasFaithfulness(),
RagasContextRelevancy(),
RagasAnswerRelevancy(),
]
# Run the evaluation suite
batch_eval_result = EvalRunner.run_suite(
evals=eval_suite,
data=dataset,
max_parallel_evals=1, # If you increase this, you may run into rate limits
)
pd.DataFrame(batch_eval_result)
▷ Configure Ragas to run automatically on your production logs¶
If you are logging your production inferences to Athina, you can configure Ragas metrics to run automatically against your production logs.
Navigate to the Athina Dashboard
Open the Evals page (lightning icon on the left)
Click the “New Eval” button on the top right
Select the Ragas tab
Select the eval you want to configure
Learn more about Athina¶
Website: https://athina.ai
Docs: https://docs.athina.ai
Github Library: https://github.com/athina-ai/athina-evals
Sandbox: https://demo.athina.ai