Update existing experiment
Sometimes you may want to update an existing experiment with new scores, or update existing scores for an experiment. You can do this using the evaluate_experiment
function.
This function will re-run the scoring metrics on the existing experiment items and update the scores:
The evaluate_experiment
function can be used to update existing scores for an experiment. If you use a scoring
metric with the same name as an existing score, the scores will be updated with the new values.
Example
Create an experiment
Suppose you are building a chatbot and want to compute the hallucination scores for a set of example conversations. For this you would create a first experiment with the evaluate
function:
evaluate
function in our LLM evaluation guide.Update the experiment
Once the first experiment is created, you realise that you also want to compute a moderation score for each example. You could re-run the experiment with new scoring metrics but this means re-running the output. Instead, you can simply update the experiment with the new scoring metrics: