Skip to content

Commit

Permalink
fix typos re: divergences and grond (#4405)
Browse files Browse the repository at this point in the history
  • Loading branch information
ciccolo-neon authored Aug 27, 2024
1 parent c940bf9 commit d839fc4
Show file tree
Hide file tree
Showing 3 changed files with 5 additions and 5 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ Your goal is to determine if the AI answer correctly matches, in substance, the
[END DATA]
Compare the AI answer to the human ground truth answer, if the AI correctly answers the question,
then the AI answer is "correct". If the AI answer is longer but contains the main idea of the
Human answer please answer "correct". If the AI answer divergences or does not contain the main
Human answer please answer "correct". If the AI answer diverges or does not contain the main
idea of the human answer, please answer "incorrect".
```

Expand Down
4 changes: 2 additions & 2 deletions packages/phoenix-evals/src/phoenix/evals/default_templates.py
Original file line number Diff line number Diff line change
Expand Up @@ -444,7 +444,7 @@
[END DATA]
Compare the AI answer to the human ground truth answer, if the AI correctly answers the question,
then the AI answer is "correct". If the AI answer is longer but contains the main idea of the
Human answer please answer "correct". If the AI answer divergences or does not contain the main
Human answer please answer "correct". If the AI answer diverges does not contain the main
idea of the human answer, please answer "incorrect".
"""

Expand All @@ -469,7 +469,7 @@
If the AI correctly answers the question as compared to the human answer, then
the AI answer LABEL is "correct". If the AI answer is longer but contains the
main idea of the Human answer please answer LABEL "correct". If the AI answer
divergences or does not contain the main idea of the human answer, please answer
diverges or does not contain the main idea of the human answer, please answer
LABEL "incorrect".
Example response:
Expand Down
4 changes: 2 additions & 2 deletions tutorials/evals/evaluate_human_vs_ai_classifications.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
" <a href=\"https://join.slack.com/t/arize-ai/shared_invite/zt-1px8dcmlf-fmThhDFD_V_48oU7ALan4Q\">Community</a>\n",
" </p>\n",
"</center>\n",
"<h1 align=\"center\">Human/GrondTruth Versus AI Evals</h1>\n",
"<h1 align=\"center\">Human/GroundTruth Versus AI Evals</h1>\n",
"\n",
"Arize provides tooling to evaluate LLM applications, including tools to determine whether AI answers match Human Groundtruth answers. In many Q&A systems its important to test the AI answer results as compared to Human answers prior to deployment. These help assess how often the answers are correctly generated by the AI system. \n",
"\n",
Expand Down Expand Up @@ -362,7 +362,7 @@
" [END DATA]\n",
"Compare the AI answer to the human ground truth answer, if the AI correctly answers the question,\n",
"then the AI answer is \"correct\". If the AI answer is longer but contains the main idea of the\n",
"Human answer please answer \"correct\". If the AI answer divergences or does not contain the main\n",
"Human answer please answer \"correct\". If the AI answer diverges or does not contain the main\n",
"idea of the human answer, please answer \"incorrect\".\n",
"\n"
]
Expand Down

0 comments on commit d839fc4

Please sign in to comment.