Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore: get prediction for eval dataset #414

Merged
merged 5 commits into from
Jul 15, 2024
Merged

Conversation

Yuan325
Copy link
Collaborator

@Yuan325 Yuan325 commented Jun 14, 2024

Add the function to get prediction for each of the queries from golden_dataset. Prediction is used as comparison to retrieve metrics.

Usage example:

from evaluation import run_llm_for_eval, goldens

# set up orchestration, session, set uuid
eval_list = await run_llm_for_eval(goldens, orchestration, session, session_id)

@Yuan325 Yuan325 requested a review from a team as a code owner June 14, 2024 23:44
@Yuan325 Yuan325 force-pushed the eval-implementation branch 3 times, most recently from 6c84c5d to c10073d Compare June 15, 2024 00:03
llm_demo/evaluation/evaluation.py Outdated Show resolved Hide resolved
llm_demo/evaluation/evaluation.py Outdated Show resolved Hide resolved
llm_demo/evaluation/evaluation.py Outdated Show resolved Hide resolved
@Yuan325 Yuan325 force-pushed the eval-implementation branch 3 times, most recently from 5f88f32 to 9a1e174 Compare June 24, 2024 22:14
@Yuan325 Yuan325 requested a review from kurtisvg June 24, 2024 22:15
Base automatically changed from eval-dataset to evaluation July 11, 2024 20:07
llm_demo/evaluation/eval_golden.py Outdated Show resolved Hide resolved
@Yuan325 Yuan325 merged commit cbbd98b into evaluation Jul 15, 2024
2 checks passed
@Yuan325 Yuan325 deleted the eval-implementation branch July 15, 2024 17:41
Yuan325 added a commit that referenced this pull request Jul 26, 2024
Add the function to get prediction for each of the queries from
golden_dataset. Prediction is used as comparison to retrieve metrics.

Usage example:

```
from evaluation import run_llm_for_eval, goldens

# set up orchestration, session, set uuid
eval_list = await run_llm_for_eval(goldens, orchestration, session, session_id)
```
Yuan325 added a commit that referenced this pull request Aug 8, 2024
Add the function to get prediction for each of the queries from
golden_dataset. Prediction is used as comparison to retrieve metrics.

Usage example:

```
from evaluation import run_llm_for_eval, goldens

# set up orchestration, session, set uuid
eval_list = await run_llm_for_eval(goldens, orchestration, session, session_id)
```
Yuan325 added a commit that referenced this pull request Aug 19, 2024
Add the function to get prediction for each of the queries from
golden_dataset. Prediction is used as comparison to retrieve metrics.

Usage example:

```
from evaluation import run_llm_for_eval, goldens

# set up orchestration, session, set uuid
eval_list = await run_llm_for_eval(goldens, orchestration, session, session_id)
```
Yuan325 added a commit that referenced this pull request Aug 19, 2024
Add the function to get prediction for each of the queries from
golden_dataset. Prediction is used as comparison to retrieve metrics.

Usage example:

```
from evaluation import run_llm_for_eval, goldens

# set up orchestration, session, set uuid
eval_list = await run_llm_for_eval(goldens, orchestration, session, session_id)
```
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants