Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feat: implemented human evaluations #2047

Conversation

ashrafchowdury
Copy link
Collaborator

Descriptions

Implemented human evaluations on evaluation page reusing the SingleModelEvalOverview.tsx and AbTestingEvalOverview.tsx component from overview page by conditionally made them adjustable, so that both use in overview and evaluations page.

Copy link

vercel bot commented Aug 30, 2024

The latest updates on your projects. Learn more about Vercel for Git ↗︎

Name Status Preview Comments Updated (UTC)
agenta ✅ Ready (Inspect) Visit Preview 💬 Add feedback Sep 3, 2024 11:55am
agenta-documentation ✅ Ready (Inspect) Visit Preview 💬 Add feedback Sep 3, 2024 11:55am

@dosubot dosubot bot added size:L This PR changes 100-499 lines, ignoring generated files. Frontend labels Aug 30, 2024
@ashrafchowdury ashrafchowdury requested a review from bekossy August 30, 2024 16:29
Copy link
Member

@bekossy bekossy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the work on this PR. Here are some changes to address:

  1. Let's make it clear that both the SingleModelEvaluation and AbTestingEvaluation are reusable components and import that where needed.
  2. The "Compare" button in AB Testing should navigate to the evaluation compare view.
  3. We currently don’t have new changes in the Automatic Evaluation view?
  4. Please resolve the merge conflicts.

@dosubot dosubot bot added size:XL This PR changes 500-999 lines, ignoring generated files. and removed size:L This PR changes 100-499 lines, ignoring generated files. labels Sep 2, 2024
@ashrafchowdury
Copy link
Collaborator Author

Changes summary:

  • Made it clear that both the SingleModel and AbTesting are reusable components.
  • Removed the compare button from SingleModel and AbTesting after talking with Ahmed
  • Some UI changes were made in the 'Create Human Eval Model'
  • Merge conflict resolved

Copy link
Member

@bekossy bekossy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let’s also remove the previous Annotation code
http://localhost/apps/APP_ID/annotations/single_model_test should not be a valid path

And fix the failing cypress tests

@dosubot dosubot bot added size:XXL This PR changes 1000+ lines, ignoring generated files. and removed size:XL This PR changes 500-999 lines, ignoring generated files. labels Sep 3, 2024
@ashrafchowdury
Copy link
Collaborator Author

ashrafchowdury commented Sep 3, 2024

Changes summary:

  • Removed single_model_test.tsx and human_a_b_testing.tsx pages.
  • Removed AutomaticResultsEvaluation.tsx and HumanEvaluationResults.tsx components.
  • Adjusted cypress test according to new changes.
  • Minor UI update in the 'create new human eval model'.

@dosubot dosubot bot added the lgtm This PR has been approved by a maintainer label Sep 3, 2024
@bekossy bekossy merged commit 60d237e into AGE-587/-implement-evaluation-main-page Sep 3, 2024
5 of 6 checks passed
@bekossy bekossy deleted the feat/implement-human-evaluations branch September 3, 2024 12:00
@bekossy bekossy restored the feat/implement-human-evaluations branch September 3, 2024 12:12
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Frontend lgtm This PR has been approved by a maintainer size:XXL This PR changes 1000+ lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants