Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Evaluations in backend #1137

Merged
merged 489 commits into from
Jan 17, 2024
Merged

Evaluations in backend #1137

merged 489 commits into from
Jan 17, 2024

Conversation

mmabrouk
Copy link
Member

@mmabrouk mmabrouk commented Jan 2, 2024

No description provided.

MohammedMaaz and others added 25 commits January 2, 2024 16:25
[Enhancement]: Integration Tests for Evaluation
Copy link
Member Author

@mmabrouk mmabrouk left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@MohammedMaaz @aakrem @aybruhm

Comparison view:

  • Currently, each input/row is displayed twice. We need to map the evaluation scenarios from the multiple runs. This should ideally be done in the backend. Here is a simple Python algorithm for how this could be implemented:
from collections import defaultdict

list1 = [('a', 1), ('b', 2)]
list2 = [('a', 10), ('b', 2)]
list3 = [('a', 3), ('b', 2), ('c', 5)]

merged_dict = defaultdict(list)
for lst in [list1, list2, list3]:
    for k, v in lst:
        merged_dict[k].append(v)
  • Input columns are displayed twice. They should be shown only once.
  • The order of the columns should be: Input/Expected output/Output var1/Output var2/Results var1/Results var2.

Evaluation view:

  • Remove the id column.
  • There's no option to shut down failed evaluations. We need a method to stop them.
  • There's no way to determine if an evaluation has failed or if an evaluation scenario has failed.
  • There's no option to rerun an evaluation.

Evaluators view:

  • Exact match should be available from the start. This should probably be handled in the backend. This also applies to all evaluators that don't need configuration.

mmabrouk and others added 28 commits January 16, 2024 18:18
@aakrem aakrem merged commit 0ba7342 into main Jan 17, 2024
5 of 8 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants