Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Evaluation data format #19

Open
Matthieu-Tinycoaching opened this issue Aug 30, 2022 · 1 comment
Open

Evaluation data format #19

Matthieu-Tinycoaching opened this issue Aug 30, 2022 · 1 comment

Comments

@Matthieu-Tinycoaching
Copy link

Hi,

1/ How should the evaluation data format be as passed in the evaluation_data argument? Could you provide me some example of evaluation data and how it should be formatted?

2/ How does the evaluation work on these data? What are the tests passed and labels used?

Thanks!

@adrienohana
Copy link

Haven't figured out how it works yet, I tried to feed some evaluation data with a folder containing :
corpus.jsonl, queries.jsonl and qrels/train.tsv . That doesn't work... nothing happens.

Would be nice have some training metrics that show what's happening. Plotting the loss maybe ? Or evaluating the data every hundred steps... Seems that my metrics keep improving way after 100k steps (beir metrics are NDCG, MAP, Precision and Recall @ K)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants