Conversation
|
The documentation is not available anymore as the PR was closed or merged. |
datasets/imdb/README.md
Outdated
| metrics: | ||
| - type: accuracy | ||
| - type: f1 | ||
| name: f1_macro |
There was a problem hiding this comment.
Perhaps it's better to use a "pretty name" here (and the other F1 metrics)?
| name: f1_macro | |
| name: F1 macro |
|
The tests are failing due to the changed metadata: I think you can fix this by updating the @lhoestq we are working with an arbitrary set of tags for |
Merge branch 'autoeval' of https://github.com/nrajani/datasets into autoeval
|
On the question of validating the For me, the required fields include:
Here I'm using the spec defined in https://github.com/huggingface/autonlp-backend/issues/414 as a guide. WDYT @lhoestq ? |
|
Makes sense ! Currently the metadata type validator doesn't support subfields - let me open a PR to add it |
- support YAML keys with dashes - add train-eval-index validation
|
I ended up improving the metadata validation in this PR x) In particular:
Let me know if it sounds good to you ! I think we can improve |
|
Come on windows... I didn't do anything advanced... Anyway, will try to fix this when I get back home x) |
Hehe, thanks! |
|
Thanks, @lhoestq this is great! |
|
Did I just fix it for windows and now it fails on linux ? xD |
Looks like the Heisenberg uncertainty principle is at play here - you cannot simultaneously have unit tests passing in both Linux and Windows 😅 |
|
The worst is that the tests pass locally both on my windows and my linux x) |
|
Ok fixed it, the issue came from python 3.6 that doesn't return the right |
lhoestq
left a comment
There was a problem hiding this comment.
Alright thanks for adding the first Autoeval config ! :D
Woohoo! Thank you so much 🤗 |
|
This is cool! |
Added autoeval config to imdb as pilot