Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Modified evaluation to use seqeval package #14

Merged
merged 3 commits into from
Jun 14, 2022
Merged

Conversation

jgmf-amazon
Copy link
Contributor

Issue 13

This PR is to fix a bug found in issue 13:

in the case, eg. label: what is the weather [datetime: today] prediction: what [datetime: is the weather today] treated as a correct prediction.

The problem is that only tokens in the predicted utterance aligned with slotted tokens in the ground truth were used when performing evaluation. Additionally, the calculation of F1 was wrong. Exact match accuracy was also affected.

To fix the bug, the seqeval package is now used, which follows conlleval conventions by default. To use seqeval, BIO tagging was implemented. Exact match accuracy was also updated to use the BIO-tagged sequences. Some new test cases were added, as well. All tests pass.

The paper preprint and the eval.ai leaderboards will be fixed by 6/17.

Copy link

@massive-dev-amz massive-dev-amz left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@jgmf-amazon jgmf-amazon merged commit 0552cdd into main Jun 14, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants