-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Questions about the results #11
Comments
All the parameters are already set correctly as default and there is no need to tune. If you follow the evaluation instruction in README, the results should be exactly as follows (I just checked again by downloading the current code): Did you uncomment line 28 in evaluation.py to set the model to be evaluated as the pre-trained model ? |
Yes,i uncomment line 28 in evaluation.py to set the model to be evaluated as the pre-trained model,and i just change the import lib in model.py as follow: Shouldn't I be doing this?This is the result of my downloading and running again.
Miscellaneous 0.000 0.000 0.000 0 avg / total 0.827 0.664 0.734 1490 thanks a lot! |
|
thanks!i try to use Python 2.7, and get a similar result |
Hi @ruidan, |
I also tried to replicate the results and I got the same as @ruidan. It is a close result to the paper, but not exactly. Was it a different training set or parameters used in the paper? EDIT: I checked the default parameters in the code and they are pretty much the same as the papers. The paper mentions that the reported results are an average of 10 executions; therefore @ilivans, it might explain the different results. EDIT2: disregard my other questions, I found the answer in the paper. |
@ThiagoSousa thank you! Shame on me for not noticing that detail. |
Hello, I try to evaluate the uploaded trained restaurant model by running evaluation.py directly, did not make any changes to the code, but did not get the same results as the paper, is it necessary to tune some parameters? Thank you for your answer.
That's the result I got.
--- Results on restaurant domain ---
precision recall f1-score support
Miscellaneous 0.000 0.000 0.000 0
avg / total 0.527 0.381 0.442 1490
The text was updated successfully, but these errors were encountered: