-
Notifications
You must be signed in to change notification settings - Fork 194
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Friendly hello - pingback #14
Comments
WOW! I never would have thought to try a How many epochs did it take to get those results? |
This is very promising. I plugged this rough architecture into EuroParl with 8 punctuation categories and am getting 83% precision, 48% recall after just 5 epochs. It is training 50x faster than @ottokart's model -- which gives me freedom to experiment and get results much faster. I'll post the results after some experimenting. |
Hi! that's definitely very interesting, thanks for sharing! 50x speedup is really impressive. Best, |
Hi Ottokar, I am using the TensorFlow Estimator API since it is very efficient and designed to take models into large-scale distributed production. It works well with large volumes of static data. Training went perfectly. But I am running into an issue since predicting punctuation needs to be done one line at a time since we are using EOS tokens from the previous results to partition. I need to figure out how to create an I've done some manual testing and results are very promising...
|
This looks very nice indeed. |
I didn't measure f-scores without no-punct. I am very sceptical regarding the precision of the app. BTW, if you like the project you can link back to it in your readme. |
I got it working (after hacking the TensorFlow prediction function)!!!
It doesn't reach the full accuracy of your model, but still quite impressive for training 40-50x faster. It peaked at about 17 epochs (a few hours on my CPU) but I didn't have early stopping on and lost the checkpoint (peak precision was a few percent higher). There's also definite room for improvement by tuning hyper-parameters. |
@vackosar There's and alternatives section in the readme now that refers to your work as well. |
The problem is that the precision/recall calculation in keras-punctuator is simply wrong. The real numbers are much lower and thats why you get impression it trains faster. The overall approach with CNN is about the same as CNN-2A from X. Che, C. Wang, H. Yang, and C. Meinel, “Punctuation predic- tion for unsegmented transcript based on word vector" paper also referenced in punctuator2 paper and the numbers match the results in the punctuator2 paper table 2 (about 54% overall f-score instead of 64% for punctuator2) |
I would definitely like to fix the precision/recall calculation. I agree that there is obviously something wrong with it. |
@nshmyrev interesting article about RNN vs Feed Forward http://www.offconvex.org/2018/07/27/approximating-recurrent/ |
@vackosar this article has multiple factual mistakes and misinterpretations. For example take:
Wavenet is a different codec operated on sample-by-sample basis thats why it got more quality than previous vocoder-based architectures, they do not compare apples to apples here.
Actually perplexity difference is significant. 46 vs 43 is a meaningful difference in many applications. And so forth. |
Hello!
I was inspired by your project and created simplified alternative. I cited you in the readme: https://github.com/vackosar/keras-punctuator
Let me know if you find this interesting.
Vaclav
The text was updated successfully, but these errors were encountered: