-
Notifications
You must be signed in to change notification settings - Fork 96
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[WIP] Albert #203
[WIP] Albert #203
Conversation
@aviks |
added Sentencepiece unigram support |
completed trainable Albert structure. |
fine-tuning Training Tutorial (it's not supported GPU so far)- |
The above code is pretty messy and not yet refractor (for the experiment) |
53c5b88
to
e2172d2
Compare
f60ae8f
to
d214d47
Compare
808c8a8
to
3937533
Compare
Hi @tejasvaidhyadev can you move this PR to TextModels now please? |
Is it okay, if I will do it the coming weekend? |
Yes, of course, only whenever you have time. |
Hi everyone
I am adding ALBERT [WIP]
Currently only raw code is given in PR.
Dependencies - Transformers.jl , WordTokenizer.jl
I am not exporting any function.I am still in middle of deciding what is the best way to use it.
But i am adding some important codes which is used for conversion of pretrained checkpoints and in Demo file below
Roadmap
apart from above refactoring and cleaning of code is remaining
Important links
Pretrained weights link .
tfckpt2bsonforalbert.jl
For detail refer - link
Demo - link
PS
All the suggestions are welcome