Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bert model training procedure #73

Open
camilomarino opened this issue May 31, 2022 · 1 comment
Open

Bert model training procedure #73

camilomarino opened this issue May 31, 2022 · 1 comment

Comments

@camilomarino
Copy link

Hey there,

I wanted to confirm some doubts that came to my mind when reviewing together the source code (https://github.com/nilmtk/nilmtk-contrib/blob/master/nilmtk_contrib/disaggregate/bert.py) and the paper (http://nilmworkshop.org/2020/proceedings/nilm20-final88.pdf):

  1. The loss function: I think I understand that the loss implemented in the code is the MSE, but in the paper they propose more terms besides MSE.
  2. The masking of the training data: I have not found that the masking of the input sequence is done as proposed in the paper.

I wanted to confirm if these two differences occur or I have misinterpreted the source code and/or the paper.

Best.

@xuuurq
Copy link

xuuurq commented Aug 12, 2022

@camilomarino Hello, sorry to bother you.I would like to ask if you have solved this problem.The bert model in nilmtk-contrib is different from the code in the BERT4NILM paper, which is reflected in the loss function and mask processing. Is the bert model in nilmtk-contrib without mask processing?
thank you very much.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants