Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Exception: batching error - examples do not produce identical results under batching #11

Open
linamy85 opened this issue Jan 5, 2018 · 1 comment

Comments

@linamy85
Copy link

linamy85 commented Jan 5, 2018

Hello,
Thank you for releasing such an amazing work. I've tried to run the process by following steps.

  1. Clone this repo
  2. mkdir -p $DATA_DIR and uncompressed Glove vectors into $DATA_DIR/word_vectors
  3. Download glove.6B.300d_onebil.txt and glove.6B.300d_yelp.txt from https://worksheets.codalab.org/bundles/0x89bc0497bbb14ee489d33e032fa43a2e/
  4. Download onebillion_split dataset from https://worksheets.codalab.org/bundles/0x017b7af92956458abc7f4169830a6537/ and put them in $DATA_DIR/onebillion_split

However, I received an error while executing python textmorph/edit_model/main.py configs/edit_model/edit_onebil.txt --gpu 0 within docker

individually:
Variable containing:
205.0143
[torch.cuda.FloatTensor of size 1 (GPU 0)]

batched:
Variable containing:
205.0143
[torch.cuda.FloatTensor of size 1 (GPU 0)]

Traceback (most recent call last):
File "textmorph/edit_model/main.py", line 40, in
exp.train()
File "/code/textmorph/edit_model/training_run.py", line 265, in train
self._train(self.config, self._train_state, self._examples, self.workspace, self.metadata, self.tb_logger)
File "/code/textmorph/edit_model/training_run.py", line 399, in _train
editor.test_batch(noiser(train_batches[0]))
File "/code/textmorph/edit_model/editor.py", line 128, in test_batch
raise Exception('batching error - examples do not produce identical results under batching')
Exception: batching error - examples do not produce identical results under batching

After some checks, I found the individually is 205.01431 while batched is 205.01433, what should I do with this?

Appreciate any help! Thank you 😄

@jdxyw
Copy link

jdxyw commented Mar 2, 2018

have the same issue

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants