-
Notifications
You must be signed in to change notification settings - Fork 246
Open
Description
'maxlen' is one of the parameters in 'train_nmt.py', set to 50 by default.
I get the following message during the training process: "Minibatch with zero sample under length 100"
Investigating the source code shows that this message is appear when there is a batch size that the length of the source and target is greater than 'maxlen'.
On the other hand, in 'data_iterator.py' training samples have been skipped when the length of source and target is greater than 'maxlen'.
- Why such a contradiction is exist? -passing samples in data_iterator and then filter them in 'prpare-data'
- If I set maxlen to a large value (1000 for example), the updating time is significantly increase, would you describe why?
Metadata
Metadata
Assignees
Labels
No labels