Skip to content

1.18.85

Compare
Choose a tag to compare
@fhieber fhieber released this 15 Mar 14:07
c1b1da8

[1.18.85]

Changed

  • Disabled dynamic batching for Translator.translate() by default due to increased memory usage. The default is to
    fill-up batches to Translator.max_batch_size.
    Dynamic batching can still be enabled if fill_up_batches is set to False.

Added

  • Added parameter to force training to stop after a given number of checkpoints. Useful when forced to share limited GPU resources.

[1.18.84]

Fixed

  • Fixed lexical constraints bugs that broke batching and caused large drop in BLEU.
    These were introduced with sampling (1.18.64).

[1.18.83]

Changed

  • The embedding size is automatically adjusted to the Transformer model size in case it is not specified on the command line.

[1.18.82]

Fixed

  • Fixed type conversion in metrics file reading introduced in 1.18.79.

[1.18.81]

Fixed

  • Making sure the training pickled training state contains the checkpoint decoder's BLEU score of the last checkpoint.

[1.18.80]

Fixed

  • Fixed a bug introduced in 1.18.77 where blank lines in the training data resulted in failure.

[1.18.79]

Added

  • Writing of the convergence/divergence status to the metrics file and guarding against numpy.histogram's errors for NaNs during divergent behaviour.