Releases: hfxunlp/transformer
v0.2.8
In this release, we:
support relative postion in self-attention;
add cnfg/hyp.py
and support cached data compression;
support swish activation function and several optimizers (optm/
);
implement context-aware Transformer proposed in Improving the Transformer Translation Model with Document-Level Context;
replace serialization backend from torch to utils.h5serial
based on h5py.
v0.2.7
v0.2.6
In this release, we:
clean the code by reduce common lines into utils
;
fix bugs (transformer.SC
and inconsistent order of ranking.);
add an argument which can disable the second linear bias of PositionwiseFFN
;
add parallel.parallelMTFP
which can accelerate the fine-tuning when part of parameters are frozen;
add loss.NLLLoss
as an additional option for ranking.
v0.2.5
In this release, we:
update the implementation of parallel modules to fix the synchronization of gradient state on GPUs and to support torch.jit;
add the support of FP16 training based on APEX, but only useful for new GPUs;
set the computation order of transformer v2 as default, since it normally performs better on large datasets;
try to add the support of sentential context under transformer/SC/
.
v0.2.4
v0.2.3
v0.2.2
v0.2.1
v0.2.0
Pre-Release v0.1.7
Set the Transformer model in the original paper as default for performance.