Chainer-based Python implementation of Transformer, an attention-based seq2seq model without convolution and recurrence.
If you want to see the architecture, please see net.py.
See "Attention Is All You Need", Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, arxiv, 2017.
This repository is partly derived from my convolutional seq2seq repo, which is also derived from Chainer's official seq2seq example.
pip
)You can use any parallel corpus.
For example, run
sh download_wmt.sh
which downloads and decompresses training dataset and development dataset from WMT/europal into your current directory. These files and their paths are set in training script train.py
as default.
PYTHONIOENCODING=utf-8 python -u train.py -g=0 -i DATA_DIR -o SAVE_DIR
During training, logs for loss, perplexity, word accuracy and time are printed at a certain internval, in addition to validation tests (perplexity and BLEU for generation) every half epoch. And also, generation test is performed and printed for checking training progress.
Some of them is as follows:
-g
: your gpu id. If cpu, set -1
.-i DATA_DIR
, -s SOURCE
, -t TARGET
, -svalid SVALID
, -tvalid TVALID
:DATA_DIR
directory needs to include a pair of training dataset SOURCE
and TARGET
with a pair of validation dataset SVALID
and TVALID
. Each pair should be parallell corpus with line-by-line sentence alignment.-o SAVE_DIR
: JSON log report file and a model snapshot will be saved in SAVE_DIR
directory (if it does not exist, it will be automatically made).-e
: max epochs of training corpus.-b
: minibatch size.-u
: size of units and word embeddings.-l
: number of layers in both the encoder and the decoder.--source-vocab
: max size of vocabulary set of source language--target-vocab
: max size of vocabulary set of target languagePlease see the others by python train.py -h
.
This repository does not aim for complete validation of results in the paper, so I have not eagerly confirmed validity of performance. But, I expect my implementation is almost compatible with a model described in the paper. Some differences where I am aware are as follows:
relu
into leaky relu
in feedforward net layers for easy gradient propagation.