|  | |
|  | |
|  | |
| # NeuralGPU | |
| Code for the Neural GPU model described in http://arxiv.org/abs/1511.08228. | |
| The extended version was described in https://arxiv.org/abs/1610.08613. | |
| Requirements: | |
| * TensorFlow (see tensorflow.org for how to install) | |
| The model can be trained on the following algorithmic tasks: | |
| * `sort` - Sort a symbol list | |
| * `kvsort` - Sort symbol keys in dictionary | |
| * `id` - Return the same symbol list | |
| * `rev` - Reverse a symbol list | |
| * `rev2` - Reverse a symbol dictionary by key | |
| * `incr` - Add one to a symbol value | |
| * `add` - Long decimal addition | |
| * `left` - First symbol in list | |
| * `right` - Last symbol in list | |
| * `left-shift` - Left shift a symbol list | |
| * `right-shift` - Right shift a symbol list | |
| * `bmul` - Long binary multiplication | |
| * `mul` - Long decimal multiplication | |
| * `dup` - Duplicate a symbol list with padding | |
| * `badd` - Long binary addition | |
| * `qadd` - Long quaternary addition | |
| * `search` - Search for symbol key in dictionary | |
| It can also be trained on the WMT English-French translation task: | |
| * `wmt` - WMT English-French translation (data will be downloaded) | |
| The value range for symbols are defined by the `vocab_size` flag. | |
| In particular, the values are in the range `vocab_size - 1`. | |
| So if you set `--vocab_size=16` (the default) then `--problem=rev` | |
| will be reversing lists of 15 symbols, and `--problem=id` will be identity | |
| on a list of up to 15 symbols. | |
| To train the model on the binary multiplication task run: | |
| ``` | |
| python neural_gpu_trainer.py --problem=bmul | |
| ``` | |
| This trains the Extended Neural GPU, to train the original model run: | |
| ``` | |
| python neural_gpu_trainer.py --problem=bmul --beam_size=0 | |
| ``` | |
| While training, interim / checkpoint model parameters will be | |
| written to `/tmp/neural_gpu/`. | |
| Once the amount of error gets down to what you're comfortable | |
| with, hit `Ctrl-C` to stop the training process. The latest | |
| model parameters will be in `/tmp/neural_gpu/neural_gpu.ckpt-<step>` | |
| and used on any subsequent run. | |
| To evaluate a trained model on how well it decodes run: | |
| ``` | |
| python neural_gpu_trainer.py --problem=bmul --mode=1 | |
| ``` | |
| To interact with a model (experimental, see code) run: | |
| ``` | |
| python neural_gpu_trainer.py --problem=bmul --mode=2 | |
| ``` | |
| To train on WMT data, set a larger --nmaps and --vocab_size and avoid curriculum: | |
| ``` | |
| python neural_gpu_trainer.py --problem=wmt --vocab_size=32768 --nmaps=256 | |
| --vec_size=256 --curriculum_seq=1.0 --max_length=60 --data_dir ~/wmt | |
| ``` | |
| With less memory, try lower batch size, e.g. `--batch_size=4`. With more GPUs | |
| in your system, there will be a batch on every GPU so you can run larger models. | |
| For example, `--batch_size=4 --num_gpus=4 --nmaps=512 --vec_size=512` will | |
| run a large model (512-size) on 4 GPUs, with effective batches of 4*4=16. | |
| Maintained by Lukasz Kaiser (lukaszkaiser) | |