Skip to content
This repository has been archived by the owner on Jul 7, 2023. It is now read-only.

Commit

Permalink
better walkthrough README
Browse files Browse the repository at this point in the history
The default batch_size for transformer_base_single_gpu
has been decreased to 2048 in transform.py
(it used to be 4096 I think).
So if it is too much, the user must use a smaller value.
  • Loading branch information
martinpopel committed Aug 1, 2017
1 parent 1bf3b44 commit 8839cf9
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -89,7 +89,7 @@ t2t-datagen \
--problem=$PROBLEM
# Train
# * If you run out of memory, add --hparams='batch_size=2048' or even 1024.
# * If you run out of memory, add --hparams='batch_size=1024'.
t2t-trainer \
--data_dir=$DATA_DIR \
--problems=$PROBLEM \
Expand Down Expand Up @@ -166,7 +166,7 @@ python -c "from tensor2tensor.models.transformer import Transformer"
with `Modality` objects, which are specified per-feature in the dataset/task
specification.
* Support for multi-GPU machines and synchronous (1 master, many workers) and
asynchrounous (independent workers synchronizing through a parameter server)
asynchronous (independent workers synchronizing through a parameter server)
[distributed training](https://github.com/tensorflow/tensor2tensor/tree/master/docs/distributed_training.md).
* Easily swap amongst datasets and models by command-line flag with the data
generation script `t2t-datagen` and the training script `t2t-trainer`.
Expand Down

0 comments on commit 8839cf9

Please sign in to comment.