Skip to content
This repository was archived by the owner on May 22, 2020. It is now read-only.

Latest commit

 

History

History
31 lines (23 loc) · 868 Bytes

File metadata and controls

31 lines (23 loc) · 868 Bytes

team004-MachineLearningModel

In order to get all the files, you will have to execute some of the shell script, which are in the folder (You might need to sudo in order to execute some of the code below)

  • save_transformer
  • save_word2vec
  • dataset

To startup the server/docker container execute

sh startup_docker.sh

And the everything is ready !!! The code for transformer is based on https://tensorflow.org/alpha/tutorials/sequences/transformer However, I have modularized for future usage and down-grade the tensorflow to a version 1.12 (instead of 2.0 as in tutorials)

Running Transformer

python3 training_transformer.py --vocabulary-size 7500 \
  --max-sen-length 20 \
  --num-layer 4 \
  --model-dim 128 \
  --pointwise-dim 512 \
  --num-head 8 \
  --dropout-rate 0.1 \
  --batch-size 64 \
  --warmup-length 400 \