-
Notifications
You must be signed in to change notification settings - Fork 12
/
Copy pathhyperparameters.txt
32 lines (27 loc) · 944 Bytes
/
hyperparameters.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
#######################################################
# Specify all the model hyperparameters here,
# then run the training/testing scripts.
#######################################################
# Main working directory
workdir=/home/rmarvin/external/LM_syneval
# Directory where the Pytorch word language model code lives (with modifications for the multitask LMs)
lm_dir=$workdir/word-language-model
# Paths to LM & CCG data
lm_data_dir=$workdir/data/lm_data
ccg_data_dir=$workdir/data/ccg_data
train=train.small.txt
valid=valid.small.txt
test=test.small.txt
# Path to save the model
model_dir=$workdir/models
# TRAINING: Default LM hyperparameters
epochs=40
model=LSTM
log_freq=1000 # how often to display training progress
batch_size=128 # batch size
num_hid=650 # number of hidden units
lr=20.0 # learning rate
seed=2223 # random seed
nlayers=2 # number of layers in network
order=5 # default order for an ngram model
dropout=0.2