-
Notifications
You must be signed in to change notification settings - Fork 9
/
Run_Model_Demos-2016-INFINITE.sh
93 lines (52 loc) · 4.7 KB
/
Run_Model_Demos-2016-INFINITE.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
#!/bin/bash
declare -i samples=444
printf "\n\nWavenet for Poem Generation.\nInitializing.\nPlease wait.\n\nPreparing to make 30 poems of $samples letters each.\n\nEach poem is generated from a mathematical model of what poetry is.\n\nThe models were generated using a neural-net (Wavenet) \nin November 2016 in Hong Kong.\n\nThe poems are generated now in realtime.\n\n"
while :
do
############## Using a neural net based on Google's Wavenet algorithm.\npoems are generated letter by letter as patterns \nwithout any dictionary, syntax, or preset rules: unsupervised.\n\n"
DATE=`date +%Y-%m-%d_%H-%M-%S`
################
declare -i var=0
################
# 39-18
wavenet_params="wavenet_params_ORIG_dilations1024_skipChannels4096_qc1024_dc8.json"
declare -a LOSS=(0.359 0.331 0.342 0.647 0.321 0.501 0.319 0.615 0.32 )
for i in ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-42013 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-84953 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-93539 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-94028 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-102129 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-110714 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-119305 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-124484 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc8/model.ckpt-127893
do
for word in $i
do
python generate_Poems_2016-folder.py --start_time=$DATE --samples=$samples --wavenet_params=$wavenet_params --loss=${LOSS[$var]} $word
var=$((var+1))
done
done
################
declare -i var=0
################
# 38-59
wavenet_params="wavenet_params_ORIG_dilations1024_skipChannels4096_qc1024_dc16.json"
declare -a LOSS=(0.369 0.306 0.327 0.338 0.556 )
for i in ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc16/model.ckpt-24835 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc16/model.ckpt-84953 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc16/model.ckpt-102130 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc16/model.ckpt-119303 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc16/model.ckpt-119542
do
for word in $i
do
python generate_Poems_2016-folder.py --start_time=$DATE --samples=$samples --wavenet_params=$wavenet_params --loss=${LOSS[$var]} $word
var=$((var+1))
done
done
################
declare -i var=0
################
# 26-03
wavenet_params="wavenet_params_ORIG_dilations1024_skipChannels4096_qc1024_dc32.json"
declare -a LOSS=(0.497 0.303 0.524 0.521 0.504 0.284 0.293 0.303 0.278 0.288 0.293 0.280 0.607 0.549 0.625 0.506)
for i in ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-49523 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-50601 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-50603 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-58102 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-58111 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-67777 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-67778 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-84951 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-93541 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-93542 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-102127 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-102129 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-102616 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-103431 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-107721 ./logdir/train/demos/dilations1024_skipChannels4096_qc1024_dc32/model.ckpt-107723
do
for word in $i
do
python generate_Poems_2016-folder.py --start_time=$DATE --samples=$samples --wavenet_params=$wavenet_params --loss=${LOSS[$var]} $word
var=$((var+1))
done
done
sleep 10s
printf "\n\n************************\n\nLOOPING THRU DEMOS AGAIN\n\nWavenet for Poem Generation.\n\nPreparing to make another 30 poems of $samples letters each.\n\n************************\n\n"
done