-
Notifications
You must be signed in to change notification settings - Fork 0
/
log.txt
169 lines (144 loc) · 7.37 KB
/
log.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
EchoStateNetwork (normalization: True, num samples: 4)
num_units: 0.13976189827119723
leakage: 0.00029876661771628216
connectivity: 8.612761615983677e-06
spectral_radius: 0.0001309925104538241
learning_rate: 0.17964205845593098
EchoStateNetwork (normalization: True, num samples: 40)
num_units: 0.01935747429429748
leakage: 8.818226223732691e-05
connectivity: 4.94149039040629e-05
spectral_radius: 3.855358960872978e-05
learning_rate: 0.02647025779110094
MLP (normalization: 1049.288386106491, num samples: 43, time: N/A)
learning_rate: 6.380568507245867e-05
batch_size: 1.6498285153654165e-05
activation_function: 0.0011703417210019906
num_layers: 2.983129480897791e-06
num_neurons: 4.319816711748486e-05
MLP (normalization: False, num samples: 43, time: 00:17:51)
learning_rate: 6.428807632496672e-05
batch_size: 1.5612965137350684e-05
activation_function: 0.0032115370224535143
num_layers: 7.782473695669979e-06
num_neurons: 3.05817606527341e-05
MLP (normalization: False, num samples: 5, time: 00:02:43)
learning_rate: 0.00036937373591086956
batch_size: 4.301193699607353e-05
activation_function: 0.002272151042738695
num_layers: 2.2952508912155167e-05
num_neurons: 0.00015373201071554378
MLP (normalization: False, num samples: 5, time: 00:01:21)
learning_rate: 0.00043177379809065995
batch_size: 7.41549907639083e-05
activation_function: 0.0036811974677224855
num_layers: 2.017252506281153e-05
num_neurons: 8.6844030243185e-05
MLP (normalization: False, num samples: 5, time: 00:01:17)
learning_rate: 0.0003647815566048962
batch_size: 4.682737843775264e-05
activation_function: 0.0021088007587889046
num_layers: 2.0300067377299058e-05
num_neurons: 0.00016147766838315648
MLP (normalization: False, num samples: 5, time: 00:01:18)
learning_rate: 0.0003666137321152974
batch_size: 4.513234340453032e-05
activation_function: 0.0045479640484406
num_layers: 3.5772415437840266e-05
num_neurons: 0.00016858635142682487
MLP (normalization: 455.47265911102295, num samples: 43, time: N/A)
learning_rate: 0.00013533976191048785
batch_size: 3.801755227100989e-05
activation_function: 0.002606258338056233
num_layers: 3.249915664651222e-05
num_neurons: 8.851299262548139e-05
MLP (num samples: 43, time: 00:07:07)
learning_rate: 0.000124542937173418
batch_size: 3.1582461749621554e-05
activation_function: 0.00426785148180025
num_layers: 5.295426470354413e-05
num_neurons: 8.801767628430164e-05
EchoStateNetwork(num samples: 4, time: 00:04:39)
num_units: 2062.311738021832
leakage: 2726.030590639959
connectivity: 92.6631726869964
spectral_radius: 3970.6439704961813
learning_rate: 3199673.3713146746
MLP(num samples: 23, time: 00:26:30)
learning_rate: 0.00019632258326433982
batch_size: 1.5606094361899343e-05
activation_function: 0.00046773567604081225
num_layers: 2.8903972431351125e-05
num_neurons: 2.8811577727216777e-05
MLP(num samples: 23, time: 00:26:35)
learning_rate: 0.05600422931288214
batch_size: 1.1793814123036733e-05
activation_function: 2.6926386064332066e-05
num_layers: 1.5772705713557952e-05
num_neurons: 2.0806467762124638e-05
MLP(num samples: 23, time: 00:25:57)
learning_rate: Variance=0.00033211476307412326, fANOVA=0.8715478115735549
batch_size: Variance=1.7980684986798776e-05, fANOVA=0.025755045206485385
activation_function: Variance=7.940451442323138e-05, fANOVA=0.04435276802261638
num_layers: Variance=1.9302089247576403e-05, fANOVA=0.0033963754971984444
num_neurons: Variance=3.969386279708946e-05, fANOVA=0.05494799970014491
LucasModel(num samples: 21, time: 00:19:49)
Conv1D_filters: Variance=0.0, fANOVA=0.0
kernel_size: Variance=0.0, fANOVA=0.0
Dense1_units: Variance=0.0, fANOVA=0.0
Dense2_units: Variance=0.0, fANOVA=0.0
LucasModel(num samples: 2, time: 00:00:22)
Conv1D_filters: Variance=0.0017530418269053882, fANOVA=0.3072277356198734
kernel_size: Variance=0.0014113772372050803, fANOVA=0.18717008665722523
Dense1_units: Variance=0.003280841811829305, fANOVA=0.3311572982981821
Dense2_units: Variance=0.001017974375091768, fANOVA=0.1744448794247193
LucasModel(num samples: 21, time: 00:20:37)
Conv1D_filters: Variance=0.013830460053415386, fANOVA=0.4166588815938626
kernel_size: Variance=0.0009262201419920485, fANOVA=0.044108966995431174
Dense1_units: Variance=0.0057197349459134975, fANOVA=0.36708107388407657
Dense2_units: Variance=0.002678016534158775, fANOVA=0.1721510775266297
LucasModel(num samples: 2, time: 00:00:16, avg performance: 0.8441942632198334)
Conv1D_filters: Variance=0.00362354195204162, fANOVA=0.36039431515467724
kernel_size: Variance=0.0002100277139085449, fANOVA=0.06072507878158971
Dense1_units: Variance=0.000690689351622531, fANOVA=0.1388097088470888
Dense2_units: Variance=0.004269130271695509, fANOVA=0.4400708972166444
esn(num samples: 2, time: 00:00:07, avg performance: 9055.283203125)
num_units: Variance=72.1338357925415, fANOVA=0.03957433118297886
leakage: Variance=18037.224460601807, fANOVA=0.14101739001558897
connectivity: Variance=84337.20804429054, fANOVA=0.2840180336167243
spectral_radius: Variance=19590.157485961914, fANOVA=0.09951902703676359
learning_rate: Variance=44405.2725687027, fANOVA=0.43587121814794416
esn(num samples: 2, time: 00:00:07, avg performance: 8909.08544921875)
num_units: Variance=63443.967319488525, fANOVA=0.34149208967327493
leakage: Variance=1396.854126214981, fANOVA=0.027311646648027564
connectivity: Variance=1331.109619140625, fANOVA=0.2265003549561577
spectral_radius: Variance=8888.9541015625, fANOVA=0.05828977069817899
learning_rate: Variance=39052.74578118324, fANOVA=0.3464061380243609
esn(num samples: 2, time: 00:00:07, avg performance: 8220.577880859375)
num_units: Variance=7601.149300575256, fANOVA=0.08118887714372042
leakage: Variance=3732.505951166153, fANOVA=0.11142320227202732
connectivity: Variance=371.91725158691406, fANOVA=0.015114532251389166
spectral_radius: Variance=158.62713718414307, fANOVA=0.22122368547772003
learning_rate: Variance=262587.4373689294, fANOVA=0.571049702855143
lucas_epochs_5(num samples: 21, time: 00:10:53, avg performance: 0.6513898315883818)
Conv1D_filters: Variance=0.002691060952203176, fANOVA=0.22669989303089974
kernel_size: Variance=0.008159389634296799, fANOVA=0.414981164972262
Dense1_units: Variance=0.005901564096981602, fANOVA=0.3111255980742546
Dense2_units: Variance=0.000583286401769813, fANOVA=0.047193343922583536
lucas_epochs_20(num samples: 21, time: 00:35:39, avg performance: 0.5613525226002648)
Conv1D_filters: Variance=0.07304033616597613, fANOVA=0.5340304159247481
kernel_size: Variance=0.06497990832565649, fANOVA=0.32295460373706625
Dense1_units: Variance=0.011138898173615153, fANOVA=0.12763588678781518
Dense2_units: Variance=0.0009091749000706426, fANOVA=0.015379093550370568
esn_samples_20(num samples: 20, time: 04:57:07, avg performance: 136.4200798034668)
num_units: Variance=296.4198717704048, fANOVA=0.004910231149847307
leakage: Variance=4391.895481798163, fANOVA=0.08080627222473738
connectivity: Variance=141.25734953878506, fANOVA=0.002434941409952328
spectral_radius: Variance=17367.341957948425, fANOVA=0.6970892662695467
learning_rate: Variance=19848.550610899052, fANOVA=0.21475928894591614
esn_samples_15(num samples: 16, time: 03:44:09, avg performance: 85.91047525405884)
num_units: Variance=353.29585332804606, fANOVA=0.01884744715526467
leakage: Variance=1328.9155016120458, fANOVA=0.057080586737022716
connectivity: Variance=54.77460443851305, fANOVA=0.0030771977226936474
spectral_radius: Variance=8073.443452324973, fANOVA=0.71605566686878
learning_rate: Variance=5557.8784701434515, fANOVA=0.20493910151623898