- While logistic sigmoid neurons are more bi-ologically plausible than hyperbolic tangent neurons, the latter work better for train-ing multi-layer neural networks. Deep Sparse Rectifier Neural Networks
- tanh vs ReLU
- this blog gives a clear explanation, Global average pooling--
- alexisbcook's professional explanation
- an explanation of average pooling with tensorflow codes
- What are Max Pooling, Average Pooling, Global Max Pooling and Global Average Pooling?
- dropout layer is used to overcome overfitting, refer to:Improving neural networks by preventing co-adaptation of feature detectors by Hinton
- In pytorch, the use of WeightedRandomSampler makes the learning more smoothly quicker but the test dataset cannot be used with WeightedRandomSampler(TODO why ) Note: shuffe should be False when using WeightedRandomSampler
- Financial Time Series Forecasting with Deep Learning -ASystematic Literature Review- 2005-2019
- Comparison between DeepESNs and gated RNNs on multivariate time-series prediction
- Multivariate Temporal Convolutional Network--A Deep Neural Networks Approach for Multivariate Time Series Forecasting