The experiments on the stacked autoencoder suggest that pretraining should be done with low values of learning rate. This ensures better convergence and performance during the fine tuning.
The experiments on the stacked autoencoder suggest that pretraining should be done with low values of learning rate. This ensures better convergence and performance during the fine tuning.