I am new to neural networks and, to understand this, I have implemented basic MLP, which I am currently training with backpropagation. I know that there are more complex and better ways to do this, but in Introduction to Machine Learning, they suggest that with one or two tricks, a basic gradient descent can be effective for studying real-world data. One of the tricks is adaptive learning speed.
The idea is to increase the learning speed by a constant value of a when the error becomes smaller, and reduce it by a fraction b of the learning speed when the error becomes larger. Thus, basically, the change in learning speed is determined as follows:
+(a)
if we study in the right direction, and
-(b * <learning rate>)
if we ruin our learning. However, in the above book there are no recommendations on how to set these parameters. I would not expect an exact offer, since setting parameters is a whole topic in itself, but just a hint, at least an order of magnitude. Any ideas?
Thanks,
Tunnuz
source share