标签:
neural network and deep learning 这本书看了陆陆续续看了好几遍了,但每次都会有不一样的收获。DL领域的paper日新月异,每天都会有很多新的idea出来,我想,深入阅读经典书籍和paper,一定可以从中发现remian open的问题,从而有不一样的视角。
PS:blog主要摘取书中重要内容简述。
摘要部分
Deep learning, a powerful set of techniques for learning in neural networks.
CHAPTER 1 Using neural nets to recognize handwritten digits
the neural network uses the examples to automatically infer rules for recognizing handwritten digits.
The exact form of active function isn’t so important - what really matters is the shape of the function when plotted.
4.The architecture of neural networks
The design of the input and output layers of a neural network is often straightforward, there can be quite an art to the design of the hidden layers. But researchers have developed many design heuristics for the hidden layers, which help people get the behaviour they want out of their nets.
Learning with gradient descent
Thinking about hyper-parameter choosing
”If we were coming to this problem for the first time then there wouldn’t be much in the output to guide us on what to do. We might worry not only about the learning rate, but about every other aspect of our neural network. We might wonder if we’ve initialized the weights and biases in a way that makes it hard for the network to learn? Or maybe we don’t have enough training data to get meaningful learning? Perhaps we haven’t run for enough epochs? Or maybe it’s impossible for a neural network with this architecture to learn to recognize handwritten digits? Maybe the learning rate is too low? Or, maybe, the learning rate is too high? When you’re coming to a problem for the first time, you’re not always sure.
The lesson to take away from this is that debugging a neural network is not trivial, and, just as for ordinary programming, there is an art to it. You need to learn that art of debugging in order to get good results from neural networks. More generally, we need to develop heuristics for choosing good hyper-parameters and a good architecture.”
CHAPTER 2 How the backpropagation algorithm works
Backpropagation(BP): a fast algorithm for computing the gradient of the cost function.
For backpropagation to work we need to make two main assumptions about the form of the cost function.
The four fundamental equations behind backpropagation
What’s clever about BP is that it enables us to simultaneously compute all the partial derivatives using just one forward pass through the network, followed by one backward pass through the network.
What indeed the BP do and how someone could ever have discovered BP?
A small perturbations will cause a change in the activation,then next and so on all the way through to causing a change in the final layer,and then the cost function.
A clever way of keeping track of small perturbations to the weights (and biases) as they propagate through the network, reach the output, and then affect the cost.
(未完待续……)
neural network and deep learning笔记(1)
标签:
原文地址:http://blog.csdn.net/dp_bupt/article/details/50990388