标签:
awesome-very-deep-learning is a curated list for papers and code about implementing and training very deep neural networks.
Deep Residual Networks are a family of extremely deep architectures (up to 1000 layers) showing compelling accuracy and nice convergence behaviors. Instead of learning a new representation at each layer, deep residual networks use identity mappings to learn residuals.
In addition, this code by Ryan Dahl helps to convert the pre-trained models to TensorFlow.
Highway Networks take inspiration from Long Short Term Memory (LSTM) and allow training of deep, efficient networks (with hundreds of layers) with conventional gradient-based methods
Theories in very deep learning concentrate on the ideas that very deep networks with skip connections are able to efficiently approximate recurrent computations (similar to the recurrent connections in the visual cortex) or are actually exponential ensembles of shallow networks
标签:
原文地址:http://www.cnblogs.com/wangxiaocvpr/p/5824177.html