标签:
http://handong1587.github.io/deep_learning/2015/10/09/rnn-and-lstm.html //RNN and LSTM
http://handong1587.github.io/deep_learning/2015/10/09/saliency-prediction.html //saliency Predection
http://handong1587.github.io/deep_learning/2015/10/09/scene-labeling.html //Scene Label
Published: 09 Oct 2015 Category: deep_learning
1) Plain Tanh Recurrent Nerual Networks
2) Gated Recurrent Neural Networks (GRU)
3) Long Short-Term Memory (LSTM)
A Beginner’s Guide to Recurrent Networks and LSTMs
http://deeplearning4j.org/lstm.html
A Deep Dive into Recurrent Neural Nets
http://nikhilbuduma.com/2015/01/11/a-deep-dive-into-recurrent-neural-networks/
Long Short-Term Memory: Tutorial on LSTM Recurrent Networks
http://people.idsia.ch/~juergen/lstm/index.htm
LSTM implementation explained
http://apaszke.github.io/lstm-explained.html
Recurrent Neural Networks Tutorial
Understanding LSTM Networks
Recurrent Neural Networks in DL4J
http://deeplearning4j.org/usingrnns.html
A Simple Way to Initialize Recurrent Networks of Rectified Linear Units
Sequence Level Training with Recurrent Neural Networks
Generating Sequences With Recurrent Neural Networks
DRAW: A Recurrent Neural Network For Image Generation
Unsupervised Learning of Video Representations using LSTMs(ICML2015)
LSTM: A Search Space Odyssey
Inferring Algorithmic Patterns with Stack-Augmented Recurrent Nets
A Critical Review of Recurrent Neural Networks for Sequence Learning
Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks(Winner of MSCOCO image captioning challenge, 2015)
Visualizing and Understanding Recurrent Networks(Andrej Karpathy, Justin Johnson, Fei-Fei Li)
Grid Long Short-Term Memory
Depth-Gated LSTM
Deep Knowledge Tracing
Top-down Tree Long Short-Term Memory Networks
Alternative structures for character-level RNNs(INRIA & Facebook AI Research)
Pixel Recurrent Neural Networks (Google DeepMind)
Long Short-Term Memory-Networks for Machine Reading
Lipreading with Long Short-Term Memory
Associative Long Short-Term Memory
Representation of linguistic form and function in recurrent neural networks
Architectural Complexity Measures of Recurrent Neural Networks
Easy-First Dependency Parsing with Hierarchical Tree LSTMs
Training Input-Output Recurrent Neural Networks through Spectral Methods
Learning to Execute
Neural Programmer-Interpreters (Google DeepMind)
A Programmer-Interpreter Neural Network Architecture for Prefrontal Cognitive Control
Convolutional RNN: an Enhanced Model for Extracting Features from Sequential Data
Recurrent Models of Visual Attention (Google DeepMind. NIPS2014)
Recurrent Model of Visual Attention(Google DeepMind)
Show, Attend and Tell: Neural Image Caption Generation with Visual Attention
A Neural Attention Model for Abstractive Sentence Summarization(EMNLP 2015. Facebook AI Research)
Effective Approaches to Attention-based Neural Machine Translation(EMNLP2015)
Generating Images from Captions with Attention
Attention and Memory in Deep Learning and NLP
Survey on the attention based RNN model and its applications in computer vision
Training Recurrent Neural Networks (PhD thesis)
Deep learning for control using augmented Hessian-free optimization
Hierarchical Conflict Propagation: Sequence Learning in a Recurrent Deep Neural Network
Recurrent Batch Normalization
Optimizing Performance of Recurrent Neural Networks on GPUs
NeuralTalk (Deprecated): a Python+numpy project for learning Multimodal Recurrent Neural Networks that describe images with sentences
NeuralTalk2: Efficient Image Captioning code in Torch, runs on GPU
char-rnn in Blocks
Project: pycaffe-recurrent
Using neural networks for password cracking
Recurrent neural networks for decoding CAPTCHAS
torch-rnn: Efficient, reusable RNNs and LSTMs for torch
Deploying a model trained with GPU in Torch into JavaScript, for everyone to use
LSTM implementation on Caffe
Survey on Attention-based Models Applied in NLP
http://yanran.li/peppypapers/2015/10/07/survey-attention-model-1.html
Survey on Advanced Attention-based Models
http://yanran.li/peppypapers/2015/10/07/survey-attention-model-2.html
Online Representation Learning in Recurrent Neural Language Models
http://www.marekrei.com/blog/online-representation-learning-in-recurrent-neural-language-models/
Fun with Recurrent Neural Nets: One More Dive into CNTK and TensorFlow
Materials to understand LSTM
https://medium.com/@shiyan/materials-to-understand-lstm-34387d6454c1#.4mt3bzoau
Understanding LSTM and its diagrams (★★★★★)
Persistent RNNs: 30 times faster RNN layers at small mini-batch sizes (Greg Diamos, Baidu Silicon Valley AI Lab)
http://svail.github.io/persistent_rnns/
All of Recurrent Neural Networks
https://medium.com/@jianqiangma/all-about-recurrent-neural-networks-9e5ae2936f6e#.q4s02elqg
Awesome Recurrent Neural Networks - A curated list of resources dedicated to RNN
Jürgen Schmidhuber’s page on Recurrent Neural Networks
http://people.idsia.ch/~juergen/rnn.html
Are there any Recurrent convolutional neural network network implementations out there ?
« Reinforcement LearningSaliency Prediction »
This task involves predicting the salient regions of an image given by human eye fixations.
Large-scale optimization of hierarchical features for saliency prediction in natural images
Predicting Eye Fixations using Convolutional Neural Networks
DeepFix: A Fully Convolutional Neural Network for predicting Human Eye Fixations
DeepSaliency: Multi-Task Deep Neural Network Model for Salient Object Detection
SuperCNN: A Superpixelwise Convolutional Neural Network for Salient Object Detection
Shallow and Deep Convolutional Networks for Saliency Prediction
Learning hierarchical features for scene labeling
Indoor Semantic Segmentation using depth information
Multi-modal unsupervised feature learning for rgb-d scene labeling
Using neon for Scene Recognition: Mini-Places2
Attend, Infer, Repeat: Fast Scene Understanding with Generative Models
Large-scale Scene Understanding Challenge
RNN and LSTM saliency Predection Scene Label
标签:
原文地址:http://www.cnblogs.com/hansjorn/p/5396668.html