24. Learning to solve a problem
24 http://neuralnetworksanddeeplearning.com/chap1.html
25. Forward and Backward passes
25
http://caffe.berkeleyvision.org/tutorial/forward_backward.html
26. The Unstable Gradient Problem
26
Why it is difficult to train an RNN
Why are deep neural networks hard to train?
27. Practical advices when training neural networks
(by Ilya Sutskever)
27
• Get good data
• Preprocessing
• Minibatches
• Gradient normalization
• Learning rate schedule
• Learning rate
• Weight Initialization
• Data augmentation
• Dropout
• Ensembling
A Brief Overview of Deep Learning
28. Outline
• Intro to DL (A. Ng)
• Intro to Neural Nets
• Training NN
• Conv Nets
• Autoencoders
• Word Embeddings
• DL@TRD
• Bonus
28
41. 41
Word embeddings and Shared representations
Deep Learning, NLP, and Representations Deep Visual-Semantic Alignments for Generating Image
Descriptions
43. 43
Word embeddings and Reversible Sentence Representation
Deep Learning, NLP, and Representations
Rich Rashid in Tianjin, October, 25, 2012
44. Telenor Norway Network topology
Word embeddings applied to Network operations
Use cases:
• Predict failures of Network components.
• Predict congestion levels on Network links.
• Detect mal-functioning devices.
45. Outline
• Intro to DL (A. Ng)
• Intro to Neural Nets
• Training NN
• Conv Nets
• Autoencoders
• Word Embeddings
• DL@TRD
• Bonus
45
46. DL@TRD - Motivations
Personal observations:
– DL is hot (hyped?)
– DL supremacy seems ineluctable
– DL can solve a whole bunch of problems
– DL is frontier technology (difficult)
– Little DL competence @ Telenor Research
Personal implications:
– Career development
– Network with partners to get momentum
– Great if this happens in Trondheim
46
47. DL@TRD - Vision
Establish a strong DL competence center in Trondheim
– A place where
• competence is gathered
• experiences are exchanged
• collaborations are fostered
– Benefits
• Share passion with others near you
• Get momentum for your work
• Funding (SFI, EU money)
– Ideally
• Collaborate across companies on problems
• Common publications
47
Next workshop: 27th March
49. Stuff we could discuss at DL@TRD
• Training Recurrent Neural Networks
• Long Short Term Memory Networks
• Echo State Networks
• Neural Turing Machines
• Hopfield Nets
• Restricted Bolzman Machines
• Deep beliefs Networks
• Teacher – Student Nets
• Momentum
• Dropout
• Full Bayesian learning
• Hessian free optimization
• Stuff I don´t know I don´t know
49
50. Conclusion & Forecast
50
• DL techniques can be applied to all sorts of data:
– Could you apply some of these techniques to your data?
• DL models are better than humans at some tasks if fed with enough
data & trained properly
• Within 5-10 years, “information work” tasks will be augmented or even
fully automated
– See Peter Norvig´s talk at InfoQ: Machine Learning for Programming
– Models can take decisions based on millions of records while removing human
biases
Big data + Deep Learning = unemployment
– New policies and economic measures will be needed to manage the adverse
effects of job computerization
– Schooling will need reforms: routine tasks non-routine tasks