SlideShare una empresa de Scribd logo
1 de 37
Descargar para leer sin conexión
[course site]
#DLUPC
Transfer learning and
domain adaptation
Day 5 Lecture 2
Kevin McGuinness
kevin.mcguinness@dcu.ie
Research Fellow
Insight Centre for Data Analytics
Dublin City University
Eric Arazo
eric.arazosanchez@dcu.ie
PhD Candidate
Insight Centre for Data Analytics
Dublin City University
Ramon Morros
ramon.morros@upc.edu
Associate Professor
Universitat Politecnica de Catalunya
Technical University of Catalonia
Many slides from:
The ability to apply knowledge learned in previous tasks to novel tasks
● Based on human learning. People can often transfer knowledge learnt
previously to novel situations
○ Play classic piano → Play jazz piano
○ Maths → Machine Learning
○ Ride motorbike → Drive a car
Transfer Learning
2
● Isolated, single task learning:
○ Knowledge is not retained or
accumulated. Learning is performed
w.o. considering past learned
knowledge in other tasks
● Learning of a new tasks relies on
the previous learned tasks:
○ Learning process can be faster, more
accurate and/or need less training data
Traditional ML vs Transfer Learning
Dataset 1
Learning
System
Task 1
Dataset 2
Learning
System
Task 2
Dataset 1
Dataset
2
Learning
System
Task 1
Knowledge
Learning
System
Task 2
3
Transfer learning in DL
Myth: you can’t do deep learning unless you have a million labelled examples for
your problem.
Reality
● You can learn useful representations from unlabelled data
● You can train on a nearby surrogate objective for which it is easy to generate
labels
● You can transfer learned representations from a related task
4
Transfer learning: idea
Instead of training a deep network from scratch for your task:
● Take a network trained on a different domain for a different source task
● Adapt it for your domain and your target task
This lecture will talk about how to do this.
Variations:
● Same domain, different task
● Different domain, same task
5
Source data
E.g. ImageNet
Source
model
Source labels
Target data
E.g. PASCAL
Target
model
Target
labels
Transfer Learned
Knowledge
Large amount
of data/labels
Small amount
of data/labels
Example: PASCAL VOC 2007
● Standard classification benchmark, 20 classes, ~10K images, 50% train, 50% test
● Deep networks can have many parameters (e.g. 60M in Alexnet)
● Direct training (from scratch) using only 5K training images can be problematic. Model overfits.
● How can we use deep networks in this setting?
6
Transfer Learning
Figure extracted from Sebastian Rudder’s blog: “Transfer Learning - Machine Learning's Next Frontier” http://ruder.io/transfer-learning/index.html 7
A Domain consists of two components:
● Feature space:
● Marginal distribution: ,
For a given domain D, a Task is defined by two components:
● A label space:
● A predictive function , learned from feature vector/label pairs, ,
● For each feature vector in the domain, predicts its corresponding label:
Notation (I)
8
Notation (II)
9
If two domains are different, they may have different
feature spaces or different marginal distributions
If two tasks are different, they may have different
label spaces or different conditional distributions
Digit 5, dataset A
Digit 5, dataset B
Dataset X
Vladimir
Barak
Angela Lionel
Dataset Y
Nuria
Garbiñe
Different label
space!
Notation (III)
10
For simplicity, only two domains or two tasks are usually considered
● Source domain
● Task on the source domain
● Target domain
● Task on the target domain
Domain adaptation
Consider a classification task where is the input space and is the set of labels.
Given two sets of samples drawn from the source and target domains:
The goal of the learning algorithm is to build a classifier with a low
target risk on the target domain
11
(Target space labeled data may not be
present in unsupervised case)
Domain biass
● Datasets are samples of the world
● In many cases, there is a shift or bias in the distributions of the source and
target data representations
12
A. Torralba, A. Efros. Unbiased Look at Dataset Bias. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2011.
Domain adaptation
When there is a domain shift
● The size of this shift is often measured by the distance between source and target
subspaces
● A typical approach is to learn a feature space transformation to align the source and
target representations (reduce domain divergence)
13
Digit 5, dataset ADigit 5, dataset B
Digit 5, dataset A
Digit 5, dataset B
Align
“Off-the-shelf”
Idea: use outputs of one or more layers of a network trained on a different task as
generic feature detectors. Train a new shallow model on these features.
Assumes that
conv2
conv3
fc1
conv1
loss
Data and labels (e.g. ImageNet)
fc2
softmax
TRANSFER
Shallow classifier (e.g. SVM)
conv2
conv3
fc1
conv1
Target data and labels
features
14
Off-the-shelf features
Works surprisingly well in practice!
Surpassed or on par with state-of-the-art in several
tasks in 2014
Image classification:
● PASCAL VOC 2007
● Oxford flowers
● CUB Bird dataset
● MIT indoors
Image retrieval:
● Paris 6k
● Holidays
● UKBench
Razavian et al, CNN Features off-the-shelf: an Astounding Baseline for Recognition, CVPRW 2014 http://arxiv.org/abs/1403.6382
15
Oxford 102 flowers dataset
(Trained to perform object classification on ILSVRC13)
Can we do better than off the shelf features?
16
Fine-tuning: supervised domain adaptation
Train deep net on “nearby” task for which it is easy
to get labels using standard backprop
● E.g. ImageNet classification
● Pseudo classes from augmented data
● Slow feature learning, ego-motion
Cut off top layer(s) of network and replace with
supervised objective for target domain
Fine-tune network using backprop with labels for
target domain until validation loss starts to increase
Aligns with
labels
conv2
conv3
fc1
conv1
surrogate loss
surrogate data
fc2 + softmax
real labelsreal data
real loss
my_fc2 + softmax
17
Freeze or fine-tune?
Bottom n layers can be frozen or fine tuned.
● Frozen: not updated during backprop
● Fine-tuned: updated during backprop
Which to do depends on target task:
● Freeze: target task labels are scarce, and we
want to avoid overfitting
● Fine-tune: target task labels are more plentiful
In general, we can set learning rates to be different
for each layer to find a tradeoff between freezing
and fine tuning labels
conv2
conv3
fc1
conv1
loss
data
fc2 + softmax
Finetunedfrozen
LR = 0
LR > 0
18
How transferable are features?
Lower layers: more general features. Transfer very
well to other tasks.
Higher layers: more task specific.
19
data
Matthew D Zeiler, Rob Fergus, “Visualizing and Understanding Convolutional Networks”, CVPR 2013
20
Yosinki et al. How transferable are features in deep neural networks. NIPS 2014. https://arxiv.org/abs/1411.1792
How transferable are features?
How transferable are features?
Transferability is negatively affected by two distinct
issues:
● The specialization of higher layer neurons
● Optimization difficulties related to splitting networks
between co-adapted neurons
Fine-tuning improves generalization when sufficient
examples are available.
Transfer learning and fine tuning often lead to better
performance than training from scratch on the target
dataset.
Even features transferred from distant tasks are
often better than random initial weights!
Yosinki et al. How transferable are features in deep neural networks. NIPS 2014. https://arxiv.org/abs/1411.1792
21
Unsupervised domain adaptation
Also possible to do domain adaptation without labels in target set.
Y Ganin and V Lempitsky, Unsupervised Domain Adaptation by Backpropagation, ICML 2015 https://arxiv.org/abs/1409.7495 22
Y. Ganin et al. "Domain-Adversarial Training of Neural Networks." Journal of Machine Learning research 17 (2016).
(Source domain)
(Source and target
domains)
Unsupervised domain adaptation
The approach promotes the emergence of "deep" features that are (i) discriminative for the main learning
task on DS
and (ii) invariant with respect to the shift between the domains
Y Ganin and V Lempitsky, Unsupervised Domain Adaptation by Backpropagation, ICML 2015 https://arxiv.org/abs/1409.7495
23
loss for label prediction
maximize the loss of the domain classifier (by making
the two feature distributions as similar as possible)
Minimize the loss of the domain
classifier and the label predictor
loss for the domain
classification
Y. Ganin,et al. "Domain-Adversarial Training of Neural Networks." Journal of Machine Learning research 17 (2016).
Unsupervised domain adaptation
Y Ganin and V Lempitsky, Unsupervised Domain Adaptation by Backpropagation, ICML 2015 https://arxiv.org/abs/1409.7495 24
Task transfer
Maximizing domain confusion does not necessarily align the classes in the target
with those in the source
Possible solution: Transfer the similarity structure amongst categories from
the source to the target, using distillation
(Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS
2014 DL Workshop, 1–9)
25
Original application was to transfer the knowledge from a large, easy to train model
into a smaller/faster model more suitable for deployment
Bucilua1
demonstrated that this can be done reliably when transferring from a large
ensemble of models to a single small model
Distillation
261
C.Bucilua, R. Caruana, and A. Niculescu-Mizil. “Model compression”. In ACMSIG KDD ’06, 2006
Idea: use the class probabilities produced by the large model as “soft targets” for
training the small model
○ The ratios of probabilities in the soft targets provide information about the learned function
○ These ratios carry information about the structure of the data
○ Train by replacing the hard labels with the softmax activations from the original large model
Distillation
Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9.
0.05
0.8
0.1
0.05
0
1
0
0
Yn
Yn
0.09
0.05
0.85
0.01
0.4
0.1
0.2
0.3
Y0Y0
27
Distillation lossMultinomial logistic loss
Distillation
● To increase the influence of non-target class probabilities in the cross entropy, the
temperature of the final softmax is raised to “soften” the final probability distribution
over classes
● Transfer can be obtained by using the same large model training set or a separate
training set
● If the ground-truth labels of the transfer set are known, standard loss and distillation
loss can be combined
Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9. 28
0.09
0.05
0.85
0.01
0.15
0.10
0.70
0.05
T=1 T>1
Semi-supervised task & domain adaptation
When some labels are available in the target domain, then we can use these when
doing domain adaptation. I.e. combine fine tuning and unsupervised domain
adaptation.
Tzeng et al. take this a step further and try to simultaneously optimize a loss that
maximizes:
1. classification accuracy on both source and target datasets
2. domain confusion of a domain classifier
3. agreement of classifier score distributions across domains
Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015.
29
Semi-supervised domain adaptation
Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015.
30
Semi-supervised domain adaptation
Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015.
Classifier loss
Soft label loss to
align classifier
scores across
domains
Domain confusion
loss
31
Domain confusion loss
Alternate optimization of two objectives (like adversarial training). First makes
domain classifier as good as possible. Standard binary cross entropy loss:
Second makes features as confusing as possible for the discriminator:
Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015.
32
Alignment of source and target predictions
Calculate per class
average softmax
activations in
source domain
Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015.
“bottle”
33
Alignment of source and target predictions
Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015.
Use these as the target
distribution for target
domain.
Minimizing cross entropy
loss same as minimizing
KL divergence!
34
Summary
● Possible to train very large models on small data by using transfer learning and
domain adaptation
● Off the shelf features work very well in various domains and tasks
● Lower layers of network contain very generic features, higher layers more task
specific features
● Supervised domain adaptation via fine tuning almost always improves
performance
● Possible to do unsupervised domain adaptation by matching feature
distributions
35
Questions?
36
Additional resources
● Lluis Castrejon, “Domain adaptation and zero-shot learning”. University of Toronto 2016.
● Hoffman, J., Guadarrama, S., Tzeng, E. S., Hu, R., Donahue, J., Girshick, R., ... & Saenko, K. (2014). LSDA: Large
scale detection through adaptation. NIPS 2014. (Slides by Xavier Giró-i-Nieto)
● Yosinski, Jason, Jeff Clune, Yoshua Bengio, and Hod Lipson. "How transferable are features in deep neural
networks?." In Advances in Neural Information Processing Systems, pp. 3320-3328. 2014.
● Shao, Ling, Fan Zhu, and Xuelong Li. "Transfer learning for visual categorization: A survey." Neural Networks and
Learning Systems, IEEE Transactions on 26, no. 5 (2015): 1019-1034.
● Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. "Net2Net: Accelerating Learning via Knowledge Transfer." ICLR
2016. [code] [Notes by Hugo Larrochelle
● Gani, Yaroslav, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario
Marchand, and Victor Lempitsky. "Domain-Adversarial Training of Neural Networks." arXiv preprint arXiv:1505.07818
(2015).
● [Hinton2015] Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9.
37

Más contenido relacionado

La actualidad más candente

GAN - Theory and Applications
GAN - Theory and ApplicationsGAN - Theory and Applications
GAN - Theory and ApplicationsEmanuele Ghelfi
 
Deep Learning in Computer Vision
Deep Learning in Computer VisionDeep Learning in Computer Vision
Deep Learning in Computer VisionSungjoon Choi
 
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...Universitat Politècnica de Catalunya
 
Introduction to Deep Learning
Introduction to Deep LearningIntroduction to Deep Learning
Introduction to Deep LearningOswald Campesato
 
Domain adaptation: A Theoretical View
Domain adaptation: A Theoretical ViewDomain adaptation: A Theoretical View
Domain adaptation: A Theoretical ViewChia-Ching Lin
 
Deep neural networks
Deep neural networksDeep neural networks
Deep neural networksSi Haem
 
Deep Generative Models
Deep Generative Models Deep Generative Models
Deep Generative Models Chia-Wen Cheng
 
Image Classification using deep learning
Image Classification using deep learning Image Classification using deep learning
Image Classification using deep learning Asma-AH
 
Variational Autoencoder
Variational AutoencoderVariational Autoencoder
Variational AutoencoderMark Chang
 
Final thesis presentation
Final thesis presentationFinal thesis presentation
Final thesis presentationPawan Singh
 
Convolution Neural Network (CNN)
Convolution Neural Network (CNN)Convolution Neural Network (CNN)
Convolution Neural Network (CNN)Basit Rafiq
 
Convolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep LearningConvolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep LearningMohamed Loey
 
Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...
Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...
Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...Universitat Politècnica de Catalunya
 
Transfer learning-presentation
Transfer learning-presentationTransfer learning-presentation
Transfer learning-presentationBushra Jbawi
 

La actualidad más candente (20)

GAN - Theory and Applications
GAN - Theory and ApplicationsGAN - Theory and Applications
GAN - Theory and Applications
 
Deep Learning in Computer Vision
Deep Learning in Computer VisionDeep Learning in Computer Vision
Deep Learning in Computer Vision
 
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
 
Introduction to Deep Learning
Introduction to Deep LearningIntroduction to Deep Learning
Introduction to Deep Learning
 
Domain adaptation: A Theoretical View
Domain adaptation: A Theoretical ViewDomain adaptation: A Theoretical View
Domain adaptation: A Theoretical View
 
Deep neural networks
Deep neural networksDeep neural networks
Deep neural networks
 
Deep learning
Deep learningDeep learning
Deep learning
 
Deep Generative Models
Deep Generative Models Deep Generative Models
Deep Generative Models
 
Image Classification using deep learning
Image Classification using deep learning Image Classification using deep learning
Image Classification using deep learning
 
Domain adaptation
Domain adaptationDomain adaptation
Domain adaptation
 
Machine Learning for Dummies
Machine Learning for DummiesMachine Learning for Dummies
Machine Learning for Dummies
 
Variational Autoencoder
Variational AutoencoderVariational Autoencoder
Variational Autoencoder
 
Final thesis presentation
Final thesis presentationFinal thesis presentation
Final thesis presentation
 
Transfer Learning (D2L4 Insight@DCU Machine Learning Workshop 2017)
Transfer Learning (D2L4 Insight@DCU Machine Learning Workshop 2017)Transfer Learning (D2L4 Insight@DCU Machine Learning Workshop 2017)
Transfer Learning (D2L4 Insight@DCU Machine Learning Workshop 2017)
 
Convolution Neural Network (CNN)
Convolution Neural Network (CNN)Convolution Neural Network (CNN)
Convolution Neural Network (CNN)
 
Convolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep LearningConvolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep Learning
 
Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...
Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...
Transfer Learning and Domain Adaptation (D2L3 2017 UPC Deep Learning for Comp...
 
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
 
Transfer learning-presentation
Transfer learning-presentationTransfer learning-presentation
Transfer learning-presentation
 
Resnet
ResnetResnet
Resnet
 

Similar a Transfer Learning and Domain Adaptation - Ramon Morros - UPC Barcelona 2018

Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...
Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...
Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...Universitat Politècnica de Catalunya
 
Resnet.pdf
Resnet.pdfResnet.pdf
Resnet.pdfYanhuaSi
 
Resnet.pptx
Resnet.pptxResnet.pptx
Resnet.pptxYanhuaSi
 
Learn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
Learn to Build an App to Find Similar Images using Deep Learning- Piotr TeterwakLearn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
Learn to Build an App to Find Similar Images using Deep Learning- Piotr TeterwakPyData
 
Recent Progress on Object Detection_20170331
Recent Progress on Object Detection_20170331Recent Progress on Object Detection_20170331
Recent Progress on Object Detection_20170331Jihong Kang
 
Understanding Convolutional Neural Networks
Understanding Convolutional Neural NetworksUnderstanding Convolutional Neural Networks
Understanding Convolutional Neural NetworksJeremy Nixon
 
Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry: Statisti...
Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry:Statisti...Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry:Statisti...
Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry: Statisti...Takayuki Nishio
 
Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...
Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...
Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...Universitat Politècnica de Catalunya
 
RunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkRunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkPutra Wanda
 
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)Universitat Politècnica de Catalunya
 
Artificial Intelligence for Vision: A walkthrough of recent breakthroughs
Artificial Intelligence for Vision:  A walkthrough of recent breakthroughsArtificial Intelligence for Vision:  A walkthrough of recent breakthroughs
Artificial Intelligence for Vision: A walkthrough of recent breakthroughsNikolas Markou
 
Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018
Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018
Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018Universitat Politècnica de Catalunya
 
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Universitat Politècnica de Catalunya
 
[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...
[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...
[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...Ziyuan Zhao
 
Semantic Segmentation on Satellite Imagery
Semantic Segmentation on Satellite ImagerySemantic Segmentation on Satellite Imagery
Semantic Segmentation on Satellite ImageryRAHUL BHOJWANI
 
Learning visual representation without human label
Learning visual representation without human labelLearning visual representation without human label
Learning visual representation without human labelKai-Wen Zhao
 
From Conventional Machine Learning to Deep Learning and Beyond.pptx
From Conventional Machine Learning to Deep Learning and Beyond.pptxFrom Conventional Machine Learning to Deep Learning and Beyond.pptx
From Conventional Machine Learning to Deep Learning and Beyond.pptxChun-Hao Chang
 
DLD meetup 2017, Efficient Deep Learning
DLD meetup 2017, Efficient Deep LearningDLD meetup 2017, Efficient Deep Learning
DLD meetup 2017, Efficient Deep LearningBrodmann17
 

Similar a Transfer Learning and Domain Adaptation - Ramon Morros - UPC Barcelona 2018 (20)

Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...
Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...
Deep Learning for Computer Vision: Transfer Learning and Domain Adaptation (U...
 
Resnet.pdf
Resnet.pdfResnet.pdf
Resnet.pdf
 
Resnet.pptx
Resnet.pptxResnet.pptx
Resnet.pptx
 
Learn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
Learn to Build an App to Find Similar Images using Deep Learning- Piotr TeterwakLearn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
Learn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
 
Recent Progress on Object Detection_20170331
Recent Progress on Object Detection_20170331Recent Progress on Object Detection_20170331
Recent Progress on Object Detection_20170331
 
Understanding Convolutional Neural Networks
Understanding Convolutional Neural NetworksUnderstanding Convolutional Neural Networks
Understanding Convolutional Neural Networks
 
Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry: Statisti...
Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry:Statisti...Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry:Statisti...
Tutorial @ IEEE ICC 2019 : Machine Learning and Stochastic Geometry: Statisti...
 
Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...
Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...
Life-long / Incremental Learning (DLAI D6L1 2017 UPC Deep Learning for Artifi...
 
RunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkRunPool: A Dynamic Pooling Layer for Convolution Neural Network
RunPool: A Dynamic Pooling Layer for Convolution Neural Network
 
Large Scale Distributed Deep Networks
Large Scale Distributed Deep NetworksLarge Scale Distributed Deep Networks
Large Scale Distributed Deep Networks
 
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
 
Artificial Intelligence for Vision: A walkthrough of recent breakthroughs
Artificial Intelligence for Vision:  A walkthrough of recent breakthroughsArtificial Intelligence for Vision:  A walkthrough of recent breakthroughs
Artificial Intelligence for Vision: A walkthrough of recent breakthroughs
 
Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018
Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018
Lifelong / Incremental Deep Learning - Ramon Morros - UPC Barcelona 2018
 
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
 
[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...
[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...
[BMVC 2022] DA-CIL: Towards Domain Adaptive Class-Incremental 3D Object Detec...
 
Mnist report
Mnist reportMnist report
Mnist report
 
Semantic Segmentation on Satellite Imagery
Semantic Segmentation on Satellite ImagerySemantic Segmentation on Satellite Imagery
Semantic Segmentation on Satellite Imagery
 
Learning visual representation without human label
Learning visual representation without human labelLearning visual representation without human label
Learning visual representation without human label
 
From Conventional Machine Learning to Deep Learning and Beyond.pptx
From Conventional Machine Learning to Deep Learning and Beyond.pptxFrom Conventional Machine Learning to Deep Learning and Beyond.pptx
From Conventional Machine Learning to Deep Learning and Beyond.pptx
 
DLD meetup 2017, Efficient Deep Learning
DLD meetup 2017, Efficient Deep LearningDLD meetup 2017, Efficient Deep Learning
DLD meetup 2017, Efficient Deep Learning
 

Más de Universitat Politècnica de Catalunya

The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...Universitat Politècnica de Catalunya
 
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-NietoTowards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-NietoUniversitat Politècnica de Catalunya
 
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...Universitat Politècnica de Catalunya
 
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in VideosGeneration of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in VideosUniversitat Politècnica de Catalunya
 
Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...Universitat Politècnica de Catalunya
 
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...Universitat Politècnica de Catalunya
 
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020Universitat Politècnica de Catalunya
 
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...Universitat Politècnica de Catalunya
 
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020Universitat Politècnica de Catalunya
 
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)Universitat Politècnica de Catalunya
 
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...Universitat Politècnica de Catalunya
 
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020Universitat Politècnica de Catalunya
 

Más de Universitat Politècnica de Catalunya (20)

Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
 
Deep Generative Learning for All
Deep Generative Learning for AllDeep Generative Learning for All
Deep Generative Learning for All
 
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
 
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-NietoTowards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
 
The Transformer - Xavier Giró - UPC Barcelona 2021
The Transformer - Xavier Giró - UPC Barcelona 2021The Transformer - Xavier Giró - UPC Barcelona 2021
The Transformer - Xavier Giró - UPC Barcelona 2021
 
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
 
Open challenges in sign language translation and production
Open challenges in sign language translation and productionOpen challenges in sign language translation and production
Open challenges in sign language translation and production
 
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in VideosGeneration of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
 
Discovery and Learning of Navigation Goals from Pixels in Minecraft
Discovery and Learning of Navigation Goals from Pixels in MinecraftDiscovery and Learning of Navigation Goals from Pixels in Minecraft
Discovery and Learning of Navigation Goals from Pixels in Minecraft
 
Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...
 
Intepretability / Explainable AI for Deep Neural Networks
Intepretability / Explainable AI for Deep Neural NetworksIntepretability / Explainable AI for Deep Neural Networks
Intepretability / Explainable AI for Deep Neural Networks
 
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
 
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
 
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
 
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
 
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
 
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
 
Curriculum Learning for Recurrent Video Object Segmentation
Curriculum Learning for Recurrent Video Object SegmentationCurriculum Learning for Recurrent Video Object Segmentation
Curriculum Learning for Recurrent Video Object Segmentation
 
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
 
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
 

Último

Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...
Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...
Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...nirzagarg
 
DATA SUMMIT 24 Building Real-Time Pipelines With FLaNK
DATA SUMMIT 24  Building Real-Time Pipelines With FLaNKDATA SUMMIT 24  Building Real-Time Pipelines With FLaNK
DATA SUMMIT 24 Building Real-Time Pipelines With FLaNKTimothy Spann
 
Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...
Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...
Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...nirzagarg
 
Harnessing the Power of GenAI for BI and Reporting.pptx
Harnessing the Power of GenAI for BI and Reporting.pptxHarnessing the Power of GenAI for BI and Reporting.pptx
Harnessing the Power of GenAI for BI and Reporting.pptxParas Gupta
 
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...Valters Lauzums
 
Reconciling Conflicting Data Curation Actions: Transparency Through Argument...
Reconciling Conflicting Data Curation Actions:  Transparency Through Argument...Reconciling Conflicting Data Curation Actions:  Transparency Through Argument...
Reconciling Conflicting Data Curation Actions: Transparency Through Argument...Bertram Ludäscher
 
Ranking and Scoring Exercises for Research
Ranking and Scoring Exercises for ResearchRanking and Scoring Exercises for Research
Ranking and Scoring Exercises for ResearchRajesh Mondal
 
Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...
Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...
Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...gajnagarg
 
Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...
Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...
Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...nirzagarg
 
In Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi Arabia
In Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi ArabiaIn Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi Arabia
In Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi Arabiaahmedjiabur940
 
Digital Transformation Playbook by Graham Ware
Digital Transformation Playbook by Graham WareDigital Transformation Playbook by Graham Ware
Digital Transformation Playbook by Graham WareGraham Ware
 
Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...
Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...
Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...Klinik kandungan
 
Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...
Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...
Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...gajnagarg
 
Discover Why Less is More in B2B Research
Discover Why Less is More in B2B ResearchDiscover Why Less is More in B2B Research
Discover Why Less is More in B2B Researchmichael115558
 
+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...
+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...
+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...Health
 
怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制
怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制
怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制vexqp
 
Capstone in Interprofessional Informatic // IMPACT OF COVID 19 ON EDUCATION
Capstone in Interprofessional Informatic  // IMPACT OF COVID 19 ON EDUCATIONCapstone in Interprofessional Informatic  // IMPACT OF COVID 19 ON EDUCATION
Capstone in Interprofessional Informatic // IMPACT OF COVID 19 ON EDUCATIONLakpaYanziSherpa
 
怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制
怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制
怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制vexqp
 

Último (20)

Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...
Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...
Top profile Call Girls In Bihar Sharif [ 7014168258 ] Call Me For Genuine Mod...
 
DATA SUMMIT 24 Building Real-Time Pipelines With FLaNK
DATA SUMMIT 24  Building Real-Time Pipelines With FLaNKDATA SUMMIT 24  Building Real-Time Pipelines With FLaNK
DATA SUMMIT 24 Building Real-Time Pipelines With FLaNK
 
Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...
Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...
Top profile Call Girls In Satna [ 7014168258 ] Call Me For Genuine Models We ...
 
Harnessing the Power of GenAI for BI and Reporting.pptx
Harnessing the Power of GenAI for BI and Reporting.pptxHarnessing the Power of GenAI for BI and Reporting.pptx
Harnessing the Power of GenAI for BI and Reporting.pptx
 
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
 
Cytotec in Jeddah+966572737505) get unwanted pregnancy kit Riyadh
Cytotec in Jeddah+966572737505) get unwanted pregnancy kit RiyadhCytotec in Jeddah+966572737505) get unwanted pregnancy kit Riyadh
Cytotec in Jeddah+966572737505) get unwanted pregnancy kit Riyadh
 
Reconciling Conflicting Data Curation Actions: Transparency Through Argument...
Reconciling Conflicting Data Curation Actions:  Transparency Through Argument...Reconciling Conflicting Data Curation Actions:  Transparency Through Argument...
Reconciling Conflicting Data Curation Actions: Transparency Through Argument...
 
Ranking and Scoring Exercises for Research
Ranking and Scoring Exercises for ResearchRanking and Scoring Exercises for Research
Ranking and Scoring Exercises for Research
 
Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...
Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...
Top profile Call Girls In dimapur [ 7014168258 ] Call Me For Genuine Models W...
 
Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...
Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...
Top profile Call Girls In Purnia [ 7014168258 ] Call Me For Genuine Models We...
 
In Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi Arabia
In Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi ArabiaIn Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi Arabia
In Riyadh ((+919101817206)) Cytotec kit @ Abortion Pills Saudi Arabia
 
Digital Transformation Playbook by Graham Ware
Digital Transformation Playbook by Graham WareDigital Transformation Playbook by Graham Ware
Digital Transformation Playbook by Graham Ware
 
Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...
Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...
Jual obat aborsi Bandung ( 085657271886 ) Cytote pil telat bulan penggugur ka...
 
Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...
Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...
Top profile Call Girls In Vadodara [ 7014168258 ] Call Me For Genuine Models ...
 
Discover Why Less is More in B2B Research
Discover Why Less is More in B2B ResearchDiscover Why Less is More in B2B Research
Discover Why Less is More in B2B Research
 
+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...
+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...
+97470301568>>weed for sale in qatar ,weed for sale in dubai,weed for sale in...
 
怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制
怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制
怎样办理旧金山城市学院毕业证(CCSF毕业证书)成绩单学校原版复制
 
Abortion pills in Doha Qatar (+966572737505 ! Get Cytotec
Abortion pills in Doha Qatar (+966572737505 ! Get CytotecAbortion pills in Doha Qatar (+966572737505 ! Get Cytotec
Abortion pills in Doha Qatar (+966572737505 ! Get Cytotec
 
Capstone in Interprofessional Informatic // IMPACT OF COVID 19 ON EDUCATION
Capstone in Interprofessional Informatic  // IMPACT OF COVID 19 ON EDUCATIONCapstone in Interprofessional Informatic  // IMPACT OF COVID 19 ON EDUCATION
Capstone in Interprofessional Informatic // IMPACT OF COVID 19 ON EDUCATION
 
怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制
怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制
怎样办理圣路易斯大学毕业证(SLU毕业证书)成绩单学校原版复制
 

Transfer Learning and Domain Adaptation - Ramon Morros - UPC Barcelona 2018

  • 1. [course site] #DLUPC Transfer learning and domain adaptation Day 5 Lecture 2 Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University Eric Arazo eric.arazosanchez@dcu.ie PhD Candidate Insight Centre for Data Analytics Dublin City University Ramon Morros ramon.morros@upc.edu Associate Professor Universitat Politecnica de Catalunya Technical University of Catalonia Many slides from:
  • 2. The ability to apply knowledge learned in previous tasks to novel tasks ● Based on human learning. People can often transfer knowledge learnt previously to novel situations ○ Play classic piano → Play jazz piano ○ Maths → Machine Learning ○ Ride motorbike → Drive a car Transfer Learning 2
  • 3. ● Isolated, single task learning: ○ Knowledge is not retained or accumulated. Learning is performed w.o. considering past learned knowledge in other tasks ● Learning of a new tasks relies on the previous learned tasks: ○ Learning process can be faster, more accurate and/or need less training data Traditional ML vs Transfer Learning Dataset 1 Learning System Task 1 Dataset 2 Learning System Task 2 Dataset 1 Dataset 2 Learning System Task 1 Knowledge Learning System Task 2 3
  • 4. Transfer learning in DL Myth: you can’t do deep learning unless you have a million labelled examples for your problem. Reality ● You can learn useful representations from unlabelled data ● You can train on a nearby surrogate objective for which it is easy to generate labels ● You can transfer learned representations from a related task 4
  • 5. Transfer learning: idea Instead of training a deep network from scratch for your task: ● Take a network trained on a different domain for a different source task ● Adapt it for your domain and your target task This lecture will talk about how to do this. Variations: ● Same domain, different task ● Different domain, same task 5 Source data E.g. ImageNet Source model Source labels Target data E.g. PASCAL Target model Target labels Transfer Learned Knowledge Large amount of data/labels Small amount of data/labels
  • 6. Example: PASCAL VOC 2007 ● Standard classification benchmark, 20 classes, ~10K images, 50% train, 50% test ● Deep networks can have many parameters (e.g. 60M in Alexnet) ● Direct training (from scratch) using only 5K training images can be problematic. Model overfits. ● How can we use deep networks in this setting? 6
  • 7. Transfer Learning Figure extracted from Sebastian Rudder’s blog: “Transfer Learning - Machine Learning's Next Frontier” http://ruder.io/transfer-learning/index.html 7
  • 8. A Domain consists of two components: ● Feature space: ● Marginal distribution: , For a given domain D, a Task is defined by two components: ● A label space: ● A predictive function , learned from feature vector/label pairs, , ● For each feature vector in the domain, predicts its corresponding label: Notation (I) 8
  • 9. Notation (II) 9 If two domains are different, they may have different feature spaces or different marginal distributions If two tasks are different, they may have different label spaces or different conditional distributions Digit 5, dataset A Digit 5, dataset B Dataset X Vladimir Barak Angela Lionel Dataset Y Nuria Garbiñe Different label space!
  • 10. Notation (III) 10 For simplicity, only two domains or two tasks are usually considered ● Source domain ● Task on the source domain ● Target domain ● Task on the target domain
  • 11. Domain adaptation Consider a classification task where is the input space and is the set of labels. Given two sets of samples drawn from the source and target domains: The goal of the learning algorithm is to build a classifier with a low target risk on the target domain 11 (Target space labeled data may not be present in unsupervised case)
  • 12. Domain biass ● Datasets are samples of the world ● In many cases, there is a shift or bias in the distributions of the source and target data representations 12 A. Torralba, A. Efros. Unbiased Look at Dataset Bias. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2011.
  • 13. Domain adaptation When there is a domain shift ● The size of this shift is often measured by the distance between source and target subspaces ● A typical approach is to learn a feature space transformation to align the source and target representations (reduce domain divergence) 13 Digit 5, dataset ADigit 5, dataset B Digit 5, dataset A Digit 5, dataset B Align
  • 14. “Off-the-shelf” Idea: use outputs of one or more layers of a network trained on a different task as generic feature detectors. Train a new shallow model on these features. Assumes that conv2 conv3 fc1 conv1 loss Data and labels (e.g. ImageNet) fc2 softmax TRANSFER Shallow classifier (e.g. SVM) conv2 conv3 fc1 conv1 Target data and labels features 14
  • 15. Off-the-shelf features Works surprisingly well in practice! Surpassed or on par with state-of-the-art in several tasks in 2014 Image classification: ● PASCAL VOC 2007 ● Oxford flowers ● CUB Bird dataset ● MIT indoors Image retrieval: ● Paris 6k ● Holidays ● UKBench Razavian et al, CNN Features off-the-shelf: an Astounding Baseline for Recognition, CVPRW 2014 http://arxiv.org/abs/1403.6382 15 Oxford 102 flowers dataset (Trained to perform object classification on ILSVRC13)
  • 16. Can we do better than off the shelf features? 16
  • 17. Fine-tuning: supervised domain adaptation Train deep net on “nearby” task for which it is easy to get labels using standard backprop ● E.g. ImageNet classification ● Pseudo classes from augmented data ● Slow feature learning, ego-motion Cut off top layer(s) of network and replace with supervised objective for target domain Fine-tune network using backprop with labels for target domain until validation loss starts to increase Aligns with labels conv2 conv3 fc1 conv1 surrogate loss surrogate data fc2 + softmax real labelsreal data real loss my_fc2 + softmax 17
  • 18. Freeze or fine-tune? Bottom n layers can be frozen or fine tuned. ● Frozen: not updated during backprop ● Fine-tuned: updated during backprop Which to do depends on target task: ● Freeze: target task labels are scarce, and we want to avoid overfitting ● Fine-tune: target task labels are more plentiful In general, we can set learning rates to be different for each layer to find a tradeoff between freezing and fine tuning labels conv2 conv3 fc1 conv1 loss data fc2 + softmax Finetunedfrozen LR = 0 LR > 0 18
  • 19. How transferable are features? Lower layers: more general features. Transfer very well to other tasks. Higher layers: more task specific. 19 data Matthew D Zeiler, Rob Fergus, “Visualizing and Understanding Convolutional Networks”, CVPR 2013
  • 20. 20 Yosinki et al. How transferable are features in deep neural networks. NIPS 2014. https://arxiv.org/abs/1411.1792 How transferable are features?
  • 21. How transferable are features? Transferability is negatively affected by two distinct issues: ● The specialization of higher layer neurons ● Optimization difficulties related to splitting networks between co-adapted neurons Fine-tuning improves generalization when sufficient examples are available. Transfer learning and fine tuning often lead to better performance than training from scratch on the target dataset. Even features transferred from distant tasks are often better than random initial weights! Yosinki et al. How transferable are features in deep neural networks. NIPS 2014. https://arxiv.org/abs/1411.1792 21
  • 22. Unsupervised domain adaptation Also possible to do domain adaptation without labels in target set. Y Ganin and V Lempitsky, Unsupervised Domain Adaptation by Backpropagation, ICML 2015 https://arxiv.org/abs/1409.7495 22 Y. Ganin et al. "Domain-Adversarial Training of Neural Networks." Journal of Machine Learning research 17 (2016). (Source domain) (Source and target domains)
  • 23. Unsupervised domain adaptation The approach promotes the emergence of "deep" features that are (i) discriminative for the main learning task on DS and (ii) invariant with respect to the shift between the domains Y Ganin and V Lempitsky, Unsupervised Domain Adaptation by Backpropagation, ICML 2015 https://arxiv.org/abs/1409.7495 23 loss for label prediction maximize the loss of the domain classifier (by making the two feature distributions as similar as possible) Minimize the loss of the domain classifier and the label predictor loss for the domain classification Y. Ganin,et al. "Domain-Adversarial Training of Neural Networks." Journal of Machine Learning research 17 (2016).
  • 24. Unsupervised domain adaptation Y Ganin and V Lempitsky, Unsupervised Domain Adaptation by Backpropagation, ICML 2015 https://arxiv.org/abs/1409.7495 24
  • 25. Task transfer Maximizing domain confusion does not necessarily align the classes in the target with those in the source Possible solution: Transfer the similarity structure amongst categories from the source to the target, using distillation (Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9) 25
  • 26. Original application was to transfer the knowledge from a large, easy to train model into a smaller/faster model more suitable for deployment Bucilua1 demonstrated that this can be done reliably when transferring from a large ensemble of models to a single small model Distillation 261 C.Bucilua, R. Caruana, and A. Niculescu-Mizil. “Model compression”. In ACMSIG KDD ’06, 2006
  • 27. Idea: use the class probabilities produced by the large model as “soft targets” for training the small model ○ The ratios of probabilities in the soft targets provide information about the learned function ○ These ratios carry information about the structure of the data ○ Train by replacing the hard labels with the softmax activations from the original large model Distillation Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9. 0.05 0.8 0.1 0.05 0 1 0 0 Yn Yn 0.09 0.05 0.85 0.01 0.4 0.1 0.2 0.3 Y0Y0 27 Distillation lossMultinomial logistic loss
  • 28. Distillation ● To increase the influence of non-target class probabilities in the cross entropy, the temperature of the final softmax is raised to “soften” the final probability distribution over classes ● Transfer can be obtained by using the same large model training set or a separate training set ● If the ground-truth labels of the transfer set are known, standard loss and distillation loss can be combined Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9. 28 0.09 0.05 0.85 0.01 0.15 0.10 0.70 0.05 T=1 T>1
  • 29. Semi-supervised task & domain adaptation When some labels are available in the target domain, then we can use these when doing domain adaptation. I.e. combine fine tuning and unsupervised domain adaptation. Tzeng et al. take this a step further and try to simultaneously optimize a loss that maximizes: 1. classification accuracy on both source and target datasets 2. domain confusion of a domain classifier 3. agreement of classifier score distributions across domains Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015. 29
  • 30. Semi-supervised domain adaptation Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015. 30
  • 31. Semi-supervised domain adaptation Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015. Classifier loss Soft label loss to align classifier scores across domains Domain confusion loss 31
  • 32. Domain confusion loss Alternate optimization of two objectives (like adversarial training). First makes domain classifier as good as possible. Standard binary cross entropy loss: Second makes features as confusing as possible for the discriminator: Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015. 32
  • 33. Alignment of source and target predictions Calculate per class average softmax activations in source domain Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015. “bottle” 33
  • 34. Alignment of source and target predictions Tzeng, Eric, et al. Simultaneous deep transfer across domains and tasks. ICCV. 2015. Use these as the target distribution for target domain. Minimizing cross entropy loss same as minimizing KL divergence! 34
  • 35. Summary ● Possible to train very large models on small data by using transfer learning and domain adaptation ● Off the shelf features work very well in various domains and tasks ● Lower layers of network contain very generic features, higher layers more task specific features ● Supervised domain adaptation via fine tuning almost always improves performance ● Possible to do unsupervised domain adaptation by matching feature distributions 35
  • 37. Additional resources ● Lluis Castrejon, “Domain adaptation and zero-shot learning”. University of Toronto 2016. ● Hoffman, J., Guadarrama, S., Tzeng, E. S., Hu, R., Donahue, J., Girshick, R., ... & Saenko, K. (2014). LSDA: Large scale detection through adaptation. NIPS 2014. (Slides by Xavier Giró-i-Nieto) ● Yosinski, Jason, Jeff Clune, Yoshua Bengio, and Hod Lipson. "How transferable are features in deep neural networks?." In Advances in Neural Information Processing Systems, pp. 3320-3328. 2014. ● Shao, Ling, Fan Zhu, and Xuelong Li. "Transfer learning for visual categorization: A survey." Neural Networks and Learning Systems, IEEE Transactions on 26, no. 5 (2015): 1019-1034. ● Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. "Net2Net: Accelerating Learning via Knowledge Transfer." ICLR 2016. [code] [Notes by Hugo Larrochelle ● Gani, Yaroslav, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky. "Domain-Adversarial Training of Neural Networks." arXiv preprint arXiv:1505.07818 (2015). ● [Hinton2015] Hinton, G., Vinyals, O., & Dean, J. “Distilling the Knowledge in a Neural Network”. NIPS 2014 DL Workshop, 1–9. 37