Speaker: Pierre Richemond, Data Science Institute of Imperial College
Title: Cutting edge generative models: Applications and implications
Abstract: This talk will examine recent developments in deep learning content generation at scale. Whether it be images or text, the latest methods have now reached a level of quality making it hard to discriminate between human- and AI-generated content. We will review recent examples of such generative models, and put their significance in a broader context, in light of such powerful tools’ potential for dual use.
Bio: Pierre is currently researching his PhD in deep reinforcement learning at the Data Science Institute of Imperial College. He also teaches Deep Learning at the Graduate School, and helps to run the Deep Learning Network and organises thematic reading groups. His background is in mathematics - he has studied electrical engineering at ENST, probability theory and stochastic processes at Universite Paris VI - Ecole Polytechnique, and business management at HEC.
4. Background : everything old is new again
• Advent of ’modern’ deep learning (resnets, batchnorm) : 2015
(arbitrary depth + large scale training)
• If stabilized, deeper is better.
• Invention of neural network training rule : 1970
(Linnainmaa’s masters thesis on backpropagation, applied to
neural networks)
2
6. Why does it work now ?
• In between : GPUs, fibre broadband, handheld megapixel
cameras, automatic differentiation software.
• All contributed to the explosion of applications we see today.
• Deep networks tend to work better and automatically find good
features for data classification & explanation on natural data:
images/video and sound.
• Other domains like text, structured data (chemistry...) can work
too with careful network design.
4
7. Deep Learning breakthroughs - 2014
• VGG architecture
• DeepDream
• First formulation of GANs
• First formulation of VAEs
• Black-box Variational Inference
• Neural Turing Machines
• Attention mechanisms
5
8. Deep Learning breakthroughs - 2015
• Residual Networks
• Batch Normalization
• ADAM optimizer
• ELU activation function
• Neural style transfer
• Graph convolutional networks
• Visual question answering (Karpathy)
• Normalizing Flows
• Deep Q-learning on Atari games
• Keras, TensorFlow
6
10. Deep Learning breakthroughs - 2017
• Wasserstein GAN
• Progressive Growing of GANs
• Capsule Networks
• Restarted SGD - cyclical learning rates
• AlphaGo Zero (MCTS as policy improvement ; no human
knowledge involved)
• Distributional Reinforcement Learning
• Equivalence of softmax Q-learning and entropic policy gradients
• TensorFlow distributions, Pyro
8
11. Deep Learning breakthroughs - 2018
• Spectral Normalization for GANs
• Fast initialization of convolutional neural networks
• Deep Video Portraits
• Differentiable neural architecture search
• OpenAI Five
• Video-to-video synthesis
• StyleGAN
9
12. A success story : Generative models
• Generative modelling aims at automatically explain features,
and generate new instances, of datapoints in a dataset
• Two main approaches that both use 2 different neural networks
instead of 1.
• Applications of the field to deep learning are 4 years old
• A success story of engineering
10
13. Generative models : GANs
Figure 1: Generative Adversarial Networks (GANs). A generator
(counterfeinter) and a discriminator (police) network play an adversarial
game, whose Nash equilibrium is perfect replication of the data distribution.
11
14. Generative models : VAEs
Figure 2: A typical variational autoencoder (VAE) architecture. An encoder
pushes input data through a (low) dimensionality bottleneck, that learns
relevant features in a latent code (optimized probabilistically). The decoder
network attempts to reconstruct inputs.
12
16. Illustrations - 2016
Figure 4: Interpolations varying the strength of a smile vector computed by
doing latent space averaging of labeled pictures. Courtesy of Tom White.
14
25. From fake news to un-real people
www.thispersondoesnotexist.com
A blog post by Kyle McDonald on how to detect fake portraits:
21
26. Some photorealistic transformations - 2018
Deep Video Portraits (Portrait reenactment)
Vid2Vid (Photorealistic video-to-video translation)
See also TacoTron for realistic speech generation.
22
27. OpenAI GPT-2 - Principle
• A generative model trained at scale
• A relatively ’simple’ architecture and prediction concept, but the
scaling makes all the difference (8 million webpages training set)
• While it can be used in a multitask context ( question answering,
text summarization, translation...), it really shines in text
generation.
• For the first time, OpenAI has declined releasing the weights of
the full trained model, citing risks
• This has caused an unprecedented stir in the (self-regulated)
machine learning research community
23
28. OpenAI GPT-2 - Coherent paragraph generation - 1
When initalized with the following prompt text :
’In a shocking finding, scientist discovered a herd of unicorns living
in a remote, previously unexplored valley, in the Andes Mountains.
Even more surprising to the researchers was the fact that the
unicorns spoke perfect English.’
the model generated as a follow-up...
24
29. OpenAI GPT-2 - Coherent paragraph generation - 2
’ “The scientist named the population, after their distinctive horn,
Ovid’s Unicorn. These four-horned, silver-white unicorns were
previously unknown to science. Now, after almost two centuries, the
mystery of what sparked this odd phenomenon is finally solved.
Dr. Jorge Pérez, an evolutionary biologist from the University of La
Paz, and several companions, were exploring the Andes Mountains
when they found a small valley, with no other animals or humans.
Pérez noticed that the valley had what appeared to be a natural
fountain, surrounded by two peaks of rock and silver snow.
Pérez and the others then ventured further into the valley. ‘By the
time we reached the top of one peak, the water looked blue, with
some crystals on top,’ said Pérez.
Pérez and his friends were astonished to see the unicorn herd. These
creatures could be seen from the air without having to move too
much to see them – they were so close they could touch their horns.”’
25
30. OpenAI GPT-2 : Test 1 - Neural news generation
GPT2 seed ’DeepMind should have been a UK champion says
ex-Google CFO who took it to America’:
Google executives Mark Zuckerberg and Sergey Brin,formerly head of
the social media giant, have been accused of lying to the US
government about the extent of their ties to the Kremlin.
26
31. OpenAI GPT-2 : Test 2 - Scientific Generation !
’We compute a stochastic scaling function for data with a weighted
average of the mean as the value for which to apply an Likert
distribution over the data in the gradient descent. We then evaluate
the prediction accuracy and find that the weights are not completely
uniform in this way. Instead, the weight of this residual model is a
weighted average of the data in the gradient descent on the average,
which is also very accurate.’
27
32. Towards post-truth economics
• Shattering the realism barrier of the uncanny valley
• Legislation will be required (don’t hold your breath)
• Content industries could be notably disrupted
• Is this a serious use case for blockchain tech ?
Is the marginal cost of content creation going to zero ?
28
34. Are we at AGI yet ?
3 different views :
1. Killer robots are coming, let’s ban research !
2. AI is just matrix multiplication...
3. AI/Deep Learning is in fact letting an optimization algorithm
automatically write code for you
29
35. ’Science sans conscience’ : TBC...
Figure 11: Human DNA data embedding and clustering (UMAP), categorized
by ethnicity. BiorXiv 2018, ’Revealing multi-scale population structure in
large cohorts’.
30