Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I...

16
Introduction Deep Learning in Practice Guillaume Charpiat Victor Berger & Wenzhuo Liu TAU team, LRI, Paris-Sud / INRIA Saclay . . . and guests! G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud Deep Learning in Practice

Transcript of Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I...

Page 1: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Deep Learning in Practice

Guillaume CharpiatVictor Berger & Wenzhuo Liu

TAU team, LRI, Paris-Sud / INRIA Saclay

. . . and guests!

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 2: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Overview

I Course summary and organization

I Chapters overview

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 3: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Context

I Deep learning: impressive results in the machine learning literature

I yet difficult to train, and still poorly understood;results = black-boxes missing explanations.

I Huge societal impact of ML today (assistance in medicine, hiring process,bank loans...)=⇒ explain their decisions, offer guarantees?

I Real world problems: usually do not fit standard academic assumptions(data quantity and quality, expert knowledge availability...).

I This course: aims at providing insights and tools to address thesepractical aspects, based on mathematical concepts and practical exercises.

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 4: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Organisation and evaluation

I Most courses: a lesson + practical exercises (evaluated)

I Extras: guest talks, Jean Zay visit (1000 GPUs cluster), . . .

Schedule8 classes of 3 hours, most often on Monday mornings (9h – 12h15 with abreak) at CentraleSupelec (not every week, check the webpage for details).

Webpage & mailing-list: https://www.lri.fr/~gcharpia/deeppractice/

Prerequisite

I The introduction to Deep Learning course by Vincent Lepetit (1stsemester)

I Notions in information theory, Bayesian statistics, analysis, differentialcalculus

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 5: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Links with other Deep Learning coursesI Introduction to Deep Learning (V. Lepetit) : prerequisite

I L’apprentissage par reseaux de neurones profonds (S. Mallat)

I Fondements Theoriques du deep learning (F. Malgouyres & al)

I Apprentissage Profond pour la Restauration et la Synthese d’Images(A. Almansa & al)

I Modelisation en neurosciences et ailleurs (J-P Nadal)

I Object recognition and computer vision (Willow team & al)

I Our course: understanding and tools to make NN work in practicewith a focus on architecture design, explainability, societal impact, realdatasets and tasks (e.g. small data, limited computational powervs. scaling up, RL...).=⇒ negligible overlap

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 6: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Outline

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 7: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Deep learning vs. classical ML and optimization – January 13th

I Going Deep or not?I Examples of successes and failures of deep learning vs. classical

techniques (random forests)I Approximation theorems vs. generalization [3, 4]I Why deep: ex. of depth vs. layer size compromises (explicit bounds)

I Gap between classical Machine Learning and Deep LearningI Forgotten Machine Learning basics (Minimum Description Length

principle, regularizers, objective function different from evaluationcriterion) and incidental palliatives (drop-out, early stopping, noise)

I Hyper-parametersand training basics

I List of practical tricks

I Practical session (exercise to give before earlyFebruary)(bring your laptop!)

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 8: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Architectures – February 3rd

I Architectures as priors on function space

I Change of design paradigmI Random initialization

I Architecture zoo

I Reminder (CNN, auto-encoder, LSTM, adversarial...)I Dealing with scale & resolution (fully-convolutional, U-nets,

pyramidal approaches...)I Dealing with depth (ResNet, Highway networks) and mixing blocks

(Inception)I Attention mechanismsI GraphCNN

I Problem modeling

I Guest talk by Yaroslav Nikulin (start-up Therapixel): Deep learning forbreast cancer detection

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 9: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Interpretability – February 10th

I At stake: the example of medical diagnosis, and societal issues withblack-box algorithms [5]

I Interpretability of neural networksI Analyzing the black-box

I at the neuron level: filter visualisation, impact analysisI at the layer level: layer statistics...I at the net level: low-dimensional representation (t-SNE) + IBI by sub-task design: “explainable AI”

I Adversarial examples & remedies

I Issues with datasetsI Biases in datasets : 4 definitions of fairness

I Getting invariant to undesirable dataset biases (e.g. gender inCVs / job offers matching)

I Ensuring errors are uniform over the datasetI Differential privacy (database client protection)

I Visualization toolsG. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 10: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Small data, weak supervision and robustness

I Modes of supervision

I Learning from synthetic data [12]I Learning from scratch vs. Transfer learningI Semi-supervised learning [11]I Self-supervised learning (ex: video prediction)I Multi-tasking

I Exploiting known invariances or priors

I Example of permutation invariance: “deep sets” [8], applied topeople genetics

I Spatial/Temporal coherenceI Choosing physically meaningful metrics, e.g. optimal transport

(Sinkhorn approximation)[9]I Dealing with noisy supervision (noisy labels)

I Transfer learning

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 11: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Incorporating physical knowledge / Learning physics

I Course featuring Michele Alessandro Bucci (LRI, TAU team) and Lionel

Mathelin (LIMSI, Paris-Sud)I Data assimilationI Learning a PDE (equation not known)I Incorporating invariances/symmetries of the problemI Knowing an equation that the solution has to satisfy: solving PDEs!I Form of the solution knownI Deep for physic dynamics : learning and controlling the dynamics

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 12: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Deep Reinforcement Learning

I Guest course by Olivier Teytaud (Facebook FAIR Paris)

I Crash-course about deep RL. . .

I . . . until alpha-0!

GPU clusters

I Presentation of Jean Zay, the GPU super-cluster for French academia, byIDRIS

I Optional visit to the cluster and practical session (job scheduler, etc.)

Generative models & Variation inference

I GAN and VAE (Variational Auto-Encoder)

I GAN vs. VAE

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 13: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Auto-DeepLearning – March 16th

I Guest talk by Zhengying Liu (LRI, TAU team, Isabelle Guyon’s group)

I Overview of recent approaches for automatic hyper-parameter tuning(architecture, learning rate, etc.): classical blackbox optimisation,Reinforcement Learning approaches, constrained computational timebudget, self-adaptive architectures...

I Additional real-world difficulties: missing data, unstructured data

I Presentation of the Auto-ML & Auto-DL challenges

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 14: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

To attend the course

I go see the website and subscribe to the mailing-list

I bring your laptop, and install PyTorch, Jupyter andmatplotlib beforehand!

I See you on Monday at 9h, amphi Janet (CentraleSupelec)

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 15: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Biographies

I Guillaume Charpiat is an INRIA researcher in the TAU team (INRIASaclay/LRI/Paris-Sud). He has worked mainly in computer vision,optimization and machine learning, and now focuses on deep learning. Heconducts studies on neural networks both in theory (self-adaptivearchitectures, formal proofs) and in applications (remote sensing, peoplegenetics, molecular dynamics simulation, brain imagery, weather forecast,skin lesion medical diagnosis, ...).

I Victor Berger and Wenzhuo Liu are PhD students in the TAU team,working on deep generative models and on graph neural networks.

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice

Page 16: Deep Learning in Practice - lri.frgcharpia/deeppractice/2020/... · Introduction Overview Context I Deep learning: impressive results in the machine learning literature I yet di cult

Introduction

Overview

Bibliographies

Why does deep and cheap learning work so well?, Henry W.Lin, Max Tegmark, David Rolnick.https://arxiv.org/abs/1608.08225

Representation Benefits of Deep Feedforward Networks, MatusTelgarsky. https://arxiv.org/abs/1509.08101

On the structure of continuous functions of several variables,David A. Sprecher.

Representation properties of networks: Kolmogorov’s theoremis irrelevant, Federico Girosi and Tomaso Poggio.

Weapons of Math Destruction, Cathy O’Neil.

Practical Variational Inference for Neu-ral Networks, Alex Graves. https://papers.nips.cc/paper/4329-practical-variational-inference-for-neural-networks

Auto-Encoding Variational Bayes, Diederik P. Kingma andMax Welling. https://arxiv.org/abs/1312.6114

Deep Sets, Manzil Zaheer, Satwik Kottur, SiamakRavanbakhsh, Barnabas Poczos, Ruslan Salakhutdinov,Alexander Smola. https://arxiv.org/abs/1703.06114

Learning Generative Models with Sinkhorn Divergences, AudeGenevay, Gabriel Peyre, Marco Cuturi.https://arxiv.org/abs/1706.00292

Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour,Priya Goyal, Piotr Dollar, Ross Girshick, Pieter Noordhuis,Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, YangqingJia, Kaiming He. https://arxiv.org/abs/1706.02677

Temporal Ensembling for Semi-Supervised Learning, SamuliLaine, Timo Aila. https://arxiv.org/abs/1610.02242

Learning from Simulated and Unsupervised Images throughAdversarial Training, Ashish Shrivastava, Tomas Pfister, OncelTuzel, Josh Susskind, Wenda Wang, Russ Webb.https://arxiv.org/abs/1612.07828

https://himanshusahni.github.io/2018/02/23/

reinforcement-learning-never-worked.html

G. Charpiat TAU team, INRIA Saclay / LRI - Paris-Sud

Deep Learning in Practice