Domain adaptation for deep learning

Similar documents
Two at Once: Enhancing Learning and Generalization Capacities via IBN-Net

Deep Domain Adaptation by Geodesic Distance Minimization

arxiv: v1 [cs.lg] 25 Jul 2017

Importance Reweighting Using Adversarial-Collaborative Training

Visual meta-learning for planning and control

Asaf Bar Zvi Adi Hayat. Semantic Segmentation

Modular Vehicle Control for Transferring Semantic Information Between Weather Conditions Using GANs

Wasserstein Distance Guided Representation Learning for Domain Adaptation

Unsupervised Domain Adaptation with Distribution Matching Machines

Fine-grained Classification

Beyond Spatial Pyramids

Predicting Deeper into the Future of Semantic Segmentation Supplementary Material

Unsupervised Domain Adaptation with Residual Transfer Networks

Transferable Attention for Domain Adaptation

Towards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks

Kernel Density Topic Models: Visual Topics Without Visual Words

A Little History of Machine Learning

Course Structure. Psychology 452 Week 12: Deep Learning. Chapter 8 Discussion. Part I: Deep Learning: What and Why? Rufus. Rufus Processed By Fetch

UNSUPERVISED LEARNING

CS 188: Artificial Intelligence Spring Announcements

Tasks ADAS. Self Driving. Non-machine Learning. Traditional MLP. Machine-Learning based method. Supervised CNN. Methods. Deep-Learning based

Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks

Kai Yu NEC Laboratories America, Cupertino, California, USA

Very Deep Residual Networks with Maxout for Plant Identification in the Wild Milan Šulc, Dmytro Mishkin, Jiří Matas

Towards Fully-automated Driving

Neural networks and optimization

Deep Learning for Computer Vision

Tutorial on Methods for Interpreting and Understanding Deep Neural Networks. Part 3: Applications & Discussion

Announcements. CS 188: Artificial Intelligence Spring Classification. Today. Classification overview. Case-Based Reasoning

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

Analyzing the Performance of Multilayer Neural Networks for Object Recognition

Jakub Hajic Artificial Intelligence Seminar I

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines

Distinguish between different types of scenes. Matching human perception Understanding the environment

Anticipating Visual Representations from Unlabeled Data. Carl Vondrick, Hamed Pirsiavash, Antonio Torralba

Grundlagen der Künstlichen Intelligenz

Deep learning on 3D geometries. Hope Yao Design Informatics Lab Department of Mechanical and Aerospace Engineering

CS 188: Artificial Intelligence. Outline

Feature Design. Feature Design. Feature Design. & Deep Learning

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

CS 188: Artificial Intelligence Spring Announcements

GENERALIZING ACROSS DOMAINS VIA CROSS-GRADIENT TRAINING

SELF-SUPERVISION. Nate Russell, Christian Followell, Pratik Lahiri

OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS

Hidden CRFs for Human Activity Classification from RGBD Data

TUTORIAL PART 1 Unsupervised Learning

arxiv: v1 [cs.cv] 30 Nov 2018

Toward Correlating and Solving Abstract Tasks Using Convolutional Neural Networks Supplementary Material

Singing Voice Separation using Generative Adversarial Networks

Lecture 14: Deep Generative Learning

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

Supplementary Material for Superpixel Sampling Networks

Convolutional Neural Networks

Introduction to Deep Neural Networks

Laconic: Label Consistency for Image Categorization

Self-Tuning Semantic Image Segmentation

Probabilistic and Logistic Circuits: A New Synthesis of Logic and Machine Learning

CS230: Lecture 9 Deep Reinforcement Learning

Compressed Fisher vectors for LSVR

A Unified Framework for Domain Adaptation using Metric Learning on Manifolds

arxiv: v4 [cs.lg] 29 Dec 2018

Introduction to Machine Learning

Neural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /9/17

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

CS230: Lecture 5 Advanced topics in Object Detection

Generalized Zero-Shot Learning with Deep Calibration Network

Deep Learning Basics Lecture 7: Factor Analysis. Princeton University COS 495 Instructor: Yingyu Liang

Generative Adversarial Networks

CS 343: Artificial Intelligence

Introduction to Convolutional Neural Networks (CNNs)

CS 4700: Foundations of Artificial Intelligence

Transfer Learning through Greedy Subset Selection

Loss Functions and Optimization. Lecture 3-1

How to do backpropagation in a brain

Cross-Domain Sentiment Classification with Target Domain Specific Information

Improved Local Coordinate Coding using Local Tangents

An EoS-meter of QCD transition from deep learning

CSCI 315: Artificial Intelligence through Deep Learning

Understanding How ConvNets See

Approximate Q-Learning. Dan Weld / University of Washington

Deep Learning (CNNs)

Lecture 12. Talk Announcement. Neural Networks. This Lecture: Advanced Machine Learning. Recap: Generalized Linear Discriminants

Training Generative Adversarial Networks Via Turing Test

Classification Semi-supervised learning based on network. Speakers: Hanwen Wang, Xinxin Huang, and Zeyu Li CS Winter

Artificial Intelligence

Approximation Methods in Reinforcement Learning

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Multi-Label Learning with Weak Label

Spatial Transformer Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

GENERATIVE ADVERSARIAL LEARNING

Image Analysis. PCA and Eigenfaces

Enforcing constraints for interpolation and extrapolation in Generative Adversarial Networks

Two transfer learning approaches for domain adaptation

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji

The Success of Deep Generative Models

DANIEL WILSON AND BEN CONKLIN. Integrating AI with Foundation Intelligence for Actionable Intelligence

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

Fantope Regularization in Metric Learning

Transcription:

What you saw is not what you get Domain adaptation for deep learning Kate Saenko

Successes of Deep Learning in AI A Learning Advance in Artificial Intelligence Rivals Human Abilities Deep Learning for self-driving cars Translate FLOWER FIELD SKY CLOUDS Google's DeepMind Masters Atari Games Face Recognition 2

So is AI solved? pedestrian detection FAIL https://www.youtube.com/watch?v=w2pwxv8rfku 3

Major limitation of deep learning Not data efficient: Learning requires millions of labeled examples, models do not generalize well to new domains; not like humans! New domain?

What you saw is not what you get What your net is trained on Dataset Bias Domain Shift Domain Adaptation Domain Transfer What it s asked to label

Example: scene segmentation sky building road building people road Train on Cityscapes, Test on Cityscapes FCNs in the Wild: Pixel-level Adversarial and Constraint-based Adaptation, Judy Hoffman, Dequan Wang, Fisher Yu, Trevor Darrell, Arxiv 2016

Domain shift: Cityscapes to SF FCNs in the Wild: Pixel-level Adversarial and Constraint-based Adaptation, Judy Hoffman, Dequan Wang, Fisher Yu, Trevor Darrell, Arxiv 2016 tree car sign road car tree Train on Cityscapes, Test on San Francisco Dashcam

No tunnels in CityScapes?... FCNs in the Wild: Pixel-level Adversarial and Constraint-based Adaptation, Judy Hoffman, Dequan Wang, Fisher Yu, Trevor Darrell, Arxiv 2016 tree building

Applications to different types of domain shift From dataset to dataset From RGB to depth From simulated to real control From CAD models to real images

Today Show that deep models can be adapted without labels Propose two deep adaptation methods: adversarial alignment correlation alignment Show applications

Background: Domain Adaptation from source to target distribution Adapt backpack chair bike?? bike Source Domain lots of labeled data Target Domain unlabeled or limited labels?

Background: unsupervised domain adaptation Sample re-weighting NO labels in target domain Roughly, three categories of methods Sample re-weighting Subspace matching Deep methods Subspace alignment Deep alignment B. Femando, A. Habrard, M. Sebban, and T. Tuytelaars. Unsupervised visual domain adaptation using subspace alignment. In ICCV, 2013. B. Gong, Y. Shi, F. Sha, and K. Grauman. Geodesic flow kernel for unsupervised domain adaptation. In CVPR, 2012 Y. Ganin and V. Lempitsky. Unsupervised domain adaptation by back propagation. In ICML 2015 13

How to adapt a deep network? backpack chair bike conv1 conv5 fc 6 fc 7 fc 8 classification loss Source Data

How to adapt a deep network? backpack chair bike conv1 conv5 fc 6 fc 7 fc 8 classification loss Source Data backpack Target Data Applying source classifier to target domain can yield inferior performance?

How to adapt a deep network? IDEA: align feature distributions backpack chair bike conv1 conv5 fc 6 fc 7 source data fc 8 Source Data shared shared shared shared shared classification loss backpack? Target Data same conv1 conv5 fc 6 fc 7 labeled target data Fine tune?..zero or few labels in target domain Siamese network?..no paired / aligned instance examples! fc 8

Deep distribution alignment by minimizing distance between distributions, e.g. Maximum Mean Discrepancy M. Long, et al. ICML 2015 CORrelation ALignment Sun and Saenko, AAAI 2016 or by adversarial domain alignment, e.g. Domain Confusion E. Tzeng et al. ICCV 2015 Reverse Gradient Y. Ganin and V. Lempitsky ICML 2015 17

Adversarial Domain Adaptation Eric Tzeng UC Berkeley Judy Hoffman UC Berkeley Trevor Darrell UC Berkeley 18

Adversarial networks

Adversarial networks P P Q Q

Adversarial domain adaptation backpack chair bike Source Data + Labels conv1 conv5 fc 6 fc 7 Classifier classification loss Unlabeled Target Data conv1 conv5 fc 6 fc 7?

Adversarial domain adaptation backpack chair bike Source Data + Labels Encoder Classifier classification loss can be shared Unlabeled Target Data Encoder?

Adversarial domain adaptation backpack chair bike Source Data + Labels Encoder Classifier classification loss can be shared Unlabeled Target Data Encoder Discriminator Adversarial loss?

Adversarial domain adaptation backpack chair bike Source Data + Labels Encoder Classifier classification loss can be shared Unlabeled Target Data Encoder Discriminator Adversarial loss?

Design choices in adversarial adaptation backpack chair bike Source Data + Labels Generative or discriminative? Unlabeled Target Data Encoder Shared or not? Encoder Classifier Discriminator classification loss confusion Which loss? Adversarial loss? Ming-Yu Liu and Oncel Tuzel. Coupled generative adversarial networks, NIPS 2016

Deep domain confusion [Tzeng ICCV15 ] backpack chair bike conv1 conv5 fc 6 fc 7 source data fc 8 Source Data Unlabeled Target Data shared shared shared shared fc D domain confusion loss domain classifier loss shared classification loss? conv1 conv5 fc 6 fc 7 labeled target data fc 8 Adversarial Training of domain label predictor and domain confusion loss: Domain Label Cross-entropy with uniform distribution

Deep domain confusion [Tzeng ICCV15 ] Train a network to minimize classification loss AND confuse two domains source inputs target inputs network parameters (fixed) domain classifier (learn) domain classifier loss domain classifier prediction domain classifier (fixed) network parameters (learn) = pp(yy DD = 1 xx) domain confusion loss iterate (cross-entropy with uniform distribution) 27

Applications to different types of domain shift From dataset to dataset From RGB to depth From simulated to real control From CAD models to real images

ImageNet adapted to Caltech [Tzeng ICCV15 ] Multiclass Accuracy 0 40 120 200 Number of labeled target examples

FCNs in the Wild: Pixel-level Adversarial and Constraint-based Adaptation, Judy Hoffman, Dequan Wang, Fisher Yu, Trevor Darrell, Arxiv 2016 Results on Cityscapes to SF adaptation Before domain confusion After domain confusion

Adversarial Loss Functions Confusion loss [Tzeng 2015] Minimax loss [Ganin 2015] GAN loss [Goodfellow 2014] stronger gradients

Adversarial Discriminative Domain Adaptation (ADDA) (in submission) backpack chair bike Source Data + Labels Generative or discriminative? Unlabeled Target Data Encoder Shared or not? Encoder Classifier Discriminator classification loss Which loss? Adversarial loss GAN?

ADDA: Adaptation on digits (in submission)

Applications to different types of domain shift From dataset to dataset From RGB to depth From simulated to real control From CAD models to real images

ADDA: Adaptation on RGB-D (in submission) Train on RGB Test on depth

ADDA: Adaptation on RGB-D (in submission)

Applications to different types of domain shift From dataset to dataset From RGB to depth From simulated to real control From CAD models to real images

Adapting Deep Visuomotor Representations with Weak Pairwise Constraints Eric Tzeng1, Coline Devin1, Judy Hoffman1, Chelsea Finn1, Pieter Abbeel1, Sergey Levine1, Kate Saenko2, Trevor Darrell1 1 University of California, Berkeley 2 Boston University

From simulation to real world control [Tzeng, Devin, et al 16]

Weak pairwise constraints [Tzeng, Devin, et al 16]

Robotic task: place rope on scale [Tzeng, Devin, et al 16]

Applications to different types of domain shift From dataset to dataset From RGB to depth From simulated to real control From CAD models to real images

Domain Adaptation via Correlation Alignment Baochen Sun Microsoft Xingchao Peng Boston University 49

Deep CORAL: Correlation Alignment for Deep Domain Adaptation [Sun 2016]

Deep CORAL: Correlation Alignment for Deep Domain Adaptation [Sun 2016]

Generative CORAL Network (in submission)

Synthetic to real adaptation for object recognition (in submission) Train on synthetic Test on real

Synthetic to Real Adaptation with Deep Generative Correlation Alignment Networks (in submission)

Synthetic to Real Adaptation with Deep Generative Correlation Alignment Networks (in submission)

Synthetic to Real Adaptation with Deep Generative Correlation Alignment Networks (in submission)

Summary Deep models can be adapted to new domains without labels Proposed two deep feature alignment methods: adversarial alignment correlation alignment Many potential applications

Thank you References Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko, Simultaneous Deep Transfer Across Domains and Tasks, ICCV 2015 Eric Tzeng, Coline Devin, Judy Hoffman, Chelsea Finn, Pieter Abbeel, Sergey Levine, Kate Saenko, Trevor Darrell, Adapting Deep Visuomotor Representations with Weak Pairwise Constraints, WAFR 2016 Baochen Sun, Jiashi Feng, Kate Saenko, Return of Frustratingly Easy Domain Adaptation, AAAI 2016 Baochen Sun, Kate Saenko, Deep CORAL: Correlation Alignment for Deep Domain Adaptation, TASK-CV Workshop at ICCV 2016 Adversarial Discriminative Domain Adaptation, in submission Synthetic to Real Adaptation with Deep Generative Correlation Alignment Networks, in submission