A Video from Google DeepMind.

Size: px
Start display at page:

Download "A Video from Google DeepMind."

Transcription

1 A Video from Google DeepMind

2 Can machine learning teach us cluster updates? Lei Wang Institute of Physics, CAS Li Huang and LW, LW,

3 Local vs Cluster algorithms

4 Local vs Cluster algorithms is slower than

5 Local vs Cluster algorithms Algorithmic innovation outperforms Moore s law!

6 Recommender Systems Learn preferences Recommendations

7 Recommender Systems Learn preferences Recommendations

8 Restricted Boltzmann Machine Smolensky 1986 Hinton and Sejnowski 1986 NX M X N X MX E (x, h) = a i x i b j h j x i W ij h j i=1 j=1 i=1 j=1 Energy based model h 2 {0, 1} M h 1 h 2 h 3... hm x 1 x 2... x N x 2 {0, 1} N

9 Restricted Boltzmann Machine Smolensky 1986 Hinton and Sejnowski 1986 NX M X N X MX E (x, h) = a i x i b j h j x i W ij h j i=1 j=1 i=1 j=1 Energy based model h 2 {0, 1} M h 1 h 2 h 3... hm x 1 x 2... x N x 2 {0, 1} N

10 Restricted Boltzmann Machine Smolensky 1986 Hinton and Sejnowski 1986 NX M X N X MX E (x, h) = a i x i b j h j x i W ij h j i=1 j=1 i=1 j=1 Energy based model h 2 {0, 1} M Probabilities p(x, h) e E(x,h) h 1 h 2 h 3... hm p(x) = X h p(x, h) x 1 x 2... x N x 2 {0, 1} N p(h x) =p(x, h)/p(x)

11 Restricted Boltzmann Machine Smolensky 1986 Hinton and Sejnowski 1986 NX M X N X MX E (x, h) = a i x i b j h j x i W ij h j i=1 j=1 i=1 j=1 Energy based model h 2 {0, 1} M Probabilities p(x, h) e E(x,h) h 1 h 2 h 3... hm p(x) = X h p(x, h) x 1 x 2... x N x 2 {0, 1} N p(h x) =p(x, h)/p(x) Universal approximator of probability distributions. Freund and Haussler, Le Roux and Bengio, 2008

12 Generative Learning Learning: Fit to the target distribution p(x) (x) target probability Learn the model from data Hinton 2002 Generating: Blocked Gibbs sampling h p(h x) x p(x 0 h) x 0 Generate more data from the learned model

13 MNIST database of handwritten digits learned weight W ij Theano deep learning tutorial

14 MNIST database of handwritten digits learned weight W ij h 1 h 2 h 3... hm x 1 x 2... x N Theano deep learning tutorial

15 MNIST database of handwritten digits learned weight W ij h 1 h 2 h 3... hm x 1 x 2... x N Theano deep learning tutorial

16 Challenge Which one is written by human? Salakhutdinov, 2015 Deep learning summer school

17 Idea Model the QMC data with an RBM, then sample!om the RBM Li Huang and LW, cf. Liu, Qi, Meng, Fu, Liu, Shen, Qi, Meng, Fu, Xu, Qi, Liu, Fu, Meng,

18 Why is it useful? When the fitting is perfect, we can completely bypass the quantum part of the QMC cf. Torlai, Melko, Even with an imperfect fitting, the RBM can still guide the QMC sampling RBM is a recommender system for the QMC simulations

19 Why is it useful? When the fitting is perfect, we can completely bypass the quantum part of the QMC cf. Torlai, Melko, Even with an imperfect fitting, the RBM can still guide the QMC sampling RBM is a recommender system for the QMC simulations Bonus: it is also fun to see what it discovers for the physical model

20 Learned Weight of a 8*8 Falicov-Kimball model

21 Accept or not? A(x! x 0 )=min apple 1, T (x 0! x) T (x! x 0 ) (x 0 ) (x) The art of Monte Carlo methods Sample the RBM, and propose the move to QMC p(h x) x h h 0 min 1, p(h0 ) p(h) p(x 0 h 0 ) x 0

22 Accept or not? A(x! x 0 )=min apple 1, T (x 0! x) T (x! x 0 ) (x 0 ) (x) Detailed balance condition for the RBM T (x 0! x) T (x! x 0 ) = p(x) p(x 0 ) Sample the RBM, and propose the move to QMC p(h x) x h h 0 min 1, p(h0 ) p(h) p(x 0 h 0 ) x 0

23 Accept or not? Acceptance rate of recommended update Li Huang and LW, A(x! x 0 )=min apple 1, p(x) p(x 0 ) (x0 ) (x) RBM Physical model surrogate function force bias R. M. Neal, Bayesian learning for neural networks, J. S. Liu, Monte Carlo strategies in scientific computing, 2008 S. Zhang, Auxiliary-field QMC for correlated electron systems, 2013

24 Question Can Boltzmann Machines Discover Cluster Updates? LW,

25 Cluster Update in a Nutshell Swendsen and Wang,1987

26 Cluster Update in a Nutshell Sample auxiliary bond variables Swendsen and Wang,1987

27 Cluster Update in a Nutshell Build clusters Swendsen and Wang,1987

28 Cluster Update in a Nutshell Flip clusters randomly Swendsen and Wang,1987

29 Cluster Update in a Nutshell Rejection free cluster update! Swendsen and Wang,1987

30 visible s i = ±1 p(h s) hidden h` =0, 1 flip clusters Swendsen-Wang Boltzmann Machine Ising spins Visible units Auxiliary bond variables Build clusters Flip clusters Hidden units Sample h given s Sample s given h

31 Boltzmann Machine for cluster update E(s, h) = X` (Ws i`s j` + b) h` s i` s j` s 2 { 1, +1} Vertex h` =1 h 2 {0, 1} Edge

32 Boltzmann Machine for cluster update E(s, h) = X` (Ws i`s j` + b) h` s i` s j` s 2 { 1, +1} Vertex h` =0 h 2 {0, 1} Edge

33 Boltzmann Machine for cluster update E(s, h) = X` (Ws i`s j` + b) h` s i` s j` s 2 { 1, +1} Vertex h` =0 h 2 {0, 1} Edge

34 Cluster Update in the BM language

35 Cluster Update in the BM language Sample hidden variables given visible units

36 Cluster Update in the BM language Inactive hidden units break the links

37 Cluster Update in the BM language Randomly flip clusters of visible units

38 Cluster Update in the BM language Voila!

39 Boltzmann Machine for cluster update E(s, h) = X` (Ws i`s j` + b) h` s i` s j` s 2 { 1, +1} Vertex h` =0 h 2 {0, 1} Edge Rejection free Monte Carlo updates when 1+eb+W 1+e b W = e2 J

40 Boltzmann Machine for cluster update E(s, h) = X` (Ws i`s j` + b) h` s i` s j` s 2 { 1, +1} Vertex h` =0 h 2 {0, 1} Edge Rejection free Monte Carlo updates when 1+eb+W 1+e b W = e2 J Encompass general cluster algorithm frameworks Niedermeyer,1988 Kandel and Domany,1991 Kawashima and Gubernatis,1995

41 Boltzmann Machine for cluster update E(s, h) = X` (Ws i`s j` + b) h` s i` s j` s 2 { 1, +1} Vertex h` =0 h 2 {0, 1} Edge Rejection free Monte Carlo updates when 1+eb+W 1+e b W = e2 J Encompass general cluster algorithm frameworks In general Niedermeyer,1988 E(s, h) =E(s) Kandel and Domany,1991 Kawashima and Gubernatis,1995 X [W F (s)+b ] h feature

42 Learn Generate Boltzmann Machines are learnable so they can adapt to various physical problems Boltzmann Machines parametrize Monte Carlo policies which can be optimized for efficiency The hidden units learn to play smart roles: Fortuin-Kasteleyn and Hubbard-Stratonovich transformations

43 Plaquette Ising model K However, for K 0, NO simple and efficient global update method is known

44 Plaquette Ising model K e Ks 1s 2 s 3 s 4! X h e W (s 1s 2 +s 3 s 4 )h

45 Plaquette Ising model K e Ks 1s 2 s 3 s 4! X h e W (s 1s 2 +s 3 s 4 )h Given s, sample h on each plaquette independently Given h, the Boltzmann Machine is an ordinary Ising model with modulated interactions, which can be sampled efficiently Rejection free cluster update!

46 Plaquette Ising model K e Ks 1s 2 s 3 s 4! X h e W (s 1s 2 +s 3 s 4 )h LW,

47 Machine learning for many-body physics New tools for (quantum) many-body problems Moreover, it offers a new way of thinking Can we make new scientific discovery with it? Can one design better algorithms with it? LW, Li Huang and LW, Li Huang, Yi-feng Yang and LW, LW,

48 Machine learning for many-body physics New tools for (quantum) many-body problems Moreover, it offers a new way of thinking Can we make new scientific discovery with it? Can one design better algorithms with it? The fun just starts! LW, Li Huang and LW, Li Huang, Yi-feng Yang and LW, LW,

49 Quantum many-body physics for ML Quantum entanglement perspective on deep learning Jing Chen, Song Cheng, Haidong Xie, LW, and Tao Xiang, Dong-Ling Deng, Xiaopeng Li and S. Das Sarma, Xun Gao, L.-M. Duan, Y. Huang and J. E. Moore, MNIST database random images from the Deep Learning book by Goodfellow, Bengio, Courville

50 A(x! x 0 )=min apple 1, p(x) p(x 0 ) (x0 ) (x) h 2 {0, 1} M x 2 {0, 1} N h 1 h 2 h 3... hm x 1 x 2... x N RBM Physical model e.g. Falikov-Kimball (classical field + fermions) model ln[ (x)] = U 2 NX x i + ln det h 1+e H(x) i while for the RBM i=1 H ij = K ij + ij U (x i 1/2) ln[p(x)] = NX a i x i + MX ln 1+e b j+ P N i=1 x iw ij i=1 j=1

51 Train the RBM p(x) (x) Supervised learning of the RBM NX i=1 = U 2 a i x i + NX i=1 MX j=1 ln 1+e b j+ P N i=1 x iw ij x i + ln det 1+e H x 1 x 2 x N......

Accelerated Monte Carlo simulations with restricted Boltzmann machines

Accelerated Monte Carlo simulations with restricted Boltzmann machines Accelerated simulations with restricted Boltzmann machines Li Huang 1 Lei Wang 2 1 China Academy of Engineering Physics 2 Institute of Physics, Chinese Academy of Sciences July 6, 2017 Machine Learning

More information

Neural Network Representation of Tensor Network and Chiral States

Neural Network Representation of Tensor Network and Chiral States Neural Network Representation of Tensor Network and Chiral States Yichen Huang ( 黄溢辰 ) 1 and Joel E. Moore 2 1 Institute for Quantum Information and Matter California Institute of Technology 2 Department

More information

Tensor network vs Machine learning. Song Cheng ( 程嵩 ) IOP, CAS

Tensor network vs Machine learning. Song Cheng ( 程嵩 ) IOP, CAS Tensor network vs Machine learning Song Cheng ( 程嵩 ) IOP, CAS physichengsong@iphy.ac.cn Outline Tensor network in a nutshell TN concepts in machine learning TN methods in machine learning Outline Tensor

More information

Guiding Monte Carlo Simulations with Machine Learning

Guiding Monte Carlo Simulations with Machine Learning Guiding Monte Carlo Simulations with Machine Learning Yang Qi Department of Physics, Massachusetts Institute of Technology Joining Fudan University in 2017 KITS, June 29 2017. 1/ 33 References Works led

More information

Restricted Boltzmann Machines

Restricted Boltzmann Machines Restricted Boltzmann Machines http://deeplearning4.org/rbm-mnist-tutorial.html Slides from Hugo Larochelle, Geoffrey Hinton, and Yoshua Bengio CSC321: Intro to Machine Learning and Neural Networks, Winter

More information

Parallel Tempering is Efficient for Learning Restricted Boltzmann Machines

Parallel Tempering is Efficient for Learning Restricted Boltzmann Machines Parallel Tempering is Efficient for Learning Restricted Boltzmann Machines KyungHyun Cho, Tapani Raiko, Alexander Ilin Abstract A new interest towards restricted Boltzmann machines (RBMs) has risen due

More information

Bias-Variance Trade-Off in Hierarchical Probabilistic Models Using Higher-Order Feature Interactions

Bias-Variance Trade-Off in Hierarchical Probabilistic Models Using Higher-Order Feature Interactions - Trade-Off in Hierarchical Probabilistic Models Using Higher-Order Feature Interactions Simon Luo The University of Sydney Data61, CSIRO simon.luo@data61.csiro.au Mahito Sugiyama National Institute of

More information

Learning Deep Architectures for AI. Part II - Vijay Chakilam

Learning Deep Architectures for AI. Part II - Vijay Chakilam Learning Deep Architectures for AI - Yoshua Bengio Part II - Vijay Chakilam Limitations of Perceptron x1 W, b 0,1 1,1 y x2 weight plane output =1 output =0 There is no value for W and b such that the model

More information

Self-learning Monte Carlo Method

Self-learning Monte Carlo Method Self-learning Monte Carlo Method Zi Yang Meng ( 孟子杨 ) http://ziyangmeng.iphy.ac.cn/ Know thyself "Know thyself" (Greek: γνῶθι σεαυτόν, gnothi seauton) one of the Delphic maxims and was inscribed in the

More information

Deep Learning the Ising Model Near Criticality

Deep Learning the Ising Model Near Criticality Deep Learning the Ising Model ear Criticality It is empirically well supported that neural networks with deep architectures perform better than shallow networks for certain machine learning tasks involving

More information

Kyle Reing University of Southern California April 18, 2018

Kyle Reing University of Southern California April 18, 2018 Renormalization Group and Information Theory Kyle Reing University of Southern California April 18, 2018 Overview Renormalization Group Overview Information Theoretic Preliminaries Real Space Mutual Information

More information

Deep Learning Basics Lecture 8: Autoencoder & DBM. Princeton University COS 495 Instructor: Yingyu Liang

Deep Learning Basics Lecture 8: Autoencoder & DBM. Princeton University COS 495 Instructor: Yingyu Liang Deep Learning Basics Lecture 8: Autoencoder & DBM Princeton University COS 495 Instructor: Yingyu Liang Autoencoder Autoencoder Neural networks trained to attempt to copy its input to its output Contain

More information

Jakub Hajic Artificial Intelligence Seminar I

Jakub Hajic Artificial Intelligence Seminar I Jakub Hajic Artificial Intelligence Seminar I. 11. 11. 2014 Outline Key concepts Deep Belief Networks Convolutional Neural Networks A couple of questions Convolution Perceptron Feedforward Neural Network

More information

arxiv: v1 [cond-mat.stat-mech] 10 Apr 2018

arxiv: v1 [cond-mat.stat-mech] 10 Apr 2018 Supervised machine learning algorithms based on generalized Gibbs ensembles Tatjana Puškarov and Axel Cortés Cubero Institute for Theoretical Physics, Center for Extreme Matter and Emergent Phenomena,

More information

Stochastic Gradient Estimate Variance in Contrastive Divergence and Persistent Contrastive Divergence

Stochastic Gradient Estimate Variance in Contrastive Divergence and Persistent Contrastive Divergence ESANN 0 proceedings, European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning. Bruges (Belgium), 7-9 April 0, idoc.com publ., ISBN 97-7707-. Stochastic Gradient

More information

UNSUPERVISED LEARNING

UNSUPERVISED LEARNING UNSUPERVISED LEARNING Topics Layer-wise (unsupervised) pre-training Restricted Boltzmann Machines Auto-encoders LAYER-WISE (UNSUPERVISED) PRE-TRAINING Breakthrough in 2006 Layer-wise (unsupervised) pre-training

More information

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics Chapter 11. Stochastic Methods Rooted in Statistical Mechanics Neural Networks and Learning Machines (Haykin) Lecture Notes on Self-learning Neural Algorithms Byoung-Tak Zhang School of Computer Science

More information

Deep Generative Models. (Unsupervised Learning)

Deep Generative Models. (Unsupervised Learning) Deep Generative Models (Unsupervised Learning) CEng 783 Deep Learning Fall 2017 Emre Akbaş Reminders Next week: project progress demos in class Describe your problem/goal What you have done so far What

More information

Based on the original slides of Hung-yi Lee

Based on the original slides of Hung-yi Lee Based on the original slides of Hung-yi Lee Google Trends Deep learning obtains many exciting results. Can contribute to new Smart Services in the Context of the Internet of Things (IoT). IoT Services

More information

Lecture 16 Deep Neural Generative Models

Lecture 16 Deep Neural Generative Models Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed

More information

Variational Autoencoders for Classical Spin Models

Variational Autoencoders for Classical Spin Models Variational Autoencoders for Classical Spin Models Benjamin Nosarzewski Stanford University bln@stanford.edu Abstract Lattice spin models are used to study the magnetic behavior of interacting electrons

More information

Machine learning quantum phases of matter

Machine learning quantum phases of matter Machine learning quantum phases of matter Summer School on Emergent Phenomena in Quantum Materials Cornell, May 2017 Simon Trebst University of Cologne Machine learning quantum phases of matter Summer

More information

Restricted Boltzmann Machines

Restricted Boltzmann Machines Restricted Boltzmann Machines Boltzmann Machine(BM) A Boltzmann machine extends a stochastic Hopfield network to include hidden units. It has binary (0 or 1) visible vector unit x and hidden (latent) vector

More information

Machine Learning with Quantum-Inspired Tensor Networks

Machine Learning with Quantum-Inspired Tensor Networks Machine Learning with Quantum-Inspired Tensor Networks E.M. Stoudenmire and David J. Schwab Advances in Neural Information Processing 29 arxiv:1605.05775 RIKEN AICS - Mar 2017 Collaboration with David

More information

Dynamical Systems and Deep Learning: Overview. Abbas Edalat

Dynamical Systems and Deep Learning: Overview. Abbas Edalat Dynamical Systems and Deep Learning: Overview Abbas Edalat Dynamical Systems The notion of a dynamical system includes the following: A phase or state space, which may be continuous, e.g. the real line,

More information

Deep Boltzmann Machines

Deep Boltzmann Machines Deep Boltzmann Machines Ruslan Salakutdinov and Geoffrey E. Hinton Amish Goel University of Illinois Urbana Champaign agoel10@illinois.edu December 2, 2016 Ruslan Salakutdinov and Geoffrey E. Hinton Amish

More information

Enhanced Gradient and Adaptive Learning Rate for Training Restricted Boltzmann Machines

Enhanced Gradient and Adaptive Learning Rate for Training Restricted Boltzmann Machines Enhanced Gradient and Adaptive Learning Rate for Training Restricted Boltzmann Machines KyungHyun Cho KYUNGHYUN.CHO@AALTO.FI Tapani Raiko TAPANI.RAIKO@AALTO.FI Alexander Ilin ALEXANDER.ILIN@AALTO.FI Department

More information

Revealing fermionic quantum criticality from new Monte Carlo techniques. Zi Yang Meng ( 孟子杨 )

Revealing fermionic quantum criticality from new Monte Carlo techniques. Zi Yang Meng ( 孟子杨 ) Revealing fermionic quantum criticality from new Monte Carlo techniques Zi Yang Meng ( 孟子杨 ) http://ziyangmeng.iphy.ac.cn Collaborators and References Xiao Yan Xu Zi Hong Liu Chuang Chen Gao Pei Pan Yang

More information

Knowledge Extraction from DBNs for Images

Knowledge Extraction from DBNs for Images Knowledge Extraction from DBNs for Images Son N. Tran and Artur d Avila Garcez Department of Computer Science City University London Contents 1 Introduction 2 Knowledge Extraction from DBNs 3 Experimental

More information

The Origin of Deep Learning. Lili Mou Jan, 2015

The Origin of Deep Learning. Lili Mou Jan, 2015 The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets

More information

Restricted Boltzmann Machines for Collaborative Filtering

Restricted Boltzmann Machines for Collaborative Filtering Restricted Boltzmann Machines for Collaborative Filtering Authors: Ruslan Salakhutdinov Andriy Mnih Geoffrey Hinton Benjamin Schwehn Presentation by: Ioan Stanculescu 1 Overview The Netflix prize problem

More information

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, 2017 Spis treści Website Acknowledgments Notation xiii xv xix 1 Introduction 1 1.1 Who Should Read This Book?

More information

Quantum Monte Carlo investigations of correlated electron systems, present and future. Zi Yang Meng ( 孟子杨 )

Quantum Monte Carlo investigations of correlated electron systems, present and future. Zi Yang Meng ( 孟子杨 ) Quantum Monte Carlo investigations of correlated electron systems, present and future Zi Yang Meng ( 孟子杨 ) http://ziyangmeng.iphy.ac.cn Collaborators Xiao Yan Xu Yoni Schattner Zi Hong Liu Erez Berg Chuang

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Neural networks Daniel Hennes 21.01.2018 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Logistic regression Neural networks Perceptron

More information

Machine Learning for Physicists Lecture 1

Machine Learning for Physicists Lecture 1 Machine Learning for Physicists Lecture 1 Summer 2017 University of Erlangen-Nuremberg Florian Marquardt (Image generated by a net with 20 hidden layers) OUTPUT INPUT (Picture: Wikimedia Commons) OUTPUT

More information

Learning Deep Architectures

Learning Deep Architectures Learning Deep Architectures Yoshua Bengio, U. Montreal Microsoft Cambridge, U.K. July 7th, 2009, Montreal Thanks to: Aaron Courville, Pascal Vincent, Dumitru Erhan, Olivier Delalleau, Olivier Breuleux,

More information

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding Large-Scale Feature Learning with Spike-and-Slab Sparse Coding Ian J. Goodfellow, Aaron Courville, Yoshua Bengio ICML 2012 Presented by Xin Yuan January 17, 2013 1 Outline Contributions Spike-and-Slab

More information

Deep Feedforward Networks

Deep Feedforward Networks Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 10-708 Probabilistic Graphical Models Homework 3 (v1.1.0) Due Apr 14, 7:00 PM Rules: 1. Homework is due on the due date at 7:00 PM. The homework should be submitted via Gradescope. Solution to each problem

More information

ARestricted Boltzmann machine (RBM) [1] is a probabilistic

ARestricted Boltzmann machine (RBM) [1] is a probabilistic 1 Matrix Product Operator Restricted Boltzmann Machines Cong Chen, Kim Batselier, Ching-Yun Ko, and Ngai Wong chencong@eee.hku.hk, k.batselier@tudelft.nl, cyko@eee.hku.hk, nwong@eee.hku.hk arxiv:1811.04608v1

More information

Classical Monte Carlo Simulations

Classical Monte Carlo Simulations Classical Monte Carlo Simulations Hyejin Ju April 17, 2012 1 Introduction Why do we need numerics? One of the main goals of condensed matter is to compute expectation values O = 1 Z Tr{O e βĥ} (1) and

More information

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari Deep Belief Nets Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines for continuous

More information

using Low-Depth Quantum Circuits Guillaume Verdon Co-Founder & Chief Scientific Officer

using Low-Depth Quantum Circuits Guillaume Verdon Co-Founder & Chief Scientific Officer Training Neural Networks using Low-Depth Quantum Circuits QC Meetup @YC Guillaume Verdon Co-Founder & Chief Scientific Officer guillaume@everettian.com Overview Classical Boltzmann machines? Energy-based

More information

Algorithms other than SGD. CS6787 Lecture 10 Fall 2017

Algorithms other than SGD. CS6787 Lecture 10 Fall 2017 Algorithms other than SGD CS6787 Lecture 10 Fall 2017 Machine learning is not just SGD Once a model is trained, we need to use it to classify new examples This inference task is not computed with SGD There

More information

Probabilistic Models in Theoretical Neuroscience

Probabilistic Models in Theoretical Neuroscience Probabilistic Models in Theoretical Neuroscience visible unit Boltzmann machine semi-restricted Boltzmann machine restricted Boltzmann machine hidden unit Neural models of probabilistic sampling: introduction

More information

A Unified View of Deep Generative Models

A Unified View of Deep Generative Models SAILING LAB Laboratory for Statistical Artificial InteLigence & INtegreative Genomics A Unified View of Deep Generative Models Zhiting Hu and Eric Xing Petuum Inc. Carnegie Mellon University 1 Deep generative

More information

arxiv: v1 [stat.ml] 30 Aug 2010

arxiv: v1 [stat.ml] 30 Aug 2010 Sparse Group Restricted Boltzmann Machines arxiv:008.4988v [stat.ml] 30 Aug 200 Heng Luo Department of Computer Science Shanghai Jiao Tong University hengluo@sjtu.edu.cn Changyong Niu Department of Computer

More information

Advanced Computation for Complex Materials

Advanced Computation for Complex Materials Advanced Computation for Complex Materials Computational Progress is brainpower limited, not machine limited Algorithms Physics Major progress in algorithms Quantum Monte Carlo Density Matrix Renormalization

More information

Mathematics for Artificial Intelligence

Mathematics for Artificial Intelligence Mathematics for Artificial Intelligence Reading Course Elena Agliari Dipartimento di Matematica Sapienza Università di Roma (TENTATIVE) PLAN OF THE COURSE Introduction Chapter : Basics of statistical mechanics

More information

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight) CSCE 636 Neural Networks Instructor: Yoonsuck Choe Deep Learning What Is Deep Learning? Learning higher level abstractions/representations from data. Motivation: how the brain represents sensory information

More information

Kartik Audhkhasi, Osonde Osoba, Bart Kosko

Kartik Audhkhasi, Osonde Osoba, Bart Kosko To appear: 2013 International Joint Conference on Neural Networks (IJCNN-2013 Noise Benefits in Backpropagation and Deep Bidirectional Pre-training Kartik Audhkhasi, Osonde Osoba, Bart Kosko Abstract We

More information

Opportunities and challenges in quantum-enhanced machine learning in near-term quantum computers

Opportunities and challenges in quantum-enhanced machine learning in near-term quantum computers Opportunities and challenges in quantum-enhanced machine learning in near-term quantum computers Alejandro Perdomo-Ortiz Senior Research Scientist, Quantum AI Lab. at NASA Ames Research Center and at the

More information

Deep Belief Networks are compact universal approximators

Deep Belief Networks are compact universal approximators 1 Deep Belief Networks are compact universal approximators Nicolas Le Roux 1, Yoshua Bengio 2 1 Microsoft Research Cambridge 2 University of Montreal Keywords: Deep Belief Networks, Universal Approximation

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Deep Learning Autoencoder Models

Deep Learning Autoencoder Models Deep Learning Autoencoder Models Davide Bacciu Dipartimento di Informatica Università di Pisa Intelligent Systems for Pattern Recognition (ISPR) Generative Models Wrap-up Deep Learning Module Lecture Generative

More information

Introduction to Restricted Boltzmann Machines

Introduction to Restricted Boltzmann Machines Introduction to Restricted Boltzmann Machines Ilija Bogunovic and Edo Collins EPFL {ilija.bogunovic,edo.collins}@epfl.ch October 13, 2014 Introduction Ingredients: 1. Probabilistic graphical models (undirected,

More information

Introduction to Deep Neural Networks

Introduction to Deep Neural Networks Introduction to Deep Neural Networks Presenter: Chunyuan Li Pattern Classification and Recognition (ECE 681.01) Duke University April, 2016 Outline 1 Background and Preliminaries Why DNNs? Model: Logistic

More information

A = {(x, u) : 0 u f(x)},

A = {(x, u) : 0 u f(x)}, Draw x uniformly from the region {x : f(x) u }. Markov Chain Monte Carlo Lecture 5 Slice sampler: Suppose that one is interested in sampling from a density f(x), x X. Recall that sampling x f(x) is equivalent

More information

How to do backpropagation in a brain

How to do backpropagation in a brain How to do backpropagation in a brain Geoffrey Hinton Canadian Institute for Advanced Research & University of Toronto & Google Inc. Prelude I will start with three slides explaining a popular type of deep

More information

Annealing Between Distributions by Averaging Moments

Annealing Between Distributions by Averaging Moments Annealing Between Distributions by Averaging Moments Chris J. Maddison Dept. of Comp. Sci. University of Toronto Roger Grosse CSAIL MIT Ruslan Salakhutdinov University of Toronto Partition Functions We

More information

Learning Deep Boltzmann Machines using Adaptive MCMC

Learning Deep Boltzmann Machines using Adaptive MCMC Ruslan Salakhutdinov Brain and Cognitive Sciences and CSAIL, MIT 77 Massachusetts Avenue, Cambridge, MA 02139 rsalakhu@mit.edu Abstract When modeling high-dimensional richly structured data, it is often

More information

An Empirical Investigation of Minimum Probability Flow Learning Under Different Connectivity Patterns

An Empirical Investigation of Minimum Probability Flow Learning Under Different Connectivity Patterns An Empirical Investigation of Minimum Probability Flow Learning Under Different Connectivity Patterns Daniel Jiwoong Im, Ethan Buchman, and Graham W. Taylor School of Engineering University of Guelph Guelph,

More information

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction ECE 521 Lecture 11 (not on midterm material) 13 February 2017 K-means clustering, Dimensionality reduction With thanks to Ruslan Salakhutdinov for an earlier version of the slides Overview K-means clustering

More information

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight) CSCE 636 Neural Networks Instructor: Yoonsuck Choe Deep Learning What Is Deep Learning? Learning higher level abstractions/representations from data. Motivation: how the brain represents sensory information

More information

Tempered Markov Chain Monte Carlo for training of Restricted Boltzmann Machines

Tempered Markov Chain Monte Carlo for training of Restricted Boltzmann Machines Tempered Markov Chain Monte Carlo for training of Restricted Boltzmann Machines Desjardins, G., Courville, A., Bengio, Y., Vincent, P. and Delalleau, O. Technical Report 1345, Dept. IRO, Université de

More information

Learning to Disentangle Factors of Variation with Manifold Learning

Learning to Disentangle Factors of Variation with Manifold Learning Learning to Disentangle Factors of Variation with Manifold Learning Scott Reed Kihyuk Sohn Yuting Zhang Honglak Lee University of Michigan, Department of Electrical Engineering and Computer Science 08

More information

BIAS-VARIANCE DECOMPOSITION FOR BOLTZMANN MACHINES

BIAS-VARIANCE DECOMPOSITION FOR BOLTZMANN MACHINES BIAS-VARIANCE DECOMPOSITION FOR BOLTZMANN MACHINES Anonymous authors Paper under double-blind review ABSTRACT We achieve bias-variance decomposition for Boltzmann machines using an information geometric

More information

arxiv: v1 [stat.ml] 14 Oct 2014

arxiv: v1 [stat.ml] 14 Oct 2014 An exact mapping between the Variational Renormalization Group and Deep Learning Pankaj Mehta Dept. of Physics, Boston University, Boston, MA David J. Schwab Dept. of Physics, Northwestern University,

More information

Invaded cluster dynamics for frustrated models

Invaded cluster dynamics for frustrated models PHYSICAL REVIEW E VOLUME 57, NUMBER 1 JANUARY 1998 Invaded cluster dynamics for frustrated models Giancarlo Franzese, 1, * Vittorio Cataudella, 1, * and Antonio Coniglio 1,2, * 1 INFM, Unità di Napoli,

More information

arxiv: v3 [cs.lg] 18 Mar 2013

arxiv: v3 [cs.lg] 18 Mar 2013 Hierarchical Data Representation Model - Multi-layer NMF arxiv:1301.6316v3 [cs.lg] 18 Mar 2013 Hyun Ah Song Department of Electrical Engineering KAIST Daejeon, 305-701 hyunahsong@kaist.ac.kr Abstract Soo-Young

More information

arxiv: v2 [cs.ne] 22 Feb 2013

arxiv: v2 [cs.ne] 22 Feb 2013 Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint arxiv:1301.3533v2 [cs.ne] 22 Feb 2013 Xanadu C. Halkias DYNI, LSIS, Universitè du Sud, Avenue de l Université - BP20132, 83957 LA

More information

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single

More information

Greedy Layer-Wise Training of Deep Networks

Greedy Layer-Wise Training of Deep Networks Greedy Layer-Wise Training of Deep Networks Yoshua Bengio, Pascal Lamblin, Dan Popovici, Hugo Larochelle NIPS 2007 Presented by Ahmed Hefny Story so far Deep neural nets are more expressive: Can learn

More information

Simon Trebst. Quantitative Modeling of Complex Systems

Simon Trebst. Quantitative Modeling of Complex Systems Quantitative Modeling of Complex Systems Machine learning In computer science, machine learning is concerned with algorithms that allow for data analytics, most prominently dimensional reduction and feature

More information

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann Neural Networks with Applications to Vision and Language Feedforward Networks Marco Kuhlmann Feedforward networks Linear separability x 2 x 2 0 1 0 1 0 0 x 1 1 0 x 1 linearly separable not linearly separable

More information

Sum-Product Networks: A New Deep Architecture

Sum-Product Networks: A New Deep Architecture Sum-Product Networks: A New Deep Architecture Pedro Domingos Dept. Computer Science & Eng. University of Washington Joint work with Hoifung Poon 1 Graphical Models: Challenges Bayesian Network Markov Network

More information

The flip-the-state transition operator for restricted Boltzmann machines

The flip-the-state transition operator for restricted Boltzmann machines Mach Learn 2013) 93:53 69 DOI 10.1007/s10994-013-5390-3 The flip-the-state transition operator for restricted Boltzmann machines Kai Brügge Asja Fischer Christian Igel Received: 14 February 2013 / Accepted:

More information

Bayesian Mixtures of Bernoulli Distributions

Bayesian Mixtures of Bernoulli Distributions Bayesian Mixtures of Bernoulli Distributions Laurens van der Maaten Department of Computer Science and Engineering University of California, San Diego Introduction The mixture of Bernoulli distributions

More information

Measuring the Usefulness of Hidden Units in Boltzmann Machines with Mutual Information

Measuring the Usefulness of Hidden Units in Boltzmann Machines with Mutual Information Measuring the Usefulness of Hidden Units in Boltzmann Machines with Mutual Information Mathias Berglund, Tapani Raiko, and KyungHyun Cho Department of Information and Computer Science Aalto University

More information

A Study of Complex Deep Learning Networks on High Performance, Neuromorphic, and Quantum Computers

A Study of Complex Deep Learning Networks on High Performance, Neuromorphic, and Quantum Computers A Study of Complex Deep Learning Networks on High Performance, Neuromorphic, and Quantum Computers Thomas E. Potok, Ph.D. Computational Data Analytics Group Oak Ridge National Laboratory ORNL is managed

More information

Energy Based Models. Stefano Ermon, Aditya Grover. Stanford University. Lecture 13

Energy Based Models. Stefano Ermon, Aditya Grover. Stanford University. Lecture 13 Energy Based Models Stefano Ermon, Aditya Grover Stanford University Lecture 13 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 13 1 / 21 Summary Story so far Representation: Latent

More information

Classification with Perceptrons. Reading:

Classification with Perceptrons. Reading: Classification with Perceptrons Reading: Chapters 1-3 of Michael Nielsen's online book on neural networks covers the basics of perceptrons and multilayer neural networks We will cover material in Chapters

More information

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 14. Sinan Kalkan

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 14. Sinan Kalkan CENG 783 Special topics in Deep Learning AlchemyAPI Week 14 Sinan Kalkan Today Hopfield Networks Boltzmann Machines Deep BM, Restricted BM Generative Adversarial Networks Variational Auto-encoders Autoregressive

More information

Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network

Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network Srinjoy Das Department of Electrical and Computer Engineering University of California, San Diego srinjoyd@gmail.com Bruno Umbria

More information

Inductive Principles for Restricted Boltzmann Machine Learning

Inductive Principles for Restricted Boltzmann Machine Learning Inductive Principles for Restricted Boltzmann Machine Learning Benjamin Marlin Department of Computer Science University of British Columbia Joint work with Kevin Swersky, Bo Chen and Nando de Freitas

More information

STA 414/2104: Lecture 8

STA 414/2104: Lecture 8 STA 414/2104: Lecture 8 6-7 March 2017: Continuous Latent Variable Models, Neural networks Delivered by Mark Ebden With thanks to Russ Salakhutdinov, Jimmy Ba and others Outline Continuous latent variable

More information

Density functionals from deep learning

Density functionals from deep learning Density functionals from deep learning Jeffrey M. McMahon Department of Physics & Astronomy March 15, 2016 Jeffrey M. McMahon (WSU) March 15, 2016 1 / 18 Kohn Sham Density-functional Theory (KS-DFT) The

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Course Structure. Psychology 452 Week 12: Deep Learning. Chapter 8 Discussion. Part I: Deep Learning: What and Why? Rufus. Rufus Processed By Fetch

Course Structure. Psychology 452 Week 12: Deep Learning. Chapter 8 Discussion. Part I: Deep Learning: What and Why? Rufus. Rufus Processed By Fetch Psychology 452 Week 12: Deep Learning What Is Deep Learning? Preliminary Ideas (that we already know!) The Restricted Boltzmann Machine (RBM) Many Layers of RBMs Pros and Cons of Deep Learning Course Structure

More information

arxiv: v1 [stat.ml] 2 Sep 2014

arxiv: v1 [stat.ml] 2 Sep 2014 On the Equivalence Between Deep NADE and Generative Stochastic Networks Li Yao, Sherjil Ozair, Kyunghyun Cho, and Yoshua Bengio Département d Informatique et de Recherche Opérationelle Université de Montréal

More information

Chapter 16. Structured Probabilistic Models for Deep Learning

Chapter 16. Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe

More information

Critical Dynamics of Two-Replica Cluster Algorithms

Critical Dynamics of Two-Replica Cluster Algorithms University of Massachusetts Amherst From the SelectedWorks of Jonathan Machta 2001 Critical Dynamics of Two-Replica Cluster Algorithms X. N. Li Jonathan Machta, University of Massachusetts Amherst Available

More information

Graphical Representations and Cluster Algorithms

Graphical Representations and Cluster Algorithms Graphical Representations and Cluster Algorithms Jon Machta University of Massachusetts Amherst Newton Institute, March 27, 2008 Outline Introduction to graphical representations and cluster algorithms

More information

10. Artificial Neural Networks

10. Artificial Neural Networks Foundations of Machine Learning CentraleSupélec Fall 217 1. Artificial Neural Networks Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech chloe-agathe.azencott@mines-paristech.fr Learning

More information

7 Monte Carlo Simulations of Quantum Spin Models

7 Monte Carlo Simulations of Quantum Spin Models 7 Monte Carlo Simulations of Quantum Spin Models Stefan Wessel Institute for Theoretical Solid State Physics RWTH Aachen University Contents 1 Introduction 2 2 World lines and local updates 2 2.1 Suzuki-Trotter

More information

Representational Power of Restricted Boltzmann Machines and Deep Belief Networks. Nicolas Le Roux and Yoshua Bengio Presented by Colin Graber

Representational Power of Restricted Boltzmann Machines and Deep Belief Networks. Nicolas Le Roux and Yoshua Bengio Presented by Colin Graber Representational Power of Restricted Boltzmann Machines and Deep Belief Networks Nicolas Le Roux and Yoshua Bengio Presented by Colin Graber Introduction Representational abilities of functions with some

More information

Auto-Encoding Variational Bayes

Auto-Encoding Variational Bayes Auto-Encoding Variational Bayes Diederik P Kingma, Max Welling June 18, 2018 Diederik P Kingma, Max Welling Auto-Encoding Variational Bayes June 18, 2018 1 / 39 Outline 1 Introduction 2 Variational Lower

More information

arxiv: v1 [cs.ne] 6 May 2014

arxiv: v1 [cs.ne] 6 May 2014 Training Restricted Boltzmann Machine by Perturbation arxiv:1405.1436v1 [cs.ne] 6 May 2014 Siamak Ravanbakhsh, Russell Greiner Department of Computing Science University of Alberta {mravanba,rgreiner@ualberta.ca}

More information

Deep Feedforward Networks

Deep Feedforward Networks Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3

More information

Basic Principles of Unsupervised and Unsupervised

Basic Principles of Unsupervised and Unsupervised Basic Principles of Unsupervised and Unsupervised Learning Toward Deep Learning Shun ichi Amari (RIKEN Brain Science Institute) collaborators: R. Karakida, M. Okada (U. Tokyo) Deep Learning Self Organization

More information