Identifying QCD transition using Deep Learning

Similar documents
An EoS-meter of QCD transition from deep learning

Memory-Augmented Attention Model for Scene Text Recognition

Feature Design. Feature Design. Feature Design. & Deep Learning

Introduction to Deep Neural Networks

Introduction to Neural Networks

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Deep Learning: Pre- Requisites. Understanding gradient descent, autodiff, and softmax

Convolutional Neural Networks II. Slides from Dr. Vlad Morariu

Classification of Higgs Boson Tau-Tau decays using GPU accelerated Neural Networks

Jakub Hajic Artificial Intelligence Seminar I

Introduction to Convolutional Neural Networks 2018 / 02 / 23

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Introduction to Convolutional Neural Networks (CNNs)

Neural Networks. David Rosenberg. July 26, New York University. David Rosenberg (New York University) DS-GA 1003 July 26, / 35

Based on the original slides of Hung-yi Lee

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Deep Learning: a gentle introduction

Neural Network Tutorial & Application in Nuclear Physics. Weiguang Jiang ( 蒋炜光 ) UTK / ORNL

Comparing Initial Conditions in a (3+1)d Boltzmann + Hydrodynamics Transport Approach

Neural Networks. Yan Shao Department of Linguistics and Philology, Uppsala University 7 December 2016

Machine learning in Higgs analyses

CSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer

Neural Networks and Deep Learning

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Machine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 3: Introduction to Deep Learning (continued)

Deep Learning Lab Course 2017 (Deep Learning Practical)

<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)

SGD and Deep Learning

Neural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /9/17

Supervised Deep Learning

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

Deep Learning Autoencoder Models

Machine Learning. Boris

arxiv: v1 [cs.lg] 10 Aug 2018

Grundlagen der Künstlichen Intelligenz

Quantum Artificial Intelligence and Machine Learning: The Path to Enterprise Deployments. Randall Correll. +1 (703) Palo Alto, CA

Lecture 35: Optimization and Neural Nets

Neural Networks 2. 2 Receptive fields and dealing with image inputs

Lecture 7 Convolutional Neural Networks

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Recurrent Neural Networks with Flexible Gates using Kernel Activation Functions

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Index. Santanu Pattanayak 2017 S. Pattanayak, Pro Deep Learning with TensorFlow,

Demystifying deep learning. Artificial Intelligence Group Department of Computer Science and Technology, University of Cambridge, UK

COMPARING FIXED AND ADAPTIVE COMPUTATION TIME FOR RE-

Deep Learning (CNNs)

10. Artificial Neural Networks

Google s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning

Introduction to Neural Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

An overview of deep learning methods for genomics

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

The connection of dropout and Bayesian statistics

arxiv: v2 [cs.sd] 7 Feb 2018

Classifaction of gg gh against qg qh

Introduction to Neural Networks

Introduction to Deep Learning

Summary and discussion of: Dropout Training as Adaptive Regularization

Deep Learning Architectures and Algorithms

Tasks ADAS. Self Driving. Non-machine Learning. Traditional MLP. Machine-Learning based method. Supervised CNN. Methods. Deep-Learning based

Convolutional Neural Networks

CSC321 Lecture 16: ResNets and Attention

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Lecture 17: Neural Networks and Deep Learning

CS 4700: Foundations of Artificial Intelligence

Deep Learning for NLP

Machine Learning for Physicists Lecture 1

Recurrent Neural Networks (RNN) and Long-Short-Term-Memory (LSTM) Yuan YAO HKUST

Architecture Multilayer Perceptron (MLP)

From perceptrons to word embeddings. Simon Šuster University of Groningen

Introduction to Deep Learning

Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

Making Deep Learning Understandable for Analyzing Sequential Data about Gene Regulation

Learning Deep Architectures for AI. Part II - Vijay Chakilam

CSE446: Neural Networks Winter 2015

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

Source localization in an ocean waveguide using supervised machine learning

PATTERN RECOGNITION AND MACHINE LEARNING

AUDIO SET CLASSIFICATION WITH ATTENTION MODEL: A PROBABILISTIC PERSPECTIVE. Qiuqiang Kong*, Yong Xu*, Wenwu Wang, Mark D. Plumbley

Lecture 8: Introduction to Deep Learning: Part 2 (More on backpropagation, and ConvNets)

Administration. Registration Hw3 is out. Lecture Captioning (Extra-Credit) Scribing lectures. Questions. Due on Thursday 10/6

Simon Trebst. Quantitative Modeling of Complex Systems

End-to-End Event Classification of High-Energy Physics Data

>TensorFlow and deep learning_

Neural Architectures for Image, Language, and Speech Processing

Normalization Techniques in Training of Deep Neural Networks

Fast Learning with Noise in Deep Neural Nets

P-TELU : Parametric Tan Hyperbolic Linear Unit Activation for Deep Neural Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Introduction to Deep Learning CMPT 733. Steven Bergner

CS230: Lecture 8 Word2Vec applications + Recurrent Neural Networks with Attention

CSC 578 Neural Networks and Deep Learning

@SoyGema GEMA PARREÑO PIQUERAS

Machine Learning Lecture 14

Transcription:

Identifying QCD transition using Deep Learning Kai Zhou Long-Gang Pang, Nan Su, Hannah Peterson, Horst Stoecker, Xin-Nian Wang Collaborators: arxiv:1612.04262

Outline 2

What is deep learning? Artificial Intelligence (AI) Machine Learning (ML) Deep Learning (DL) Booming in theory and applications! 1. Big Data 2. GPU parallel 3. New architecture 2006 Geoffrey Hinton 3

Supervised learning---most commen DL RL RNN/LSTM/GRU Reinforcement Learning (or Deep Q-learning) Playing Games Recurrent Neural Network Long Short Term Memory Gated Recurrent Unit Natural Language processing CNN Convolution Neural Network Image recognition/ classification 4

Fancy applications (RL) AlphaGo (by Google DeepMind) beat human master Self driving cars Robot control Playing Games like flappybird, Starcraft 5

RNN/LSTM for natural language generation arxiv: 1610.09889v1 6

Object detection using CNN Slides from KaiMing, He s recent talk 7

Artistic style transfer using CNN arxiv:1508.06576 style can be modelled and transferred! 8

Hybrid CNN + LSTM for image captioning By Google Brain Using concepts learned from similar scenes in the training set. 9

An example of DL structure : 10

hello world example of deep neural network Fig from CS231N, Stanford 1, Feed Forward Linear operations:rotating, boosting,... increasing or decreasing dimensions non linear activation function: correlation/links 11

hello world example of deep neural network Fig from CS231N, Stanford 1, Feed Forward Linear operations:rotating, boosting,... increasing or decreasing dimensions 12

hello world example of deep neural network Fig from CS231N, Stanford 1, Feed Forward Transformation of feature space : 13

hello world example of deep neural network Fig from CS231N, Stanford 2, Back Propagation Mean square error (simplest loss function) function with the predicted value and true value SGD for parameter update to minimize loss function 14

Overfitting problem in fully connected network Fixed data size n tio ro er r g in in ro er Early stopping a lid Va a Tr Cost or Loss r # of parameters or training time Too many parameters may easily over fit to training dataset 15

Ways to reduce overfitting 1.Early stopping 2.Increase training dataset by a.preparing more data. b.data augmentation (crop, scale, rotate, flip ). 3.Reduce number of parameters a.dropout: randomly discarding neurons. b.drop connection: randomly discarding connections. c.cnn: CNN locally connected to a small chunk of neurons in the previous layer. d.go deeper. 4.Regularization, weight decay 16

Convolution neural network Fig from CS231N, Stanford Advantage: scaling, rotating, translation invariant features can be learned since only subregion is connected to the filter/kernel which scan the whole input to feel the 2 d structure and local statistics, and Weight shared. 17

An example for CNN structure AlexNet, Krizhevsky et al. 2012 8 layers, 60 million parameters Data augmentation and Dropout are used to reduce overfitting, first use ReLU Removing any hidden layer results in 2% loss. 18

What CNN did after the training? Distributed representations are learned! 19

Open Source Libraries Keras + TensorFlow in the present study Keras is a high level neural network library, written in Python and capable of running on top of either TensorFlow or Theano. # Build one fully connected neural network (100->64->10 neurons) in Keras from keras.models import Sequential from keras.layers import Dense, Activation model = Sequential() model.add(dense(output_dim=64, input_dim=100)) model.add(activation("relu")) model.add(dense(output_dim=10)) model.add(activation( softmax")) model.compile(loss='categorical_crossentropy', optimizer='sgd', metrics=['accuracy']) 20

Deep learning in Physics (Particle Physics) Searching for Exotic Particles in High Energy Physics with Deep Learning P. Baldi, P. Sadowski, and D. Whiteson Nature Communications Higgs benchmark Signal Background 21

Deep learning in Physics (HeavyIon Jet) 22

Deep learning in Physics (Cond-Mat Ising) Extract order parameter from the spin configurations, can encode phases and discriminate Phase Transition for spin, Coulomb and even topological phases. 23

Identifying QCD transition in HIC using DL 24

QCD transition: 1st order or crossover? EOSL EOSQ 25

Uncertainties in the simulation of HIC Initial stage Pre-Equilibrium dynamics? Fast thermalization/tau0? Initial entropy deposition? Baryon stopping? Initial state fluctuations and flow? EoS and transport coefficients (shear viscosity, bulk viscosity) some unfixed scattering cross sections Particle momentum distribution Big uncertainty when comparing with Exp. data 26

New perspective --- Deep Learning Dog Brain/CNN Cat CNN Crossover (0,1)? 1st order (1,0) 27

Key idea for our prototype study Supervised learning using deep convolution neural network with huge amount of labeled training data (spectra, EoS type) from event-byevent relativistic hydrodynamic simulations. 28

Training dataset for charged pions at mid rapidity CLVisc: (3+1D viscous hydrodynamic program parallelized on GPU using OpenCL) with AMPT initial condition (~22000 events, doubled by left-right flipping, 10% for validation). is 0.4 fm for Au+Au and 0.2 fm for Pb+Pb collisions T_dec=0.137 GeV 29

Testing dataset iebe-vishnu: another viscous hydro package with different numerical solver for the partial differential equations and with different initial conditions, eta/s. is 0.6 fm for all the testing dataset. T_dec in [0.115GeV, 0.142GeV] for iebe-vishnu 30

First try with fully connected neural network Overfit to the training dataset! Does not work for testing dataset. no generalization. 31

CNN architecture for EoS-meter Batch normalization, Dropout, L2 Rregularization, PreLU are used to prevent the overfitting. Train 500 epochs, in mini-batch with size=64 Learning rate: 0.0001, decay with rate: 1.0E-6 32

Results ~95% prediction accuracy on avg. in independent testing dataset solely from the raw spectra The performance is robust against : initial conditions/fluctuations, tao_0, eta/s, T_dec ---- model independent (general rule is learned) 33

EoS-meter from CNN phi p_t Each hotspot represents correlations in a small phase space (32 features in the 2nd convolution layer)! Correlation of these features from the next fully connected layer 34

Conventional observables Strongly depends on initial stage fluctuations and other parameters 35

CNN provides novel perspective in connecting QCD theory with HIC directly : There do exist 'mapping/encoders/projection' from QCD transition onto the final state raw spectra, although they are not intuitive to conventional interpretation yet. ---They are clean (robust to other uncertainties and parameters) The deep CNN can provide a powerful and efficient 'Decoder' for the above Encoder/mapping ------the high-level representations act as 'EoS-meter' 36

Outlook Try to find out the underlying mechanism (guided back-propagation) Extend the model to work with real Exp. data Extract other dynamical parameters like eta/s. 37

What DL is doing? Slides by Long-gang