Appendix for Stochastic Gradient Monomial Gamma Sampler

Size: px
Start display at page:

Download "Appendix for Stochastic Gradient Monomial Gamma Sampler"

Transcription

1 Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing theorem to characterize the stationary distribution of the stochastic process with SDEs in (3) Theorem 3 The stochastic process generated from SDEs (3) converges to a stationary distribution p( ) / exp( H( )), where H( ) is defined as in () Proof We first show that the Fokker-Panck equation hods for the proposed SDE and probabiity density p( ), r p( )V ( ) = r r T :[p( )D( )] Here the r, (/, /p, / ) represents a vector inner product and : denotes the matrix doube dot product, ie, X : Y = Tr(X T Y ) In order to show FP equation hods, we ook at both side of the equation The eft hand side can be written as r p( )V ( ) = [ V ( ) H( ) V ( )]p( ) = { [(ru( )) r U( )] For the right hand side, p [(rk(p)) r K(p)] [(rf ( )) r F ( )]}p( ) r r T :[p( )D( )] = rr T : p( ) p rr T : p( ) rr T : p( ) = { [(ru( )) r U( )] p [(rk(p)) r K(p)] [(rf ( )) r F ( )]}p( ) For stationary distribution, p(,t) t = As a resut, the equaity in (4) hods The stochastic process defined by (3) is preserved by the dynamic Aternativey, one can everage the recipe from (Ma et a, 5) to recover the same concusion, by setting semi-definite matrix D = Diag([, p, ]) and skew-symmetric Q to be I I rk(p) rk(p) A Agorithm SGMGT/SGMGT-D with Euer integrator Input: Monomia parameter a, variances {, p, }, thermostat variance, resamping frequency {T p,t }, and earning rate h Initiaized, p and (a to zero) for t =,, do Evauate stochastic gradient rũ( ) on mini-batch After each T p iterations, resampe momentum, p t, from / exp( K c (p)) using rejection samping For each T iterations, resampe thermostats, t, from N (, p ) Generate random variabes, p, N D (, ) Compute first and second order derivatives for kinetics K c (p) as in () t = t rũ( )h rk c(p)h p h p t = p t rũ( )h (rk c(p))h p p h p t = t [rk c (p) rk c (p) (r K c (p))]h ( p)h p h Note that under the softened kinetics, the K c (p) is twice differentiabe, and rk c (p) is Lipschitz continuous Thus the Fokker-Panck equation hods, eading to a stationary distribution invariant to target distribution Another remark is that the resamping process for p and wi sti ead to the same invariante distribution p( ), since the resamping process is directy drawing sampe from the margina distribution Finay, it can be proved that the corresponding Itô diffusion of our agorithm in (3) is non-reversibe This speed up the convergence speed to equiibrium, because it is known that a reversibe process convergences sower than its non-reversibe counter part (Hwang et a, 5) B SGMGT-D/SGMGT-D with Euer integrator C Detais for softened kinetics We provide the detais for the derivation of softened kinetics Note that in the SDE (3), ony rk c (p) and r K c (p) is invoved For a =, we consider K c (p) = g(p)/c og( e cg(p) ), (4) g(p) =p/m

2 Stochastic Gradient Monomia Gamma Samper which gives rk c (p) = m (g(p)), r K c (p) = m (g(p)) Where, (x) = ecx e cx (tanh) with the softening parameter c, (x) = For a =, we consider which gives K c (p) =g(p) is the hyperboic tangent function g(p) = p / /m cecx (e cx ) 4 c( e cg(p) ), (5) rk c (p) = m sign(p) (g(p)) p /, r K c (p) = m (g(p)) (g(p)) p 4m (g(p)) p 3/ In genera, for arbitrary a, we consider setting the rk c (p) = a m (g(p))a p /a Such specification wi yied a differentiabe softened kinetics function by computing the integra, which is tractabe for positive vaue of a However, in practice, as suggested by Zhang et a (6) the optima a woud usuay between [5, ] We woud suggest consider using a =or a = for genera inference tasks D Synthetic muti-we potentia probem The five-we potentia is defined as: U( ), e P i= ci sin( 4 i( 4)), where c =( 47, 83, 7,, 4,, 7, 37, 87, 37) is a vector, c i is the i-th eement of c E Symmetric Spitting Integrators for SGMGT The first-ordered Euer integration resuts in high discretization error in Hamitonian dynamic updating of HMC In Chen et a (6), a symmetric spitting scheme is everaged to reduce the numerica error We appied the ( p) softened kinetics K c (p), and set F ( ) as In this symmetric spitting scheme, the Hamitonian is spit into sub-componenents, and for each sub-componenents an individua SDE is appied on The resuting discretization is sympectic and second-ordered: A : d = B : d = O : d = rũ( )rk c(p) f( ) rk c (p) rũ( ) A dt/ A dt D( )dw A dt/ Here we denote f( ), [(rk c (p)) (r K c (p))] ( p) for carity The sub-sde under sub-sde B is anayticay sovabe Foowing (Chen et a, 5), for a 6= /, the updating procedure foows an ABOBA scheme, given by A : t/3 = t rk c (p)h/, t/3 = t f( )h/ B : p t/3 =[p (a )/a a a/(a ) t a t/ h/] O : t/3 = t/3 p p t/3 = p t/3 rũ( ))h/p p p, t/3 = t/3 p B : p t =[p (a )/a a a/(a ) t/3 a t/3 h/] A : t = t/3 rk c (p)h/, t = t/3 f( )h/ When a =/, it foows the spitting scheme with standard SGNHT (Chen et a, 5) F Experimenta setups for DRBM The hyper-parameter setups for the DRBM experiments are provided as beow We seect the hyperparameters based on the performance on vaidation dataset The agorithm wi be eary stopped if the vaidation error starts to increase The seection is based on a grid search For p, and we seect from {,,,, } For the softening parameter c we seect from {3, 5, 8} We fixed m =and = The stepsize is chosen from {e 5, e 5, 5e 5, e 4, e 4, 5e 4} The T p and T are set as and, respectivey For SGLD, we use a stepsize of e 5 Tabe 4 Experimenta setup for discriminative RBM Agorithms p c h SGNHT e-4 SGNHT-D - e-4 SGMGT-D (a=) 3 e-5 SGMGT-D (a=) 5 5e-5

3 Stochastic Gradient Monomia Gamma Samper Tabe 5 Experimenta setup for RNNs Agorithms m p c SGNHT SGNHT-D - SGMGT/SGMGT-D (a=) 5 SGMGT/SGMGT-D (a=) 3 G Experimenta setups for RNN The hyper-parameter setups for the RNN experiments are simiar to the DRBM experiments For p, and we seect from {,,, } For the softening parameter c we seect from {3, 5, 8} We fixed the m =and = The stepsize of SGMGT-D/SGMGT is chosen from {e 3, 5e 3, e 3, 5e 3, 3e 3} The T p and T are set as and, respectivey We aso incorporate a decay scheme for stepsize, ie the stepsize is divided by a decaying factor = for each scan of dataset (ie each epoch) The gradient estimated on a subset of data is cipped to have a maximum vaue of 5 as in Chen et a (6) for each dimension to prevent updates from a arge gradient vaue to bow up the objective oss For JSB we use a stepsize of e 3 for SGMGT, for other three datasets (Piano, Muse, Nott) we use a stepsize of 3e 3 For SGLD, we use a stepsize of e 3, for SGNHT the stepsize is set as 5e 5 The other hyperparameters are provided in Tabe 5 H Additiona figure for RNN experiments We provide the tracepot of one parameter in RNN experiment of JSB dataset We choose this parameter at random Generay, the SGMGT with a =seems to demonstrate more random wak behavior than SGMGT with a = Parameter vaue SGMGT (a=) SGMGT (a=) SGMGT-D (a=) SGMGT-D (a=) 4 # of iterations Figure 6 Tracepot for RNN experiments I Additiona resuts for RNN experiments Here we provide the resuts of severa optimization methods, the resuts are taken from Chen et a (6) Tabe 6 Test negative og-ikeihood resuts on poyphonic music datasets using RNN Agorithms Piano Nott Muse JSB Adam RMSprop SGD-M SGD HF SGD-M J Convergence property Proof This foows the proof for genera SG-MCMC agorithms Specificay, in SGMGT, the generator of the corresponding SDE is defined as: where Lf(x), F (x) r T : rr T f(x), x =(, p, ), ru( )rk c (p) F (x) = rũ( ) ( p rf ( ))rk c (p) (rkc (p)) r K c (p) A, rf ( ) p p = p A p After introducing stochastic gradients, in each iteration t, the generator is perturbed by: V t = rũ( ) ru( ) (r r), such that L t = L V t, where L t is the oca generator for the SDE in iterator t After defining these notation, we foows the proofs of Theorem and Theorem 3 in Chen et a (5) The proof for the bias: Foowing Theorem in Chen et a (5), in the decreasing step size setting, the spit fow can be written as: E ( (X h )) = I h L (X ( )h ) KX k= h k k! L (X ( )h )O(h K )

4 Simiary, the expected difference between and can be simpified using the step size sequence (h ) as: E (6) = (E ( (X Lh )) (X )) (7) KX P h k L Lk k!s (X ( )h )O( hk ) (8) L k= Simiar to the derivation in Chen et a (5), we can derive the foowing bounds k =(,,K): h k E L k (X ( )h ) (9) = O = O (h k h k ) L! k (X ( )h )h K h K! Stochastic Gradient Monomia Gamma Samper () Substitute (9) into (6) and coect ow order terms, we have: E () = (E ( (X Lh )) (X )) O( hk ) As a resut, the bias can be expressed as: () E appe (E [ (X Lh )] (X )) O( hk ) =O hk ) hk Taking L!, both terms go to zero by assumption The proof for the MSE: Foowing simiar derivations as in Theorem in Chen et a (5), we have that E ( (X h )) =! (X ( )h ) h L (X ( )h ) h V (X ( )h ) KX k= h k k! L k (X ( )h )C h K Substitute the Poisson equation into the above equation and divided both sides by, we have ˆ E (X Lh ) (x ) = E (X ( )h ) (X ( )h ) KX k= h V (X ( )h ) h k k! Lk (X ( )h )C h3 As a resut, there exists some positive constant C, such that: E ˆ B appe CE ( (X ) E (X Lh )) (3) {z } A E (X ( )h ) (X ( )h ) {z } A KX h S L h3 k V k! h k! Lk k!s (X ( )h ) L k= {z } A 3 A (4) A can be bounded by assumptions, and A is shown to be bounded by using the fact that E (X ( )h ) (X ( )h ) = O( p h ) from Theorem in Chen et a (5) Furthermore, simiar to the proof of Theorem in Chen et a (5), the expectation of A 3 can aso be bounded by using the formua E[X ] = (E X) E[(X E X) ] and (9) It turns out that the resuting terms have order higher than those from the other terms, thus can be ignored in the expression beow After some simpifications, (3) is bounded by: E ˆ (5) X h = C X E k V k h P! L hk E k V k ( hk )! (6) for some C>, this competes the first part of the theorem We can see that according to the assumption, the ast two

5 Stochastic Gradient Monomia Gamma Samper terms in (5) approach to when L! If we further P assume h to because: X h E k =, then the first term in (5) approaches V k appe sup E k V k P h S L As a resut, we have im L! E ˆ =! Let the number of sampes in each resamping period to be (T ) L, and denote T, T Further denote the sampe average in the -th resamping period to be: ˆT, XT T (x (T ) ), where {x (T ) } denotes sampes in the -th resamping period The fina sampe average is defined as: K Proof for Lemma To prove Lemma, we first introduce the foowing emma from Geyer (5) Lemma 3 (Geyer (5)) Suppose µ is a probabiity distribution and for each z in the domain of domain of µ there is a Markov kerne P z satisfying = P z, and suppose that the map (z,x) P z (x, A) is jointy measurabe for each A Then Z Q(x, A) = µ(dz)p z (x, A) defines a kerne Q that is Markov and satisfies = Q Proof Detaied proof can be found in Chapter 3 of Geyer (5) ˆT, T P T ˆT = As a resut, the bias can be bounded as: E ˆT = P T appe X = X = X X L T = E P T ˆT = X T E ˆT T P T E ˆT T T O P T h h T O P h T h Now it is ready to prove Lemma Proof of Lemma First, we note that the momentum (or other auxiiary variabes) is resamped from the stationary distribution of the Itô diffusion As a resut, for each mode parameter, it corresponds to a Markov kerne P with the stationary Gaussian density According to Lemma 3, the composition of the numerica integrator in SGMGT and the resamping forms a Markov kerne Q(, A), such that h = h Q The above equation means that h is aso the stationary distribution of the Markov kerne Q, which competes the proof L Proof for Lemma Proof First, the optima bias and MSE bounds in Proposition are given by: Bias: E ˆT MSE: E ˆ = O T /, = O T /3 Optimizing over h, we have appeo E ˆT = X ( P T ) / P T T O P T / = O T /, which is the same as the optima bias bound for SGMGT The proof for the optima MSE bound foows simiary M Stochastic sice samping In this section, we everage the connection between sice samping and HMC (Zhang et a, 6), to investigate the approach to perform sice samping with subset of data Sice samping (Nea, 3) augments the density p( )/C (where C> is a normaization constant) with sice variabes u, such that the joint distribution p(, u) =/C, st < u < p( ) To sampe from the target distribution, sice samping is performed in a Gibbs samping manner, ie, aternating between uniformy samping the sice variabe (sice samping step) u, and uniformy generating new

6 Stochastic Gradient Monomia Gamma Samper sampes (conditiona samping step), from a restricted domain such that the (unnormaized) density function vaues for are ess than the samped sice variabe u Sice samping aows moves that can adaptivey fit the scae of the oca density structure, thus yieding rapid mixing When the dataset is arge, however, fu-data density evauations can be very expensive One recent attempt to use subset data for sice samping incorporates a hypothesis test sub-procedure when performing the conditiona samping step (DuBois et a, 4) However, the rejection rate coud be arge if the mini-batch size is sma Furthermore, sampes from the agorithm are biased due to the hypothesis test step One straightforward approach to perform stochastic sice samping is by evauating the ikeihood on a subset of data during the conditiona samping step when performing standard sice samping This approach, detaied in the SM is referred as naïve stochastic sice samping (Naïve stochastic SS) As shown in Figure 7 in the SM, appying this naïve impementation to a Bayesian inear regression probem woud yied over-dispersed sampes The reason why naïve stochastic sice samping fais can be expained by foowing the ogic of Zhang et a (6) and Chen et a (4); Betancourt (5) In Zhang et a (6), the authors demonstrate the connection between sice samping and Hamitonian Monte Caro, reveaed by Hamitonian-Jacobi Equation As a resut, performing sice samping can be equivaenty reaized in an HMC formuation We consider mapping naïve stochastic sice samping to its equivaent HMC space parameterized by mode parameter and momentum p as in (7) (where the monomia parameter a =, with notation from Zhang et a (6)) This resuts in an HMC formuation that is equivaent to the naïve stochastic gradient HMC in Chen et a (4), but with different kinetic function, as in (3) when a = Simiar to (Chen et a, 4), the entropy of the joint distribution of (, p) woud aways increase due to the stochastic noise, expaining the over-dispersion distribution that we observe in Figure 7 in the SM Fortunatey, one can everage the connection between sice samping and HMC from Zhang et a (6) to perform an improved stochastic sice samping This is done by adopting the SDE of SGHMC in (7) and substituting the Gaussian kinetic with a softened Lapace kinetic (ie K c (p) when a =) as in () A friction term ArK c (p) is incorporate to offset the stochastic noise, resuting in d = rk c (p)dt, (7) q dp = [rũ( )ArK c(p)]dt (AI ˆB( ))dw The resuting stochastic Lapace HMC agorithm (detaied in the SM) from (7) is (asymptoticay) invariant to the target distribution, and performs equivaenty to a correct stochastic sice samping in one-dimensiona cases, as c! In Figure 7, the stochastic Lapace HMC samper can ameiorate the over-dispersion of samped posterior distribution than naïve stochastic sice samping N Naive stochastic sice samping and Stochastic Lapacian HMC The naïve stochastic sice samping can be described in Agorithm Agorithm Naïve stochastic SS Input: Initia parameter for t =,, do Samping a mini-batch x t Evauate stochastic negative og-density Ũ x t ( t ), N exp[ og p( t ) N Px x t og p(x t )] Uniformy sampe u t from (, exp[ Ũ xt ( t )]) Sampe t from { : Ũ x t ( ) < og( u t )} using doubing and shrinking (Nea, 3) According to Zhang et a (6), Agorithm has deep connection to Agorithm 3 in HMC formuation, in univariate scenarios Agorithm 3 Naïve stochastic SS in HMC space Input: Initia parameter for t =,, do Samping a mini-batch x t Samping each momentum p independenty (for each dimension) from a Lapacian distribution L(m), where m>is the mass parameter for s =,, do Evauate stochastic gradient, rũ( ), from (5) on mini-batch x t Perform eap-frog updates using (4) by substituting the ru( ) with rũ( ) and substituting rk(p) with sign(p)/m By adding a friction term as in Chen et a (4) we provide the corrected SG-MCMC Agorithm 3 that corresponds to stochastic sice samping

7 Stochastic Gradient Monomia Gamma Samper Stochastic Lapace HMC Naive stochastic SS Orace density Probabity density Figure 7 Naïve stochastic sice samping vs Stochastic Lapacian HMC Agorithm 4 Stochastic Lapacian HMC Input: Initia parameter for t =,, do Samping a momentum p from a distribution / exp( K c (p)) with K c (p) defined in (), where c is the softened parameter for s =,, do Samping a mini-batch x t Evauating stochastic gradient rũ( ) from (5) on mini-batch x t Performing eap-frog updating using SDE in (7) We provide the empirica density drawn by naïve stochastic sice samping and stochastic Lapacian HMC on a synthetic Bayesian inear regression probem with one feature dimension For each instance i, y i N( x i, ) We estimate the posterior of the singe parameter The synthetic dataset has training sampes We use a minibatch size of 3 for each method, and coect, Monte Caro iterations For stochastic Lapacian HMC we use a stepsize of the diffusion parameter A is set to be 7 and the soften parameter is set to be From Figure 7, the stochastic Lapacian HMC can better recover the target distribution

Appendix for Stochastic Gradient Monomial Gamma Sampler

Appendix for Stochastic Gradient Monomial Gamma Sampler 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 3 3 33 34 35 36 37 38 39 4 4 4 43 44 45 46 47 48 49 5 5 5 53 54 Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing

More information

A. Distribution of the test statistic

A. Distribution of the test statistic A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch

More information

Stochastic Variational Inference with Gradient Linearization

Stochastic Variational Inference with Gradient Linearization Stochastic Variationa Inference with Gradient Linearization Suppementa Materia Tobias Pötz * Anne S Wannenwetsch Stefan Roth Department of Computer Science, TU Darmstadt Preface In this suppementa materia,

More information

6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17. Solution 7

6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17. Solution 7 6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17 Soution 7 Probem 1: Generating Random Variabes Each part of this probem requires impementation in MATLAB. For the

More information

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones ASummaryofGaussianProcesses Coryn A.L. Baier-Jones Cavendish Laboratory University of Cambridge caj@mrao.cam.ac.uk Introduction A genera prediction probem can be posed as foows. We consider that the variabe

More information

arxiv: v1 [stat.ml] 21 Oct 2016

arxiv: v1 [stat.ml] 21 Oct 2016 On the Convergence of Stochastic Gradient MCMC Agorithms with High-Order Integrators arxiv:60.06665v [stat.m] 2 Oct 206 Changyou Chen Nan Ding awrence Carin Dept. of Eectrica and Computer Engineering,

More information

MONTE CARLO SIMULATIONS

MONTE CARLO SIMULATIONS MONTE CARLO SIMULATIONS Current physics research 1) Theoretica 2) Experimenta 3) Computationa Monte Caro (MC) Method (1953) used to study 1) Discrete spin systems 2) Fuids 3) Poymers, membranes, soft matter

More information

STA 216 Project: Spline Approach to Discrete Survival Analysis

STA 216 Project: Spline Approach to Discrete Survival Analysis : Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view

More information

Bridging the Gap between Stochastic Gradient MCMC and Stochastic Optimization

Bridging the Gap between Stochastic Gradient MCMC and Stochastic Optimization Bridging the Gap between Stochastic Gradient MCMC and Stochastic Optimization Changyou Chen, David Carlson, Zhe Gan, Chunyuan Li, Lawrence Carin May 2, 2016 1 Changyou Chen Bridging the Gap between Stochastic

More information

Expectation-Maximization for Estimating Parameters for a Mixture of Poissons

Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Brandon Maone Department of Computer Science University of Hesini February 18, 2014 Abstract This document derives, in excrutiating

More information

Math 124B January 31, 2012

Math 124B January 31, 2012 Math 124B January 31, 212 Viktor Grigoryan 7 Inhomogeneous boundary vaue probems Having studied the theory of Fourier series, with which we successfuy soved boundary vaue probems for the homogeneous heat

More information

Do Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix

Do Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix VOL. NO. DO SCHOOLS MATTER FOR HIGH MATH ACHIEVEMENT? 43 Do Schoos Matter for High Math Achievement? Evidence from the American Mathematics Competitions Genn Eison and Ashey Swanson Onine Appendix Appendix

More information

Jackson 4.10 Homework Problem Solution Dr. Christopher S. Baird University of Massachusetts Lowell

Jackson 4.10 Homework Problem Solution Dr. Christopher S. Baird University of Massachusetts Lowell Jackson 4.10 Homework Probem Soution Dr. Christopher S. Baird University of Massachusetts Lowe PROBLEM: Two concentric conducting spheres of inner and outer radii a and b, respectivey, carry charges ±.

More information

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with? Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine

More information

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After

More information

A Brief Introduction to Markov Chains and Hidden Markov Models

A Brief Introduction to Markov Chains and Hidden Markov Models A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,

More information

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model Appendix of the Paper The Roe of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Mode Caio Ameida cameida@fgv.br José Vicente jose.vaentim@bcb.gov.br June 008 1 Introduction In this

More information

Learning Fully Observed Undirected Graphical Models

Learning Fully Observed Undirected Graphical Models Learning Fuy Observed Undirected Graphica Modes Sides Credit: Matt Gormey (2016) Kayhan Batmangheich 1 Machine Learning The data inspires the structures we want to predict Inference finds {best structure,

More information

Separation of Variables and a Spherical Shell with Surface Charge

Separation of Variables and a Spherical Shell with Surface Charge Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation

More information

SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS

SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS ISEE 1 SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS By Yingying Fan and Jinchi Lv University of Southern Caifornia This Suppementary Materia

More information

Inductive Bias: How to generalize on novel data. CS Inductive Bias 1

Inductive Bias: How to generalize on novel data. CS Inductive Bias 1 Inductive Bias: How to generaize on nove data CS 478 - Inductive Bias 1 Overfitting Noise vs. Exceptions CS 478 - Inductive Bias 2 Non-Linear Tasks Linear Regression wi not generaize we to the task beow

More information

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries c 26 Noninear Phenomena in Compex Systems First-Order Corrections to Gutzwier s Trace Formua for Systems with Discrete Symmetries Hoger Cartarius, Jörg Main, and Günter Wunner Institut für Theoretische

More information

NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS

NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS TONY ALLEN, EMILY GEBHARDT, AND ADAM KLUBALL 3 ADVISOR: DR. TIFFANY KOLBA 4 Abstract. The phenomenon of noise-induced stabiization occurs

More information

Integrating Factor Methods as Exponential Integrators

Integrating Factor Methods as Exponential Integrators Integrating Factor Methods as Exponentia Integrators Borisav V. Minchev Department of Mathematica Science, NTNU, 7491 Trondheim, Norway Borko.Minchev@ii.uib.no Abstract. Recenty a ot of effort has been

More information

<C 2 2. λ 2 l. λ 1 l 1 < C 1

<C 2 2. λ 2 l. λ 1 l 1 < C 1 Teecommunication Network Contro and Management (EE E694) Prof. A. A. Lazar Notes for the ecture of 7/Feb/95 by Huayan Wang (this document was ast LaT E X-ed on May 9,995) Queueing Primer for Muticass Optima

More information

Efficiently Generating Random Bits from Finite State Markov Chains

Efficiently Generating Random Bits from Finite State Markov Chains 1 Efficienty Generating Random Bits from Finite State Markov Chains Hongchao Zhou and Jehoshua Bruck, Feow, IEEE Abstract The probem of random number generation from an uncorreated random source (of unknown

More information

Some Measures for Asymmetry of Distributions

Some Measures for Asymmetry of Distributions Some Measures for Asymmetry of Distributions Georgi N. Boshnakov First version: 31 January 2006 Research Report No. 5, 2006, Probabiity and Statistics Group Schoo of Mathematics, The University of Manchester

More information

(This is a sample cover image for this issue. The actual cover is not yet available at this time.)

(This is a sample cover image for this issue. The actual cover is not yet available at this time.) (This is a sampe cover image for this issue The actua cover is not yet avaiabe at this time) This artice appeared in a journa pubished by Esevier The attached copy is furnished to the author for interna

More information

Lecture Note 3: Stationary Iterative Methods

Lecture Note 3: Stationary Iterative Methods MATH 5330: Computationa Methods of Linear Agebra Lecture Note 3: Stationary Iterative Methods Xianyi Zeng Department of Mathematica Sciences, UTEP Stationary Iterative Methods The Gaussian eimination (or

More information

17 Lecture 17: Recombination and Dark Matter Production

17 Lecture 17: Recombination and Dark Matter Production PYS 652: Astrophysics 88 17 Lecture 17: Recombination and Dark Matter Production New ideas pass through three periods: It can t be done. It probaby can be done, but it s not worth doing. I knew it was

More information

Statistical Astronomy

Statistical Astronomy Lectures for the 7 th IAU ISYA Ifrane, nd 3 rd Juy 4 p ( x y, I) p( y x, I) p( x, I) p( y, I) Statistica Astronomy Martin Hendry, Dept of Physics and Astronomy University of Gasgow, UK http://www.astro.ga.ac.uk/users/martin/isya/

More information

Joel Broida UCSD Fall 2009 Phys 130B QM II. Homework Set 2 DO ALL WORK BY HAND IN OTHER WORDS, DON T USE MATHEMAT- ICA OR ANY CALCULATORS.

Joel Broida UCSD Fall 2009 Phys 130B QM II. Homework Set 2 DO ALL WORK BY HAND IN OTHER WORDS, DON T USE MATHEMAT- ICA OR ANY CALCULATORS. Joe Broida UCSD Fa 009 Phys 30B QM II Homework Set DO ALL WORK BY HAND IN OTHER WORDS, DON T USE MATHEMAT- ICA OR ANY CALCULATORS. You may need to use one or more of these: Y 0 0 = 4π Y 0 = 3 4π cos Y

More information

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA) 1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using

More information

LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL HARMONICS

LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL HARMONICS MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Physics 8.07: Eectromagnetism II October 7, 202 Prof. Aan Guth LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL

More information

A proposed nonparametric mixture density estimation using B-spline functions

A proposed nonparametric mixture density estimation using B-spline functions A proposed nonparametric mixture density estimation using B-spine functions Atizez Hadrich a,b, Mourad Zribi a, Afif Masmoudi b a Laboratoire d Informatique Signa et Image de a Côte d Opae (LISIC-EA 4491),

More information

FFTs in Graphics and Vision. Spherical Convolution and Axial Symmetry Detection

FFTs in Graphics and Vision. Spherical Convolution and Axial Symmetry Detection FFTs in Graphics and Vision Spherica Convoution and Axia Symmetry Detection Outine Math Review Symmetry Genera Convoution Spherica Convoution Axia Symmetry Detection Math Review Symmetry: Given a unitary

More information

Week 6 Lectures, Math 6451, Tanveer

Week 6 Lectures, Math 6451, Tanveer Fourier Series Week 6 Lectures, Math 645, Tanveer In the context of separation of variabe to find soutions of PDEs, we encountered or and in other cases f(x = f(x = a 0 + f(x = a 0 + b n sin nπx { a n

More information

AALBORG UNIVERSITY. The distribution of communication cost for a mobile service scenario. Jesper Møller and Man Lung Yiu. R June 2009

AALBORG UNIVERSITY. The distribution of communication cost for a mobile service scenario. Jesper Møller and Man Lung Yiu. R June 2009 AALBORG UNIVERSITY The distribution of communication cost for a mobie service scenario by Jesper Møer and Man Lung Yiu R-29-11 June 29 Department of Mathematica Sciences Aaborg University Fredrik Bajers

More information

Mat 1501 lecture notes, penultimate installment

Mat 1501 lecture notes, penultimate installment Mat 1501 ecture notes, penutimate instament 1. bounded variation: functions of a singe variabe optiona) I beieve that we wi not actuay use the materia in this section the point is mainy to motivate the

More information

Nonlinear Gaussian Filtering via Radial Basis Function Approximation

Nonlinear Gaussian Filtering via Radial Basis Function Approximation 51st IEEE Conference on Decision and Contro December 10-13 01 Maui Hawaii USA Noninear Gaussian Fitering via Radia Basis Function Approximation Huazhen Fang Jia Wang and Raymond A de Caafon Abstract This

More information

Stochastic Gradient MCMC with Stale Gradients

Stochastic Gradient MCMC with Stale Gradients Stochastic Gradient MCMC with Stae Gradients Changyou Chen Nan Ding Chunyuan i Yizhe Zhang awrence Carin Dept. of Eectrica and Computer Engineering, Duke University, Durham, NC, USA Googe Inc., Venice,

More information

Alberto Maydeu Olivares Instituto de Empresa Marketing Dept. C/Maria de Molina Madrid Spain

Alberto Maydeu Olivares Instituto de Empresa Marketing Dept. C/Maria de Molina Madrid Spain CORRECTIONS TO CLASSICAL PROCEDURES FOR ESTIMATING THURSTONE S CASE V MODEL FOR RANKING DATA Aberto Maydeu Oivares Instituto de Empresa Marketing Dept. C/Maria de Moina -5 28006 Madrid Spain Aberto.Maydeu@ie.edu

More information

Testing for the Existence of Clusters

Testing for the Existence of Clusters Testing for the Existence of Custers Caudio Fuentes and George Casea University of Forida November 13, 2008 Abstract The detection and determination of custers has been of specia interest, among researchers

More information

1D Heat Propagation Problems

1D Heat Propagation Problems Chapter 1 1D Heat Propagation Probems If the ambient space of the heat conduction has ony one dimension, the Fourier equation reduces to the foowing for an homogeneous body cρ T t = T λ 2 + Q, 1.1) x2

More information

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING 8 APPENDIX 8.1 EMPIRICAL EVALUATION OF SAMPLING We wish to evauate the empirica accuracy of our samping technique on concrete exampes. We do this in two ways. First, we can sort the eements by probabiity

More information

AST 418/518 Instrumentation and Statistics

AST 418/518 Instrumentation and Statistics AST 418/518 Instrumentation and Statistics Cass Website: http://ircamera.as.arizona.edu/astr_518 Cass Texts: Practica Statistics for Astronomers, J.V. Wa, and C.R. Jenkins, Second Edition. Measuring the

More information

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES Separation of variabes is a method to sove certain PDEs which have a warped product structure. First, on R n, a inear PDE of order m is

More information

Quantum Mechanical Models of Vibration and Rotation of Molecules Chapter 18

Quantum Mechanical Models of Vibration and Rotation of Molecules Chapter 18 Quantum Mechanica Modes of Vibration and Rotation of Moecues Chapter 18 Moecuar Energy Transationa Vibrationa Rotationa Eectronic Moecuar Motions Vibrations of Moecues: Mode approximates moecues to atoms

More information

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com

More information

Rate-Distortion Theory of Finite Point Processes

Rate-Distortion Theory of Finite Point Processes Rate-Distortion Theory of Finite Point Processes Günther Koiander, Dominic Schuhmacher, and Franz Hawatsch, Feow, IEEE Abstract We study the compression of data in the case where the usefu information

More information

14 Separation of Variables Method

14 Separation of Variables Method 14 Separation of Variabes Method Consider, for exampe, the Dirichet probem u t = Du xx < x u(x, ) = f(x) < x < u(, t) = = u(, t) t > Let u(x, t) = T (t)φ(x); now substitute into the equation: dt

More information

The distribution of the number of nodes in the relative interior of the typical I-segment in homogeneous planar anisotropic STIT Tessellations

The distribution of the number of nodes in the relative interior of the typical I-segment in homogeneous planar anisotropic STIT Tessellations Comment.Math.Univ.Caroin. 51,3(21) 53 512 53 The distribution of the number of nodes in the reative interior of the typica I-segment in homogeneous panar anisotropic STIT Tesseations Christoph Thäe Abstract.

More information

Model-based Clustering by Probabilistic Self-organizing Maps

Model-based Clustering by Probabilistic Self-organizing Maps EEE TRANATN N NERA NETRK, V. XX, N., ode-based ustering by robabiistic ef-organizing aps hih-ian heng, Hsin-hia u, ember, EEE, and Hsin-in ang, ember, EEE Abstract n this paper, we consider the earning

More information

Math 124B January 17, 2012

Math 124B January 17, 2012 Math 124B January 17, 212 Viktor Grigoryan 3 Fu Fourier series We saw in previous ectures how the Dirichet and Neumann boundary conditions ead to respectivey sine and cosine Fourier series of the initia

More information

6 Wave Equation on an Interval: Separation of Variables

6 Wave Equation on an Interval: Separation of Variables 6 Wave Equation on an Interva: Separation of Variabes 6.1 Dirichet Boundary Conditions Ref: Strauss, Chapter 4 We now use the separation of variabes technique to study the wave equation on a finite interva.

More information

Steepest Descent Adaptation of Min-Max Fuzzy If-Then Rules 1

Steepest Descent Adaptation of Min-Max Fuzzy If-Then Rules 1 Steepest Descent Adaptation of Min-Max Fuzzy If-Then Rues 1 R.J. Marks II, S. Oh, P. Arabshahi Λ, T.P. Caude, J.J. Choi, B.G. Song Λ Λ Dept. of Eectrica Engineering Boeing Computer Services University

More information

Technical Appendix for Voting, Speechmaking, and the Dimensions of Conflict in the US Senate

Technical Appendix for Voting, Speechmaking, and the Dimensions of Conflict in the US Senate Technica Appendix for Voting, Speechmaking, and the Dimensions of Confict in the US Senate In Song Kim John Londregan Marc Ratkovic Juy 6, 205 Abstract We incude here severa technica appendices. First,

More information

Two-Stage Least Squares as Minimum Distance

Two-Stage Least Squares as Minimum Distance Two-Stage Least Squares as Minimum Distance Frank Windmeijer Discussion Paper 17 / 683 7 June 2017 Department of Economics University of Bristo Priory Road Compex Bristo BS8 1TU United Kingdom Two-Stage

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Schoo of Computer Science Probabiistic Graphica Modes Gaussian graphica modes and Ising modes: modeing networks Eric Xing Lecture 0, February 0, 07 Reading: See cass website Eric Xing @ CMU, 005-07 Network

More information

Active Learning & Experimental Design

Active Learning & Experimental Design Active Learning & Experimenta Design Danie Ting Heaviy modified, of course, by Lye Ungar Origina Sides by Barbara Engehardt and Aex Shyr Lye Ungar, University of Pennsyvania Motivation u Data coection

More information

PARSIMONIOUS VARIATIONAL-BAYES MIXTURE AGGREGATION WITH A POISSON PRIOR. Pierrick Bruneau, Marc Gelgon and Fabien Picarougne

PARSIMONIOUS VARIATIONAL-BAYES MIXTURE AGGREGATION WITH A POISSON PRIOR. Pierrick Bruneau, Marc Gelgon and Fabien Picarougne 17th European Signa Processing Conference (EUSIPCO 2009) Gasgow, Scotand, August 24-28, 2009 PARSIMONIOUS VARIATIONAL-BAYES MIXTURE AGGREGATION WITH A POISSON PRIOR Pierric Bruneau, Marc Gegon and Fabien

More information

Haar Decomposition and Reconstruction Algorithms

Haar Decomposition and Reconstruction Algorithms Jim Lambers MAT 773 Fa Semester 018-19 Lecture 15 and 16 Notes These notes correspond to Sections 4.3 and 4.4 in the text. Haar Decomposition and Reconstruction Agorithms Decomposition Suppose we approximate

More information

arxiv: v1 [math.fa] 23 Aug 2018

arxiv: v1 [math.fa] 23 Aug 2018 An Exact Upper Bound on the L p Lebesgue Constant and The -Rényi Entropy Power Inequaity for Integer Vaued Random Variabes arxiv:808.0773v [math.fa] 3 Aug 08 Peng Xu, Mokshay Madiman, James Mebourne Abstract

More information

Notes on Backpropagation with Cross Entropy

Notes on Backpropagation with Cross Entropy Notes on Backpropagation with Cross Entropy I-Ta ee, Dan Gowasser, Bruno Ribeiro Purue University October 3, 07. Overview This note introuces backpropagation for a common neura network muti-cass cassifier.

More information

(f) is called a nearly holomorphic modular form of weight k + 2r as in [5].

(f) is called a nearly holomorphic modular form of weight k + 2r as in [5]. PRODUCTS OF NEARLY HOLOMORPHIC EIGENFORMS JEFFREY BEYERL, KEVIN JAMES, CATHERINE TRENTACOSTE, AND HUI XUE Abstract. We prove that the product of two neary hoomorphic Hece eigenforms is again a Hece eigenform

More information

A Sparse Covariance Function for Exact Gaussian Process Inference in Large Datasets

A Sparse Covariance Function for Exact Gaussian Process Inference in Large Datasets A Covariance Function for Exact Gaussian Process Inference in Large Datasets Arman ekumyan Austraian Centre for Fied Robotics The University of Sydney NSW 26, Austraia a.mekumyan@acfr.usyd.edu.au Fabio

More information

Copyright information to be inserted by the Publishers. Unsplitting BGK-type Schemes for the Shallow. Water Equations KUN XU

Copyright information to be inserted by the Publishers. Unsplitting BGK-type Schemes for the Shallow. Water Equations KUN XU Copyright information to be inserted by the Pubishers Unspitting BGK-type Schemes for the Shaow Water Equations KUN XU Mathematics Department, Hong Kong University of Science and Technoogy, Cear Water

More information

u(x) s.t. px w x 0 Denote the solution to this problem by ˆx(p, x). In order to obtain ˆx we may simply solve the standard problem max x 0

u(x) s.t. px w x 0 Denote the solution to this problem by ˆx(p, x). In order to obtain ˆx we may simply solve the standard problem max x 0 Bocconi University PhD in Economics - Microeconomics I Prof M Messner Probem Set 4 - Soution Probem : If an individua has an endowment instead of a monetary income his weath depends on price eves In particuar,

More information

Akaike Information Criterion for ANOVA Model with a Simple Order Restriction

Akaike Information Criterion for ANOVA Model with a Simple Order Restriction Akaike Information Criterion for ANOVA Mode with a Simpe Order Restriction Yu Inatsu * Department of Mathematics, Graduate Schoo of Science, Hiroshima University ABSTRACT In this paper, we consider Akaike

More information

The Binary Space Partitioning-Tree Process Supplementary Material

The Binary Space Partitioning-Tree Process Supplementary Material The inary Space Partitioning-Tree Process Suppementary Materia Xuhui Fan in Li Scott. Sisson Schoo of omputer Science Fudan University ibin@fudan.edu.cn Schoo of Mathematics and Statistics University of

More information

Asymptotic Properties of a Generalized Cross Entropy Optimization Algorithm

Asymptotic Properties of a Generalized Cross Entropy Optimization Algorithm 1 Asymptotic Properties of a Generaized Cross Entropy Optimization Agorithm Zijun Wu, Michae Koonko, Institute for Appied Stochastics and Operations Research, Caustha Technica University Abstract The discrete

More information

Introduction to Simulation - Lecture 13. Convergence of Multistep Methods. Jacob White. Thanks to Deepak Ramaswamy, Michal Rewienski, and Karen Veroy

Introduction to Simulation - Lecture 13. Convergence of Multistep Methods. Jacob White. Thanks to Deepak Ramaswamy, Michal Rewienski, and Karen Veroy Introduction to Simuation - Lecture 13 Convergence of Mutistep Methods Jacob White Thans to Deepa Ramaswamy, Micha Rewiensi, and Karen Veroy Outine Sma Timestep issues for Mutistep Methods Loca truncation

More information

High Spectral Resolution Infrared Radiance Modeling Using Optimal Spectral Sampling (OSS) Method

High Spectral Resolution Infrared Radiance Modeling Using Optimal Spectral Sampling (OSS) Method High Spectra Resoution Infrared Radiance Modeing Using Optima Spectra Samping (OSS) Method J.-L. Moncet and G. Uymin Background Optima Spectra Samping (OSS) method is a fast and accurate monochromatic

More information

From Margins to Probabilities in Multiclass Learning Problems

From Margins to Probabilities in Multiclass Learning Problems From Margins to Probabiities in Muticass Learning Probems Andrea Passerini and Massimiiano Ponti 2 and Paoo Frasconi 3 Abstract. We study the probem of muticass cassification within the framework of error

More information

Partial permutation decoding for MacDonald codes

Partial permutation decoding for MacDonald codes Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics

More information

HYDROGEN ATOM SELECTION RULES TRANSITION RATES

HYDROGEN ATOM SELECTION RULES TRANSITION RATES DOING PHYSICS WITH MATLAB QUANTUM PHYSICS Ian Cooper Schoo of Physics, University of Sydney ian.cooper@sydney.edu.au HYDROGEN ATOM SELECTION RULES TRANSITION RATES DOWNLOAD DIRECTORY FOR MATLAB SCRIPTS

More information

Homework 5 Solutions

Homework 5 Solutions Stat 310B/Math 230B Theory of Probabiity Homework 5 Soutions Andrea Montanari Due on 2/19/2014 Exercise [5.3.20] 1. We caim that n 2 [ E[h F n ] = 2 n i=1 A i,n h(u)du ] I Ai,n (t). (1) Indeed, integrabiity

More information

XSAT of linear CNF formulas

XSAT of linear CNF formulas XSAT of inear CN formuas Bernd R. Schuh Dr. Bernd Schuh, D-50968 Kön, Germany; bernd.schuh@netcoogne.de eywords: compexity, XSAT, exact inear formua, -reguarity, -uniformity, NPcompeteness Abstract. Open

More information

A MODEL FOR ESTIMATING THE LATERAL OVERLAP PROBABILITY OF AIRCRAFT WITH RNP ALERTING CAPABILITY IN PARALLEL RNAV ROUTES

A MODEL FOR ESTIMATING THE LATERAL OVERLAP PROBABILITY OF AIRCRAFT WITH RNP ALERTING CAPABILITY IN PARALLEL RNAV ROUTES 6 TH INTERNATIONAL CONGRESS OF THE AERONAUTICAL SCIENCES A MODEL FOR ESTIMATING THE LATERAL OVERLAP PROBABILITY OF AIRCRAFT WITH RNP ALERTING CAPABILITY IN PARALLEL RNAV ROUTES Sakae NAGAOKA* *Eectronic

More information

Global sensitivity analysis using low-rank tensor approximations

Global sensitivity analysis using low-rank tensor approximations Goba sensitivity anaysis using ow-rank tensor approximations K. Konaki 1 and B. Sudret 1 1 Chair of Risk, Safety and Uncertainty Quantification, arxiv:1605.09009v1 [stat.co] 29 May 2016 ETH Zurich, Stefano-Franscini-Patz

More information

Ant Colony Algorithms for Constructing Bayesian Multi-net Classifiers

Ant Colony Algorithms for Constructing Bayesian Multi-net Classifiers Ant Coony Agorithms for Constructing Bayesian Muti-net Cassifiers Khaid M. Saama and Aex A. Freitas Schoo of Computing, University of Kent, Canterbury, UK. {kms39,a.a.freitas}@kent.ac.uk December 5, 2013

More information

Learning Structured Weight Uncertainty in Bayesian Neural Networks

Learning Structured Weight Uncertainty in Bayesian Neural Networks Learning Structured Weight Uncertainty in Bayesian Neura Networks Shengyang Sun Changyou Chen Lawrence Carin Tsinghua University Duke University Duke University Abstract Deep neura networks DNNs are increasingy

More information

Explicit overall risk minimization transductive bound

Explicit overall risk minimization transductive bound 1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,

More information

An approximate method for solving the inverse scattering problem with fixed-energy data

An approximate method for solving the inverse scattering problem with fixed-energy data J. Inv. I-Posed Probems, Vo. 7, No. 6, pp. 561 571 (1999) c VSP 1999 An approximate method for soving the inverse scattering probem with fixed-energy data A. G. Ramm and W. Scheid Received May 12, 1999

More information

$, (2.1) n="# #. (2.2)

$, (2.1) n=# #. (2.2) Chapter. Eectrostatic II Notes: Most of the materia presented in this chapter is taken from Jackson, Chap.,, and 4, and Di Bartoo, Chap... Mathematica Considerations.. The Fourier series and the Fourier

More information

Formulas for Angular-Momentum Barrier Factors Version II

Formulas for Angular-Momentum Barrier Factors Version II BNL PREPRINT BNL-QGS-06-101 brfactor1.tex Formuas for Anguar-Momentum Barrier Factors Version II S. U. Chung Physics Department, Brookhaven Nationa Laboratory, Upton, NY 11973 March 19, 2015 abstract A

More information

arxiv:nlin/ v2 [nlin.cd] 30 Jan 2006

arxiv:nlin/ v2 [nlin.cd] 30 Jan 2006 expansions in semicassica theories for systems with smooth potentias and discrete symmetries Hoger Cartarius, Jörg Main, and Günter Wunner arxiv:nin/0510051v [nin.cd] 30 Jan 006 1. Institut für Theoretische

More information

SydU STAT3014 (2015) Second semester Dr. J. Chan 18

SydU STAT3014 (2015) Second semester Dr. J. Chan 18 STAT3014/3914 Appied Stat.-Samping C-Stratified rand. sampe Stratified Random Samping.1 Introduction Description The popuation of size N is divided into mutuay excusive and exhaustive subpopuations caed

More information

FRIEZE GROUPS IN R 2

FRIEZE GROUPS IN R 2 FRIEZE GROUPS IN R 2 MAXWELL STOLARSKI Abstract. Focusing on the Eucidean pane under the Pythagorean Metric, our goa is to cassify the frieze groups, discrete subgroups of the set of isometries of the

More information

Iterative Decoding Performance Bounds for LDPC Codes on Noisy Channels

Iterative Decoding Performance Bounds for LDPC Codes on Noisy Channels Iterative Decoding Performance Bounds for LDPC Codes on Noisy Channes arxiv:cs/060700v1 [cs.it] 6 Ju 006 Chun-Hao Hsu and Achieas Anastasopouos Eectrica Engineering and Computer Science Department University

More information

A NOTE ON QUASI-STATIONARY DISTRIBUTIONS OF BIRTH-DEATH PROCESSES AND THE SIS LOGISTIC EPIDEMIC

A NOTE ON QUASI-STATIONARY DISTRIBUTIONS OF BIRTH-DEATH PROCESSES AND THE SIS LOGISTIC EPIDEMIC (January 8, 2003) A NOTE ON QUASI-STATIONARY DISTRIBUTIONS OF BIRTH-DEATH PROCESSES AND THE SIS LOGISTIC EPIDEMIC DAMIAN CLANCY, University of Liverpoo PHILIP K. POLLETT, University of Queensand Abstract

More information

Statistics for Applications. Chapter 7: Regression 1/43

Statistics for Applications. Chapter 7: Regression 1/43 Statistics for Appications Chapter 7: Regression 1/43 Heuristics of the inear regression (1) Consider a coud of i.i.d. random points (X i,y i ),i =1,...,n : 2/43 Heuristics of the inear regression (2)

More information

Reliability: Theory & Applications No.3, September 2006

Reliability: Theory & Applications No.3, September 2006 REDUNDANCY AND RENEWAL OF SERVERS IN OPENED QUEUING NETWORKS G. Sh. Tsitsiashvii M.A. Osipova Vadivosto, Russia 1 An opened queuing networ with a redundancy and a renewa of servers is considered. To cacuate

More information

SEMINAR 2. PENDULUMS. V = mgl cos θ. (2) L = T V = 1 2 ml2 θ2 + mgl cos θ, (3) d dt ml2 θ2 + mgl sin θ = 0, (4) θ + g l

SEMINAR 2. PENDULUMS. V = mgl cos θ. (2) L = T V = 1 2 ml2 θ2 + mgl cos θ, (3) d dt ml2 θ2 + mgl sin θ = 0, (4) θ + g l Probem 7. Simpe Penduum SEMINAR. PENDULUMS A simpe penduum means a mass m suspended by a string weightess rigid rod of ength so that it can swing in a pane. The y-axis is directed down, x-axis is directed

More information

Incremental Reformulated Automatic Relevance Determination

Incremental Reformulated Automatic Relevance Determination IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 60, NO. 9, SEPTEMBER 22 4977 Incrementa Reformuated Automatic Reevance Determination Dmitriy Shutin, Sanjeev R. Kukarni, and H. Vincent Poor Abstract In this

More information

4 Separation of Variables

4 Separation of Variables 4 Separation of Variabes In this chapter we describe a cassica technique for constructing forma soutions to inear boundary vaue probems. The soution of three cassica (paraboic, hyperboic and eiptic) PDE

More information

Supplementary Appendix (not for publication) for: The Value of Network Information

Supplementary Appendix (not for publication) for: The Value of Network Information Suppementary Appendix not for pubication for: The Vaue of Network Information Itay P. Fainmesser and Andrea Gaeotti September 6, 03 This appendix incudes the proof of Proposition from the paper "The Vaue

More information

Chapter 5. Wave equation. 5.1 Physical derivation

Chapter 5. Wave equation. 5.1 Physical derivation Chapter 5 Wave equation In this chapter, we discuss the wave equation u tt a 2 u = f, (5.1) where a > is a constant. We wi discover that soutions of the wave equation behave in a different way comparing

More information

Chemical Kinetics Part 2

Chemical Kinetics Part 2 Integrated Rate Laws Chemica Kinetics Part 2 The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates the rate

More information