Active Learning & Experimental Design
|
|
- Marilyn Webster
- 5 years ago
- Views:
Transcription
1 Active Learning & Experimenta Design Danie Ting Heaviy modified, of course, by Lye Ungar Origina Sides by Barbara Engehardt and Aex Shyr Lye Ungar, University of Pennsyvania
2 Motivation u Data coection may be expensive Cost of time and materias for an experiment Cheap vs. expensive data n Raw images vs. annotated images u Not a abes are equay usefu u Want to coect the best data at minima cost Where do abes come from? What observations shoud one abe?
3 Toy exampes Assume you are earning y=f(x) for x a scaar You are earning inear regression on [-1,1]. You can pick two x s to get y s for. What two vaues woud you pick? A) -1/3, 1/3 B) -1, 1 C) 0, 1 D) Something ese
4 Toy exampes Assume you are earning y = f(x) for x a scaar You are earning an SVM cassifier on [-1.1]. You can pick 4 x s to get y s for. What strategy woud you use to pick x s? A) Pick -1, -1/3, 1/3, 1 B) Pick -1, 1, see what the answer is, then pick next x C) Pick -1/3, 1/3, see what the answer is, then pick next x D) Something ese
5 Toy Exampe: 1D cassifier x x x x x x x x x x Unabeed data: abes are a 0 then a 1 (eft to right) Cassifier (threshod function): h w (x) = 1 if x > w (0 otherwise) Goa: find transition between 0 and 1 abes in minimum steps Naïve method: choose points to abe at random on ine Requires O(n) training data to find underying cassifier Better method: binary search for transition between 0 and 1 Requires O(og n) training data to find underying cassifier Exponentia reduction in training data size!
6 Exampe: coaborative fitering Users usuay rate ony a few movies ratings are expensive Which movies do you show users to best extrapoate movie preferences? [Yu et a. 2006]
7 Exampe: coaborative fitering Users usuay rate ony a few movies ratings are expensive Which movies do you show users to best extrapoate movie preferences? u Baseine agorithms: Random: m movies randomy Most Popuar Movies: m most frequenty rated movies u Most popuar movies is not better than random design! u Popuar movies rated highy by a users; do not discriminate tastes [Yu et a. 2006]
8 What criteria might you use to seect which items to abe?
9 Active Learning Methods u Active earning Uncertainty samping Query by committee Information-based oss functions u Optima experimenta design u Non-inear optima experimenta design u Summary
10 Active earning u Setup: Given existing knowedge (X,y), choose where to coect more data Assume access to cheap unabeed points Make a query to obtain expensive abe Want to find abes that are informative u Output: Cassifier / predictor trained on ess abeed data u Simiar to active earning in cassrooms Students ask questions, receive a response, and ask further questions Contrast: passive earning: student just istens to ecturer u This ecture covers: how to measure the vaue of data agorithms to choose the data
11 Exampe: Gene expression and Cancer cassification u Active earning takes 31 points to achieve same accuracy as passive earning with 174 Liu 2004
12 Active Learning Setup u Active earner picks which data point x to query u Receive response y from an orace u Update parameters q of the mode u Repeat u Query seected to minimize some oss function ( risk )
13 Active Learning u Heuristic methods for reducing risk: Seect most uncertain data point Seect most informative data point n n for distinguishing different hypotheses to optimize expected information gain u Some computationa considerations: May be many queries to cacuate oss for n n Subsampe points Probabiity far from the true min decreases exponentiay May not be easy to cacuate oss
14 Uncertainty Samping u Query the item (x) that the current cassifier is most uncertain about u Needs measure of uncertainty u Exampes: Entropy Least confident predicted abe Eucidean distance (e.g. point cosest to margin in SVM)
15 Exampe: Gene expression and Cancer cassification u Data: Cancerous ung tissue sampes Cheap unabeed data n gene expression profies from Affymetrix microarray Labeed data: n 0-1 abe for adenocarcinoma or maignant peura mesotheioma u Method: Linear SVM Measure of uncertainty n distance to SVM hyperpane Liu 2004
16 Exampe: Gene expression and Cancer cassification u Active earning takes 31 points to achieve same accuracy as passive earning with 174 Liu 2004
17 Query by Committee u Which unabeed point shoud you choose? a) Red arrow point b) Green arrow point
18 Query by Committee u Yeow = vaid hypotheses
19 Query by Committee u Point on max-margin hyperpane does not reduce the number of vaid hypotheses by much
20 Query by Committee u Queries an exampe based on the degree of disagreement between committee of cassifiers
21 Query by Committee u Prior distribution over cassifiers/hypotheses u Sampe a set of cassifiers from distribution u Natura for ensembe methods which are aready sampes Random forests, Bagged cassifiers, etc. u Measures of disagreement Entropy of predicted responses KL-divergence of predictive distributions
22 Query by Committee Appication u Use Naïve Bayes mode for text cassification (20 Newsgroups dataset) [McCaum & Nigam, 1998]
23 Information-based Loss Function u Previous methods ooked at uncertainty at a singe point Does not ook at whether you can actuay reduce uncertainty or if adding the point makes a difference in the mode u Want to mode notions of information gained Maximize KL divergence between posterior and prior Maximize reduction in mode entropy between posterior and prior (reduce number of bits required to describe distribution) u A of these can be extended to optima design agorithms u Must decide how to hande uncertainty about query response, mode parameters [MacKay, 1992]
24 Entropy u A measure of information in random event X with possibe outcomes {x 1,,x n } H(x) = - Σ i p(x i ) og 2 p(x i ) u The average minimum number of yes/no questions to answer some question Reated to binary search [Shannon, 1948]
25 Kuback Leiber divergence u P = true distribution; u Q = aternative distribution that is used to encode data u KL divergence is the expected extra message ength per datum that must be transmitted using Q D KL (P Q) = Σ i P(x i ) og (P(x i )/Q(x i )) = Σ i P(x i ) og Q(x i ) + Σ i P(x i ) og P(x i ) = H(P,Q) - H(P) = Cross-entropy - entropy u Measures how different the two distributions are
26 KL divergence properties u Non-negative: D(P Q) 0 u Divergence 0 if and ony if P and Q are equa: D(P Q) = 0 iff P = Q u Non-symmetric: D(P Q) D(Q P) u Does not satisfy triange inequaity D(P Q) D(P R) + D(R Q)
27 KL divergence properties u Non-negative: D(P Q) 0 u Divergence 0 if and ony if P and Q are equa: D(P Q) = 0 iff P = Q u Non-symmetric: D(P Q) D(Q P) u Does not satisfy triange inequaity D(P Q) D(P R) + D(R Q) Not a distance metric
28 KL divergence as info gain u The KL divergence of the posteriors measures the information gain expected from query (x ): u Goa: choose a query that maximizes the KL divergence between the updated posterior probabiity and the current posterior probabiity D( p(θ x, x ) p(θ x)) This represents the argest expected information gain
29 Active earning warning u Choice of data is ony as good as the mode itsef u Assume a inear mode, then two data points are sufficient u What happens when data are not inear?
30 Active Learning = Sequentia Experimenta Design u Active earning Uncertainty samping Query by committee Information-based oss functions u Optima experimenta design A-optima design D-optima design E-optima design Non-inear optima experimenta design u Summary
31 Experimenta Design u Many considerations in designing an experiment Deaing with confounders Feasibiity Choice of variabes to measure Size of experiment ( # of data points ) Conduction of experiment Choice of interventions/queries to make Etc.
32 Experimenta Design u Many considerations in designing an experiment Deaing with confounders Feasibiity Choice of variabes to measure Size of experiment ( # of data points ) Conduction of experiment Choice of interventions/queries to make Etc. u We wi ony ook at one of them
33 Optima Experimenta Design? u Heuristics give empiricay good performance but Not that much theory on how good the heuristics are u Optima experimenta design gives theoretica criteria for choosing a set of points to abe for a specific set of assumptions and objectives u Theory is good when you ony get to run (a series of) experiments once
34 Optima Experimenta Design u Given a mode M with parameters β, What queries are maximay informative i.e. wi yied the best estimate of β u Best minimizes variance of estimate Equivaenty, maximizes the Fisher Information Different methods use different ways of computing a scaar from the variance matrix u Linear modes Optima design does not depend on β! u Non-inear modes Depends on β, but can use Tayor expansion to inear mode
35 Goa: Minimize variance of w If y = x T β + ε then w = (X T X) -1 X T y w ~ N(β, σ 2 (X T X) -1 ) We want to minimize the variance of of our parameter estimate w, so pick training data X to minimize (X T X) -1 But that is a matrix, so we need to reduce it to a scaar A-optima (average) design minimizes trace(x T X) -1 D-optima (determinant) design minimizes og det(x T X) -1 E-optima (extreme) design minimizes max eigenvaue of (X T X) -1 Aphabet soup of other criteria (C-, G-, L-, V-, etc.)
36 Fisher Information Matrix The negative of the expected vaue of the Hessian of the og of the probabiity of an observation given the parameters θ. If y = x T w + ε Then p(y t ) ~ exp(-(x tt w ) 2 /2σ 2) m ij = - E[d 2 \dw i dw j (og exp(-(x tt w) 2 /2σ 2 ))] = E[d 2 \dw i dw j ((x tt w) 2 /2σ 2 )] ~ E[x it x j ] So M ~ X T X
37 A-Optima Design u A-optima design minimizes the trace of (X T X) -1 Minimizing trace (sum of diagona eements) essentiay chooses maximay independent coumns (sma correations between interventions) u Tends to choose points on the border of the dataset Exampe: mixture of four Gaussians [Yu et a., 2006]
38 A-Optima Design A-optima design minimizes the trace of (X T X) -1 Exampe: 20 candidate data points, minima eipsoid that contains a points [Boyd & Vandenberghe, 2004
39 D-Optima design u D-optima design minimizes og determinant of (X T X) -1 u Equivaent to choosing the confidence eipsoid with minimum voume ( most powerfu hypothesis test in some sense) Minimizing entropy of the estimated parameters u Most commony used optima design [Boyd & Vandenberghe, 2004
40 E-Optima design u E-optima design minimizes argest eigenvaue of (X T X) -1 u Minimizes the diameter of the confidence eipsoid [Boyd & Vandenberghe, 2004
41 Summary of Optima Design [Boyd & Vandenberghe, 2004
42 Practicaities u Sometimes you can generate an x arbitrariy u More often you need to seect from a set of given x s This can be an expensive search!
43 Experimenta Design u Introduction: information theory u Active earning Uncertainty samping Query by committee Information-based oss functions u Optima experimenta design A-optima design D-optima design E-optima design Non-inear optima experimenta design u Summary
44 Optima design in non-inear modes u Given a non-inear mode y = g(x, θ) u Mode is described by a Tayor expansion around a a j ( x, ) = g(x,θ) / θ j, evauated at u Maximization of Fisher information matrix is now the same as the inear mode u Yieds a ocay optima design, optima for the particuar vaue of θ u Yieds no information on the (ack of) fit of the mode [Atkinson, 1996]
45 Optima design in non-inear modes u Probem: parameter vaue θ, used to choose experiments F, is unknown u Three genera techniques to address this probem, usefu for many possibe notions of gain 1) Sequentia experimenta design: iterate between choosing experiment x and updating parameter estimates θ 2) Bayesian experimenta design: put a prior distribution on parameter θ, choose a best data x 3) Maximin experimenta design: assume worst case scenario for parameter θ, choose a best data x
46 Response Surface Methods u Estimate effects of oca changes to the interventions (queries) In particuar, estimate how to maximize the response u Appications: Find optima conditions for growing ce cutures Deveop robust process for chemica manufacturing u Procedure for maximizing response Given a set of datapoints, interpoate a oca surface (This oca surface is caed the response surface ) n Typicay use a quadratic poynomia to obtain a Hessian Hi-cimb or take Newton step on the response surface to find next x Use next x to interpoate subsequent response surface
47 Response Surface Modeing Goa: Approximate the function f(c) = score(minimize(c)) 1. Fit a smoothed response surface to the data points 2. Minimize response surface to find new candidate 3. Use method to find nearby oca minimum of score function 4. Add candidate to data points 5. Re-fit surface, repeat [Bum, unpubished]
48 Summary Active earning (sequentia) Query by committee Uncertainty samping Information-based oss functions Optima experimenta design A-optima design D-optima design E-optima design Non-inear optima experimenta design Sequentia experimenta design Bayesian experimenta design Response surface methods Mutipe modes Predictive distribution on points Maximize info gain Minimize trace of information matrix Minimize og det of information matrix Minimize argest eigenvaue of information matrix Mutipe-shot experiments; Litte known of parameters Singe-shot experiment; Some idea of parameter distribution Sequentia experiments for optimization
Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?
Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine
More informationFRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)
1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using
More informationAST 418/518 Instrumentation and Statistics
AST 418/518 Instrumentation and Statistics Cass Website: http://ircamera.as.arizona.edu/astr_518 Cass Texts: Practica Statistics for Astronomers, J.V. Wa, and C.R. Jenkins, Second Edition. Measuring the
More informationOn the Goal Value of a Boolean Function
On the Goa Vaue of a Booean Function Eric Bach Dept. of CS University of Wisconsin 1210 W. Dayton St. Madison, WI 53706 Lisa Heerstein Dept of CSE NYU Schoo of Engineering 2 Metrotech Center, 10th Foor
More informationExplicit overall risk minimization transductive bound
1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,
More informationCS229 Lecture notes. Andrew Ng
CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view
More informationStochastic Variational Inference with Gradient Linearization
Stochastic Variationa Inference with Gradient Linearization Suppementa Materia Tobias Pötz * Anne S Wannenwetsch Stefan Roth Department of Computer Science, TU Darmstadt Preface In this suppementa materia,
More informationExpectation-Maximization for Estimating Parameters for a Mixture of Poissons
Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Brandon Maone Department of Computer Science University of Hesini February 18, 2014 Abstract This document derives, in excrutiating
More informationMONTE CARLO SIMULATIONS
MONTE CARLO SIMULATIONS Current physics research 1) Theoretica 2) Experimenta 3) Computationa Monte Caro (MC) Method (1953) used to study 1) Discrete spin systems 2) Fuids 3) Poymers, membranes, soft matter
More informationSVM: Terminology 1(6) SVM: Terminology 2(6)
Andrew Kusiak Inteigent Systems Laboratory 39 Seamans Center he University of Iowa Iowa City, IA 54-57 SVM he maxima margin cassifier is simiar to the perceptron: It aso assumes that the data points are
More informationA. Distribution of the test statistic
A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch
More informationA Brief Introduction to Markov Chains and Hidden Markov Models
A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,
More informationBourgain s Theorem. Computational and Metric Geometry. Instructor: Yury Makarychev. d(s 1, s 2 ).
Bourgain s Theorem Computationa and Metric Geometry Instructor: Yury Makarychev 1 Notation Given a metric space (X, d) and S X, the distance from x X to S equas d(x, S) = inf d(x, s). s S The distance
More informationASummaryofGaussianProcesses Coryn A.L. Bailer-Jones
ASummaryofGaussianProcesses Coryn A.L. Baier-Jones Cavendish Laboratory University of Cambridge caj@mrao.cam.ac.uk Introduction A genera prediction probem can be posed as foows. We consider that the variabe
More informationStatistical Learning Theory: A Primer
Internationa Journa of Computer Vision 38(), 9 3, 2000 c 2000 uwer Academic Pubishers. Manufactured in The Netherands. Statistica Learning Theory: A Primer THEODOROS EVGENIOU, MASSIMILIANO PONTIL AND TOMASO
More informationInductive Bias: How to generalize on novel data. CS Inductive Bias 1
Inductive Bias: How to generaize on nove data CS 478 - Inductive Bias 1 Overfitting Noise vs. Exceptions CS 478 - Inductive Bias 2 Non-Linear Tasks Linear Regression wi not generaize we to the task beow
More informationC. Fourier Sine Series Overview
12 PHILIP D. LOEWEN C. Fourier Sine Series Overview Let some constant > be given. The symboic form of the FSS Eigenvaue probem combines an ordinary differentia equation (ODE) on the interva (, ) with a
More informationBayesian Unscented Kalman Filter for State Estimation of Nonlinear and Non-Gaussian Systems
Bayesian Unscented Kaman Fiter for State Estimation of Noninear and Non-aussian Systems Zhong Liu, Shing-Chow Chan, Ho-Chun Wu and iafei Wu Department of Eectrica and Eectronic Engineering, he University
More information6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17. Solution 7
6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17 Soution 7 Probem 1: Generating Random Variabes Each part of this probem requires impementation in MATLAB. For the
More informationGeneral Certificate of Education Advanced Level Examination June 2010
Genera Certificate of Education Advanced Leve Examination June 2010 Human Bioogy HBI6T/P10/task Unit 6T A2 Investigative Skis Assignment Task Sheet The effect of temperature on the rate of photosynthesis
More informationMARKOV CHAINS AND MARKOV DECISION THEORY. Contents
MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After
More informationFFTs in Graphics and Vision. Spherical Convolution and Axial Symmetry Detection
FFTs in Graphics and Vision Spherica Convoution and Axia Symmetry Detection Outine Math Review Symmetry Genera Convoution Spherica Convoution Axia Symmetry Detection Math Review Symmetry: Given a unitary
More informationSeparation of Variables and a Spherical Shell with Surface Charge
Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation
More informationStatistics for Applications. Chapter 7: Regression 1/43
Statistics for Appications Chapter 7: Regression 1/43 Heuristics of the inear regression (1) Consider a coud of i.i.d. random points (X i,y i ),i =1,...,n : 2/43 Heuristics of the inear regression (2)
More informationSome Measures for Asymmetry of Distributions
Some Measures for Asymmetry of Distributions Georgi N. Boshnakov First version: 31 January 2006 Research Report No. 5, 2006, Probabiity and Statistics Group Schoo of Mathematics, The University of Manchester
More informationCryptanalysis of PKP: A New Approach
Cryptanaysis of PKP: A New Approach Éiane Jaumes and Antoine Joux DCSSI 18, rue du Dr. Zamenhoff F-92131 Issy-es-Mx Cedex France eiane.jaumes@wanadoo.fr Antoine.Joux@ens.fr Abstract. Quite recenty, in
More informationFrom Margins to Probabilities in Multiclass Learning Problems
From Margins to Probabiities in Muticass Learning Probems Andrea Passerini and Massimiiano Ponti 2 and Paoo Frasconi 3 Abstract. We study the probem of muticass cassification within the framework of error
More informationLecture Note 3: Stationary Iterative Methods
MATH 5330: Computationa Methods of Linear Agebra Lecture Note 3: Stationary Iterative Methods Xianyi Zeng Department of Mathematica Sciences, UTEP Stationary Iterative Methods The Gaussian eimination (or
More informationOptimality of Inference in Hierarchical Coding for Distributed Object-Based Representations
Optimaity of Inference in Hierarchica Coding for Distributed Object-Based Representations Simon Brodeur, Jean Rouat NECOTIS, Département génie éectrique et génie informatique, Université de Sherbrooke,
More informationTarget Location Estimation in Wireless Sensor Networks Using Binary Data
Target Location stimation in Wireess Sensor Networks Using Binary Data Ruixin Niu and Pramod K. Varshney Department of ectrica ngineering and Computer Science Link Ha Syracuse University Syracuse, NY 344
More informationMATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES
MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES Separation of variabes is a method to sove certain PDEs which have a warped product structure. First, on R n, a inear PDE of order m is
More informationSmoothness equivalence properties of univariate subdivision schemes and their projection analogues
Numerische Mathematik manuscript No. (wi be inserted by the editor) Smoothness equivaence properties of univariate subdivision schemes and their projection anaogues Phiipp Grohs TU Graz Institute of Geometry
More informationProblem set 6 The Perron Frobenius theorem.
Probem set 6 The Perron Frobenius theorem. Math 22a4 Oct 2 204, Due Oct.28 In a future probem set I want to discuss some criteria which aow us to concude that that the ground state of a sef-adjoint operator
More informationStatistical Astronomy
Lectures for the 7 th IAU ISYA Ifrane, nd 3 rd Juy 4 p ( x y, I) p( y x, I) p( x, I) p( y, I) Statistica Astronomy Martin Hendry, Dept of Physics and Astronomy University of Gasgow, UK http://www.astro.ga.ac.uk/users/martin/isya/
More informationBP neural network-based sports performance prediction model applied research
Avaiabe onine www.jocpr.com Journa of Chemica and Pharmaceutica Research, 204, 6(7:93-936 Research Artice ISSN : 0975-7384 CODEN(USA : JCPRC5 BP neura networ-based sports performance prediction mode appied
More informationThe EM Algorithm applied to determining new limit points of Mahler measures
Contro and Cybernetics vo. 39 (2010) No. 4 The EM Agorithm appied to determining new imit points of Maher measures by Souad E Otmani, Georges Rhin and Jean-Marc Sac-Épée Université Pau Veraine-Metz, LMAM,
More informationHigh Spectral Resolution Infrared Radiance Modeling Using Optimal Spectral Sampling (OSS) Method
High Spectra Resoution Infrared Radiance Modeing Using Optima Spectra Samping (OSS) Method J.-L. Moncet and G. Uymin Background Optima Spectra Samping (OSS) method is a fast and accurate monochromatic
More informationGeneral Certificate of Education Advanced Level Examination June 2010
Genera Certificate of Education Advanced Leve Examination June 2010 Human Bioogy HBI6T/Q10/task Unit 6T A2 Investigative Skis Assignment Task Sheet The effect of using one or two eyes on the perception
More informationSupport Vector Machine and Its Application to Regression and Classification
BearWorks Institutiona Repository MSU Graduate Theses Spring 2017 Support Vector Machine and Its Appication to Regression and Cassification Xiaotong Hu As with any inteectua project, the content and views
More informationTHE THREE POINT STEINER PROBLEM ON THE FLAT TORUS: THE MINIMAL LUNE CASE
THE THREE POINT STEINER PROBLEM ON THE FLAT TORUS: THE MINIMAL LUNE CASE KATIE L. MAY AND MELISSA A. MITCHELL Abstract. We show how to identify the minima path network connecting three fixed points on
More informationAsynchronous Control for Coupled Markov Decision Systems
INFORMATION THEORY WORKSHOP (ITW) 22 Asynchronous Contro for Couped Marov Decision Systems Michae J. Neey University of Southern Caifornia Abstract This paper considers optima contro for a coection of
More informationPartial permutation decoding for MacDonald codes
Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics
More informationFRIEZE GROUPS IN R 2
FRIEZE GROUPS IN R 2 MAXWELL STOLARSKI Abstract. Focusing on the Eucidean pane under the Pythagorean Metric, our goa is to cassify the frieze groups, discrete subgroups of the set of isometries of the
More informationVI.G Exact free energy of the Square Lattice Ising model
VI.G Exact free energy of the Square Lattice Ising mode As indicated in eq.(vi.35), the Ising partition function is reated to a sum S, over coections of paths on the attice. The aowed graphs for a square
More informationAsymptotic Properties of a Generalized Cross Entropy Optimization Algorithm
1 Asymptotic Properties of a Generaized Cross Entropy Optimization Agorithm Zijun Wu, Michae Koonko, Institute for Appied Stochastics and Operations Research, Caustha Technica University Abstract The discrete
More informationXSAT of linear CNF formulas
XSAT of inear CN formuas Bernd R. Schuh Dr. Bernd Schuh, D-50968 Kön, Germany; bernd.schuh@netcoogne.de eywords: compexity, XSAT, exact inear formua, -reguarity, -uniformity, NPcompeteness Abstract. Open
More informationStatistical Learning Theory: a Primer
??,??, 1 6 (??) c?? Kuwer Academic Pubishers, Boston. Manufactured in The Netherands. Statistica Learning Theory: a Primer THEODOROS EVGENIOU AND MASSIMILIANO PONTIL Center for Bioogica and Computationa
More informationPrimal and dual active-set methods for convex quadratic programming
Math. Program., Ser. A 216) 159:469 58 DOI 1.17/s117-15-966-2 FULL LENGTH PAPER Prima and dua active-set methods for convex quadratic programming Anders Forsgren 1 Phiip E. Gi 2 Eizabeth Wong 2 Received:
More informationAppendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model
Appendix of the Paper The Roe of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Mode Caio Ameida cameida@fgv.br José Vicente jose.vaentim@bcb.gov.br June 008 1 Introduction In this
More informationSVM-based Supervised and Unsupervised Classification Schemes
SVM-based Supervised and Unsupervised Cassification Schemes LUMINITA STATE University of Pitesti Facuty of Mathematics and Computer Science 1 Targu din Vae St., Pitesti 110040 ROMANIA state@cicknet.ro
More informationConvolutional Networks 2: Training, deep convolutional networks
Convoutiona Networks 2: Training, deep convoutiona networks Hakan Bien Machine Learning Practica MLP Lecture 8 30 October / 6 November 2018 MLP Lecture 8 / 30 October / 6 November 2018 Convoutiona Networks
More informationUnconditional security of differential phase shift quantum key distribution
Unconditiona security of differentia phase shift quantum key distribution Kai Wen, Yoshihisa Yamamoto Ginzton Lab and Dept of Eectrica Engineering Stanford University Basic idea of DPS-QKD Protoco. Aice
More informationBALANCING REGULAR MATRIX PENCILS
BALANCING REGULAR MATRIX PENCILS DAMIEN LEMONNIER AND PAUL VAN DOOREN Abstract. In this paper we present a new diagona baancing technique for reguar matrix pencis λb A, which aims at reducing the sensitivity
More informationResearch of Data Fusion Method of Multi-Sensor Based on Correlation Coefficient of Confidence Distance
Send Orders for Reprints to reprints@benthamscience.ae 340 The Open Cybernetics & Systemics Journa, 015, 9, 340-344 Open Access Research of Data Fusion Method of Muti-Sensor Based on Correation Coefficient
More informationII. PROBLEM. A. Description. For the space of audio signals
CS229 - Fina Report Speech Recording based Language Recognition (Natura Language) Leopod Cambier - cambier; Matan Leibovich - matane; Cindy Orozco Bohorquez - orozcocc ABSTRACT We construct a rea time
More informationConvergence Property of the Iri-Imai Algorithm for Some Smooth Convex Programming Problems
Convergence Property of the Iri-Imai Agorithm for Some Smooth Convex Programming Probems S. Zhang Communicated by Z.Q. Luo Assistant Professor, Department of Econometrics, University of Groningen, Groningen,
More informationLecture 4: Types of errors. Bayesian regression models. Logistic regression
Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture
More informationParagraph Topic Classification
Paragraph Topic Cassification Eugene Nho Graduate Schoo of Business Stanford University Stanford, CA 94305 enho@stanford.edu Edward Ng Department of Eectrica Engineering Stanford University Stanford, CA
More information8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING
8 APPENDIX 8.1 EMPIRICAL EVALUATION OF SAMPLING We wish to evauate the empirica accuracy of our samping technique on concrete exampes. We do this in two ways. First, we can sort the eements by probabiity
More informationFast Spectral Clustering via the Nyström Method
Fast Spectra Custering via the Nyström Method Anna Choromanska, Tony Jebara, Hyungtae Kim, Mahesh Mohan 3, and Caire Monteeoni 3 Department of Eectrica Engineering, Coumbia University, NY, USA Department
More informationEfficiently Generating Random Bits from Finite State Markov Chains
1 Efficienty Generating Random Bits from Finite State Markov Chains Hongchao Zhou and Jehoshua Bruck, Feow, IEEE Abstract The probem of random number generation from an uncorreated random source (of unknown
More informationLecture 9. Stability of Elastic Structures. Lecture 10. Advanced Topic in Column Buckling
Lecture 9 Stabiity of Eastic Structures Lecture 1 Advanced Topic in Coumn Bucking robem 9-1: A camped-free coumn is oaded at its tip by a oad. The issue here is to find the itica bucking oad. a) Suggest
More informationDetermining The Degree of Generalization Using An Incremental Learning Algorithm
Determining The Degree of Generaization Using An Incrementa Learning Agorithm Pabo Zegers Facutad de Ingeniería, Universidad de os Andes San Caros de Apoquindo 22, Las Condes, Santiago, Chie pzegers@uandes.c
More informationLearning Gaussian Processes from Multiple Tasks
Kai Yu kai.yu@siemens.com Information and Communication, Corporate Technoogy, Siemens AG, Munich, Germany Voker Tresp voker.tresp@siemens.com Information and Communication, Corporate Technoogy, Siemens
More informationSUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS
ISEE 1 SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS By Yingying Fan and Jinchi Lv University of Southern Caifornia This Suppementary Materia
More informationEmmanuel Abbe Colin Sandon
Detection in the stochastic bock mode with mutipe custers: proof of the achievabiity conjectures, acycic BP, and the information-computation gap Emmanue Abbe Coin Sandon Abstract In a paper that initiated
More informationUniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete
Uniprocessor Feasibiity of Sporadic Tasks with Constrained Deadines is Strongy conp-compete Pontus Ekberg and Wang Yi Uppsaa University, Sweden Emai: {pontus.ekberg yi}@it.uu.se Abstract Deciding the feasibiity
More informationProbabilistic Graphical Models
Schoo of Computer Science Probabiistic Graphica Modes Gaussian graphica modes and Ising modes: modeing networks Eric Xing Lecture 0, February 0, 07 Reading: See cass website Eric Xing @ CMU, 005-07 Network
More informationCompletion. is dense in H. If V is complete, then U(V) = H.
Competion Theorem 1 (Competion) If ( V V ) is any inner product space then there exists a Hibert space ( H H ) and a map U : V H such that (i) U is 1 1 (ii) U is inear (iii) UxUy H xy V for a xy V (iv)
More informationData Mining Technology for Failure Prognostic of Avionics
IEEE Transactions on Aerospace and Eectronic Systems. Voume 38, #, pp.388-403, 00. Data Mining Technoogy for Faiure Prognostic of Avionics V.A. Skormin, Binghamton University, Binghamton, NY, 1390, USA
More informationTight Approximation Algorithms for Maximum Separable Assignment Problems
MATHEMATICS OF OPERATIONS RESEARCH Vo. 36, No. 3, August 011, pp. 416 431 issn 0364-765X eissn 156-5471 11 3603 0416 10.187/moor.1110.0499 011 INFORMS Tight Approximation Agorithms for Maximum Separabe
More informationSequential Decoding of Polar Codes with Arbitrary Binary Kernel
Sequentia Decoding of Poar Codes with Arbitrary Binary Kerne Vera Miosavskaya, Peter Trifonov Saint-Petersburg State Poytechnic University Emai: veram,petert}@dcn.icc.spbstu.ru Abstract The probem of efficient
More informationPREDICTION OF DEFORMED AND ANNEALED MICROSTRUCTURES USING BAYESIAN NEURAL NETWORKS AND GAUSSIAN PROCESSES
PREDICTION OF DEFORMED AND ANNEALED MICROSTRUCTURES USING BAYESIAN NEURAL NETWORKS AND GAUSSIAN PROCESSES C.A.L. Baier-Jones, T.J. Sabin, D.J.C. MacKay, P.J. Withers Department of Materias Science and
More informationAnalysis of Emerson s Multiple Model Interpolation Estimation Algorithms: The MIMO Case
Technica Report PC-04-00 Anaysis of Emerson s Mutipe Mode Interpoation Estimation Agorithms: The MIMO Case João P. Hespanha Dae E. Seborg University of Caifornia, Santa Barbara February 0, 004 Anaysis
More informationhttps://doi.org/ /epjconf/
HOW TO APPLY THE OPTIMAL ESTIMATION METHOD TO YOUR LIDAR MEASUREMENTS FOR IMPROVED RETRIEVALS OF TEMPERATURE AND COMPOSITION R. J. Sica 1,2,*, A. Haefee 2,1, A. Jaai 1, S. Gamage 1 and G. Farhani 1 1 Department
More informationConsistent linguistic fuzzy preference relation with multi-granular uncertain linguistic information for solving decision making problems
Consistent inguistic fuzzy preference reation with muti-granuar uncertain inguistic information for soving decision making probems Siti mnah Binti Mohd Ridzuan, and Daud Mohamad Citation: IP Conference
More informationTwo view learning: SVM-2K, Theory and Practice
Two view earning: SVM-2K, Theory and Practice Jason D.R. Farquhar jdrf99r@ecs.soton.ac.uk Hongying Meng hongying@cs.york.ac.uk David R. Hardoon drh@ecs.soton.ac.uk John Shawe-Tayor jst@ecs.soton.ac.uk
More informationData Discovery and Anomaly Detection Using Atypicality: Theory
Data Discovery and Anomay Detection Using Atypicaity: Theory Anders Høst-Madsen, Feow, IEEE, Eyas Sabeti, Member, IEEE, Chad Waton Abstract A centra question in the era of big data is what to do with the
More informationBDD-Based Analysis of Gapped q-gram Filters
BDD-Based Anaysis of Gapped q-gram Fiters Marc Fontaine, Stefan Burkhardt 2 and Juha Kärkkäinen 2 Max-Panck-Institut für Informatik Stuhsatzenhausweg 85, 6623 Saarbrücken, Germany e-mai: stburk@mpi-sb.mpg.de
More informationarxiv: v1 [cs.db] 1 Aug 2012
Functiona Mechanism: Regression Anaysis under Differentia Privacy arxiv:208.029v [cs.db] Aug 202 Jun Zhang Zhenjie Zhang 2 Xiaokui Xiao Yin Yang 2 Marianne Winsett 2,3 ABSTRACT Schoo of Computer Engineering
More informationSTA 216 Project: Spline Approach to Discrete Survival Analysis
: Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing
More informationChapter 1 Decomposition methods for Support Vector Machines
Chapter 1 Decomposition methods for Support Vector Machines Support Vector Machines (SVM) are widey used as a simpe and efficient too for inear and noninear cassification as we as for regression probems.
More informationFINAL: CS 6375 (Machine Learning) Fall 2014
FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for
More informationChemical Kinetics Part 2
Integrated Rate Laws Chemica Kinetics Part 2 The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates the rate
More informationDIGITAL FILTER DESIGN OF IIR FILTERS USING REAL VALUED GENETIC ALGORITHM
DIGITAL FILTER DESIGN OF IIR FILTERS USING REAL VALUED GENETIC ALGORITHM MIKAEL NILSSON, MATTIAS DAHL AND INGVAR CLAESSON Bekinge Institute of Technoogy Department of Teecommunications and Signa Processing
More informationLecture 1: Introduction, Entropy and ML estimation
0-704: Information Processing and Learning Spring 202 Lecture : Introduction, Entropy and ML estimation Lecturer: Aarti Singh Scribes: Min Xu Disclaimer: These notes have not been subjected to the usual
More informationStatistical learning. Chapter 20, Sections 1 4 1
Statistical learning Chapter 20, Sections 1 4 Chapter 20, Sections 1 4 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete
More information$, (2.1) n="# #. (2.2)
Chapter. Eectrostatic II Notes: Most of the materia presented in this chapter is taken from Jackson, Chap.,, and 4, and Di Bartoo, Chap... Mathematica Considerations.. The Fourier series and the Fourier
More informationEfficient Generation of Random Bits from Finite State Markov Chains
Efficient Generation of Random Bits from Finite State Markov Chains Hongchao Zhou and Jehoshua Bruck, Feow, IEEE Abstract The probem of random number generation from an uncorreated random source (of unknown
More informationLearning Fully Observed Undirected Graphical Models
Learning Fuy Observed Undirected Graphica Modes Sides Credit: Matt Gormey (2016) Kayhan Batmangheich 1 Machine Learning The data inspires the structures we want to predict Inference finds {best structure,
More information( ) is just a function of x, with
II. MULTIVARIATE CALCULUS The first ecture covered functions where a singe input goes in, and a singe output comes out. Most economic appications aren t so simpe. In most cases, a number of variabes infuence
More informationCONGRUENCES. 1. History
CONGRUENCES HAO BILLY LEE Abstract. These are notes I created for a seminar tak, foowing the papers of On the -adic Representations and Congruences for Coefficients of Moduar Forms by Swinnerton-Dyer and
More informationTrainable fusion rules. I. Large sample size case
Neura Networks 19 (2006) 1506 1516 www.esevier.com/ocate/neunet Trainabe fusion rues. I. Large sampe size case Šarūnas Raudys Institute of Mathematics and Informatics, Akademijos 4, Vinius 08633, Lithuania
More informationKernel pea and De-Noising in Feature Spaces
Kerne pea and De-Noising in Feature Spaces Sebastian Mika, Bernhard Schokopf, Aex Smoa Kaus-Robert Muer, Matthias Schoz, Gunnar Riitsch GMD FIRST, Rudower Chaussee 5, 12489 Berin, Germany {mika, bs, smoa,
More informationThe Group Structure on a Smooth Tropical Cubic
The Group Structure on a Smooth Tropica Cubic Ethan Lake Apri 20, 2015 Abstract Just as in in cassica agebraic geometry, it is possibe to define a group aw on a smooth tropica cubic curve. In this note,
More informationDiscrete Techniques. Chapter Introduction
Chapter 3 Discrete Techniques 3. Introduction In the previous two chapters we introduced Fourier transforms of continuous functions of the periodic and non-periodic (finite energy) type, as we as various
More informationc 2016 Georgios Rovatsos
c 2016 Georgios Rovatsos QUICKEST CHANGE DETECTION WITH APPLICATIONS TO LINE OUTAGE DETECTION BY GEORGIOS ROVATSOS THESIS Submitted in partia fufiment of the requirements for the degree of Master of Science
More informationA Sparse Covariance Function for Exact Gaussian Process Inference in Large Datasets
A Covariance Function for Exact Gaussian Process Inference in Large Datasets Arman ekumyan Austraian Centre for Fied Robotics The University of Sydney NSW 26, Austraia a.mekumyan@acfr.usyd.edu.au Fabio
More informationFast Blind Recognition of Channel Codes
Fast Bind Recognition of Channe Codes Reza Moosavi and Erik G. Larsson Linköping University Post Print N.B.: When citing this work, cite the origina artice. 213 IEEE. Persona use of this materia is permitted.
More information(Refer Slide Time: 2:34) L C V
Microwave Integrated Circuits Professor Jayanta Mukherjee Department of Eectrica Engineering Indian Intitute of Technoogy Bombay Modue 1 Lecture No 2 Refection Coefficient, SWR, Smith Chart. Heo wecome
More information