VC Dimension and Sauer s Lemma
|
|
- Shonda Beasley
- 5 years ago
- Views:
Transcription
1 CMSC (Spring 2008) Learning Theory Lecture: VC Diension and Sauer s Lea Instructors: Sha Kakade and Abuj Tewari Radeacher Averages and Growth Function Theore Let F be a class of ±-valued functions Then we have, Proof We have, 2 ln ΠF () R (F) [ [ R (F) = E E sup ɛ i a i X a F X i= E 2 ln F X [ ] 2 ln ΠF () E ]] = 2 ln ΠF () Since f(x i ) {±}, any a F X has a = The first inequality above therefore follows fro Massart s finite class lea The second inequality follows fro the definition of the growth function Π F () Note that plugging in the trivial bound Π F () 2 does not give us any interesting bound This is quite reasonable since this bound would hold for any function class no atter how coplicated it is To easure the coplexity of F, let us look at the first natural nuber such that Π F () falls below 2 This brings us to the definition of the Vapnik-Chervonenkis diension 2 Vapnik-Chervonenkis Diension The Vapnik-Chervonenkis diension (or siply the VC-diension) of a function class F {±} X is defined as VCdi(F) := ax { > 0 Π F () = 2 } An equivalent definition is that VCdi(F) is the size of the largest set shattered by F A set {x,, x } is said to be shattered by F if for any labelling b = (b,, b ) {±}, there is a function f F such that (f(x ),, f(x )) = (b,, b ) Note that a function f {±} X can de identified with the subset of X on which it is equal to + So, we often talk about the VC-diension of a collection of subsets of X The table below gives the VC-diensions for a few exaples The proofs of the clais in the first three rows of the table are left as exercises Here we prove only the last clai: the VC-diension of halfspaces in R d is d +
2 X F VCdi(F) R 2 convex polygons R 2 axis-aligned rectangles 4 R 2 convex polygons with d vertices 2d + R d halfspaces d + Theore 2 Let X = R d Define the set of ±-valued functions associated with halfspaces, Then, VCdi(F) = d + Proof We have to prove two inequalities F = { x sgn (w x θ) w R d, θ R } VCdi(F) d +, () VCdi(F) d + (2) To prove the first inequality, we need to exhibit a particular set of size d + that is shattered by F Proving the second inequality is a bit ore tricky: we need to show that for all sets of size d + 2, there is labelling that cannot be realized using halfpsaces Let us first prove () Consider the set X = {0, e,, e d } which consists of the origin along with the vectors in the standard basis of R d Given a labelling b 0,, b d of these points, set θ = b 0, w i = θ + b i, i [d] With these definitions, it iediately follows that w 0 θ = b 0 and for all i [d], w e i θ = b i Thus, X is shattered by F Since, X = d +, we have proved () Before we prove (2), we need the following result fro convex geoetry Radon s Lea Let X R d be a set of size d + 2 Then there exist two disjoint subsets X, X 2 of X such that conv(x ) conv(x 2 ) Here conv(xenotes the convex hull of X Proof Let X = {x,, x d+2 } Consider the following syste of d + equations in the variables λ,, λ d+2, λ λ 2 ( x x 2 x d+2 ) λ d+2 = 0 (3) Since, there are ore variables than equations, there is a non-trivial solution λ 0 Define the set of indices, Since λ 0, both P and N and non-epty and P = {i λ i > 0}, N = { j λ j < 0 } λ i = j N Moreover, since λ satisfies d+2 i= λ i x i = 0, we have λ i x i = j N 2 ( λ j ) 0 ( λ j )x j
3 Defining X = {x i X i P } and X 2 = {x i X i N}, we see that the point λ i x i j N = ( λ j )x j λ i j N ( λ j lies both in conv(x ) as well as conv(x 2 ) Given Radon s lea, the proof of () is quite easy We have to show that given a set X R d of size d + 2, there is a labelling that cannot be realized using halfspaces Obtain disjoint subsets X, X 2 of X whose existence is guaranteed by Radon s lea Now consider a labelling in which all the points in X are labelled + and those in X 2 are labelled We clai that such a labelling cannot be realized using a halfspace Suppose there is such a halfspace H Note that if a halfspace assigns a particular label to a set of points, then every point in their convex hull is also assigned the sae label Thus every point in conv(x ) is labelled + by H while every point in conv(x 2 ) is labelled But conv(x ) conv(x 2 ) giving us a contradiction We often work with ±-valued functions obtained by thresholding real valued functions at 0 If these real valued functions coe fro a finite diensional vector space, the next result gives an upper bound on the VC diension Theore 22 Let G be a finite diensional vector space of functions on R d Define, If the diension of G is k then VCdi(F) k F = {x sgn(g(x)) g G} Proof Fix an arbitrary set of k + points x,, x k+ We show that this set cannot be shattered by F Consider the linear transforation T : G R k+ defined as T (g) = (g(x ),, g(x k+ ) The diension of the iage of G under T is at ost k Thus, there exists a non-zero vector λ R k+ that is orthogonal to it That is, for all g G, k+ λ i g(x i ) = 0 (4) At least one of the sets, i= P := {i λ i > 0}, N := {j λ j < 0}, is non-epty Without loss of generality assue it is P Consider a labelling of x,, x k+ that assigns the label + to all x i such that i P and to the rest If this labelling is realized by a function in F then there exists g 0 G such that λ i g 0 (x i ) > 0, λ i g 0 (x i ) 0 But this contradicts (4) Therefore x,, x k+ cannot be shattered by F i N 3
4 3 Growth Function and VC Diension Suppose VCdi(F) = d Then for all d, Π F () = 2 The lea below, due to Sauer, iplies that for > d, Π F () = O( d ), a polynoial rate of growth This result is rearkable for it iplies that the growth function exhibits just two kinds of behavior If VCdi(F) = then Π F grows exponentially with On the other hand, if VCdi(F) = d < then the growth function is O( d ) Sauer s Lea Let F be such that VCdi(F) d Then, we have Π F () d ( ) i Proof We prove this by induction on + d For = d =, the above inequality holds as both sides are equal to 2 Assue that it holds for and d and for and d We will prove it for and d Define the function, h(, d) := d ( ) i so that our induction hypothesis is: for F with VCdi(F) d, Π F () h(, d) Since ( ) ( ) ( ) = +, i i i is is easy to verify that h satisfies the recurrence h(, d) = h(, d) + h(, d ) Fix a class F with VCdi(F) = d and a set X = {x,, x } X Let F = F X and X 2 = {x 2,, x } and define the function classes, F := F X F 2 := F X2 F 3 := {f X2 f F & f F st x X 2, f (x) = f(x) & f (x ) = f(x )} Note that VCdi(F ) VCdi(F) d and we wish to bound F By the definitions above, we have F = F 2 + F 3 It is easy to see that VCdi(F 2 ) d Also, VCdi(F 3 ) d because if F 3 shatters a set, we can always add x to it to get a set that is shattered by F By induction hypothesis, F 2 h(, d) and F 3 h(, d ) Thus, we have F x = F h(, d) + h(, d ) = h(, d) Since x,, x were arbitrary, we have and the induction step is coplete Π F () = sup F x x X h(, d) Corollary 3 Let F be such that VCdi(F) d Then, we have, for d, Π F () ( e d 4
5 Proof Since n d, we have d ( ) ( d ( ) ( d i d i ( ( ) ( d d i ( ) ( d + d ) d ( e d d ) i ) i 5
1 Generalization bounds based on Rademacher complexity
COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #0 Scribe: Suqi Liu March 07, 08 Last tie we started proving this very general result about how quickly the epirical average converges
More informationE0 370 Statistical Learning Theory Lecture 5 (Aug 25, 2011)
E0 370 Statistical Learning Theory Lecture 5 Aug 5, 0 Covering Nubers, Pseudo-Diension, and Fat-Shattering Diension Lecturer: Shivani Agarwal Scribe: Shivani Agarwal Introduction So far we have seen how
More information1 Rademacher Complexity Bounds
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #10 Scribe: Max Goer March 07, 2013 1 Radeacher Coplexity Bounds Recall the following theore fro last lecture: Theore 1. With probability
More informationComputational and Statistical Learning Theory
Coputational and Statistical Learning Theory TTIC 31120 Prof. Nati Srebro Lecture 2: PAC Learning and VC Theory I Fro Adversarial Online to Statistical Three reasons to ove fro worst-case deterinistic
More informationSymmetrization and Rademacher Averages
Stat 928: Statistical Learning Theory Lecture: Syetrization and Radeacher Averages Instructor: Sha Kakade Radeacher Averages Recall that we are interested in bounding the difference between epirical and
More information1 Bounding the Margin
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost
More informationComputational and Statistical Learning Theory
Coputational and Statistical Learning Theory Proble sets 5 and 6 Due: Noveber th Please send your solutions to learning-subissions@ttic.edu Notations/Definitions Recall the definition of saple based Radeacher
More informationIntroduction to Machine Learning
Introduction to Machine Learning Vapnik Chervonenkis Theory Barnabás Póczos Empirical Risk and True Risk 2 Empirical Risk Shorthand: True risk of f (deterministic): Bayes risk: Let us use the empirical
More informationE0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis
E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds
More informationList Scheduling and LPT Oliver Braun (09/05/2017)
List Scheduling and LPT Oliver Braun (09/05/207) We investigate the classical scheduling proble P ax where a set of n independent jobs has to be processed on 2 parallel and identical processors (achines)
More informationarxiv: v1 [math.co] 19 Apr 2017
PROOF OF CHAPOTON S CONJECTURE ON NEWTON POLYTOPES OF q-ehrhart POLYNOMIALS arxiv:1704.0561v1 [ath.co] 19 Apr 017 JANG SOO KIM AND U-KEUN SONG Abstract. Recently, Chapoton found a q-analog of Ehrhart polynoials,
More information1 Proof of learning bounds
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a
More informationPREPRINT 2006:17. Inequalities of the Brunn-Minkowski Type for Gaussian Measures CHRISTER BORELL
PREPRINT 2006:7 Inequalities of the Brunn-Minkowski Type for Gaussian Measures CHRISTER BORELL Departent of Matheatical Sciences Division of Matheatics CHALMERS UNIVERSITY OF TECHNOLOGY GÖTEBORG UNIVERSITY
More informationUnderstanding Machine Learning Solution Manual
Understanding Machine Learning Solution Manual Written by Alon Gonen Edited by Dana Rubinstein Noveber 17, 2014 2 Gentle Start 1. Given S = ((x i, y i )), define the ultivariate polynoial p S (x) = i []:y
More informationComputational and Statistical Learning theory
Computational and Statistical Learning theory Problem set 2 Due: January 31st Email solutions to : karthik at ttic dot edu Notation : Input space : X Label space : Y = {±1} Sample : (x 1, y 1,..., (x n,
More informationIntroduction to Discrete Optimization
Prof. Friedrich Eisenbrand Martin Nieeier Due Date: March 9 9 Discussions: March 9 Introduction to Discrete Optiization Spring 9 s Exercise Consider a school district with I neighborhoods J schools and
More information13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices
CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay
More informationQuestion 1. Question 3. Question 4. Graduate Analysis I Exercise 4
Graduate Analysis I Exercise 4 Question 1 If f is easurable and λ is any real nuber, f + λ and λf are easurable. Proof. Since {f > a λ} is easurable, {f + λ > a} = {f > a λ} is easurable, then f + λ is
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU10701 11. Learning Theory Barnabás Póczos Learning Theory We have explored many ways of learning from data But How good is our classifier, really? How much data do we
More information1 Proving the Fundamental Theorem of Statistical Learning
THEORETICAL MACHINE LEARNING COS 5 LECTURE #7 APRIL 5, 6 LECTURER: ELAD HAZAN NAME: FERMI MA ANDDANIEL SUO oving te Fundaental Teore of Statistical Learning In tis section, we prove te following: Teore.
More informationConstruction of the Electronic Angular Wave Functions and Probability Distributions of the Hydrogen Atom
Construction of the Electronic Angular Wave Functions and Probability Distributions of the Hydrogen Ato Thoas S. Kuntzlean Mark Ellison John Tippin Departent of Cheistry Departent of Cheistry Departent
More informationQuantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search
Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths
More informationORIGAMI CONSTRUCTIONS OF RINGS OF INTEGERS OF IMAGINARY QUADRATIC FIELDS
#A34 INTEGERS 17 (017) ORIGAMI CONSTRUCTIONS OF RINGS OF INTEGERS OF IMAGINARY QUADRATIC FIELDS Jürgen Kritschgau Departent of Matheatics, Iowa State University, Aes, Iowa jkritsch@iastateedu Adriana Salerno
More information1 Identical Parallel Machines
FB3: Matheatik/Inforatik Dr. Syaantak Das Winter 2017/18 Optiizing under Uncertainty Lecture Notes 3: Scheduling to Miniize Makespan In any standard scheduling proble, we are given a set of jobs J = {j
More informationLecture 21. Interior Point Methods Setup and Algorithm
Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,
More informationPrerequisites. We recall: Theorem 2 A subset of a countably innite set is countable.
Prerequisites 1 Set Theory We recall the basic facts about countable and uncountable sets, union and intersection of sets and iages and preiages of functions. 1.1 Countable and uncountable sets We can
More informationThe Vapnik-Chervonenkis Dimension
The Vapnik-Chervonenkis Dimension Prof. Dan A. Simovici UMB 1 / 91 Outline 1 Growth Functions 2 Basic Definitions for Vapnik-Chervonenkis Dimension 3 The Sauer-Shelah Theorem 4 The Link between VCD and
More informationSome Classical Ergodic Theorems
Soe Classical Ergodic Theores Matt Rosenzweig Contents Classical Ergodic Theores. Mean Ergodic Theores........................................2 Maxial Ergodic Theore.....................................
More informationMACHINE LEARNING - CS671 - Part 2a The Vapnik-Chervonenkis Dimension
MACHINE LEARNING - CS671 - Part 2a The Vapnik-Chervonenkis Dimension Prof. Dan A. Simovici UMB Prof. Dan A. Simovici (UMB) MACHINE LEARNING - CS671 - Part 2a The Vapnik-Chervonenkis Dimension 1 / 30 The
More informationThe Weierstrass Approximation Theorem
36 The Weierstrass Approxiation Theore Recall that the fundaental idea underlying the construction of the real nubers is approxiation by the sipler rational nubers. Firstly, nubers are often deterined
More information3.8 Three Types of Convergence
3.8 Three Types of Convergence 3.8 Three Types of Convergence 93 Suppose that we are given a sequence functions {f k } k N on a set X and another function f on X. What does it ean for f k to converge to
More informationA Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness
A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,
More informationlecture 36: Linear Multistep Mehods: Zero Stability
95 lecture 36: Linear Multistep Mehods: Zero Stability 5.6 Linear ultistep ethods: zero stability Does consistency iply convergence for linear ultistep ethods? This is always the case for one-step ethods,
More informationFeature Extraction Techniques
Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that
More informationA1. Find all ordered pairs (a, b) of positive integers for which 1 a + 1 b = 3
A. Find all ordered pairs a, b) of positive integers for which a + b = 3 08. Answer. The six ordered pairs are 009, 08), 08, 009), 009 337, 674) = 35043, 674), 009 346, 673) = 3584, 673), 674, 009 337)
More informationVC Dimension Review. The purpose of this document is to review VC dimension and PAC learning for infinite hypothesis spaces.
VC Dimension Review The purpose of this document is to review VC dimension and PAC learning for infinite hypothesis spaces. Previously, in discussing PAC learning, we were trying to answer questions about
More informationFAST DYNAMO ON THE REAL LINE
FAST DYAMO O THE REAL LIE O. KOZLOVSKI & P. VYTOVA Abstract. In this paper we show that a piecewise expanding ap on the interval, extended to the real line by a non-expanding ap satisfying soe ild hypthesis
More informationChapter 6 1-D Continuous Groups
Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:
More informationA := A i : {A i } S. is an algebra. The same object is obtained when the union in required to be disjoint.
59 6. ABSTRACT MEASURE THEORY Having developed the Lebesgue integral with respect to the general easures, we now have a general concept with few specific exaples to actually test it on. Indeed, so far
More informationLecture 9 November 23, 2015
CSC244: Discrepancy Theory in Coputer Science Fall 25 Aleksandar Nikolov Lecture 9 Noveber 23, 25 Scribe: Nick Spooner Properties of γ 2 Recall that γ 2 (A) is defined for A R n as follows: γ 2 (A) = in{r(u)
More informationMetric Entropy of Convex Hulls
Metric Entropy of Convex Hulls Fuchang Gao University of Idaho Abstract Let T be a precopact subset of a Hilbert space. The etric entropy of the convex hull of T is estiated in ters of the etric entropy
More informationFoundations of Machine Learning Boosting. Mehryar Mohri Courant Institute and Google Research
Foundations of Machine Learning Boosting Mehryar Mohri Courant Institute and Google Research ohri@cis.nyu.edu Weak Learning Definition: concept class C is weakly PAC-learnable if there exists a (weak)
More informationLecture Learning infinite hypothesis class via VC-dimension and Rademacher complexity;
CSCI699: Topics in Learning and Game Theory Lecture 2 Lecturer: Ilias Diakonikolas Scribes: Li Han Today we will cover the following 2 topics: 1. Learning infinite hypothesis class via VC-dimension and
More informationTHE VAPNIK- CHERVONENKIS DIMENSION and LEARNABILITY
THE VAPNIK- CHERVONENKIS DIMENSION and LEARNABILITY Dan A. Simovici UMB, Doctoral Summer School Iasi, Romania What is Machine Learning? The Vapnik-Chervonenkis Dimension Probabilistic Learning Potential
More informationLearnability and Stability in the General Learning Setting
Learnability and Stability in the General Learning Setting Shai Shalev-Shwartz TTI-Chicago shai@tti-c.org Ohad Shair The Hebrew University ohadsh@cs.huji.ac.il Nathan Srebro TTI-Chicago nati@uchicago.edu
More informationA Quantum Observable for the Graph Isomorphism Problem
A Quantu Observable for the Graph Isoorphis Proble Mark Ettinger Los Alaos National Laboratory Peter Høyer BRICS Abstract Suppose we are given two graphs on n vertices. We define an observable in the Hilbert
More informationComputational Learning Theory
CS 446 Machine Learning Fall 2016 OCT 11, 2016 Computational Learning Theory Professor: Dan Roth Scribe: Ben Zhou, C. Cervantes 1 PAC Learning We want to develop a theory to relate the probability of successful
More informationLORENTZ SPACES AND REAL INTERPOLATION THE KEEL-TAO APPROACH
LORENTZ SPACES AND REAL INTERPOLATION THE KEEL-TAO APPROACH GUILLERMO REY. Introduction If an operator T is bounded on two Lebesgue spaces, the theory of coplex interpolation allows us to deduce the boundedness
More informationIntelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes
More informationMath Reviews classifications (2000): Primary 54F05; Secondary 54D20, 54D65
The Monotone Lindelöf Property and Separability in Ordered Spaces by H. Bennett, Texas Tech University, Lubbock, TX 79409 D. Lutzer, College of Willia and Mary, Williasburg, VA 23187-8795 M. Matveev, Irvine,
More informationMath Real Analysis The Henstock-Kurzweil Integral
Math 402 - Real Analysis The Henstock-Kurzweil Integral Steven Kao & Jocelyn Gonzales April 28, 2015 1 Introduction to the Henstock-Kurzweil Integral Although the Rieann integral is the priary integration
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October
More informationCHAPTER 3 REVIEW QUESTIONS MATH 3034 Spring a 1 b 1
. Let U = { A M (R) A = and b 6 }. CHAPTER 3 REVIEW QUESTIONS MATH 334 Spring 7 a b a and b are integers and a 6 (a) Let S = { A U det A = }. List the elements of S; that is S = {... }. (b) Let T = { A
More informationON SOME PROBLEMS OF GYARMATI AND SÁRKÖZY. Le Anh Vinh Mathematics Department, Harvard University, Cambridge, Massachusetts
#A42 INTEGERS 12 (2012) ON SOME PROLEMS OF GYARMATI AND SÁRKÖZY Le Anh Vinh Matheatics Departent, Harvard University, Cabridge, Massachusetts vinh@ath.harvard.edu Received: 12/3/08, Revised: 5/22/11, Accepted:
More informationA Bernstein-Markov Theorem for Normed Spaces
A Bernstein-Markov Theore for Nored Spaces Lawrence A. Harris Departent of Matheatics, University of Kentucky Lexington, Kentucky 40506-0027 Abstract Let X and Y be real nored linear spaces and let φ :
More informationOn the Inapproximability of Vertex Cover on k-partite k-uniform Hypergraphs
On the Inapproxiability of Vertex Cover on k-partite k-unifor Hypergraphs Venkatesan Guruswai and Rishi Saket Coputer Science Departent Carnegie Mellon University Pittsburgh, PA 1513. Abstract. Coputing
More informationlecture 37: Linear Multistep Methods: Absolute Stability, Part I lecture 38: Linear Multistep Methods: Absolute Stability, Part II
lecture 37: Linear Multistep Methods: Absolute Stability, Part I lecture 3: Linear Multistep Methods: Absolute Stability, Part II 5.7 Linear ultistep ethods: absolute stability At this point, it ay well
More informationSingularities of divisors on abelian varieties
Singularities of divisors on abelian varieties Olivier Debarre March 20, 2006 This is joint work with Christopher Hacon. We work over the coplex nubers. Let D be an effective divisor on an abelian variety
More informationMonochromatic images
CHAPTER 8 Monochroatic iages 1 The Central Sets Theore Lea 11 Let S,+) be a seigroup, e be an idepotent of βs and A e There is a set B A in e such that, for each v B, there is a set C A in e with v+c A
More informationarxiv: v1 [math.nt] 14 Sep 2014
ROTATION REMAINDERS P. JAMESON GRABER, WASHINGTON AND LEE UNIVERSITY 08 arxiv:1409.411v1 [ath.nt] 14 Sep 014 Abstract. We study properties of an array of nubers, called the triangle, in which each row
More informationHandout 7. and Pr [M(x) = χ L (x) M(x) =? ] = 1.
Notes on Coplexity Theory Last updated: October, 2005 Jonathan Katz Handout 7 1 More on Randoized Coplexity Classes Reinder: so far we have seen RP,coRP, and BPP. We introduce two ore tie-bounded randoized
More informationStatistical and Computational Learning Theory
Statistical and Computational Learning Theory Fundamental Question: Predict Error Rates Given: Find: The space H of hypotheses The number and distribution of the training examples S The complexity of the
More informationCSE525: Randomized Algorithms and Probabilistic Analysis May 16, Lecture 13
CSE55: Randoied Algoriths and obabilistic Analysis May 6, Lecture Lecturer: Anna Karlin Scribe: Noah Siegel, Jonathan Shi Rando walks and Markov chains This lecture discusses Markov chains, which capture
More information#A52 INTEGERS 10 (2010), COMBINATORIAL INTERPRETATIONS OF BINOMIAL COEFFICIENT ANALOGUES RELATED TO LUCAS SEQUENCES
#A5 INTEGERS 10 (010), 697-703 COMBINATORIAL INTERPRETATIONS OF BINOMIAL COEFFICIENT ANALOGUES RELATED TO LUCAS SEQUENCES Bruce E Sagan 1 Departent of Matheatics, Michigan State University, East Lansing,
More informationComputational Learning Theory. Definitions
Computational Learning Theory Computational learning theory is interested in theoretical analyses of the following issues. What is needed to learn effectively? Sample complexity. How many examples? Computational
More informationMath 262A Lecture Notes - Nechiporuk s Theorem
Math 6A Lecture Notes - Nechiporuk s Theore Lecturer: Sa Buss Scribe: Stefan Schneider October, 013 Nechiporuk [1] gives a ethod to derive lower bounds on forula size over the full binary basis B The lower
More informationOn Certain C-Test Words for Free Groups
Journal of Algebra 247, 509 540 2002 doi:10.1006 jabr.2001.9001, available online at http: www.idealibrary.co on On Certain C-Test Words for Free Groups Donghi Lee Departent of Matheatics, Uni ersity of
More informationRademacher Complexity Margin Bounds for Learning with a Large Number of Classes
Radeacher Coplexity Margin Bounds for Learning with a Large Nuber of Classes Vitaly Kuznetsov Courant Institute of Matheatical Sciences, 25 Mercer street, New York, NY, 002 Mehryar Mohri Courant Institute
More informationPolygonal Designs: Existence and Construction
Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G
More informationLecture Notes on Metric Spaces
Lecture Notes on Metric Spaces Math 117: Summer 2007 John Douglas Moore Our goal of these notes is to explain a few facts regarding metric spaces not included in the first few chapters of the text [1],
More informationA Simple Regression Problem
A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where
More informationIntroduction to Robotics (CS223A) (Winter 2006/2007) Homework #5 solutions
Introduction to Robotics (CS3A) Handout (Winter 6/7) Hoework #5 solutions. (a) Derive a forula that transfors an inertia tensor given in soe frae {C} into a new frae {A}. The frae {A} can differ fro frae
More informationThe concavity and convexity of the Boros Moll sequences
The concavity and convexity of the Boros Moll sequences Ernest X.W. Xia Departent of Matheatics Jiangsu University Zhenjiang, Jiangsu 1013, P.R. China ernestxwxia@163.co Subitted: Oct 1, 013; Accepted:
More informationHandout 6 Solutions to Problems from Homework 2
CS 85/185 Fall 2003 Lower Bounds Handout 6 Solutions to Probles fro Hoewor 2 Ait Charabarti Couter Science Dartouth College Solution to Proble 1 1.2: Let f n stand for A 111 n. To decide the roerty f 3
More informationCSE 648: Advanced algorithms
CSE 648: Advanced algorithms Topic: VC-Dimension & ɛ-nets April 03, 2003 Lecturer: Piyush Kumar Scribed by: Gen Ohkawa Please Help us improve this draft. If you read these notes, please send feedback.
More informationε-nets and VC Dimension
ε-nets and VC Dimension Sampling is a powerful idea applied widely in many disciplines, including CS. There are at least two important uses of sampling: estimation and detection. CNN, Nielsen, NYT etc
More informationUniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval
Unifor Approxiation and Bernstein Polynoials with Coefficients in the Unit Interval Weiang Qian and Marc D. Riedel Electrical and Coputer Engineering, University of Minnesota 200 Union St. S.E. Minneapolis,
More informationImproved Guarantees for Agnostic Learning of Disjunctions
Iproved Guarantees for Agnostic Learning of Disjunctions Pranjal Awasthi Carnegie Mellon University pawasthi@cs.cu.edu Avri Blu Carnegie Mellon University avri@cs.cu.edu Or Sheffet Carnegie Mellon University
More informationLecture October 23. Scribes: Ruixin Qiang and Alana Shine
CSCI699: Topics in Learning and Gae Theory Lecture October 23 Lecturer: Ilias Scribes: Ruixin Qiang and Alana Shine Today s topic is auction with saples. 1 Introduction to auctions Definition 1. In a single
More informationActive Learning: Disagreement Coefficient
Advanced Course in Machine Learning Spring 2010 Active Learning: Disagreement Coefficient Handouts are jointly prepared by Shie Mannor and Shai Shalev-Shwartz In previous lectures we saw examples in which
More informationCS Lecture 13. More Maximum Likelihood
CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood
More informationComputational Learning Theory: Shattering and VC Dimensions. Machine Learning. Spring The slides are mainly from Vivek Srikumar
Computational Learning Theory: Shattering and VC Dimensions Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 This lecture: Computational Learning Theory The Theory of Generalization
More informationMachine Learning. VC Dimension and Model Complexity. Eric Xing , Fall 2015
Machine Learning 10-701, Fall 2015 VC Dimension and Model Complexity Eric Xing Lecture 16, November 3, 2015 Reading: Chap. 7 T.M book, and outline material Eric Xing @ CMU, 2006-2015 1 Last time: PAC and
More informationSolutions of some selected problems of Homework 4
Solutions of soe selected probles of Hoework 4 Sangchul Lee May 7, 2018 Proble 1 Let there be light A professor has two light bulbs in his garage. When both are burned out, they are replaced, and the next
More informationPhysics 215 Winter The Density Matrix
Physics 215 Winter 2018 The Density Matrix The quantu space of states is a Hilbert space H. Any state vector ψ H is a pure state. Since any linear cobination of eleents of H are also an eleent of H, it
More informationMODULAR HYPERBOLAS AND THE CONGRUENCE ax 1 x 2 x k + bx k+1 x k+2 x 2k c (mod m)
#A37 INTEGERS 8 (208) MODULAR HYPERBOLAS AND THE CONGRUENCE ax x 2 x k + bx k+ x k+2 x 2k c (od ) Anwar Ayyad Departent of Matheatics, Al Azhar University, Gaza Strip, Palestine anwarayyad@yahoo.co Todd
More informationThis article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and
This article appeared in a ournal published by Elsevier. The attached copy is furnished to the author for internal non-coercial research and education use, including for instruction at the authors institution
More informationSupplementary to Learning Discriminative Bayesian Networks from High-dimensional Continuous Neuroimaging Data
Suppleentary to Learning Discriinative Bayesian Networks fro High-diensional Continuous Neuroiaging Data Luping Zhou, Lei Wang, Lingqiao Liu, Philip Ogunbona, and Dinggang Shen Proposition. Given a sparse
More informationVapnik-Chervonenkis Dimension of Neural Nets
P. L. Bartlett and W. Maass: Vapnik-Chervonenkis Dimension of Neural Nets 1 Vapnik-Chervonenkis Dimension of Neural Nets Peter L. Bartlett BIOwulf Technologies and University of California at Berkeley
More informationf(1), and so, if f is continuous, f(x) = f(1)x.
2.2.35: Let f be a additive fuctio. i Clearly fx = fx ad therefore f x = fx for all Z+ ad x R. Hece, for ay, Z +, f = f, ad so, if f is cotiuous, fx = fx. ii Suppose that f is bouded o soe o-epty ope set.
More information4 = (0.02) 3 13, = 0.25 because = 25. Simi-
Theore. Let b and be integers greater than. If = (. a a 2 a i ) b,then for any t N, in base (b + t), the fraction has the digital representation = (. a a 2 a i ) b+t, where a i = a i + tk i with k i =
More informationHomework 3 Solutions CSE 101 Summer 2017
Hoework 3 Solutions CSE 0 Suer 207. Scheduling algoriths The following n = 2 jobs with given processing ties have to be scheduled on = 3 parallel and identical processors with the objective of iniizing
More information12 th Annual Johns Hopkins Math Tournament Saturday, February 19, 2011 Power Round-Poles and Polars
1 th Annual Johns Hopkins Math Tournaent Saturday, February 19, 011 Power Round-Poles and Polars 1. Definition and Basic Properties 1. Note that the unit circles are not necessary in the solutions. They
More informationComputable Shell Decomposition Bounds
Coputable Shell Decoposition Bounds John Langford TTI-Chicago jcl@cs.cu.edu David McAllester TTI-Chicago dac@autoreason.co Editor: Leslie Pack Kaelbling and David Cohn Abstract Haussler, Kearns, Seung
More informationConsistent Multiclass Algorithms for Complex Performance Measures. Supplementary Material
Consistent Multiclass Algoriths for Coplex Perforance Measures Suppleentary Material Notations. Let λ be the base easure over n given by the unifor rando variable (say U over n. Hence, for all easurable
More informationOn the Existence of Pure Nash Equilibria in Weighted Congestion Games
MATHEMATICS OF OPERATIONS RESEARCH Vol. 37, No. 3, August 2012, pp. 419 436 ISSN 0364-765X (print) ISSN 1526-5471 (online) http://dx.doi.org/10.1287/oor.1120.0543 2012 INFORMS On the Existence of Pure
More informationLearnability of Gaussians with flexible variances
Learnability of Gaussians with flexible variances Ding-Xuan Zhou City University of Hong Kong E-ail: azhou@cityu.edu.hk Supported in part by Research Grants Council of Hong Kong Start October 20, 2007
More information2 Q 10. Likewise, in case of multiple particles, the corresponding density in 2 must be averaged over all
Lecture 6 Introduction to kinetic theory of plasa waves Introduction to kinetic theory So far we have been odeling plasa dynaics using fluid equations. The assuption has been that the pressure can be either
More informationOn Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40
On Poset Merging Peter Chen Guoli Ding Steve Seiden Abstract We consider the follow poset erging proble: Let X and Y be two subsets of a partially ordered set S. Given coplete inforation about the ordering
More information