Generative learning methods for bags of features

Similar documents
Kernel Density Topic Models: Visual Topics Without Visual Words

Lecture 2: Correlated Topic Model

Probabilistic Latent Semantic Analysis

Latent Dirichlet Allocation in Web Spam Filtering

The Effect of a Finite Measurement Volume on Power Spectra from a Burst Type LDA

Global Behaviour Inference using Probabilistic Latent Semantic Analysis

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

Nonlinear Estimation. Professor David H. Staelin

Bayesian Networks Practice

Lecture 6 : Dimensionality Reduction

Distinguish between different types of scenes. Matching human perception Understanding the environment

Colin Cameron: Asymptotic Theory for OLS

Bayesian Networks Practice

Colin Cameron: Brief Asymptotic Theory for 240A

Probabilistic Learning

CS145: INTRODUCTION TO DATA MINING

Spectral Hashing: Learning to Leverage 80 Million Images

Lecture 13 Visual recognition

Convergence of random variables, and the Borel-Cantelli lemmas

Outline. Markov Chains and Markov Models. Outline. Markov Chains. Markov Chains Definitions Huizhen Yu

Consistency and asymptotic normality

Collaborative Place Models Supplement 1

Introduction to Machine Learning

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling

Note for plsa and LDA-Version 1.1

Probabilistic Latent Semantic Analysis

Capacity Allocation. Outline Two-class allocation Multiple-class allocation Demand dependence Bid prices. Based on Phillips (2005) Chapter 7.

topic modeling hanna m. wallach

Topic Modeling: Beyond Bag-of-Words

The Bayes classifier

Expectation Maximization Mixture Models HMMs

Probabilistic Learning

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling

Topics. Bayesian Learning. What is Bayesian Learning? Objectives for Bayesian Learning

PER-PATCH METRIC LEARNING FOR ROBUST IMAGE MATCHING. Sezer Karaoglu, Ivo Everts, Jan C. van Gemert, and Theo Gevers

Named Entity Recognition using Maximum Entropy Model SEEM5680

CS9840 Learning and Computer Vision Prof. Olga Veksler. Lecture 2. Some Concepts from Computer Vision Curse of Dimensionality PCA

Bayesian Methods: Naïve Bayes

COS 424: Interacting with Data. Lecturer: Dave Blei Lecture #11 Scribe: Andrew Ferguson March 13, 2007

Information retrieval LSI, plsi and LDA. Jian-Yun Nie

7. Introduction to Large Sample Theory

CS6220: DATA MINING TECHNIQUES

Rasterization: Bresenham s Midpoint Algorithm

Document and Topic Models: plsa and LDA

Topic Models. Brandon Malone. February 20, Latent Dirichlet Allocation Success Stories Wrap-up

Generative Clustering, Topic Modeling, & Bayesian Inference

Deep Learning Basics Lecture 10: Neural Language Models. Princeton University COS 495 Instructor: Yingyu Liang

Categorization ANLP Lecture 10 Text Categorization with Naive Bayes

ANLP Lecture 10 Text Categorization with Naive Bayes

CS Lecture 18. Topic Models and LDA

Learning Markov Graphs Up To Edit Distance

Level Construction of Decision Trees in a Partition-based Framework for Classification

Bayesian Learning. CSL603 - Fall 2017 Narayanan C Krishnan

Class-Specific Simplex-Latent Dirichlet Allocation for Image Classification

Consistency and asymptotic normality

Class-Specific Simplex-Latent Dirichlet Allocation for Image Classification

The basic model of decision theory under risk. Theory of expected utility (Bernoulli-Principle) Introduction to Game Theory

CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, and Tony Wu

Collaborative topic models: motivations cont

Markov Chain Analysis of the Sequential Probability Ratio Test for Automatic Track Maintenance

Expectation maximization

N-gram N-gram Language Model for Large-Vocabulary Continuous Speech Recognition

VCMC: Variational Consensus Monte Carlo

Relative Performance Guarantees for Approximate Inference in Latent Dirichlet Allocation

Shared Segmentation of Natural Scenes. Dependent Pitman-Yor Processes

Parameter estimation: A new approach to weighting a priori information

Topic Modeling: Beyond Bag-of-Words

An Algorithm for Fast Calculation of Back-off N-gram Probabilities with Unigram Rescaling

Modeling Complex Temporal Composition of Actionlets for Activity Prediction

IMAGE classification is a topic of significant interest within

Econometrics I. September, Part I. Department of Economics Stanford University

GRAPH LIMITS AND EXCHANGEABLE RANDOM GRAPHS

Fisher Vector image representation

CuO/ Water Nanofluid Heat Transfer Through Triangular Ducts

Probabilistic Machine Learning

The Naïve Bayes Classifier. Machine Learning Fall 2017

Lecture 2: Probability, Naive Bayes

Detecting Humans via Their Pose

Machine Learning 4. week

Asian International School Curriculum Mapping Level : Starter Subject : Mathematics_ School Year:

Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text

Biitlli Biointelligence Laboratory Lb School of Computer Science and Engineering. Seoul National University

AI*IA 2003 Fusion of Multiple Pattern Classifiers PART III

Logistic Regression. Machine Learning Fall 2018

Global Scene Representations. Tilke Judd

A method of constructing the half-rate QC-LDPC codes with linear encoder, maximum column weight three and inevitable girth 26

Bivariate distributions characterized by one family of conditionals and conditional percentile or mode functions

Naïve Bayes. Vibhav Gogate The University of Texas at Dallas

Graph Cut based Inference with Co-occurrence Statistics. Ľubor Ladický, Chris Russell, Pushmeet Kohli, Philip Torr

Pattern Recognition and Machine Learning. Learning and Evaluation of Pattern Recognition Processes

Knowledge Discovery and Data Mining 1 (VO) ( )

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Neural Networks. Tobias Scheffer

Graphical Models. Mark Gales. Lent Machine Learning for Language Processing: Lecture 3. MPhil in Advanced Computer Science

PROBABILISTIC LATENT SEMANTIC ANALYSIS

INVESTIGATION OF THE DEISGN AND PERFORMANCE OF REPEATING SPACE TRACK CONSTELLATIONS

General Linear Model Introduction, Classes of Linear models and Estimation

Unsupervised Activity Perception by Hierarchical Bayesian Models

On the optimal control of linear complementarity systems

CHS GUSSET PLATE CONNECTIONS ANALYSES Theoretical and Experimental Approaches

Replicated Softmax: an Undirected Topic Model. Stephen Turner

Transcription:

Generative learning methos for bags of features Moel the robability of a bag of features given a class Many slies aate from Fei-Fei Li, Rob Fergus, an Antonio Torralba

Generative methos We ill cover to moels, both insire by text ocument analysis: Naïve Bayes Probabilistic Latent Semantic Analysis

The Naïve Bayes moel Assume that each feature is conitionally ineenent given the class N f, K, f c = f c 1 N i i= 1 f i : ith feature in the image N: number of features in the image Csurka et al. 2004

The Naïve Bayes moel Assume that each feature is conitionally ineenent given the class N M n 1 c i= 1 = 1 f, K, f = = N c fi c f i : ith feature in the image N: number of features in the image : th visual or in the vocabulary M: sie of visual vocabulary n : number of features of tye in the image Csurka et al. 2004

The Naïve Bayes moel Assume that each feature is conitionally ineenent given the class N M n 1 c i= 1 = 1 f, K, f = = N c fi c No. of features of tye in training images of class c c = Total no. of features in training images of class c Csurka et al. 2004

The Naïve Bayes moel Assume that each feature is conitionally ineenent given the class N M n 1 c i= 1 = 1 f, K, f = = N c fi c No. of features of tye in training images of class c + 1 c = Total no. of features in training images of class c + M Lalace l smoothing to avoi ero counts Csurka et al. 2004

The Naïve Bayes moel Maximum A Posteriori ecision: c* = arg max c c M = 1 c n = arg max c log c + M = 1 n log c you shoul comute the log of the likelihoo instea of the likelihoo itself in orer to avoi unerflo Csurka et al. 2004

The Naïve Bayes moel Grahical moel : c N Csurka et al. 2004

Probabilistic Latent Semantic Analysis = 1 + 2 + 3 Image ebra grass tree visual toics T. Hofmann, Probabilistic Latent Semantic Analysis, UAI 1999

Probabilistic Latent Semantic Analysis Unsuervise technique To-level generative moel: a ocument is a mixture of toics, an each toic has its on characteristic or istribution ocument toic or P P T. Hofmann, Probabilistic Latent Semantic Analysis, UAI 1999

Probabilistic Latent Semantic Analysis Unsuervise technique To-level generative moel: a ocument is a mixture of toics, an each toic has its on characteristic or istribution i K = k = 1 i k k T. Hofmann, Probabilistic Latent Semantic Analysis, UAI 1999

The LSA moel i K = k = 1 i k k Probability of or i Probability of Probability of in ocument knon or i given toic k unknon toic k given ocument unknon

The LSA moel i K = k = 1 i k k ocuments toics ocuments ors ors toics k i = i k Observe coeor Coeor istributions Class istributions istributions er toic class er image M N M K K N

Learning LSA arameters Maximie likelihoo of ata: Observe counts of or i in ocument M number of coeors N number of images Slie creit: Josef Sivic

Inference Fining the most likely toic class for an image: = arg max

Inference Inference Fining the most likely toic class for an image: max arg = Fining the most likely toic class for a visual Fining the most likely toic class for a visual or in a given image: = = arg max, arg max

Toic iscovery in images J. Sivic, B. Russell, A. Efros, A. Zisserman, B. Freeman, Discovering Obects an their Location in Images, ICCV 2005

Alication of LSA: Action recognition Sace-time interest oints Juan Carlos Niebles, Hongcheng Wang an Li Fei-Fei, Unsuervise Learning of Human Action Categories Using Satial-Temoral Wors, IJCV 2008.

Alication of LSA: Action recognition Juan Carlos Niebles, Hongcheng Wang an Li Fei-Fei, Unsuervise Learning of Human Action Categories Using Satial-Temoral Wors, IJCV 2008.

LSA moel i K = k = 1 i k k Probability bilit of or i Probability of Probability of in vieo knon or i given toic k unknon toic k given vieo unknon i = satial-temoral or = vieo n i, = co-occurrence table # of occurrences of or i in vieo = toic, corresoning to an action

Action recognition examle

Multile Actions

Multile Actions

Summary: Generative moels Naïve Bayes Unigram moels in ocument analysis Assumes conitional ineenence of ors given class Parameter estimation: frequency counting Probabilistic Latent Semantic Analysis Unsuervise technique Each ocument is a mixture of toics image is a mixture of classes Can be thought of as matrix ecomosition Parameter estimation: Exectation-Maximiation