Uncovering the Latent Structures of Crowd Labeling
|
|
- Dana Hampton
- 5 years ago
- Views:
Transcription
1 Uncovering the Latent Structures of Crowd Labeling Tian Tian and Jun Zhu Presenter:XXX Tsinghua University 1 / 26
2 Motivation Outline 1 Motivation 2 Related Works 3 Crowdsourcing Latent Class 4 Experiments 5 Conclusion 2 / 26
3 Motivation Motivation: Background Artificial intelligence are relying more and more on large-scale training datasets. Expert labeling are expensive and time-consuming. 3 / 26
4 Motivation Motivation: Crowdsourcing Solution Using multiple web workers to label each item. Recovery the ground truth from the noisy data. Table: Different workers may give inconsistent labels to a same item. worker a worker b worker c worker d item item item item / 26
5 Related Works Outline 1 Motivation 2 Related Works 3 Crowdsourcing Latent Class 4 Experiments 5 Conclusion 5 / 26
6 Related Works Majority Voting Majority Voting Assumption For every worker, the ground truth label is always the most common to be given, and the labels for each item are given independently. P(Y m = d) = (n,m) I δ W nm,d d,(n,m) I δ, m, (1) W nm,d 6 / 26
7 Related Works Dawid-Skene Estimator Dawid-Skene estimator Assumption The performance of a worker is consistent across different items, and his or her behavior can be measured by a confusion matrix. Table: An example of binary classification confusion matrix. label A label B label A 1 0 label B 1/3 2/3 7 / 26
8 Related Works Dawid-Skene Estimator Dawid-Skene estimator Assumption The performance of a worker is consistent across different items, and his or her behavior can be measured by a confusion matrix. P(W q, p) = m d q d n,l p ndl δ W nm,l, (2) p ndl : the probability that worker n labels an item as l when it s ground truth label is d. 8 / 26
9 Crowdsourcing Latent Class Outline 1 Motivation 2 Related Works 3 Crowdsourcing Latent Class 4 Experiments 5 Conclusion 9 / 26
10 Crowdsourcing Latent Class Latent Class Assumptions Latent Class Assumptions Assumption I. Each item belongs to one specific latent class only. Assumption II. Items in the same latent class should be classified into the same category. 10 / 26
11 Crowdsourcing Latent Class Latent Class Assumptions Latent Class Assumptions Each item belongs to one specific latent class only. Items in the same latent class in the same category. Figure: Latent classes of a binary classification task to classify fruit and vegetable. 11 / 26
12 Crowdsourcing Latent Class Latent Class Dawid-Skene Estimator Latent Class Confusion Matrix Extend the original confusion matrix. An entry p nkl in this matrix represents the probability that worker n gives label l to an item which belongs to latent class k. Table: An example of latent class confusion matrix. tomato pumpkin cuke apple Fruit 2/3 0 1/3 1 Vegetable 1/3 1 2/ / 26
13 Crowdsourcing Latent Class Latent Class Dawid-Skene Estimator Chinese Restaurant Process The number of latent classes K is unknown in advance. we choose a nonparametric prior for the latent class confusion matrix to learn the number of classes. Z i is the latent class assignment of item i, with the CRP prior, the probability Old Class : P(Z i = k) n k + α, k = 1 K, (3) New Class : P(Z i = K + 1) α. (4) 13 / 26
14 Crowdsourcing Latent Class Latent Class Dawid-Skene Estimator Nonparametric Dawid-Skene Model Full Bayesian Model: Assign : Z α c CRP(α c ), (5) Entries : p nk α d Dirichlet(α d ), n, k, (6) Obervation : W nm Z, p n Multinomial(A nm ), n, m, (7) Here A nm = {A nm1,, A nmd }, A nmd = K k=1 p nkd δ Zm,k 14 / 26
15 Crowdsourcing Latent Class Latent Class Dawid-Skene Estimator Inference We use Gibbs Sampling to infer the hidden variables. Conditional Distributions: confusion matrix parameter: p nk Z, W Dirichlet(p nk B nk ), n, k, (8) where B nkd = M m=1 δ W nm,dδ Zm,k + α d /D. hidden variables, when k K, P(Z m = k Z m, p, W) n k When generating a new class, P(Z m = k new Z m, p, W) α c N n=1 N n=1 d=1 D δ p Wnm,d nkd, (9) D d=1 Γ(δ W nm,d + α d /D). (10) Γ(1 + α d ) 15 / 26
16 Crowdsourcing Latent Class Latent Class Minimax Entropy Estimator Latent Class Minimax Entropy Estimator We also extend the minimax entropy estimator. Idea: change the category constraints. New objective: min Z max p nmd log p nmd α m τmd 2 p,τ,σ 2 n,m,d m,d n,m,d ( ) pnmd δ Wnm,d = τmd, m, d, β n σ 2 ndk 2 m s.t. n ( ) pnmd δ Wnm,d δzm,k = σ ndk, n, k, p nmd = 1, n, m. d (11) 16 / 26
17 Crowdsourcing Latent Class Category Recovery Category Recovery Regard items in a same class as one imaginary item, here we call it a hyper-item. We use a generalized Dawid-Skene estimator with hyper-items to estimate the category assignments. P(W q, p) = n q d p nkd ndl, (12) k d n,l where n nkd = m δ W nm,dδ Zm,k is the count of labels that worker n gives to hyper-item k. 17 / 26
18 Experiments Outline 1 Motivation 2 Related Works 3 Crowdsourcing Latent Class 4 Experiments 5 Conclusion 18 / 26
19 Experiments Synthetic Dataset 4 latent classes, 40 items parameters for each latent class. 2 types of workers, 20 workers of each type. K α =0.1 c α =0.2 c α =0.8 c α c =0.4 α =1.6 c Error Rate (%) # Iterations (a) α c Figure: (a) shows the numbers of latent classes found by NDS with different color. (b) shows the average category recovery error rates. (b) 19 / 26
20 Experiments Real Flower Dataset 4 flower species, 50 pics for each. 2 species for each category. 36 workers, 2366 labels in total. (a) NDS (b) LC-ME (c) NDS+CR (d) LC-ME+CR Figure: entry corresponding to image, column corresponding to flower species. (a)(b): color denotes latent class. (c)(d): denotes category. 20 / 26
21 Experiments Real Flower Dataset 4 flower species, 50 pics for each. 2 species for each category. 36 workers, 2366 labels in total. Figure: Representative pictures for different latent classes.(best viewed in color). 21 / 26
22 Experiments Category Recovery Table: Performance of several models on flowers dataset. the average error rate of 10 trials, together with standard deviation, are presented. # MV ± ± ± ± DS ± ± ± ± NDS ± ± ± ± ME ± ± ± ± LC-ME ± ± ± ± The first line means the number of workers we used on estimation tasks. 22 / 26
23 Conclusion Outline 1 Motivation 2 Related Works 3 Crowdsourcing Latent Class 4 Experiments 5 Conclusion 23 / 26
24 Conclusion Conclusion We show a latent class structures in crowdsourcing. We propose algorithms to disclose meaningful latent classes. We show that latent structures can help to achieve higher accuracies on category recovery tasks. 24 / 26
25 References Outline 1 Motivation 2 Related Works 3 Crowdsourcing Latent Class 4 Experiments 5 Conclusion 25 / 26
26 References Selected Reference Dawid, A.P., and Skene, A.M.: Maximum likelihood estimation of observer error-rates using the em algorithm. Applied statistics (1979), Dempster, A.P., Laird, N.M., Rubin, D.B., etal.: Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal statistical Society 39, 1 (1977), Li, H., Yu, B., and Zhou, D.: Error rate analysis of labeling by crowdsourcing. Raykar, V.C., Yu, S., Zhao, L.H., Valadez, G.H., Florin, C., Bogoni, L., and Moy, L.: Learning from crowds. The Journal of Machine Learning Research 11 (2010), Sheshadri, A., and Lease, M.: Square: A benchmark for research on computing crowd consensus. In First AAAI Conference on Human Computation and Crowdsourcing (2013). Tian, Y., and Zhu, J.: Learning from crowds in the presence of schools of thought. In Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining (2012), ACM, Welinder, P., Branson, S., Belongie, S., and Perona, P.: The multidimensional wisdom of crowds. In Neural Information Processing Systems (2010), vol.10, Zhou, D., Platt, J.C., Basu, S., and Mao, Y.: Learning from the wisdom of crowds by minimax entropy. In Neural Information Processing Systems (2012), / 26
Learning from the Wisdom of Crowds by Minimax Entropy. Denny Zhou, John Platt, Sumit Basu and Yi Mao Microsoft Research, Redmond, WA
Learning from the Wisdom of Crowds by Minimax Entropy Denny Zhou, John Platt, Sumit Basu and Yi Mao Microsoft Research, Redmond, WA Outline 1. Introduction 2. Minimax entropy principle 3. Future work and
More informationCrowdsourcing via Tensor Augmentation and Completion (TAC)
Crowdsourcing via Tensor Augmentation and Completion (TAC) Presenter: Yao Zhou joint work with: Dr. Jingrui He - 1 - Roadmap Background Related work Crowdsourcing based on TAC Experimental results Conclusion
More informationImproving Quality of Crowdsourced Labels via Probabilistic Matrix Factorization
Human Computation AAAI Technical Report WS-12-08 Improving Quality of Crowdsourced Labels via Probabilistic Matrix Factorization Hyun Joon Jung School of Information University of Texas at Austin hyunjoon@utexas.edu
More informationAggregating Ordinal Labels from Crowds by Minimax Conditional Entropy. Denny Zhou Qiang Liu John Platt Chris Meek
Aggregating Ordinal Labels from Crowds by Minimax Conditional Entropy Denny Zhou Qiang Liu John Platt Chris Meek 2 Crowds vs experts labeling: strength Time saving Money saving Big labeled data More data
More informationarxiv: v2 [cs.lg] 17 Nov 2016
Approximating Wisdom of Crowds using K-RBMs Abhay Gupta Microsoft India R&D Pvt. Ltd. abhgup@microsoft.com arxiv:1611.05340v2 [cs.lg] 17 Nov 2016 Abstract An important way to make large training sets is
More informationAdaptive Crowdsourcing via EM with Prior
Adaptive Crowdsourcing via EM with Prior Peter Maginnis and Tanmay Gupta May, 205 In this work, we make two primary contributions: derivation of the EM update for the shifted and rescaled beta prior and
More informationTruth Discovery and Crowdsourcing Aggregation: A Unified Perspective
Truth Discovery and Crowdsourcing Aggregation: A Unified Perspective Jing Gao 1, Qi Li 1, Bo Zhao 2, Wei Fan 3, and Jiawei Han 4 1 SUNY Buffalo; 2 LinkedIn; 3 Baidu Research Big Data Lab; 4 University
More informationBudget-Optimal Task Allocation for Reliable Crowdsourcing Systems
Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems Sewoong Oh Massachusetts Institute of Technology joint work with David R. Karger and Devavrat Shah September 28, 2011 1 / 13 Crowdsourcing
More informationPermuation Models meet Dawid-Skene: A Generalised Model for Crowdsourcing
Permuation Models meet Dawid-Skene: A Generalised Model for Crowdsourcing Ankur Mallick Electrical and Computer Engineering Carnegie Mellon University amallic@andrew.cmu.edu Abstract The advent of machine
More informationLearning Medical Diagnosis Models from Multiple Experts
Learning Medical Diagnosis Models from Multiple Experts Hamed Valizadegan, Quang Nguyen, Milos Hauskrecht 1 Department of Computer Science, University of Pittsburgh, email: hamed, quang, milos@cs.pitt.edu
More informationOn the Impossibility of Convex Inference in Human Computation
On the Impossibility of Convex Inference in Human Computation Nihar B. Shah U.C. Berkeley nihar@eecs.berkeley.edu Dengyong Zhou Microsoft Research dengyong.zhou@microsoft.com Abstract Human computation
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationLearning with Noisy Labels. Kate Niehaus Reading group 11-Feb-2014
Learning with Noisy Labels Kate Niehaus Reading group 11-Feb-2014 Outline Motivations Generative model approach: Lawrence, N. & Scho lkopf, B. Estimating a Kernel Fisher Discriminant in the Presence of
More informationLearning from Crowds in the Presence of Schools of Thought
Learning from Crowds in the Presence of Schools of Thought ABSTRACT Yuandong Tian School of Computer Science Carnegie Mellon University Pittsburgh, PA 3, USA yuandong@andrew.cmu.edu Crowdsourcing has recently
More informationIterative Learning for Reliable Crowdsourcing Systems
Iterative Learning for Reliable Crowdsourcing Systems David R. Karger Sewoong Oh Devavrat Shah Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology Abstract Crowdsourcing
More informationCrowdsourcing label quality: a theoretical analysis
. RESEARCH PAPER. SCIENCE CHINA Information Sciences November 5, Vol. 58 xxxxxx: xxxxxx: doi: xxxxxxxxxxxxxx Crowdsourcing label quality: a theoretical analysis WANG Wei & ZHOU Zhi-Hua * National Key Laboratory
More informationPerformance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project
Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore
More informationCrowdsourcing with Multi-Dimensional Trust
18th International Conference on Information Fusion Washington, DC - July 6-9, 2015 Crowdsourcing with Multi-Dimensional Trust Xiangyang Liu, He He, John S. Baras Institute for Systems Research and Dept.
More informationFinite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier
Finite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier Kaizhu Huang, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More informationarxiv: v1 [cs.lg] 2 Jan 2019
A Full Probabilistic Model for Yes/No Type Crowdsourcing in Multi-Class Classification Belen Saldias-Fuentes Pavlos Protopapas Karim Pichara B arxiv:190100397v1 [cslg] 2 Jan 2019 Abstract Crowdsourcing
More informationOnline Bayesian Passive-Agressive Learning
Online Bayesian Passive-Agressive Learning International Conference on Machine Learning, 2014 Tianlin Shi Jun Zhu Tsinghua University, China 21 August 2015 Presented by: Kyle Ulrich Introduction Online
More informationArticle from. Predictive Analytics and Futurism. July 2016 Issue 13
Article from Predictive Analytics and Futurism July 2016 Issue 13 Regression and Classification: A Deeper Look By Jeff Heaton Classification and regression are the two most common forms of models fitted
More informationAlgorithmisches Lernen/Machine Learning
Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines
More informationCSCI-567: Machine Learning (Spring 2019)
CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March
More informationMixtures of Gaussians. Sargur Srihari
Mixtures of Gaussians Sargur srihari@cedar.buffalo.edu 1 9. Mixture Models and EM 0. Mixture Models Overview 1. K-Means Clustering 2. Mixtures of Gaussians 3. An Alternative View of EM 4. The EM Algorithm
More informationVariational Inference for Crowdsourcing
Variational Inference for Crowdsourcing Qiang Liu ICS, UC Irvine qliu1@ics.uci.edu Jian Peng TTI-C & CSAIL, MIT jpeng@csail.mit.edu Alexander Ihler ICS, UC Irvine ihler@ics.uci.edu Abstract Crowdsourcing
More informationLatent Variable Models and EM algorithm
Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic
More informationLecture 10. Announcement. Mixture Models II. Topics of This Lecture. This Lecture: Advanced Machine Learning. Recap: GMMs as Latent Variable Models
Advanced Machine Learning Lecture 10 Mixture Models II 30.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ Announcement Exercise sheet 2 online Sampling Rejection Sampling Importance
More informationCollaborative topic models: motivations cont
Collaborative topic models: motivations cont Two topics: machine learning social network analysis Two people: " boy Two articles: article A! girl article B Preferences: The boy likes A and B --- no problem.
More informationAggregating Crowdsourced Ordinal Labels via Bayesian Clustering
Aggregating Crowdsourced Ordinal Labels via Bayesian Clustering Xiawei Guo (B) and James T. Kwok Department of Computer Science and Engineering, Hong Kong University of Science and Technology, Clear Water
More informationarxiv: v2 [cs.lg] 20 May 2018
LEARNING FROM NOISY SINGLY-LABELED DATA Ashish Khetan University of Illinois at Urbana-Champaign Urbana, IL 61801 khetan2@illinois.edu Zachary C. Lipton Amazon Web Services Seattle, WA 98101 liptoz@amazon.com
More informationConvergence Rate of Expectation-Maximization
Convergence Rate of Expectation-Maximiation Raunak Kumar University of British Columbia Mark Schmidt University of British Columbia Abstract raunakkumar17@outlookcom schmidtm@csubcca Expectation-maximiation
More informationApplying Latent Dirichlet Allocation to Group Discovery in Large Graphs
Lawrence Livermore National Laboratory Applying Latent Dirichlet Allocation to Group Discovery in Large Graphs Keith Henderson and Tina Eliassi-Rad keith@llnl.gov and eliassi@llnl.gov This work was performed
More informationDeep Poisson Factorization Machines: a factor analysis model for mapping behaviors in journalist ecosystem
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationMachine Learning for Signal Processing Bayes Classification
Machine Learning for Signal Processing Bayes Classification Class 16. 24 Oct 2017 Instructor: Bhiksha Raj - Abelino Jimenez 11755/18797 1 Recap: KNN A very effective and simple way of performing classification
More informationStat 542: Item Response Theory Modeling Using The Extended Rank Likelihood
Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal
More informationAutomated Segmentation of Low Light Level Imagery using Poisson MAP- MRF Labelling
Automated Segmentation of Low Light Level Imagery using Poisson MAP- MRF Labelling Abstract An automated unsupervised technique, based upon a Bayesian framework, for the segmentation of low light level
More informationStatistical Debugging with Latent Topic Models
Statistical Debugging with Latent Topic Models David Andrzejewski, Anne Mulhern, Ben Liblit, Xiaojin Zhu Department of Computer Sciences University of Wisconsin Madison European Conference on Machine Learning,
More informationParts 3-6 are EXAMPLES for cse634
1 Parts 3-6 are EXAMPLES for cse634 FINAL TEST CSE 352 ARTIFICIAL INTELLIGENCE Fall 2008 There are 6 pages in this exam. Please make sure you have all of them INTRODUCTION Philosophical AI Questions Q1.
More informationA Generative Block-Diagonal Model for Clustering
A Generative Block-Diagonal Model for Clustering Junxiang Chen Dept. of Electrical & Computer Engineering Northeastern University Boston, MA 02115 jchen@ece.neu.edu Jennifer Dy Dept. of Electrical & Computer
More informationLearning From Crowds. Presented by: Bei Peng 03/24/15
Learning From Crowds Presented by: Bei Peng 03/24/15 1 Supervised Learning Given labeled training data, learn to generalize well on unseen data Binary classification ( ) Multi-class classification ( y
More informationRaRE: Social Rank Regulated Large-scale Network Embedding
RaRE: Social Rank Regulated Large-scale Network Embedding Authors: Yupeng Gu 1, Yizhou Sun 1, Yanen Li 2, Yang Yang 3 04/26/2018 The Web Conference, 2018 1 University of California, Los Angeles 2 Snapchat
More informationContent-based Recommendation
Content-based Recommendation Suthee Chaidaroon June 13, 2016 Contents 1 Introduction 1 1.1 Matrix Factorization......................... 2 2 slda 2 2.1 Model................................. 3 3 flda 3
More informationarxiv: v3 [cs.lg] 25 Aug 2017
Achieving Budget-optimality with Adaptive Schemes in Crowdsourcing Ashish Khetan and Sewoong Oh arxiv:602.0348v3 [cs.lg] 25 Aug 207 Abstract Crowdsourcing platforms provide marketplaces where task requesters
More informationLearning Binary Classifiers for Multi-Class Problem
Research Memorandum No. 1010 September 28, 2006 Learning Binary Classifiers for Multi-Class Problem Shiro Ikeda The Institute of Statistical Mathematics 4-6-7 Minami-Azabu, Minato-ku, Tokyo, 106-8569,
More informationHYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH
HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi
More informationCS 188: Artificial Intelligence. Outline
CS 188: Artificial Intelligence Lecture 21: Perceptrons Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. Outline Generative vs. Discriminative Binary Linear Classifiers Perceptron Multi-class
More informationGenerative Clustering, Topic Modeling, & Bayesian Inference
Generative Clustering, Topic Modeling, & Bayesian Inference INFO-4604, Applied Machine Learning University of Colorado Boulder December 12-14, 2017 Prof. Michael Paul Unsupervised Naïve Bayes Last week
More informationChapter 14 Combining Models
Chapter 14 Combining Models T-61.62 Special Course II: Pattern Recognition and Machine Learning Spring 27 Laboratory of Computer and Information Science TKK April 3th 27 Outline Independent Mixing Coefficients
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 218 Outlines Overview Introduction Linear Algebra Probability Linear Regression 1
More informationMIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,
MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run
More informationPattern Recognition and Machine Learning. Bishop Chapter 9: Mixture Models and EM
Pattern Recognition and Machine Learning Chapter 9: Mixture Models and EM Thomas Mensink Jakob Verbeek October 11, 27 Le Menu 9.1 K-means clustering Getting the idea with a simple example 9.2 Mixtures
More informationA Randomized Approach for Crowdsourcing in the Presence of Multiple Views
A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion
More informationWe Prediction of Geological Characteristic Using Gaussian Mixture Model
We-07-06 Prediction of Geological Characteristic Using Gaussian Mixture Model L. Li* (BGP,CNPC), Z.H. Wan (BGP,CNPC), S.F. Zhan (BGP,CNPC), C.F. Tao (BGP,CNPC) & X.H. Ran (BGP,CNPC) SUMMARY The multi-attribute
More informationParametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a
Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a
More informationTopic Models. Brandon Malone. February 20, Latent Dirichlet Allocation Success Stories Wrap-up
Much of this material is adapted from Blei 2003. Many of the images were taken from the Internet February 20, 2014 Suppose we have a large number of books. Each is about several unknown topics. How can
More informationBandit-Based Task Assignment for Heterogeneous Crowdsourcing
Neural Computation, vol.27, no., pp.2447 2475, 205. Bandit-Based Task Assignment for Heterogeneous Crowdsourcing Hao Zhang Department of Computer Science, Tokyo Institute of Technology, Japan Yao Ma Department
More informationStructure Learning in Sequential Data
Structure Learning in Sequential Data Liam Stewart liam@cs.toronto.edu Richard Zemel zemel@cs.toronto.edu 2005.09.19 Motivation. Cau, R. Kuiper, and W.-P. de Roever. Formalising Dijkstra's development
More informationSmall-variance Asymptotics for Dirichlet Process Mixtures of SVMs
Small-variance Asymptotics for Dirichlet Process Mixtures of SVMs Yining Wang Jun Zhu Tsinghua University July, 2014 Y. Wang and J. Zhu (Tsinghua University) Max-Margin DP-means July, 2014 1 / 25 Outline
More informationChapter 16. Structured Probabilistic Models for Deep Learning
Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe
More informationStudy Notes on the Latent Dirichlet Allocation
Study Notes on the Latent Dirichlet Allocation Xugang Ye 1. Model Framework A word is an element of dictionary {1,,}. A document is represented by a sequence of words: =(,, ), {1,,}. A corpus is a collection
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationLINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,
More informationCS145: INTRODUCTION TO DATA MINING
CS145: INTRODUCTION TO DATA MINING Text Data: Topic Model Instructor: Yizhou Sun yzsun@cs.ucla.edu December 4, 2017 Methods to be Learnt Vector Data Set Data Sequence Data Text Data Classification Clustering
More informationAn Approach to Classification Based on Fuzzy Association Rules
An Approach to Classification Based on Fuzzy Association Rules Zuoliang Chen, Guoqing Chen School of Economics and Management, Tsinghua University, Beijing 100084, P. R. China Abstract Classification based
More informationMachine Learning for natural language processing
Machine Learning for natural language processing Hidden Markov Models Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 33 Introduction So far, we have classified texts/observations
More informationarxiv: v3 [stat.ml] 1 Nov 2014
Spectral Methods meet EM: A Provably Optimal Algorithm for Crowdsourcing Yuchen Zhang Xi Chen Dengyong Zhou Michael I. Jordan arxiv:406.3824v3 [stat.ml] Nov 204 November 4, 204 Abstract Crowdsourcing is
More informationDEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE
Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures
More informationOnline Bayesian Passive-Aggressive Learning
Online Bayesian Passive-Aggressive Learning Full Journal Version: http://qr.net/b1rd Tianlin Shi Jun Zhu ICML 2014 T. Shi, J. Zhu (Tsinghua) BayesPA ICML 2014 1 / 35 Outline Introduction Motivation Framework
More informationExpectation Maximisation (EM) CS 486/686: Introduction to Artificial Intelligence University of Waterloo
Expectation Maximisation (EM) CS 486/686: Introduction to Artificial Intelligence University of Waterloo 1 Incomplete Data So far we have seen problems where - Values of all attributes are known - Learning
More informationLecture 4 Discriminant Analysis, k-nearest Neighbors
Lecture 4 Discriminant Analysis, k-nearest Neighbors Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se fredrik.lindsten@it.uu.se
More informationClick Prediction and Preference Ranking of RSS Feeds
Click Prediction and Preference Ranking of RSS Feeds 1 Introduction December 11, 2009 Steven Wu RSS (Really Simple Syndication) is a family of data formats used to publish frequently updated works. RSS
More informationBayes Classifiers. CAP5610 Machine Learning Instructor: Guo-Jun QI
Bayes Classifiers CAP5610 Machine Learning Instructor: Guo-Jun QI Recap: Joint distributions Joint distribution over Input vector X = (X 1, X 2 ) X 1 =B or B (drinking beer or not) X 2 = H or H (headache
More informationMachine Learning Basics
Security and Fairness of Deep Learning Machine Learning Basics Anupam Datta CMU Spring 2019 Image Classification Image Classification Image classification pipeline Input: A training set of N images, each
More informationSparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference
Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Shunsuke Horii Waseda University s.horii@aoni.waseda.jp Abstract In this paper, we present a hierarchical model which
More informationA Bayesian model for fusing biomedical labels
Chapter 7 A Bayesian model for fusing biomedical labels Tingting Zhu, Gari D. Clifford and David A. Clifton 7.1 Background In manual annotation of data, significant intra- and inter-observer disagreements
More informationAn Empirical Study of Building Compact Ensembles
An Empirical Study of Building Compact Ensembles Huan Liu, Amit Mandvikar, and Jigar Mody Computer Science & Engineering Arizona State University Tempe, AZ 85281 {huan.liu,amitm,jigar.mody}@asu.edu Abstract.
More informationMachine Learning Basics Lecture 7: Multiclass Classification. Princeton University COS 495 Instructor: Yingyu Liang
Machine Learning Basics Lecture 7: Multiclass Classification Princeton University COS 495 Instructor: Yingyu Liang Example: image classification indoor Indoor outdoor Example: image classification (multiclass)
More informationAdvanced Machine Learning
Advanced Machine Learning Nonparametric Bayesian Models --Learning/Reasoning in Open Possible Worlds Eric Xing Lecture 7, August 4, 2009 Reading: Eric Xing Eric Xing @ CMU, 2006-2009 Clustering Eric Xing
More informationRandomized Decision Trees
Randomized Decision Trees compiled by Alvin Wan from Professor Jitendra Malik s lecture Discrete Variables First, let us consider some terminology. We have primarily been dealing with real-valued data,
More informationContents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)
Contents Lecture Lecture Linear Discriminant Analysis Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University Email: fredriklindsten@ituuse Summary of lecture
More informationLatent Dirichlet Allocation (LDA)
Latent Dirichlet Allocation (LDA) D. Blei, A. Ng, and M. Jordan. Journal of Machine Learning Research, 3:993-1022, January 2003. Following slides borrowed ant then heavily modified from: Jonathan Huang
More informationMidterm, Fall 2003
5-78 Midterm, Fall 2003 YOUR ANDREW USERID IN CAPITAL LETTERS: YOUR NAME: There are 9 questions. The ninth may be more time-consuming and is worth only three points, so do not attempt 9 unless you are
More informationINFINITE MIXTURES OF MULTIVARIATE GAUSSIAN PROCESSES
INFINITE MIXTURES OF MULTIVARIATE GAUSSIAN PROCESSES SHILIANG SUN Department of Computer Science and Technology, East China Normal University 500 Dongchuan Road, Shanghai 20024, China E-MAIL: slsun@cs.ecnu.edu.cn,
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationDirichlet Process Based Evolutionary Clustering
Dirichlet Process Based Evolutionary Clustering Tianbing Xu 1 Zhongfei (Mark) Zhang 1 1 Dept. of Computer Science State Univ. of New York at Binghamton Binghamton, NY 13902, USA {txu,zhongfei,blong}@cs.binghamton.edu
More informationA Bayesian Approach to Concept Drift
A Bayesian Approach to Concept Drift Stephen H. Bach Marcus A. Maloof Department of Computer Science Georgetown University Washington, DC 20007, USA {bach, maloof}@cs.georgetown.edu Abstract To cope with
More informationChapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems
LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 08: Direct Maximum Likelihood/MAP Estimation and Incomplete Data Problems Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute
More informationGaussian Mixture Model
Case Study : Document Retrieval MAP EM, Latent Dirichlet Allocation, Gibbs Sampling Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 5 th,
More informationCS446: Machine Learning Fall Final Exam. December 6 th, 2016
CS446: Machine Learning Fall 2016 Final Exam December 6 th, 2016 This is a closed book exam. Everything you need in order to solve the problems is supplied in the body of this exam. This exam booklet contains
More informationTUTORIAL PART 1 Unsupervised Learning
TUTORIAL PART 1 Unsupervised Learning Marc'Aurelio Ranzato Department of Computer Science Univ. of Toronto ranzato@cs.toronto.edu Co-organizers: Honglak Lee, Yoshua Bengio, Geoff Hinton, Yann LeCun, Andrew
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationThe Knowledge Gradient for Sequential Decision Making with Stochastic Binary Feedbacks
The Knowledge Gradient for Sequential Decision Making with Stochastic Binary Feedbacks Yingfei Wang, Chu Wang and Warren B. Powell Princeton University Yingfei Wang Optimal Learning Methods June 22, 2016
More informationShort Note: Naive Bayes Classifiers and Permanence of Ratios
Short Note: Naive Bayes Classifiers and Permanence of Ratios Julián M. Ortiz (jmo1@ualberta.ca) Department of Civil & Environmental Engineering University of Alberta Abstract The assumption of permanence
More informationCourse 395: Machine Learning
Course 395: Machine Learning Lecturers: Maja Pantic (maja@doc.ic.ac.uk) Stavros Petridis (sp104@doc.ic.ac.uk) Goal (Lectures): To present basic theoretical concepts and key algorithms that form the core
More informationArtificial Neural Networks
Introduction ANN in Action Final Observations Application: Poverty Detection Artificial Neural Networks Alvaro J. Riascos Villegas University of los Andes and Quantil July 6 2018 Artificial Neural Networks
More informationBayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington
Bayesian Classifiers and Probability Estimation Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1 Data Space Suppose that we have a classification problem The
More informationA Hybrid Generative/Discriminative Approach to Semi-supervised Classifier Design
A Hybrid Generative/Discriminative Approach to Semi-supervised Classifier Design Akinori Fujino, Naonori Ueda, and Kazumi Saito NTT Communication Science Laboratories, NTT Corporation 2-4, Hikaridai, Seika-cho,
More informationThe Benefits of a Model of Annotation
The Benefits of a Model of Annotation Rebecca J. Passonneau and Bob Carpenter Columbia University Center for Computational Learning Systems Department of Statistics LAW VII, August 2013 Conventional Approach
More information