AUTHORSHIP RECOGNITION USING THE DYNAMICS OF CO-OCCURRENCE NETWORKS. Camilo Akimushkin Valencia
|
|
- Verity Townsend
- 5 years ago
- Views:
Transcription
1 AUTHORSHIP RECOGNITION USING THE DYNAMICS OF CO-OCCURRENCE NETWORKS Camilo Akimushkin Valencia Instituto de Física de São Carlos Universidade de São Paulo 11 November 2016 C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
2 Authorship recognition Lexical features (frequency) Words, n-grams, functional words, types of words, discourse-connecting expressions, slang, contractions, dialects, orthography mistakes, proper names, semantic features (polysemy). Character-level features Character n-grams, frequent suffixes, punctuation. Text format Lengths of lines, words and phrases, formatting (white spaces), capitalization, nonalphanumeric characters, beginnings and ends of texts. Other Syntactic features: n-grams syntactic function, kinds of phrases, perplexity, morphological complexity. C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
3 Complex networks Natural C. Akimushkin (IFSC-USP) Artificial Social SoFiA Other 11 November / 24
4 Complexity of language Zipf s law (1939) f j = Ar q j, q = 1 P (k) k γ, γ = 1 + q 1 Biemann, Quasthoff. Networks generated from natural language text. In: Dynamics on and of Complex Networks Dorogovtsev, Mendes. Language as an Evolving Word Web. P. Roy. Soc. Lond. B. Bio C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
5 Word co-occurrence networks Construction It was the best of times, it was the worst of times, it was the age of wisdom, it was the age of foolishness... A Tale of Two Cities - Charles Dickens C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
6 Word co-occurrence networks Construction best worst age times times wisdom age foolishness C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
7 Word co-occurrence networks Construction best times good worst times foolishness time age wisdom age foolishness wisdom bad age C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
8 Network metrics C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
9 Network metrics 1 Clustering: c i = e i k i(k i 1) 2 Diameter: D = max{d ij } 3 Radius: r = min{d ij } 4 Cliques: Number of complete subgraphs 5 Load centrality: Betweenness centrality with loads on edges 6 Transitivity: T = 3 triangles triads 7 Betweenness centrality: B i = Σ s i t g i st g st 8 Shortest path: l ij = [A n ] ij 9 Connectivity: k i = [ A 2] ii 10 Intermittency: I i = var( )/ 11 Number of nodes: N 12 Number of edges: E A: Adjacency matrix; g st = Σl st; i : Distance between two appearances of a word. C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
10 Dynamics of networks for authorship recognition Authorship of books Few books per author Depends on style Small networks Uneven networks C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
11 Dynamics of networks for authorship recognition Authorship of books Few books per author Depends on style Small networks Uneven networks Dynamics OF the network C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
12 Time series Clustering Radius Load centrality Betweenness centrality Degree Nodes Diameter Cliques Transitivity Shortest path Intermittency Edges Time series for Moby Dick by H. Melville C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
13 Time series 2 1 Frequency Arthur Conan Doyle Bernard Shaw Fyodor Dostoyevsky Herman Melville Jack London Jonathan Swift Leo Tolstoy Nathaniel Hawthorne Connectivity Clustering Radius Load centrality Betweenness centrality Degree Nodes Diameter Cliques Transitivity Shortest path Intermittency Edges Time series for Moby Dick by H. Melville C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
14 Time series 2 1 Frequency Arthur Conan Doyle Bernard Shaw Fyodor Dostoyevsky Herman Melville Jack London Jonathan Swift Leo Tolstoy Nathaniel Hawthorne [ µ i = 1 T Connectivity T j=1 (x j µ 1 ) i ] 1/i 3 Clustering Radius Load centrality Betweenness centrality Degree Nodes Diameter Cliques Transitivity Shortest path Intermittency Edges Time series for Moby Dick by H. Melville C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
15 Time series Autocorrelation r(x, y) = Σ T i=1 (xi x)(yi ȳ) Σ T i=1 (x i x) 2 Σ T i=1 (yi ȳ)2 ACF (τ) = r(x, x τ ) ±5% Wiener-Khinchin theorem C(τ) x (t)x(τ + t)dt = x ν 2 e 2πıντ dν = F[ x ν 2 ](τ) C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
16 Time series Stationarity tests and ARIMA fittings Auto-regressive model AR(p) x t = a 1 x t 1 + a 2 x t a p x t p + ε t, t > p Characteristic equation: 1 a 1 z a 2 z a p z p = 0 Unit root tests: z = 1? Auto-Regressive Integrated Moving Average model ARIMA(p,d,q) ( ) ( ) p q 1 φ i L i (1 L) d x t = 1 + θ i L i ε t i=0 Lag operator: Lx t = x t 1 i=0 C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
17 Time series Stationarity tests Phillips-Perron KPSS Dickey-Fuller McKinnon Clustering Betweenness centrality Cliques Diameter Intermittency Load centrality Degree Radius Shortest path Edges Nodes Transitivity p value > 0.05 p value < 0.05 C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
18 Time series ARIMA fittings Network metric Value of d Clustering Betweenness centrality Cliques Diameter Intermittency Load centrality Degree Radius Shortest path Edges Nodes Transitivity ARIMA(p,0,q) Stationary 73% ARIMA(p,1,q) First order integrated 27% Total C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
19 Time series ARIMA fittings Table: Series fitted with an ARIMA(p,d,q) model having the biggest values of the sum p + d + q. ARIMA(p,d,q) Book Measure Sum p d q The Poems of Jonathan Swift, D.D., Volume 2 Load centrality The Journal to Stella Clustering The Iron Heel Clustering Typee: A Romance of the South Seas Edges C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
20 Data analysis Dimensionality reduction Feature selection Feature extraction Supervised learning Zero Rule: 1/8 = 12.5% One Rule Naive Bayes K-Nearest Neighbors J48 (tree) Radial Basis Function Networks 48 Attributes 80 Books 8 Authors Precision T P A P A = T P A + F P A Recall T P A R A = T P A + F N A T P : True Positives F P : False Positives F N: False Negatives C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
21 Feature selection Variance treshold Success score % J48 K-Nearest Neighbors 20 Naive Bayes RBFNetwork Variance treshold Success scores using variance threshold C D R Cq L T B S K I N E Attributes Features using variance threshold. Success score % J48 K-Nearest Neighbors Naive Bayes RBFNetwork Number of attributes Success scores using score-based criteria. J48 KNN NB RBFN C D R Cq L T B S K I N E Attributes Features using score-based criteria Success scores and combinations of features using feature selection. In the upper figures maximum values are marked with circles. In the lower figures if an attribute is present in the combination the corresponding cell is painted black. C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
22 Feature extraction Arthur Conan Doyle Bernard Shaw Fyodor Dostoyevsky Herman Melville Jack London Jonathan Swift Leo Tolstoy Nathaniel Hawthorne Feature extraction using ISOMAP. C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
23 Feature extraction Scores Original precision Original recall PCA precision PCA recall Isomap precision Isomap recall OneR J48 KNN NB RBFN Scores using feature extraction. C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
24 Summary of classification success scores Attributes J48 (%) KNN (%) NB (%) RBFN (%) Original set Variance threshold best Score-based best {µ 1 } {µ 2, µ 3, µ 4 } PCA ISOMAP C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
25 The role of words The Memoirs of Sherlock Holmes The Return of Sherlock Holmes Only one different word out of the 20 highest ranked! C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
26 Dissimilarity matrix C. Akimushkin (IFSC-USP) SoFiA November / 24
27 Dissimilarity matrix C. Akimushkin (IFSC-USP) SoFiA November / 24
28 Projection Success scores > 90% C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
29 Summary Time series are stationary. Global sample statistics can be obtained. Dynamic measures are author-dependent. Weight on edges is relevant. Dimensionality reduction enhances classification. Books are located on a curved manifold in attribute space. A word s role in a network is author-dependent. Network metrics must be jointly used for classification. Many hidden features of networks. C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
30 Muchas gracias! C. Akimushkin (IFSC-USP) SoFiA 11 November / 24
Univariate, Nonstationary Processes
Univariate, Nonstationary Processes Jamie Monogan University of Georgia March 20, 2018 Jamie Monogan (UGA) Univariate, Nonstationary Processes March 20, 2018 1 / 14 Objectives By the end of this meeting,
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationFace Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi
Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold
More informationFINAL: CS 6375 (Machine Learning) Fall 2014
FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for
More informationEconometric Methods for Panel Data
Based on the books by Baltagi: Econometric Analysis of Panel Data and by Hsiao: Analysis of Panel Data Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies
More informationCircle the single best answer for each multiple choice question. Your choice should be made clearly.
TEST #1 STA 4853 March 6, 2017 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. There are 32 multiple choice
More informationECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests
ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests Patrick Herb Brandeis University Spring 2016 Patrick Herb (Brandeis University) Unit Root Tests ECON/FIN
More informationChapter 2 The Naïve Bayes Model in the Context of Word Sense Disambiguation
Chapter 2 The Naïve Bayes Model in the Context of Word Sense Disambiguation Abstract This chapter discusses the Naïve Bayes model strictly in the context of word sense disambiguation. The theoretical model
More informationLECTURE NOTE #11 PROF. ALAN YUILLE
LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationNonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.
Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all
More informationEconometrics of financial markets, -solutions to seminar 1. Problem 1
Econometrics of financial markets, -solutions to seminar 1. Problem 1 a) Estimate with OLS. For any regression y i α + βx i + u i for OLS to be unbiased we need cov (u i,x j )0 i, j. For the autoregressive
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu November 16, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision
More informationScenario 5: Internet Usage Solution. θ j
Scenario : Internet Usage Solution Some more information would be interesting about the study in order to know if we can generalize possible findings. For example: Does each data point consist of the total
More informationRetrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1
Retrieval by Content Part 2: Text Retrieval Term Frequency and Inverse Document Frequency Srihari: CSE 626 1 Text Retrieval Retrieval of text-based information is referred to as Information Retrieval (IR)
More informationEmpirical Market Microstructure Analysis (EMMA)
Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction
More informationOnline appendix for the paper Scarcity without Leviathan: The Violent Effects of Cocaine Supply Shortages in the Mexican Drug War.
Online appendix for the paper Scarcity without Leviathan: The Violent Effects of Cocaine Supply Shortages in the Mexican Drug War. Juan Camilo Castillo Daniel Mejía Pascual Restrepo August 2016 Online
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu March 16, 2016 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision
More informationNotion of Distance. Metric Distance Binary Vector Distances Tangent Distance
Notion of Distance Metric Distance Binary Vector Distances Tangent Distance Distance Measures Many pattern recognition/data mining techniques are based on similarity measures between objects e.g., nearest-neighbor
More informationUnivariate ARIMA Models
Univariate ARIMA Models ARIMA Model Building Steps: Identification: Using graphs, statistics, ACFs and PACFs, transformations, etc. to achieve stationary and tentatively identify patterns and model components.
More informationIntroduction to Machine Learning. Introduction to ML - TAU 2016/7 1
Introduction to Machine Learning Introduction to ML - TAU 2016/7 1 Course Administration Lecturers: Amir Globerson (gamir@post.tau.ac.il) Yishay Mansour (Mansour@tau.ac.il) Teaching Assistance: Regev Schweiger
More informationTime Series Modeling. Shouvik Mani April 5, /688: Practical Data Science Carnegie Mellon University
Time Series Modeling Shouvik Mani April 5, 2018 15-388/688: Practical Data Science Carnegie Mellon University Goals After this lecture, you will be able to: Explain key properties of time series data Describe,
More informationLink Prediction. Eman Badr Mohammed Saquib Akmal Khan
Link Prediction Eman Badr Mohammed Saquib Akmal Khan 11-06-2013 Link Prediction Which pair of nodes should be connected? Applications Facebook friend suggestion Recommendation systems Monitoring and controlling
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Matrix Data: Clustering: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu November 3, 2015 Methods to Learn Matrix Data Text Data Set Data Sequence Data Time Series Graph
More informationBCT Lecture 3. Lukas Vacha.
BCT Lecture 3 Lukas Vacha vachal@utia.cas.cz Stationarity and Unit Root Testing Why do we need to test for Non-Stationarity? The stationarity or otherwise of a series can strongly influence its behaviour
More informationOn Bias, Variance, 0/1-Loss, and the Curse-of-Dimensionality. Weiqiang Dong
On Bias, Variance, 0/1-Loss, and the Curse-of-Dimensionality Weiqiang Dong 1 The goal of the work presented here is to illustrate that classification error responds to error in the target probability estimates
More informationTrending Models in the Data
April 13, 2009 Spurious regression I Before we proceed to test for unit root and trend-stationary models, we will examine the phenomena of spurious regression. The material in this lecture can be found
More informationMidterm: CS 6375 Spring 2018
Midterm: CS 6375 Spring 2018 The exam is closed book (1 cheat sheet allowed). Answer the questions in the spaces provided on the question sheets. If you run out of room for an answer, use an additional
More informationCS 340 Lec. 6: Linear Dimensionality Reduction
CS 340 Lec. 6: Linear Dimensionality Reduction AD January 2011 AD () January 2011 1 / 46 Linear Dimensionality Reduction Introduction & Motivation Brief Review of Linear Algebra Principal Component Analysis
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationEstimating trends using filters
Estimating trends using filters... contd. 3. Exponential smoothing of data to estimate the trend m[k] ˆm[k] = v[k]+(1 )ˆm[k 1], k =2,, n ˆm[1] = v[1] The choice of has to be fine tuned according to the
More informationEstimation and application of best ARIMA model for forecasting the uranium price.
Estimation and application of best ARIMA model for forecasting the uranium price. Medeu Amangeldi May 13, 2018 Capstone Project Superviser: Dongming Wei Second reader: Zhenisbek Assylbekov Abstract This
More information15.082J & 6.855J & ESD.78J. Algorithm Analysis
15.082J & 6.855J & ESD.78J Algorithm Analysis 15.082 Overview of subject Importance of Algorithm Analysis Importance of homework Midterms Moving forward 2 Overview of lecture Proof techniques Proof by
More informationCS540 ANSWER SHEET
CS540 ANSWER SHEET Name Email 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 1 2 Final Examination CS540-1: Introduction to Artificial Intelligence Fall 2016 20 questions, 5 points
More informationConsider the trend-cycle decomposition of a time series y t
1 Unit Root Tests Consider the trend-cycle decomposition of a time series y t y t = TD t + TS t + C t = TD t + Z t The basic issue in unit root testing is to determine if TS t = 0. Two classes of tests,
More informationat least 50 and preferably 100 observations should be available to build a proper model
III Box-Jenkins Methods 1. Pros and Cons of ARIMA Forecasting a) need for data at least 50 and preferably 100 observations should be available to build a proper model used most frequently for hourly or
More informationBoolean and Vector Space Retrieval Models
Boolean and Vector Space Retrieval Models Many slides in this section are adapted from Prof. Joydeep Ghosh (UT ECE) who in turn adapted them from Prof. Dik Lee (Univ. of Science and Tech, Hong Kong) 1
More informationNonlinear Dimensionality Reduction
Nonlinear Dimensionality Reduction Piyush Rai CS5350/6350: Machine Learning October 25, 2011 Recap: Linear Dimensionality Reduction Linear Dimensionality Reduction: Based on a linear projection of the
More informationSparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.
ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Previous lectures: Sparse vectors recap How to represent
More informationANLP Lecture 22 Lexical Semantics with Dense Vectors
ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Henry S. Thompson ANLP Lecture 22 5 November 2018 Previous
More informationClassification for High Dimensional Problems Using Bayesian Neural Networks and Dirichlet Diffusion Trees
Classification for High Dimensional Problems Using Bayesian Neural Networks and Dirichlet Diffusion Trees Rafdord M. Neal and Jianguo Zhang Presented by Jiwen Li Feb 2, 2006 Outline Bayesian view of feature
More informationMachine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.
Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted
More informationProf. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis
Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation
More informationPermutation-invariant regularization of large covariance matrices. Liza Levina
Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work
More informationModern Information Retrieval
Modern Information Retrieval Chapter 8 Text Classification Introduction A Characterization of Text Classification Unsupervised Algorithms Supervised Algorithms Feature Selection or Dimensionality Reduction
More information7. Integrated Processes
7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider
More informationMCMC analysis of classical time series algorithms.
MCMC analysis of classical time series algorithms. mbalawata@yahoo.com Lappeenranta University of Technology Lappeenranta, 19.03.2009 Outline Introduction 1 Introduction 2 3 Series generation Box-Jenkins
More informationIntroduction to Machine Learning Midterm Exam Solutions
10-701 Introduction to Machine Learning Midterm Exam Solutions Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes,
More informationCorrelation Preserving Unsupervised Discretization. Outline
Correlation Preserving Unsupervised Discretization Jee Vang Outline Paper References What is discretization? Motivation Principal Component Analysis (PCA) Association Mining Correlation Preserving Discretization
More informationConditional Random Fields and beyond DANIEL KHASHABI CS 546 UIUC, 2013
Conditional Random Fields and beyond DANIEL KHASHABI CS 546 UIUC, 2013 Outline Modeling Inference Training Applications Outline Modeling Problem definition Discriminative vs. Generative Chain CRF General
More informationStephen Scott.
1 / 35 (Adapted from Ethem Alpaydin and Tom Mitchell) sscott@cse.unl.edu In Homework 1, you are (supposedly) 1 Choosing a data set 2 Extracting a test set of size > 30 3 Building a tree on the training
More informationProximity data visualization with h-plots
The fifth international conference user! 2009 Proximity data visualization with h-plots Irene Epifanio Dpt. Matemàtiques, Univ. Jaume I (SPAIN) epifanio@uji.es; http://www3.uji.es/~epifanio Outline Motivating
More informationRiemann Manifold Methods in Bayesian Statistics
Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes
More informationA TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED
A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED by W. Robert Reed Department of Economics and Finance University of Canterbury, New Zealand Email: bob.reed@canterbury.ac.nz
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2010 Lecture 22: Nearest Neighbors, Kernels 4/18/2011 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!)
More information7. Integrated Processes
7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider
More informationMIRA, SVM, k-nn. Lirong Xia
MIRA, SVM, k-nn Lirong Xia Linear Classifiers (perceptrons) Inputs are feature values Each feature has a weight Sum is the activation activation w If the activation is: Positive: output +1 Negative, output
More informationRandomized Algorithms
Randomized Algorithms Saniv Kumar, Google Research, NY EECS-6898, Columbia University - Fall, 010 Saniv Kumar 9/13/010 EECS6898 Large Scale Machine Learning 1 Curse of Dimensionality Gaussian Mixture Models
More informationCLUe Training An Introduction to Machine Learning in R with an example from handwritten digit recognition
CLUe Training An Introduction to Machine Learning in R with an example from handwritten digit recognition Ad Feelders Universiteit Utrecht Department of Information and Computing Sciences Algorithmic Data
More informationMIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,
MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run
More informationSOME BASICS OF TIME-SERIES ANALYSIS
SOME BASICS OF TIME-SERIES ANALYSIS John E. Floyd University of Toronto December 8, 26 An excellent place to learn about time series analysis is from Walter Enders textbook. For a basic understanding of
More informationTesting for non-stationarity
20 November, 2009 Overview The tests for investigating the non-stationary of a time series falls into four types: 1 Check the null that there is a unit root against stationarity. Within these, there are
More informationClassic Time Series Analysis
Classic Time Series Analysis Concepts and Definitions Let Y be a random number with PDF f Y t ~f,t Define t =E[Y t ] m(t) is known as the trend Define the autocovariance t, s =COV [Y t,y s ] =E[ Y t t
More informationRegularization. CSCE 970 Lecture 3: Regularization. Stephen Scott and Vinod Variyam. Introduction. Outline
Other Measures 1 / 52 sscott@cse.unl.edu learning can generally be distilled to an optimization problem Choose a classifier (function, hypothesis) from a set of functions that minimizes an objective function
More informationForecasting using R. Rob J Hyndman. 2.4 Non-seasonal ARIMA models. Forecasting using R 1
Forecasting using R Rob J Hyndman 2.4 Non-seasonal ARIMA models Forecasting using R 1 Outline 1 Autoregressive models 2 Moving average models 3 Non-seasonal ARIMA models 4 Partial autocorrelations 5 Estimation
More informationFORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL
FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL B. N. MANDAL Abstract: Yearly sugarcane production data for the period of - to - of India were analyzed by time-series methods. Autocorrelation
More informationMetric-based classifiers. Nuno Vasconcelos UCSD
Metric-based classifiers Nuno Vasconcelos UCSD Statistical learning goal: given a function f. y f and a collection of eample data-points, learn what the function f. is. this is called training. two major
More information22/04/2014. Economic Research
22/04/2014 Economic Research Forecasting Models for Exchange Rate Tuesday, April 22, 2014 The science of prognostics has been going through a rapid and fruitful development in the past decades, with various
More informationPROBABILISTIC LATENT SEMANTIC ANALYSIS
PROBABILISTIC LATENT SEMANTIC ANALYSIS Lingjia Deng Revised from slides of Shuguang Wang Outline Review of previous notes PCA/SVD HITS Latent Semantic Analysis Probabilistic Latent Semantic Analysis Applications
More informationTime Series Outlier Detection
Time Series Outlier Detection Tingyi Zhu July 28, 2016 Tingyi Zhu Time Series Outlier Detection July 28, 2016 1 / 42 Outline Time Series Basics Outliers Detection in Single Time Series Outlier Series Detection
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationWeighted Least Squares
Weighted Least Squares ST 430/514 Recall the linear regression equation E(Y ) = β 0 + β 1 x 1 + β 2 x 2 + + β k x k We have estimated the parameters β 0, β 1, β 2,..., β k by minimizing the sum of squared
More informationFinal Examination 7/6/2011
The Islamic University of Gaza Faculty of Commerce Department of Economics & Applied Statistics Time Series Analysis - Dr. Samir Safi Spring Semester 211 Final Examination 7/6/211 Name: ID: INSTRUCTIONS:
More informationRETRIEVAL MODELS. Dr. Gjergji Kasneci Introduction to Information Retrieval WS
RETRIEVAL MODELS Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Retrieval models Boolean model Vector space model Probabilistic
More informationCS249: ADVANCED DATA MINING
CS249: ADVANCED DATA MINING Clustering Evaluation and Practical Issues Instructor: Yizhou Sun yzsun@cs.ucla.edu May 2, 2017 Announcements Homework 2 due later today Due May 3 rd (11:59pm) Course project
More informationBayes Decision Theory - I
Bayes Decision Theory - I Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statistical Learning from Data Goal: Given a relationship between a feature vector and a vector y, and iid data samples ( i,y i ), find
More informationAlternative Parameterizations of Markov Networks. Sargur Srihari
Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models with Energy functions
More informationArtificial Neural Networks Examination, June 2004
Artificial Neural Networks Examination, June 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Chapter 9 Multivariate time series 2 Transfer function
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Matrix Data: Clustering: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu October 19, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network
More informationLecture 10: Dimension Reduction Techniques
Lecture 10: Dimension Reduction Techniques Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 17, 2018 Input Data It is assumed that there is a set
More informationDimension Reduction Techniques. Presented by Jie (Jerry) Yu
Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage
More informationChapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis
Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive
More informationNANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION MAS451/MTH451 Time Series Analysis TIME ALLOWED: 2 HOURS
NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION 2012-2013 MAS451/MTH451 Time Series Analysis May 2013 TIME ALLOWED: 2 HOURS INSTRUCTIONS TO CANDIDATES 1. This examination paper contains FOUR (4)
More informationCentrality Measures. Leonid E. Zhukov
Centrality Measures Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Network Science Leonid E.
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationPart-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287
Part-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287 Review: Neural Networks One-layer multi-layer perceptron architecture, NN MLP1 (x) = g(xw 1 + b 1 )W 2 + b 2 xw + b; perceptron x is the
More informationAdvanced Econometrics
Advanced Econometrics Marco Sunder Nov 04 2010 Marco Sunder Advanced Econometrics 1/ 25 Contents 1 2 3 Marco Sunder Advanced Econometrics 2/ 25 Music Marco Sunder Advanced Econometrics 3/ 25 Music Marco
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationarxiv: v1 [physics.soc-ph] 17 Feb 2013
Identification of Literary Movements Using Complex Networks Represent Texts arxiv:1302.4099v1 [physics.soc-ph] 17 Feb 2013 Diego Raphael Amancio 1, Osvaldo N. Oliveira Jr. 1, Luciano da Fonura Costa 1
More informationNeural Networks DWML, /25
DWML, 2007 /25 Neural networks: Biological and artificial Consider humans: Neuron switching time 0.00 second Number of neurons 0 0 Connections per neuron 0 4-0 5 Scene recognition time 0. sec 00 inference
More informationChapter 2: Unit Roots
Chapter 2: Unit Roots 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and undeconometrics II. Unit Roots... 3 II.1 Integration Level... 3 II.2 Nonstationarity
More informationStochastic Processes
Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.
More informationTable of Contents. Multivariate methods. Introduction II. Introduction I
Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation
More informationAnnouncements. CS 188: Artificial Intelligence Spring Classification. Today. Classification overview. Case-Based Reasoning
CS 188: Artificial Intelligence Spring 21 Lecture 22: Nearest Neighbors, Kernels 4/18/211 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!) Remaining
More informationA Test of Cointegration Rank Based Title Component Analysis.
A Test of Cointegration Rank Based Title Component Analysis Author(s) Chigira, Hiroaki Citation Issue 2006-01 Date Type Technical Report Text Version publisher URL http://hdl.handle.net/10086/13683 Right
More informationComputer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)
Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori
More informationBANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1
BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR Data Mining I Lecture
More informationNonlinear Dimensionality Reduction
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Kernel PCA 2 Isomap 3 Locally Linear Embedding 4 Laplacian Eigenmap
More information