Paragraph Topic Classification

Size: px
Start display at page:

Download "Paragraph Topic Classification"

Transcription

1 Paragraph Topic Cassification Eugene Nho Graduate Schoo of Business Stanford University Stanford, CA Edward Ng Department of Eectrica Engineering Stanford University Stanford, CA Introduction The project attempts to predict the topics of paragraph-ength texts. We posit that machine earning agorithms can extend a person s capabiity to consume information through the automated categorization of text and through an increased capabiity to ocate reevant information. We formay define the probem as constructing a mode that, given an input of text, outputs a vector of reevant topics. The project uses approximatey Wikipedia artice summaries categorized by Wikipedia users a priori. A number of approaches were taken incuding Naive Bayes, One-vs-Rest Cassifier (OvR with GoVe Vectors, Latent Dirichet Aocation (LDA/OvR, GoVe Vectors/LDA/OvR, Convoution Neura Networks (CNN, and Long Short Term Memory networks (LSTM. 2 Reated Work We found a number of papers associated with topic cassification, which we wi segment into traditiona, topic modeing, and finay the use of neura networks approaches to topic cassification. 2.1 Traditiona Approaches The use of Naive Bayes in topic cassification has been preeminent given its dua properties of being accurate and quick to train (Ting et. a (2011. However, there are variations of Naive Bayes that woud have ikey improved our performance, specificay an agorithm deveoped by Wang et. a (2012 which used Naive Bayes og-count ratios as features into an SVM mode (known as NB-SVM, and using bigrams rather than tf-idf. This combined mode (with a sma modification, known as NB-LM is considered to be one of the state-of-the-art supervised earning agorithms for topic cassification. Further research indicated the use of bigrams (and n-grams in genera does not revea any improved performance (Tan et. a, 2002, eaving the impementation of NB-LM as the primary focus on any future impementation using traditiona earning agorithms. 2.2 Topic Modeing Another approach to topic cassification is to presume that for a given dataset of corpora, there are n number of topics distributed, and that for each given document there is some probabiity that they beong to a given topic. We impement Latent Dirichet Aocation

2 2.3 Neura Networks The recent insurgent interest of neura networks have aso pushed the appication of neura networks to incude topic cassification. Convoution neura networks, initiay targeted for image cassification appications, have been modified to aso perform natura anguage processing with a high rate of success and considered to be state-of-the-art for topic cassification. We impement a rudimentary convoution neura network based on Kim et. a (2014 (with the sight modification of using GoVe vectors rather than word2vec which provided the highest rate of accuracy out of a our agorithms. However, there are a coupe of variations that may have improved our performance on this task, incuding the use of individua characters as inputs (Zhang et. a (2015 and training embeddings through the use parae convoution ayers with differing region sizes (Johnson et. a (2014. The increased compexity of neura network topoogy is often associated with the need for arger training data sets, with Zhang et. a (2015 expicity mentioning that a dataset of severa miions needs to be provided for character-eve convoutiona neura nets to outperform more traditiona approaches. In order for these other topoogies to function we, our dataset of Wikipedia artices woud have been required to be much arger. However, using parae convoution ayers may have been an area of future exporation. 3 Dataset We used Wikipedia artices summaries as our data. Each artice s first summary paragraph served as our input, and its topic assignments as our abes. We scraped Wikipedia for tota 400, 000 inputs, a beonging to one or more of the foowing five topics: mathematics, poitics, computer science, fim, and music. We imited our scope of topics to five to test the efficacy of our modes without requiring an unwiedy amount of data, and thus computationa resources. Our training and test sets were and entries respectivey, with an average word count of 93 for each summary. We used Wikipedia because it is one of the few avaiabe sources of topicay abeed, mutiabe dataset. We insisted on soving a mutiabe probem because, in rea-word appications, a piece of text is rarey about one topic ony. An exampe entry is: In mathematics, the minimum k-cut, is a combinatoria optimization probem that requires finding a set of edges whose remova woud partition the graph to k connected components. These edges are referred to as k-cut. The goa is to find the minimum-weight k-cut. This partitioning can have appications in VLSI design, data-mining, finite eements and communication in parae computing. 4 Features We used term frequency-inverse document frequency (tf-idf and GoVe as our main features. 4.1 Term Frequency-Inverse Document Frequency We used tf-idf because it captures how important a word is to a document in a corpus by cacuating the frequency of a word in a given document offset by its frequency in the overa corpus, de-emphasizing common words ike is and the. It is aso a common baseine feature for information retrieva tasks. The weighting scheme used for tf-idf is the defaut provided by skearn. 4.2 Goba Vectors for Word Representation (GoVe Later in the project, we used pre-trained GoVe to feed richer representation of input text into our modes. GoVe represents words as vectors, with the vector representing the co-occurrence of the word with other words over a goba set. The cost function which GoVe minimizes is J(θ = 1 W f(p ij(u T i v j og P ij 2 2 i,j=1 where P ij is the probabiity of word i occurring in the context of j, W represents the entire vocabuary, f provides is a given weighting function, and u and v are the vector representations of i and j respectivey. 2

3 Our hypothesis for using GoVe vectors was that they woud more effectivey capture the underying reationships between words characterizing the same topic because reated words are ikey represented by simiar vectors. We used pre-trained GoVe vectors from Stanford s NLP group 1 (trained on Wikipedia 2014 corpus, 100-dimensions. 5 Methods Our overa approach to text cassification was starting with a simpe baseine mode and graduay adding more sophisticated features or mode architecture to see if and how the resuts improve. Tabe 1 summarizes our approaches and rationae. Tabe 1: Summary of approaches and rationae Mode Approach [features] Naive Bayes [tf-idf] OvR [GoVe] LDA + OvR [tf] LDA + GoVE + OvR [tf] CNN [GoVe] Rationae Common baseine mode for text cassification One-vs-Rest supports mutiabe earning; richer feature (GoVe To capture atent topics more effectivey Compementary behavior (word embedding + topic distribution Fast; generaizes we; oca word order is not important 5.1 Baseine: Naive Bayes We used Naive Bayes for our baseine impementation. Naive Bayes is often a go-to strategy for baseine given its speed and ease of its training. We trained Naive Bayes from skearn for muticass topic cassification, where paragraphs that fit into mutipe topics were cassified to be under a singe compound topic. Since we imit ourseves to 5 topics, there are 2 5 = 32 possibe abes to choose from. Input feature was a tf-idf matrix. 5.2 GoVe + One-vs-Rest SVM cassifier To test our hypothesis that feeding richer representation of our input text woud improve the resuts, we used GoVe as our feature and fed it into a inear Support Vector Machine (SVM cassifier using One-vs-Rest strategy (OvR. We chose OvR because it enabes us to do mutiabe cassification by constructing k SVMs, where k is the number of casses. Given m training data (x (1, y (1,..., (x (m, y (m, where x (i R n and y (i R k indicates casses x (i beongs to, the th SVM soves the foowing probem: 1 min w, b,ξ 2 wt w + c m i=1 ξ (i s.t. w T x (i + b 1 ξ (i, if y (i = 1 w T x (i + b 1 + ξ (i, if y (i = 0 where ξ i 0 for a i and c is a penaty parameter. In our impementation, x (i was constructed by first vectorizing each word of the ith exampe into a R 100 vector using the pre-trained GoVe weights, and then coapsing the matrix into a singe vector using tf-idf as the weighting factor. For a new exampe x, OvR conducts the foowing for each : y = sgn( w T x + b 5.3 LDA + One-vs-Rest SVM cassifier LDA is a 3-eve hierarchica Bayesian mode often used for topic modeing in natura anguage processing. Fuy expaining the mode is outside the scope of this paper, but one can get an intuition by observing how a singe document is modeed: ( N p(w α, β = p(θ α p(z n θp(w n z n, β dθ n=1 z n where α and β are parameters, θ~dirichet(α is a random variabe representing topic mixture, w is the document, z n is a topic for n th word of the document, and w n is n t h word. Intuition is that the term inside parenthesis, representing each document, is modeed as p(z zn p(w dn z dn. In other words, each document is modeed as random mixtures over atent topics, where each topic is characterized by a distribution over words. Our hypothesis for choosing LDA was that modeing 1 3

4 documents this way woud hep us addresses head-on a core chaenge we were facing: How can we cassify documents based on underying themes, rather than presence or absence of certain key words?" Since LDA is an unsupervised earning mode, we used its output, document topic matrix (number of inputs number of topics, as our design matrix X that gets fed into the OvR cassifier. 5.4 Convoutiona Neura Network (CNN Convoutiona Neura Networks (CNN appy ayers of convoving fiters to features. Traditionay CNNs have been to computer vision, but recenty they are increasingy appied to natura anguage processing probems. Instead of appying fiters to image pixes, CNNs in NLP appy the fiters to the matrices representing paragraphs, typicay consisting of rows of word embeddings such as word2vec or GoVe. We wi not define CNNs in mathematica formaism in this paper, but the essence of CNNs can be seen at the convoution ayers. Given the document matrix x 1 : n R n k, where x i s (word embedding vectors ike GoVe for a n words are stacked together, convoution occurs when a fiter w R hk is appied to a window of h words to produce a new feature. A feature c i is generated from window of words x i : i + h 1 by c i = f(w x i:i+h 1 + b where b R is a bias term and f is a non-inear activation function. This fiter is appied to each possibe window of words in the paragraph, producing n h + 1 of c i s above. We chose CNN for two main reasons. First, we beieve CNNs do what LDA does capturing atent topics but in a much more powerfu way. By appying convoving fiters and abstracting away from the raw word embeddings, CNNs can capture the "shapes of topics" the same way CNNs appied to image recognition capture the shapes of objects. Second, CNNs are fast and reativey easy to train with imited GPU resources, which enabed us to experiment with various hyperparameter settings without extensive infrastructure setup. The mode architecture is a muti-abe variant of the CNN architecture proposed in Kim et. a (2014, using a static embedding ayer, and an extra fuy-connected ayer. In order to do muti-abe cassification we used a binary cross entropy oss function and sigmoid activation. Figure 1: Graph of CNN Mode Architecture 5.5 Long Short Term Memory (LSTM LSTM is a Recurrent Neura Network (RNN architecture that is we-suited to earn from sequentia experience when ong time ags of unknown size exist between important events. It accompishes this by having gates: computing the foowing i t = σ (W (i x t + U (i h t 1, f t = σ (W (f x t + U (f h t 1, o t = σ (W (o x t + U (o h t 1 c t = tanh (W (c x t + U (c h t 1, c t = f t c t 1 + i t c+ t, h t = o t tanh(c t Intuition is that at each time t, given input x t and hidden state from t 1, LSTM ce has the power to decide whether to input, forget, or output the data based on function σ, thereby enabing the overa network to earn when "important events" may be cose or far apart in a given sequence of data. We tried LSTM mainy as a benchmark against our CNN resuts. RNN architecture is widey viewed as the go-to strategy for natura anguage processing (NLP tasks because anguage is a sequentia type of data. Our mode architecture is defined by an embedding ayer, a singe ayer of 200 LSTM ces, and a feed-forward ayer using sigmoid activation. As with CNN, it is muti-abe variant using a binary cross entropy oss function. 4

5 6 Resuts and Discussion The overa resuts for our various approaches at paragraph topic cassification are summarized in Figure 2 beow. There are a few points worth noting about these resuts. Figure 2: Metrics for a Approaches First, Naive Bayes was very robust despite its simpicity, and given its preeminence in topic cassification, this is not at a surprising. Second, the GoVe and One-vs-Rest mode performed very poory on accuracy. We beieve this has to do with the fact that we coapsed the GoVe document matrix into a vector to feed into the cassifier. We suspect they began to converge toward simiar, overapping pattern the more data we used. This coud be because GoVe synthesizes the reationship between words with vector distances, and with more data, different types of differences a get refected with the simiar distances. As a data point, GoVe + OvR achieved 40% accuracy when trained over ony sampes. Third, contrary to the initia hypothesis, LDA did not work we in topic prediction. We beieve the principa reason is the difficuty to contro the way an unsupervised agorithm earns. The output of LDA we used for cassification document topic matrix is LDA s best guess about the ikeihood of a given sampe beonging to each of the five custers it earned to differentiate between. A corpus with m documents coud be custered into five groups in 5 m different ways, and it is neary impossibe to exacty aign the way an unsupervised earning agorithm custers them with the way humans did a priori. Fourth, both neura networks far outperformed other approaches with 97% accuracy. It is noteworthy that CNN and LSTM both ended up with very simiar metrics, but CNN was much faster and robust. CNN s training time for sampes was on average 30 minutes on a aptop CPU, whereas LSTM took 3 hours for haf the number of sampes. Lasty, as the next section on hyperparameter tuning wi discuss, CNN was very robust, showing consistenty effective resuts across various hyperparameter settings. In contrast, LSTM s resuts varied widey depending on hyperparameters Hyperparameter Tuning Our CNN mode proved reativey stabe over various hyperparameter settings. For exampe, deeper networks performed ony marginay better, but at much higher train and test time cost. Going from 1 ayer to 3, we saw an increase of 0.5 percentage points for both reca and precision. Changing dropout ayers, number of fuy-connected ayers, number and size of fiters, and activation function between ReLU and tanh did not yied more than 5% difference in accuracy. On the other hand, LSTM was very sensitive to the number of epochs, going from 40% to 86% as we increased the epochs from 1 to 5. The resuts aso varied depending on whether we aowed embeddings to be continuousy updated ( 10% higher when aowed, as we as the number of LSTM ces per ayer (80% for 100 ces vs 86% 200 ces. In the end, however, LSTM performed marginay better than CNN. 7 Future Work We woud ike to further perform hyperparamter tuning of the modes impemented, perhaps with grid search. In addition, given additiona time, we woud have ideay expored the weights assigned to our CNN and LSTM mode to make sense of how a given corpus is weighted. We woud aso re-estabish our baseine using the NB-SVM mode proposed by Wang et. a (

6 References [1] Ting, S. L., Ip, W. H., Tsang, A. H. (2011. Is Naive Bayes a good cassifier for document cassification?. Internationa Journa of Software Engineering and Its Appications, 5(3, [2] Yoo, J. Y., Yang, D. (2015. Cassification Scheme of Unstructured Text Document using TF-IDF and Naive Bayes Cassifier. [3] Wang, S., Manning, C. D. (2012, Juy. Baseines and bigrams: Simpe, good sentiment and topic cassification. In Proceedings of the 50th Annua Meeting of the Association for Computationa Linguistics: Short Papers-Voume 2 (pp Association for Computationa Linguistics. [4] Bei, D. M., Ng, A. Y., Jordan, M. I. (2003. Latent dirichet aocation. Journa of machine Learning research, 3(Jan, [5] Ramage, D., Ha, D., Naapati, R., Manning, C. D. (2009, August. Labeed LDA: A supervised topic mode for credit attribution in muti-abeed corpora. In Proceedings of the 2009 Conference on Empirica Methods in Natura Language Processing: Voume 1-Voume 1 (pp Association for Computationa Linguistics. [6] Tan, C. M., Wang, Y. F., Lee, C. D. (2002. The use of bigrams to enhance text categorization. Information processing management, 38(4, [7] Kim, Y. (2014. Convoutiona neura networks for sentence cassification. arxiv preprint arxiv: [8] Zhang, X., Zhao, J., LeCun, Y. (2015. Character-eve convoutiona networks for text cassification. In Advances in Neura Information Processing Systems (pp [9] Johnson, R., Zhang, T. (2014. Effective use of word order for text categorization with convoutiona neura networks. arxiv preprint arxiv:

Convolutional Networks 2: Training, deep convolutional networks

Convolutional Networks 2: Training, deep convolutional networks Convoutiona Networks 2: Training, deep convoutiona networks Hakan Bien Machine Learning Practica MLP Lecture 8 30 October / 6 November 2018 MLP Lecture 8 / 30 October / 6 November 2018 Convoutiona Networks

More information

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with? Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine

More information

Stochastic Variational Inference with Gradient Linearization

Stochastic Variational Inference with Gradient Linearization Stochastic Variationa Inference with Gradient Linearization Suppementa Materia Tobias Pötz * Anne S Wannenwetsch Stefan Roth Department of Computer Science, TU Darmstadt Preface In this suppementa materia,

More information

A Brief Introduction to Markov Chains and Hidden Markov Models

A Brief Introduction to Markov Chains and Hidden Markov Models A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,

More information

BP neural network-based sports performance prediction model applied research

BP neural network-based sports performance prediction model applied research Avaiabe onine www.jocpr.com Journa of Chemica and Pharmaceutica Research, 204, 6(7:93-936 Research Artice ISSN : 0975-7384 CODEN(USA : JCPRC5 BP neura networ-based sports performance prediction mode appied

More information

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com

More information

Statistical Learning Theory: A Primer

Statistical Learning Theory: A Primer Internationa Journa of Computer Vision 38(), 9 3, 2000 c 2000 uwer Academic Pubishers. Manufactured in The Netherands. Statistica Learning Theory: A Primer THEODOROS EVGENIOU, MASSIMILIANO PONTIL AND TOMASO

More information

Inductive Bias: How to generalize on novel data. CS Inductive Bias 1

Inductive Bias: How to generalize on novel data. CS Inductive Bias 1 Inductive Bias: How to generaize on nove data CS 478 - Inductive Bias 1 Overfitting Noise vs. Exceptions CS 478 - Inductive Bias 2 Non-Linear Tasks Linear Regression wi not generaize we to the task beow

More information

Two view learning: SVM-2K, Theory and Practice

Two view learning: SVM-2K, Theory and Practice Two view earning: SVM-2K, Theory and Practice Jason D.R. Farquhar jdrf99r@ecs.soton.ac.uk Hongying Meng hongying@cs.york.ac.uk David R. Hardoon drh@ecs.soton.ac.uk John Shawe-Tayor jst@ecs.soton.ac.uk

More information

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones ASummaryofGaussianProcesses Coryn A.L. Baier-Jones Cavendish Laboratory University of Cambridge caj@mrao.cam.ac.uk Introduction A genera prediction probem can be posed as foows. We consider that the variabe

More information

Optimality of Inference in Hierarchical Coding for Distributed Object-Based Representations

Optimality of Inference in Hierarchical Coding for Distributed Object-Based Representations Optimaity of Inference in Hierarchica Coding for Distributed Object-Based Representations Simon Brodeur, Jean Rouat NECOTIS, Département génie éectrique et génie informatique, Université de Sherbrooke,

More information

II. PROBLEM. A. Description. For the space of audio signals

II. PROBLEM. A. Description. For the space of audio signals CS229 - Fina Report Speech Recording based Language Recognition (Natura Language) Leopod Cambier - cambier; Matan Leibovich - matane; Cindy Orozco Bohorquez - orozcocc ABSTRACT We construct a rea time

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view

More information

Multilayer Kerceptron

Multilayer Kerceptron Mutiayer Kerceptron Zotán Szabó, András Lőrincz Department of Information Systems, Facuty of Informatics Eötvös Loránd University Pázmány Péter sétány 1/C H-1117, Budapest, Hungary e-mai: szzoi@csetehu,

More information

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After

More information

From Margins to Probabilities in Multiclass Learning Problems

From Margins to Probabilities in Multiclass Learning Problems From Margins to Probabiities in Muticass Learning Probems Andrea Passerini and Massimiiano Ponti 2 and Paoo Frasconi 3 Abstract. We study the probem of muticass cassification within the framework of error

More information

A Novel Learning Method for Elman Neural Network Using Local Search

A Novel Learning Method for Elman Neural Network Using Local Search Neura Information Processing Letters and Reviews Vo. 11, No. 8, August 2007 LETTER A Nove Learning Method for Eman Neura Networ Using Loca Search Facuty of Engineering, Toyama University, Gofuu 3190 Toyama

More information

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com

More information

Supervised i-vector Modeling - Theory and Applications

Supervised i-vector Modeling - Theory and Applications Supervised i-vector Modeing - Theory and Appications Shreyas Ramoji, Sriram Ganapathy Learning and Extraction of Acoustic Patterns LEAP) Lab, Eectrica Engineering, Indian Institute of Science, Bengauru,

More information

STA 216 Project: Spline Approach to Discrete Survival Analysis

STA 216 Project: Spline Approach to Discrete Survival Analysis : Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing

More information

An Algorithm for Pruning Redundant Modules in Min-Max Modular Network

An Algorithm for Pruning Redundant Modules in Min-Max Modular Network An Agorithm for Pruning Redundant Modues in Min-Max Moduar Network Hui-Cheng Lian and Bao-Liang Lu Department of Computer Science and Engineering, Shanghai Jiao Tong University 1954 Hua Shan Rd., Shanghai

More information

Neural Networks Compression for Language Modeling

Neural Networks Compression for Language Modeling Neura Networks Compression for Language Modeing Artem M. Grachev 1,2, Dmitry I. Ignatov 2, and Andrey V. Savchenko 3 arxiv:1708.05963v1 [stat.ml] 20 Aug 2017 1 Samsung R&D Institute Rus, Moscow, Russia

More information

Appendix A: MATLAB commands for neural networks

Appendix A: MATLAB commands for neural networks Appendix A: MATLAB commands for neura networks 132 Appendix A: MATLAB commands for neura networks p=importdata('pn.xs'); t=importdata('tn.xs'); [pn,meanp,stdp,tn,meant,stdt]=prestd(p,t); for m=1:10 net=newff(minmax(pn),[m,1],{'tansig','purein'},'trainm');

More information

Active Learning & Experimental Design

Active Learning & Experimental Design Active Learning & Experimenta Design Danie Ting Heaviy modified, of course, by Lye Ungar Origina Sides by Barbara Engehardt and Aex Shyr Lye Ungar, University of Pennsyvania Motivation u Data coection

More information

The EM Algorithm applied to determining new limit points of Mahler measures

The EM Algorithm applied to determining new limit points of Mahler measures Contro and Cybernetics vo. 39 (2010) No. 4 The EM Agorithm appied to determining new imit points of Maher measures by Souad E Otmani, Georges Rhin and Jean-Marc Sac-Épée Université Pau Veraine-Metz, LMAM,

More information

A Better Way to Pretrain Deep Boltzmann Machines

A Better Way to Pretrain Deep Boltzmann Machines A Better Way to Pretrain Deep Botzmann Machines Rusan Saakhutdino Department of Statistics and Computer Science Uniersity of Toronto rsaakhu@cs.toronto.edu Geoffrey Hinton Department of Computer Science

More information

Turbo Codes. Coding and Communication Laboratory. Dept. of Electrical Engineering, National Chung Hsing University

Turbo Codes. Coding and Communication Laboratory. Dept. of Electrical Engineering, National Chung Hsing University Turbo Codes Coding and Communication Laboratory Dept. of Eectrica Engineering, Nationa Chung Hsing University Turbo codes 1 Chapter 12: Turbo Codes 1. Introduction 2. Turbo code encoder 3. Design of intereaver

More information

A Comparison Study of the Test for Right Censored and Grouped Data

A Comparison Study of the Test for Right Censored and Grouped Data Communications for Statistica Appications and Methods 2015, Vo. 22, No. 4, 313 320 DOI: http://dx.doi.org/10.5351/csam.2015.22.4.313 Print ISSN 2287-7843 / Onine ISSN 2383-4757 A Comparison Study of the

More information

Combining reaction kinetics to the multi-phase Gibbs energy calculation

Combining reaction kinetics to the multi-phase Gibbs energy calculation 7 th European Symposium on Computer Aided Process Engineering ESCAPE7 V. Pesu and P.S. Agachi (Editors) 2007 Esevier B.V. A rights reserved. Combining reaction inetics to the muti-phase Gibbs energy cacuation

More information

SVM: Terminology 1(6) SVM: Terminology 2(6)

SVM: Terminology 1(6) SVM: Terminology 2(6) Andrew Kusiak Inteigent Systems Laboratory 39 Seamans Center he University of Iowa Iowa City, IA 54-57 SVM he maxima margin cassifier is simiar to the perceptron: It aso assumes that the data points are

More information

SVM-based Supervised and Unsupervised Classification Schemes

SVM-based Supervised and Unsupervised Classification Schemes SVM-based Supervised and Unsupervised Cassification Schemes LUMINITA STATE University of Pitesti Facuty of Mathematics and Computer Science 1 Targu din Vae St., Pitesti 110040 ROMANIA state@cicknet.ro

More information

Moreau-Yosida Regularization for Grouped Tree Structure Learning

Moreau-Yosida Regularization for Grouped Tree Structure Learning Moreau-Yosida Reguarization for Grouped Tree Structure Learning Jun Liu Computer Science and Engineering Arizona State University J.Liu@asu.edu Jieping Ye Computer Science and Engineering Arizona State

More information

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA) 1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using

More information

Support Vector Machine and Its Application to Regression and Classification

Support Vector Machine and Its Application to Regression and Classification BearWorks Institutiona Repository MSU Graduate Theses Spring 2017 Support Vector Machine and Its Appication to Regression and Cassification Xiaotong Hu As with any inteectua project, the content and views

More information

Distributed average consensus: Beyond the realm of linearity

Distributed average consensus: Beyond the realm of linearity Distributed average consensus: Beyond the ream of inearity Usman A. Khan, Soummya Kar, and José M. F. Moura Department of Eectrica and Computer Engineering Carnegie Meon University 5 Forbes Ave, Pittsburgh,

More information

Chemical Kinetics Part 2

Chemical Kinetics Part 2 Integrated Rate Laws Chemica Kinetics Part 2 The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates the rate

More information

Statistical Learning Theory: a Primer

Statistical Learning Theory: a Primer ??,??, 1 6 (??) c?? Kuwer Academic Pubishers, Boston. Manufactured in The Netherands. Statistica Learning Theory: a Primer THEODOROS EVGENIOU AND MASSIMILIANO PONTIL Center for Bioogica and Computationa

More information

Approach to Identifying Raindrop Vibration Signal Detected by Optical Fiber

Approach to Identifying Raindrop Vibration Signal Detected by Optical Fiber Sensors & Transducers, o. 6, Issue, December 3, pp. 85-9 Sensors & Transducers 3 by IFSA http://www.sensorsporta.com Approach to Identifying Raindrop ibration Signa Detected by Optica Fiber ongquan QU,

More information

Unconditional security of differential phase shift quantum key distribution

Unconditional security of differential phase shift quantum key distribution Unconditiona security of differentia phase shift quantum key distribution Kai Wen, Yoshihisa Yamamoto Ginzton Lab and Dept of Eectrica Engineering Stanford University Basic idea of DPS-QKD Protoco. Aice

More information

BDD-Based Analysis of Gapped q-gram Filters

BDD-Based Analysis of Gapped q-gram Filters BDD-Based Anaysis of Gapped q-gram Fiters Marc Fontaine, Stefan Burkhardt 2 and Juha Kärkkäinen 2 Max-Panck-Institut für Informatik Stuhsatzenhausweg 85, 6623 Saarbrücken, Germany e-mai: stburk@mpi-sb.mpg.de

More information

Scalable Spectrum Allocation for Large Networks Based on Sparse Optimization

Scalable Spectrum Allocation for Large Networks Based on Sparse Optimization Scaabe Spectrum ocation for Large Networks ased on Sparse Optimization innan Zhuang Modem R&D Lab Samsung Semiconductor, Inc. San Diego, C Dongning Guo, Ermin Wei, and Michae L. Honig Department of Eectrica

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Schoo of Computer Science Probabiistic Graphica Modes Posterior Reguarization: an integrative paradigm for earning GMs p Eric Xing (courtesy to Jun Zhu) Lecture 29, Apri 30, 2014 Reading: 1 Learning GMs

More information

Discrete Techniques. Chapter Introduction

Discrete Techniques. Chapter Introduction Chapter 3 Discrete Techniques 3. Introduction In the previous two chapters we introduced Fourier transforms of continuous functions of the periodic and non-periodic (finite energy) type, we as various

More information

Ant Colony Algorithms for Constructing Bayesian Multi-net Classifiers

Ant Colony Algorithms for Constructing Bayesian Multi-net Classifiers Ant Coony Agorithms for Constructing Bayesian Muti-net Cassifiers Khaid M. Saama and Aex A. Freitas Schoo of Computing, University of Kent, Canterbury, UK. {kms39,a.a.freitas}@kent.ac.uk December 5, 2013

More information

Discrete Techniques. Chapter Introduction

Discrete Techniques. Chapter Introduction Chapter 3 Discrete Techniques 3. Introduction In the previous two chapters we introduced Fourier transforms of continuous functions of the periodic and non-periodic (finite energy) type, as we as various

More information

Power Control and Transmission Scheduling for Network Utility Maximization in Wireless Networks

Power Control and Transmission Scheduling for Network Utility Maximization in Wireless Networks ower Contro and Transmission Scheduing for Network Utiity Maximization in Wireess Networks Min Cao, Vivek Raghunathan, Stephen Hany, Vinod Sharma and. R. Kumar Abstract We consider a joint power contro

More information

Research of Data Fusion Method of Multi-Sensor Based on Correlation Coefficient of Confidence Distance

Research of Data Fusion Method of Multi-Sensor Based on Correlation Coefficient of Confidence Distance Send Orders for Reprints to reprints@benthamscience.ae 340 The Open Cybernetics & Systemics Journa, 015, 9, 340-344 Open Access Research of Data Fusion Method of Muti-Sensor Based on Correation Coefficient

More information

Nonlinear Gaussian Filtering via Radial Basis Function Approximation

Nonlinear Gaussian Filtering via Radial Basis Function Approximation 51st IEEE Conference on Decision and Contro December 10-13 01 Maui Hawaii USA Noninear Gaussian Fitering via Radia Basis Function Approximation Huazhen Fang Jia Wang and Raymond A de Caafon Abstract This

More information

Partial permutation decoding for MacDonald codes

Partial permutation decoding for MacDonald codes Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics

More information

The Group Structure on a Smooth Tropical Cubic

The Group Structure on a Smooth Tropical Cubic The Group Structure on a Smooth Tropica Cubic Ethan Lake Apri 20, 2015 Abstract Just as in in cassica agebraic geometry, it is possibe to define a group aw on a smooth tropica cubic curve. In this note,

More information

Efficiently Generating Random Bits from Finite State Markov Chains

Efficiently Generating Random Bits from Finite State Markov Chains 1 Efficienty Generating Random Bits from Finite State Markov Chains Hongchao Zhou and Jehoshua Bruck, Feow, IEEE Abstract The probem of random number generation from an uncorreated random source (of unknown

More information

Adaptive Noise Cancellation Using Deep Cerebellar Model Articulation Controller

Adaptive Noise Cancellation Using Deep Cerebellar Model Articulation Controller daptive Noise Canceation Using Deep Cerebear Mode rticuation Controer Yu Tsao, Member, IEEE, Hao-Chun Chu, Shih-Wei an, Shih-Hau Fang, Senior Member, IEEE, Junghsi ee*, and Chih-Min in, Feow, IEEE bstract

More information

New Efficiency Results for Makespan Cost Sharing

New Efficiency Results for Makespan Cost Sharing New Efficiency Resuts for Makespan Cost Sharing Yvonne Beischwitz a, Forian Schoppmann a, a University of Paderborn, Department of Computer Science Fürstenaee, 3302 Paderborn, Germany Abstract In the context

More information

A proposed nonparametric mixture density estimation using B-spline functions

A proposed nonparametric mixture density estimation using B-spline functions A proposed nonparametric mixture density estimation using B-spine functions Atizez Hadrich a,b, Mourad Zribi a, Afif Masmoudi b a Laboratoire d Informatique Signa et Image de a Côte d Opae (LISIC-EA 4491),

More information

Can Active Learning Experience Be Transferred?

Can Active Learning Experience Be Transferred? Can Active Learning Experience Be Transferred? Hong-Min Chu Department of Computer Science and Information Engineering, Nationa Taiwan University E-mai: r04922031@csie.ntu.edu.tw Hsuan-Tien Lin Department

More information

Data Mining Technology for Failure Prognostic of Avionics

Data Mining Technology for Failure Prognostic of Avionics IEEE Transactions on Aerospace and Eectronic Systems. Voume 38, #, pp.388-403, 00. Data Mining Technoogy for Faiure Prognostic of Avionics V.A. Skormin, Binghamton University, Binghamton, NY, 1390, USA

More information

Collective organization in an adaptative mixture of experts

Collective organization in an adaptative mixture of experts Coective organization in an adaptative mixture of experts Vincent Vigneron, Christine Fuchen, Jean-Marc Martinez To cite this version: Vincent Vigneron, Christine Fuchen, Jean-Marc Martinez. Coective organization

More information

Reliability Improvement with Optimal Placement of Distributed Generation in Distribution System

Reliability Improvement with Optimal Placement of Distributed Generation in Distribution System Reiabiity Improvement with Optima Pacement of Distributed Generation in Distribution System N. Rugthaicharoencheep, T. Langtharthong Abstract This paper presents the optima pacement and sizing of distributed

More information

A Solution to the 4-bit Parity Problem with a Single Quaternary Neuron

A Solution to the 4-bit Parity Problem with a Single Quaternary Neuron Neura Information Processing - Letters and Reviews Vo. 5, No. 2, November 2004 LETTER A Soution to the 4-bit Parity Probem with a Singe Quaternary Neuron Tohru Nitta Nationa Institute of Advanced Industria

More information

Determining The Degree of Generalization Using An Incremental Learning Algorithm

Determining The Degree of Generalization Using An Incremental Learning Algorithm Determining The Degree of Generaization Using An Incrementa Learning Agorithm Pabo Zegers Facutad de Ingeniería, Universidad de os Andes San Caros de Apoquindo 22, Las Condes, Santiago, Chie pzegers@uandes.c

More information

Efficient Part-of-Speech Tagging with a Min-Max Modular Neural-Network Model

Efficient Part-of-Speech Tagging with a Min-Max Modular Neural-Network Model Appied Inteigence 19, 65 81, 2003 c 2003 Kuwer Academic Pubishers. Manufactured in The Netherands. Efficient Part-of-Speech Tagging with a Min-Max Moduar Neura-Network Mode BAO-LIANG LU Department of Computer

More information

Melodic contour estimation with B-spline models using a MDL criterion

Melodic contour estimation with B-spline models using a MDL criterion Meodic contour estimation with B-spine modes using a MDL criterion Damien Loive, Ney Barbot, Oivier Boeffard IRISA / University of Rennes 1 - ENSSAT 6 rue de Kerampont, B.P. 80518, F-305 Lannion Cedex

More information

Chemical Kinetics Part 2. Chapter 16

Chemical Kinetics Part 2. Chapter 16 Chemica Kinetics Part 2 Chapter 16 Integrated Rate Laws The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates

More information

A Bayesian Framework for Community Detection Integrating Content and Link

A Bayesian Framework for Community Detection Integrating Content and Link A Bayesian Framewor for Community Detection Integrating Content and Lin Tianbao Yang, Rong Jin Dept. of Computer Science and Engineering Michigan State University East Lansing, MI 48824 Yun Chi, Shenghuo

More information

(This is a sample cover image for this issue. The actual cover is not yet available at this time.)

(This is a sample cover image for this issue. The actual cover is not yet available at this time.) (This is a sampe cover image for this issue The actua cover is not yet avaiabe at this time) This artice appeared in a journa pubished by Esevier The attached copy is furnished to the author for interna

More information

Optimal Control of Assembly Systems with Multiple Stages and Multiple Demand Classes 1

Optimal Control of Assembly Systems with Multiple Stages and Multiple Demand Classes 1 Optima Contro of Assemby Systems with Mutipe Stages and Mutipe Demand Casses Saif Benjaafar Mohsen EHafsi 2 Chung-Yee Lee 3 Weihua Zhou 3 Industria & Systems Engineering, Department of Mechanica Engineering,

More information

Learning Structured Weight Uncertainty in Bayesian Neural Networks

Learning Structured Weight Uncertainty in Bayesian Neural Networks Learning Structured Weight Uncertainty in Bayesian Neura Networks Shengyang Sun Changyou Chen Lawrence Carin Tsinghua University Duke University Duke University Abstract Deep neura networks DNNs are increasingy

More information

Evolutionary Product-Unit Neural Networks for Classification 1

Evolutionary Product-Unit Neural Networks for Classification 1 Evoutionary Product-Unit Neura Networs for Cassification F.. Martínez-Estudio, C. Hervás-Martínez, P. A. Gutiérrez Peña A. C. Martínez-Estudio and S. Ventura-Soto Department of Management and Quantitative

More information

Explicit overall risk minimization transductive bound

Explicit overall risk minimization transductive bound 1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,

More information

A simple reliability block diagram method for safety integrity verification

A simple reliability block diagram method for safety integrity verification Reiabiity Engineering and System Safety 92 (2007) 1267 1273 www.esevier.com/ocate/ress A simpe reiabiity bock diagram method for safety integrity verification Haitao Guo, Xianhui Yang epartment of Automation,

More information

12.2. Maxima and Minima. Introduction. Prerequisites. Learning Outcomes

12.2. Maxima and Minima. Introduction. Prerequisites. Learning Outcomes Maima and Minima 1. Introduction In this Section we anayse curves in the oca neighbourhood of a stationary point and, from this anaysis, deduce necessary conditions satisfied by oca maima and oca minima.

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabiistic Graphica Modes Inference & Learning in DL Zhiting Hu Lecture 19, March 29, 2017 Reading: 1 Deep Generative Modes Expicit probabiistic modes Provide an expicit parametric specification of the

More information

Technical Appendix for Voting, Speechmaking, and the Dimensions of Conflict in the US Senate

Technical Appendix for Voting, Speechmaking, and the Dimensions of Conflict in the US Senate Technica Appendix for Voting, Speechmaking, and the Dimensions of Confict in the US Senate In Song Kim John Londregan Marc Ratkovic Juy 6, 205 Abstract We incude here severa technica appendices. First,

More information

Fast Blind Recognition of Channel Codes

Fast Blind Recognition of Channel Codes Fast Bind Recognition of Channe Codes Reza Moosavi and Erik G. Larsson Linköping University Post Print N.B.: When citing this work, cite the origina artice. 213 IEEE. Persona use of this materia is permitted.

More information

arxiv: v2 [stat.ap] 28 Mar 2016

arxiv: v2 [stat.ap] 28 Mar 2016 Bayesian modeing of networks in compex business inteigence probems Daniee Durante, Say Paganin, Bruno Scarpa and David B. Dunson arxiv:0.00v [stat.ap] Mar 0 University of Padua, Department of Statistica

More information

arxiv: v2 [cond-mat.stat-mech] 14 Nov 2008

arxiv: v2 [cond-mat.stat-mech] 14 Nov 2008 Random Booean Networks Barbara Drosse Institute of Condensed Matter Physics, Darmstadt University of Technoogy, Hochschustraße 6, 64289 Darmstadt, Germany (Dated: June 27) arxiv:76.335v2 [cond-mat.stat-mech]

More information

Effective Appearance Model and Similarity Measure for Particle Filtering and Visual Tracking

Effective Appearance Model and Similarity Measure for Particle Filtering and Visual Tracking Effective Appearance Mode and Simiarity Measure for Partice Fitering and Visua Tracking Hanzi Wang David Suter and Konrad Schinder Institute for Vision Systems Engineering Department of Eectrica and Computer

More information

arxiv: v1 [cs.cv] 25 Oct 2017

arxiv: v1 [cs.cv] 25 Oct 2017 Crop Panning using Stochastic Visua Optimization Gunjan Sehga *, Bindu Gupta, Kausha Paneri, Karamjit Singh, Geetika Sharma, Gautam Shroff TCS Research, India arxiv:1710.09077v1 [cs.cv] 25 Oct 2017 ABSTRACT

More information

Crystallisation of a supercooled spherical nodule in a flow

Crystallisation of a supercooled spherical nodule in a flow EUROTHERM 69 Heat and Mass Transfer in Soid-Liquid Phase hange Processes June 25-27, 2003, Bistra caste, Ljubjana, Sovenia Eds.: B. Sarer, D. Gobin rystaisation of a supercooed spherica nodue in a fow

More information

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model Appendix of the Paper The Roe of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Mode Caio Ameida cameida@fgv.br José Vicente jose.vaentim@bcb.gov.br June 008 1 Introduction In this

More information

Limited magnitude error detecting codes over Z q

Limited magnitude error detecting codes over Z q Limited magnitude error detecting codes over Z q Noha Earief choo of Eectrica Engineering and Computer cience Oregon tate University Corvais, OR 97331, UA Emai: earief@eecsorstedu Bea Bose choo of Eectrica

More information

MONTE CARLO SIMULATIONS

MONTE CARLO SIMULATIONS MONTE CARLO SIMULATIONS Current physics research 1) Theoretica 2) Experimenta 3) Computationa Monte Caro (MC) Method (1953) used to study 1) Discrete spin systems 2) Fuids 3) Poymers, membranes, soft matter

More information

NEW DEVELOPMENT OF OPTIMAL COMPUTING BUDGET ALLOCATION FOR DISCRETE EVENT SIMULATION

NEW DEVELOPMENT OF OPTIMAL COMPUTING BUDGET ALLOCATION FOR DISCRETE EVENT SIMULATION NEW DEVELOPMENT OF OPTIMAL COMPUTING BUDGET ALLOCATION FOR DISCRETE EVENT SIMULATION Hsiao-Chang Chen Dept. of Systems Engineering University of Pennsyvania Phiadephia, PA 904-635, U.S.A. Chun-Hung Chen

More information

pp in Backpropagation Convergence Via Deterministic Nonmonotone Perturbed O. L. Mangasarian & M. V. Solodov Madison, WI Abstract

pp in Backpropagation Convergence Via Deterministic Nonmonotone Perturbed O. L. Mangasarian & M. V. Solodov Madison, WI Abstract pp 383-390 in Advances in Neura Information Processing Systems 6 J.D. Cowan, G. Tesauro and J. Aspector (eds), Morgan Kaufmann Pubishers, San Francisco, CA, 1994 Backpropagation Convergence Via Deterministic

More information

DISTRIBUTION OF TEMPERATURE IN A SPATIALLY ONE- DIMENSIONAL OBJECT AS A RESULT OF THE ACTIVE POINT SOURCE

DISTRIBUTION OF TEMPERATURE IN A SPATIALLY ONE- DIMENSIONAL OBJECT AS A RESULT OF THE ACTIVE POINT SOURCE DISTRIBUTION OF TEMPERATURE IN A SPATIALLY ONE- DIMENSIONAL OBJECT AS A RESULT OF THE ACTIVE POINT SOURCE Yury Iyushin and Anton Mokeev Saint-Petersburg Mining University, Vasiievsky Isand, 1 st ine, Saint-Petersburg,

More information

TRAVEL TIME ESTIMATION FOR URBAN ROAD NETWORKS USING LOW FREQUENCY PROBE VEHICLE DATA

TRAVEL TIME ESTIMATION FOR URBAN ROAD NETWORKS USING LOW FREQUENCY PROBE VEHICLE DATA TRAVEL TIME ESTIMATIO FOR URBA ROAD ETWORKS USIG LOW FREQUECY PROBE VEHICLE DATA Erik Jeneius Corresponding author KTH Roya Institute of Technoogy Department of Transport Science Emai: erik.jeneius@abe.kth.se

More information

Incremental Reformulated Automatic Relevance Determination

Incremental Reformulated Automatic Relevance Determination IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 60, NO. 9, SEPTEMBER 22 4977 Incrementa Reformuated Automatic Reevance Determination Dmitriy Shutin, Sanjeev R. Kukarni, and H. Vincent Poor Abstract In this

More information

A. Distribution of the test statistic

A. Distribution of the test statistic A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch

More information

https://doi.org/ /epjconf/

https://doi.org/ /epjconf/ HOW TO APPLY THE OPTIMAL ESTIMATION METHOD TO YOUR LIDAR MEASUREMENTS FOR IMPROVED RETRIEVALS OF TEMPERATURE AND COMPOSITION R. J. Sica 1,2,*, A. Haefee 2,1, A. Jaai 1, S. Gamage 1 and G. Farhani 1 1 Department

More information

PREDICTION OF DEFORMED AND ANNEALED MICROSTRUCTURES USING BAYESIAN NEURAL NETWORKS AND GAUSSIAN PROCESSES

PREDICTION OF DEFORMED AND ANNEALED MICROSTRUCTURES USING BAYESIAN NEURAL NETWORKS AND GAUSSIAN PROCESSES PREDICTION OF DEFORMED AND ANNEALED MICROSTRUCTURES USING BAYESIAN NEURAL NETWORKS AND GAUSSIAN PROCESSES C.A.L. Baier-Jones, T.J. Sabin, D.J.C. MacKay, P.J. Withers Department of Materias Science and

More information

Consistent linguistic fuzzy preference relation with multi-granular uncertain linguistic information for solving decision making problems

Consistent linguistic fuzzy preference relation with multi-granular uncertain linguistic information for solving decision making problems Consistent inguistic fuzzy preference reation with muti-granuar uncertain inguistic information for soving decision making probems Siti mnah Binti Mohd Ridzuan, and Daud Mohamad Citation: IP Conference

More information

Cryptanalysis of PKP: A New Approach

Cryptanalysis of PKP: A New Approach Cryptanaysis of PKP: A New Approach Éiane Jaumes and Antoine Joux DCSSI 18, rue du Dr. Zamenhoff F-92131 Issy-es-Mx Cedex France eiane.jaumes@wanadoo.fr Antoine.Joux@ens.fr Abstract. Quite recenty, in

More information

A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes

A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes A Fundamenta Storage-Communication Tradeoff in Distributed Computing with Stragging odes ifa Yan, Michèe Wigger LTCI, Téécom ParisTech 75013 Paris, France Emai: {qifa.yan, michee.wigger} @teecom-paristech.fr

More information

Two-Stage Least Squares as Minimum Distance

Two-Stage Least Squares as Minimum Distance Two-Stage Least Squares as Minimum Distance Frank Windmeijer Discussion Paper 17 / 683 7 June 2017 Department of Economics University of Bristo Priory Road Compex Bristo BS8 1TU United Kingdom Two-Stage

More information

A Bayesian Framework for Learning Rule Sets for Interpretable Classification

A Bayesian Framework for Learning Rule Sets for Interpretable Classification Journa of Machine Learning Research 18 (2017) 1-37 Submitted 1/16; Revised 2/17; Pubished 8/17 A Bayesian Framework for Learning Rue Sets for Interpretabe Cassification Tong Wang Cynthia Rudin Finae Doshi-Veez

More information

The influence of temperature of photovoltaic modules on performance of solar power plant

The influence of temperature of photovoltaic modules on performance of solar power plant IOSR Journa of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vo. 05, Issue 04 (Apri. 2015), V1 PP 09-15 www.iosrjen.org The infuence of temperature of photovotaic modues on performance

More information

Mixed Volume Computation, A Revisit

Mixed Volume Computation, A Revisit Mixed Voume Computation, A Revisit Tsung-Lin Lee, Tien-Yien Li October 31, 2007 Abstract The superiority of the dynamic enumeration of a mixed ces suggested by T Mizutani et a for the mixed voume computation

More information

Bayesian Unscented Kalman Filter for State Estimation of Nonlinear and Non-Gaussian Systems

Bayesian Unscented Kalman Filter for State Estimation of Nonlinear and Non-Gaussian Systems Bayesian Unscented Kaman Fiter for State Estimation of Noninear and Non-aussian Systems Zhong Liu, Shing-Chow Chan, Ho-Chun Wu and iafei Wu Department of Eectrica and Eectronic Engineering, he University

More information

Theory and implementation behind: Universal surface creation - smallest unitcell

Theory and implementation behind: Universal surface creation - smallest unitcell Teory and impementation beind: Universa surface creation - smaest unitce Bjare Brin Buus, Jaob Howat & Tomas Bigaard September 15, 218 1 Construction of surface sabs Te aim for tis part of te project is

More information

Stochastic Complement Analysis of Multi-Server Threshold Queues. with Hysteresis. Abstract

Stochastic Complement Analysis of Multi-Server Threshold Queues. with Hysteresis. Abstract Stochastic Compement Anaysis of Muti-Server Threshod Queues with Hysteresis John C.S. Lui The Dept. of Computer Science & Engineering The Chinese University of Hong Kong Leana Goubchik Dept. of Computer

More information