Language Information RetrievalÓ j

Size: px
Start display at page:

Download "Language Information RetrievalÓ j"

Transcription

1 ± i jschang@cs.nthu.edu.tw o ± k j ± j z ± wyw j ± j ÒCross Language Information RetrievalÓ j h ±ÒQuery TranslationÓ ± k h ± ± ± k ± h ± j ± ± ± 1. ± j y k j ± j z ± y p wywj ± 1

2 j ÒCross-Language Information RetrievalÓ q z ± ± j ä ±åòdocument translationó ä hi ±åòquery translationó ÒMcCarley 1999Ó h ± NTCIR-2 i ÒKando et al. 2001Ó h h Assembly Parade Law Parade and Demonstration Constitution Freedom of speech Indemnification Communism Country separation Council of Grand Justices Legislation Amendments h ± dg ÒWord Sense DisambiguationÓ ÒIde and Veronis 1998, Chen and Chang 1998Ó ±ÒPhrase TranslationÓ ± ± ÒText CollectionÓ ± d g ± ÒLexical ChoiceÓ demonstration ± ä å ± ä å 2

3 j u 1. ± ÒGey and Chen 1997, Kwok 2001Ó 2. ÒOard 1999, Kwok 2001Ó u ± ± Kwok ± Michael Jordan ±ä å j ± Chen Ò1999Ó Òoccurrence statisticsó ± µ ± j ± ÒKnight and Graehl 1997, Chang et al. 2001Ó j j p j h ± ± ± ÒExample-based ApproachÓ ± ji Òdata-drivenÓ j i ± IBM Watson Brown Ò1988, 1990, 1993Ó ± j p j k u ± ÒStatistical Alignment and Machine TranslationÓ h ± j h ± j g u ± ÒAlignment ProbabilityÓ 3

4 ± 2. ± ± ± ÒDirect ApproachÓ r ÒTransfer ApproachÓ 1980 j ÒEmpirical ApproachÓ ± ± Brown p Brown S ± T ± ÒTranslation ProbabilityÓ Pr( T S ) g (a) ± ÒLexical Translation ProbabilityÓ Pr( S i T j ) (b) ÒFertility ProbabilityÓ Pr( a b ) (c) ÒDistortion ProbabilityÓ Pr(i j, k, m ) S i S i T j T j a S i b T j k S m T Brown ± k ä å ä åqv ÒExpectation and Maximization 4

5 AlgorithmÓ ä å º ± ä å º ± k EM qv ± ± k ÒNoisy Channel ModelÓ ± N-gram ÒLanguage ModelÓ qv ÒBeam SearchÓ ± 3. ± Brown ± ± ÒalignmentÓ y ± ± S i, i i T j S i j 1 Pr(j i, k, m) 1 Pr(j i, k, m) =0,i i k ± ± A* S 1 S 2 S 3 ± S 1 {T 1, T 2 } S 2 {T 3, T 4 } S 3 {T 5 } A* = (0, 12, 34, 5)Ò 0 Ó k =3 m =5 ± A* 35% A* ÒMaximum Likelihood EstimationÓ 5

6 Pr MLE ( A* )=0.35 Pr( A* )=P(1 1,3,5) P(2 1,3,5) P(3 2,3,5) P(4 2,3,5) P(5 3,3,5) j Ò0.6Ó P(j i,3,5) k Pr( A* ) < (0.6) 5 = << 0.35 Pr(A*)<<Pr MLE ( A* ) w ± ± ± ÒAlignment ProbabilityÓ ± ± S ± T ± PrÒT SÓ g (a) ± ÒLexical Translation ProbabilityÓ Pr( T(A i ) S i ) (b) ± ÒAlignment ProbabilityÓ Pr ( A k, m )=Pr ( A 0, A 1, A 2,,A k k, m ) S i S i T(A i ) T S i A 0 T S f A i T S i f, i>0 k S m T S i T 6

7 B(A, i) E(A, i) k S i T(A i )={T B(A,i), T B(A,i)+1, T E(A,i) } A i ={B(A,i), B(A,i)+1,, E(A,i) } 4. ± k g 1. ± j 2. ± u ± k kn 3. ± j k 4. ± j 4.1 BDCr ÒBDC 1992Ó 3 ± ± i i k 96,156 ± ÒP n, Q n Ó,n=1,N h EM qv ª ± ± Och Ò2000Ó 7

8 IBM 1. Brown EM qv ± 2. ± j jv Pr(j i, k, m )=1/m j 0.5 i 0.5 3I ( i j, k, m) = 1 [1] m k i = k = j = m = S T i k j M Pr( j i,k,m) flight flight flight flight eight eight eight eight ± 2 ± Pr( j i,k,m) Òflight eight, Ó 1 v 1 8

9 E C ± Pr (C E) Pr( C E) N k m n= 1 i= 1 j= 1 = N δ ( E, P ( i)) δ ( C, Q ( j)) Pr( j i, k, m) k m n= 1 i= 1 j= 1 n δ ( E, P ( i)) Pr( j i, k, m) P n (i) P n i Q n (j) Q n j k = P n m = Q n δ(x, y) =1 x = y, δ(x, y) =0 x y n n [2] S i T j i k j m Pr(j i,k,m) Pr(T j S i ) Pr(T j S i ) Pr(j i,k,m) flight flight flight flight eight eight eight eight ± 2 E C E Pr (C E) ± 9

10 4.2 EM qv v EM qv º ÒGreedy MethodÓ ÒP n, Q n Ó 0 ± Ò 2Ó º (threshold) k y ¼ flight eight 2 3 Ò0, 34, 12Ó S i T i i j k m Pr(j i,k,m) Pr(T j S i ) Pr(T j S i ) Pr(j i,k,m) flight flight eight eight Òflight eight, Ó Ò0, 34, 12Ó v - ± k 10

11 ± k m u A count( A ( S, T) ) Pr( A k, m) = [3] count( k = S, m = T ) 4 k m A A 0 A 1 A 2 Pr(A k,m) ± 15 EM qv 601 u 38 u ¼ 11

12 1. ± ± ± % y S T T(A 0 ) S 1 T(A 1 ) S 2 T(A 2 ) T-shaped antenna ƒ T-shaped antenna ƒ X-ray examination X-ray examination irresistible force irresistible force Unwritten law unwritten law Central Asia Central Asia mutual non-interference mutual non-interference undesirable element undesirable element unalterable truth unalterable truth come soon come soon a desperado a desperado 5 5 u v ± 4.2 ± g 12

13 Òbound morphemeó $empty$ i Òdata sparsenessó $any$ Good-Turing Òsmoothing methodó $any$ ± E C PrÒC EÓ flight flight flight flight $empty$ flight flight flight flight flight flight flight flight flight d flight $any$ flight ± 6 flight ± $empty$ $any$ flight $empty$ k ± 4 Ò0,0,1234Ó 13

14 Ò1,0,234Ó $any$ EM qv $empty$ r 4.3 EM qv v jv 1 ± S T A 0 A 1 A 2 T(A 0 ) T(A 1 ) T(A 2 ) Pr(T S,A) flight eight flight eight flight eight $empty$ flight eight flight eight Pr( flight eight) 5 jv ÒS, TÓ A v ± Pr(T S, A) A Pr(T S, A) A A (S i, T(A i )) Pr( T S) = max Pr( T S, A) = max Pr( A k, m) A A* A k i= 1 Pr( T ( A ) S ) i i 14

15 A * = arg max Pr( T S, A) A = arg max Pr( A k, m) A k i= 1 Pr( T( A ) S ) i i [4] k = S, m = T (S, T) =Òflight eight, Ó A ± v A = (0, 12, 34) Pr(T S,A) =Pr(0,12,34 2,4) Pr( flight) Pr( eight) A = (0, 34, 12) Pr(T S,A) =Pr(0,34,12 2,4) Pr( flight) Pr( eight) A = (0, 3, 124) Pr(T S,A) =Pr(0, 3, 124 2,4) Pr( flight) Pr( eight) A = (2, 34, 1) Pr(T S,A) =Pr(2, 34, 1 2,4) Pr( flight) Pr( eight) A = (12, 34, 0) Pr(T S,A) =Pr(12, 34, 0 2,4) Pr( flight) Pr($empty$ eight) 7 Òflight eight, Ó ± 7 A* = (0, 34, 12) 5. ± ± k 10 i EM qv j j 15

16 S T T(A 0 ) T(A 1 ) T(A 2 ) T(A 0 ) T(A 1 ) T(A 2 ) association football delay flip-flop I demodulator fg fg f g Disgraceful act disregard to secret ballot bearer stock false retrieval used car infix operation jv jv jv 8 jv ± ¼ ± j j 1. j n ± association ± ä å ä å association football ± ± äa å 2. ± ± Òlight verbó make take to 3. ± Òflight eight, Ó ¼ u j Brown u 16

17 8 Och (2000) 100 ( ) 2 u S (sure) P (possible) S P S P j (recall) (precision) (error rate) A Ι S recall = S 185 = = A Ι P precision = A 226 = = A Ι S + AΙ P AER( S, P; A) = 1 = 1 = A + S ± 6.1 ± ± i k i Òflight attendant Ó Òflight, Ó Òflight attendant Ó flight $any$ $any$ j 17

18 j œ~ d d œ~ Òflight, Ó ±Òflight, Ó Òflight, Ó ± Òflight, Ó 0 Òattendant, Ó ±Òattendant, Ó Òattendant, Ó Òattendant, Ó Òattendant, $any$ó ± j Òpreservation, Ó Òpreservation, Ó Òpreservation, Ó ± Òflight attendant Ó ± Pr($any$ flight) Pr($any$ eight) m ± Pr(A 2,3) u Ò0, 12, 3Ó Òflight, Ó Òattendant, Ó ± Òflight, Ó Òattendant, Ó j j Òflight, Ó Òattendant, Ó 18

19 LTP º k LTP j 6.2 ± ± ±j ± ± v j j h ± ± ± h ± u 1. ± Òdocument frequencyó ± ± 2. k k ± N-gram qv ± ± ª ± ± h S ± T* Pr( T S) = max Pr( T S, A) A 19

20 T * = arg max Pr( T S) Pr( T ) T = arg max max Pr( T S, A) Pr( T ) T = arg max max Pr( A k, m) T A A k i= 1 Pr( T ( A ) S ) Pr( T ) Pr(T) = ± T k = S m = T Pr(A k, m) Pr(T S) v qv ÒBranch and Bound AlgorithmÓ T* Pr(A k, *) Pr(* S i ) gòsolutionó Òupper boundó Pr(* S i ) Pr(T) N-gram k r g œ ± ƒ ± ÒKupiec 1993Ó ± k º P v ± A ± T j+1, j+m i i arg max Pr( A n( i), m) Pr( T A, j, m k= 1 j+ 1, j+ m = arg max Pr( A n( i), m) Pr( T A, j, m n( i ) = arg max Pr( A n( i), m) Pr( T A, j, m j+ 1, j+ m P, A) S b, e, A) j+ B( A, K ), j+ E( A, k ) 20 S b+ k 1 A n m )

21 Pr( A n, m) A S b+k-1 P k T j+b(a,k),j+e(a,k) P k ± A ± qv S T k (P 1, Q 1 ), (P 2, Q 2 ),,(P k, Q k ) (1). ÒPart of speech taggeró S ÒlemmaÓ (2). Òshallow parsingó Òbasic phraseó P 1, P 2,,P k P I = S b(i), e(i), P b =n(i) (3). P I v ± A ± T j ð (i) 1, j m*(i) ÒBranch and Bound AlgorithmÓ A*(i), j*(i), m*(i) argmax Pr( A n( i), m) A, j, m n( i) k = 1 Pr( T j+ B( A, k ), j+ E ( A, k ) S b+ k 1 ) (4). P I (P I, T j ð (i) 1, j m*(i) ) 7. ± ± h ± j 21

22 Brown v ±ª ƒ ± ± ± j h 1. ± 2. dg h œ ± 3. d Wordnet ± dg ƒf H g ~ ± 1. BDC 1992 The BDC Chinese-English electronic dictionary (version 2.0), Behavior Design Corporation, Taiwan. 2. Brown, P. F., Cocke J., Della Pietra S. A., Della Pietra V. J., Jelinek F., Mercer R. L., and Roosin P. S A Statistical Approach to Language Translation, In Proceedings of the 12th International Conference on Computational Linguistics, Budapest, Hungary, pp

23 3. Brown, P. F., Cocke J., Della Pietra S. A., Della Pietra V. J., Jelinek F., Lafferty J. D., Mercer R. L., and Roosin P. S A Statistical Approach to Machine Translation, Computational Linguistics, 16/2, pp Brown, P. F., Della Pietra S. A., Della Pietra V. J., and Mercer R. L The Mathematics of Statistical Machine Translation: Parameter Estimation, Computational Linguistics, 19/2, pp Chang, J. S. et al Nathu IR System at NTCIR-2. In Proceedings of the Second NTCIR Workshop Meeting on Evaluation of Chinese and Japanese Text Retrieval and Text Summarization, pp. (5) 49-52, National Institute of Informatics, Japan. 6. Chang, J. S., Ker S. J., and Chen M. H Taxonomy and Lexical Semantics From the Perspective of Machine Readable Dictionary, In Proceedings of the third Conference of the Association for Machine Translation in the Americas (AMTA), pp Chen, H.H., G.W. Bian and W.C. Lin Resolving Translation Ambiguity and Target Polysemy in Cross-Language Information Retrieval. In Proceedings of the 37 th Annual Meeting of the Association for Computation Linguistics, pp Dagan, I., Church K. W. and Gale W. A Robust Bilingual Word Alignment or Machine Aided Translation, In Proceedings of the Workshop on Very Large Corpora Academic and Industrial Perspectives, pp Fung, P. and McKeown K Aligning Noisy Parallel Corpora across Language Groups: Word Pair Feature Matching by Dynamic Time Warping, In Proceedings of the First Conference of the Association for Machine Translation in the Americas (AMTA), pp , Columbia, Maryland, USA. 10. Gale, W. A. and Church K. W Identifying Word Correspondences in Parallel Texts, In Proceedings of the Fourth DARPA Speech and Natural Language Workshop, pp Gey, F C and A. Chen Phrase Discovery for English and Cross-Language Retrieval at TREC-6. In Proceedings of the 6 th Text Retrieval Evaluation Conference, pp

24 12. Ide, N. and J Veronis Special Issue on Word Sense Disambiguation, editors, Computational Linguistics, 24/ Isabelle, P Machine Translation at the TAUM Group, In M. King, editor, Machine Translation Today: The State of the Art, Proceedings of the Third Lugano Tutorial, pp Kando, Noriko, Kenro Aihara, Koji Eguchi and Hiroyuki Kato Proceedings of the Second NTCIR Workshop Meeting on Evaluation of Chinese and Japanese Text Retrieval and Text Summarization, National Institute of Informatics, Japan. 15. Kay, M. and Röscheisen M Text-Translation Alignment, Technical Report P , Xerox Palo Alto Research Center. 16. Ker, S. J. and Chang J. S A Class-base Approach to Word Alignment, Computational Linguistics, 23/2, pp Knight, K. and J Graehl Machine Transliteration, In Proceedings of the 35 th Annual Meeting of the Association for Computational Linguistics and 8 th Conference of ACL European Chapter, pp Kupiec, Julian An Algorithm for finding noun phrase correspondence in bilingual corpus, In ACL 31, 23/2, pp Kwok, K L NTCIR-2 Chinese, Cross-Language Retrieval Experiments Using PIRCS. In Proceedings of the Second NTCIR Workshop Meeting on Evaluation of Chinese and Japanese Text Retrieval and Text Summarization, pp. (5) 14-20, National Institute of Informatics, Japan. 20. Longman Group 1992 Longman English-Chinese Dictionary of Contemporary English, Published by Longman Group (Far East) Ltd., Hong Kong. 21. McCarley, J. Scott Should we Translate the Documents or the Queries in Cross-Language Information Retrieval? In Proceedings of the 37 th Annual Meeting of the Association for Computation Linguistics, pp Melamed, I. D Automatic Construction of Clean Broad-Coverage Translation Lexicons, In Proceedings of the second Conference of the Association for Machine Translation in the Americas (AMTA), pp Nagao, M Machine Translation: How Far Can it Go? Oxford University Press, Oxford. 24

25 24. Oard, D W and J. Wang Effect of Term Segmentation on Chinese/English Cross-Language Information Retrieval. In Proceedings of the Symposium on String and Processing and Information Retrieval Och, Franz Josef and Hermann Ney Improved Statistical Alignment Models. In Proceedings of the 38 th Annual Meeting of the Association for Computation Linguistics. 26. Pirkola, A The Effect of Query Structure and Dictionary Setups in Dictionary-based Cross-Language Retrieval. In Proceedings of the 21 st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp Smadja, F., McKeown K., and Hatzivassiloglou V Translating Collocations for Bilingual Lexicons: A Statistical Approach, Computational Linguistics, 22/1, pp Utsuro, T., Ikeda H., Yamane M., Matsumoto M., and Nagao M Bilingual Text Matching Using Bilingual Dictionary and Statistics, In Proceedings of the 15th International Conference on Computational Linguistics, pp Wu, D. and Xia X Learning an English-Chinese Lexicon from a Parallel Corpus, In Proceedings of the first Conference of the Association for Machine Translation in the Americas (AMTA), pp

Utilizing Portion of Patent Families with No Parallel Sentences Extracted in Estimating Translation of Technical Terms

Utilizing Portion of Patent Families with No Parallel Sentences Extracted in Estimating Translation of Technical Terms 1 1 1 2 2 30% 70% 70% NTCIR-7 13% 90% 1,000 Utilizing Portion of Patent Families with No Parallel Sentences Extracted in Estimating Translation of Technical Terms Itsuki Toyota 1 Yusuke Takahashi 1 Kensaku

More information

Structure and Complexity of Grammar-Based Machine Translation

Structure and Complexity of Grammar-Based Machine Translation Structure and of Grammar-Based Machine Translation University of Padua, Italy New York, June 9th, 2006 1 2 Synchronous context-free grammars Definitions Computational problems 3 problem SCFG projection

More information

Speech Recognition Lecture 5: N-gram Language Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri

Speech Recognition Lecture 5: N-gram Language Models. Eugene Weinstein Google, NYU Courant Institute Slide Credit: Mehryar Mohri Speech Recognition Lecture 5: N-gram Language Models Eugene Weinstein Google, NYU Courant Institute eugenew@cs.nyu.edu Slide Credit: Mehryar Mohri Components Acoustic and pronunciation model: Pr(o w) =

More information

Cross-Lingual Language Modeling for Automatic Speech Recogntion

Cross-Lingual Language Modeling for Automatic Speech Recogntion GBO Presentation Cross-Lingual Language Modeling for Automatic Speech Recogntion November 14, 2003 Woosung Kim woosung@cs.jhu.edu Center for Language and Speech Processing Dept. of Computer Science The

More information

Automatic Evaluation in Text Summarization

Automatic Evaluation in Text Summarization 1 Automatic Evaluation in Text Summarization Hidetsugu Nanba Tsutomu Hirao Hiroshima City University nanba@hiroshima-cuacjp, http://wwwnlpitshiroshima-cuacjp/~nanba/ NTT NTT Communication Science Laboratories

More information

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister A Syntax-based Statistical Machine Translation Model Alexander Friedl, Georg Teichtmeister 4.12.2006 Introduction The model Experiment Conclusion Statistical Translation Model (STM): - mathematical model

More information

Triplet Lexicon Models for Statistical Machine Translation

Triplet Lexicon Models for Statistical Machine Translation Triplet Lexicon Models for Statistical Machine Translation Saša Hasan, Juri Ganitkevitch, Hermann Ney and Jesús Andrés Ferrer lastname@cs.rwth-aachen.de CLSP Student Seminar February 6, 2009 Human Language

More information

Word-Transliteration Alignment

Word-Transliteration Alignment Word-Transliteration Alignment Tracy Lin Dep. of Communication Engineering National Chiao Tung University, 1001, Ta Hsueh Road, Hsinchu, 300, Taiwan tracylin@cm.nctu.edu.tw Chien-Cheng Wu Department of

More information

statistical machine translation

statistical machine translation statistical machine translation P A R T 3 : D E C O D I N G & E V A L U A T I O N CSC401/2511 Natural Language Computing Spring 2019 Lecture 6 Frank Rudzicz and Chloé Pou-Prom 1 University of Toronto Statistical

More information

An Algorithm for Fast Calculation of Back-off N-gram Probabilities with Unigram Rescaling

An Algorithm for Fast Calculation of Back-off N-gram Probabilities with Unigram Rescaling An Algorithm for Fast Calculation of Back-off N-gram Probabilities with Unigram Rescaling Masaharu Kato, Tetsuo Kosaka, Akinori Ito and Shozo Makino Abstract Topic-based stochastic models such as the probabilistic

More information

Natural Language Processing (CSEP 517): Machine Translation

Natural Language Processing (CSEP 517): Machine Translation Natural Language Processing (CSEP 57): Machine Translation Noah Smith c 207 University of Washington nasmith@cs.washington.edu May 5, 207 / 59 To-Do List Online quiz: due Sunday (Jurafsky and Martin, 2008,

More information

N-grams. Motivation. Simple n-grams. Smoothing. Backoff. N-grams L545. Dept. of Linguistics, Indiana University Spring / 24

N-grams. Motivation. Simple n-grams. Smoothing. Backoff. N-grams L545. Dept. of Linguistics, Indiana University Spring / 24 L545 Dept. of Linguistics, Indiana University Spring 2013 1 / 24 Morphosyntax We just finished talking about morphology (cf. words) And pretty soon we re going to discuss syntax (cf. sentences) In between,

More information

An Empirical Study on Dimensionality Optimization in Text Mining for Linguistic Knowledge Acquisition

An Empirical Study on Dimensionality Optimization in Text Mining for Linguistic Knowledge Acquisition An Empirical Study on Dimensionality Optimization in Text Mining for Linguistic Knowledge Acquisition Yu-Seop Kim 1, Jeong-Ho Chang 2, and Byoung-Tak Zhang 2 1 Division of Information and Telecommunication

More information

Investigating Connectivity and Consistency Criteria for Phrase Pair Extraction in Statistical Machine Translation

Investigating Connectivity and Consistency Criteria for Phrase Pair Extraction in Statistical Machine Translation Investigating Connectivity and Consistency Criteria for Phrase Pair Extraction in Statistical Machine Translation Spyros Martzoukos Christophe Costa Florêncio and Christof Monz Intelligent Systems Lab

More information

Word Alignment by Thresholded Two-Dimensional Normalization

Word Alignment by Thresholded Two-Dimensional Normalization Word Alignment by Thresholded Two-Dimensional Normalization Hamidreza Kobdani, Alexander Fraser, Hinrich Schütze Institute for Natural Language Processing University of Stuttgart Germany {kobdani,fraser}@ims.uni-stuttgart.de

More information

N-gram N N-gram. N-gram. Detection and Correction for Errors in Hiragana Sequences by a Hiragana Character N-gram.

N-gram N N-gram. N-gram. Detection and Correction for Errors in Hiragana Sequences by a Hiragana Character N-gram. Vol. 40 No. 6 June 1999 N N 3 N N 5 36-gram 5 4-gram Detection and Correction for Errors in Hiragana Sequences by a Hiragana Character Hiroyuki Shinnou In this paper, we propose the hiragana character

More information

Learning Features from Co-occurrences: A Theoretical Analysis

Learning Features from Co-occurrences: A Theoretical Analysis Learning Features from Co-occurrences: A Theoretical Analysis Yanpeng Li IBM T. J. Watson Research Center Yorktown Heights, New York 10598 liyanpeng.lyp@gmail.com Abstract Representing a word by its co-occurrences

More information

The Noisy Channel Model and Markov Models

The Noisy Channel Model and Markov Models 1/24 The Noisy Channel Model and Markov Models Mark Johnson September 3, 2014 2/24 The big ideas The story so far: machine learning classifiers learn a function that maps a data item X to a label Y handle

More information

Multi-Task Word Alignment Triangulation for Low-Resource Languages

Multi-Task Word Alignment Triangulation for Low-Resource Languages Multi-Task Word Alignment Triangulation for Low-Resource Languages Tomer Levinboim and David Chiang Department of Computer Science and Engineering University of Notre Dame {levinboim.1,dchiang}@nd.edu

More information

ACS Introduction to NLP Lecture 2: Part of Speech (POS) Tagging

ACS Introduction to NLP Lecture 2: Part of Speech (POS) Tagging ACS Introduction to NLP Lecture 2: Part of Speech (POS) Tagging Stephen Clark Natural Language and Information Processing (NLIP) Group sc609@cam.ac.uk The POS Tagging Problem 2 England NNP s POS fencers

More information

Multi-Source Neural Translation

Multi-Source Neural Translation Multi-Source Neural Translation Barret Zoph and Kevin Knight Information Sciences Institute Department of Computer Science University of Southern California {zoph,knight}@isi.edu Abstract We build a multi-source

More information

A Systematic Comparison of Training Criteria for Statistical Machine Translation

A Systematic Comparison of Training Criteria for Statistical Machine Translation A Systematic Comparison of Training Criteria for Statistical Machine Translation Richard Zens and Saša Hasan and Hermann Ney Human Language Technology and Pattern Recognition Lehrstuhl für Informatik 6

More information

Computing Optimal Alignments for the IBM-3 Translation Model

Computing Optimal Alignments for the IBM-3 Translation Model Computing Optimal Alignments for the IBM-3 Translation Model Thomas Schoenemann Centre for Mathematical Sciences Lund University, Sweden Abstract Prior work on training the IBM-3 translation model is based

More information

Multi-Source Neural Translation

Multi-Source Neural Translation Multi-Source Neural Translation Barret Zoph and Kevin Knight Information Sciences Institute Department of Computer Science University of Southern California {zoph,knight}@isi.edu In the neural encoder-decoder

More information

Machine Learning for natural language processing

Machine Learning for natural language processing Machine Learning for natural language processing N-grams and language models Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 25 Introduction Goals: Estimate the probability that a

More information

Phrasetable Smoothing for Statistical Machine Translation

Phrasetable Smoothing for Statistical Machine Translation Phrasetable Smoothing for Statistical Machine Translation George Foster and Roland Kuhn and Howard Johnson National Research Council Canada Ottawa, Ontario, Canada firstname.lastname@nrc.gc.ca Abstract

More information

A Recursive Statistical Translation Model

A Recursive Statistical Translation Model A Recursive Statistical Translation Model Juan Miguel Vilar Dpto. de Lenguajes y Sistemas Informáticos Universitat Jaume I Castellón (Spain) jvilar@lsi.uji.es Enrique Vidal Dpto. de Sistemas Informáticos

More information

Maximal Lattice Overlap in Example-Based Machine Translation

Maximal Lattice Overlap in Example-Based Machine Translation Maximal Lattice Overlap in Example-Based Machine Translation Rebecca Hutchinson Paul N. Bennett Jaime Carbonell Peter Jansen Ralf Brown June 6, 2003 CMU-CS-03-138 School of Computer Science Carnegie Mellon

More information

IBM Model 1 for Machine Translation

IBM Model 1 for Machine Translation IBM Model 1 for Machine Translation Micha Elsner March 28, 2014 2 Machine translation A key area of computational linguistics Bar-Hillel points out that human-like translation requires understanding of

More information

Theory of Alignment Generators and Applications to Statistical Machine Translation

Theory of Alignment Generators and Applications to Statistical Machine Translation Theory of Alignment Generators and Applications to Statistical Machine Translation Raghavendra Udupa U Hemanta K Mai IBM India Research Laboratory, New Delhi {uraghave, hemantkm}@inibmcom Abstract Viterbi

More information

Comparison of Log-Linear Models and Weighted Dissimilarity Measures

Comparison of Log-Linear Models and Weighted Dissimilarity Measures Comparison of Log-Linear Models and Weighted Dissimilarity Measures Daniel Keysers 1, Roberto Paredes 2, Enrique Vidal 2, and Hermann Ney 1 1 Lehrstuhl für Informatik VI, Computer Science Department RWTH

More information

Natural Language Processing. Statistical Inference: n-grams

Natural Language Processing. Statistical Inference: n-grams Natural Language Processing Statistical Inference: n-grams Updated 3/2009 Statistical Inference Statistical Inference consists of taking some data (generated in accordance with some unknown probability

More information

Toponym Disambiguation using Ontology-based Semantic Similarity

Toponym Disambiguation using Ontology-based Semantic Similarity Toponym Disambiguation using Ontology-based Semantic Similarity David S Batista 1, João D Ferreira 2, Francisco M Couto 2, and Mário J Silva 1 1 IST/INESC-ID Lisbon, Portugal {dsbatista,msilva}@inesc-id.pt

More information

Out of GIZA Efficient Word Alignment Models for SMT

Out of GIZA Efficient Word Alignment Models for SMT Out of GIZA Efficient Word Alignment Models for SMT Yanjun Ma National Centre for Language Technology School of Computing Dublin City University NCLT Seminar Series March 4, 2009 Y. Ma (DCU) Out of Giza

More information

P R + RQ P Q: Transliteration Mining Using Bridge Language

P R + RQ P Q: Transliteration Mining Using Bridge Language P R + RQ P Q: Transliteration Mining Using Bridge Language Mitesh M. Khapra Raghavendra Udupa n Institute of Technology Microsoft Research, Bombay, Bangalore, Powai, Mumbai 400076 raghavu@microsoft.com

More information

Tuning as Linear Regression

Tuning as Linear Regression Tuning as Linear Regression Marzieh Bazrafshan, Tagyoung Chung and Daniel Gildea Department of Computer Science University of Rochester Rochester, NY 14627 Abstract We propose a tuning method for statistical

More information

On Using Selectional Restriction in Language Models for Speech Recognition

On Using Selectional Restriction in Language Models for Speech Recognition On Using Selectional Restriction in Language Models for Speech Recognition arxiv:cmp-lg/9408010v1 19 Aug 1994 Joerg P. Ueberla CMPT TR 94-03 School of Computing Science, Simon Fraser University, Burnaby,

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa Recap: vector space model Represent both doc and query by concept vectors Each concept defines one dimension K concepts define a high-dimensional space Element

More information

Natural Language Processing (CSE 490U): Language Models

Natural Language Processing (CSE 490U): Language Models Natural Language Processing (CSE 490U): Language Models Noah Smith c 2017 University of Washington nasmith@cs.washington.edu January 6 9, 2017 1 / 67 Very Quick Review of Probability Event space (e.g.,

More information

Language Modelling. Marcello Federico FBK-irst Trento, Italy. MT Marathon, Edinburgh, M. Federico SLM MT Marathon, Edinburgh, 2012

Language Modelling. Marcello Federico FBK-irst Trento, Italy. MT Marathon, Edinburgh, M. Federico SLM MT Marathon, Edinburgh, 2012 Language Modelling Marcello Federico FBK-irst Trento, Italy MT Marathon, Edinburgh, 2012 Outline 1 Role of LM in ASR and MT N-gram Language Models Evaluation of Language Models Smoothing Schemes Discounting

More information

Foundations of Natural Language Processing Lecture 5 More smoothing and the Noisy Channel Model

Foundations of Natural Language Processing Lecture 5 More smoothing and the Noisy Channel Model Foundations of Natural Language Processing Lecture 5 More smoothing and the Noisy Channel Model Alex Lascarides (Slides based on those from Alex Lascarides, Sharon Goldwater and Philipop Koehn) 30 January

More information

Statistical Substring Reduction in Linear Time

Statistical Substring Reduction in Linear Time Statistical Substring Reduction in Linear Time Xueqiang Lü Institute of Computational Linguistics Peking University, Beijing lxq@pku.edu.cn Le Zhang Institute of Computer Software & Theory Northeastern

More information

Decoding Complexity in. Word-Replacement Translation Models. Kevin Knight. University of Southern California

Decoding Complexity in. Word-Replacement Translation Models. Kevin Knight. University of Southern California Decoding Complexity in Word-Replacement Translation Models Kevin Knight University of Southern California Statistical machine translation is a relatively new approach to the longstanding problem of translating

More information

Machine Translation. CL1: Jordan Boyd-Graber. University of Maryland. November 11, 2013

Machine Translation. CL1: Jordan Boyd-Graber. University of Maryland. November 11, 2013 Machine Translation CL1: Jordan Boyd-Graber University of Maryland November 11, 2013 Adapted from material by Philipp Koehn CL1: Jordan Boyd-Graber (UMD) Machine Translation November 11, 2013 1 / 48 Roadmap

More information

Word Alignment via Quadratic Assignment

Word Alignment via Quadratic Assignment Word Alignment via Quadratic Assignment Simon Lacoste-Julien UC Berkeley, Berkeley, CA 94720 slacoste@cs.berkeley.edu Ben Taskar UC Berkeley, Berkeley, CA 94720 taskar@cs.berkeley.edu Dan Klein UC Berkeley,

More information

Ranked Retrieval (2)

Ranked Retrieval (2) Text Technologies for Data Science INFR11145 Ranked Retrieval (2) Instructor: Walid Magdy 31-Oct-2017 Lecture Objectives Learn about Probabilistic models BM25 Learn about LM for IR 2 1 Recall: VSM & TFIDF

More information

Natural Language Processing (CSEP 517): Machine Translation (Continued), Summarization, & Finale

Natural Language Processing (CSEP 517): Machine Translation (Continued), Summarization, & Finale Natural Language Processing (CSEP 517): Machine Translation (Continued), Summarization, & Finale Noah Smith c 2017 University of Washington nasmith@cs.washington.edu May 22, 2017 1 / 30 To-Do List Online

More information

Exploring Asymmetric Clustering for Statistical Language Modeling

Exploring Asymmetric Clustering for Statistical Language Modeling Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL, Philadelphia, July 2002, pp. 83-90. Exploring Asymmetric Clustering for Statistical Language Modeling Jianfeng

More information

Phrase-Based Statistical Machine Translation with Pivot Languages

Phrase-Based Statistical Machine Translation with Pivot Languages Phrase-Based Statistical Machine Translation with Pivot Languages N. Bertoldi, M. Barbaiani, M. Federico, R. Cattoni FBK, Trento - Italy Rovira i Virgili University, Tarragona - Spain October 21st, 2008

More information

SYNTHER A NEW M-GRAM POS TAGGER

SYNTHER A NEW M-GRAM POS TAGGER SYNTHER A NEW M-GRAM POS TAGGER David Sündermann and Hermann Ney RWTH Aachen University of Technology, Computer Science Department Ahornstr. 55, 52056 Aachen, Germany {suendermann,ney}@cs.rwth-aachen.de

More information

Word Alignment via Quadratic Assignment

Word Alignment via Quadratic Assignment University of Pennsylvania ScholarlyCommons Departmental Papers (CIS) Department of Computer & Information Science 2006 Word Alignment via Quadratic Assignment Simon Lacoste-Julien University of California

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa VS model in practice Document and query are represented by term vectors Terms are not necessarily orthogonal to each other Synonymy: car v.s. automobile Polysemy:

More information

Empirical Methods in Natural Language Processing Lecture 10a More smoothing and the Noisy Channel Model

Empirical Methods in Natural Language Processing Lecture 10a More smoothing and the Noisy Channel Model Empirical Methods in Natural Language Processing Lecture 10a More smoothing and the Noisy Channel Model (most slides from Sharon Goldwater; some adapted from Philipp Koehn) 5 October 2016 Nathan Schneider

More information

Prenominal Modifier Ordering via MSA. Alignment

Prenominal Modifier Ordering via MSA. Alignment Introduction Prenominal Modifier Ordering via Multiple Sequence Alignment Aaron Dunlop Margaret Mitchell 2 Brian Roark Oregon Health & Science University Portland, OR 2 University of Aberdeen Aberdeen,

More information

cmp-lg/ Aug 1996

cmp-lg/ Aug 1996 Automatic Alignment of nglish-hinese Bilingual Texts of NS News Donghua Xu hew Lim Tan xudonghu@iscs.nus.sg tancl@iscs.nus.sg Department of Information System and omputer Science, National University of

More information

Midterm sample questions

Midterm sample questions Midterm sample questions CS 585, Brendan O Connor and David Belanger October 12, 2014 1 Topics on the midterm Language concepts Translation issues: word order, multiword translations Human evaluation Parts

More information

Word Alignment via Submodular Maximization over Matroids

Word Alignment via Submodular Maximization over Matroids Word Alignment via Submodular Maximization over Matroids Hui Lin, Jeff Bilmes University of Washington, Seattle Dept. of Electrical Engineering June 21, 2011 Lin and Bilmes Submodular Word Alignment June

More information

The OntoNL Semantic Relatedness Measure for OWL Ontologies

The OntoNL Semantic Relatedness Measure for OWL Ontologies The OntoNL Semantic Relatedness Measure for OWL Ontologies Anastasia Karanastasi and Stavros hristodoulakis Laboratory of Distributed Multimedia Information Systems and Applications Technical University

More information

Probabilistic Language Modeling

Probabilistic Language Modeling Predicting String Probabilities Probabilistic Language Modeling Which string is more likely? (Which string is more grammatical?) Grill doctoral candidates. Regina Barzilay EECS Department MIT November

More information

DT2118 Speech and Speaker Recognition

DT2118 Speech and Speaker Recognition DT2118 Speech and Speaker Recognition Language Modelling Giampiero Salvi KTH/CSC/TMH giampi@kth.se VT 2015 1 / 56 Outline Introduction Formal Language Theory Stochastic Language Models (SLM) N-gram Language

More information

Graphical models for part of speech tagging

Graphical models for part of speech tagging Indian Institute of Technology, Bombay and Research Division, India Research Lab Graphical models for part of speech tagging Different Models for POS tagging HMM Maximum Entropy Markov Models Conditional

More information

Natural Language Processing SoSe Language Modelling. (based on the slides of Dr. Saeedeh Momtazi)

Natural Language Processing SoSe Language Modelling. (based on the slides of Dr. Saeedeh Momtazi) Natural Language Processing SoSe 2015 Language Modelling Dr. Mariana Neves April 20th, 2015 (based on the slides of Dr. Saeedeh Momtazi) Outline 2 Motivation Estimation Evaluation Smoothing Outline 3 Motivation

More information

Theory of Alignment Generators and Applications to Statistical Machine Translation

Theory of Alignment Generators and Applications to Statistical Machine Translation Theory of Alignment Generators and Applications to Statistical Machine Translation Hemanta K Maji Raghavendra Udupa U IBM India Research Laboratory, New Delhi {hemantkm, uraghave}@inibmcom Abstract Viterbi

More information

Bayesian Learning of Non-compositional Phrases with Synchronous Parsing

Bayesian Learning of Non-compositional Phrases with Synchronous Parsing Bayesian Learning of Non-compositional Phrases with Synchronous Parsing Hao Zhang Computer Science Department University of Rochester Rochester, NY 14627 zhanghao@cs.rochester.edu Chris Quirk Microsoft

More information

Statistical NLP Spring Corpus-Based MT

Statistical NLP Spring Corpus-Based MT Statistical NLP Spring 2010 Lecture 17: Word / Phrase MT Dan Klein UC Berkeley Corpus-Based MT Modeling correspondences between languages Sentence-aligned parallel corpus: Yo lo haré mañana I will do it

More information

Corpus-Based MT. Statistical NLP Spring Unsupervised Word Alignment. Alignment Error Rate. IBM Models 1/2. Problems with Model 1

Corpus-Based MT. Statistical NLP Spring Unsupervised Word Alignment. Alignment Error Rate. IBM Models 1/2. Problems with Model 1 Statistical NLP Spring 2010 Corpus-Based MT Modeling correspondences between languages Sentence-aligned parallel corpus: Yo lo haré mañana I will do it tomorrow Hasta pronto See you soon Hasta pronto See

More information

arxiv: v1 [cs.cl] 21 May 2017

arxiv: v1 [cs.cl] 21 May 2017 Spelling Correction as a Foreign Language Yingbo Zhou yingbzhou@ebay.com Utkarsh Porwal uporwal@ebay.com Roberto Konow rkonow@ebay.com arxiv:1705.07371v1 [cs.cl] 21 May 2017 Abstract In this paper, we

More information

Using Conservative Estimation for Conditional Probability instead of Ignoring Infrequent Case

Using Conservative Estimation for Conditional Probability instead of Ignoring Infrequent Case Using Conservative Estimation for Conditional Probability instead of Ignoring Infrequent Case Masato Kikuchi, Eiko Yamamoto, Mitsuo Yoshida, Masayuki Okabe, Kyoji Umemura Department of Computer Science

More information

Lecture 13: More uses of Language Models

Lecture 13: More uses of Language Models Lecture 13: More uses of Language Models William Webber (william@williamwebber.com) COMP90042, 2014, Semester 1, Lecture 13 What we ll learn in this lecture Comparing documents, corpora using LM approaches

More information

A Study for Evaluating the Importance of Various Parts of Speech (POS) for Information Retrieval (IR)

A Study for Evaluating the Importance of Various Parts of Speech (POS) for Information Retrieval (IR) A Study for Evaluating the Importance of Various Parts of Speech (POS) for Information Retrieval (IR) Chirag Shah Dept. of CSE IIT Bombay, Powai Mumbai - 400 076, Maharashtra, India. Email: chirag@cse.iitb.ac.in

More information

Learning to translate with neural networks. Michael Auli

Learning to translate with neural networks. Michael Auli Learning to translate with neural networks Michael Auli 1 Neural networks for text processing Similar words near each other France Spain dog cat Neural networks for text processing Similar words near each

More information

PROBABILISTIC LATENT SEMANTIC ANALYSIS

PROBABILISTIC LATENT SEMANTIC ANALYSIS PROBABILISTIC LATENT SEMANTIC ANALYSIS Lingjia Deng Revised from slides of Shuguang Wang Outline Review of previous notes PCA/SVD HITS Latent Semantic Analysis Probabilistic Latent Semantic Analysis Applications

More information

COMS 4705 (Fall 2011) Machine Translation Part II

COMS 4705 (Fall 2011) Machine Translation Part II COMS 4705 (Fall 2011) Machine Translation Part II 1 Recap: The Noisy Channel Model Goal: translation system from French to English Have a model p(e f) which estimates conditional probability of any English

More information

Machine Translation: Examples. Statistical NLP Spring Levels of Transfer. Corpus-Based MT. World-Level MT: Examples

Machine Translation: Examples. Statistical NLP Spring Levels of Transfer. Corpus-Based MT. World-Level MT: Examples Statistical NLP Spring 2009 Machine Translation: Examples Lecture 17: Word Alignment Dan Klein UC Berkeley Corpus-Based MT Levels of Transfer Modeling correspondences between languages Sentence-aligned

More information

Hidden Markov Modelling

Hidden Markov Modelling Hidden Markov Modelling Introduction Problem formulation Forward-Backward algorithm Viterbi search Baum-Welch parameter estimation Other considerations Multiple observation sequences Phone-based models

More information

Automatic Speech Recognition and Statistical Machine Translation under Uncertainty

Automatic Speech Recognition and Statistical Machine Translation under Uncertainty Outlines Automatic Speech Recognition and Statistical Machine Translation under Uncertainty Lambert Mathias Advisor: Prof. William Byrne Thesis Committee: Prof. Gerard Meyer, Prof. Trac Tran and Prof.

More information

HMM Expanded to Multiple Interleaved Chains as a Model for Word Sense Disambiguation

HMM Expanded to Multiple Interleaved Chains as a Model for Word Sense Disambiguation HMM Expanded to Multiple Interleaved Chains as a Model for Word Sense Disambiguation Denis Turdakov and Dmitry Lizorkin Institute for System Programming of the Russian Academy of Sciences, 25 Solzhenitsina

More information

Generalized Stack Decoding Algorithms for Statistical Machine Translation

Generalized Stack Decoding Algorithms for Statistical Machine Translation Generalized Stack Decoding Algorithms for Statistical Machine Translation Daniel Ortiz Martínez Inst. Tecnológico de Informática Univ. Politécnica de Valencia 4607 Valencia, Spain dortiz@iti.upv.es Ismael

More information

Variational Decoding for Statistical Machine Translation

Variational Decoding for Statistical Machine Translation Variational Decoding for Statistical Machine Translation Zhifei Li, Jason Eisner, and Sanjeev Khudanpur Center for Language and Speech Processing Computer Science Department Johns Hopkins University 1

More information

N-gram Language Modeling Tutorial

N-gram Language Modeling Tutorial N-gram Language Modeling Tutorial Dustin Hillard and Sarah Petersen Lecture notes courtesy of Prof. Mari Ostendorf Outline: Statistical Language Model (LM) Basics n-gram models Class LMs Cache LMs Mixtures

More information

Automatic Evaluation of Machine Translation Quality Using Longest Common Subsequence and Skip-Bigram Statistics

Automatic Evaluation of Machine Translation Quality Using Longest Common Subsequence and Skip-Bigram Statistics Automatic Evaluation of Machine Translation Quality Using Longest Common Subsequence and Skip-Bigram Statistics Chin-Yew Lin and Franz Josef Och Information Sciences Institute University of Southern California

More information

Determining Word Sense Dominance Using a Thesaurus

Determining Word Sense Dominance Using a Thesaurus Determining Word Sense Dominance Using a Thesaurus Saif Mohammad and Graeme Hirst Department of Computer Science University of Toronto EACL, Trento, Italy (5th April, 2006) Copyright cfl2006, Saif Mohammad

More information

TALP Phrase-Based System and TALP System Combination for the IWSLT 2006 IWSLT 2006, Kyoto

TALP Phrase-Based System and TALP System Combination for the IWSLT 2006 IWSLT 2006, Kyoto TALP Phrase-Based System and TALP System Combination for the IWSLT 2006 IWSLT 2006, Kyoto Marta R. Costa-jussà, Josep M. Crego, Adrià de Gispert, Patrik Lambert, Maxim Khalilov, José A.R. Fonollosa, José

More information

Language Models, Smoothing, and IDF Weighting

Language Models, Smoothing, and IDF Weighting Language Models, Smoothing, and IDF Weighting Najeeb Abdulmutalib, Norbert Fuhr University of Duisburg-Essen, Germany {najeeb fuhr}@is.inf.uni-due.de Abstract In this paper, we investigate the relationship

More information

Word Alignment. Chris Dyer, Carnegie Mellon University

Word Alignment. Chris Dyer, Carnegie Mellon University Word Alignment Chris Dyer, Carnegie Mellon University John ate an apple John hat einen Apfel gegessen John ate an apple John hat einen Apfel gegessen Outline Modeling translation with probabilistic models

More information

Lecture 12: Algorithms for HMMs

Lecture 12: Algorithms for HMMs Lecture 12: Algorithms for HMMs Nathan Schneider (some slides from Sharon Goldwater; thanks to Jonathan May for bug fixes) ENLP 17 October 2016 updated 9 September 2017 Recap: tagging POS tagging is a

More information

Statistical NLP Spring HW2: PNP Classification

Statistical NLP Spring HW2: PNP Classification Statistical NLP Spring 2010 Lecture 16: Word Alignment Dan Klein UC Berkeley HW2: PNP Classification Overall: good work! Top results: 88.1: Matthew Can (word/phrase pre/suffixes) 88.1: Kurtis Heimerl (positional

More information

HW2: PNP Classification. Statistical NLP Spring Levels of Transfer. Phrasal / Syntactic MT: Examples. Lecture 16: Word Alignment

HW2: PNP Classification. Statistical NLP Spring Levels of Transfer. Phrasal / Syntactic MT: Examples. Lecture 16: Word Alignment Statistical NLP Spring 2010 Lecture 16: Word Alignment Dan Klein UC Berkeley HW2: PNP Classification Overall: good work! Top results: 88.1: Matthew Can (word/phrase pre/suffixes) 88.1: Kurtis Heimerl (positional

More information

This kind of reordering is beyond the power of finite transducers, but a synchronous CFG can do this.

This kind of reordering is beyond the power of finite transducers, but a synchronous CFG can do this. Chapter 12 Synchronous CFGs Synchronous context-free grammars are a generalization of CFGs that generate pairs of related strings instead of single strings. They are useful in many situations where one

More information

Topic Models and Applications to Short Documents

Topic Models and Applications to Short Documents Topic Models and Applications to Short Documents Dieu-Thu Le Email: dieuthu.le@unitn.it Trento University April 6, 2011 1 / 43 Outline Introduction Latent Dirichlet Allocation Gibbs Sampling Short Text

More information

Advanced Natural Language Processing Syntactic Parsing

Advanced Natural Language Processing Syntactic Parsing Advanced Natural Language Processing Syntactic Parsing Alicia Ageno ageno@cs.upc.edu Universitat Politècnica de Catalunya NLP statistical parsing 1 Parsing Review Statistical Parsing SCFG Inside Algorithm

More information

Automatically Evaluating Text Coherence using Anaphora and Coreference Resolution

Automatically Evaluating Text Coherence using Anaphora and Coreference Resolution 1 1 Barzilay 1) Automatically Evaluating Text Coherence using Anaphora and Coreference Resolution Ryu Iida 1 and Takenobu Tokunaga 1 We propose a metric for automatically evaluating discourse coherence

More information

Language Models. Hongning Wang

Language Models. Hongning Wang Language Models Hongning Wang CS@UVa Notion of Relevance Relevance (Rep(q), Rep(d)) Similarity P(r1 q,d) r {0,1} Probability of Relevance P(d q) or P(q d) Probabilistic inference Different rep & similarity

More information

Improved Decipherment of Homophonic Ciphers

Improved Decipherment of Homophonic Ciphers Improved Decipherment of Homophonic Ciphers Malte Nuhn and Julian Schamper and Hermann Ney Human Language Technology and Pattern Recognition Computer Science Department, RWTH Aachen University, Aachen,

More information

Statistical Machine Translation of Natural Languages

Statistical Machine Translation of Natural Languages 1/37 Statistical Machine Translation of Natural Languages Rule Extraction and Training Probabilities Matthias Büchse, Toni Dietze, Johannes Osterholzer, Torsten Stüber, Heiko Vogler Technische Universität

More information

Name: Matriculation Number: Tutorial Group: A B C D E

Name: Matriculation Number: Tutorial Group: A B C D E Name: Matriculation Number: Tutorial Group: A B C D E Question: 1 (5 Points) 2 (6 Points) 3 (5 Points) 4 (5 Points) Total (21 points) Score: General instructions: The written test contains 4 questions

More information

Annotation tasks and solutions in CLARIN-PL

Annotation tasks and solutions in CLARIN-PL Annotation tasks and solutions in CLARIN-PL Marcin Oleksy, Ewa Rudnicka Wrocław University of Technology marcin.oleksy@pwr.edu.pl ewa.rudnicka@pwr.edu.pl CLARIN ERIC Common Language Resources and Technology

More information

Domain Adaptation for Word Sense Disambiguation under the Problem of Covariate Shift

Domain Adaptation for Word Sense Disambiguation under the Problem of Covariate Shift Domain Adaptation for Word Sense Disambiguation under the Problem of Covariate Shift HIRONORI KIKUCHI 1,a) HIROYUKI SHINNOU 1,b) Abstract: Word sense disambiguation(wsd) is the task of identifying the

More information

Exploitation of Machine Learning Techniques in Modelling Phrase Movements for Machine Translation

Exploitation of Machine Learning Techniques in Modelling Phrase Movements for Machine Translation Journal of Machine Learning Research 12 (2011) 1-30 Submitted 5/10; Revised 11/10; Published 1/11 Exploitation of Machine Learning Techniques in Modelling Phrase Movements for Machine Translation Yizhao

More information

ORANGE: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation

ORANGE: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation ORANGE: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation Chin-Yew Lin and Franz Josef Och Information Sciences Institute University of Southern California 4676 Admiralty Way

More information