Statistical Machine Translation of Natural Languages
|
|
- Dylan Hines
- 5 years ago
- Views:
Transcription
1 1/26 Statistical Machine Translation of Natural Languages Heiko Vogler Technische Universität Dresden Germany Graduiertenkolleg Quantitative Logics and Automata Dresden, November, 2012
2 1/26 Weighted Tree Automata and Weighted Tree Transducers can help in Statistical Machine Translation of Natural Languages Heiko Vogler Technische Universität Dresden Germany Graduiertenkolleg Quantitative Logics and Automata Dresden, November, 2012
3 2/26 outline of the talk: Statistical machine translation
4 2/26 outline of the talk: Statistical machine translation Modeling with wta and wtt
5 2/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling
6 2/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary
7 3/26 outline of the talk: Statistical machine translation no survey! Modeling with wta and wtt Using automata theoretic results to improve modeling Summary
8 4/26 given: source language SL target language TL find: translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel.
9 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel.
10 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation
11 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation modeling - training - evaluation
12 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation modeling - training - evaluation
13 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation modeling - training - evaluation assumptions modeling H hypothesis space assumptions: mental work, experience, no data hypothesis space: H {h h : SL TL}
14 log-linear modeling: 5/26
15 5/26 log-linear modeling: set Y of correspondence structures [Liang et al. 06] π SL : Y SL π TL : Y TL
16 5/26 log-linear modeling: set Y of correspondence structures [Liang et al. 06] π SL : Y SL π TL : Y TL hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s )
17 5/26 log-linear modeling: set Y of correspondence structures [Liang et al. 06] π SL : Y SL π TL : Y TL hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s ) λ 1 Φ(y) λ m Φ(y) m
18 6/26 hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s ) λ 1 Φ(y) λ m Φ(y) m
19 6/26 hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s ) λ 1 Φ(y) λ m Φ(y) m here: m = 2 Φ(y) = (h TM (y), h LM (y)) h TM : translation model h LM : language model
20 7/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary
21 8/26 first assumption: SL and TL are the yields of weighted recognizable tree languages
22 8/26 first assumption: SL and TL are the yields of weighted recognizable tree languages weighted tree language: L : T Σ R
23 8/26 first assumption: SL and TL are the yields of weighted recognizable tree languages weighted tree language: L : T Σ R L is recognizable: if there is a wta A which recognizes (computes) L
24 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states)
25 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ)
26 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ)
27 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ
28 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ
29 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ
30 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ
31 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ
32 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ
33 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) weighted tree language recognized by A: σ: label of ξ at w k: rank of σ L A : T Σ R, L A (ξ) = max r R A (ξ) wt(r)
34 10/26 second assumption: translation from SL and TL is specified by a weighted tree transducer [Yamada, Knight 01] translation from English to Japanese
35 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols
36 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights linear, nondeleting in x 1,..., x k
37 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights
38 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights (leftmost) derivation: d = ρ 1 ρ n
39 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights (leftmost) derivation: weight of a derivation d: d = ρ 1 ρ n wt(d) = n i=1 wt(ρ i)
40 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights (leftmost) derivation: weight of a derivation d: d = ρ 1 ρ n wt(d) = n i=1 wt(ρ i) weighted tree transformation computed by M: τ M : T Σ T Σ R, τ M (ξ 1, ξ 2 ) = max D M : set of all derivations d D M : π(d)=(ξ 1,ξ 2) wt(d)
41 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06]
42 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06] wtt M as translation model; d D M
43 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06] wtt M as translation model; d D M wta A as language model; r R A
44 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06] wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))} π SL (d, r) = yield(first(d)) π TL (d, r) = yield(last(d))
45 13/26 wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))}
46 13/26 wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))} here: m = ( 2 ) Φ(d, r) = log wt M (d), log wt A (r)
47 13/26 wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))} here: m = ( 2 ) Φ(d, r) = log wt M (d), log wt A (r) H = {h λ,m,a λ R 2 0, wtt M, wta A} h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s
48 14/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary Weight exponentiation Output product
49 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s
50 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s
51 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s Lemma: Let A be a wta and λ R 0.
52 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s Lemma: Let A be a wta and λ R 0. There is a wta A s.t. Q A = Q A and wt A (r) = wt A (r) λ for every r.
53 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s Lemma: Let A be a wta and λ R 0. There is a wta A s.t. Q A = Q A and wt A (r) = wt A (r) λ for every r. = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL (d,r)=s )
54 16/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary Weight exponentiation Output product
55 Let τ : T Σ T Σ R and L : T Σ R 17/26
56 17/26 Let τ : T Σ T Σ R and L : T Σ R output product of τ and L: τ L : T Σ T Σ R (τ L)(ξ, ζ) τ(ξ, ζ) L(ζ)
57 17/26 Let τ : T Σ T Σ R and L : T Σ R output product of τ and L: τ L : T Σ T Σ R (τ L)(ξ, ζ) τ(ξ, ζ) L(ζ) input product of L and τ: L τ : T Σ T Σ R (L τ)(ξ, ζ) L(ξ) τ(ξ, ζ)
58 Theorem [Maletti 06]: Let M wtt and A wta. 18/26
59 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A 18/26
60 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02]
61 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02] rule of M:
62 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02] rule of M: states of A: p, p 1, p 2
63 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02] rule of M: states of A: p, p 1, p 2 rule of M A:
64 19/26
65 19/26
66 19/26
67 ϕ : D M {(d, p) d D M, p R partial A (d)} bijection wt(d ) = wt(d) max r completion( p) wt(r) 19/26
68 20/26 recall: h λ,m,a : SL TL s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s )
69 20/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s )
70 20/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s ) = π TL (argmax d DM wt M (d) p max p R A π SL(d)=s (d) max r compl.( p) wt A (r) )
71 20/26 h λ,m,a : SL TL s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s ) = π TL (argmax d DM wt M (d) p max p R A π SL(d)=s (d) max r compl.( p) wt A (r) = π TL argmax d DM wt M (d) max r compl.( p) wt A (r) π SL(d)=s p R p A (d) ) )
72 20/26 h λ,m,a : SL TL s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s ) = π TL (argmax d DM wt M (d) p max p R A π SL(d)=s (d) max r compl.( p) wt A (r) = π TL argmax d DM wt M (d) max r compl.( p) wt A (r) π SL(d)=s p R p A (d) = π TL (argmax d DM A wt M A (d) π SL(d)=s ) ) )
73 21/26 recall: Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A
74 21/26 recall: Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A generalization to mildly context-sensitive languages Theorem [Büchse, Nederhof, V. 11]: Let M synchronized tree-adjoining grammar (STAG) and A wta. There is an STAG M A such that: τ M A = τ M L A
75 21/26 recall: Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A generalization to mildly context-sensitive languages Theorem [Büchse, Nederhof, V. 11]: Let M synchronized tree-adjoining grammar (STAG) and A wta. There is an STAG M A such that: τ M A = τ M L A Theorem [Nederhof, V. 12]: Let M synchronized context-free tree grammar (SCFTG) and A wta. There is an SCFTG M A such that: τ M A = τ M L A
76 22/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary
77 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M
78 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s
79 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s )
80 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s (output product) = π TL (argmax d DM A : π SL(d)=s wt M A (d) ) )
81 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s (output product) = π TL (argmax d DM A : wt M A (d) π SL(d)=s ) (B,S,P; input product) = π TL (argmax d DAs (M A) wt As (M A )(d) ) )
82 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s (output product) = π TL (argmax d DM A : wt M A (d) π SL(d)=s ) (B,S,P; input product) = π TL (argmax d DAs (M A) wt As (M A )(d) = π TL ( Knuth(As (M A ))) ) ) )
83 impression: SMT is playing with formulas 24/26
84 24/26 impression: SMT is playing with formulas but: SMT is an engineering task!
85 24/26 impression: SMT is playing with formulas but: SMT is an engineering task! weighted tree automata and weighted tree transducers can help in modeling statistical machine translation of natural languages
86 25/26 References: [Baker 79] Composition of top-down and bottom-up tree transductions [Bar-Hillel, Shamir, Perles 61] On formal properties of simple phrase structure grammars [Büchse, Nederhof, V. 11] Tree Parsing with Synchronous Tree-Adjoining Grammars [Engelfriet, Fülöp, V. 02] Bottom-up and Top-down Tree Series Transformations [Fülöp, V. 09] Weighted tree automata and tree transducers [Knight et al ]... [Knuth 77] A generalization of Dijkstra s algorithm [Lopez 08] Statistical Machine Translation [Liang, Bouchard-Côté, Klein, Taskar 06] An End-to-End Discriminative Approach to Machine Translation [Maletti 06] Compositions of Tree Series Transformations [Maletti, Satta 09] Parsing Algorithms based on Tree Automata [Nederhof, V. 12] Synchronous Context-Free Tree Grammars [Thatcher 67] Characterizing derivation trees of context-free grammars through a generalization of finite automata theory [Yamada, Knight 01] A syntax-based statistical translation model
87 26/26 [Knight et al ] [Charniak, Knight, Yamada 03] Syntax-based language models for statistical machine translation [Galley, Hopkins, Knight, Marcu 04] What s in a translation rule? [Graehl, Knight 04] Training tree transducers [Graehl, Knight 05] An overview of probabilistic tree transducers for natural language processing
88 26/26 [Knight et al ] [Charniak, Knight, Yamada 03] Syntax-based language models for statistical machine translation [Galley, Hopkins, Knight, Marcu 04] What s in a translation rule? [Graehl, Knight 04] Training tree transducers [Graehl, Knight 05] An overview of probabilistic tree transducers for natural language processing [Maletti 11] Survey: Weighted Extended Top-down Tree Transducers Part III: Applications in Machine Translation
89 26/26 [Knight et al ] [Charniak, Knight, Yamada 03] Syntax-based language models for statistical machine translation [Galley, Hopkins, Knight, Marcu 04] What s in a translation rule? [Graehl, Knight 04] Training tree transducers [Graehl, Knight 05] An overview of probabilistic tree transducers for natural language processing [Maletti 11] Survey: Weighted Extended Top-down Tree Transducers Part III: Applications in Machine Translation Thank you!
Statistical Machine Translation of Natural Languages
1/37 Statistical Machine Translation of Natural Languages Rule Extraction and Training Probabilities Matthias Büchse, Toni Dietze, Johannes Osterholzer, Torsten Stüber, Heiko Vogler Technische Universität
More informationApplications of Tree Automata Theory Lecture VI: Back to Machine Translation
Applications of Tree Automata Theory Lecture VI: Back to Machine Translation Andreas Maletti Institute of Computer Science Universität Leipzig, Germany on leave from: Institute for Natural Language Processing
More informationLecture 7: Introduction to syntax-based MT
Lecture 7: Introduction to syntax-based MT Andreas Maletti Statistical Machine Translation Stuttgart December 16, 2011 SMT VII A. Maletti 1 Lecture 7 Goals Overview Tree substitution grammars (tree automata)
More informationHierarchies of Tree Series TransducersRevisited 1
Hierarchies of Tree Series TransducersRevisited 1 Andreas Maletti 2 Technische Universität Dresden Fakultät Informatik June 27, 2006 1 Financially supported by the Friends and Benefactors of TU Dresden
More informationThe Power of Tree Series Transducers
The Power of Tree Series Transducers Andreas Maletti 1 Technische Universität Dresden Fakultät Informatik June 15, 2006 1 Research funded by German Research Foundation (DFG GK 334) Andreas Maletti (TU
More informationCompositions of Tree Series Transformations
Compositions of Tree Series Transformations Andreas Maletti a Technische Universität Dresden Fakultät Informatik D 01062 Dresden, Germany maletti@tcs.inf.tu-dresden.de December 03, 2004 1. Motivation 2.
More informationLecture 9: Decoding. Andreas Maletti. Stuttgart January 20, Statistical Machine Translation. SMT VIII A. Maletti 1
Lecture 9: Decoding Andreas Maletti Statistical Machine Translation Stuttgart January 20, 2012 SMT VIII A. Maletti 1 Lecture 9 Last time Synchronous grammars (tree transducers) Rule extraction Weight training
More informationCompositions of Bottom-Up Tree Series Transformations
Compositions of Bottom-Up Tree Series Transformations Andreas Maletti a Technische Universität Dresden Fakultät Informatik D 01062 Dresden, Germany maletti@tcs.inf.tu-dresden.de May 17, 2005 1. Motivation
More informationApplications of Tree Automata Theory Lecture V: Theory of Tree Transducers
Applications of Tree Automata Theory Lecture V: Theory of Tree Transducers Andreas Maletti Institute of Computer Science Universität Leipzig, Germany on leave from: Institute for Natural Language Processing
More informationHow to train your multi bottom-up tree transducer
How to train your multi bottom-up tree transducer Andreas Maletti Universität tuttgart Institute for Natural Language Processing tuttgart, Germany andreas.maletti@ims.uni-stuttgart.de Portland, OR June
More informationMinimization of Weighted Automata
Minimization of Weighted Automata Andreas Maletti Universitat Rovira i Virgili Tarragona, Spain Wrocław May 19, 2010 Minimization of Weighted Automata Andreas Maletti 1 In Collaboration with ZOLTÁN ÉSIK,
More informationComposition Closure of ε-free Linear Extended Top-down Tree Transducers
Composition Closure of ε-free Linear Extended Top-down Tree Transducers Zoltán Fülöp 1, and Andreas Maletti 2, 1 Department of Foundations of Computer Science, University of Szeged Árpád tér 2, H-6720
More informationRandom Generation of Nondeterministic Tree Automata
Random Generation of Nondeterministic Tree Automata Thomas Hanneforth 1 and Andreas Maletti 2 and Daniel Quernheim 2 1 Department of Linguistics University of Potsdam, Germany 2 Institute for Natural Language
More informationPreservation of Recognizability for Weighted Linear Extended Top-Down Tree Transducers
Preservation of Recognizability for Weighted Linear Extended Top-Down Tree Transducers Nina eemann and Daniel Quernheim and Fabienne Braune and Andreas Maletti University of tuttgart, Institute for Natural
More informationA Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister
A Syntax-based Statistical Machine Translation Model Alexander Friedl, Georg Teichtmeister 4.12.2006 Introduction The model Experiment Conclusion Statistical Translation Model (STM): - mathematical model
More informationCompositions of Bottom-Up Tree Series Transformations
Compositions of Bottom-Up Tree Series Transformations Andreas Maletti a Technische Universität Dresden Fakultät Informatik D 01062 Dresden, Germany maletti@tcs.inf.tu-dresden.de May 27, 2005 1. Motivation
More informationStructure and Complexity of Grammar-Based Machine Translation
Structure and of Grammar-Based Machine Translation University of Padua, Italy New York, June 9th, 2006 1 2 Synchronous context-free grammars Definitions Computational problems 3 problem SCFG projection
More informationCompositions of Top-down Tree Transducers with ε-rules
Compositions of Top-down Tree Transducers with ε-rules Andreas Maletti 1, and Heiko Vogler 2 1 Universitat Rovira i Virgili, Departament de Filologies Romàniques Avinguda de Catalunya 35, 43002 Tarragona,
More informationRELATING TREE SERIES TRANSDUCERS AND WEIGHTED TREE AUTOMATA
International Journal of Foundations of Computer Science Vol. 16, No. 4 (2005) 723 741 c World Scientific Publishing Company RELATING TREE SERIES TRANSDUCERS AND WEIGHTED TREE AUTOMATA ANDREAS MALETTI
More informationPure and O-Substitution
International Journal of Foundations of Computer Science c World Scientific Publishing Company Pure and O-Substitution Andreas Maletti Department of Computer Science, Technische Universität Dresden 01062
More informationSyntax-Directed Translations and Quasi-alphabetic Tree Bimorphisms Revisited
Syntax-Directed Translations and Quasi-alphabetic Tree Bimorphisms Revisited Andreas Maletti and C t lin Ionuµ Tîrn uc Universitat Rovira i Virgili Departament de Filologies Romàniques Av. Catalunya 35,
More informationAspects of Tree-Based Statistical Machine Translation
Aspects of Tree-Based Statistical Machine Translation Marcello Federico Human Language Technology FBK 2014 Outline Tree-based translation models: Synchronous context free grammars Hierarchical phrase-based
More informationThe Power of Weighted Regularity-Preserving Multi Bottom-up Tree Transducers
International Journal of Foundations of Computer cience c World cientific Publishing Company The Power of Weighted Regularity-Preserving Multi Bottom-up Tree Transducers ANDREA MALETTI Universität Leipzig,
More informationMusic as a Formal Language
Music as a Formal Language Finite-State Automata and Pd Bryan Jurish moocow@ling.uni-potsdam.de Universität Potsdam, Institut für Linguistik, Potsdam, Germany pd convention 2004 / Jurish / Music as a formal
More informationWeighted Regular Tree Grammars with Storage
Discrete Mathematics and Theoretical Computer Science DMTCS vol. 20:1, 2018, #26 Weighted Regular Tree Grammars with Storage Zoltán Fülöp 1 Luisa Herrmann 2 Heiko Vogler 2 arxiv:1705.06681v5 [cs.fl] 2
More informationWhat we have done so far
What we have done so far DFAs and regular languages NFAs and their equivalence to DFAs Regular expressions. Regular expressions capture exactly regular languages: Construct a NFA from a regular expression.
More informationEinführung in die Computerlinguistik
Einführung in die Computerlinguistik Context-Free Grammars (CFG) Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 22 CFG (1) Example: Grammar G telescope : Productions: S NP VP NP
More informationComputational Models - Lecture 4 1
Computational Models - Lecture 4 1 Handout Mode Iftach Haitner. Tel Aviv University. November 21, 2016 1 Based on frames by Benny Chor, Tel Aviv University, modifying frames by Maurice Herlihy, Brown University.
More informationThis kind of reordering is beyond the power of finite transducers, but a synchronous CFG can do this.
Chapter 12 Synchronous CFGs Synchronous context-free grammars are a generalization of CFGs that generate pairs of related strings instead of single strings. They are useful in many situations where one
More informationAlgorithms for Syntax-Aware Statistical Machine Translation
Algorithms for Syntax-Aware Statistical Machine Translation I. Dan Melamed, Wei Wang and Ben Wellington ew York University Syntax-Aware Statistical MT Statistical involves machine learning (ML) seems crucial
More informationPrefix Probability for Probabilistic Synchronous Context-Free Grammars
Prefix Probability for Probabilistic Synchronous Context-Free rammars Mark-Jan Nederhof School of Computer Science University of St Andrews North Haugh, St Andrews, Fife KY16 9SX United Kingdom markjan.nederhof@googlemail.com
More informationCross-Entropy and Estimation of Probabilistic Context-Free Grammars
Cross-Entropy and Estimation of Probabilistic Context-Free Grammars Anna Corazza Department of Physics University Federico II via Cinthia I-8026 Napoli, Italy corazza@na.infn.it Giorgio Satta Department
More informationBackward and Forward Bisimulation Minimisation of Tree Automata
Backward and Forward Bisimulation Minimisation of Tree Automata Johanna Högberg 1, Andreas Maletti 2, and Jonathan May 3 1 Dept. of Computing Science, Umeå University, S 90187 Umeå, Sweden johanna@cs.umu.se
More informationComputation Histories
208 Computation Histories The computation history for a Turing machine on an input is simply the sequence of configurations that the machine goes through as it processes the input. An accepting computation
More informationLinking Theorems for Tree Transducers
Linking Theorems for Tree Transducers Andreas Maletti maletti@ims.uni-stuttgart.de peyer October 1, 2015 Andreas Maletti Linking Theorems for MBOT Theorietag 2015 1 / 32 tatistical Machine Translation
More informationExtended Multi Bottom-Up Tree Transducers
Extended Multi Bottom-Up Tree Transducers Joost Engelfriet 1, Eric Lilin 2, and Andreas Maletti 3;? 1 Leiden Instite of Advanced Comper Science Leiden University, P.O. Box 9512, 2300 RA Leiden, The Netherlands
More informationKey words. tree transducer, natural language processing, hierarchy, composition closure
THE POWER OF EXTENDED TOP-DOWN TREE TRANSDUCERS ANDREAS MALETTI, JONATHAN GRAEHL, MARK HOPKINS, AND KEVIN KNIGHT Abstract. Extended top-down tree transducers (transducteurs généralisés descendants [Arnold,
More informationHarvard CS 121 and CSCI E-207 Lecture 10: Ambiguity, Pushdown Automata
Harvard CS 121 and CSCI E-207 Lecture 10: Ambiguity, Pushdown Automata Salil Vadhan October 4, 2012 Reading: Sipser, 2.2. Another example of a CFG (with proof) L = {x {a, b} : x has the same # of a s and
More informationWeighted tree automata and weighted logics
Theoretical Computer Science 366 (2006) 228 247 www.elsevier.com/locate/tcs Weighted tree automata and weighted logics Manfred Droste a,, Heiko Vogler b a Institute of Computer Science, Leipzig University,
More informationNew Results on Context-Free Tree Languages
Technische Universität Dresden New Results on Context-Free Tree Languages Dissertation zur Erlangung des akademischen Grades Doctor rerum naturalium (Dr. rer. nat.) vorgelegt an der Technischen Universität
More informationBisimulation Minimisation for Weighted Tree Automata
Bisimulation Minimisation for Weighted Tree Automata Johanna Högberg 1, Andreas Maletti 2, and Jonathan May 3 1 Department of Computing Science, Umeå University S 90187 Umeå, Sweden johanna@cs.umu.se 2
More informationFoundations of Informatics: a Bridging Course
Foundations of Informatics: a Bridging Course Week 3: Formal Languages and Semantics Thomas Noll Lehrstuhl für Informatik 2 RWTH Aachen University noll@cs.rwth-aachen.de http://www.b-it-center.de/wob/en/view/class211_id948.html
More informationCompositions of Extended Top-down Tree Transducers
Compositions of Extended Top-down Tree Transducers Andreas Maletti ;1 International Computer Science Institute 1947 Center Street, Suite 600, Berkeley, CA 94704, USA Abstract Unfortunately, the class of
More informationOn the Myhill-Nerode Theorem for Trees. Dexter Kozen y. Cornell University
On the Myhill-Nerode Theorem for Trees Dexter Kozen y Cornell University kozen@cs.cornell.edu The Myhill-Nerode Theorem as stated in [6] says that for a set R of strings over a nite alphabet, the following
More informationSyntactical analysis. Syntactical analysis. Syntactical analysis. Syntactical analysis
Context-free grammars Derivations Parse Trees Left-recursive grammars Top-down parsing non-recursive predictive parsers construction of parse tables Bottom-up parsing shift/reduce parsers LR parsers GLR
More informationComputability and Complexity
Computability and Complexity Lecture 5 Reductions Undecidable problems from language theory Linear bounded automata given by Jiri Srba Lecture 5 Computability and Complexity 1/14 Reduction Informal Definition
More informationBringing machine learning & compositional semantics together: central concepts
Bringing machine learning & compositional semantics together: central concepts https://githubcom/cgpotts/annualreview-complearning Chris Potts Stanford Linguistics CS 244U: Natural language understanding
More informationTheory of Languages and Automata
Theory of Languages and Automata Chapter 1- Regular Languages & Finite State Automaton Sharif University of Technology Finite State Automaton We begin with the simplest model of Computation, called finite
More informationCOM364 Automata Theory Lecture Note 2 - Nondeterminism
COM364 Automata Theory Lecture Note 2 - Nondeterminism Kurtuluş Küllü March 2018 The FA we saw until now were deterministic FA (DFA) in the sense that for each state and input symbol there was exactly
More informationMaschinelle Sprachverarbeitung
Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other
More informationMaschinelle Sprachverarbeitung
Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other
More informationFuzzy Tree Transducers
Advances in Fuzzy Mathematics. ISSN 0973-533X Volume 11, Number 2 (2016), pp. 99-112 Research India Publications http://www.ripublication.com Fuzzy Tree Transducers S. R. Chaudhari and M. N. Joshi Department
More informationSpectral Learning for Non-Deterministic Dependency Parsing
Spectral Learning for Non-Deterministic Dependency Parsing Franco M. Luque 1 Ariadna Quattoni 2 Borja Balle 2 Xavier Carreras 2 1 Universidad Nacional de Córdoba 2 Universitat Politècnica de Catalunya
More informationRoger Levy Probabilistic Models in the Study of Language draft, October 2,
Roger Levy Probabilistic Models in the Study of Language draft, October 2, 2012 224 Chapter 10 Probabilistic Grammars 10.1 Outline HMMs PCFGs ptsgs and ptags Highlight: Zuidema et al., 2008, CogSci; Cohn
More informationExpectation Maximization (EM)
Expectation Maximization (EM) The EM algorithm is used to train models involving latent variables using training data in which the latent variables are not observed (unlabeled data). This is to be contrasted
More informationTuning as Linear Regression
Tuning as Linear Regression Marzieh Bazrafshan, Tagyoung Chung and Daniel Gildea Department of Computer Science University of Rochester Rochester, NY 14627 Abstract We propose a tuning method for statistical
More informationIntroduction and Motivation. Introduction and Motivation. Introduction to Computability. Introduction and Motivation. Theory. Lecture5: Context Free
ntroduction to Computability Theory Lecture5: Context Free Languages Prof. Amos sraeli 1 ntroduction and Motivation n our study of RL-s we Covered: 1. Motivation and definition of regular languages. 2.
More informationThis lecture covers Chapter 5 of HMU: Context-free Grammars
This lecture covers Chapter 5 of HMU: Context-free rammars (Context-free) rammars (Leftmost and Rightmost) Derivations Parse Trees An quivalence between Derivations and Parse Trees Ambiguity in rammars
More informationStructured Prediction Models via the Matrix-Tree Theorem
Structured Prediction Models via the Matrix-Tree Theorem Terry Koo Amir Globerson Xavier Carreras Michael Collins maestro@csail.mit.edu gamir@csail.mit.edu carreras@csail.mit.edu mcollins@csail.mit.edu
More informationThe Geometry of Statistical Machine Translation
The Geometry of Statistical Machine Translation Presented by Rory Waite 16th of December 2015 ntroduction Linear Models Convex Geometry The Minkowski Sum Projected MERT Conclusions ntroduction We provide
More informationMonadic Second Order Logic and Automata on Infinite Words: Büchi s Theorem
Monadic Second Order Logic and Automata on Infinite Words: Büchi s Theorem R. Dustin Wehr December 18, 2007 Büchi s theorem establishes the equivalence of the satisfiability relation for monadic second-order
More informationExam: Synchronous Grammars
Exam: ynchronous Grammars Duration: 3 hours Written documents are allowed. The numbers in front of questions are indicative of hardness or duration. ynchronous grammars consist of pairs of grammars whose
More informationCISC4090: Theory of Computation
CISC4090: Theory of Computation Chapter 2 Context-Free Languages Courtesy of Prof. Arthur G. Werschulz Fordham University Department of Computer and Information Sciences Spring, 2014 Overview In Chapter
More informationTree transformations and dependencies
Tree transformations and deendencies Andreas Maletti Universität Stuttgart Institute for Natural Language Processing Azenbergstraße 12, 70174 Stuttgart, Germany andreasmaletti@imsuni-stuttgartde Abstract
More informationCS243, Logic and Computation Nondeterministic finite automata
CS243, Prof. Alvarez NONDETERMINISTIC FINITE AUTOMATA (NFA) Prof. Sergio A. Alvarez http://www.cs.bc.edu/ alvarez/ Maloney Hall, room 569 alvarez@cs.bc.edu Computer Science Department voice: (67) 552-4333
More informationLanguages. Languages. An Example Grammar. Grammars. Suppose we have an alphabet V. Then we can write:
Languages A language is a set (usually infinite) of strings, also known as sentences Each string consists of a sequence of symbols taken from some alphabet An alphabet, V, is a finite set of symbols, e.g.
More informationChapter 14 (Partially) Unsupervised Parsing
Chapter 14 (Partially) Unsupervised Parsing The linguistically-motivated tree transformations we discussed previously are very effective, but when we move to a new language, we may have to come up with
More informationWeighted Tree-Walking Automata
Acta Cybernetica 19 (2009) 275 293. Weighted Tree-Walking Automata Zoltán Fülöp and Loránd Muzamel Abstract We define weighted tree-walking automata. We show that the class of tree series recognizable
More informationContext-free Grammars and Languages
Context-free Grammars and Languages COMP 455 002, Spring 2019 Jim Anderson (modified by Nathan Otterness) 1 Context-free Grammars Context-free grammars provide another way to specify languages. Example:
More informationNatural Language Processing : Probabilistic Context Free Grammars. Updated 5/09
Natural Language Processing : Probabilistic Context Free Grammars Updated 5/09 Motivation N-gram models and HMM Tagging only allowed us to process sentences linearly. However, even simple sentences require
More informationStatistical Machine Translation
Statistical Machine Translation -tree-based models (cont.)- Artem Sokolov Computerlinguistik Universität Heidelberg Sommersemester 2015 material from P. Koehn, S. Riezler, D. Altshuler Bottom-Up Decoding
More informationTree Transducers in Machine Translation
Tree Transducers in Machine Translation Andreas Maletti Universitat Rovira i Virgili Tarragona, pain andreas.maletti@urv.cat Jena August 23, 2010 Tree Transducers in MT A. Maletti 1 Machine translation
More informationQuasi-Synchronous Phrase Dependency Grammars for Machine Translation. lti
Quasi-Synchronous Phrase Dependency Grammars for Machine Translation Kevin Gimpel Noah A. Smith 1 Introduction MT using dependency grammars on phrases Phrases capture local reordering and idiomatic translations
More informationClarifications from last time. This Lecture. Last Lecture. CMSC 330: Organization of Programming Languages. Finite Automata.
CMSC 330: Organization of Programming Languages Last Lecture Languages Sets of strings Operations on languages Finite Automata Regular expressions Constants Operators Precedence CMSC 330 2 Clarifications
More informationChapter 1. Formal Definition and View. Lecture Formal Pushdown Automata on the 28th April 2009
Chapter 1 Formal and View Lecture on the 28th April 2009 Formal of PA Faculty of Information Technology Brno University of Technology 1.1 Aim of the Lecture 1 Define pushdown automaton in a formal way
More informationProbabilistic Parsing Strategies
Probabilistic Parsing Strategies Mark-Jan Nederhof Faculty of Arts University of Groningen P.O. Box 716 NL-9700 AS Groningen The Netherlands markjan@let.rug.nl Giorgio Satta Dept. of Information Engineering
More informationMildly Context-Sensitive Grammar Formalisms: Embedded Push-Down Automata
Mildly Context-Sensitive Grammar Formalisms: Embedded Push-Down Automata Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Sommersemester 2011 Intuition (1) For a language L, there is a TAG G with
More informationMyhill-Nerode Theorem for Recognizable Tree Series Revisited?
Myhill-Nerode Theorem for Recognizable Tree Series Revisited? Andreas Maletti?? International Comper Science Instite 1947 Center Street, Suite 600, Berkeley, CA 94704, USA Abstract. In this contribion
More informationA Polynomial Time Algorithm for Parsing with the Bounded Order Lambek Calculus
A Polynomial Time Algorithm for Parsing with the Bounded Order Lambek Calculus Timothy A. D. Fowler Department of Computer Science University of Toronto 10 King s College Rd., Toronto, ON, M5S 3G4, Canada
More informationShift-Reduce Word Reordering for Machine Translation
Shift-Reduce Word Reordering for Machine Translation Katsuhiko Hayashi, Katsuhito Sudoh, Hajime Tsukada, Jun Suzuki, Masaaki Nagata NTT Communication Science Laboratories, NTT Corporation 2-4 Hikaridai,
More informationIntroduction to Computers & Programming
16.070 Introduction to Computers & Programming Theory of computation: What is a computer? FSM, Automata Prof. Kristina Lundqvist Dept. of Aero/Astro, MIT Models of Computation What is a computer? If you
More informationNondeterministic Finite Automata
Nondeterministic Finite Automata Mahesh Viswanathan Introducing Nondeterminism Consider the machine shown in Figure. Like a DFA it has finitely many states and transitions labeled by symbols from an input
More informationC2.1 Regular Grammars
Theory of Computer Science March 22, 27 C2. Regular Languages: Finite Automata Theory of Computer Science C2. Regular Languages: Finite Automata Malte Helmert University of Basel March 22, 27 C2. Regular
More informationAutomata Theory - Quiz II (Solutions)
Automata Theory - Quiz II (Solutions) K. Subramani LCSEE, West Virginia University, Morgantown, WV {ksmani@csee.wvu.edu} 1 Problems 1. Induction: Let L denote the language of balanced strings over Σ =
More informationSCHEME FOR INTERNAL ASSESSMENT TEST 3
SCHEME FOR INTERNAL ASSESSMENT TEST 3 Max Marks: 40 Subject& Code: Automata Theory & Computability (15CS54) Sem: V ISE (A & B) Note: Answer any FIVE full questions, choosing one full question from each
More informationIntroduction to Automata
Introduction to Automata Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr 1 /
More informationCS:4330 Theory of Computation Spring Regular Languages. Finite Automata and Regular Expressions. Haniel Barbosa
CS:4330 Theory of Computation Spring 2018 Regular Languages Finite Automata and Regular Expressions Haniel Barbosa Readings for this lecture Chapter 1 of [Sipser 1996], 3rd edition. Sections 1.1 and 1.3.
More informationCSE 105 THEORY OF COMPUTATION
CSE 105 THEORY OF COMPUTATION Fall 2016 http://cseweb.ucsd.edu/classes/fa16/cse105-abc/ Today's learning goals Sipser Ch 3 Trace the computation of a Turing machine using its transition function and configurations.
More informationDM17. Beregnelighed. Jacob Aae Mikkelsen
DM17 Beregnelighed Jacob Aae Mikkelsen January 12, 2007 CONTENTS Contents 1 Introduction 2 1.1 Operations with languages...................... 2 2 Finite Automata 3 2.1 Regular expressions/languages....................
More informationGrammar formalisms Tree Adjoining Grammar: Formal Properties, Parsing. Part I. Formal Properties of TAG. Outline: Formal Properties of TAG
Grammar formalisms Tree Adjoining Grammar: Formal Properties, Parsing Laura Kallmeyer, Timm Lichte, Wolfgang Maier Universität Tübingen Part I Formal Properties of TAG 16.05.2007 und 21.05.2007 TAG Parsing
More informationReal-Time Systems. Lecture 15: The Universality Problem for TBA Dr. Bernd Westphal. Albert-Ludwigs-Universität Freiburg, Germany
Real-Time Systems Lecture 15: The Universality Problem for TBA 2013-06-26 15 2013-06-26 main Dr. Bernd Westphal Albert-Ludwigs-Universität Freiburg, Germany Contents & Goals Last Lecture: Extended Timed
More informationC2.1 Regular Grammars
Theory of Computer Science March 6, 26 C2. Regular Languages: Finite Automata Theory of Computer Science C2. Regular Languages: Finite Automata Malte Helmert University of Basel March 6, 26 C2. Regular
More informationIBM Model 1 for Machine Translation
IBM Model 1 for Machine Translation Micha Elsner March 28, 2014 2 Machine translation A key area of computational linguistics Bar-Hillel points out that human-like translation requires understanding of
More informationFORMAL LANGUAGES, AUTOMATA AND COMPUTATION
FORMAL LANGUAGES, AUTOMATA AND COMPUTATION DECIDABILITY ( LECTURE 15) SLIDES FOR 15-453 SPRING 2011 1 / 34 TURING MACHINES-SYNOPSIS The most general model of computation Computations of a TM are described
More informationMANFRED DROSTE AND WERNER KUICH
Logical Methods in Computer Science Vol. 14(1:21)2018, pp. 1 14 https://lmcs.episciences.org/ Submitted Jan. 31, 2017 Published Mar. 06, 2018 WEIGHTED ω-restricted ONE-COUNTER AUTOMATA Universität Leipzig,
More informationComputational Models - Lecture 1 1
Computational Models - Lecture 1 1 Handout Mode Ronitt Rubinfeld and Iftach Haitner. Tel Aviv University. February 29/ March 02, 2016 1 Based on frames by Benny Chor, Tel Aviv University, modifying frames
More informationCS21 Decidability and Tractability
CS21 Decidability and Tractability Lecture 8 January 24, 2018 Outline Turing Machines and variants multitape TMs nondeterministic TMs Church-Turing Thesis So far several models of computation finite automata
More informationLecture Notes on Inductive Definitions
Lecture Notes on Inductive Definitions 15-312: Foundations of Programming Languages Frank Pfenning Lecture 2 September 2, 2004 These supplementary notes review the notion of an inductive definition and
More informationTheory of Computation
Theory of Computation (Feodor F. Dragan) Department of Computer Science Kent State University Spring, 2018 Theory of Computation, Feodor F. Dragan, Kent State University 1 Before we go into details, what
More informationDecidable and undecidable languages
The Chinese University of Hong Kong Fall 2011 CSCI 3130: Formal languages and automata theory Decidable and undecidable languages Andrej Bogdanov http://www.cse.cuhk.edu.hk/~andrejb/csc3130 Problems about
More information