Statistical Machine Translation of Natural Languages

Size: px
Start display at page:

Download "Statistical Machine Translation of Natural Languages"

Transcription

1 1/26 Statistical Machine Translation of Natural Languages Heiko Vogler Technische Universität Dresden Germany Graduiertenkolleg Quantitative Logics and Automata Dresden, November, 2012

2 1/26 Weighted Tree Automata and Weighted Tree Transducers can help in Statistical Machine Translation of Natural Languages Heiko Vogler Technische Universität Dresden Germany Graduiertenkolleg Quantitative Logics and Automata Dresden, November, 2012

3 2/26 outline of the talk: Statistical machine translation

4 2/26 outline of the talk: Statistical machine translation Modeling with wta and wtt

5 2/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling

6 2/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary

7 3/26 outline of the talk: Statistical machine translation no survey! Modeling with wta and wtt Using automata theoretic results to improve modeling Summary

8 4/26 given: source language SL target language TL find: translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel.

9 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel.

10 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation

11 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation modeling - training - evaluation

12 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation modeling - training - evaluation

13 4/26 given: source language SL target language TL find: machine translation h : SL TL e.g. SL = English TL = German s = I saw the man with the telescope h(s) = Ich sah den Mann durch das Tel. machine translation statistical machine translation modeling - training - evaluation assumptions modeling H hypothesis space assumptions: mental work, experience, no data hypothesis space: H {h h : SL TL}

14 log-linear modeling: 5/26

15 5/26 log-linear modeling: set Y of correspondence structures [Liang et al. 06] π SL : Y SL π TL : Y TL

16 5/26 log-linear modeling: set Y of correspondence structures [Liang et al. 06] π SL : Y SL π TL : Y TL hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s )

17 5/26 log-linear modeling: set Y of correspondence structures [Liang et al. 06] π SL : Y SL π TL : Y TL hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s ) λ 1 Φ(y) λ m Φ(y) m

18 6/26 hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s ) λ 1 Φ(y) λ m Φ(y) m

19 6/26 hypothesis space: H = {h λ,φ λ R m 0, Φ : Y Rm } h λ,φ : SL TL s π TL (argmax y Y : λ Φ(y) π SL (y)=s ) λ 1 Φ(y) λ m Φ(y) m here: m = 2 Φ(y) = (h TM (y), h LM (y)) h TM : translation model h LM : language model

20 7/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary

21 8/26 first assumption: SL and TL are the yields of weighted recognizable tree languages

22 8/26 first assumption: SL and TL are the yields of weighted recognizable tree languages weighted tree language: L : T Σ R

23 8/26 first assumption: SL and TL are the yields of weighted recognizable tree languages weighted tree language: L : T Σ R L is recognizable: if there is a wta A which recognizes (computes) L

24 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states)

25 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ)

26 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ)

27 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ

28 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ

29 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ

30 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ

31 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ

32 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) σ: label of ξ at w k: rank of σ

33 9/26 weighted tree automaton (wta) A = (Q, Σ, δ, F ) Q finite set (states) Σ ranked alphabet (input symbols) δ = (δ σ σ Σ) δ σ : Q k Q R δ σ (q 1 q k, q) R F Q (final states) run on ξ T Σ : r : pos(ξ) Q set of runs on ξ: R A (ξ) weight of r: wt(r) = w pos(ξ) δ σ(r(w1) r(wk), r(w)) weighted tree language recognized by A: σ: label of ξ at w k: rank of σ L A : T Σ R, L A (ξ) = max r R A (ξ) wt(r)

34 10/26 second assumption: translation from SL and TL is specified by a weighted tree transducer [Yamada, Knight 01] translation from English to Japanese

35 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols

36 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights linear, nondeleting in x 1,..., x k

37 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights

38 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights (leftmost) derivation: d = ρ 1 ρ n

39 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights (leftmost) derivation: weight of a derivation d: d = ρ 1 ρ n wt(d) = n i=1 wt(ρ i)

40 11/26 weighted tree transducer (wtt) M = (Q, Σ, q 0, R) Q, Σ as for wta q 0 Q (initial state) Σ: input and output symbols R finite set of particular term rewrite rules with weights (leftmost) derivation: weight of a derivation d: d = ρ 1 ρ n wt(d) = n i=1 wt(ρ i) weighted tree transformation computed by M: τ M : T Σ T Σ R, τ M (ξ 1, ξ 2 ) = max D M : set of all derivations d D M : π(d)=(ξ 1,ξ 2) wt(d)

41 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06]

42 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06] wtt M as translation model; d D M

43 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06] wtt M as translation model; d D M wta A as language model; r R A

44 12/26 log-linear modeling with wtt and wta: set Y of correspondence structures [Liang et al. 06] wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))} π SL (d, r) = yield(first(d)) π TL (d, r) = yield(last(d))

45 13/26 wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))}

46 13/26 wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))} here: m = ( 2 ) Φ(d, r) = log wt M (d), log wt A (r)

47 13/26 wtt M as translation model; d D M wta A as language model; r R A Y = {(d, r) D M R A r R A (last(d))} here: m = ( 2 ) Φ(d, r) = log wt M (d), log wt A (r) H = {h λ,m,a λ R 2 0, wtt M, wta A} h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s

48 14/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary Weight exponentiation Output product

49 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s

50 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s

51 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s Lemma: Let A be a wta and λ R 0.

52 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s Lemma: Let A be a wta and λ R 0. There is a wta A s.t. Q A = Q A and wt A (r) = wt A (r) λ for every r.

53 15/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ 2 π SL (d,r)=s ) = π TL (argmax (d,r) Y : wt M (d) wt A (r) λ 2/λ 1 π SL (d,r)=s Lemma: Let A be a wta and λ R 0. There is a wta A s.t. Q A = Q A and wt A (r) = wt A (r) λ for every r. = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL (d,r)=s )

54 16/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary Weight exponentiation Output product

55 Let τ : T Σ T Σ R and L : T Σ R 17/26

56 17/26 Let τ : T Σ T Σ R and L : T Σ R output product of τ and L: τ L : T Σ T Σ R (τ L)(ξ, ζ) τ(ξ, ζ) L(ζ)

57 17/26 Let τ : T Σ T Σ R and L : T Σ R output product of τ and L: τ L : T Σ T Σ R (τ L)(ξ, ζ) τ(ξ, ζ) L(ζ) input product of L and τ: L τ : T Σ T Σ R (L τ)(ξ, ζ) L(ξ) τ(ξ, ζ)

58 Theorem [Maletti 06]: Let M wtt and A wta. 18/26

59 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A 18/26

60 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02]

61 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02] rule of M:

62 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02] rule of M: states of A: p, p 1, p 2

63 18/26 Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A Proof: [Baker 79, Engelfriet, Fülöp,V. 02] rule of M: states of A: p, p 1, p 2 rule of M A:

64 19/26

65 19/26

66 19/26

67 ϕ : D M {(d, p) d D M, p R partial A (d)} bijection wt(d ) = wt(d) max r completion( p) wt(r) 19/26

68 20/26 recall: h λ,m,a : SL TL s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s )

69 20/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s )

70 20/26 h λ,m,a : SL TL ) s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s ) = π TL (argmax d DM wt M (d) p max p R A π SL(d)=s (d) max r compl.( p) wt A (r) )

71 20/26 h λ,m,a : SL TL s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s ) = π TL (argmax d DM wt M (d) p max p R A π SL(d)=s (d) max r compl.( p) wt A (r) = π TL argmax d DM wt M (d) max r compl.( p) wt A (r) π SL(d)=s p R p A (d) ) )

72 20/26 h λ,m,a : SL TL s π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s = π TL (argmax d DM wt M (d) max r RA (last(d)) wt A (r) π SL(d)=s ) = π TL (argmax d DM wt M (d) p max p R A π SL(d)=s (d) max r compl.( p) wt A (r) = π TL argmax d DM wt M (d) max r compl.( p) wt A (r) π SL(d)=s p R p A (d) = π TL (argmax d DM A wt M A (d) π SL(d)=s ) ) )

73 21/26 recall: Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A

74 21/26 recall: Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A generalization to mildly context-sensitive languages Theorem [Büchse, Nederhof, V. 11]: Let M synchronized tree-adjoining grammar (STAG) and A wta. There is an STAG M A such that: τ M A = τ M L A

75 21/26 recall: Theorem [Maletti 06]: Let M wtt and A wta. There is a wtt M A such that: τ M A = τ M L A generalization to mildly context-sensitive languages Theorem [Büchse, Nederhof, V. 11]: Let M synchronized tree-adjoining grammar (STAG) and A wta. There is an STAG M A such that: τ M A = τ M L A Theorem [Nederhof, V. 12]: Let M synchronized context-free tree grammar (SCFTG) and A wta. There is an SCFTG M A such that: τ M A = τ M L A

76 22/26 outline of the talk: Statistical machine translation Modeling with wta and wtt Using automata theoretic results to improve modeling Summary

77 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M

78 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s

79 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s )

80 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s (output product) = π TL (argmax d DM A : π SL(d)=s wt M A (d) ) )

81 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s (output product) = π TL (argmax d DM A : wt M A (d) π SL(d)=s ) (B,S,P; input product) = π TL (argmax d DAs (M A) wt As (M A )(d) ) )

82 23/26 model for translation from SL to TL: model for TL: wta A sentence s of SL wtt M ) h λ,m,a (s) = π TL (argmax (d,r) Y : wt M (d) λ1 wt A (r) λ2 π SL(d,r)=s (weight exp.) = π TL (argmax (d,r) Y : wt M (d) wt A (r) π SL(d,r)=s (output product) = π TL (argmax d DM A : wt M A (d) π SL(d)=s ) (B,S,P; input product) = π TL (argmax d DAs (M A) wt As (M A )(d) = π TL ( Knuth(As (M A ))) ) ) )

83 impression: SMT is playing with formulas 24/26

84 24/26 impression: SMT is playing with formulas but: SMT is an engineering task!

85 24/26 impression: SMT is playing with formulas but: SMT is an engineering task! weighted tree automata and weighted tree transducers can help in modeling statistical machine translation of natural languages

86 25/26 References: [Baker 79] Composition of top-down and bottom-up tree transductions [Bar-Hillel, Shamir, Perles 61] On formal properties of simple phrase structure grammars [Büchse, Nederhof, V. 11] Tree Parsing with Synchronous Tree-Adjoining Grammars [Engelfriet, Fülöp, V. 02] Bottom-up and Top-down Tree Series Transformations [Fülöp, V. 09] Weighted tree automata and tree transducers [Knight et al ]... [Knuth 77] A generalization of Dijkstra s algorithm [Lopez 08] Statistical Machine Translation [Liang, Bouchard-Côté, Klein, Taskar 06] An End-to-End Discriminative Approach to Machine Translation [Maletti 06] Compositions of Tree Series Transformations [Maletti, Satta 09] Parsing Algorithms based on Tree Automata [Nederhof, V. 12] Synchronous Context-Free Tree Grammars [Thatcher 67] Characterizing derivation trees of context-free grammars through a generalization of finite automata theory [Yamada, Knight 01] A syntax-based statistical translation model

87 26/26 [Knight et al ] [Charniak, Knight, Yamada 03] Syntax-based language models for statistical machine translation [Galley, Hopkins, Knight, Marcu 04] What s in a translation rule? [Graehl, Knight 04] Training tree transducers [Graehl, Knight 05] An overview of probabilistic tree transducers for natural language processing

88 26/26 [Knight et al ] [Charniak, Knight, Yamada 03] Syntax-based language models for statistical machine translation [Galley, Hopkins, Knight, Marcu 04] What s in a translation rule? [Graehl, Knight 04] Training tree transducers [Graehl, Knight 05] An overview of probabilistic tree transducers for natural language processing [Maletti 11] Survey: Weighted Extended Top-down Tree Transducers Part III: Applications in Machine Translation

89 26/26 [Knight et al ] [Charniak, Knight, Yamada 03] Syntax-based language models for statistical machine translation [Galley, Hopkins, Knight, Marcu 04] What s in a translation rule? [Graehl, Knight 04] Training tree transducers [Graehl, Knight 05] An overview of probabilistic tree transducers for natural language processing [Maletti 11] Survey: Weighted Extended Top-down Tree Transducers Part III: Applications in Machine Translation Thank you!

Statistical Machine Translation of Natural Languages

Statistical Machine Translation of Natural Languages 1/37 Statistical Machine Translation of Natural Languages Rule Extraction and Training Probabilities Matthias Büchse, Toni Dietze, Johannes Osterholzer, Torsten Stüber, Heiko Vogler Technische Universität

More information

Applications of Tree Automata Theory Lecture VI: Back to Machine Translation

Applications of Tree Automata Theory Lecture VI: Back to Machine Translation Applications of Tree Automata Theory Lecture VI: Back to Machine Translation Andreas Maletti Institute of Computer Science Universität Leipzig, Germany on leave from: Institute for Natural Language Processing

More information

Lecture 7: Introduction to syntax-based MT

Lecture 7: Introduction to syntax-based MT Lecture 7: Introduction to syntax-based MT Andreas Maletti Statistical Machine Translation Stuttgart December 16, 2011 SMT VII A. Maletti 1 Lecture 7 Goals Overview Tree substitution grammars (tree automata)

More information

Hierarchies of Tree Series TransducersRevisited 1

Hierarchies of Tree Series TransducersRevisited 1 Hierarchies of Tree Series TransducersRevisited 1 Andreas Maletti 2 Technische Universität Dresden Fakultät Informatik June 27, 2006 1 Financially supported by the Friends and Benefactors of TU Dresden

More information

The Power of Tree Series Transducers

The Power of Tree Series Transducers The Power of Tree Series Transducers Andreas Maletti 1 Technische Universität Dresden Fakultät Informatik June 15, 2006 1 Research funded by German Research Foundation (DFG GK 334) Andreas Maletti (TU

More information

Compositions of Tree Series Transformations

Compositions of Tree Series Transformations Compositions of Tree Series Transformations Andreas Maletti a Technische Universität Dresden Fakultät Informatik D 01062 Dresden, Germany maletti@tcs.inf.tu-dresden.de December 03, 2004 1. Motivation 2.

More information

Lecture 9: Decoding. Andreas Maletti. Stuttgart January 20, Statistical Machine Translation. SMT VIII A. Maletti 1

Lecture 9: Decoding. Andreas Maletti. Stuttgart January 20, Statistical Machine Translation. SMT VIII A. Maletti 1 Lecture 9: Decoding Andreas Maletti Statistical Machine Translation Stuttgart January 20, 2012 SMT VIII A. Maletti 1 Lecture 9 Last time Synchronous grammars (tree transducers) Rule extraction Weight training

More information

Compositions of Bottom-Up Tree Series Transformations

Compositions of Bottom-Up Tree Series Transformations Compositions of Bottom-Up Tree Series Transformations Andreas Maletti a Technische Universität Dresden Fakultät Informatik D 01062 Dresden, Germany maletti@tcs.inf.tu-dresden.de May 17, 2005 1. Motivation

More information

Applications of Tree Automata Theory Lecture V: Theory of Tree Transducers

Applications of Tree Automata Theory Lecture V: Theory of Tree Transducers Applications of Tree Automata Theory Lecture V: Theory of Tree Transducers Andreas Maletti Institute of Computer Science Universität Leipzig, Germany on leave from: Institute for Natural Language Processing

More information

How to train your multi bottom-up tree transducer

How to train your multi bottom-up tree transducer How to train your multi bottom-up tree transducer Andreas Maletti Universität tuttgart Institute for Natural Language Processing tuttgart, Germany andreas.maletti@ims.uni-stuttgart.de Portland, OR June

More information

Minimization of Weighted Automata

Minimization of Weighted Automata Minimization of Weighted Automata Andreas Maletti Universitat Rovira i Virgili Tarragona, Spain Wrocław May 19, 2010 Minimization of Weighted Automata Andreas Maletti 1 In Collaboration with ZOLTÁN ÉSIK,

More information

Composition Closure of ε-free Linear Extended Top-down Tree Transducers

Composition Closure of ε-free Linear Extended Top-down Tree Transducers Composition Closure of ε-free Linear Extended Top-down Tree Transducers Zoltán Fülöp 1, and Andreas Maletti 2, 1 Department of Foundations of Computer Science, University of Szeged Árpád tér 2, H-6720

More information

Random Generation of Nondeterministic Tree Automata

Random Generation of Nondeterministic Tree Automata Random Generation of Nondeterministic Tree Automata Thomas Hanneforth 1 and Andreas Maletti 2 and Daniel Quernheim 2 1 Department of Linguistics University of Potsdam, Germany 2 Institute for Natural Language

More information

Preservation of Recognizability for Weighted Linear Extended Top-Down Tree Transducers

Preservation of Recognizability for Weighted Linear Extended Top-Down Tree Transducers Preservation of Recognizability for Weighted Linear Extended Top-Down Tree Transducers Nina eemann and Daniel Quernheim and Fabienne Braune and Andreas Maletti University of tuttgart, Institute for Natural

More information

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister A Syntax-based Statistical Machine Translation Model Alexander Friedl, Georg Teichtmeister 4.12.2006 Introduction The model Experiment Conclusion Statistical Translation Model (STM): - mathematical model

More information

Compositions of Bottom-Up Tree Series Transformations

Compositions of Bottom-Up Tree Series Transformations Compositions of Bottom-Up Tree Series Transformations Andreas Maletti a Technische Universität Dresden Fakultät Informatik D 01062 Dresden, Germany maletti@tcs.inf.tu-dresden.de May 27, 2005 1. Motivation

More information

Structure and Complexity of Grammar-Based Machine Translation

Structure and Complexity of Grammar-Based Machine Translation Structure and of Grammar-Based Machine Translation University of Padua, Italy New York, June 9th, 2006 1 2 Synchronous context-free grammars Definitions Computational problems 3 problem SCFG projection

More information

Compositions of Top-down Tree Transducers with ε-rules

Compositions of Top-down Tree Transducers with ε-rules Compositions of Top-down Tree Transducers with ε-rules Andreas Maletti 1, and Heiko Vogler 2 1 Universitat Rovira i Virgili, Departament de Filologies Romàniques Avinguda de Catalunya 35, 43002 Tarragona,

More information

RELATING TREE SERIES TRANSDUCERS AND WEIGHTED TREE AUTOMATA

RELATING TREE SERIES TRANSDUCERS AND WEIGHTED TREE AUTOMATA International Journal of Foundations of Computer Science Vol. 16, No. 4 (2005) 723 741 c World Scientific Publishing Company RELATING TREE SERIES TRANSDUCERS AND WEIGHTED TREE AUTOMATA ANDREAS MALETTI

More information

Pure and O-Substitution

Pure and O-Substitution International Journal of Foundations of Computer Science c World Scientific Publishing Company Pure and O-Substitution Andreas Maletti Department of Computer Science, Technische Universität Dresden 01062

More information

Syntax-Directed Translations and Quasi-alphabetic Tree Bimorphisms Revisited

Syntax-Directed Translations and Quasi-alphabetic Tree Bimorphisms Revisited Syntax-Directed Translations and Quasi-alphabetic Tree Bimorphisms Revisited Andreas Maletti and C t lin Ionuµ Tîrn uc Universitat Rovira i Virgili Departament de Filologies Romàniques Av. Catalunya 35,

More information

Aspects of Tree-Based Statistical Machine Translation

Aspects of Tree-Based Statistical Machine Translation Aspects of Tree-Based Statistical Machine Translation Marcello Federico Human Language Technology FBK 2014 Outline Tree-based translation models: Synchronous context free grammars Hierarchical phrase-based

More information

The Power of Weighted Regularity-Preserving Multi Bottom-up Tree Transducers

The Power of Weighted Regularity-Preserving Multi Bottom-up Tree Transducers International Journal of Foundations of Computer cience c World cientific Publishing Company The Power of Weighted Regularity-Preserving Multi Bottom-up Tree Transducers ANDREA MALETTI Universität Leipzig,

More information

Music as a Formal Language

Music as a Formal Language Music as a Formal Language Finite-State Automata and Pd Bryan Jurish moocow@ling.uni-potsdam.de Universität Potsdam, Institut für Linguistik, Potsdam, Germany pd convention 2004 / Jurish / Music as a formal

More information

Weighted Regular Tree Grammars with Storage

Weighted Regular Tree Grammars with Storage Discrete Mathematics and Theoretical Computer Science DMTCS vol. 20:1, 2018, #26 Weighted Regular Tree Grammars with Storage Zoltán Fülöp 1 Luisa Herrmann 2 Heiko Vogler 2 arxiv:1705.06681v5 [cs.fl] 2

More information

What we have done so far

What we have done so far What we have done so far DFAs and regular languages NFAs and their equivalence to DFAs Regular expressions. Regular expressions capture exactly regular languages: Construct a NFA from a regular expression.

More information

Einführung in die Computerlinguistik

Einführung in die Computerlinguistik Einführung in die Computerlinguistik Context-Free Grammars (CFG) Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 22 CFG (1) Example: Grammar G telescope : Productions: S NP VP NP

More information

Computational Models - Lecture 4 1

Computational Models - Lecture 4 1 Computational Models - Lecture 4 1 Handout Mode Iftach Haitner. Tel Aviv University. November 21, 2016 1 Based on frames by Benny Chor, Tel Aviv University, modifying frames by Maurice Herlihy, Brown University.

More information

This kind of reordering is beyond the power of finite transducers, but a synchronous CFG can do this.

This kind of reordering is beyond the power of finite transducers, but a synchronous CFG can do this. Chapter 12 Synchronous CFGs Synchronous context-free grammars are a generalization of CFGs that generate pairs of related strings instead of single strings. They are useful in many situations where one

More information

Algorithms for Syntax-Aware Statistical Machine Translation

Algorithms for Syntax-Aware Statistical Machine Translation Algorithms for Syntax-Aware Statistical Machine Translation I. Dan Melamed, Wei Wang and Ben Wellington ew York University Syntax-Aware Statistical MT Statistical involves machine learning (ML) seems crucial

More information

Prefix Probability for Probabilistic Synchronous Context-Free Grammars

Prefix Probability for Probabilistic Synchronous Context-Free Grammars Prefix Probability for Probabilistic Synchronous Context-Free rammars Mark-Jan Nederhof School of Computer Science University of St Andrews North Haugh, St Andrews, Fife KY16 9SX United Kingdom markjan.nederhof@googlemail.com

More information

Cross-Entropy and Estimation of Probabilistic Context-Free Grammars

Cross-Entropy and Estimation of Probabilistic Context-Free Grammars Cross-Entropy and Estimation of Probabilistic Context-Free Grammars Anna Corazza Department of Physics University Federico II via Cinthia I-8026 Napoli, Italy corazza@na.infn.it Giorgio Satta Department

More information

Backward and Forward Bisimulation Minimisation of Tree Automata

Backward and Forward Bisimulation Minimisation of Tree Automata Backward and Forward Bisimulation Minimisation of Tree Automata Johanna Högberg 1, Andreas Maletti 2, and Jonathan May 3 1 Dept. of Computing Science, Umeå University, S 90187 Umeå, Sweden johanna@cs.umu.se

More information

Computation Histories

Computation Histories 208 Computation Histories The computation history for a Turing machine on an input is simply the sequence of configurations that the machine goes through as it processes the input. An accepting computation

More information

Linking Theorems for Tree Transducers

Linking Theorems for Tree Transducers Linking Theorems for Tree Transducers Andreas Maletti maletti@ims.uni-stuttgart.de peyer October 1, 2015 Andreas Maletti Linking Theorems for MBOT Theorietag 2015 1 / 32 tatistical Machine Translation

More information

Extended Multi Bottom-Up Tree Transducers

Extended Multi Bottom-Up Tree Transducers Extended Multi Bottom-Up Tree Transducers Joost Engelfriet 1, Eric Lilin 2, and Andreas Maletti 3;? 1 Leiden Instite of Advanced Comper Science Leiden University, P.O. Box 9512, 2300 RA Leiden, The Netherlands

More information

Key words. tree transducer, natural language processing, hierarchy, composition closure

Key words. tree transducer, natural language processing, hierarchy, composition closure THE POWER OF EXTENDED TOP-DOWN TREE TRANSDUCERS ANDREAS MALETTI, JONATHAN GRAEHL, MARK HOPKINS, AND KEVIN KNIGHT Abstract. Extended top-down tree transducers (transducteurs généralisés descendants [Arnold,

More information

Harvard CS 121 and CSCI E-207 Lecture 10: Ambiguity, Pushdown Automata

Harvard CS 121 and CSCI E-207 Lecture 10: Ambiguity, Pushdown Automata Harvard CS 121 and CSCI E-207 Lecture 10: Ambiguity, Pushdown Automata Salil Vadhan October 4, 2012 Reading: Sipser, 2.2. Another example of a CFG (with proof) L = {x {a, b} : x has the same # of a s and

More information

Weighted tree automata and weighted logics

Weighted tree automata and weighted logics Theoretical Computer Science 366 (2006) 228 247 www.elsevier.com/locate/tcs Weighted tree automata and weighted logics Manfred Droste a,, Heiko Vogler b a Institute of Computer Science, Leipzig University,

More information

New Results on Context-Free Tree Languages

New Results on Context-Free Tree Languages Technische Universität Dresden New Results on Context-Free Tree Languages Dissertation zur Erlangung des akademischen Grades Doctor rerum naturalium (Dr. rer. nat.) vorgelegt an der Technischen Universität

More information

Bisimulation Minimisation for Weighted Tree Automata

Bisimulation Minimisation for Weighted Tree Automata Bisimulation Minimisation for Weighted Tree Automata Johanna Högberg 1, Andreas Maletti 2, and Jonathan May 3 1 Department of Computing Science, Umeå University S 90187 Umeå, Sweden johanna@cs.umu.se 2

More information

Foundations of Informatics: a Bridging Course

Foundations of Informatics: a Bridging Course Foundations of Informatics: a Bridging Course Week 3: Formal Languages and Semantics Thomas Noll Lehrstuhl für Informatik 2 RWTH Aachen University noll@cs.rwth-aachen.de http://www.b-it-center.de/wob/en/view/class211_id948.html

More information

Compositions of Extended Top-down Tree Transducers

Compositions of Extended Top-down Tree Transducers Compositions of Extended Top-down Tree Transducers Andreas Maletti ;1 International Computer Science Institute 1947 Center Street, Suite 600, Berkeley, CA 94704, USA Abstract Unfortunately, the class of

More information

On the Myhill-Nerode Theorem for Trees. Dexter Kozen y. Cornell University

On the Myhill-Nerode Theorem for Trees. Dexter Kozen y. Cornell University On the Myhill-Nerode Theorem for Trees Dexter Kozen y Cornell University kozen@cs.cornell.edu The Myhill-Nerode Theorem as stated in [6] says that for a set R of strings over a nite alphabet, the following

More information

Syntactical analysis. Syntactical analysis. Syntactical analysis. Syntactical analysis

Syntactical analysis. Syntactical analysis. Syntactical analysis. Syntactical analysis Context-free grammars Derivations Parse Trees Left-recursive grammars Top-down parsing non-recursive predictive parsers construction of parse tables Bottom-up parsing shift/reduce parsers LR parsers GLR

More information

Computability and Complexity

Computability and Complexity Computability and Complexity Lecture 5 Reductions Undecidable problems from language theory Linear bounded automata given by Jiri Srba Lecture 5 Computability and Complexity 1/14 Reduction Informal Definition

More information

Bringing machine learning & compositional semantics together: central concepts

Bringing machine learning & compositional semantics together: central concepts Bringing machine learning & compositional semantics together: central concepts https://githubcom/cgpotts/annualreview-complearning Chris Potts Stanford Linguistics CS 244U: Natural language understanding

More information

Theory of Languages and Automata

Theory of Languages and Automata Theory of Languages and Automata Chapter 1- Regular Languages & Finite State Automaton Sharif University of Technology Finite State Automaton We begin with the simplest model of Computation, called finite

More information

COM364 Automata Theory Lecture Note 2 - Nondeterminism

COM364 Automata Theory Lecture Note 2 - Nondeterminism COM364 Automata Theory Lecture Note 2 - Nondeterminism Kurtuluş Küllü March 2018 The FA we saw until now were deterministic FA (DFA) in the sense that for each state and input symbol there was exactly

More information

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other

More information

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other

More information

Fuzzy Tree Transducers

Fuzzy Tree Transducers Advances in Fuzzy Mathematics. ISSN 0973-533X Volume 11, Number 2 (2016), pp. 99-112 Research India Publications http://www.ripublication.com Fuzzy Tree Transducers S. R. Chaudhari and M. N. Joshi Department

More information

Spectral Learning for Non-Deterministic Dependency Parsing

Spectral Learning for Non-Deterministic Dependency Parsing Spectral Learning for Non-Deterministic Dependency Parsing Franco M. Luque 1 Ariadna Quattoni 2 Borja Balle 2 Xavier Carreras 2 1 Universidad Nacional de Córdoba 2 Universitat Politècnica de Catalunya

More information

Roger Levy Probabilistic Models in the Study of Language draft, October 2,

Roger Levy Probabilistic Models in the Study of Language draft, October 2, Roger Levy Probabilistic Models in the Study of Language draft, October 2, 2012 224 Chapter 10 Probabilistic Grammars 10.1 Outline HMMs PCFGs ptsgs and ptags Highlight: Zuidema et al., 2008, CogSci; Cohn

More information

Expectation Maximization (EM)

Expectation Maximization (EM) Expectation Maximization (EM) The EM algorithm is used to train models involving latent variables using training data in which the latent variables are not observed (unlabeled data). This is to be contrasted

More information

Tuning as Linear Regression

Tuning as Linear Regression Tuning as Linear Regression Marzieh Bazrafshan, Tagyoung Chung and Daniel Gildea Department of Computer Science University of Rochester Rochester, NY 14627 Abstract We propose a tuning method for statistical

More information

Introduction and Motivation. Introduction and Motivation. Introduction to Computability. Introduction and Motivation. Theory. Lecture5: Context Free

Introduction and Motivation. Introduction and Motivation. Introduction to Computability. Introduction and Motivation. Theory. Lecture5: Context Free ntroduction to Computability Theory Lecture5: Context Free Languages Prof. Amos sraeli 1 ntroduction and Motivation n our study of RL-s we Covered: 1. Motivation and definition of regular languages. 2.

More information

This lecture covers Chapter 5 of HMU: Context-free Grammars

This lecture covers Chapter 5 of HMU: Context-free Grammars This lecture covers Chapter 5 of HMU: Context-free rammars (Context-free) rammars (Leftmost and Rightmost) Derivations Parse Trees An quivalence between Derivations and Parse Trees Ambiguity in rammars

More information

Structured Prediction Models via the Matrix-Tree Theorem

Structured Prediction Models via the Matrix-Tree Theorem Structured Prediction Models via the Matrix-Tree Theorem Terry Koo Amir Globerson Xavier Carreras Michael Collins maestro@csail.mit.edu gamir@csail.mit.edu carreras@csail.mit.edu mcollins@csail.mit.edu

More information

The Geometry of Statistical Machine Translation

The Geometry of Statistical Machine Translation The Geometry of Statistical Machine Translation Presented by Rory Waite 16th of December 2015 ntroduction Linear Models Convex Geometry The Minkowski Sum Projected MERT Conclusions ntroduction We provide

More information

Monadic Second Order Logic and Automata on Infinite Words: Büchi s Theorem

Monadic Second Order Logic and Automata on Infinite Words: Büchi s Theorem Monadic Second Order Logic and Automata on Infinite Words: Büchi s Theorem R. Dustin Wehr December 18, 2007 Büchi s theorem establishes the equivalence of the satisfiability relation for monadic second-order

More information

Exam: Synchronous Grammars

Exam: Synchronous Grammars Exam: ynchronous Grammars Duration: 3 hours Written documents are allowed. The numbers in front of questions are indicative of hardness or duration. ynchronous grammars consist of pairs of grammars whose

More information

CISC4090: Theory of Computation

CISC4090: Theory of Computation CISC4090: Theory of Computation Chapter 2 Context-Free Languages Courtesy of Prof. Arthur G. Werschulz Fordham University Department of Computer and Information Sciences Spring, 2014 Overview In Chapter

More information

Tree transformations and dependencies

Tree transformations and dependencies Tree transformations and deendencies Andreas Maletti Universität Stuttgart Institute for Natural Language Processing Azenbergstraße 12, 70174 Stuttgart, Germany andreasmaletti@imsuni-stuttgartde Abstract

More information

CS243, Logic and Computation Nondeterministic finite automata

CS243, Logic and Computation Nondeterministic finite automata CS243, Prof. Alvarez NONDETERMINISTIC FINITE AUTOMATA (NFA) Prof. Sergio A. Alvarez http://www.cs.bc.edu/ alvarez/ Maloney Hall, room 569 alvarez@cs.bc.edu Computer Science Department voice: (67) 552-4333

More information

Languages. Languages. An Example Grammar. Grammars. Suppose we have an alphabet V. Then we can write:

Languages. Languages. An Example Grammar. Grammars. Suppose we have an alphabet V. Then we can write: Languages A language is a set (usually infinite) of strings, also known as sentences Each string consists of a sequence of symbols taken from some alphabet An alphabet, V, is a finite set of symbols, e.g.

More information

Chapter 14 (Partially) Unsupervised Parsing

Chapter 14 (Partially) Unsupervised Parsing Chapter 14 (Partially) Unsupervised Parsing The linguistically-motivated tree transformations we discussed previously are very effective, but when we move to a new language, we may have to come up with

More information

Weighted Tree-Walking Automata

Weighted Tree-Walking Automata Acta Cybernetica 19 (2009) 275 293. Weighted Tree-Walking Automata Zoltán Fülöp and Loránd Muzamel Abstract We define weighted tree-walking automata. We show that the class of tree series recognizable

More information

Context-free Grammars and Languages

Context-free Grammars and Languages Context-free Grammars and Languages COMP 455 002, Spring 2019 Jim Anderson (modified by Nathan Otterness) 1 Context-free Grammars Context-free grammars provide another way to specify languages. Example:

More information

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09 Natural Language Processing : Probabilistic Context Free Grammars Updated 5/09 Motivation N-gram models and HMM Tagging only allowed us to process sentences linearly. However, even simple sentences require

More information

Statistical Machine Translation

Statistical Machine Translation Statistical Machine Translation -tree-based models (cont.)- Artem Sokolov Computerlinguistik Universität Heidelberg Sommersemester 2015 material from P. Koehn, S. Riezler, D. Altshuler Bottom-Up Decoding

More information

Tree Transducers in Machine Translation

Tree Transducers in Machine Translation Tree Transducers in Machine Translation Andreas Maletti Universitat Rovira i Virgili Tarragona, pain andreas.maletti@urv.cat Jena August 23, 2010 Tree Transducers in MT A. Maletti 1 Machine translation

More information

Quasi-Synchronous Phrase Dependency Grammars for Machine Translation. lti

Quasi-Synchronous Phrase Dependency Grammars for Machine Translation. lti Quasi-Synchronous Phrase Dependency Grammars for Machine Translation Kevin Gimpel Noah A. Smith 1 Introduction MT using dependency grammars on phrases Phrases capture local reordering and idiomatic translations

More information

Clarifications from last time. This Lecture. Last Lecture. CMSC 330: Organization of Programming Languages. Finite Automata.

Clarifications from last time. This Lecture. Last Lecture. CMSC 330: Organization of Programming Languages. Finite Automata. CMSC 330: Organization of Programming Languages Last Lecture Languages Sets of strings Operations on languages Finite Automata Regular expressions Constants Operators Precedence CMSC 330 2 Clarifications

More information

Chapter 1. Formal Definition and View. Lecture Formal Pushdown Automata on the 28th April 2009

Chapter 1. Formal Definition and View. Lecture Formal Pushdown Automata on the 28th April 2009 Chapter 1 Formal and View Lecture on the 28th April 2009 Formal of PA Faculty of Information Technology Brno University of Technology 1.1 Aim of the Lecture 1 Define pushdown automaton in a formal way

More information

Probabilistic Parsing Strategies

Probabilistic Parsing Strategies Probabilistic Parsing Strategies Mark-Jan Nederhof Faculty of Arts University of Groningen P.O. Box 716 NL-9700 AS Groningen The Netherlands markjan@let.rug.nl Giorgio Satta Dept. of Information Engineering

More information

Mildly Context-Sensitive Grammar Formalisms: Embedded Push-Down Automata

Mildly Context-Sensitive Grammar Formalisms: Embedded Push-Down Automata Mildly Context-Sensitive Grammar Formalisms: Embedded Push-Down Automata Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Sommersemester 2011 Intuition (1) For a language L, there is a TAG G with

More information

Myhill-Nerode Theorem for Recognizable Tree Series Revisited?

Myhill-Nerode Theorem for Recognizable Tree Series Revisited? Myhill-Nerode Theorem for Recognizable Tree Series Revisited? Andreas Maletti?? International Comper Science Instite 1947 Center Street, Suite 600, Berkeley, CA 94704, USA Abstract. In this contribion

More information

A Polynomial Time Algorithm for Parsing with the Bounded Order Lambek Calculus

A Polynomial Time Algorithm for Parsing with the Bounded Order Lambek Calculus A Polynomial Time Algorithm for Parsing with the Bounded Order Lambek Calculus Timothy A. D. Fowler Department of Computer Science University of Toronto 10 King s College Rd., Toronto, ON, M5S 3G4, Canada

More information

Shift-Reduce Word Reordering for Machine Translation

Shift-Reduce Word Reordering for Machine Translation Shift-Reduce Word Reordering for Machine Translation Katsuhiko Hayashi, Katsuhito Sudoh, Hajime Tsukada, Jun Suzuki, Masaaki Nagata NTT Communication Science Laboratories, NTT Corporation 2-4 Hikaridai,

More information

Introduction to Computers & Programming

Introduction to Computers & Programming 16.070 Introduction to Computers & Programming Theory of computation: What is a computer? FSM, Automata Prof. Kristina Lundqvist Dept. of Aero/Astro, MIT Models of Computation What is a computer? If you

More information

Nondeterministic Finite Automata

Nondeterministic Finite Automata Nondeterministic Finite Automata Mahesh Viswanathan Introducing Nondeterminism Consider the machine shown in Figure. Like a DFA it has finitely many states and transitions labeled by symbols from an input

More information

C2.1 Regular Grammars

C2.1 Regular Grammars Theory of Computer Science March 22, 27 C2. Regular Languages: Finite Automata Theory of Computer Science C2. Regular Languages: Finite Automata Malte Helmert University of Basel March 22, 27 C2. Regular

More information

Automata Theory - Quiz II (Solutions)

Automata Theory - Quiz II (Solutions) Automata Theory - Quiz II (Solutions) K. Subramani LCSEE, West Virginia University, Morgantown, WV {ksmani@csee.wvu.edu} 1 Problems 1. Induction: Let L denote the language of balanced strings over Σ =

More information

SCHEME FOR INTERNAL ASSESSMENT TEST 3

SCHEME FOR INTERNAL ASSESSMENT TEST 3 SCHEME FOR INTERNAL ASSESSMENT TEST 3 Max Marks: 40 Subject& Code: Automata Theory & Computability (15CS54) Sem: V ISE (A & B) Note: Answer any FIVE full questions, choosing one full question from each

More information

Introduction to Automata

Introduction to Automata Introduction to Automata Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr 1 /

More information

CS:4330 Theory of Computation Spring Regular Languages. Finite Automata and Regular Expressions. Haniel Barbosa

CS:4330 Theory of Computation Spring Regular Languages. Finite Automata and Regular Expressions. Haniel Barbosa CS:4330 Theory of Computation Spring 2018 Regular Languages Finite Automata and Regular Expressions Haniel Barbosa Readings for this lecture Chapter 1 of [Sipser 1996], 3rd edition. Sections 1.1 and 1.3.

More information

CSE 105 THEORY OF COMPUTATION

CSE 105 THEORY OF COMPUTATION CSE 105 THEORY OF COMPUTATION Fall 2016 http://cseweb.ucsd.edu/classes/fa16/cse105-abc/ Today's learning goals Sipser Ch 3 Trace the computation of a Turing machine using its transition function and configurations.

More information

DM17. Beregnelighed. Jacob Aae Mikkelsen

DM17. Beregnelighed. Jacob Aae Mikkelsen DM17 Beregnelighed Jacob Aae Mikkelsen January 12, 2007 CONTENTS Contents 1 Introduction 2 1.1 Operations with languages...................... 2 2 Finite Automata 3 2.1 Regular expressions/languages....................

More information

Grammar formalisms Tree Adjoining Grammar: Formal Properties, Parsing. Part I. Formal Properties of TAG. Outline: Formal Properties of TAG

Grammar formalisms Tree Adjoining Grammar: Formal Properties, Parsing. Part I. Formal Properties of TAG. Outline: Formal Properties of TAG Grammar formalisms Tree Adjoining Grammar: Formal Properties, Parsing Laura Kallmeyer, Timm Lichte, Wolfgang Maier Universität Tübingen Part I Formal Properties of TAG 16.05.2007 und 21.05.2007 TAG Parsing

More information

Real-Time Systems. Lecture 15: The Universality Problem for TBA Dr. Bernd Westphal. Albert-Ludwigs-Universität Freiburg, Germany

Real-Time Systems. Lecture 15: The Universality Problem for TBA Dr. Bernd Westphal. Albert-Ludwigs-Universität Freiburg, Germany Real-Time Systems Lecture 15: The Universality Problem for TBA 2013-06-26 15 2013-06-26 main Dr. Bernd Westphal Albert-Ludwigs-Universität Freiburg, Germany Contents & Goals Last Lecture: Extended Timed

More information

C2.1 Regular Grammars

C2.1 Regular Grammars Theory of Computer Science March 6, 26 C2. Regular Languages: Finite Automata Theory of Computer Science C2. Regular Languages: Finite Automata Malte Helmert University of Basel March 6, 26 C2. Regular

More information

IBM Model 1 for Machine Translation

IBM Model 1 for Machine Translation IBM Model 1 for Machine Translation Micha Elsner March 28, 2014 2 Machine translation A key area of computational linguistics Bar-Hillel points out that human-like translation requires understanding of

More information

FORMAL LANGUAGES, AUTOMATA AND COMPUTATION

FORMAL LANGUAGES, AUTOMATA AND COMPUTATION FORMAL LANGUAGES, AUTOMATA AND COMPUTATION DECIDABILITY ( LECTURE 15) SLIDES FOR 15-453 SPRING 2011 1 / 34 TURING MACHINES-SYNOPSIS The most general model of computation Computations of a TM are described

More information

MANFRED DROSTE AND WERNER KUICH

MANFRED DROSTE AND WERNER KUICH Logical Methods in Computer Science Vol. 14(1:21)2018, pp. 1 14 https://lmcs.episciences.org/ Submitted Jan. 31, 2017 Published Mar. 06, 2018 WEIGHTED ω-restricted ONE-COUNTER AUTOMATA Universität Leipzig,

More information

Computational Models - Lecture 1 1

Computational Models - Lecture 1 1 Computational Models - Lecture 1 1 Handout Mode Ronitt Rubinfeld and Iftach Haitner. Tel Aviv University. February 29/ March 02, 2016 1 Based on frames by Benny Chor, Tel Aviv University, modifying frames

More information

CS21 Decidability and Tractability

CS21 Decidability and Tractability CS21 Decidability and Tractability Lecture 8 January 24, 2018 Outline Turing Machines and variants multitape TMs nondeterministic TMs Church-Turing Thesis So far several models of computation finite automata

More information

Lecture Notes on Inductive Definitions

Lecture Notes on Inductive Definitions Lecture Notes on Inductive Definitions 15-312: Foundations of Programming Languages Frank Pfenning Lecture 2 September 2, 2004 These supplementary notes review the notion of an inductive definition and

More information

Theory of Computation

Theory of Computation Theory of Computation (Feodor F. Dragan) Department of Computer Science Kent State University Spring, 2018 Theory of Computation, Feodor F. Dragan, Kent State University 1 Before we go into details, what

More information

Decidable and undecidable languages

Decidable and undecidable languages The Chinese University of Hong Kong Fall 2011 CSCI 3130: Formal languages and automata theory Decidable and undecidable languages Andrej Bogdanov http://www.cse.cuhk.edu.hk/~andrejb/csc3130 Problems about

More information