Main article: IBM alignment models
The IBM models4 are used in Statistical machine translation to train a translation model and an alignment model. They are an instance of the Expectation–maximization algorithm: in the expectation-step the translation probabilities within each sentence are computed, in the maximization step they are accumulated to global translation probabilities. Features:
Vogel et al.5 developed an approach featuring lexical translation probabilities and relative alignment by mapping the problem to a Hidden Markov model. The states and observations represent the source and target words respectively. The transition probabilities model the alignment probabilities. In training the translation and alignment probabilities can be obtained from γ t ( i ) {\displaystyle \gamma _{t}(i)} and ξ t ( i , j ) {\displaystyle \xi _{t}(i,j)} in the Forward-backward algorithm.
P. F. Brown et al. 1993. The Mathematics of Statistical Machine Translation: Parameter Estimation Archived April 24, 2009, at the Wayback Machine. Computational Linguistics, 19(2):263–311. http://acl.ldc.upenn.edu/J/J93/J93-2003.pdf ↩
Och, F.J. and Tillmann, C. and Ney, H. and others 1999, Improved alignment models for statistical machine translation, Proc. of the Joint SIGDAT Conf. on Empirical Methods in Natural Language Processing and Very Large Corpora http://www.aclweb.org/anthology/W99-0604.pdf ↩
ACL 2005: Building and Using Parallel Texts for Languages with Scarce Resources Archived May 9, 2009, at the Wayback Machine http://www.cse.unt.edu/~rada/wpt05/ ↩
Philipp Koehn (2009). Statistical Machine Translation. Cambridge University Press. p. 86ff. ISBN 978-0521874151. Retrieved 21 October 2015. 978-0521874151 ↩
S. Vogel, H. Ney and C. Tillmann. 1996. HMM-based Word Alignment in Statistical Translation Archived 2018-03-02 at the Wayback Machine. In COLING ’96: The 16th International Conference on Computational Linguistics, pp. 836-841, Copenhagen, Denmark. https://aclanthology.info/pdf/C/C96/C96-2141.pdf ↩