![]() | This article may be in need of reorganization to comply with Wikipedia's layout guidelines. (January 2025) |
Statistical machine translation (SMT) is a machine translation approach where translations are generated on the basis of statistical models whose parameters are derived from the analysis of bilingual text corpora. The statistical approach contrasts with the rule-based approaches to machine translation as well as with example-based machine translation,[1] that superseded the previous rule-based approach that required explicit description of each and every linguistic rule, which was costly, and which often did not generalize to other languages. Since 2003, the statistical approach itself has been gradually superseded by the deep learning-based neural machine translation.
The first ideas of statistical machine translation were introduced by Warren Weaver in 1949,[2] including the ideas of applying Claude Shannon's information theory. Statistical machine translation was re-introduced in the late 1980s and early 1990s by researchers at IBM's Thomas J. Watson Research Center.[3][4][5] Before the introduction of neural machine translation, it was by far the most widely studied machine translation method.
Statistical machine translation is related to other data-driven methods in machine translation, such as the earlier work on example-based machine translation. Contrast this to systems that are based on hand-crafted rules.
© MMXXIII Rich X Search. We shall prevail. All rights reserved. Rich X Search