summaryrefslogtreecommitdiff
path: root/report/introduction.tex
diff options
context:
space:
mode:
Diffstat (limited to 'report/introduction.tex')
-rw-r--r--report/introduction.tex8
1 files changed, 4 insertions, 4 deletions
diff --git a/report/introduction.tex b/report/introduction.tex
index 21e0e907..3b673c8e 100644
--- a/report/introduction.tex
+++ b/report/introduction.tex
@@ -1,12 +1,12 @@
\chapter{Introduction}
-Automatically generating high quality translations for foreign texts remains a central challenge for Natural Language Processing research.
-Recent advances in Statistical Machine Translation (SMT) has enabled these technologies to move out of research labs an become viable commercial products and ubiquitous online tools. \footnote{e.g., translate.google.com, www.systran.co.uk, www.languageweaver.com}
+Automatically generating high quality translations for foreign texts remains a central challenge for natural language processing research.
+Recent advances in statistical machine translation (SMT) has enabled these technologies to move out of research labs an become viable commercial products and useful online tools. \footnote{e.g., translate.google.com, www.systran.co.uk, www.languageweaver.com}
However these successes have not been uniform;
current state-of-the-art translation output varies markedly in quality depending on the languages being translated.
-Those language pairs that are closely related language pairs (e.g., English and French) can be translated with a high degree of precision, while for distant pairs (e.g., English and Chinese) the result is far from acceptable.
+Those language pairs that are closely related language pairs (e.g., English and French) can be translated with high quality, while for distant pairs (e.g., English and Chinese) the result tends to be much lower quality.
It is tempting to argue that SMT's current limitations can be overcome simply by increasing the amount of data on which the systems are trained.
-However, large scale evaluation campaigns for Chinese~$\rightarrow$~English translation have not yielded the expected gains despite the increasing size of the models.
+However, large scale evaluation campaigns for Chinese~$\rightarrow$~English translation, such as the DARPA GALE program, have not yielded high quality translation despite providing hundreds of millions of words worth of training data.
\begin{figure}[t]
\centering \includegraphics[scale=0.55]{urdu_example_translation.pdf}