Lentis/Google Translate

Google Translate is a machine translation service for written text, speech, and text in images. Since its launch in 2006, the service has become available online and as an app, including offline functionality. There are 103 languages available and over 100 billion words being translated each day as of 2018, with more than 92% of translations coming from outside the US. In addition, there are numerous features, such as text pronunciation, single-word dictionaries and "phrasebook" translation-saving for later use.

Google Neural Machine Translation
In 2016, Google introduced its Neural Machine Translation (NMT) system, which is a framework for machine-learning algorithms to process complex data. With this design, the network first encodes the original sentence as a list of individual vectors, with each vector representing the meaning of a word. Once the sentence is read, the decoder begins generating the new sentence by having it focus on a weighted distribution over the relevant encoded vectors. The benefit of this is that the vectors take their surrounding vectors into context in order to produce the correct translation, one word at a time. In comparison to Phrase-Based Machine Translation, this algorithm takes an entire input sentence as a block for translation as opposed to smaller individual blocks.

Translation Quality
Google's Neural Machine Translation is currently only available for select languages; however, there are many others are in the process of development. With its increase in the quality of translation, the new implementation has attained a translation that is very close to that of a human.

Translate Community
As a corrective measure, Google Translate employs a "Translate Community." User can sign up as speakers of a given language, check provided translations for accuracy, and provide translations of words and phrases Google is unsure about. They also have an incentive system of badges and awards. Additionally, even those who don't sign up can click on a Google Translate translation and offer an alternative which Google uses to improve its translation quality.

2018 World Cup
Google Translate saw a large spike in usage during the 2018 World Cup in Russia. Within the country, it saw a 30% increase in total usage, and a 200% increase in queries containing the phrase "world cup". Even the word 'beer' saw an increase of 65%.

Legal cases
In 2017, a court in the United Kingdom had to rely on Google translate to inform a defendant that proceedings had to be delayed because an interpreter could not be found. While this was a good use, it shows that the court was not ready to rely on the service itself to communicate during the actual trial.

In the 2017 case US vs Cruz-Zamora, Google translate was used for communication between a Police Officer and driver during a traffic stop which led to the discovery of illicit substances and arrest of Mr. Cruz-Zamora. At issue was whether the Mr. Cruz-Zamora was able to give informed consent to have his car searched based on a question asked through Google translate. The question the officer asked, “Can I search the car?”, translated to “¿Puedo buscar el auto?” While literally correct, when translating back to English this becomes “Can I find the car?”, to which the defendant first replied with “I do not understand” and, upon repeating, “Yes, yes. Go ahead.” The court ruled that Google translate could not be used by an officer to reliably converse with and therefore obtain consent from someone. This case established precedent for Google Translate being insufficient in the US court of law.

Translation between cultures
Because the training translations attempt to make material relevant to foreign-language audiences, it is possible for Google Translate to learn interesting "translations." For example, it has translated "Ivan the Terrible" in Russian to "Abraham Lincoln" in English. This phenomenon may be due to translators using these names as examples of strong central leaders that their audiences will recognize.

Effect on human judgement
Google Translate, while very intelligent, is not a substitute for human judgement. Consider the Portuguese idiom "tirar onda," which means "to kid" or "to joke." Google Translate word-for-word translates it to "to take a wave." A human translator would recognize that this phrase is not meant to be literally translated. Reliance on machine translations by users unfamiliar with the source or target language can cause confusion and error because they are not able to use their best judgement.

Google also has the ability to capture the underlying meaning behind the language, even at times when it should not have. For instance, the expression “I’m a flat-earther” translates to “I’m a crazy person”, when translating from English to French. It found that when people use the term ‘flat-earther’, they mean an individual with unusual beliefs. However, with the resurgence of actual flat-earther believers, the system is not prepared for the literal meaning of the term for the group of people. An example of this phenomenon of misguided trust does not even involve the service. In 2008 a bilingual road sign in Wales used "I am not in the office at the moment. Send any work to be translated" as the Welsh translation of a sign's English message. Ironically, in this particular case, the road planners could have used Google Translate to uncover the error.

Cultural bias
Google Translate is based on human-created translations, which often substitute the original cultural concepts for the target audience. This presents the opportunity for human biases to appear. For example, when typing "Barcelona, Catalonia" in Catalan, the Spanish translation is "Barcelona, Spain." This translation reflects nationalist sentiment in Spanish culture regarding Catalonian independence.

Google Translate learns largely from United Nations and European Parliament transcripts, as well as popular novels, such as Harry Potter, that are translated into many languages. Furthermore, translations between two non-English languages always are translated first to English then from English to the target language. For these reasons, there is a Eurocentric, and specifically Anglocentric bias. Translating between two similar languages, such as Italian and Spanish, can inject mistakes caused by English’s distance from each language, and languages of countries with smaller roles in the UN or EU have less source material and therefore produce less accurate translations.

Google Translate also chooses what languages to add to its list based on how much translated text exists in that language on the internet. This means that languages of groups underrepresented on the internet, especially those with less access to the internet, are unlikely to see their languages offered anytime soon. This creates a feedback loop whereby groups with little text in their language on the internet cannot use Google Translate, and therefore have limited accessibility to the internet which then makes it more difficult to put texts in their language on the internet. Increasing accessibility to smaller language groups or speakers of dying languages is an issue Google Translate faces moving forward.

Gender bias
Google’s artificial intelligence has been found to reflect human stereotypes with respect to gender. This is seen in many translations, particularly from gender-neutral to gender-inclusive words. During translation, the system uses source documents as input from its databases and learns to give a result from the knowledge it gains. This allows the system to follow both current and previous patterns that are written in the English language. For example, according to a 2016 study by DATAUSA, 77.5% of computer scientists were recorded as men while 89.3% of registered nurses were women. As a result, it is common to see the male term of the title "programmer" in place of the female term, such as in the case when translating from English to Italian. Statistically speaking, the system produces a reasonable outcome given the data that it reflects.

Conclusion
Technology such as Google Translate can lower barriers to internationalization. However, as evinced in the gender bias, Anglicizing effects, and accessibility issues of Google Translate, it is important to question how the technology we use affects our interactions across cultures and what standards or cultural inequities it may unknowingly be imposing.

Artificial Intelligence has limitations, but is rapidly bridging the gap. Additionally, as seen in the Welsh cases, humans are not themselves infallible. This raises two questions: to what standard do we hold AI, and is it a fair standard? As seen in the Cruz-Zamora case, AI’s imperfections are not currently acceptable, at least by a court of law, but as the day approaches when that is no longer the case, how we handle the transition will be largely affected by the forethought we practice or fail to today.