You are on page 1of 2

The Era of Artificial Speech Translation

Artificial speech translation has long been a topic of interest, and recent technological
advancements have brought us closer to a future where language barriers may become a thing of
the past. While significant progress has been made, it is essential to critically evaluate the current
state of artificial speech translation and its potential implications (Kohn 1). This paper examines
the existing systems, their practical applications, economic drivers, challenges, and the enduring
value of human language skills.
Notable advancements in artificial speech translation technology have been achieved. For
instance, Google has incorporated translation features into its Pixel earbuds, enabling real-time
translation through its smartphone app (Kohn 1). While these systems demonstrate the feasibility
of the concept, they are currently perceived as curious novelties rather than transformative tools.
Barak Turovsky, Google Translate's director of product, acknowledges that we are still in the
early stages of voice-enabled devices, with translation being an essential aspect of their
development (Kohn 1).
One practical application of artificial speech translation is facilitating interactions in
public spaces, such as hotel check-ins or fundamental public interactions. These applications
address the need to overcome language barriers and offer alternatives to relying solely on
travelers' English (Kohn 1). Emerging Asian markets present significant economic drivers for
translation devices. Google estimates that only 20% of the world's population speaks English,
while a vast amount of internet content is in English (Kohn 1). Breaking these language barriers
becomes crucial in enabling internet access and fostering inclusive communication.
Despite the progress made, several challenges remain. Simultaneous translation, akin to
the voiceover on televised news, is necessary for fluid communication. Offline functionality is
also essential when internet access is limited (Kohn 1). Privacy concerns regarding accumulating
emotional speech data in the cloud further complicate the development of artificial speech
translation systems.
Artificial speech translation systems must address technical challenges and demonstrate
social awareness. They should be capable of adapting to cultural norms and addressing people
appropriately. However, professionals, such as linguists and language experts, are unlikely to
outsource language awareness entirely to software (Kohn 1). Seamless and ubiquitous artificial
speech translation will enhance the value of language skills, as automated translation primarily
delivers practical, low-prestige information.
As artificial speech translation becomes a ubiquitous commodity product, the value of
human language skills will only increase. Shared languages promote a sense of belonging,
community, and trust. Learning another person's language demonstrates commitment and
facilitates meaningful connections (Kohn 2). While machine translation offers convenience, it
can only partially replace the subtle understanding that comes with the knowledge of a language.
Human language skills will always be crucial in discerning nuances and extracting meaning from
complex interactions (Kohn 1).
The era of artificial speech translation is gradually emerging, offering glimpses of a
future where language barriers may be diminished. However, despite their functionality, the
current systems are still perceived as novelties rather than transformative tools. Challenges
related to simultaneous translation, offline functionality, and privacy concerns must be
overcome. Additionally, human language skills will remain valuable in facilitating genuine
connections and comprehending subtle nuances that automated systems may struggle to grasp.

Work Cited
Kohn, Marek. "Is the era of artificial speech translation upon us." The Guardian 2019.
https://isl.anthropomatik.kit.edu/downloads/Isthe-era-of-artificial-speech-translation-
upon-us_The-Guardian.pdf

You might also like