Google is working on language-to-text AR glasses. It’s a complicated idea

​​Google appears to be working on AR glasses once again, but this time, it is showing a new feature that translates speech to readable text.  At last week’s Google I/O 2022, the company demonstrated an AR glasses prototype that can translate spoken language into readable display text. Google has not hinted whether they are developing these as a product, or when, but the fact that they showed them to developers is indicative that they are thinking of how to extend the model of AR glasses to utilize their gigantic datasets and existing technologies. If Google moves forward with the product, it is likely that it will frame it as a device that would  attempt to break down language barriers. Sounds great, right? No more trying to find Google Translate on the web and pecking phrases into our mobile phones to translate things. When (or if) these hit the market, we’ll finally be able to read foreign signs, order correctly in restaurants, and even make new friends more easily when we travel. More significantly, there would be a way to quickly translate communication in the event of an emergency, when people may not all speak the same language. On another level, these “translation glasses” could also open up communication channels for the deaf and hard of hearing community, giving them a new way to communicate with those around them. However, as with all new technology ideas, Google’s translation glasses could come with a huge social cost: to our privacy, our well-being, and our cooperation with each other in our communities. What does it mean when Google becomes the translator for our lives, and are we comfortable with that notion?  The problem with any type of technology translation device is that it has to “listen” to those around it, to procure the data to translate. And if the AR glasses are listening, we will need to know what, or whom, they are listening to—and when they are listening. At the moment, we don’t know if these glasses will be able to distinguish between more than one person at a time, either. Also, we will need to know if it is legal for these glasses to listen without consent—and if one needs the consent of someone to record them in order to translate them, will one need the glasses to translate the consent? We don’t know if in the future, these glasses will have the capacity to record what they translate, nor will we know if they could identify whom they are recording at any given time, or within what range they are capable of listening. And If they are recording glasses, or even with the transcribed text, we’ll need to know if  that is stored somewhere that can be erased, and if people could opt-out in a public space without being recorded while doing so. Let’s assume for the moment that these Google glasses won’t record us, and that Google manages to figure out consent and permission. Given that, in our crowded, noisy world, the usual problems with speech to text could still abound in the form of misunderstandings, misrepresentations, etc., in what Google ‘hears’ and what it writes as a result of that hearing. The tech might also have a lot of misspellings and confusion with mixing languages. As The Verge pointed out, many of us “code switch” using words from many different languages interspersed, with the added complexity of not all of them reading from left-to-right, which will need to be accommodated, too.  Now add to that an aggregate population using these while wandering around, which invokes much of what I wrote with Dr. Catherine Flick about Meta’s pre-Ray-Ban Stories Project Aria glasses. Many of the same issues persist, except for that with these new Google glasses, people may be walking around and reading transcripts, which again, is more like what was going on in the early days of cell phones and Divided Attention, creating potentially dangerous outcomes as distracted people walk into traffic or fall into fountains. One of the main concerns with the glasses is Google’s seeming assumption here that technology can solve cultural problems—and that if the technology isn’t working, the solution is to develop and apply more technology. In this case, solving cross-cultural communication problems cannot be fully solved with language translation. Tech can help, but these glasses won’t translate culture or cultural norms such as whether someone is comfortable being direct or indirect, or any one of multitudes of cultural nuances and cues found in the ways that different people in different groups communicate with each other. For that, we need other humans to guide us.

S.A. Applin, PhD, is an anthropologist whose research explores the domains of human agency, algorithms, AI, and automation in the context of social systems and sociability. You can find more at @anthropunk and PoSR.org.

https://www.fastcompany.com/90753311/google-is-working-on-language-to-text-ar-glasses-its-a-complicated-idea?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Établi 3y | 18 mai 2022, 14:21:55


Connectez-vous pour ajouter un commentaire

Autres messages de ce groupe

Teens are still setting fire to Chromebooks for TikTok clout

Students are still setting fire to their Chromebooks for TikTok—and now they’re facing the consequences.

Fast Company first reported on the #ChromebookChallenge trend last

15 mai 2025, 10:50:03 | Fast company - tech
Google is returning to virtual reality with Android XR—and a new strategy

At its annual Google I/O developer conference in Mountain View next week, Google will try to rally developers around one of its next big bets: Android XR.

15 mai 2025, 10:50:02 | Fast company - tech
Elon Musk’s Grok AI is replying to tweets with claims about ‘white genocide’ in South Africa

X users who interacted with the chatbot Grok on Wednesday were confronted with replies about the legitimacy of white genocide in South Africa—often regardless of context.

In one post, a

14 mai 2025, 20:50:03 | Fast company - tech
How Headspace and Ozlo help people drift off with sound

Ever wonder why the sound of rain makes you instantly drowsy, but a ticking clock drives you up the wall? That’s because not all noise soothes the brain in the same way. Sleep sounds might seem li

14 mai 2025, 16:20:06 | Fast company - tech
Elon Musk’s DOGE is launching a new AI retirement system. It was built mostly under Biden

Elon Musk’s Department of Government Efficiency (DOGE) has spent its first 100 days slashing

14 mai 2025, 16:20:05 | Fast company - tech