Eggs “over easy” and reflections on the authenticity of ChatGPT’s language

Eggs cooked over easy

I was in Vietnam last fall on assignment from the US State Department (Public Diplomacy Division) to serve as an “English Language Specialist,” giving workshops to foreign language teachers (mostly of English, come Korean and Mandarin) on integrating technology into language learning. The main topic: generative AI like ChatGPT. There was a lot of interest, as well as plenty of concern (students cheating, teachers losing their jobs). AI offers a variety of opportunities for language teachers and learners, as well as bringing with it lots of issues, both in education and for society as a whole. I just finished up an upcoming column on that topic for Language Learning & Technology.

One of the topics we talked about in the workshops was the quality of the language output from AI systems. While the language generated is grammatically correct, fluent, and coherent (although sometimes inaccurate), I pointed out to the teachers that AI systems have a quite different orientation to language than do humans, with a likely result that the social and cultural aspects of AI output might be questionable. I decided to try out an experiment with ChatGPT based on an incident I witnessed at breakfast at the hotel where I was staying in Hanoi. I entered the prompt below:

Me

An American tourist is staying at a hotel in Vietnam. At breakfast, he goes to the grill station and requests, in English, fried eggs “over easy” (i.e. with runny yokes). The cook has trouble understanding what he wants. The tourist repeats his request, but the cook still does not understand. The tourist says, “Ok, I’ll come back later. Thank you.” The cook does not respond. The tourist starts to walk away, then turns back to the cook and says, “I SAID thank you.” Why did the tourist make that last statement?

ChatGPT

The tourist likely made the last statement, “I said thank you,” as a polite way of expressing gratitude and acknowledging the cook, even though there was no verbal response from the cook. In some cultures, including American culture, expressing gratitude is considered polite and is a social norm. The tourist may have wanted to convey that he appreciated the cook’s efforts or wanted to ensure that his gratitude was acknowledged, even if the cook did not respond verbally. It’s a common social convention to say “thank you” after making a request or receiving assistance.

Of course, ChatGPT was wrong and took literally what was said instead of understanding the real meaning of “I SAID thank you.” The American tourist expected (unreasonably) an “Ok” or “You’re welcome” or some kind of verbal response and was repeating his “thank you” from irritation over the lack of response (and over not getting the eggs he wanted). From the perspective of language pragmatics, the Vietnamese cook failed to complete an “adjacency pair”, a conversational turn-taking pattern that stipulates a standard reply to an utterance. Unfortunately, the cook didn’t know enough English to play his role as expected. These are formulaic sequences, without real semantic meaning, rather a form of socialization, connecting speakers together. The American English greeting “How are you?” is not asking for a health report, but just offering a greeting, with an expected reply of “Fine, thanks”. Below is an abstract for a talk I am giving (virtually) in Portugal at an upcoming conference on “Digital Learning Environments and Authenticity in English Language Teaching.” My presentation deals with the question of social and cultural authenticity in AI language production:

The ability of ChatGPT and other AI systems to generate language that resembles closely human-produced speech has led to claims that AI chatbots can “facilitate an authentic, interactional language learning environment” (Chiu et al., 2023), that AI use is “essential for promoting cultural sensitivity, intercultural competency, and global awareness” (Anis, 2023, p. 64), and that AI-based VR supplies “the benefits of in-country immersion programs without the hassle” (Divekar, 2022, p. 2354). The suggestion in these studies is that AI output is linguistically and culturally “authentic” and could substitute in language learning settings for human interlocutors or could even provide similar benefits to a study abroad experience.

Such a view ignores the process used by AI systems to reproduce language and the limitations of that process for the linguistic features and cultural content of the resulting output. AI systems break down language (their training data) into mathematical symbols and use machine learning to find patterns and regularities to form a “large language model” that enables next word prediction in a text string, which is then used to very effectively construct sentences, paragraphs, even complete discourses. Humans, in contrast, are socialized into their language abilities, learning gradually how to use language appropriately within an ever-expanding circle of social contexts. Through interactions with others, we acquire the social and cultural norms of language use, including the contextually appropriate use of nonverbal communication, i.e., facial expressions, body language, and paralanguage. The statistical model of language in AI lacks the sociocultural grounding humans have through sensorimotor interactions and from simply living in the real world.

Studies of AI’s capabilities to engage in pragmatically effective language use have shown significant limitations (Lee & Wang, 2022; Su & Goslar, 2023). While AI systems can gain pragmalinguistic knowledge and learn appropriate formulaic sequences (politeness conventions, for example) through the verbal exchanges in their training data, they have proven to be much less effective in sociopragmatic engagement, that is, in generating contextually acceptable speech reflecting an interlocutor’s state of mind, intentions, and emotional status. AI systems are likely to improve through user interactions added to their language models, through enlarging their datasets, and through multimodal incorporation (adding video and image training). However, those measures still will not supply the lived experience humans go through in negotiating common ground linguistically and culturally in social interactions and therefore the ability to deal with nuanced pragmatic scenarios. AI generated language–while valuable as a resource in language learning–will remain artificial and inauthentic in ways that cannot serve as an acceptable substitute for actual learner engagement in the L2 with peers and expert speakers.

What’s in a name? Maybe getting elected

Chinese ballot in San Francisco

Interesting piece recently in the New York Times on politicians in San Francisco, using Chinese names to help them get elected. That city, in fact, has itself an interesting Chinese name: 舊金山 (Jiù Jīn Shān), which translates as “Old Gold Mountain,” referencing the 19th-century gold rush in California. San Francisco has one of the oldest and best-known “Chinatowns” in the US, with many residents speaking Cantonese or Mandarin as their first language. Due to the large number of Chinese-speaking residents, San Francisco has mandated since 1999 that names for political candidates in local elections appear in both English and Chinese. That is no problem for candidates with a Chinese heritage; they likely already have Chinese names. For non-Chinese candidates, they can simply transliterate their English (or Spanish) names. Foreign names are usually transliterated into Chinese by phonetically selecting characters that approximate the sound of the name, as explained in this post from “Fluent in Mandarin.”  So “Anderson” becomes 安德森(Āndésēn). It’s not an automatic process, as there are often different characters that can be used for a specific sound.

As pointed out on that site, “The problem with ‘transliterating’ foreign names into Chinese is that they can often sound very unnatural, and not like a native Chinese name.” It’s also the case that the actual meaning of the name (through the characters used) chosen for phonetic similarity may be quite strange (for Chinese speakers). The Chinese transliteration of “Anderson” (安德森)means Install-Virtue-Forest. For foreigners in China, it’s sometimes better to come up with some kind of nickname, as those are frequently used informally in China. In China, I go by the name 老牛 (Lǎo niú), meaning “Old Bull”; that references my age but was particularly chosen as my oldest son (who’s married to a Chinese woman) is “Young Bull” in China.

In San Francisco, it used to be that non-Chinese politicians could come up with their own Chinese names. As it turned out, many did not simply use transliterations of their names, but instead came up with Chinese names that made them sound good. That has now changed, so that Chinese names can only be used if you were born with that name (i.e. are from a Chinese background) or have publicly used that Chinese name for at least 2 years. Otherwise, your name will be transliterated into Chinese phonetically by the election authorities. This has resulted in some candidates having to change the Chinese versions of their names. According to the Times article,

Michael Isaku Begert, who is running to keep his local judgeship, cannot use 米高義, which means in part “high” and “justice,” a name that suggests he was destined to sit on the bench. And Daniel Lurie, who is challenging Mayor London Breed, must scrap the name he had been campaigning with for months: 羅瑞德, which means “auspicious” and “virtue.” Mr. Lurie’s new name, 丹尼爾·羅偉, pronounced Daan-nei-ji Lo-wai, is a transliterated version that uses characters closer to the sound of his name in English but are meaningless when strung together.

According to a piece in the San Franciso Standard, this rule may actually disadvantage ethnic Chinese Americans: “American-born Chinese candidates may be unintended victims of the law, if they don’t have a birth certificate in Chinese that proves they were born with Chinese names.” They may have to supply alternative proof, which meant, according to the article, candidates, “were scrambling to seek old Chinese newspaper articles, Chinese school homework or Chinese family books to find the appropriate records.”

San Francisco is not the only US locale printing ballots in more than one language. According to the Timesarticle:

Certain towns in Alaska must translate ballots into Yup’ik, an Indigenous Alaskan language, while some counties in Arizona must do so in Navajo and Apache. Hundreds of jurisdictions around the nation must translate their ballots into Spanish, while 19 must print them in Chinese, 12 in Vietnamese and four in Korean.

It would be interesting to know if politicians in those areas who have English names also come up with names in those other languages, and whether in the process they try to play language games to make themselves sound good.

Language change

Jean Patrick Niambé, Ivorian rapper, Arlette Bashizi for The New York Times

In workshops I gave this fall in Vietnam on the use of AI in language learning, one of the topics we discussed was the change in English language use brought about through the AI frenzy of 2023. I used as an example a Doonesbury cartoon that highlighted the shift from the career benefits of being a “creative” to the now more lucrative role of becoming a “promptive,” i.e., doing prompt engineering. Traditional careers as lawyers and artists are presented as going away (replaced respectively by ChatGPT and Midjourney), leading to the need to find a “side hustle.” “AI” itself was named the most notable word of 2023 by Collins dictionary. Lots of other AI terms have entered this year into everyday language in English including “machine learning” and “large language model.” “Prompt” is now inevitably tied to AI use.

Language change happens in myriad ways, one being through technological and societal change, another through demographic shifts. That is illustrated in a recent article in the New York Times on changes to French through new uses of the language in the growing number of French speakers in West and central Africa:

More than 60 percent of those who speak French daily now live in Africa, and 80 percent of children studying in French are in Africa. There are as many French speakers in Kinshasa, the capital of the Democratic Republic of Congo, as in Paris. Through social media platforms like TikTok and YouTube, they are literally spreading the word, reshaping the French language from African countries, like Ivory Coast, that were once colonized by France.

The article chronicles how young people in Africa have adapted French creatively in entertainment and the arts. One rapper comments, “We’ve tried to rap in pure French, but nobody was listening to us,” so language used in rapping is adjusted to the social realities of local life. An example of an African neologism having gained wide use is the verb “enjailler” to mean “having fun,” a word originally used in the Côte d’Ivoire in the context of jumping off and on buses in Abidjan. Traditional words have been given new meanings:

In Abidjan this year, people began to call a boyfriend “mon pain” — French for “my bread.” Improvisations soon proliferated: “pain choco” is a cute boyfriend. A sugary bread, a sweet one. A bread just out of the oven is a hot partner.

Interestingly, this development comes at a time when the influence of France in Africa has declined. Some countries have evicted French ambassadors and troops and French has lost its official status in several countries (Mali, Burkina Faso). This demonstrates that language change has a dynamic of its own, often unpredictable and not controllable by political policies or government dictates.

Is this English: “Get down from the car!”?

For anyone familiar with the ethnic and linguistic make-up of Florida, it will come as no surprise that in the Miami area (and elsewhere in the state), everyday language use is characterized by a mix of English and Spanish. “Spanglish” is a term used to describe the influence of English into spoken Spanish. Now a linguist at Florida International University (Phillip Carter) has published findings that indicate a new version of English is emerging in the Miami area which is heavily influenced by Spanish. An article on the research in Scientific American provides some examples:

“We got down from the car and went inside.”
“I made the line to pay for groceries.”
“He made a party to celebrate his son’s birthday.”

For most speakers of North American Spanish, those sentences sound strange. In fact, as Carter points out, these are “literal lexical calques,” i.e., direct, word-for-word translations.

“Get down from the car” instead of “get out of the car” is based on the Spanish phrase “bajar del carro,” which translates, for speakers outside of Miami, as “get out of the car.” But “bajar” means “to get down,” so it makes sense that many Miamians think of “exiting” a car in terms of “getting down” and not “getting out.” Locals often say “married with,” as in “Alex got married with José,” based on the Spanish “casarse con” – literally translated as “married with.” They’ll also say “make a party,” a literal translation of the Spanish “hacer una fiesta.”

Carter provides additional examples that are based on phonetic transfers: “’Thanks God,’ a type of loan translation from ‘gracias a Dios,’ is common in Miami. In this case, speakers analogize the ‘s’ sound at the end of “gracias” and apply it to the English form.

A YouTube video provides further examples from Carter’s research:

Hamilton auf Deutsch

The German cast of Hamilton

Translating always means approximating the meaning of the original text. That is especially the case for translating literary texts, where it’s not just a matter of finding equivalent meanings but also of conveying the feel and style of the original. When the texts are poetry that makes the process even more complicated, as there is a need to make decisions on issues like rhyme, alliteration, meter, rhythm, etc. Then there is the cultural component. Literary texts are embedded in socio-historical contexts, which may be familiar to the intended readers of the original. Translators need to determine whether to simply convey the cultural context as in the original or add or explain so that it is understandable to the readers of the translation. Then there is humor and word play. Music complicates further, as the translated language must fit in to the time constraints built into song lyrics.

Translating the celebrated musical Hamilton into another language has all those complications and more. The story told in the musical is deeply enmeshed in the history and mythology of the founding of the United States. The story and the central figures in it are well known to anyone who has attended schools in the US. For those not having that background, the dynamics of the exchanges among characters in the musical, taken from historical accounts, will be unfamiliar. Then there is the kind of music in Hamilton, namely hip-hop or rap. While that style of music originated in the US, it has spread across the world, so the musical form will likely be familiar, at least to young theater goers. However, in the US, the cultural context of rap is tied closely to African-Americans and that is reflected in the musical, at least in its original stage version and movie, in which the main characters are Black.

So, translating Hamilton into German was no easy task, as pointed out in a recent piece in the New York Times: “Hamilton is a mouthful, even in English. Forty-seven songs; more than 20,000 words; fast-paced lyrics, abundant wordplay, complex rhyming patterns, plus allusions not only to hip-hop and musical theater but also to arcane aspects of early American history.” It wasn’t just the challenge of keeping the musical character as close as possible to the original, it was also the problem linguistically of going from English to German, as the piece states, “a language characterized by multisyllabic compound nouns and sentences that often end with verbs”. Translations from English to German often end up being considerably longer than the original. That was going to be a problem here. So, the translators had to be flexible and creative, finding ways to keep the wordage down, while maintaining the essentials of the content and including as many of the artistic effects in the lyrics as possible. The latter included the internal rhyming that is characteristic of rapping and is used extensively in Hamilton. The translators were able to work with the musical’s creator, Lin-Manuel Miranda, who monitored the translations to make sure the lyrics in German fit the spirit of the original. The New York Times article and a piece in National Public Radio provide examples of the wording in German. The clip below shows the results.

The German Hamilton started playing this month in Hamburg; it will be interesting to see how German theater goers react. One of the interesting aspects of the German production is that the makeup of the cast mirrors that of the New York production, with actors of color playing the main roles. The fact that this was possible in Germany is a demonstration of the surprising diversity in contemporary Germany, with waves of immigration having significantly changed the homogeneity of the population there. In fact, many in the cast are immigrants, or the children of immigrants. Not all are German citizens, but they all speak fluent German, mostly as their first language. For my money, the songs sound very good (and natural!) in German.

“I’m a fraud within my own identity”: Racial Impostor Syndrome

Emily Kwong at age 2 with her grandparents

There was a compelling story this week on NPR dealing with individuals from families with mixed ethnic backgrounds.  In the story, Emily Kwong discusses her decision to learn Mandarin Chinese at the age of 30. Her Dad had grown up speaking Chinese as his first language, but when he started school, he started learning English, and soon switched over completely. When she was little, Emily had learned some Chinese from her grandmother but had not retained any. As an adult, she started to feel increasingly that she needed to try to reconnect with the Chinese side of her identity and decided that necessitated learning the language of her father.

The feeling that Emily has of “feeling that I’m not Chinese enough” has sometimes been labeled racial imposter syndrome, the sense that individuals may have that their lives are in a sense inauthentic because they don’t conform to key aspects of the ethnic heritage with which they identify. That might be related to a lack of knowledge about traditional customs or aspects of the way of life, or be connected to the individual’s appearance. But quite likely there will be a language component. For Emily, learning Chinese has proven to be a challenge, as it is for many native English speakers, particularly if one starts later in life. However, the difficulties she has had with the language pale when compared to the shame she feels in not being able to connect at all with her Chinese-speaking relatives:

I’ve decided that any shame I might feel about imperfect pronunciation, fumbles with grammar is nothing compared to the shame I felt about not knowing the language at all; the shame I feel as my older relatives rattle off dim sum dishes and I stare down the menu pictures, feeling like a fraud within my own identity, missing something I never had in the first place.

This sense of feeling like a fraud within one’s own identity is likely felt by many individuals from immigrant families or those with connections to multiple cultures. NPR a few years ago did a series on such cases, which are increasing as American identities become more mixed. The accounts of experiences are revealing. A woman whose mother is Panamanian and father a white American recounted:

When I was young (20s) and living in the city, I would get asked multiple times a day where I was from, where my people were from, because Allentown, Pennsylvania, clearly wasn’t the answer they were looking for … It always felt like the undercurrent of that question was, “You aren’t white, but you aren’t black. What are you?” But truthfully, I don’t feel like I fit with Latinas either. My Spanish is atrocious and I grew up in rural PA. Even my cousin said a few weeks ago, “Well, you aren’t really Spanish, because your dad is white.” Which gutted me, truly. I identify as Latina.

Another individual featured was a light-skinned biracial woman:

White people like to believe I’m Caucasian like them; I think it makes their life less complicated. But I don’t identify as 100% white, so there always comes a time in the conversation or relationship where I need to ‘out’ myself and tell them that I’m biracial. It’s a vulnerable experience, but it becomes even harder when I’m with black Americans. It may sound strange — and there are so many layers to this that are hard to unpack — but I think what it comes down to is: they have more of a claim to ‘blackness’ than I ever will and therefore have the power to tell me I don’t belong, I’m not enough, that I should stay on the white side of the identity line.

Emily’s effort to retrieve a heritage through language learning may not apply to everyone, but for many it may be an important tool for feeling less like a “fraud”. The experiences in her family illustrate a frequent pattern in immigrant families, namely that the first generation may give up their first language, as Emily’s Dad did, in order to assimilate into mainstream US society. That was widely the case throughout the 20th century in the US. Second or third generation children, like Emily, discover often that not speaking the language of their immigrant parent or parents leaves a void in their sense of who they are.

Is Ukraine a country?

Russian troops in Ukraine

One of the issues in the war between Russia and Ukraine is the insistence by President Putin of Russia that in fact Ukraine is not a “real” country but historically and culturally belongs to Russia. That is part of his stated rationale for the war, along with the concern he continually expresses that Ukraine will join NATO and thus become an enemy on Russia’s doorstep.

Many have debunked the idea that Ukraine does not have a right to be an independent nation. The historical, cultural, and linguistic reasons are laid out nicely in a recent piece by Keith Gessen in the New York Times Magazine. He cites the conversation from one of the posted videos from the war showing a Ukrainian citizen confronting Russian soldiers after their tank runs out of gas on the road to Kyiv. The Ukrainian asks in Russian, “Can I tow you back to Russia?”

Beyond the courage and the sense of humor on display, Gessen comments that the fascinating aspect of the encounter was not so much how it documented the incompetence of the Russian invasion, but rather that the Ukrainian could communicate so easily, so freely, with the Russian soldiers. Not only do many Ukrainians speak Russian, often as their native language, but they switch easily between Russian and Ukrainian, depending on the context and the situation. Although Russian war propaganda makes much of the minority status of Russian speakers in Ukraine, that does not ring true in the everyday life of most Ukrainians:

Russian propaganda claims that the language is discriminated against, and there are people in Russia who believe that you will get shouted at, or even attacked, for speaking Russian in Kyiv. Yet in the videos now emerging from Ukraine, over and over again, people are speaking Russian. Soldiers speak Russian as they fire rocket-propelled grenades at Russian tanks. Locals speak Russian as they survey annihilated Russian columns.

Gessen points out that Americans are likely to associate war with countries far away, where a foreign language is spoken. He points out that this war is altogether different: “Russia invading Ukraine is less like our wars in Iraq or Vietnam and more like the United States invading Canada.”

The President of Ukraine himself, Volodymyr Zelenskyy, switches easily between the two languages. In a speech rebutting Putin’s claim of the shared identity of Ukraine and Russia, he switched to Russian in order to address the Russian people:

Responding to the Kremlin’s claims that it was protecting the separatist regions from Ukrainian plans to take them by force, Zelensky asked whom, exactly, Russia thought he was going to bomb. “Donetsk?” he asked incredulously. “Where I’ve been dozens of times, seen people’s faces, looked into their eyes? Artyoma Street, where I hung out with my friends? Donbas Arena, where I cheered on our boys at the Eurocup? Scherbakov Park, where we all went drinking after our boys lost? Luhansk? The house where the mother of my best friend lives? Where my best friend’s father is buried?”

Zelensky was making the argument that Ukraine is indeed a nation, formed not by history or by language, but by the histories and memories of its people and through personal connections among its citizens. Zelensky continued his speech: “Note that I’m speaking now in Russian, but no one in Russia understands what I’m talking about. These place names, these streets, these families, these events — this is all foreign to you. It’s unfamiliar. This is our land. This is our history. What are you going to fight for? And against whom?”

It is a powerful and effective argument that he is making, insisting that Ukraine is not Russia, but instead its own nation, multiethnic and multilingual, and rooted to a particular place, culture, and history. And one whose continued existence is worth fighting for.

Latin/Greek at Princeton: optional for classics

Princeton University

Princeton University recently announced that it will no longer be necessary for students majoring in classics to learn Latin and Ancient Greek. The rationale given is that such language requirements disadvantage students from high schools not offering Latin. I assume few US high schools now teach Latin, not to mention ancient Greek. It used to be that the area where I live, central Virginia, had the highest enrollment in Latin of all secondary schools in the US. At my university, VCU, we had a thriving Latin program and had difficulty finding enough Latin instructors to accommodate the demand. But those times are now past, a victim of the general decline in language learning in the US. Latin has the additional disadvantage of not being “useful”, i.e. not relating directly to job prospects.

The Princeton decision has generated controversy. While Latin and Greek will continue to be offered as electives, not requiring classics students to take them will inevitably lead to enrollment declines and to classics majors not learning those languages, so crucial for understanding classical culture and literature. Linguist John McWhorter in an article in the Atlantic argues that the decision, made not to disadvantage incoming students from non-elite schools not offering Latin, actually is likely to have the opposite effect: “By ending a requirement that classics majors learn Greek or Latin, Princeton risks amplifying racism instead of curing it.” His argument is that the decision, instead of encouraging disadvantaged students, African-Americans and Latinos, deprives them of the opportunity to expand their knowledge and their identities by learning second languages related to the content they are studying:

The Princeton classics decision also deprives students—and, to the extent that the change is racially focused, Black students in particular—of the pleasant challenge of mastering Latin or Greek. With their rich systems of case marking on nouns and flexible word order, both are difficult for an English speaker…Crucially, you often must go through a phase of drudgery—learning the rules, memorizing vocabulary—before you pass into a phase of mastery and comprehension, like dealing with scales on the piano before playing sonatas. The Princeton decision is discouraging students from even beginning this process. Professors may think of the change as a response to racism, but the implicit intention—sparing Black students the effort of learning Latin or Greek—can be interpreted as racist itself.

Whether one agrees or not with McWhorter’s argument, I find one assertion he makes to be absolutely valid, namely that reading the classics (or any literary work not written in English) in translation is far different from being able to read the text in the original language, no matter how good the literary translation is.

Big data and language learning

The big news in artificial intelligence (AI) this past year was the arrival of GPT-3, a substantially improved version of the “Generative Pre-trained Transformer” from OpenAI, an advanced AI system built on a web of artificial neural networks, deep machine learning, and massive collection of data on human language. The system has been described as a giant step towards the realization of AGI, “artificial general intelligence”, the ability of a system to use language in virtually any domain of human activity. I wrote about this development in the latest issue of Language Learning & Technology, a special journal issue on big data and language learning. I discuss the breakthrough represented by AGI:

Normally, an AI system will be able to deal effectively only within a narrowly defined domain, for which the system has been trained, so as to expect specific language patterns typically used in that context. Google Duplex, for example, does a remarkable job in conversing over the phone with human operators in making dinner reservations or reserving a ride on Uber. GPT-3, in contrast, has been shown to interact through language in a wide variety of genres and content areas: creative writing, journalism, essays, poetry, text-based gaming, and even writing software code. The Guardian newspaper ran an article written by the program, while the New York Times asked it to write about love. A blogger used GPT-3 to write multiple blog posts, subsequently receiving numerous subscribers and notice on tech websites. The fact that many readers were not able to tell that the GPT-3 generated texts were written by an AI system raises questions of trust and authenticity, mirroring the concerns raised about audio and video “deepfakes”, based on training an artificial neural network on many hours of real audio or video footage of the targeted individual.

The system represents a remarkable achievement in its ability to write in natural sounding language (idiomaticity, flow, cohesion). That ability is based on the collection and analysis of huge volumes of speech data collected by crawling the internet, including all of Wikipedia. GPT-3 translates that data into a very large (175 billion!) set of connections or “parameters”, i.e. mathematical representations of patterns. These parameters provide a model of language, based not on rules, but on actual language usage. That allows the system to predict speech sequencing, based on regularly occurring constructions of words and phrases, thereby enabling the machine production of natural-sounding language utterances. One can imagine how powerful GPT-3 could be integrated into a smart personal assistant such as Siri. We are already seeing interesting uses of chatbots and intelligent assistants in language learning. A company called LearnFromAnyone is building on top of GPT-3 a kind of automated tutor, which can take on the identity of famous scientists or writers.

While GPT-3 and other advanced AI systems represent a significant technical achievement, there are, as I discuss in the article, plenty of reasons to be cautious and thoughtful in their use, as is the case generally with big data in both social and educational contexts. While the language generated by GPT-3 mimics what a human might write in terms of language use, compositional structure, and idea development, the texts don’t always make sense in terms of lived human experience, i.e. demonstrating an understanding of social norms and cultural practices. Human beings have the advantage in communicative effectiveness of having lived in the real world and and having developed the pragmatic abilities to generate language that is contingent on human interactions and appropriate to the context. We also can use crucial non-verbal cues, unavailable to a machine: gesture, gaze, posture, intonation, etc.

I argue in the article that a human factor is a crucial mediating factor in implementations of AI systems built on top of big data, particularly in education. Learning analytics (collection of data about student academic performance) tends to treat students as data, not as human beings with complicated lives (especially these days). I discuss these and other ethical and practical issues with data collection and use in the context of D’Ignazio and Klein’s Data feminism (2020). The book explores many examples of inequities in data science, as well as providing useful suggestions for overcoming disparities in data collection (favoring standard language use, for example) and for recognizing and compensating for algorithmic bias.

Free academic speech or racial slur?

USC Professor Patton, removed from course

Earlier this month, the University of Southern California removed business professor Greg Patton from his classroom. His offense? In a lecture on linguistics, he used a Chinese word as an illustrating example of filler words (“um” or “like” in English). So far, so good, but that Chinese expression, 那个, or ne ga sounds a lot like a racial slur in English (the N word). That word is one that I have found to be tremendously useful when I’m in China. It means “that one” and comes in handy ordering in a restaurant when you can just point at a picture of a dish and say “ne ga”, i.e. I’ll have that one. Additionally, native speakers of Mandarin use it in conversation as a filler, as the USC professor was trying to illustrate, making the point that such words or sounds are common across languages. He made clear that the expression was Chinese (not English). Despite that, several African-American students took offense and complained. They wrote a letter to the dean of the School of Business, describing Patton as insensitive and suggested he be removed from his post. They wrote,

There are over 10,000 characters in the Chinese written language and to use this phrase, a clear synonym with this derogatory N-Word term, is hurtful and unacceptable to our USC Marshall community. The negligence and disregard displayed by our professor was very clear in today’s class.

In fact, the letter sent by the students is incorrect, in that the Chinese term is not a “a clear synonym with this derogatory N-Word term, ” in fact not a synonym at all, i.e. a word with the equivalent meaning. It is at most a homonym (words sounding alike), but that is not normally seen as significant or meaningful when you are dealing with two different languages.

As reported in Inside Higher Education, the complaint and removal have been controversial with a petition for Patton’s reinstatement stating:

For him to be censored simply because a Chinese word sounds like an English pejorative term is a mistake and is not appropriate, especially given the educational setting. It also dismisses the fact that Chinese is a real language and has its own pronunciations that have no relation to English.

The professor himself apologized to those students offended, but also reported to Inside Higher Education, “Given the difference in sounds, accent, context and language, I did not connect this in the moment to any English words and certainly not any racial slur.”

In a report on the incident in The Atlantic, a fellow professor (from UCLA), Eugene Volokh, suggested how the Business School Dean should have replied:

This should go without saying, but of course many languages have words that sound vaguely like English epithets or vulgarities, and vice versa … Naturally, USC students are expected to understand this, and recognize that such accidents of pronunciation have nothing to do with any actually insulting or offensive meaning. To the extent that our first reaction to hearing such a word might be shock or upset, part of language education (or education of any sort) is to learn to set that aside. The world’s nearly one billion Mandarin speakers have no obligation to organize their speech to avoid random similarities with English words, and neither do our faculty (or students or anyone else) when they are speaking Mandarin.

On the other hand, as the article discusses, this kind of reply, as reasonable as it sounds, does not take into account the real feelings of the USC students who were upset by the incident.

Advanced tech: No need to learn a language?

From Ciklopea (Juraj Močilac)

I’m currently in Belgium, attending a conference on language learning and technology (EuroCALL 2019). Many topics are presented and discussed at such conferences, but one which came up repeatedly at this one is the use of smart digital services and devices which incorporate voice recognition and voice synthesis, available in multiple languages. Those include Apple’s Siri, Amazon’s Alexa, and Google Assistant, available on mobile phones/watches, dedicated devices, and smart speakers. In addition, machine translation such as Google Translate is constantly improving, as artificial intelligence advances (especially through neural networks) and large collections of language data (corpora) are collected and tagged. There are also dedicated translation devices being marketed, such as Pocketalk and Illi.

I presented a paper on this topic at a previous conference this summer in Taiwan (PPTell 2019). I summarized current developments in this way:

All these projects and devices have been expanding continuously the number of languages supported, with as well language variations included, such as Australian English, alongside British and North American varieties. Amazon has begun an intriguing project to add additional languages to Alexa. An Alexa skill, Cleo, uses crowdsourcing, inviting users to contribute data to support incorporation of additional languages. Speech recognition and synthesis continue to show significant advancements from year to year. Synthesized voices in particular, have improved tremendously, sounding much less robotic. Google Duplex, for example, has rolled out a service which is now available on both Android and iOS devices to allow users to ask Google Assistant to book a dinner reservation at a restaurant. The user specifies the restaurant, date and time, and the number of the party. Google Assistant places a call to the restaurant and engages in an interaction with the restaurant reservation desk. Google has released audio recordings of such calls, in which the artificial voice sounds remarkably human.

Advances in natural language processing (NLP) will impact all digital language services – making the quality of machine translations more reliable, improving the accuracy of speech recognition, enhancing the quality of speech synthesis, and, finally, rendering conversational abilities more human-like. At the same time, advances in chip design, miniaturization, and batteries, will allow sophisticated language services to be made available on mobile, wearable, and implantable devices. We are already seeing devices on the market which move in this direction. Those include Google Pixel earbuds which recognize and translate user speech into a target language and translate back the partner’s speech into the user’s language.

Conference participant, Mark Pegrum, kindly summarized some of the other informationpresented in his blog.

The question I addressed at the conference was, given this scenario, will there still be a need for language learning in the future. Can’t we all just use smart devices instead? My conclusion was no:

Even as language assistants become more sophisticated and capable, few would argue that they represent a satisfactory communication scenario. Holding a phone or device, or using earbuds, creates an awkward barrier, an electronic intermediary. That might work satisfactorily for quick information seeking questions but is hardly inviting for an extended conversation, that is, even if the battery held out long enough. Furthermore, in order to have socially and emotionally fulfilling conversations with a fellow human, a device would need support far beyond transactional language situations. Real language use is not primarily transactional, but social, more about building relationships than achieving a goal. Although language consists of repeating patterns, the direction in which a conversation involves is infinitely variable. Therefore, language support needs to be very robust, to support all the twists and turns of conversational exchanges. Real language use is varied, colorful, and creative and therefore difficult to anticipate. Conversations also don’t develop logically — they progress by stops and starts, including pauses and silences. The verbal language is richly supplemented semantically by paralanguage, facial expressions, and body language. This reality makes NLP all the more difficult. Humans can hear irony and sarcasm in the tone of voice and receive messages accordingly. We understand the clues that nonverbals and the context of the conversation provide for interpreting meaning.

It remains to be seen how technology will evolve to offer language support and instant translation, but despite advances it is hard to imagine a future in which learning a second language is not needed, if not alone for insights it provides into other cultures. Smart technology will continue to improve and offer growing convenience and efficiency in providing language services but is not likely to replace the human process of person-to-person communication and the essentially social nature of language learning.

Learning a language changes lives

Linguist Daniel Everett in the Amazon

As a further example of the growth in support for STEM fields in education (science, technology, engineering, math) in the US, and the concurrent drop in support for the humanities, the state legislature in my state of Virginia this year changed the requirements for the academic diploma from high school to enable computer science to substitute for foreign language study. This aligns with the perception that employment opportunities of the future will call for skills in coding, not in speaking another language. In the US, the trend in education emphasizing vocational training is often linked to the economic downturn following the 2008 financial collapse. The result has been a decade of gradual decline in enrollment in foreign language classes at high schools and universities. As a consequence, fewer US students are learning a second language (at least formally in school). This has unfortunate consequences both individually and society-wide. On a personal level, students lose the opportunity to develop a new identity, by experiencing the world through the different cultural lens that a second language provides. That has social ramifications, as monolinguals tend to cling to the one culture they know and tend therefore to be less receptive to other ways of life.

These thoughts were prompted through a recent piece by linguistic anthropologist Daniel Everett, “Learning another language should be compulsory in every school“. Everett is Dean of Arts and Sciences at Bentley University in Massachusetts. He is best known for his work as a field linguist in the Amazon, learning Pirahã, which he described in the wonderful book, Don’t Sleep, There Are Snakes: Life and Language in the Amazonian Jungle (2008). He’s also famous for his spat with the most famous linguist on earth, Noam Chomsky over the concept of universal grammar, aspects of which Everett found to be contradicted by characteristics of the Pirahã language.

In his essay, Everett, talks about his own language experiences, growing up in California, along the border with Mexico, learning Spanish so well that he joined a Mexican rock ’n’ roll band based in Tijuana:

Learning Spanish changed my life. It taught me more about English, it gave me friendships and connections and respect I never could have otherwise received. Just as learning Portuguese, Pirahã and smatterings of other Amazonian languages continued to transform me during my entire life. Now, after spending most of my adult life in higher education, researching languages, cultures and cognition, I have become more convinced than ever that nothing teaches us about the world and how to think more effectively better than learning new languages. That is why I advocate for fluency in foreign languages. But for this to happen, language-learning needs to make a comeback as a requirement of both primary and secondary education in the United States. Learning another language benefits each learner in at least three ways – pragmatically, neurologically and culturally.

He describes the practical and potential employment advantages of knowing another language. Then he provides an example of how learning a language provides insight into how other cultures see the world:

Beyond the pragmatic benefits to learning languages are humanistic, cultural benefits. It is precisely because not all languages are the same that learning them can expand our understanding of the world. Every language has evolved in a specific geocultural niche, and thus has different ways of talking and codifying the world. But this is precisely why learning them is so beneficial. Take, for example:

John borrowed $10,000 from the bank of Mom and Dad to pay down his college loan.

The Pirahã language of the Amazon has no words for numbers or for borrow, dollars, bank, mom, dad, college or loan. So this sentence cannot be translated into Pirahã (it could be if their culture were to change and they learned about the modern economy). On the other hand, consider a common Pirahã phrase:

Piibooxio xigahapaati.

This phrase means ‘go upriver’ in Pirahã. Innocuous enough, until you realise that Pirahãs use this phrase instead of the less precise ‘turn left’ or ‘turn right’ (which depends on where the speaker and hearer are facing) – this is uttered from deep within the jungle or at the river’s edge – all Pirahãs carry a mental map of tens of thousands of acres of jungle in their heads, and thus know where all points of reference are, whether a river or a specific region of the jungle. They use absolute directions, not relative directions as, say, the average American does when he says ‘turn left’ (vs the absolute direction, ‘turn north’). So to use Pirahã phrases intelligibly requires learning about their local geography.

Everett points to the broadening horizons language learning brings with it:

Language-learning induces reflection both on how we ourselves think and communicate, and how others think. Thus it teaches culture implicitly. Languages should be at the very heart of our educational systems. Learning languages disables our easy and common habit of glossing over differences and failing to understand others and ourselves. You cannot achieve fluency in another language without learning its speakers’ perspectives on the world, and thereby enriching your own conceptual arsenal.

Everett ends with a reference to a video presentation of the method he uses in learning indigenous languages, while not having a common language. In the video, Everett uses Pirahã in eliciting language lessons from a woman speaking Hmong. After the demonstration, Everett describes some of the fascinating aspects of Pirahã, including the variety of communication forms including humming, whistle speech, musical speech, and yelling. Speech acts are radically different from Western norms, with no greeting or leave taking rituals. Men and women speaking Pirahã use different consonants.

The experience of speaking Pirahã involves a very different human experience than speaking English. This is true of all languages, ways of life are embedded deeply in how languages work. Learning to program computers is a highly valuable skill, especially in today’s world, but it shouldn’t substitute for the invaluable human experience of learning a second language.


Anger on display (or not): Ford and Kanvanaugh

Christine Blasey Ford and Brett Kavanaugh

The explosive Senate hearing last week involving US Supreme Court nominee Brett Kavanaugh and Christine Blasey Ford, who has accused Kavanaugh of sexual assault is still resonating this week. The two not only told different versions of what happened at a high school party 36 years ago, they differed significantly in how they expressed themselves. That includes paralanguage — tone, volume — and body language. The situation is of course unique and it’s problematic to extrapolate too far from the specific exchange to general differences in communication styles between men and women (a topic studied extensively by scholars such as Deborah Tannen). Nevertheless, I do believe the exchange fits into familiar patterns of communication that are gender-specific and socially-determined.

Ford’s testimony was calm, measured and deferential, while Kavanaugh was aggressive, belligerent and bullying. Some have commented that of course Kavanaugh was angry, since he was convinced he was innocent of the charges. However, there’s no doubt that Ford was just as convinced that she was telling the truth – many Republicans, including the President, found her testimony to be “credible”. Yet, she did not yell and engage in angry outbursts and accusations – despite the fact that the assault she alleges (attempted rape) would amply justify that behavior.

In a commentary yesterday on NPR, Rebecca Traister, author of Good and Mad: The Revolutionary Power of Women’s Anger, explains that contrast by pointing to the social acceptability (in the US at least) of verbal displays of anger by males, but not by women: “He had in his arsenal the ability to use anger, fury, tears in a way that he felt confident would resonate with the American people”. According to Traister, those same tools were not available to Ford, at least in the sense of being socially acceptable as used by a woman. They are even less likely to be available to women of color in the US:

I think it’s almost impossible for us to conceive of the voices of women of color as being heard loudly enough because they have been so unheard and so marginalized for so long. In fact, it’s women of color who have been the leaders and the leading thinkers of so many of our social movements, in ways that have remained invisible to us.

For Traister, the anger that Ford suppressed in the hearing is likely to re-emerge in a different form:

My argument is not that women’s anger is always righteous. It’s that it’s very often politically potent and yet we’re told not to take it seriously, still. I think that it’s the anger that women are feeling across the country that is having a catalytic connective impact. And this is part of a long process — social movements take a long time. The kind of anger that women are feeling in this moment around Kavanaugh is going to be part of a far longer story that’s going to extend deep into our future.

It’s been interesting to follow as well as the analysis of the encounter from the perspective of nonverbal communication, as in a minute by minute analysis by Jack Brown. His conclusion, from studying the body language: “Judge Kavanaugh’s nonverbal behavior indicates he’s lying as well as deliberately withholding information (lies of omission). He believes he is guilty of sexual assault.” On the other hand, Carol Kinsey Goman, another body language expert, comments:

Habitual and well-rehearsed liars can become quite comfortable with their falsehoods. But the same response is true for liars who believe their own lies. When Ford and Kavanaugh say they are “100% positive,” they may both genuinely believe it.

This echoes an assessment by another analyst, Patti Wood:

Wood says that a key piece of Dr. Ford’s testimony was revelatory; when she said that Kavanaugh and Mark Judge’s laughter — “the uproarious laughter between the two, and their having fun at my expense” — is the strongest impression in her memory. “If Kavanaugh did it and he was laughing, he may not have seen it or felt it as anything but ‘horseplay,'” says Wood. “He may not have had it register in his memory as anything wrong or bad. And if he was drunk, he may not have remembered it at all. This is important because his anger is so strong and he seems so emphatic, and he could actually feel he never did anything like this.”

We will see this week what conclusion the FBI investigation reaches. It seems likely that we will never have factual information that settles the claims once and for all. In the absence of conclusive evidence or convincing witnesses, we will have to continue to rely on our interpretations of the words and behavior of the two individuals involved.

Linguistic relativity: Insights from cognitive science

Blue in Russian: 2 different words

A recent piece in the online magazine, Undark, Scientists Probe an Enduring Question: Can Language Shape Perception?, examines an interesting, new approach to the topic of linguistic relativity, i.e. the extent to which the language we speak influences how we see the world. It starts with a familiar topic within the scholarly debate around this question, namely whether the existence or absence of words for specific colors in a language affects speakers’ perception of colors. This was famously studied by Benjamin Whorf, the very same of the Sapir-Whorf hypothesis, which first laid out this strong connection between language and culture. Whorf studied Native American languages and noted that Navaho speakers split black into two colors and lumped blue and green together into one, thus providing a slightly different color lens on the world than English. Later scholars had doubts, such as Berlin and Kay (1969), who argued that there are cross-linguistic regularities in the encoding of color such that a small number of basic color terms emerge in most languages and that these patterns stem ultimately from biology. This was put forward as evidence against the linguistic relativity hypothesis. Other researchers went further, labeling Whorf’s views as “vaporous mysticism” (Black, 1954),

In this piece in Undark, the example centers on the color blue, and the distinction between shades of blue made in Greek (and in other languages as well) with a separate word, ghalazio, for light blue (akin to goluboy in Russian) and ble for darker blue (siniy in Russian). Instead of the usual procedure in experiments of this kind, the researchers involved (at Bangor University) did not show the participant colors and ask them to name them. Instead, they attached electrodes to their scalps in order to track changes in electrical signals in the brain’s visual system as they looked at different colors. This allowed researchers to observe directly neural activity. In the experiment, native speakers of English and Greek were shown a series of circles or squares of different shades of blues and green. Electrical activity measured in the participants revealed that in the Greek speakers, the visual system responded differently to ghalazio and ble shapes, a change not seen in the English speakers. Thus, for the English speakers everything was “blue”, while the Greeks see the world in a more differentiated way, at least in references to things bluish.

The research team decided to replicate the experiment, this time with shapes:

The Spanish word taza encompasses both cups and mugs, whereas English distinguishes between the two. When Spanish and English speakers were presented with pictures of a cup, mug, or bowl, the difference between the cup and mug elicited greater electrical activity in the brains of English speakers than in Spanish speakers.

Such experiments from cognitive scientists seem likely to continue – and to continue the debate around linguistic relativity. They do seem to provide scientific evidence that our brains are wired by language, at least in some areas. It would be interesting to see experiments of related phenomena, such as the presence or absence of grammatical genders in languages. As Guy Deutscher described in the NY Times Magazine a few years ago, psychologists have compared associative characteristics of objects with different noun genders in different languages (such as the feminine word for bridge in German, die Brücke with the masculine el puente in Spanish):

When speakers were asked to grade various objects on a range of characteristics, Spanish speakers deemed bridges, clocks and violins to have more “manly properties” like strength, but Germans tended to think of them as more slender or elegant. With objects like mountains or chairs, which are “he” in German but “she” in Spanish, the effect was reversed.

Another area of interest would be time perception, as discussed in a recent article in Popular Science, “The language you speak changes your perception of time”:

Different languages frame time differently. Swedish and English speakers, for example, tend to think of time in terms of distance—what a long day, we say. Time becomes an expanse one has to traverse. Spanish and Greek speakers, on the other hand, tend to think of time in terms of volume—what a full day, they exclaim. Time becomes a container to be filled.

The article cites a study in the Journal of Experimental Psychology: General, in which Spanish and Swedish speakers were asked to estimate time based in either distance (how long a line took to grow) or volume (how long it took for a container to fill). The results showed that there were indeed clear differences in the two groups in how they measured time in terms of length or volume, in other words, that the language they spoke affected how they estimated the relative passage of time. The original article, not surprisingly, is full of cautions and caveats. That is likely to be the case in such studies and provides a rationale for cognitive scientists to continue their efforts to measure brain activity related to language and culture.

Best time ever to learn languages?

MinorMynas app for language learning

The BBC recently ran a story, “Is this the best time in history to learn a foreign language?”, with the subtitle, “Today’s youngest generation is more multilingual and wired than ever. Could the tech they’re using breed a global army of polyglots?”. I would say that depends. Certainly, the opportunities are there now for language learning through resources on the Internet, most available on mobile phones as well. The article provides an example, through this combination of collaboration and smartphones, with a profile of Hillary Yip, a 13-year-old student from Hong Kong, who created a smartphone app (MinorMynas) for enabling young people to connect with one another for the explicit goal of language learning. The article cites the increase in migration patterns worldwide that increases the multicultural make-up of the population in many countries as one of the developments that is leading to greater interest in language learning: “This increased migration, especially in cities, brings people with a wide variety of cultural and linguistic backgrounds into close contact. Could a more multicultural world lead to a more multilingual generation?”. Yes, that could be the case, certainly among the younger generation, as the article points out. Unfortunately, in older populations, the influx of newcomers from different cultures may lead to discrimination and nativist political views, as we’ve seen in the US, UK, and elsewhere.

The article references the idea of “translanguaging”, the informal mixing of languages common today on the Internet, as well as in many multicultural classrooms. The opportunities for encountering other languages online are increasing, as social media enables contact with people from around the globe. One of the new options is the availability of streaming videos–especially in English– in the target language (with native language subtitles) now available in many countries through Netflix and other services. In many countries, TV shows and movies, shown on TV networks, are dubbed. However, in some countries, such as the Netherlands and Scandinavian countries, videos are shown in the original audio soundtrack. It’s no coincidence that citizens of those countries have typically had better English language skills. The availability of target language videos on commercial services as well as on YouTube offers the possibility of learning or refreshing a language through entertainment. This is, as the article discusses, a way to learn without having the goal to learn, or even without the realization that one is learning.

Several recent studies document this process for learners in Europe and in Brazil. The study from Brazil found that “fully autonomous self-instructed learners” of English gained a high level of proficiency without formal instruction, revealing “how the new affordances for naturalistic learning through the Internet have transformed informal language learning, enabling significant numbers of independent, informal learners in foreign language contexts to achieve very high levels of proficiency” (Cole & Vanderplank, 2016, p. 31). In fact, the study showed that the autonomous learners studied had fewer “fossilized errors” than classroom-based learners at a similar proficiency level, that is, fewer persistent, baked-in errors in grammar or word usage. This may be a wake-up call for instructed language learning, to look at more ways to encourage students to make use of online language resources.

Cole, Jason, and Robert Vanderplank. 2016. “Comparing autonomous and class-based learners in Brazil: Evidence for the present-day advantages of informal, out-of-class learning.” System 61: 31-42.