The buzz around artificial intelligence (AI) technologies like ChatGPT is palpable. People are both optimistic and frightened by the possibilities of these tools. Clearly, these technologies will change how people write. But in terms of what people write, these technologies seem to be embracing the status quo.
In fact, the way these tools are currently built appears to homogenise writing – making everything sound the same. And writing that sounds the same is not just boring; it also perpetuates inequity.
When writing tools prioritise one way of writing over another, they reinforce existing hierarchies that unfairly position Standard American English (SAE) and the Queen’s English over other languages and ways of writing.
How does ChatGPT work?
Technologies like ChatGPT are called large language models (LLMs). LLMs provide textual responses to human commands, by using machine learning to study patterns of words in a massive archive of texts.
Crucially, however, ChatGPT does not know the meaning of words. ChatGPT generates definitions by sorting through a mountain of definitions and then collating those into a single response that suits the context of a query.
In other words, without meaning as its guide, ChatGPT responds to queries by relying on context clues, stylistic structures, writing forms, linguistic patterns and word frequency.
This functionality means that, by default, ChatGPT perpetuates dominant modes of writing and language use while sidelining less common ones.
Erasing diversity
Dominant modes of writing don’t become dominant by accident. They become dominant because one social group wants to assert power over another social group.
There is not, for example, one kind of English. There are many Englishes.
The decision to prioritise Standard American English in many US classrooms, for example, means that speakers of Black English – a language with its own grammar, lexicon and remarkable history of resistance – are penalised and shamed for writing as they speak.
Similarly, in Aotearoa New Zealand, the Queen’s English became dominant not because it’s intrinsically better than te reo Māori. Rather, European colonisers wanted to stamp out Māori culture, and writing in the Queen’s English became a key tool for furthering that objective. In the 20th century, students were regularly beaten for speaking Māori in schools.
‘Can I see your parts list?’ What AI’s attempted chat-up lines tell us about computer-generated language
Going against the default
Supporters of ChatGPT will be quick to note that ChatGPT can read, analyse and generate content in many languages, including in Black English and te reo Māori.
But the concern is not about what ChatGPT can do.
It’s about what its default settings are. It’s about how ChatGPT is configured to treat some forms of writing as normal, typical and expected. And it’s about how ChatGPT requires a special request to generate non-normative forms of writing.
This problematic default behaviour also occurs in ChatGPT’s sister programme, Dall-E 2. This image-generating AI was asked to create an image for this article based on this prompt: “close up photo of hands typing on a laptop.” The programme created four images. All had white masculine hands.
The programme needed a more specific prompt to generate an image that included a person of colour because even the ways that AI visualises writing is dominated by white men.
Ultimately, this kind of algorithmic bias continues to make white English-speaking men the standard of writing culture, while ushering everyone else to the margins.
How did it get like this?
It’s no surprise that ChatGPT’s default functionality seems to prioritise forms of English writing developed by white people. White English-speaking men have long dominated many writing-intensive sectors, including journalism, law, politics, medicine, computer science and academia.
These white English-speaking men have collectively written billions of words, many times more than their colleagues of colour. The sheer volume of words these authors have written means that they likely constitute the majority of ChatGPT’s learning models, even though ChatGPT’s parent company, OpenAI, doesn’t publicly reveal its source material.
So when users ask ChatGPT to generate content in any of these disciplines, the default output is written in the voice, style and language of those same white English-speaking men.
AI and the future of work: 5 experts on what ChatGPT, DALL-E and other AI tools mean for artists and knowledge workers
Challenging the norm
Some people will say that we need defaults and standards in writing. They argue that we need to teach people to write in the Queen’s English or SAE so that people don’t miss out on jobs and promotions because they write in a different way.
But that line of thinking just means capitulating to workplace prejudice and reinforcing an unjust system through our participation in it. Instead, other scholars say we need to challenge those unfair writing standards and encourage writers to embrace the rich rhetorical possibilities in their linguistic diversity.
Educators who want to embrace linguistic diversity might be tempted to ban text-generating AI from their schools and universities.
But it’s worth remembering that writing itself is a technology that has been, and still is, used to further inequality. Literary scholar Alice Te Punga Somerville calls this “the inextricability of writing from historical and ongoing violence.”.
In response to this threat, however, Professor Somerville does not advocate abandoning writing altogether. Rather, she insists on using the tool of writing critically and creatively to resist oppression.
Taking her lead, educators might instead encourage students to develop new ways of deploying these tools to compose a more equitable future. Doing so means, as Professor Vershawn Young says in Black English
that good writin gone look and sound a bit different than some may now expect. And another real, real good result is we gone help reduce prejudice.
ChatGPT is a data privacy nightmare. If you’ve ever posted online, you ought to be concerned