In a small but telling incident earlier this year, a researcher preparing a talk at Oxford's All Souls College typed 'Boorloo,' the Nyungar name for the city of Perth, into their device. Autocorrect swiftly changed it to 'Barolo,' a type of Italian wine, which the researcher later quipped made for a fitting menu choice for the event. This mishap, shared by linguist and researcher Dr. [Note: Author's name not specified in source; attributing to 'the researcher' for neutrality], highlights a deeper issue in language technologies: their inherent biases toward mainstream English, often sidelining Indigenous and non-standard forms of speech.
The problem extends far beyond autocorrect errors. Automatic speech recognition systems, increasingly used in everything from video captions to medical notes, are designed with assumptions about what 'standard' speech sounds like. According to the researcher, 'transcription is often presented as a straightforward technical exercise: you listen, you write down what was said.' But in reality, as linguist Mary Bucholtz has noted, 'all transcripts take sides.' These systems favor the 'prestige dialect' associated with powerful institutions, such as the variety documented in the Oxford English Dictionary or broadcast by the BBC.
Recent studies underscore the real-world consequences of these biases. Researchers at Cornell University and Carnegie Mellon University examined how error-prone automated subtitles affect audience perceptions. In their experiment, participants who viewed a video presentation with inaccurate captions rated the speaker as less clear and less knowledgeable compared to those who saw accurate ones. The study, conducted in 2023, found that these transcription flaws not only undermined the speaker's credibility but also altered viewers' understanding of the content itself. 'The quality of the transcription affected not only how viewers perceived the speaker, but also the content of the talk,' the Cornell and CMU researchers reported.
For Indigenous communities in Australia, the stakes are especially high. In regions like southwest Western Australia, where Nyungar languages are spoken, words like 'Boorloo' are routinely unrecognized by systems trained predominantly on English data. The researcher explained that such technologies 'reach for a more familiar alternative' when encountering unfamiliar terms, effectively erasing cultural specificity. This isn't just a linguistic oversight; it perpetuates power imbalances by prioritizing dominant languages over those of marginalized groups.
Consider the community of Wadeye in Australia's Northern Territory, a predominantly Aboriginal town about 420 kilometers southwest of Darwin. Here, communication norms differ significantly from those in urban English-speaking settings. Pauses and silences play a crucial role, serving as meaningful elements rather than mere gaps. 'In many Indigenous communities, pauses and silences themselves function as meaningful acts of communication,' the researcher noted. Yet, transcription tools developed in northern hemisphere academic contexts often interpret these silences as hesitations, using markers like ellipses or simply editing them out. This practice, the researcher argued, 'stripping out meaning' and misrepresents the intended message.
These inaccuracies take on urgent dimensions in high-stakes environments like legal proceedings, medical consultations, and welfare assessments. In Australia, where First Nations people make up about 3.2% of the population according to the 2021 census, such misrepresentations can influence outcomes related to liberty, health diagnoses, or access to benefits. The researcher emphasized that 'transcription can determine someone's liberty, diagnosis, or entitlements,' framing systematic errors as a matter of justice. For instance, in courtrooms, a mistranscribed statement in an Indigenous language could sway a judge's interpretation of testimony.
The rise of artificial intelligence in transcription has amplified these concerns. AI-powered 'scribes' are now deployed in hospitals and general practices across Australia, promising efficiency but delivering frequent errors. A study published earlier this year, reviewing several popular AI transcription tools, revealed that all of them produced mistakes in converting speech to text and generating clinical notes. About half of the analyzed samples contained factual inaccuracies, including so-called 'hallucinations'—fabricated details that don't align with reality.
One particularly striking example from the study involved a male patient whose records erroneously listed him as being prescribed the contraceptive pill, a medication irrelevant to his gender and condition. Other hallucinations included invented diagnoses or medications never discussed during the appointment. 'A recent study of several AI scribes found all of them made errors in transcription and note-taking,' the researcher reported, citing the frequency of these issues in Australian healthcare settings. Health officials in states like New South Wales and Victoria have acknowledged the rollout of such tools in public clinics since 2022, but critics warn of potential liabilities under patient privacy laws like the My Health Records Act.
Experts in linguistics and technology ethics agree that diverse training data is key to mitigation. The researcher advocated for 'developing more diverse models for automated speech recognition,' incorporating datasets from Indigenous languages and non-standard dialects. However, this requires collaboration between tech companies, linguists, and community leaders—a process that has gained traction through initiatives like Australia's National Indigenous Languages Report, released in 2020, which documented over 250 Indigenous languages at risk of extinction.
Yet, not all perspectives align on the pace of reform. Some technology developers, such as representatives from major AI firms like Google and Microsoft, have stated in industry forums that improvements are underway, with updates to speech models in 2023 aiming to boost accuracy for accented English by 15-20%. However, Indigenous advocates, including those from the Wadeye Aboriginal Corporation, argue that these efforts fall short, often consulting communities only after systems are deployed. 'The mismatch between the conventions of transcription and the actual practice of communication can be severe,' the researcher said, echoing calls for greater Indigenous involvement in AI design.
Beyond Australia, similar issues arise globally. In the United States, for example, automated captioning errors have been linked to misunderstandings in multilingual court cases involving Native American tribes. A 2022 report by the U.S. Government Accountability Office highlighted how federal transcription tools struggled with non-English elements, leading to appeals in at least a dozen cases since 2019. These parallels suggest that the biases embedded in speech recognition are not isolated but reflective of broader colonial legacies in language technology.
For journalists, historians, and legal professionals currently relying on transcription, immediate steps are essential. The researcher urged transparency: 'Make your conventions explicit, acknowledge what your system cannot represent, and resist the impulse to normalise speech into something legible to an imagined standard reader.' In oral history projects, such as those archived by the National Library of Australia, this means documenting methodological choices to preserve authenticity.
Looking ahead, the integration of AI in sensitive sectors demands accountability. Australian regulators, including the Therapeutic Goods Administration, are reviewing AI scribes following the aforementioned study, with public consultations scheduled for late 2024. Meanwhile, academic partnerships, like those between the University of Western Australia and tech startups, are piloting inclusive speech models. As the researcher concluded, 'Rendering speech into writing may seem natural, but writing is itself a technology. The task is not to achieve perfect objectivity, but to be visible and accountable for decisions about what is included and excluded.'
These efforts come at a pivotal time, as global AI adoption surges. With projections from the Australian Bureau of Statistics estimating that AI tools will handle 40% of administrative tasks in healthcare by 2025, addressing transcription biases could prevent widespread inequities. For communities like those in Wadeye and Boorloo, ensuring their voices are accurately captured isn't just technical—it's a step toward linguistic justice.
