Hostname: page-component-cb9f654ff-5kfdg Total loading time: 0 Render date: 2025-08-21T08:46:09.622Z Has data issue: false hasContentIssue false

Evidence that cultural groups differ in their abilities to detect fake accents: a follow up

Published online by Cambridge University Press:  01 July 2025

Jonathan R. Goodman*
Affiliation:
Department of Psychiatry, University of Cambridge, Cambridge, UK Wellcome Sanger Institute, Hinxton, UK
Robert A. Foley
Affiliation:
Leverhulme Centre for Human Evolutionary Studies, University of Cambridge, Cambridge, UK
*
Corresponding author: Jonathan R. Goodman; Email: jrg74@cam.ac.uk

Abstract

We recently reported that cultural group membership may be a predictor of the likelihood that an individual will detect a faked accent in a recording. Here, we present follow-up data to our original study using a larger data set comprised of responses from the across the world. Our findings are in line with our previous work and suggest that native listeners perform better at this task than do non-native listeners overall, although with some between-group variation. We discuss our findings within the context of signals of trustworthiness and suggest future avenues of research.

Information

Type
Commentary
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press.

Social media summary

Native listeners from Belfast, Dublin, and Glasgow excel at detecting fake accents, unlike RP listeners and non-natives.

We recently reported (Goodman et al., Reference Goodman, Crema, Nolan, Cohen and Foley2024) that cultural group membership may be a predictor of the likelihood that an individual will detect a faked accent in a recording. Native listeners across regions proved strong at recognising attempts at mimicking their own accents.

However, this finding was not true across all listener groups: of the Belfast, Bristol, Dublin, Essex, Glasgow, Northeast England, and Received Pronunciation (RP) accents, only speakers from Glasgow, Belfast, Dublin, and Northeast England were better at the task than were those who were non-native speakers. Notably, RP listeners of RP recordings were only slightly better than chance at detecting fakers (95% credible interval: 51%–67%), similar to non-native speakers; listeners from Belfast, Dublin, Glasgow, and Northeast England had 95% credible intervals of a correct answer ranging from approximately 62% to 85%.

Following publication of this research and extensive media engagement, we duplicated the original study via newspaper websites. The response was considerable and received approval from the University of Cambridge’s Department of Archaeology Ethics Committee for analysis of data collected (approval received in February 2025); here we present a new analysis based on 28,100 responses from 2033 participants.

We reported the study design elsewhere (Goodman et al., Reference Goodman, Crema, Nolan, Cohen and Foley2024). Our design here was identical except that we allowed participation from anywhere in the world, whereas for our previous experiment a criterion for inclusion was being raised in the UK or Ireland. Of 3020 individuals who expressed online interest in participation, 2033 answered at least one question from the experiment. The vast majority (1852) did not speak in an accent we investigated originally, with varying numbers for the seven accents evaluated (Belfast, 28; Bristol, 4; Dublin, 42; Essex, 8; Glasgow, 34; Northeast England, 5; RP, 60; see the online supplementary material for details: https://github.com/jonathanrgoodman/accents3).

First, we calculated the 95% Jeffreys interval on the entire data set, which indicated a likelihood of a correct answer of between 53.83% and 55.25%, which was lower than the interval in our previously published sample (60.32%–62.47%). We next fitted a Bayesian hierarchical model using the brms package in R (R Core Team, 2025; Bürkner et al., Reference Bürkner, Gabry, Weber, Johnson, Modrak, Badr and Mills2022) and the tidyr (Wickham & Girlich, Reference Wickham and Girlich2022), ggplot2 (Wickham, Reference Wickham2016), and ggridges (Wilke, Reference Wilke2021) packages. The findings suggested that the lower probability interval for the overall sample was driven by the inclusion of the large number of participants recruited who did not speak in a study accent: in this subgroup, the 95% credible interval was 50.47%–57.25%, which was lower than among listeners who spoke in a study accent (63.99%–70.51%; difference: −15.38% to −10.64%).

A subsequent model using listener–accent group as a predictive variable showed, similar to our initial findings, that participants from Belfast, Dublin, and Glasgow drove this effect (the sample size from Northeast England was, however, too small to determine any effect). RP listeners, as with our initial findings, did not perform at a rate higher than that seen with non-study accent listeners (Table 1); these inferences are based on intervals excluding zero).

Table 1. 95% credible intervals for a correct answer by listener subgroup with differences, using non-study accent as a reference. Belfast, Dublin, and Glasgow native listeners all performed better at the task than did those who did not speak in a study accent

Finally, we obtained participant locations using longitude and latitude data collected during the online survey (however, these data points have been removed on the study’s Github page to protect participant anonymity). Of 1898 participants with location data available, 1500 were based in the United States, followed by 126 in the UK and 64 in Canada (full country-level data are available in the online supplementary material). We used this information as a proxy to group participants into one of four groups: study accent listener (169), other UK-/Ireland-based (80), other English-speaking country-based (1533), non-English-speaking country-based (116).

Using these subgroups, we fitted a final Bayesian hierarchical model using listener group as a predictive variable; the 95% credible intervals are displayed in Figure 1 and the online supplementary material. As predicted by our original findings, study accent listeners had the highest probability of success, followed by other listeners from UK or Ireland; there was, however, no notable difference between other listeners regardless of whether they were from an English-speaking country. Given that we do not know whether these listeners were from an English-speaking country originally, we cannot comment further on this finding, although we suggest this as a future avenue of inquiry.

Figure 1. 95% credible intervals for a correct response by listener subgroup. Study accent listeners had the highest probability of a correct response (62.49%–70.12%), followed by UK/Ireland (53.18%–62.44%), and other English-speaking country (50.03%–56.85%) and non-English-speaking country (49.72%–58.32%).

These follow-up findings support those from our original sample, and suggest further that local importance placed on accents in cities such as Belfast, Dublin, and Glasgow, owing to cultural and historical factors, affects the likelihood that a native listener will detect accent fakery. Furthermore, the lack of difference between RP speaker–listeners and the large cohort of participants from the USA suggests that there is nothing peculiar about the Belfast, Dublin, and Glasgow accents that makes them more difficult to fake – but rather that psychological qualities among the listeners, stemming from local cultural evolution, are driving the effects noted here.

We propose that these findings justify a broader study of what common signals determine whether a signaller appears trustworthy to receivers, following suggestions that signal theory plays a major role in social selection frameworks (Gambetta & Hamill, Reference Gambetta and Hamill2005; Goodman & Milne, Reference Goodman and Milne2024). Accents are likely to be important signals in the formation of human bonds, helping listeners to determine the social identity of the speaker and place cooperative trust accordingly (Cohen, Reference Cohen2012; Pietraszewski & Schwartz, Reference Pietraszewski and Schwartz2014a, Reference Pietraszewski and Schwartz2014b). Other signals, including body language, use of Duchenne smiles, tattoos, and dress are, however, also likely to play a role – suggesting a broader investigation of these qualities together would be helpful in determining how humans create lasting bonds based on limited information about a partner’s trustworthiness.

Acknowledgements

We are indebted to Ian Sample, William Hunter, and Eric Niiler for helping to disseminate this follow-up to our previous study.

Author Contributions

JRG and RAF conceived and produced the study; JRG collected data, performed the analyses, and wrote the article; RAF supported with editing and overall supervision.

Financial Support

This research received no specific grant from any funding agency, commercial or not-for-profit sectors.

Competing interests

JRG and RAF declare none.

Research Transparency and Reproducibility

All data and code are available at https://github.com/jonathanrgoodman/accents3.

References

Bürkner, P.-C., Gabry, J., Weber, S., Johnson, A., Modrak, M., Badr, H. S., & Mills, S. C. (2022) brms: Bayesian Regression Models using ‘Stan’. 2.18.0.Google Scholar
Cohen, E. (2012). The evolution of tag-based cooperation in humans: The case for accent. Current Anthropology, 53(5), 588616.Google Scholar
Gambetta, D., & Hamill, H. (2005). Streetwise: How taxi drivers establish their customers’ trustworthiness. Russell Sage Foundation.Google Scholar
Goodman, J. R., Crema, E., Nolan, F., Cohen, E., & Foley, R. A. (2024). Evidence that cultural groups differ in their abilities to detect fake accents. Evolutionary Human Sciences, 6, e46. https://doi.org/10.1017/ehs.2024.36Google Scholar
Goodman, J. R., & Milne, R. (2024). Signalling and rich trustworthiness in data-driven healthcare: An interdisciplinary approach. Data & Policy 6, e62. https://doi.org/10.1017/dap.2024.74Google Scholar
Pietraszewski, D., & Schwartz, A. (2014a). Evidence that accent is a dimension of social categorization, not a byproduct of perceptual salience, familiarity, or ease-of-processing. Evolution and Human Behavior, 35(1), 4350. https://doi.org/10.1016/j.evolhumbehav.2013.09.006Google Scholar
Pietraszewski, D., & Schwartz, A. (2014b) Evidence that accent is a dedicated dimension of social categorization, not a byproduct of coalitional categorization. Evolution and Human Behavior 35(1), 5157. https://doi.org/10.1016/j.evolhumbehav.2013.09.005Google Scholar
R Core Team. (2025). R: A language and environment for statistical computing. R Foundation for Statistical Computing.Google Scholar
Wickham, H. (2016). ggplot2: Elegant graphics for data analysis. Springer-Verlag.Google Scholar
Wickham, H., & Girlich, M., & R Studio. (2022). tidyr: Tidy messy data (1.2.1).Google Scholar
Wilke, C. O. (2021). ggridges: Ridgeline plots in ‘ggplot2’ (R package version 0.5.3).Google Scholar
Figure 0

Table 1. 95% credible intervals for a correct answer by listener subgroup with differences, using non-study accent as a reference. Belfast, Dublin, and Glasgow native listeners all performed better at the task than did those who did not speak in a study accent

Figure 1

Figure 1. 95% credible intervals for a correct response by listener subgroup. Study accent listeners had the highest probability of a correct response (62.49%–70.12%), followed by UK/Ireland (53.18%–62.44%), and other English-speaking country (50.03%–56.85%) and non-English-speaking country (49.72%–58.32%).