Monday 20 January 2020

Accent Bias: Voices at Work

Continuing our series of posts related to the 'Accent Bias in Britain' project, in this blog post we discuss some findings from our research which investigated current attitudes to accents in Britain.


In our last blog post, we explored some of the findings of the second part of our study which investigated how the UK public evaluated 5 different accents in mock interviews. The third part of our study, detailed here, investigated whether people in positions of power such as recruiters would exhibit the same type of accent biases. 

Our study focuses on a profession that has been previously described as lacking diversity, Law. We were interested in examining whether accent bias interferes with judgements of professional skill. In other words, would a candidate with, say a Multicultural London English accent, be perceived as less professional or competent as their Received Pronunciation speaking peers? 

To investigate this question, we played the same mock interviews as described in our last blog post to 61 legal professionals.We prepared 10 short mock interview answers, varying between ‘good’ and ‘poor’ quality. Before we conducted the experiment, these answers were independently judged as 'good' or 'poor' by a group of 25 legal professionals otherwise unrelated to the project.

To create the mock interviews, we had 10 speakers (2 of each accent) record 10 good and 10 poor interview responses. This resulted in 100 recordings. The accents we tested were: Multicultural London English (MLE), Estuary English (EE), Received Pronunciation (RP), General Northern English (GNE), and Urban West Yorkshire English (UWYE). 

From the 100 recordings, our 61 legal professionals heard a random selection of 10 interview answers. They were then asked to evaluate whether they thought the answer was a 'good' answer or a 'poor' answer. They were asked to indicate this on a 10-point scale, responding to the following questions:

  1. “How would you rate the overall quality of the candidate’s answer?”
  2. “Does the candidate’s answer show relevant expertise and knowledge?”
  3. “In your opinion, how likely is it that the candidate will succeed as a lawyer?”
  4. “Is the candidate somebody that you personally would like to work with?”
  5. “How likely would you be to recommend hiring this candidate?”
When we analysed our data, we identified a surprising effect. Whilst the general public displayed a great deal of accent bias in judging the competency of a job candidate, the lawyers did not follow this pattern. In fact, the professions did not show significant preferences for Received Pronunciation (RP) or General Northern English (GNE), nor did they show a consistent dispreference for working class or non-white accents. Instead, they showed a consistent ability to judge the quality of an answer as 'good' or 'poor' regardless of the accent the answer was presented to them in. Their answers very closely matched the answers given by the group of professionals who rated the quality of the written answers. 


The graph above shows this effect. The high quality answers are in yellow and the lower quality answers are in green. As you should be able to see, across the five different accents (see the x-axis on the bottom), the ratings remain relatively the same. At the same time, however, it is worth noting that EE & MLE receive the lowest ratings of all the accents. 

Note, however, that RP is also lower rated than some of the other accents. This is surprising given that RP was evaluated as the most prestigious accent in the label study. It's possible that this ranking might be related to the association of RP with a higher level of education, so there is a greater expectation of these individuals. 

It is also interesting to note that some of the social factors seen to effect the general public's responses do not seem to influence the professional's judgements. The age and regional origin of  legal professionals did not affect how they responded to job candidates, unlike what we found among the general public. Their Motivation to Control a Prejudiced Response (MCPR) - a psychological factor that had a strong effect on how listeners behaved in our public survey - also did not effect their ratings. 

Our findings therefore suggest that when legal professionals are asked to judge the suitability of a candidate, they are able to switch off biases and attend very well to the quality of an answer, judging the competency of the individual independently of their accent. 

Of course, however, the current study simulates just one small part of hiring candidates. It doesn't look at accent bias in other aspects of professional life, like informal interaction during the interview or everyday experiences on the job. So, it's possible that accent bias might influence the candidate's progression later on down the line. 

At least in terms of hiring though, it looks like it's relatively good news for speakers of regional and 'non-standard' accents! 

This summary was written by Christian Ilbury


Monday 6 January 2020

Accent Bias: Responses to Voices

Continuing our series of posts related to the 'Accent Bias in Britain' project, in this blog post we discuss some findings from our research which investigated current attitudes to accents in Britain.



In the most recent blog post, we explored the findings of the first part of our study which investigated attitudes to accent labels. The second part of our study, detailed here, investigated how people responded to recordings of speakers with different accents to see if the same accent bias exists in speech. 

To examine these questions, we recorded 10 speakers of 5 different accents (2 speakers each). These accents were Multicultural London English (MLE), Estuary English (EE), Received Pronunciation (RP), General Northern English (GNE), and Urban West Yorkshire English (UWYE). Speakers of these accents were recorded reading scripted mock interview answers. 

These recordings were then played to over 1,100 participants aged between 18-79 from across the country. The sample of participants was balanced for both ethnicity and gender. 

For each of the 10 mock interview answers the participants heard, they were asked to evaluate the candidate's performance, knowledge, suitability, and hireability for a job. Participants were asked to rate the candidate on a 10-point scale - where 10 is the highest. They were asked to respond to questions such as:

  1. “How would you rate the overall quality of the candidate's answer?”
  2. “Does the candidate's answer show expert knowledge?”
  3. “How likely is it that the candidate will succeed as a lawyer?”
  4. “Is the candidate somebody that you personally would like to work with?”
  5. “How would you rate the candidate overall?”
The participants also provided information on their age, social background, and education. 

When we analysed the results, we found a significant effect of the listener's age. Older listeners generally rated the two southern accents (MLE and EE) lower than all of the other accents. Younger participants, however, did not show this pattern. 

You can see this effect in the graph below. On the right are the older participants and on the left, the younger participants. The higher the line, the more positive the evaluation. As one can see, the ratings drop when you move from the younger respondents to their older peers. 

Is accent bias decreasing or is this just 'age-grading'?

This could mean one of two things. It could be that general attitudes to accents are changing, such that younger listeners will continue to exhibit the same accent preferences later on in life. On the other hand, it's possible that this could be evidence of age-grading. This is where young people might be more tolerant of accent diversity in their early years but become more critical as they get older.

A second finding of this study was that people's evaluations of accents in the responses to the interview questions depends on the type of question being answered. In questions that require a degree of technical or specialist knowledge, like those questions which asked specific details about law, all accents were rated more favourably. In more general questions, such as those which asked personal details or the work experience of the candidate, the accents were downrated much more.

Degree of expertise and accent rating

The effect of the 'expertise' required is shown in the graph above. The yellow line indicates 'expert' answers and the green line indicates 'non-expert' answers. As you should be able to see, all accents are rated much lower when the answer is a 'non-expert' answer than for an 'expert' answer. 

We also asked participants a series of questions aimed to test how prejudiced they were. We proposed that the more prejudiced people were, the lower their ratings of the different accents would be. In fact, this is exactly what we find. See the graph below. 

More prejudiced listeners were more likely to downrate all of the accents  

Those who reported they were more likely to be prejudiced towards different accents showed much lower ratings than those who were more likely to control their prejudice. The graph above shows ratings depending on MCPR (Motivation to Control a Prejudice Response). The blue line is those who reported that they are not prejudiced towards different accents, whereas the green line is those who report exhibiting more prejudice. 

What these results suggest is that there is a a systematic bias against certain accents in England (particularly Southern working-class varieties), whereas RP is evaluated much more positively and is perceived to be the most suitable for professional employment.

However, these results are reported for the general public. Would we see the same types of evaluations amongst those who are responsible for hiring candidates? In the next blog post, we explore this question. In the meantime, you can find our more about the project by visiting the project website


This summary was written by Christian Ilbury