From research to Develop Diverse: our process explained

Insights from our data team

Being confronted with subconscious biases can be a mixed experience. Sometimes it will make perfect sense why we recommend you to change a phrase, other times you might wonder how we got that particular phrase as biased. 

But don’t worry, there are reasons behind it all. One of them is that subconscious bias is.. subconscious. In a way, it is not supposed to make sense from the beginning, but using Develop Diverse should make you more conscious along the way.

Another reason is that the phrases and concepts we highlight has been proved to negatively influence how people read your job ad. 

So, if you ever wondered how we got from research papers to a multilingual bias detecting model, do read on.

1st step: Literature Review

When Develop Diverse set out to create an AI model for inclusive writing, there was no existing overview, library, or database of biases and stereotypes. We each brought our own personal experiences with stereotyping, and our own unconscious stereotypes about others. Naturally, these individual experiences couldn’t form the foundation of an AI model designed to work across multiple countries.

So, we took on the task to create that essential library of biases and stereotypes ourselves. This required extensive reading and research.

A team of linguistic researchers reviewed and summarized over 100 academic articles published by universities and independent research institutions. The studies investigate biases and stereotypes in the workplace approached the subject from a sociological, psychological, or linguistic perspective, ensuring our model is both holistic and nuanced.

Some of the most influential works we rely on include:

  • Debiasing Job Ads by Replacing Masculine Language Increases Gender Diversity of Applicant Pools, He (2025)
  • Does Gendered Wording in Job Advertisements Deter Women from Joining Start-ups? Seong et al. (2023)
  • The Agentic-Communal Model of Advantage and Disadvantage: How Inequality Produces Similarities in the Psychology of Power, Social Class, Gender and Race, Rucker et al. (2018)
  • Agentic and Communal Values: Their Scope and Measurement, Trapnell & Paulhus (2012)
  • Evidence That Gendered Wording in Job Advertisements Exists and Sustains Gender Inequality, Gaucher et al. (2011)
  • The Duality of Human Existence, Bakan (1966)

While some studies are older, they remain the most cited and thorough research on the topic to date.

Culture shape biases and stereotypes

With a multilingual model, covering English, German, French, Danish, and Swedish, we also need a multicultural approach. That’s why we investigate biases and stereotypes within that language community each time we develop a bias detection for a new language.

2nd step: Mapping Findings

While reviewing these articles, we mapped out biases and stereotypes across five demographic groups: gender, age, ethnicity, neurodiversity, and physical disability.

We exclude any biases that wasn’t sufficiently backed up by research. For example, biases relating to socioeconomic class exist, but lacked enough evidence to be included.

We organized all identified biases and stereotypes into more than 50 overarching categories. These categories represent cultural concepts like “power,” “humility,” or “influence”. Emphasizing these concepts in your job ad has been proved to affect who applies (read more about this in Trapnell &Paulhus (2012)).

3rd step: Creating a Framework and Preparing Data

Using this mapping, we developed a framework and began preparing training data for the AI model. Our dataset includes over 1,000 job ads from around the world, written in English, Danish, Swedish, German, or French, covering various industries and sectors.

Biases are everywhere, so our framework and annotation guidelines are very strict and thoroughly documented to minimize individual interpretation. All trainers are trained linguists, and the annotated data undergoes multiple rounds of validation throughout the training process.

4th step: Training the Model

We fine-tune a multilingual BERT language model on the annotated data. This approach combines BERT’s advanced contextual understanding with our in-house expertise in bias detection, enabling highly accurate identification of biased language.

5th step: Validating the Model and Framework

In 2021, we collaborated with consultancy Epinion to rigorously test our framework. We replicated parts of Gaucher et al.’s 2011 landmark study on inclusive language in job ads and confirmed the same findings:

  • Ads with the most inclusive language attracted the most applicants.
  • Ads emphasizing communal values*, such as “humility”, “compassion” and “harmony” attracted fewer applicants.
  • Ads emphasizing agentic values*, such as “power”, “influence” and “autonomy” attracted the fewest applicants.

Read more about agency and communion here.

In 2025, we evaluated the model again to ensure these effects still hold. We compared number of applicants to jobs where the job ad had an inclusivity below and above 90. Our results showed:

Inclusive towardsIncrease in applicants
All genders

52%

All ehnicities49%
All ages47%

All neurodiversities

34%
All physical abilities31%

Beyond validating our model and framework, we continuously monitor new research and update the model accordingly.

6th step: Suggesting the Right Alternatives

Now that we have an AI model capable of detecting biases and stereotypes, the next question is: what should be written instead?

As mentioned, certain cultural concepts can deter applicants if overly emphasized. Our goal is to minimize or better contextualize these concepts—not simply swap out words with synonyms, which risks reintroducing biases.

Research on high performing job ad shows that being specific, detailed and describing the context of the job can increase the number of qualified candidates.

Example: What does “impact” mean?

To find the best replacement for a biased phrase we use the surrounding context to suggest phrases that share meaning with the original words but are more specific. For example:

If you love working with data, collaborating with cross-functional teams, and making a tangible impact – keep reading.

This phrase highlights the candidate’s influence, but it doesn’t clarify what that influence entails. Influence isn’t inherently negative, but how it’s described affects how candidates perceive the role and your expectations to the candidate.

The sentence that come just before the one above goes like this:

At XXX, we’re looking for a Data Analyst – Product to bring clarity to complexity, empowering our teams with the insights they need to build world-class products.

It provides us with the context we need to suggest what kind of impact the role will have.

Instead of suggesting synonymous alternatives like “be influential,” we recommend more precise phrases like “provide data-driven solutions”, or “support team decisions,” which directly reflect the role’s responsibilities shown in the context.

Get a high performing job ad

If you want to read more about high performing job ads, we recommend these literature reviews:

  • Varying the content of job advertisements the effects of message specificity, Bearden & Hardesty (2006)
  • Efficient recruitment with effective job advertisement: an exploratory literature review and research agenda, Mahjoub A. (2021)