<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v3.0 20080202//EN" "http://dtd.nlm.nih.gov/publishing/3.0/journalpublishing3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="3.0" xml:lang="en" article-type="research article">
 <front>
  <journal-meta>
   <journal-id journal-id-type="publisher-id">
    jss
   </journal-id>
   <journal-title-group>
    <journal-title>
     Open Journal of Social Sciences
    </journal-title>
   </journal-title-group>
   <issn pub-type="epub">
    2327-5952
   </issn>
   <issn publication-format="print">
    2327-5960
   </issn>
   <publisher>
    <publisher-name>
     Scientific Research Publishing
    </publisher-name>
   </publisher>
  </journal-meta>
  <article-meta>
   <article-id pub-id-type="doi">
    10.4236/jss.2025.1310001
   </article-id>
   <article-id pub-id-type="publisher-id">
    jss-146198
   </article-id>
   <article-categories>
    <subj-group subj-group-type="heading">
     <subject>
      Articles
     </subject>
    </subj-group>
    <subj-group subj-group-type="Discipline-v2">
     <subject>
      Business 
     </subject>
     <subject>
       Economics, Social Sciences 
     </subject>
     <subject>
       Humanities
     </subject>
    </subj-group>
   </article-categories>
   <title-group>
    Detecting Bias in AI: A Multi-Label RoBERTa Classification Model to Detect Bias in LLM-Generated Diversity Reports 
   </title-group>
   <contrib-group>
    <contrib contrib-type="author" xlink:type="simple">
     <name name-style="western">
      <surname>
       Mathew Sunil
      </surname>
      <given-names>
       Abraham
      </given-names>
     </name>
    </contrib>
    <contrib contrib-type="author" xlink:type="simple">
     <name name-style="western">
      <surname>
       Nicole
      </surname>
      <given-names>
       Lee
      </given-names>
     </name>
    </contrib>
    <contrib contrib-type="author" xlink:type="simple">
     <name name-style="western">
      <surname>
       Rezza
      </surname>
      <given-names>
       Moieni
      </given-names>
     </name>
    </contrib>
   </contrib-group> 
   <aff id="affnull">
    <addr-line>
     aDiversity Atlas, Melbourne, Australia
    </addr-line> 
   </aff> 
   <pub-date pub-type="epub">
    <day>
     29
    </day> 
    <month>
     09
    </month>
    <year>
     2025
    </year>
   </pub-date> 
   <volume>
    13
   </volume> 
   <issue>
    10
   </issue>
   <fpage>
    1
   </fpage>
   <lpage>
    17
   </lpage>
   <history>
    <date date-type="received">
     <day>
      30,
     </day>
     <month>
      June
     </month>
     <year>
      2025
     </year>
    </date>
    <date date-type="published">
     <day>
      26,
     </day>
     <month>
      June
     </month>
     <year>
      2025
     </year> 
    </date> 
    <date date-type="accepted">
     <day>
      26,
     </day>
     <month>
      September
     </month>
     <year>
      2025
     </year> 
    </date>
   </history>
   <permissions>
    <copyright-statement>
     © Copyright 2014 by authors and Scientific Research Publishing Inc. 
    </copyright-statement>
    <copyright-year>
     2014
    </copyright-year>
    <license>
     <license-p>
      This work is licensed under the Creative Commons Attribution International License (CC BY). http://creativecommons.org/licenses/by/4.0/
     </license-p>
    </license>
   </permissions>
   <abstract>
    With increasing use of generative AI in creating textual summaries and dashboard reports, there is significant concern regarding diverse forms of bias such as gender, geographical, religious, cultural, and language biases. This research investigates the presence of bias in AI-generated diversity reports and presents a sentence-level bias detection model to quantify and classify different types of bias. The study focuses on five key bias categories: gender, religion, age, disability, and sexuality. We train and evaluate a multi-label bias classifier with a RoBERTa-based deep learning model, integrating manual confidence-weighted annotation practices to ensure reliable labelling. Synthetic diversity reports were generated using the Gemini 1.5-flash language model to simulate real-world corporate content. We use this model to analyse around 1000 reports (10,000+ sentences) for bias and assess the nature and distribution of different bias types. The model demonstrated high accuracy and recall rates, effectively detecting both overt and subtle biases across categories. Analysis of over 10,000 sentences revealed measurable bias in the generated reports, with disability, gender, and religion biases being the most frequently detected. These findings highlight that even when using inclusive prompts, large language models can produce biased content. The model’s strong performance and ability to detect nuanced and intersectional biases make it a practical tool for organisations aiming to audit AI-generated communications. We expect this to contribute to the growing field of ethical AI by supporting more transparent, fair, and inclusive corporate reporting. 
   </abstract>
   <kwd-group> 
    <kwd>
     Artificial Intelligence (AI)
    </kwd> 
    <kwd>
      Ethnical AI
    </kwd> 
    <kwd>
      Bias Detection
    </kwd> 
    <kwd>
      Large Language Models (LLMs)
    </kwd> 
    <kwd>
      Diversity Data Analysis 
    </kwd>
   </kwd-group>
  </article-meta>
 </front>
 <body>
  <sec id="s1">
   <title>1. Introduction</title>
   <p>The emergence of generative Artificial Intelligence (AI) tools has transformed organisational communication, particularly in creating diversity and inclusion reports. Large Language Models (LLMs) like GPT-4, GPT-4o, and Gemini are increasingly utilised to automate content for internal and external distribution. However, this reliance on AI introduces challenges related to fairness and representation, as these models may unintentionally perpetuate societal biases present in their training data <xref ref-type="bibr" rid="scirp.146198-3">
     (Bender et al., 2021).
    </xref> Such biases can appear across dimensions like gender, religion, age, disability, and sexuality, affecting how various groups are portrayed in organisational narratives. To demonstrate, consider this hypothetical sentence generated for a diversity report: “Our senior leadership team, headed by capable male executives, collaborates with supportive female team members to help our dynamic, youthful workforce navigate modern workplace challenges while addressing the needs of employees with disabilities.” This one sentence reveals several subtle biases: it links leadership with men and support roles with women (gender bias), suggests that only younger employees possess energy (age bias), and presents disability accommodations as “special” instead of being regarded as standard practice (disability bias). These types of seemingly neutral statements can subtly reinforce harmful stereotypes while appearing professional and inclusive on the surface.</p>
   <p>This is particularly problematic for diversity reports, which aim to foster inclusivity and fairness. If AI-generated content contains biases, it can contradict the intended goals of these reports. Addressing and reducing these biases is essential to uphold credibility, advance equity, and ensure alignment with organisational values.</p>
   <p>Previous studies conducted within the organisation have also addressed these concerns. <xref ref-type="bibr" rid="scirp.146198-17">
     Raichur et al. (2023)
    </xref> demonstrated the persistent gender biases in government communication despite policy-driven inclusivity efforts, while <xref ref-type="bibr" rid="scirp.146198-11">
     Mao et al. (2023)
    </xref> developed a large-scale language model specifically to detect and mitigate gender and age biases in Australian job advertisements. These foundational studies inform and support the direction of the current project.</p>
   <p>The main aim of this study is to detect and quantify bias in AI-generated reports. The objectives were to:</p>
   <p>In the context of this study, bias refers to systematic and unfair favouritism or marginalisation of particular groups within AI-generated content. The following categories of bias were specifically assessed:</p>
   <p>This project holds significant importance for several reasons. First, it introduces a structured framework for detecting bias at the sentence level within long-form AI-generated text. Second, it integrates human-centred annotation practices with modern deep learning techniques, making the resulting model both robust and grounded in real-world perceptions. Third, it provides actionable insights for organisations seeking to audit and enhance the inclusivity of their automated communication systems. Finally, by focusing on corporate diversity reporting, it addresses a gap in current NLP research, which has primarily concentrated on bias in social media and short-form text.</p>
  </sec><sec id="s2">
   <title>
    <xref ref-type="bibr" rid="scirp.146198-"></xref>2. Literature Review</title>
   <p>
    <xref ref-type="bibr" rid="scirp.146198-"></xref>Artificial Intelligence (AI), particularly Large Language Models (LLMs) such as GPT, has revolutionised how content is generated, summarised, and presented across industries. However, growing reliance on these models has also raised ethical concerns regarding the biases they inherit and propagate. These biases can influence representation, reinforce stereotypes, and affect societal equity (<xref ref-type="bibr" rid="scirp.146198-18">
     Shuford, 2024
    </xref>; <xref ref-type="bibr" rid="scirp.146198-12">
     Marinucci et al., 2023
    </xref>). Recent studies further reinforce this concern by demonstrating how gender and age biases are deeply embedded in both public communications and recruitment materials, even when organisations intend to promote inclusivity (<xref ref-type="bibr" rid="scirp.146198-17">
     Raichur et al., 2023
    </xref>; <xref ref-type="bibr" rid="scirp.146198-11">
     Mao et al., 2023
    </xref>). As institutions increasingly use AI to automate content, especially public-facing reports, identifying and mitigating biases in AI-generated text becomes essential to ensure fairness, inclusivity, and trustworthiness.</p>
   <p>This literature review explores academic and applied research related to bias detection in AI-generated language and discusses the classification of multi-dimensional biases (e.g., gender, religion or worldview, age, disability, sexuality). It draws on studies from Machine Learning (ML), Natural Language Processing (NLP), media theory, and ethical AI governance to build a foundation for bias quantification in institutional reporting.</p>
   <sec id="s2_1">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>2.1. Defining and Understanding Bias in Text</title>
    <p>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>Bias in language models is typically defined as a systematic and unfair skew in output toward or against certain groups, perspectives, or regions. <xref ref-type="bibr" rid="scirp.146198-13">
      Mehrabi et al. (2021)
     </xref> describes bias as multi-dimensional and context-dependent, often arising from biased training datasets or societal power structures. Gender bias, for instance, is observed when leadership roles are associated with masculine terms <xref ref-type="bibr" rid="scirp.146198-19">
      (Zhao et al., 2018),
     </xref> while religious bias may reinforce dominant worldviews or stereotypes <xref ref-type="bibr" rid="scirp.146198-2">
      (Abrar et al., 2025).
     </xref> Geographic bias is often observed when Global North regions (e.g., the United States, the United Kingdom, Australia) are overrepresented in training data or outputs <xref ref-type="bibr" rid="scirp.146198-10">
      (Li et al., 2024).
     </xref> These biases are not mutually exclusive and can intersect, compounding marginalisation. <xref ref-type="bibr" rid="scirp.146198-19">
      Zhao et al. (2018)
     </xref> demonstrated how word embeddings trained on common text sources reflect gender stereotypes, such as associating “man” with “programmer” and “woman” with “homemaker”. <xref ref-type="bibr" rid="scirp.146198-2">
      Abrar et al. (2025)
     </xref> explored religious narratives within NLP, identifying skewed representation of Abrahamic versus Eastern religions. <xref ref-type="bibr" rid="scirp.146198-10">
      Li et al. (2024)
     </xref> found that GPT-based summaries favoured data and references from Global North countries, limiting global contextual awareness. These studies underline the necessity for deliberate design and evaluation to ensure equitable outputs across cultural and demographic lines.</p>
    <p>The theoretical basis for understanding bias in AI-generated organisational reports can be traced to principles of strategic corporate communication. <xref ref-type="bibr" rid="scirp.146198-4">
      Cornelissen (2023)
     </xref> highlights that organisational communication extends beyond information delivery, focusing instead on shaping perceptions, building meaning, and guiding interpretation through framing techniques. In the case of AI-produced content, such as diversity reports or HR summaries, the manner of presentation can greatly affect stakeholder comprehension, trust, and involvement. Sentence-level framing, tone, and emphasis might inadvertently perpetuate existing social hierarchies or overlook marginalised viewpoints, emphasising the need for a thorough assessment of representational fairness in automated reporting.</p>
   </sec>
   <sec id="s2_2">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>2.2. Role of AI in DEI Initiatives</title>
    <p>As Diversity, Equity, and Inclusion (DEI) becomes a strategic priority across the tech sector, organisations are increasingly relying on AI tools to generate structured reports that highlight progress and gaps in workplace equity. AI-generated reports enable companies to streamline internal documentation, reduce human error, and scale communication efforts without expanding operational costs. These tools are especially useful for creating regular performance updates, recruitment analytics, inclusive policy summaries, and compliance documentation.</p>
    <p>According to <xref ref-type="bibr" rid="scirp.146198-9">
      Kondra
     </xref><xref ref-type="bibr" rid="scirp.146198-9">
      et al.
     </xref><xref ref-type="bibr" rid="scirp.146198-9">
      (2025),
     </xref> AI systems are being integrated across HR and DEI functions for tasks such as monitoring demographic representation, evaluating inclusive hiring practices, and automating the creation of diversity dashboards. This automation helps reduce bias in human decision-making but may also embed new forms of algorithmic bias, making transparency and interpretability in AI outputs essential.</p>
    <p>For example, large tech firms like Microsoft, IBM, and Accenture have reported using generative AI to assist in producing inclusion reports, employee experience summaries, and ethical compliance updates. Thus, this project’s focus on AI-generated DEI reports reflects a real-world trend: the growing dependence on language models for inclusive corporate communication and the corresponding need to ensure those outputs are equitable and unbiased.</p>
   </sec>
   <sec id="s2_3">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>2.3. Bias Detection in AI</title>
    <p>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>A variety of techniques have been employed to detect bias within AI models and their outputs. These approaches utilise diverse methodologies, from relying on lexicon-based methods to employing sophisticated embedding-based models for comprehensive detection. Lexicon-based methods use pre-compiled lists of terms to identify bias but lack contextual sensitivity (<xref ref-type="bibr" rid="scirp.146198-6">
      Fast et al., 2016
     </xref>), whereas embedding-based models like BERT and RoBERTa are more effective in detecting subtle bias through semantic analysis (<xref ref-type="bibr" rid="scirp.146198-8">
      Jentzsch &amp; Turan, 2022
     </xref>). This aligns with previous work where RoBERTa outperformed other models in bias detection tasks within Australian job advertisements, specifically for gender and age bias detection (<xref ref-type="bibr" rid="scirp.146198-11">
      Mao et al., 2023
     </xref>). Supervised models trained on datasets like StereoSet (<xref ref-type="bibr" rid="scirp.146198-14">
      Nadeem et al., 2020
     </xref>) enable sentence-level detection and classification of multiple bias types. Fairness evaluation metrics like Equality of Odds and Disparate Impact further assess model performance in bias-sensitive tasks (<xref ref-type="bibr" rid="scirp.146198-5">
      Dixon et al., 2018
     </xref>). Sentence-level analysis is crucial, as it provides granular insight into how specific lines within a larger report may carry bias, enabling better downstream filtering and correction.</p>
    <p>Building on these detection techniques, scoring and classification models then provide a structured way to quantify identified biases. Supervised ML classifiers such as logistic regression, support vector machine, or fine-tuned transformers can assign binary labels (biased/unbiased) or bias type codes (e.g., 1G = gender bias). Furthermore, probabilistic scoring allows assigning a confidence-based score between 0 and 1 to indicate the likelihood of bias, which can support nuanced mitigation strategies (<xref ref-type="bibr" rid="scirp.146198-15">
      Quan &amp; Pu, 2023
     </xref>; <xref ref-type="bibr" rid="scirp.146198-7">
      Hossin &amp; Sulaiman, 2015
     </xref>).</p>
   </sec>
   <sec id="s2_4">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>2.4. Broader Impacts and Policy Implications</title>
    <p>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>Bias in AI-generated reports can influence decision-making, amplify inequality, and reduce user trust. This is particularly concerning in education, hiring, and journalism. Governments and institutions are thus calling for transparency in AI development. For instance, Amazon’s recruitment AI system was scrapped after it was found to downgrade resumes containing the word “women’s”, reflecting gender bias in the training data. Another well-known example includes Twitter’s image-cropping algorithm, which was found to disproportionately favour lighter-skinned faces over darker ones in preview images, raising concerns about racial bias in visual AI systems. <xref ref-type="bibr" rid="scirp.146198-16">
      R’boul (2021)
     </xref> cautions that intercultural messaging, even when promoting inclusion, may reinforce existing hierarchies. Ethical AI frameworks recommend continuous auditing, inclusive dataset creation, and human-in-the-loop systems to reduce unintended consequences. Bias detection models like the one developed in this study contribute directly to mitigation efforts by identifying harmful patterns early, enabling organisations to retrain models, refine prompts, and prevent the reinforcement of systemic inequalities.</p>
   </sec>
   <sec id="s2_5">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>2.5. Gaps and Future Directions</title>
    <p>Despite significant progress, current models struggle with intersectional bias, where multiple identity-based biases compound within AI outputs affecting individuals who belong to multiple marginalised groups simultaneously (<xref ref-type="bibr" rid="scirp.146198-13">
      Mehrabi et al., 2021
     </xref>). Additionally, most bias benchmarks are based on Western language corpora <xref ref-type="bibr" rid="scirp.146198-10">
      (Li et al., 2024).
     </xref> Our project addresses these gaps by developing a multi-label bias detection model trained on a manually annotated, sentence-level dataset that covers five distinct bias categories: gender, religion, age, disability, and sexuality. By generating 990 AI-generated diversity reports using Gemini LLM, we introduced diversity in model behaviour and content sources. Furthermore, our dataset includes annotations from multiple human evaluators using a majority vote system to ensure reliability and reduce subjectivity. This approach directly confronts the issue of intersectional and cultural bias by applying consistent, multi-annotator labelling across diverse topics and prompts. In doing so, the project moves toward building a scalable, adaptable framework for detecting representational bias in real-world organisational content. Future work should include multilingual datasets, domain-specific audits, and real-time evaluation tools for AI-generated content in live systems. Integrating human annotations with automated predictions could further enhance accuracy and trustworthiness.</p>
   </sec>
  </sec><sec id="s3">
   <title>
    <xref ref-type="bibr" rid="scirp.146198-"></xref>3. Methodology</title>
   <p>This section outlines the methodology employed in this study, encompassing dataset preparation, manual annotation, and model development. We explain the generation and refinement of the synthetic diversity reports, the systematic process of human annotation for bias identification, and the subsequent training of our model. <xref ref-type="fig" rid="fig1">
     Figure 1
    </xref> offers a comprehensive overview of these stages.</p>
   <fig id="fig1" position="float">
    <label>Figure 1</label>
    <caption>
     <title>
      <xref ref-type="bibr" rid="scirp.146198-"></xref>Figure 1. Methodology workflow outlining the key stages from dataset generation to final bias prediction.</title>
    </caption>
    <graphic mimetype="image" position="float" xlink:type="simple" xlink:href="https://html.scirp.org/file/6500419-rId13.jpeg?20250929030937" />
   </fig>
   <sec id="s3_1">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>3.1. Dataset Preparation</title>
    <p>A total of 990 synthetic diversity reports were generated using prompts applied to three prominent large language models: GPT-4, GPT-4o, and Gemini 1.5-flash. Each report was based on one of nine standardised diversity-related prompts, covering topics such as multilingual workplace support, inclusive religious practices, accommodations for long-term conditions, generational diversity, data-driven diversity and inclusion strategies, planning, social media content, and internal communications to HR or executive leadership. Examples of these prompts include:</p>
    <p>These prompts were distributed across 110 major technology companies to ensure diverse industry representation.</p>
    <p>Following an evaluation of the outputs, reports generated by Gemini 1.5-flash were selected for further analysis. This decision was primarily driven by Gemini’s consistent ability to complete sentences and maintain logical coherence, an area where other models occasionally produced incomplete or abruptly ended sentences. Particularly, GPT-4 and GPT-4o often produced incomplete sentences, exhibited inconsistent formatting that made segmentation more difficult, and generated repetitive content that could distort bias detection. These models struggled to keep context clear in longer documents, leading to contradictory statements and extraneous annotation. The choice of model may have impacted the bias patterns present in the training data, as various large language models display different bias inclinations based on their respective training sets. In addition, Gemini’s outputs exhibited a more uniform structure and clearer language, contributing to more reliable annotation and enhanced performance in downstream model training.</p>
    <p>The final dataset consisted of over 10,000 sentences. Reports were segmented into individual sentences using Python-based natural language processing techniques, and relevant metadata such as report ID and prompt type were retained. To minimise cognitive bias during the annotation process, company names were anonymised prior to review.</p>
   </sec>
   <sec id="s3_2">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>3.2. Manual Annotation and Label Creation</title>
    <p>A total of 300 randomly selected reports, comprising approximately 4700 sentences, were independently annotated by four reviewers. Each sentence was assessed for the presence (labelled as 1) or absence (labelled as 0) of bias across five key dimensions: gender bias, religious bias, age bias, disability bias, and sexuality bias. Reviewers carefully analysed each sentence and categorised it into the appropriate bias classifications according to their evaluation of biased wording, stereotypes, or unjust representation.</p>
    <p>To evaluate the reliability of the annotations, inter-annotator agreement was determined for all bias categories prior to implementing the weighted majority voting method. This offered a quantitative assessment of the consistency among the four reviewers in recognising bias within the sentences they manually annotated.</p>
    <p>To enhance the reliability of the annotation process, a weighted majority voting system was employed.</p>
    <p>This approach facilitated the generation of both binary bias indicators and confidence-weighted labels, enabling more nuanced model training and evaluation.</p>
   </sec>
   <sec id="s3_3">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>3.3. Model Development</title>
    <p>The RoBERTa model was chosen for its superior performance in sentence-level classification and proven success in bias detection, attributed to its enhanced pretraining efficiency, robust contextual understanding, and stable training compared to BERT. Its ability to handle large datasets, process long sequences, and support flexible fine-tuning for multi-label tasks made it ideal for identifying nuanced biases across diverse text domains. Additionally, RoBERTa’s balance of high performance, computational efficiency, and strong community support in frameworks like Hugging Face ensures reliable and practical implementation.</p>
    <p>Training data was split into 80% for training and 20% for validation. Training was performed over 5 epochs using the AdamW optimiser with a learning rate of 2e−5 and batch size of 8. A linear learning rate scheduler was employed, and early stopping was considered to prevent overfitting. Sentences and labels were converted into PyTorch-compatible datasets with appropriate attention masks and input embeddings. Sentences were tokenised using the RoBERTa tokeniser, applying padding and truncation to a maximum length of 256 tokens. This ensured consistent input shapes and preserved key contextual information for robust sentence-level understanding.</p>
    <p>The final trained model file size was approximately 487 MB, indicating that the base model was efficiently fine-tuned without excessive parameter overhead. This balance between model capacity and storage size made it practical for sentence-level bias detection tasks.</p>
    <p>The initial analysis revealed a low prevalence of bias across most categories, with certain dimensions such as sexuality bias appearing in only approximately 1% of the annotated sentences. To address the resulting class imbalance and enhance model learning, several strategies were employed:</p>
    <p>A multi-label classification setup was used, as each sentence could simultaneously exhibit more than one type of bias. This approach allowed the model to independently assess all five bias categories, enhancing detection accuracy for complex and intersectional biases. The bias detection models produced probability scores for each bias label rather than binary outputs. Instead of applying a fixed threshold such as 0.5 across all labels, the study computed optimal thresholds for each category individually by maximising the F1-score, with a particular emphasis on improving recall. This approach enabled the model to better identify subtle forms of bias, reduce the likelihood of false negatives, and enhance the overall accuracy of sentence classification when applied in real-world scenarios.</p>
   </sec>
  </sec><sec id="s4">
   <title>
    <xref ref-type="bibr" rid="scirp.146198-"></xref>4. Results and Discussions</title>
   <p>This section presents the findings from the application of our RoBERTa-based model, showcasing the distribution of detected bias, model performance metrics, and the practical implications of these findings for organisations.</p>
   <sec id="s4_1">
    <title>4.1. Bias Detection Findings</title>
    <p>The RoBERTa-based multi-label classification model was applied to a final dataset comprising 10,862 sentences sourced from synthetic diversity reports generated using the Gemini 1.5-flash large language model. The model independently evaluated each sentence for the presence of five distinct types of bias: gender, religion, age, disability, and sexuality.</p>
    <p>The predicted distribution of bias revealed that gender bias was detected in 681 sentences (6.27%), religion bias in 675 sentences (6.21%), age bias in 549 sentences (5.05%), disability bias in 829 sentences (7.63%), and sexuality bias in 111 sentences (1.02%). <xref ref-type="fig" rid="fig2">
      Figure 2
     </xref> represents these total bias detections across the dataset. These findings indicate that, despite the use of inclusive prompts during the generation process, the LLM-generated reports contained measurable bias across multiple dimensions.</p>
    <fig id="fig2" position="float">
     <label>Figure 2</label>
     <caption>
      <title>
       <xref ref-type="bibr" rid="scirp.146198-"></xref>Figure 2. Number of biased sentences detected by bias type.</title>
     </caption>
     <graphic mimetype="image" position="float" xlink:type="simple" xlink:href="https://html.scirp.org/file/6500419-rId14.jpeg?20250929030938" />
    </fig>
    <p>The analysis revealed that disability bias was the most frequently detected bias, suggesting that even with diversity-focused prompts, the language model consistently generated content that subtly marginalised individuals with disabilities. For instance, the sentence “This plan addresses the needs of employees with long-term conditions, aiming to improve workplace wellbeing and productivity.” positions employees with disabilities as requiring special attention rather than being part of the standard organisational fabric, reinforcing the notion of separation rather than inclusion.</p>
    <p>Gender and religion biases were also notably prevalent, each appearing in over 6% of the total sentences, indicating persistent representational imbalances in these categories. Gender bias frequently manifested in the form of binary gender assumptions and demographic generalisations. For example, the sentence “Assume your workforce is 60% female, 40% male” portrays a binary framing that excludes gender-diverse individuals and oversimplifies gender representation.</p>
    <p>Religion bias often surfaced in organisational narratives that framed religious diversity as a potential management issue rather than an inclusive opportunity. The sentence “This plan aims to address potential issues arising from diverse religious and worldview beliefs within your organisation” subtly positions diversity as a challenge, which can unintentionally marginalise individuals with diverse religious backgrounds.</p>
    <p>Sexuality bias, though less frequently detected, was sometimes reflected in language that indirectly connected sexual orientation with health-related concerns. For instance, the sentence “The percentage of individuals identifying as non-heterosexual and experiencing long-term health conditions varies across generations, impacting individual needs and preferences.” subtly implies an association between non-heterosexual identities and health conditions, which risks perpetuating harmful stereotypes.</p>
    <p>Furthermore, <xref ref-type="fig" rid="fig3">
      Figure 3
     </xref> below illustrates ten companies with the highest total bias detections across the analysed synthetic diversity reports. Okta recorded the highest number of biased sentences (38), followed closely by ASML and GitHub (Microsoft) with 36 and 35 detections respectively. This distribution highlights that even among leading technology companies, measurable bias persists in AI-generated content.</p>
    <fig id="fig3" position="float">
     <label>Figure 3</label>
     <caption>
      <title>
       <xref ref-type="bibr" rid="scirp.146198-"></xref>Figure 3. Ten companies with the highest number of biased sentences detected.</title>
     </caption>
     <graphic mimetype="image" position="float" xlink:type="simple" xlink:href="https://html.scirp.org/file/6500419-rId15.jpeg?20250929030938" />
    </fig>
   </sec>
   <sec id="s4_2">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>4.2. Model Performance Metrics</title>
    <p>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>The model’s predictive performance was evaluated on a validation set comprising 20% of the manually annotated dataset. The following metrics—accuracy, precision, recall, and F1-score—were recorded and are presented in <xref ref-type="table" rid="table1">
      Table 1
     </xref> below. The overall model accuracy across all labels was 0.99.</p>
    <table-wrap id="table1">
     <label>
      <xref ref-type="table" rid="table1">
       Table 1
      </xref></label>
     <caption>
      <title>
       <xref ref-type="bibr" rid="scirp.146198-"></xref>Table 1. Values of accuracy, precision, recall, and F1-Score by bias type.</title>
     </caption>
     <table class="MsoTableGrid custom-table" border="0" cellspacing="0" cellpadding="0"> 
      <tr> 
       <td class="custom-bottom-td acenter" width="27.63%"><p style="text-align:center">Bias Type</p></td> 
       <td class="custom-bottom-td acenter" width="18.09%"><p style="text-align:center">Accuracy</p></td> 
       <td class="custom-bottom-td acenter" width="18.09%"><p style="text-align:center">Precision</p></td> 
       <td class="custom-bottom-td acenter" width="18.09%"><p style="text-align:center">Recall</p></td> 
       <td class="custom-bottom-td acenter" width="18.09%"><p style="text-align:center">F1-Score</p></td> 
      </tr> 
      <tr> 
       <td class="custom-top-td acenter" width="27.63%"><p style="text-align:center">Gender Bias</p></td> 
       <td class="custom-top-td acenter" width="18.09%"><p style="text-align:center">0.99</p></td> 
       <td class="custom-top-td acenter" width="18.09%"><p style="text-align:center">0.81</p></td> 
       <td class="custom-top-td acenter" width="18.09%"><p style="text-align:center">0.97</p></td> 
       <td class="custom-top-td acenter" width="18.09%"><p style="text-align:center">0.89</p></td> 
      </tr> 
      <tr> 
       <td class="acenter" width="27.63%"><p style="text-align:center">Religion Bias</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.98</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.90</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.82</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.86</p></td> 
      </tr> 
      <tr> 
       <td class="acenter" width="27.63%"><p style="text-align:center">Age Bias</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.99</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.93</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.89</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.91</p></td> 
      </tr> 
      <tr> 
       <td class="acenter" width="27.63%"><p style="text-align:center">Disability Bias</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.97</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.69</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.97</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.80</p></td> 
      </tr> 
      <tr> 
       <td class="acenter" width="27.63%"><p style="text-align:center">Sexuality Bias</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">1.00</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.86</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">1.00</p></td> 
       <td class="acenter" width="18.09%"><p style="text-align:center">0.92</p></td> 
      </tr> 
     </table>
    </table-wrap>
    <p>The model exhibited excellent sensitivity, particularly through consistently high recall scores across all bias types. It performed especially well in detecting gender, disability, and sexuality biases:</p>
   </sec>
   <sec id="s4_3">
    <title>4.3. Model Strengths and Considerations</title>
    <p>A key strength of this model lies in its use of a confidence-weighted annotation process. By applying varying weights based on the level of annotator agreement classified as low, medium, or high confidence, the model was trained to prioritise highly reliable examples while still learning from more ambiguous cases. This approach proved particularly valuable, as it improved the model’s ability to detect subtle bias indicators that may not have been consistently identified by all annotators. It also reduced the risk of overfitting to only the clearest examples, ensuring that the model developed sensitivity to a wide range of bias expressions, including less overt cases. The inclusion of confidence-weighted learning allowed the model to make more nuanced decisions during training, contributing to its excellent generalisation on the validation set. By learning from both highly agreed-upon and less certain examples, the model became particularly robust in handling the complex nature of bias detection.</p>
    <p>Throughout this study, the bias detection model demonstrated several significant strengths. One of its most notable advantages was its high sensitivity, with consistently strong recall scores across all bias categories. This ensured that the model successfully detected most biased sentences, making it particularly reliable for screening large volumes of AI-generated content where manual review would be impractical. The model also exhibited strong multi-label flexibility, correctly identifying intersectional bias where a single sentence could display multiple types of bias simultaneously. This capability is crucial for real-world applications, where complex layers of bias often coexist.</p>
    <p>Furthermore, the model effectively managed class imbalance within the dataset. The use of focal loss and strategies for minority class augmentation, especially for sexuality bias, allowed the model to learn better from underrepresented categories and improve its sensitivity to less frequently detected biases. Another key strength was the optimisation of decision thresholds for each bias type. Instead of applying a generic cut-off, the model was fine-tuned to select thresholds that maximised the F1-scores for each specific category, thereby enhancing its overall accuracy and practical performance.</p>
   </sec>
   <sec id="s4_4">
    <title>
     <xref ref-type="bibr" rid="scirp.146198-"></xref>4.4. Practical Implications for Organisations</title>
    <p>The predicted results highlight that even when using carefully designed, diversity-focused prompts, large language models continue to produce measurable bias across multiple dimensions. The presence of such bias in AI-generated content raises significant concerns for organisations, especially when the most frequently detected biases like disability, gender, and religion closely align with areas that are central to corporate diversity, equity, and inclusion (DEI) reporting. Failing to detect and address these biases can seriously undermine an organisation’s DEI commitments and may lead to the dissemination of content that is inconsistent with the organisation’s core values. Such inconsistencies can damage the credibility of AI-generated reports and erode trust among both internal stakeholders and the broader public. If biased content is publicly released, it also poses reputational risks that can be difficult to recover from.</p>
    <p>These findings emphasise the importance of integrating automated bias detection tools, such as the model developed in this study, into organisational workflows. Proactively auditing AI-generated content can help organisations identify and mitigate representational biases early in the content creation process, ensuring that their communications remain fair, inclusive, and aligned with their stated diversity goals. A feasible approach would include creating a review workflow that incorporates human oversight, where sentences identified by the bias detection system are sent directly to DEI specialists for assessment and modification. Organisations could establish bias threshold levels to focus their review on the most concerning content while still ensuring production efficiency.</p>
   </sec>
  </sec><sec id="s5">
   <title>
    <xref ref-type="bibr" rid="scirp.146198-"></xref>5. Limitations and Future Works</title>
   <p>While the model performed strongly in many areas, it is important to acknowledge its limitations. The very high accuracy and recall, particularly in the sexuality bias category, may indicate a degree of overfitting to the augmented data used during training. This could potentially limit the model’s ability to maintain the same level of performance when exposed to entirely new or more complex datasets. In addition, the model’s validation was conducted using a holdout set from the same synthetic dataset on which it was trained. Although the internal validation results were promising, further testing on real-world, externally sourced reports is essential to confirm the model’s generalisability and practical robustness. Another limitation stems from the exclusive use of synthetic reports generated by Gemini 1.5-flash. Broader testing on reports generated by other large language models, as well as actual organisational documents, would provide a more comprehensive assessment of the model’s reliability and adaptability across different content sources. Finally, despite efforts to enhance annotation quality through majority voting and confidence weighting, the inherent subjectivity involved in manual bias annotation remains a factor that may have influenced the final labels. Human interpretation of bias is inherently variable, and these subjective differences could have affected both the training data and the model’s subsequent learning outcomes.</p>
  </sec><sec id="s6">
   <title>
    <xref ref-type="bibr" rid="scirp.146198-"></xref>6. Conclusion</title>
   <p>This study developed a multi-label classification model based on RoBERTa to detect five key types of bias: gender, religion, age, disability, and sexuality, within AI-generated diversity reports. The model was applied to over 10,000 sentences sourced from synthetic reports generated using the Gemini 1.5-flash language model. The results revealed measurable bias across all categories, with disability, gender, and religion biases being the most frequently detected. These findings highlight that even when using carefully constructed, diversity-oriented prompts, large language models continue to produce content that reflects subtle but persistent representational biases.</p>
   <p>A key strength of this research lies in its confidence-weighted annotation process, which allowed the model to learn from both highly reliable and more ambiguous examples. By incorporating varying levels of annotator agreement, the model developed a nuanced sensitivity to both overt and subtle expressions of bias. Its strong predictive performance, particularly with consistently high recall across all bias types, makes it highly suitable for practical organisational use. The model can reliably screen large volumes of AI-generated content to proactively identify and flag potentially biased sentences before publication.</p>
   <p>The study also has important implications for organisations. It demonstrates that automated systems designed to promote diversity and inclusion are not inherently immune to bias and require careful auditing. Integrating automated bias detection tools, like the one developed in this project, is essential to safeguard organisational credibility, support inclusive communication, and mitigate reputational risks.</p>
   <p>However, several limitations should be acknowledged. The reliance on synthetic reports, the use of augmented data for minority classes, and the inherent subjectivity in manual annotation may have influenced the model’s generalisability. Future research should focus on testing the model with real-world organisational documents and reports generated by different language models, as well as expanding datasets to include more diverse cultural and linguistic contexts. Overall, this research presents a scalable and adaptable framework for bias detection and contributes meaningfully to the growing field of ethical AI.</p>
  </sec><sec id="s7">
   <title>Acknowledgements</title>
   <p>This project’s success relied on the guidance, collaboration, and support of numerous individuals and organisations. Heartfelt thanks go to Cultural Infusion and Diversity Atlas for offering the platform and opportunity to investigate the convergence of AI and diversity through this research.</p>
   <p>Appreciation is expressed to the wider research and development community at Cultural Infusion and Diversity Atlas and Cultural Infusion, whose discussions, diverse perspectives, and commitment to inclusivity profoundly shaped the project’s vision and objectives, although they may not agree with all the interpretations or conclusions of the results. The authors would like to thank Peter Mousaferiadis, Mary Legrand, MyLinh Le, Aida Hakimi, Quincy Hall, Catherine Mccredie and Osvaldo Branquinho for their support in this research.</p>
  </sec><sec id="s8">
   <title>Appendix</title>
   <p>The complete set of code for this project is securely stored in a private GitHub repository. Access has been granted to the project collaborators. The repository can be found at the following link:</p>
   <p>
    <xref ref-type="bibr" rid="scirp.146198-https://github.com/CulturalInfusion/Detecting-Bias-in-AI-A-Case-Study-of-LLM-Generated-Diversity-Reports-for-Tech-Companies-.git">
     https://github.com/CulturalInfusion/Detecting-Bias-in-AI-A-Case-Study-of-LLM-Generated-Diversity-Reports-for-Tech-Companies-.git
    </xref></p>
  </sec>
 </body><back>
  <ref-list>
   <title>References</title>
   <ref id="scirp.146198-ref1">
    <label>1</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Abrar, A., Oeshy, N. T., Kabir, M.,&amp;Ananiadou, S. (2025). Religious Bias Landscape in Language and Text-to-Image Models: Analysis, Detection, and Debiasing Strategies. arXiv: 2501.08441.
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref2">
    <label>2</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Bender, E. M., Gebru, T., McMillan-Major, A.,&amp;Shmitchell, S. (2021). On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (pp. 610-623). ACM. &gt;https://doi.org/10.1145/3442188.3445922
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref3">
    <label>3</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Cornelissen, J. P. (2023). Corporate Communication: A Guide to Theory and Practice. SAGE Publications Ltd.
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref4">
    <label>4</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Dixon, L., Li, J., Sorensen, J., Thain, N.,&amp;Vasserman, L. (2018). Measuring and Mitigating Unintended Bias in Text Classification. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society (pp. 67-73). ACM. &gt;https://doi.org/10.1145/3278721.3278729
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref5">
    <label>5</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Fast, E., Chen, B.,&amp;Bernstein, M. S. (2016). Empath: Understanding Topic Signals in Large-Scale Text. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (pp. 4647-4657). ACM. &gt;https://doi.org/10.1145/2858036.2858535
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref6">
    <label>6</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Hossin, M.,&amp;Sulaiman, M. N. (2015). A Review on Evaluation Metrics for Data Classification Evaluations. International Journal of Data Mining &amp; Knowledge Management Process, 5, 1-11. &gt;https://doi.org/10.5121/ijdkp.2015.5201
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref7">
    <label>7</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Jentzsch, S.,&amp;Turan, C. (2022). Gender Bias in BERT—Measuring and Analysing Biases through Sentiment Rating in a Realistic Downstream Classification Task. In Proceedings of the 4th Workshop on Gender Bias in Natural Language Processing (GeBNLP) (pp. 184-199). Association for Computational Linguistics. &gt;https://doi.org/10.18653/v1/2022.gebnlp-1.20
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref8">
    <label>8</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Kondra, S., Medapati, S., Koripalli, M., Chandra Nandula, S. R. S.,&amp;Zink, J. (2025). AI and Diversity, Equity, and Inclusion (DEI): Examining the Potential for AI to Mitigate Bias and Promote Inclusive Communication. Journal of Artificial intelligence and Machine Learning, 3, 1-10. &gt;https://doi.org/10.55124/jaim.v3i1.249
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref9">
    <label>9</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Li, B., Haider, S.,&amp;Callison-Burch, C. (2024). This Land Is {Your, My} Land: Evaluating Geopolitical Biases in Language Models. arXiv: 2305.14610. 
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref10">
    <label>10</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Mao, R., Tan, L.,&amp;Moieni, R. (2023). Developing a Large-Scale Language Model to Unveil and Alleviate Gender and Age Biases in Australian Job Ads. In 2023 IEEE International Conference on Big Data (BigData) (pp. 4176-4185). IEEE. &gt;https://doi.org/10.1109/bigdata59044.2023.10386083
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref11">
    <label>11</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Marinucci, L., Mazzuca, C.,&amp;Gangemi, A. (2023). Exposing Implicit Biases and Stereotypes in Human and Artificial Intelligence: State of the Art and Challenges with a Focus on Gender. AI &amp; SOCIETY, 38, 747-761. &gt;https://doi.org/10.1007/s00146-022-01474-3
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref12">
    <label>12</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K.,&amp;Galstyan, A. (2021). A Survey on Bias and Fairness in Machine Learning. ACM Computing Surveys, 54, 1-35. &gt;https://doi.org/10.1145/3457607
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref13">
    <label>13</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Nadeem, M., Bethke, A.,&amp;Reddy, S. (2020). StereoSet: Measuring Stereotypical Bias in Pretrained Language Models. arXiv: 2004.09456. 
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref14">
    <label>14</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Quan, Z.,&amp;Pu, L. (2023). An Improved Accurate Classification Method for Online Education Resources Based on Support Vector Machine (SVM): Algorithm and Experiment. Education and Information Technologies, 28, 8097-8111. &gt;https://doi.org/10.1007/s10639-022-11514-6
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref15">
    <label>15</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     R’boul, H. (2021). North/South Imbalances in Intercultural Communication Education. Language and Intercultural Communication, 21, 144-157. &gt;https://doi.org/10.1080/14708477.2020.1866593
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref16">
    <label>16</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Raichur, A., Lee, N.,&amp;Moieni, R. (2023). A Natural Language Processing Approach to Promote Gender Equality: Analysing the Progress of Gender-Inclusive Language on the Victorian Government Website. Open Journal of Social Sciences, 11, 513-529. &gt;https://doi.org/10.4236/jss.2023.119033
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref17">
    <label>17</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Shuford, J. (2024). Examining Ethical Aspects of AI: Addressing Bias and Equity in the Discipline. Journal of Artificial Intelligence General Science (JAIGS), 3, 262-280. &gt;https://doi.org/10.60087/jaigs.v3i1.119
    </mixed-citation>
   </ref>
   <ref id="scirp.146198-ref18">
    <label>18</label>
    <mixed-citation publication-type="other" xlink:type="simple">
     Zhao, J., Wang, T., Yatskar, M., Ordonez, V.,&amp;Chang, K. W. (2018). Gender Bias in Co-Reference Resolution: Evaluation and Debiasing Methods. arXiv: 1804.06876.
    </mixed-citation>
   </ref>
  </ref-list>
 </back>
</article>