This paper is published in Volume-11, Issue-5, 2025
Area
Machine Learning
Author
Sanvi Choukhani
Org/Univ
La Martiniere for Girls, West Bengal, India
Keywords
Gender Bias, Job Descriptions, Fair Recruitment, Natural Language Processing (NLP), Machine Learning, Topic Modelling, Algorithmic Fairness.
Citations
IEEE
Sanvi Choukhani. Analysing Gender Bias in Job Descriptions Using Machine Learning and NLP Techniques, International Journal of Advance Research, Ideas and Innovations in Technology, www.IJARIIT.com.
APA
Sanvi Choukhani (2025). Analysing Gender Bias in Job Descriptions Using Machine Learning and NLP Techniques. International Journal of Advance Research, Ideas and Innovations in Technology, 11(5) www.IJARIIT.com.
MLA
Sanvi Choukhani. "Analysing Gender Bias in Job Descriptions Using Machine Learning and NLP Techniques." International Journal of Advance Research, Ideas and Innovations in Technology 11.5 (2025). www.IJARIIT.com.
Sanvi Choukhani. Analysing Gender Bias in Job Descriptions Using Machine Learning and NLP Techniques, International Journal of Advance Research, Ideas and Innovations in Technology, www.IJARIIT.com.
APA
Sanvi Choukhani (2025). Analysing Gender Bias in Job Descriptions Using Machine Learning and NLP Techniques. International Journal of Advance Research, Ideas and Innovations in Technology, 11(5) www.IJARIIT.com.
MLA
Sanvi Choukhani. "Analysing Gender Bias in Job Descriptions Using Machine Learning and NLP Techniques." International Journal of Advance Research, Ideas and Innovations in Technology 11.5 (2025). www.IJARIIT.com.
Abstract
The growing use of automated recruitment systems has raised concerns about gender bias in job descriptions. Subtle linguistic cues can discourage qualified candidates from underrepresented groups, reinforcing workplace inequality. This study presents a computational framework using Natural Language Processing (NLP) and Machine Learning (ML) to detect and analyse such bias. The methodology involves text preprocessing, gender-coded word scoring, topic modelling with Latent Dirichlet Allocation (LDA), clustering via KMeans, and visualisation through t-SNE. A curated lexicon of masculine- and feminine-coded words assigns bias scores, while topic modelling uncovers latent themes in postings. Clustering groups of semantically similar descriptions enables analysis of bias distributions across occupational categories. Findings show that bias varies by job type: technical and managerial roles tend to use more masculine-coded language, while service and support roles favour feminine-coded terms. Semantic cluster visualisations confirm systemic patterns in word usage. This research underscores the need for fairness-aware audits in recruitment, offering both theoretical and practical insights into bias detection. The framework provides organisations with a scalable tool to identify and mitigate hidden biases, promoting inclusive hiring practices and supporting compliance with ethical and regulatory standards.
