Abstract
While rapid advances in artificial intelligence (AI) hiring tools promise to transform the workplace, these algorithms risk exacerbating existing biases against marginalized groups. In light of these ethical issues, AI vendors have sought to translate normative concepts such as fairness into measurable, mathematical criteria that can be optimized for. However, questions of disability and access often are omitted from these ongoing discussions about algorithmic bias. In this paper, I argue that the multiplicity of different kinds and intensities of people’s disabilities and the fluid, contextual ways in which they manifest point to the limits of algorithmic fairness initiatives. In particular, existing de-biasing measures tend to flatten variance within and among disabled people and abstract away information in ways that reinforce pathologization. While fair machine learning methods can help mitigate certain disparities, I argue that fairness alone is insufficient to secure accessible, inclusive AI. I then outline a disability justice approach, which provides a framework for centering disabled people’s experiences and attending to the structures and norms that underpin algorithmic bias.
Similar content being viewed by others
Data availability
Not applicable.
Materials availability
Not applicable.
Code availability
Not applicable.
Notes
I use “disabled person,” or identity-first language, instead of “person with disabilities,” or person-first language, as the latter can separate people from their disabilities and treat the term as negative (Ladau, 2015). While people should not be recognized solely on the basis of on their disabilities, disabilities are not merely tacked on to a person (Brown, 2015). Indeed, disability rights activists claimed disability as a source of community and pride in the 1990s (Linton, 2010), which continues in modern activist campaigns such as #SayTheWord (Andrews et al., 2019).
For example, the Social Security Administration understands disability with respect to the capacity for gainful employment; the Americans with Disabilities Act speaks of it in terms of essential functions, major life activities, and reasonable accommodations; and Worker’s Compensation defines it as percent of ability lost (Samuels, 2014). That one may be recognized as disabled under some but not all of these definitions does not just reflect bureaucratic errors but fundamental conflicts about how to classify the multiplicity of disabled people’s embodiments.
Disability is a fuzzy, porous category and whether one identifies with it hinges on a host of factors, such as the model at hand (Brown & Broido, 2020). Indeed, one of the central disputes in disability studies concerns how to define disability (Kafer, 2013; Samuels, 2014). This coalitional account is thus a minimal one that captures features of other models, i.e., it theorizes disability as at least in part non-medical and ableism as a structural power relation.
While data scientists can track the inputs and outputs of complex machine learning AI systems, they cannot do the same for its internal data processing operations, which occur in a black box (Castelvecchi, 2016). For example, an algorithm might detect predictors of job performance that vendors cannot recognize easily or explain.
Some endorse demographic parity, which equalizes selection rates in each group, regardless of outcomes (Lipton et al., 2018) For instance, an algorithm might rank job applicants within protected classes and select the top 5%. However, direct use of membership in these classes is disparate treatment and thus illegal (Romanov et al., 2019).
Indeed, similar tensions emerge when implementing calibration’s equal predictive values and the neutrality of anti-classification in legal contexts (Green & Viljoen, 2020). This suggests the incompatibility of these methods stems not from the translation of fairness into technical terms but deeper conflicts between their normative assumptions.
Even if data is anonymized, it can reveal disability status. For example, researchers found they could predict blindness from Twitter activity (Morris et al., 2016) and Parkinson’s through mouse tracking (White et al., 2018). They were also able to reidentify people by combining anonymized data with public records (Emam et al., 2011).
Indeed, the predictive validity of most hiring software has yet to be independently verified (Raghavan et al., 2020). Moreover, studies suggest that more subjective, qualitative measures of competence can be at least as accurate as ‘objective’ ones, as they better capture contextual aspects of workplace settings (Singh et al., 2016; Vij & Bedi, 2016).
Given the multitude of different disabilities, some have suggested expanding the number of disability categories that AI consider (Trewin, 2018). However, there may be too few people who have disabilities that are experienced in similar ways for AI to detect patterns, especially because many disabilities are comorbid and high-dimensionality data is difficult to capture even with large datasets (Givens & Morris, 2020).
References
Agrawal, R. (2019). AI is reducing barriers for people with disabilities to enter the Workforce: Hector Minto. Microsoft. https://news.microsoft.com/en-in/features/ai-is-reducing-barriers-for-people-with-disabilities-to-enter-workforce-hector-minto/
Altman, B. M., & Rasch, E. K. (2016). Purpose of an international comparable census disability measure. In B. M. Altman (Ed.), International measurement of disability: purpose, method, and application (pp. 55–68). Springer.
Ameri, M., Schur, L., Adya, M., Bentley, S., McKay, P., & Kruse, D. (2015). The disability employment puzzle: A field experiment on employer hiring behavior. ILR Review, 71(2), 329–364. https://doi.org/10.3386/w21560
Andrews, E., Forber-Pratt, A. J., Mona, L. R., Lund, E. M., Pilarski, C. R., & Balter, R. (2019). SaytheWord: A disability culture commentary on the erasure of ‘disability.’ Rehabilitation Psychology, 64(2), 111–118. https://doi.org/10.1037/rep0000258
Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine Bias. ProPublica. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
Barocas, S., Hardt, M., & Narayanan, A. (2018). Fairness and machine learning: Limitations and opportunities. https://fairmlbook.org/pdf/fairmlbook.pdf
Barocas, S., & Selbst, A. D. (2016). Big data’s disparate impact. California Law Review. https://doi.org/10.2139/ssrn.2477899
Baynton, D. (2001). Disability and the Justification of Inequality in American History. In P. K. Longmore & L. Umansky (Eds.), The New Disability History: American Perspectives. London: New York University Press.
Bennett, C. L., Rosner, D. K., & Taylor, A. S. (2020). The care work of access. In 2020 Proceedings of the 2020 CHI conference on human factors in computing systems. https://doi.org/10.1145/3313831.3376568
Birnbaum, E. (2019). Over 1,000 students across 17 colleges pledge not to work at Palantir over ICE work. The Hill. https://thehill.com/policy/technology/461573-over-1000-students-across-17-colleges-pledge-not-to-work-at-palantir-over
Bogart, K. R. (2014). The role of disability self-concept in adaptation to congenital or acquired disability. Rehabilitation Psychology, 59(1), 107–115. https://doi.org/10.1037/a0035800
Bogart, K. R., Rottenstein, A., Lund, E., & Bouchard, L. (2017). Who self-identifies as disabled? An examination of impairment and contextual predictors. Rehabilitation Psychology, 62(4), 553–562. https://doi.org/10.1037/rep0000132
Bogen, M., & Rieke, A. (2018). Help wanted: An examination of hiring algorithms, equity, and bias. Upturn. https://www.upturn.org/static/reports/2018/hiring-algorithms/files/Upturn%20--%20Help%20Wanted%20-%20An%20Exploration%20of%20Hiring%20Algorithms,%20Equity%20and%20Bias.pdf
Brown, L. (2015). Identity-first language. Autistic self advocacy network. https://autisticadvocacy.org/about-asan/identity-first-language/
Brown, K. R., & Broido, E. M. (2020). Ableism and assessment: Including students with disabilities. New Directions for Student Services, 169, 31–41. https://doi.org/10.1002/ss.20342
Buolamwini, J., & Gebru, T. (2018). Gender shades: Intersectional accuracy disparities in commercial gender classification. In Proceedings of the 1st conference on fairness, accountability and transparency (Vol. 81, 77–91).
Castelvecchi, D. (2016). Can we open the black box of AI? Nature News, 538(7263), 20–23. https://doi.org/10.1038/538020a
Cate, O. T., & Regehr, G. (2019). The power of subjectivity in the assessment of medical trainees. Academic Medicine, 94, 333–337. https://doi.org/10.1097/ACM.0000000000002495
Chamorro-Premuzic, T., Winsborough, D., Sherman, R. A., & Hogan, R. (2016). New talent signals: Shiny new objects or a brave new world? Industrial and Organizational Psychology, 9(3), 621–640. https://doi.org/10.1017/iop.2016.6
Chen, I. Y., Johansson, F. D., & Sontag, D. (2018). Why is my classifier discriminatory? In 32nd Conference on neural information processing systems. arXiv:1805.12002
Chouldechova, A. (2017). Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data, 5(2), 1–11. https://doi.org/10.1089/big.2016.0047
Chouldechova, A., Putnam-Hornstein, E., Benavides-Prado, D., Fialko, O., & Vaithianathan, R. (2018). A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. Proceedings of Machine Learning Research, 81, 1–15.
Chouldechova, A., & Roth, A. (2020). A Snapshot of the frontiers of fairness in machine learning. Communications of the ACM, 63(5), 82–89. https://doi.org/10.1145/3376898
Commission, E. E. O. (1978). Uniform guidelines on employee selection procedures. Federal Register, 43(166), 38290–38315.
Corbett-Davies, S., & Goel, S. (2018). The measure and mismeasure of fairness: a critical review of fair machine learning. arXiv preprint. arXiv:1808.00023
Costello, K. (2019). Gartner survey shows 37 percent of organizations have implemented AI in some form. Gartner. https://www.gartner.com/en/newsroom/press-releases/2019-01-21-gartner-survey-shows-37-percent-of-organizations-have
Crenshaw, K. (1991). Mapping the margins: Intersectionality, identity politics, and violence against women of color. Stanford Law Review, 43(6), 1241. https://doi.org/10.2307/1229039
d’Alessandro, B., O’Neil, C., & LaGatta, T. (2017). Conscientious classification: A data scientist’s guide to discrimination-aware classification. Big Data, 5(2), 1–15. https://doi.org/10.1089/big.2016.0048
Dastin, J. (2018). Amazon scraps secret AI recruiting tool that showed bias against women. Reuters. https://www.reuters.com/article/us-amazon-com-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08G
Dieterich, W., Mendoza, C., & Brennan, T. (2016). COMPAS risk scales: Demonstrating accuracy equity and predictive parity. https://go.volarisgroup.com/rs/430-MBX-989/images/ProPublica_Commentary_Final_070616.pdf
Emam, K. E., Jonker, E., Arbuckle, L., & Malin, B. (2011). A Systematic review of re-identification attacks on health data. PLoS ONE. https://doi.org/10.1371/journal.pone.0028071
Erevelles, N. (2018). Toward justice as ontology: Disability and the question of (in)difference. In E. Tuck & K. W. Yang (Eds.), Toward what justice? Describing diverse dreams of justice in education. Routledge.
Favaretto, M., Clercq, E. D., & Elger, B. S. (2019). Big data and discrimination: Perils, promises and solutions. A systematic review. Journal of Big Data. https://doi.org/10.1186/s40537-019-0177-4
Fok, R., Kaur, H., Palani, S., Mott, M. E., & Lasecki, W. S. (2018). Towards more robust speech interactions for deaf and hard of hearing users. In Proceedings of the 20th international ACM SIGACCESS conference on computers and accessibility. https://doi.org/10.1145/3234695.3236343
Friedman, B., & Hendry, D. G. (2019). Value sensitive design: Shaping technology with moral imagination. MIT.
Fruchterman, J., & Mellea, J. (2018). Expanding employment success for people with disabilities. Benetech. https://benetech.org/about/resources/expanding-employment-success-for-people-with-disabilities-2/
Ghoshal, A. (2018). This HR firm is using AI to hire without bias, negotiate salary. VC Circle. https://www.vccircle.com/this-hr-firm-is-using-ai-to-hire-without-bias-negotiate-salary/
Givens, A. R., & Morris, M. R. (2020). Centering disability perspectives in algorithmic fairness, accountability, & transparency. In Proceedings of the 2020 conference on fairness, accountability, and transparency. https://doi.org/10.1145/3351095.3375686
Goggin, G., Ellis, K., & Hawkins, W. (2019). Disability at the centre of digital inclusion: assessing a new moment in technology and rights. Communication Research and Practice, 5(3), 290–303. https://doi.org/10.1080/22041451.2019.1641061
Goodley, D. (2014). Dis/ability studies: Theorizing disableism and ableism. Routledge.
Govaerts, M., & van der Vleuten, C. P. (2013). Validity in work-based assessment: Expanding our horizons. Medical Education, 47(12), 1164–1174. https://doi.org/10.1111/medu.12289
Green, B. (2018). “Fair” risk assessments: A precarious approach for criminal justice reform. In Workshop on fairness, accountability, and transparency in machine learning.
Green, B., & Viljoen, S. (2020). Algorithmic realism: Expanding the boundaries of algorithmic thought. In Conference on fairness, accountability, and transparency. https://doi.org/10.1145/3351095.3372840
Greene, D., Hoffmann, A. L., & Stark, L. (2019). Better, nicer, clearer, fairer: A critical assessment of the movement for ethical artificial intelligence and machine learning. In Proceedings of the 52nd Hawaii international conference on system sciences.
Guo, A., Vaughan, J. W., Wallach, H., & Morris, M. R. (2019). Toward Fairness in AI for people with disabilities SBG@a research roadmap. SIGACCESS Accessibility and Computing. https://doi.org/10.1145/3386296.3386298
Hoffman, A. L. (2019a). Where fairness fails: Data, algorithms, and the limits of antidiscrimination discourse. Information, Communication & Society, 22(7), 900–915. https://doi.org/10.1080/1369118x.2019.1573912
Hoffmann, S. (2019b). What genetic testing teaches about long-term predictive health analytics regulation. North Carolina Law Review, 123, 1–38.
Houser, K. A. (2019). Can AI solve the diversity problem in the tech industry? Mitigating noise and bias in employment decision-making. Stanford Technology Law Review, 22, 90.
Houtenville, A., & Kalargyrou, V. (2015). Employers’ perspectives about employing people with disabilities. Cornell Hospitality Quarterly, 56(2), 168–179. https://doi.org/10.1177/1938965514551633
Hutchinson, B., Prabhakaran, V., Denton, E., Webster, K., Zhong, Y., & Denuyl, S. (2020). Social biases in NLP models as barriers for persons with disabilities. In Proceedings of the association for computational linguistics. arXiv:2005.00813
Hutchinson, B., Prabhakaran, V., Denton, E., Webster, K., Zhong, Y., & Denuyl, S. (2019). Unintended machine learning biases as social barriers for persons with disabilities. SIGACCESS Accessibility and Computing, 125, 1–10. https://doi.org/10.1145/3386296.3386305
Jacobs, A. Z., & Wallach, H. (2021). Measurement and fairness. In Conference on fairness, accountability, and transparency. https://doi.org/10.1145/3442188.3445901
Kafer, A. (2013). Feminist, queer. Indiana University Press.
Korsgaard, H., Nylandsted, C., & Bodker, S. (2016) Computational alternatives in participatory design—putting the T back in socio-technical research. In Proceedings of the 14th participatory design conference (Vol. 1, pp. 71–79). https://doi.org/10.1145/2940299.2940314
Krahn, G. L. (2011). WHO world report on disability: A review. Disability and Health Journal, 4(3), 141–142. https://doi.org/10.1016/j.dhjo.2011.05.001
Kruse, D., Schur, L., Rogers, S., & Ameri, M. (2018). Why do workers with disabilities earn less? Occupational job requirements and disability discrimination. British Journal of Industrial Relations, 56(4), 798–834. https://doi.org/10.1111/bjir.12257
Ladau, E. (2015). Why person-first language doesn't always put the person first. Think Inclusive. http://www.thinkinclusive.us/why-person-first-language-doesnt-always-put-the-person-first/
Larsen, L. (2019). HireVue assessments and preventing algorithmic bias. HireVue. http://www.hirevue.com/blog/hirevue-assessments-and-preventing-algorithmic-bias
Lillywhite, A., & Wolbring, G. (2019). Coverage of ethics within the artificial intelligence and machine learning academic literature: The case of disabled people. Assistive Technology. https://doi.org/10.1080/10400435.2019.1593259
Lindsay, S., Cagliostro, E., Albarico, M., Mortaji, N., & Karon, L. (2018). A systematic review of the benefits of hiring people with disabilities. Journal of Occupational Rehabilitation, 28(4), 634–655. https://doi.org/10.1007/s10926-018-9756-z
Linton, S. (2010). Claiming disability: Knowledge and identity. New York University Press.
Lipton, Z. C., Chouldechova, A., & McAuley, J. (2018). Does mitigating ML's impact disparity require treatment disparity? In 32nd Conference on neural information processing systems.
Maritz, R., Aronsky, D., & Prodinger, B. (2017). The International Classification of Functioning, Disability and Health (ICF) in electronic health records: A systematic review. Applied Clinical Informatics, 8(3), 964–980. https://doi.org/10.4338/ACI2017050078
Mingus, M. (2017). Moving toward the ugly: A politic beyond desirability. In L. K. Davis, J. Dolmage, N. Erevelles, S. P. Harris, A. Luft, S. Shweik, & L. Ware (Eds.), Beginning with disability: A primer. Routledge.
Morris, M. R., Zolyomi, A., Yao, C., Bahram, S., Bigham, J. P., & Kane, S. K. (2016). With most of it being pictures now, I rarely use it. In CHI conference on human factors in computing systems. https://doi.org/10.1145/2858036.2858116
Mott, M. E., & Wobbrock, J. O. (2019). Cluster touch: Improving touch accuracy on smartphones for people with motor and situational impairments. In CHI Conference on human factors in computing systems. https://doi.org/10.1145/3290605.3300257
Mulligan, D. K., Kroll, J. A., Kohli, N., & Wong, R. Y. (2019). This thing called fairness: Disciplinary confusion realizing a value in technology. Proceedings of the ACM on Human–computer Interaction. https://doi.org/10.1145/3359221
Nario-Redmond, M. R., Kemerling, A. A., & Silverman, A. (2019). Hostile, benevolent, and ambivalent ableism: Contemporary manifestations. Journal of Social Issues, 75(3), 726–756. https://doi.org/10.1111/josi.12337
Nario-Redmond, M. R., Noel, J. G., & Fern, E. (2013). Redefining disability, re-imagining the self: Disability identification predicts self-esteem and strategic responses to stigma. Self and Identity, 12(5), 468–488. https://doi.org/10.1080/15298868.2012.681118
Nielsen, K. E. (2012). A disability history of the United States. Beacon Press.
Ochigame, R. (2019b). The invention of ‘ethical AI:’ How big tech manipulates academia to avoid regulation. The Intercept. http://theintercept.com/2019b/12/20/mit-ethical-ai-artificial-intelligence/
Ochigame, R. (2019a). The illusion of algorithmic fairness. Zotero. https://www.zotero.org/groups/2311907/joi-public/items/CTBJQYME
Office of Civil Rights. (2021). HIPAA for Professionals. US Department of Health and Human Services. https://www.hhs.gov/hipaa/for-professionals/index.html
Okoro, C. A., Hollis, N. D., Cyrus, A. C., & Griffin-Blake, S. (2017). Prevalence of disabilities and health care access by disability status and type among adults—United States. Morbidity and Morality Weekly Report, 67(32), 882–887. https://doi.org/10.15585/mmwr.mm6732a3
Oliver, M. (2013). The social model of disability: Thirty years on. Disability & Society, 28(7), 1024–1026. https://doi.org/10.1080/09687599.2013.818773
Passi, S., & Barocas, S. (2019). Problem formulation and fairness. In Proceedings of the conference on fairness, accountability, and transparency. https://doi.org/10.1145/3287560.3287567
Peña, E., Stapleton, L., Brown, K. R., Broido, E., Stygles, K., & Rankin, S. (2018). A universal research design for student affairs scholars and practitioners. College Student Affairs Journal, 36(2), 1–14. https://doi.org/10.1353/csj.2018.0012
Prince, A. E. R., & Schwarcz, D. (2019). Proxy discrimination in the age of artificial intelligence and big data. Iowa Law Review, 105, 1257.
Raghavan, M., Barocas, S., Kleinberg, J., & Levy, K. (2020). Mitigating bias in algorithmic hiring: Evaluating claims and practices. In Proceedings of the conference on fairness, accountability, and transparency. https://doi.org/10.1145/3351095.3372828
Romanov, A., De-Arteaga, M., Wallach, H., Chayes, J., Borgs, C., Chouldechova, A., Geyik, S., Kenthapadi, K., Rumshisky, A., Kalai, A. T. (2019). What's in a name? Reducing bias in bios without access to protected attributes. In Proceedings of the North American Association for Computational Linguistics. arXiv:1904.05233
Rose, S. F. (2017). No right to be idle: The invention of disability, 1840s–1930s. University of North Carolina Press.
Rothstein, M. A. (2020). Predictive health information and employment discrimination under the ADA and GINA. The Journal of Law, Medicine, and Ethics, 48, 595–602. https://doi.org/10.1177/1073110520958887
Rotolo, C. T., Church, A. H., Adler, S., Smither, J. W., & Colquitt, A. L. (2018). Putting an end to bad talent management: A call to action for the field of industrial and organizational psychology. Industrial and Organizational Psychology, 11(2), 176–219. https://doi.org/10.1017/iop.2018.6
Samuels, E. (2014). Fantasies of identification: Disability, gender, race. New York University Press.
Sánchez-Monedero, J., Dencik, L., & Edwards, L. (2020). What does it mean to 'solve' the problem of discrimination in hiring? In Proceedings of the 2020 conference on fairness, accountability, and transparency. https://doi.org/10.1145/3351095.3372849
Santuzzi, A. M., & Waltz, P. R. (2016). Disability in the workplace: A unique and variable identity. Journal of Management, 42(5), 1111–1135. https://doi.org/10.1177/0149206315626269
Scholz, T. (2016). Platform cooperativism: Challenging the corporate sharing economy. http://eticasfoundation.org/wp-content/uploads/2019/03/Scholz_Platform-Cooperativism.pdf
Selbst, A. D., boyd, d., Friedler, S. A., Venkatasubramanian, S., & Vertesi, J. (2019). Fairness and abstraction in sociotechnical systems. In Proceedings of the conference on fairness, accountability, and transparency. https://doi.org/10.1145/3287560.3287598
Selbst, A. D., & Barocas, S. (2018). The intuitive appeal of explainable machines. Fordham Law Review. https://doi.org/10.2139/ssrn.3126971
Shakespeare, T. (2006). The social model of disability. In L. J. Davis (Ed.), The disability studies reader (2nd ed., pp. 197–204). Psychology Press.
Shields, J. (2018). Over 98% of Fortune 500 companies use applicant tracking systems (ATS). https://www.jobscan.co/blog/fortune-500-use-applicant-tracking-systems/
Siebers, T. (2008). Disability theory. University of Michigan Press.
Simon, J., Wong, P., & Rieder, G. (2020). Algorithmic bias and the Value Sensitive Design approach. Internet Policy Review, 9(4), 1–16. https://doi.org/10.14763/2020.4.1534
Singh, S., Darwish, T. K., & Potocnik, K. (2016). Measuring organizational performance: a case for subjective measures. British Journal of Management, 27, 214–244. https://doi.org/10.1111/1467-8551.12126
Suur, J. (2019). Preventing unconscious bias with Tribepad. TribePad. http://www.tribepad.com/new-features/preventing-unconscious-bias-with-tribepad/
Tilmes, N. (2020). The limits of fair machine learning: An analysis of structural ableism in algorithmic hiring. New York University.
Todd, S. (2019). People are terrible judges of talent. Can algorithms do better? Quartz. http://qz.com/work/1742847/pymetrics-ceo-frida-polli-on-the-ai-solution-to-hiring-bias/?utm_source=Solutions+Story+Tracker
Treviranus, J. (2018). Sidewalk Toronto and why smarter is not better. Medium. https://medium.datadriveninvestor.com/sidewalk-toronto-and-why-smarter-is-not-better-b233058d01c8
Trewin, S. (2018). AI fairness for people with disabilities: Point of view. arXiv preprint. arXiv:1811.10670
Trewin, S., Basson, S., Muller, M., Branham, S., Treviranus, J., Gruen, D., Hebert, D., Lyckowski, N., & Manser, E. (2019). Considerations for AI fairness for people with disabilities. AI Matters, 5(3), 1–24. https://doi.org/10.1145/3362077.3362086
van der Bijl-Brouwer, M., & Malcom, B. (2020). Systemic design principles in social innovation: A study of expert practices and design rationales. She Ji: THe Journal of Design, Economics, and Innovation, 6(2), 386–407. https://doi.org/10.1016/j.sheji.2020.06.001
Vij, S., & Bedi, H. S. (2016). Are subjective business performance measures justified? International Journal of Productivity and Performance Management, 65(5), 603–621. https://doi.org/10.1108/IJPPM-12-2014-0196
Wakabayashi, D., & Shane, S. (2018). Google will not renew pentagon contract that upset employees. https://www.nytimes.com/2018/06/01/technology/google-pentagon-project-maven.html
White, R. W., Doraiswamy, P. M., & Horvitz, E. (2018). Detecting neurodegenerative disorders from web search signals. NPJ Digital Medicine. https://doi.org/10.1038/s41746-018-0016-6
Whittaker, M., Alper, M., Bennett, C. L., Hendren, S., Kaziunas, L., Mills, M., M. R. Morris, J. Rankin, E. Rogers, M. Salas, West, S. M. (2019). Disability, bias, and AI. AI Now Institute. https://wecount.inclusivedesign.ca/uploads/Disability-bias-AI.pdf
Yoo, J. (2017). Pymetrics with Dr. Julie Yoo—YouTube. YouTube. http://www.youtube.com/watch?v=9fF1FDLyEmM
Zafar, M. B., Valera, I., Rodriguez, M. G., & Gummadi, K. P. (2017). Fairness beyond disparate treatment & disparate impact. In Proceedings of the 26th international conference on World Wide Web. https://doi.org/10.1145/3038912.3052660
Zuloaga, L. (2021). The latest leap in HireVue’s assessment technology. HireVue. https://www.hirevue.com/blog/hiring/the-latest-leap-in-hirevues-assessment-technology
Zyskowski, K., Morris, M. R., Bigham, J. P., Gray, M. L., & Kane, S. K. (2015). Accessible crowdwork? Understanding the value in challenge of microtask employment for people with disabilities. Computer-Supported Cooperative Work and Social Computing. https://doi.org/10.1145/2675133.2675158
Funding
No funding was received to assist with the preparation of this manuscript.
Author information
Authors and Affiliations
Contributions
Not applicable.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no conflict of interest to declare that are relevant to the content of this article.
Ethical approval
Not applicable.
Consent to publish
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Tilmes, N. Disability, fairness, and algorithmic bias in AI recruitment. Ethics Inf Technol 24, 21 (2022). https://doi.org/10.1007/s10676-022-09633-2
Accepted:
Published:
DOI: https://doi.org/10.1007/s10676-022-09633-2