AI Hiring Tools Under Attack: ACLU Files Claims with Feds Over Common Hiring Tools
Insights
6.11.24
The ACLU recently fired a clear warning shot to employers by asking the FTC to investigate a personality assessment test, a video interview tool, and a cognitive ability assessment screening device – all powered by artificial intelligence – because of alleged discrimination. The May 30 complaint follows the ACLU filing a charge with the EEOC over the same AI tools, alleging they unfairly screen out applicants with disabilities and unfairly target those with certain racial backgrounds. What do employers and AI developers need to know about these aggressive maneuvers and how should employers adjust when it comes to hiring? We’ll discuss this issue and more at AI IMPACT – an FP Conference for Business Leaders this June 26-28 in Washington, D.C. Learn more and register here.
3 Hiring Tools Targeted by ACLU
The three tools under attack were developed by Aon Consulting, a major player in the assessment field.
|
|
The ACLU claims that these tools assess general personality traits such as positivity, emotional awareness, liveliness, ambition, and drive that are not job-related or necessary for a specific job. It claims they can unfairly screen out people based on disabilities, as those with depression, anxiety, and autism (to name a few) can be scored low without regard to their actual job skills.
The ACLU specifically says that the use of AI to fuel these tools does nothing but exacerbate these fundamental problems. In fact, it alleges that the video interviewing tool is likely to discriminate based on disability, race, and other protected characteristics because of the AI data relied upon by the programs.
|
The ACLU cites data showing that the average gridChallenge scores for assessment-takers of different races reveals troubling disparities. It claims that a study revealed Asian, Black, Hispanic or Latino, and other test-takers scored lower than white test-takers on average.
It also claims that gridChallenge can discriminate against those with cognitive impairments and various mental health disabilities. For example, it contends that autistic people score significantly lower on various measures of working memory compared to individuals in the general population.
Why Did the ACLU File With the FTC?
The ACLU’s complaint with the FTC centers around the developer’s claim that these hiring tools are “fair,” “bias free,” and “without adverse impact” – which the ACLU alleges amounts to a deceptive marketing tactic. It claims there is a representation, omission, or practice that is likely to mislead consumers – in this case, employers looking for an efficient and effective hiring experience – to their detriment.
It asks the agency to launch an investigation into the developer to determine whether it is engaging in unfair and deceptive acts and practices. It also seeks an order from the agency to force Aon to stop selling or administering these tools until these problems are fixed.
Employers Aren’t Necessarily Off the Hook
While the ACLU’s complaint with the FTC doesn’t name any employers who might have relied upon these tools, the ACLU is not necessarily letting them off the hook quite yet. While it claims that the developer may have misled employers into a sense of comfort through its allegedly deceptive statements about the tools, it also says that “employers are legally obligated to ensure that the assessments they use do not discriminate, including when using technologically complex assessments.” Plaintiffs’ lawyers will certainly watch the FTC investigation closely.
EEOC Process Also Underway
In fact, the ACLU has already initiated an employment-related attack against Aon and one of the employers using these tools. It claims that a biracial autistic job applicant who was required to take the ADEPT-15 and gridChallenge assessments as part of the employer’s hiring process faced racial and disability discrimination as a result. It alleges violations of the Americans with Disabilities Act (ADA) and Title VII of the Civil Rights Act of 1964 on behalf of not only that applicant but all others similarly situated.
What’s Next?
The EEOC charge was filed late last year and is still pending with the agency. That administrative process could take several more months before any action is revealed, but it is likely that the charge was simply a precursor to the ACLU filing federal litigation against Aon and the employer on these same claims.
The FTC complaint is very recent, having just been filed in the past few weeks. The agency will soon determine what steps to take as a result of the allegations presented in the complaint and the developer’s response, though the administrative process could take months to unfold.
What Should Employers Do?
If you currently engage an AI developer or seek to use an AI-fueled tool to improve your hiring process, you should consider the following steps in light of these recent actions:
- Establish an AI Governance System –Develop clear systems and procedures – including a human oversight component – before deploying AI in the workplace. This is consistent with the guidelines issued by the National Institute for Standards and Technology (NIST) on artificial intelligence (NIST AI RMF). You should also evaluate the processes once in action to make sure they effectively govern the AI systems. Your Fisher Phillips AI team can assist with creating an AI Governance program specifically tailored for your organization. This program is based on the guidelines in NIST AI RMF.
- Vet Your Vendors – As the ACLU noted in its complaint, employers cannot simply play dumb and point fingers at AI vendors should problems arise. The EEOC said as much when it issued guidance about AI bias last year. Plaintiffs will increasingly seek to hold employers accountable for any resulting discrimination, so make sure you ask the right questions (with the help of your lawyers) when you are vetting products and at the time of implementation.
- Be Transparent – Employers should be transparent with workers and candidates about AI systems being used in the workplace. This is an increasingly common theme, and one that may soon be required in states like Colorado, California, and New York.
- Offer Accommodations if Feasible – Some states could soon require employers to allow workers and applicants to opt out of any AI process that makes determinations about their workplace future. You might want to offer the same. Some options could include specialized equipment, alternative tests or exam formats, or the opportunity to work in a quieter setting.
- Align Your Questioning – To the extent you develop a set of questions to be fed into the AI system and posed to applicants, make sure they are closely aligned with the job requirements for the position and don’t range too far afield.
Do You Want to Learn More?
If you are interested in learning more, sign up for AI IMPACT – an FP Conference for Business Leaders this June 26-28 in Washington, D.C. Learn more and register here.
Conclusion
We will continue to monitor these developments and provide the most up-to-date information directly to your inbox, so make sure you are subscribed to Fisher Phillips’ Insight System. If you have questions, contact your Fisher Phillips attorney, the authors of this Insight, or any attorney in our AI, Data, and Analytics Practice Group or Government Relations Team.
Related People
-
- John M. Polson
- Chairman & Managing Partner