Update Applicable to:
All employers utilizing artificial intelligence (AI) tools to interview and screen job applicants
On May 12, 2022, the EEOC issued long-awaited guidance on the use of such AI tools (the Guidance), examining how employers can seek to prevent AI-related disability discrimination.
Download Our Free Benefits Guide
Download our Benefits Brochure to see how we can provide Fortune 500-level benefits at a fraction of the cost.Download Guide
What are the details?
The Guidance identifies a number of ways in which employment-related use of AI can, even unintentionally, violate the Americans with Disabilities Act (ADA), including if employers are more frequently relying on the use of AI tools to automate employment decision-making, such as software that can review resumes and “chatbots” that interview and screen job applicants.
- “[t]he employer does not provide a ‘reasonable accommodation’ that is necessary for a job applicant or employee to be rated fairly and accurately by” the AI;
- “[t]he employer relies on an algorithmic decision-making tool that intentionally or unintentionally ‘screens out’ an individual with a disability, even though that individual is able to do the job with a reasonable accommodation”; or
- “[t]he employer adopts an [AI] tool for use with its job applicants or employees that violates the ADA’s restrictions on disability-related inquiries and medical examinations.”
The Guidance further states that “in many cases” employers are liable under the ADA for use of AI even if the tools are designed and administered by a separate vendor, noting that “employers may be held responsible for the actions of their agents … if the employer has given them authority to act on its behalf.”
The Guidance also identifies various best practices for employers, including:
- Announcing generally that employees and applicants subject to an AI tool may request reasonable accommodations and be provided instructions as to how to ask for accommodations.
- Providing information about the AI tool, how it works, and what it is used for to the employees and applicants subjected to it. For example, an employer that uses keystroke-monitoring software may choose to disclose this software as part of new employees’ onboarding and explain that it is intended to measure employee productivity.
- If the software was developed by a third party, ask the vendor whether:
- the AI software was developed to accommodate people with disabilities, and if so, how;
- there are alternative formats available for disabled individuals; and
- the AI software asks questions likely to elicit medical or disability-related information.
- If an employer is developing its own software, engage experts to analyze the algorithm for potential biases at different steps of the development process, such as a psychologist if the tool is intended to test cognitive traits.
- Only using AI tools that directly measure traits that are actually necessary for performing the job’s duties.
- Additionally, it is always a best practice to train staff, especially supervisors and managers, how to recognize requests for reasonable accommodations and to respond promptly and effectively to those requests. If the AI tool is used by a third party on the employer’s behalf, that third party’s staff should also be trained to recognize requests for reasonable accommodation and forward them promptly to the employer.
Schedule a Call
Learn more about VensureHR and how we can make an impact on your business.Contact VensureHR
For more information, please see the links below:
Article 1 – Article 2 – Article 3
What do employers need to do?
Employers should review the links provided above and the guidance and carefully audit them to ensure that this technology is not creating discriminatory outcomes.