As the EEOC is well-aware, employers now have a wide variety of algorithmic decision-making tools available to assist them in making employment decisions, including recruitment, hiring, retention, promotion, transfer, performance monitoring, demotion, dismissal, and referral. Employers increasingly utilize these tools in an attempt to save time and effort, increase objectivity, optimize employee performance, or decrease bias. The EEOC’s Resource seeks to inform employers how to monitor the newer algorithmic decision-making tools and ensure compliance with Title VII.

To set the parameters for the Resource, the EEOC first defines a few key terms:

Taken together, employers sometimes utilize different types of software that incorporate algorithmic decision-making at a number of stages of the employment process. Some of the examples provided by the EEOC in terms of how employers can utilize artificial intelligence include: resume scanners that prioritize applications using certain keywords; employee monitoring software that rates employees on the basis of their keystrokes; “virtual assistants” or “chatbots” that ask job candidates about their qualifications and reject candidates who do not meet pre-defined requirements; video interviewing software that evaluates candidates based on their speech patterns and facial expressions; and testing software that provides “job fit” scores for applicants or employees regarding their personalities, aptitudes, cognitive skills, or perceived “cultural fit,” which is typically based on their performance on a game or on a more traditional test.

“Questions And Answers” About AI

After summarizing the pertinent provisions of Title VII, the heart of the EEOC’s Resource is presented in a question and answer format. First, the EEOC defines a “selection procedure” to be any “measure, combination of measures, or procedure” if it is used as a basis for an employment decision. Employers can assess whether a selection procedure has an adverse impact on a particular protected group by checking whether use of the procedure causes a selection rate for individuals in the group that is “substantially” less than the selection rate for individuals in another group. If there is an adverse impact, then use of the tool will run afoul of Title VII unless the employer can demonstrate that, pursuant to Title VII, such use is “job related and consistent with business necessity.”

The EEOC then posits the critical question of whether an employer is responsible under Title VII for its use of algorithmic decision-making tools even if the tools are designed or administered by another entity, such as a software vendor. This is an important issue since many companies seek the assistance of third-party technologies to facilitate some of their employment-decision processes. The EEOC indicates that “in many cases, yes,” employers are responsible for the actions of their agents, such as third-party vendors. Ultimately, if the employer is making the final employment decision, the buck would likely stop with the employer in terms of Title VII liability .

The EEOC also defines the term, “selection rate,” which refers to the proportion of applicants or candidates who are hired, promoted, or otherwise selected. The selection rate for a group of applicants or candidates is calculated by dividing the number of persons hired, promoted, or otherwise selected from the group by the total number of candidates in that group. By virtue of including this definition in the Resource, a reading of the tea leaves suggests that the EEOC will be monitoring selection rates to determine whether there is an adverse impact in employment decisions that were catalyze from the use of artificial intelligence.

In terms of what is an acceptable selection rate, the EEOC relies on the “four-fifths rule,” which is a general rule of thumb for determining whether the selection rate for one group is “substantially” different than the selection rate of another group. The rule states that one rate is substantially different than another if their ratio is less than four-fifths (or 80%). For example, if the selection rate for Black applicants was 30% and the selection rate for White applicants was 60%, the ratio of the two rates is thus 30/60 (or 50%). Because 30/60 (or 50%) is lower than 4/5 (or 80%), the four-fifths rule dictates that the selection rate for Black applicants is substantially different than the selection rate for White applicants, which may be evidence of discrimination against Black applicants.

The EEOC does note that the, “four-fifths rule” is a general suggestion, and may not be appropriate in every circumstance. Some courts have also found this rule to be inapplicable. Nonetheless, employers would be prudent to ask whether artificial intelligence vendors deployed the “four-fifths rule” in their algorithms. Statistics matter here.

Finally, the EEOC posits the issue of what an employers should do when they discover that the use of an algorithmic decision-making tool would result in an adverse impact. The EEOC explains that one advantage of algorithmic decision-making tools is that the process of developing the tool may itself produce a variety of comparably effective alternative algorithms. Accordingly, employers’ failure to adopt a less discriminatory algorithm that may have been considered during the development process could give rise to liability. Employers should thus take heed to document the steps they take to utilize non-discriminatory algorithms.

Implications For Employers

The use of artificial intelligence in employment decisions may be the new frontier for future EEOC investigations. While these technologies can have tremendous cost-benefits, the risk is undeniable. Inevitably, some employer using AI will be the subject of a test case in the future.

Employers should monitor the results of their own use of artificial intelligence. This can be accomplished by conducting self-analyses on an ongoing basis, to determine whether employment practices are disproportionately having a negative impact on certain protected classes.

As the EEOC notes, employers can proactively change the practices going forward. Given the agility of the artificial intelligence software, employers who do find the technologies’ “employment decisions” to be problematic can and should work with vendors to remedy such defects.

Source: Duane Morris LLP – Alex W. Karasik and Gerald L. Maatman, Jr.

Leave a Reply

%d bloggers like this: