FTC Outlines Approach to Discrimination in AI and Foreshadows Potential Enforcement

A recent blog post from the Federal Trade Commission (FTC) provides direct guidance on the FTC’s approach to Artificial Intelligence (AI) bias and signals potential enforcement in this area. The post—entitled Aiming for truth, fairness, and equity in your company’s use of AI—makes clear that the FTC will use its FTC Act authority to pursue sale and use of biased algorithms, and lays out a roadmap for its compliance expectations. The post is a step forward on one of Acting FTC Chairwoman Slaughter’s key policy priorities of promoting racial equity through consumer protection, as she has spoken often of potential harms from discriminatory algorithms. 

In the post, the FTC lays out the legal framework for evaluating AI bias, which includes the Fair Credit Reporting Act (FCRA), the Equal Credit Opportunity Act (ECOA), and Section 5 of the FTC Act. Both the FCRA and ECOA are applicable when AI is used for certain purposes, such as credit. The FTC Act has a broader sweep, and the explanation of Section 5 is particularly notable: “The FTC Act prohibits unfair or deceptive practices. That would include the sale or use of – for example – racially biased algorithms.” That would cover use of AI in a wide range of contexts, including advertising decisions.

Indeed, the FTC suggests that enforcement could be on the horizon if use of AI results in discrimination, noting that companies should “keep in mind that if you don’t hold yourself accountable, the FTC may do it.” As one example, the FTC states that an algorithm practice could be “unfair” under the FTC Act if the model “pinpoints [] consumers by considering race, color, religion, and sex – and the result is digital redlining.”

The rest of the post lays out a roadmap for compliance expectations for companies using AI and related algorithms. Among other things, companies are expected to:

  • Rely on inclusive data sets: “[f]rom the start, [companies should] think about ways to improve [their] data set, design [their] model to account for data gaps, and – in light of any shortcomings – limit where or how [they] use the model.”

  • Test their algorithms “both before [companies] use it and periodically after that—to make sure that it doesn’t discriminate on the basis of race, gender, or other protected class.”

  • Be truthful with customers about how their data is used and not exaggerate what an algorithm can deliver.

  • Be transparent and independent, “for example, by using transparency frameworks and independent standards.”

The FTC’s statements are a starting point for companies to prevent AI bias in practice, and companies that develop and use AI should be aware of the many resources in this space. For example, in the area of transparency, the National Institute of Standards and Technology (NIST) is currently developing standards for explainable AI that would be helpful for companies to consult – though that work remains ongoing. Similarly, NIST is in the process of considering ways to measure and test for bias in AI. Laws like ECOA have detailed implementing regulations and examination guidance, which may provide some guideposts—while much of the guidance is credit-specific, it could be used analogously in different contexts. In short, while the FTC has begun to outline the principles for companies to follow, companies will need to be creative and forward-thinking as they evaluate and address potential AI bias risks.

Tags

Wiley Connect

Sign up for updates

Wiley Rein LLP Cookie Preference Center

Your Privacy

When you visit our website, we use cookies on your browser to collect information. The information collected might relate to you, your preferences, or your device, and is mostly used to make the site work as you expect it to and to provide a more personalized web experience. For more information about how we use Cookies, please see our Privacy Policy.

Strictly Necessary Cookies

Always Active

Necessary cookies enable core functionality such as security, network management, and accessibility. These cookies may only be disabled by changing your browser settings, but this may affect how the website functions.

Functional Cookies

Always Active

Some functions of the site require remembering user choices, for example your cookie preference, or keyword search highlighting. These do not store any personal information.

Form Submissions

Always Active

When submitting your data, for example on a contact form or event registration, a cookie might be used to monitor the state of your submission across pages.

Performance Cookies

Performance cookies help us improve our website by collecting and reporting information on its usage. We access and process information from these cookies at an aggregate level.

Powered by Firmseek