NYC Bias Audit: What You Need To Know

December 1, 2022

Join Credo AI’s Navrina Singh and Susannah Shattuck, and AIethicist.org's Merve Hickok, as we dive into what this law means, who it applies to, and steps employers, agencies, and HR vendors can take to meet new bias audit requirements by April 15, 2023.

Merve Hickok
Research Director at Center for AI and Digital Policy & Founder of AIethicist.org
Navrina Singh
Founder & CEO
Susannah Shattuck
Head of Product

You may also like

AI Bias Audit: What You Need to Know About the Updated NYC Algorithmic Hiring Law

Starting April 15, 2023, New York City's Local Law No. 144 will require any automated employment decision tools (AEDT) used on NYC-based candidates or employees to undergo annual independent bias audits. With many details left undefined by the law, ensuring compliance can be challenging. To help, AIethicist.org's Merve Hickok, BABL AI's Shea Brown, and Credo AI's Susannah Shattuck and Ehrik Aldana discussed recent changes to the proposed law, updates from the January 23rd hearing, and steps employers, agencies, and HR vendors can take to meet new bias audit requirements. Watch it now! 🔥

NYC Bias Audit: Employer & HR Vendor Collaboration for Compliance

Join Credo AI’s Susannah Shattuck, Seekout’s Sam Shaddox, and Pymetrics’ Frida Polli, as we discuss the audit requirements for employers, how to engage your HR vendors and team during due diligence and the audit process, and ways Seekout and Pymetrics are helping their customers get in compliance.

AI Risk Management 101: A Practical Guide to Adopting the NIST AI RMF in 2023

On January 26th, the National Institute of Standards and Technology (NIST) announced the launch of the NIST AI Risk Management Framework (AI RMF) 1.0. This comprehensive set of guidelines and best practices is aimed at helping organizations proactively manage risks in designing, developing, and using AI products, services, and systems. Join Ms. Elham Tabassi, Chief of Staff in the Information Technology Laboratory (ITL) at NIST, Dr. Amanda C. Muller, Ph.D., Chief of Responsible Technology & Consulting Artificial Intelligence Systems Engineer at Northrop Grumman, and Susannah Shattuck and Evi Fuelle from Credo AI as they discuss what the NIST AI RMF 1.0 is designed to do, how your organization can benefit from the comprehensive framework, and tools to accelerate and demonstrate adoption.

Join the movement to make
Responsible Al a reality