The only AI Governance Platform deeply integrated into the global policy and standards ecosystem.

Credo AI is a trusted partner for global policymakers, regulators, and standard setters. Our team includes staff with prior experience at the European Commission, technology trade associations, and the U.S. Department of Commerce, as well as some of the largest enterprises in Europe (including our on-the-ground staff in Europe). Credo AI’s Policy team connects with key stakeholders in the U.S. Congress, as well as Mayors, Governors, and State-Level legislators.

Credo AI’s CEO and Founder, Navrina Singh, sits on the National Artificial Intelligence Advisory Committee (NAIAC), which advises President Biden and the National AI Initiative Office and is a young global leader with the World Economic Forum as well as an OECD AI Expert serving on the OECD’s Expert Group on AI Risk and Accountability. Navrina previously served as an executive board member of the Mozilla Foundation and Mozilla AI, supporting its trustworthy AI charter.

A Glimpse into our Global Impact and Ecosystem:

June 5, 2023

EU AI Act OpenLoop Sandbox

Credo AI was chosen as one of a select group of small and medium-sized enterprises (SMEs) to participate in a trailblazing initiative called Open Loop (a global policy experimentation program supported by Meta), designed to test various aspects of the EU AI Act in practice. Credo AI had the chance to apply design thinking to the policy, testing it "in practice" by asking ourselves difficult questions & sharing ideas across our product, policy, and data science teams at Credo AI. Credo AI provided preeminent thought leadership to a robust discussion on the implementation of the EU AI Act, including our experience working with industry, as well as our expertise in creating governance artifacts (transparency reporting, algorithmic impact assessments, algorithm design evaluations, model cards and more) for enterprises of all sizes, with a variety of AI use cases. Learn more here.

May 1, 2023

National AI Advisory Committee (NAIAC) Public Hearing at U.S. Department of Commerce

As a member of the National Artificial Intelligence Advisory Committee (NAIAC), Credo AI's CEO Navrina Singh spoke about the year-long work of the NAIAC at a public hearing hosted by the U.S. Department of Commerce, announcing the release of the Committee’s Year 1 Report. The NAIAC - launched in April 2022 - is tasked with advising the President and the National AI Initiative Office on topics related to the National AI Initiative.

April 18, 2023

UK Centre for Data Ethics and Innovation (CDEI) Algorithmic Impact Assessments Workshop

Hosted by CDEI and Ada Lovelace Institute, Credo AI was the only enterprise selected to present at the UK Government Centre for Data Ethics and Innovation workshop entitled “Exploring Tools for Trustworthy AI: Impact Assessments,” at White Hall in London.

In this workshop, Credo AI showcased our RAI Governance Platform, and research conducted into Algorithmic Impact Assessment prototypes for generative AI and human resources, to an audience of global regulators and enterprises, including the UK Information Commissioner’s Office, Ada Lovelace Institute, The Alan Turing Institute, British Standards Institute (BSI), DeepMind, Mastercard, Northrop Grumman, NHS AI Lab, and more.

This workshop provided an interactive opportunity for regulators and legislators, as well as standard-setting bodies and impacted enterprises, to exchange dialogue over best practices for algorithmic transparency reporting.

April 17, 2023

OECD AI Risk and Accountability Expert Working Group

Credo AI’s CEO Navrina Singh spoke at  OECD.AI AI Risk and Accountability Expert Working Group meeting at OECD Headquarters in Paris, France, as part of critical discussions on the impact of generative AI on AI policy worldwide, including discussions of the NIST Risk Management Framework, International Standards on AI, and the European Union AI Act. 

Through the OECD.AI Network of Experts workstream on AI risk, the OECD is engaging with partner organizations, including the International Organization for Standardization (ISO), Institute of Electrical and Electronics Engineers (IEEE), National Institute of Standards and Technology (NIST), European Committee for Electrotechnical Standardization (CEN-CENELEC), the European Commission (EC), Council of Europe (CoE), UNESCO, OECD, EU-US Trade and Technology Council (TTC) and Responsible AI Institute (RAII)-WEF to identify common guideposts to assess AI risk and impact for Trustworthy AI. The goal is to help implement effective and accountable trustworthy AI systems by promoting global consistency.

April 13, 2023

EU AI Act Insights

Credo AI’s Head of Business Development, Giovanni Leoni, and Global Policy Director Evi Fuelle engage with the European Parliament co-rapporteur of the EU AI Act, MEP Dragos Tudorache, & Chief of Staff Dan Nechita, to discuss how businesses can be best prepared for the European Union Artificial Intelligence Act and the future of general-purpose AI.

February 27, 2023

National Artificial Intelligence Initiative Panel hosted by SeedAI

Credo AI’s Global Policy Director  Evi Fuelle represented Credo AI's views and research on Responsible AI Governance on a panel alongside Elham Tabassi (Chief of Staff, Information Technology Laboratory, NIST), Janet Haven (Executive Director, Data & Society Research Institute), Nicol Turner Lee (Senior Fellow, Center for Technology Innovation, The Brookings Institution), and Christine Curtis (Partnership on AI) at a public engagement event hosted by SeedAI on the National AI Research Resource (NAIRR) Task Force's final report.

Policy Intelligence: Translating Policy and Standards to Code

Drawing from our experiences and discussions with global policymakers and standard setters, Credo AI has developed extensive and deep “Policy Intelligence.” Credo AI integrates this expertise and the most up-to-date insights into our Responsible AI Governance Platform. This work combines a profound technical grasp of AI risks with extensive policy and regulatory knowledge.

Our Policy Intelligence feeds into our Policy Packs, Credo AI’s technical requirements developed in collaboration with our research team to translate high-level concepts into checklists of actionable steps to ensure your AI systems are responsible, safe, and compliant.

Credo AI is trusted by the ones that build trust

Recognized as 2022 World Economic Forum Technology Pioneer
Received SOC 2 Type II-Certification
Named Key Responsible AI Governance Platform by IDC
CEO appointed to National AI Advisory Committee (NAIAC)
Named one of the Next Big Things In Tech by Fast Company
Analyst Coverage
Analyst Coverage
CB Insights’ annual list of the 100 most promising private AI companies in the world
Recognized by Madrona Ventures, Pitchbook and Goldman Sachs as a 2022 Intelligent Applications Top 40 winner for our work in Responsible AI. Learn more about the award and see a full list of recipients here.

It doesn’t end there

The knowledge we gather is shared through our expert content in our Resource Center. To learn more about topics such as the EU AI Act, NIST AI Risk Management Framework, and how to embark on your AI governance journey, please visit our Resource Center.

Adopt AI with confidence today

The Responsible AI Governance Platform enables AI, data, and business teams to track, prioritize, and control AI projects to ensure AI remains profitable, compliant, and safe.