Credobility:
The only AI Governance Platform deeply integrated into the global policy and standards ecosystem.
Credo AI is a trusted partner for global policymakers, regulators, and standard setters. Our team includes staff with prior experience at the European Commission, technology trade associations, and the U.S. Department of Commerce, as well as some of the largest enterprises in Europe (including our on-the-ground staff in Europe). Credo AI’s Policy team connects with key stakeholders in the U.S. Congress, as well as Mayors, Governors, and State-Level legislators.
Credo AI’s CEO and Founder, Navrina Singh, sits on the National Artificial Intelligence Advisory Committee (NAIAC), which advises President Biden and the National AI Initiative Office and is a young global leader with the World Economic Forum as well as an OECD AI Expert serving on the OECD’s Expert Group on AI Risk and Accountability. Navrina previously served as an executive board member of the Mozilla Foundation and Mozilla AI, supporting its trustworthy AI charter.

A Glimpse into our Global Impact and Ecosystem:

AI Commitments Bloomberg Interview
Founder and CEO of Credo AI, Navrina Singh, joined Caroline Hyde on Bloomberg TV to share her insights on the Voluntary “AI Commitments” from seven leading Large Language Model (LLM) providers and AI companies (Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI), announced by the Biden-Harris Administration on 21 July 2023. Learn more here, and watch here.

EU AI Act OpenLoop Sandbox
Credo AI was chosen as one of a select group of small and medium-sized enterprises (SMEs) to participate in a trailblazing initiative called Open Loop (a global policy experimentation program supported by Meta), designed to test various aspects of the EU AI Act in practice. Credo AI had the chance to apply design thinking to the policy, testing it "in practice" by asking ourselves difficult questions & sharing ideas across our product, policy, and data science teams at Credo AI. Credo AI provided preeminent thought leadership to a robust discussion on the implementation of the EU AI Act, including our experience working with industry, as well as our expertise in creating governance artifacts (transparency reporting, algorithmic impact assessments, algorithm design evaluations, model cards and more) for enterprises of all sizes, with a variety of AI use cases. Learn more here.

National AI Advisory Committee (NAIAC) Public Hearing at U.S. Department of Commerce
As a member of the National Artificial Intelligence Advisory Committee (NAIAC), Credo AI's CEO Navrina Singh spoke about the year-long work of the NAIAC at a public hearing hosted by the U.S. Department of Commerce, announcing the release of the Committee’s Year 1 Report. The NAIAC - launched in April 2022 - is tasked with advising the President and the National AI Initiative Office on topics related to the National AI Initiative.

UK Centre for Data Ethics and Innovation (CDEI) Algorithmic Impact Assessments Workshop
Hosted by CDEI and Ada Lovelace Institute, Credo AI was the only enterprise selected to present at the UK Government Centre for Data Ethics and Innovation workshop entitled “Exploring Tools for Trustworthy AI: Impact Assessments,” at White Hall in London.
In this workshop, Credo AI showcased our RAI Governance Platform, and research conducted into Algorithmic Impact Assessment prototypes for generative AI and human resources, to an audience of global regulators and enterprises, including the UK Information Commissioner’s Office, Ada Lovelace Institute, The Alan Turing Institute, British Standards Institute (BSI), DeepMind, Mastercard, Northrop Grumman, NHS AI Lab, and more.
This workshop provided an interactive opportunity for regulators and legislators, as well as standard-setting bodies and impacted enterprises, to exchange dialogue over best practices for algorithmic transparency reporting.

OECD AI Risk and Accountability Expert Working Group
Credo AI’s CEO Navrina Singh spoke at OECD.AI AI Risk and Accountability Expert Working Group meeting at OECD Headquarters in Paris, France, as part of critical discussions on the impact of generative AI on AI policy worldwide, including discussions of the NIST Risk Management Framework, International Standards on AI, and the European Union AI Act.
Through the OECD.AI Network of Experts workstream on AI risk, the OECD is engaging with partner organizations, including the International Organization for Standardization (ISO), Institute of Electrical and Electronics Engineers (IEEE), National Institute of Standards and Technology (NIST), European Committee for Electrotechnical Standardization (CEN-CENELEC), the European Commission (EC), Council of Europe (CoE), UNESCO, OECD, EU-US Trade and Technology Council (TTC) and Responsible AI Institute (RAII)-WEF to identify common guideposts to assess AI risk and impact for Trustworthy AI. The goal is to help implement effective and accountable trustworthy AI systems by promoting global consistency.

EU AI Act Insights
Credo AI’s Head of Business Development, Giovanni Leoni, and Global Policy Director Evi Fuelle engage with the European Parliament co-rapporteur of the EU AI Act, MEP Dragos Tudorache, & Chief of Staff Dan Nechita, to discuss how businesses can be best prepared for the European Union Artificial Intelligence Act and the future of general-purpose AI.

National Artificial Intelligence Initiative Panel hosted by SeedAI
Credo AI’s Global Policy Director Evi Fuelle represented Credo AI's views and research on Responsible AI Governance on a panel alongside Elham Tabassi (Chief of Staff, Information Technology Laboratory, NIST), Janet Haven (Executive Director, Data & Society Research Institute), Nicol Turner Lee (Senior Fellow, Center for Technology Innovation, The Brookings Institution), and Christine Curtis (Partnership on AI) at a public engagement event hosted by SeedAI on the National AI Research Resource (NAIRR) Task Force's final report.
Policy Intelligence: Translating Policy and Standards to Code
Drawing from our experiences and discussions with global policymakers and standard setters, Credo AI has developed extensive and deep “Policy Intelligence.” Credo AI integrates this expertise and the most up-to-date insights into our Responsible AI Governance Platform. This work combines a profound technical grasp of AI risks with extensive policy and regulatory knowledge.
Our Policy Intelligence feeds into our Policy Packs, Credo AI’s technical requirements developed in collaboration with our research team to translate high-level concepts into checklists of actionable steps to ensure your AI systems are responsible, safe, and compliant.
Credo AI is trusted by the ones that build trust











It doesn’t end there
The knowledge we gather is shared through our expert content in our Resource Center. To learn more about topics such as the EU AI Act, NIST AI Risk Management Framework, and how to embark on your AI governance journey, please visit our Resource Center.

Adopt AI with confidence today
The Responsible AI Governance Platform enables AI, data, and business teams to track, prioritize, and control AI projects to ensure AI remains profitable, compliant, and safe.