ORCAA’s dual mission is to help define accountability for algorithms, and to keep people safe from harmful consequences of AI and automated systems. Whether it’s a hiring algorithm, healthcare AI, predictive scoring system, or generative AI platform, we are here to think about how it could fail, for whom, and what you can do to monitor and mitigate these risks. We are actively developing frameworks and governance approaches to help companies and organizations use algorithms and AI safely – confirming these technologies perform as intended and operate within sensible guardrails. We help our clients realize the transformative benefits of AI, while avoiding discrimination, bias, and other problems.
Services
Algorithmic Audit
A comprehensive assessment of risks associated with a specific use case of an algorithmic system. We audit systems of all kinds, including generative AI, automated decision systems, predictive models, and facial recognition.
Uses our Ethical Matrix framework
Identifies high-priority issues around fairness and performance
Delivers recommendations for measuring and mitigating risks
Outputs / Deliverables: Algorithmic Audit Report
Pilot: Quantitative testing for regulatory compliance and more
Analysis to measure bias in algorithmic systems, using data from live deployments or test data.
Proprietary, patent-pending cloud based analysis platform
Incorporates inference methodologies, so we can measure gender and race/ethnicity bias even if you do not possess this data
Double-firewall privacy protection: we never see personally-identifying information; client never sees individual inferences
Outputs / Deliverables: Bias Audit Report (e.g. for NYC Local Law 144); custom analysis reports
AI Governance + Risk Management Consultation
We help build core infrastructure to help you use AI responsibly.
Develop an organization-level approach to measuring and managing the risks of deployed AI systems, especially with regard to bias/discrimination
Review and recommend governance policies, processes, and structures
Assistance with procurement of AI technologies from vendors and other 3rd parties
Outputs / Deliverables: Documentation of risk management structures; diligence reports on prospective vendors
Cockpit Design
Running a deployed AI system is like piloting a plane: to fly safely, you need critical information in real time. To make an effective cockpit, you must understand what can go wrong in flight, and include dials/gauges that monitor for those risks. We help build cockpits for your AI systems.
Identify and/or construct metrics that address key risks
Calibrate thresholds and develop mitigation tactics
Can work with existing analytics platforms or assist with your custom implementation
Education + Training
We teach people how to think critically about AI and algorithmic systems, how to identify blind spots, and about the craft of auditing.
Workshops
Bespoke training for audit and risk management teams
Industries + Clients
Our partners range from silicon valley to the DMV, including private and public companies of all sizes and public agencies, in the US and internationally. We have a special interest in algorithms being used for hiring, insurance, credit, education, and healthcare – because we think algorithms in these sectors have major impacts on people’s lives, and need to be monitored closely.
Past and current clients include:
Explainable Fairness
When is an algorithm fair?
We propose the Explainable Fairness framework: a process to define and measure when an algorithm is complying with an anti-discrimination law. The central question is, Does the algorithm treat certain groups of people differently than others? The framework has three steps: choose a protected class, choose an outcome of interest, and measure and explain differences.
Example from hiring algorithms
Step 1: Identify protected stakeholder groups. Fair hiring rules prohibit discrimination by employers according to gender, race, national origin, and disability, among other protected classes. So all these could be considered specific groups for whom fairness needs to be verified.
Step 2: Identify outcomes of interest. In hiring, being offered a job is an obvious topline outcome. Other outcomes could also be considered employment decisions: for instance, whether a candidate gets screened out at the resume stage, or whether they are invited to interview, or who applied in the first place, which might reflect bias in recruitment.
Step 3: Measure and Explain Loop. Measure the outcomes of interest for different categories of the protected class. For example, are fewer women getting interviews? If so, is there a legitimate factor that explains that difference? For example, are men who apply more likely to have a relevant credential or more years of experience? If so, account for those legitimate factors and remeasure the outcomes. If you end up with unexplained large differences, you have a problem.
The process can be applied more generally, and looks like this:
Bias Audits for NYC Local Law 144
The new NYC law Int 1894-2020 (“Local Law 144”) requires annual bias audits of all automatic hiring decision tools used to hire candidates in or from NYC.
We offer a Bias Audit service to help companies comply with this law. We use our Pilot analysis platform to conduct disparate-impact-style and other analyses on real-world data arising from a specific use of a given decision tool. To show what we mean, here is a mock Bias Audit report for a fictitious company NewCo, which is using fictitious ToolX in its hiring process.
We can conduct a Bias Audit whether you are a vendor building hiring tools to be used by others, or a company using an AI tool in your own hiring process. If you already have race/ethnicity or gender information about candidates, we can use it in the audit; if you do not, we offer inference methods to model this information.
Please contact us to learn more about getting a Bias Audit.
HTI-1 Final Rule Reporting
How We Can Help
ORCAA and DIHI are proud to offer HTI-1 compliance reporting research and support services. ONC’s December 2023 HTI-1 Final Rule requires health IT developers to conduct more diligence than ever before about their Predictive Decision Support Interventions (Predictive DSIs) -- and provide more detailed reporting. We offer end-to-end support to navigate these requirements, including:
Developing use-case-specific metrics and processes for ensuring fairness and validating performance of Predictive DSIs
Curating data for and conducting external research on the validity and fairness of Predictive DSIs, and
Preparing and reviewing compliance reports.
Whether you are starting from scratch to meet HTI-1 requirements or have your compliance reports drafted and simply want an independent expert review, we can help. Please see below our research and service offering details to meet each section of the requirements:
-
image01
About the Partnership
This partnership represents a unique combination of expertise. ORCAA is a leading algorithmic auditing consultancy, focused on developing and applying standards for algorithmic systems. Our experience -- with diverse clients including private firms, regulatory agencies, and AGs; and across industries with different regulatory regimes -- gives us a broad perspective on how to demonstrate that algorithmic systems are safe and fair. ORCAA is an inaugural member of the US AI Safety Institute Consortium. Duke Institute for Health Innovation (DIHI) brings over ten years of real-world experience translating ideas into sustainable health innovations, including the sourcing, design, development, and implementation of more than 20 AI-based solutions into clinical care. DIHI is also the coordinating center for Health AI Partnership, a multi-stakeholder collaborative to advance responsible and equitable use of AI in healthcare.
Principles
Context Matters
An algorithm isn’t good or bad per se – it is just a tool. Who is using it, and to what end? What are the consequences? We go beyond traditional accuracy metrics to ask how the algorithm is used to make decisions, and who is affected.
Putting the science into data science
Graphs and statistical tests can be useful – if they answer a meaningful question. We translate specific ideas and concerns about fairness into statements we can test empirically. Then we design experiments to say whether those standards are being met.
AI Ethics cannot be automated
There cannot be a universal algorithmic auditing checklist or a fully automated review process – context is too important. An audit for a specific purpose may be repeatable, but human discussion and judgment will always be essential. We will tailor a tool to your needs, not repurpose what worked in another setting.
Contact