
What Is an AI Bias Audit and Does Your Business Need One?
Two-Sentence Summary
A bias audit is a statistical test that measures whether an AI hiring tool produces different selection rates for different demographic groups — and New York City requires one every year for any employer using automated tools to screen job candidates. This article explains how the audit works, what 'selection rate' and 'impact ratio' actually mean, who has to do it, what the penalties are for skipping it, and how federal discrimination law connects to the same underlying analysis.
If you've recently heard the phrase "bias audit" and your instinct was to assume it was either very technical or very optional, you're in good company. Most businesses encounter this term when they're already behind — either because a vendor mentioned it, a lawyer flagged it, or they found out New York City has been fining employers for not conducting one since July 2023.
The good news is that a bias audit is more understandable than it sounds. It's a structured test of whether an AI hiring tool produces different outcomes for different demographic groups. That's essentially the whole concept. The law that requires it, New York City's Local Law 144, is specific about what the test has to measure, who has to do it, and what you have to do with the results.
This article explains all of it — what a bias audit is, what the law requires, who it applies to, what the penalties are, and where this is heading beyond New York City. Every factual claim about the law is drawn from the enacted statute and implementing regulations. Source links are at the bottom.
What is an AI bias audit under NYC Local Law 144?
A bias audit under NYC Local Law 144 is an impartial evaluation of an automated employment decision tool (AEDT) that measures whether the tool produces different selection rates for different demographic groups — a concept called disparate impact. The key metric is the impact ratio: a group's selection rate divided by the rate of the most-selected group. A ratio below 0.80 (the federal "four-fifths rule" from the EEOC Uniform Guidelines) signals potential adverse impact requiring attention.
A bias audit, in the context of New York City's Local Law 144, is an impartial evaluation of an automated employment decision tool (AEDT) to assess the tool's performance across racial, ethnic, and sex categories. The technical term for what the audit measures is "disparate impact" — whether the tool produces meaningfully different selection rates for different demographic groups.
Think of it this way. Your AI hiring tool looks at applications and decides who gets through to the next stage. Some percentage of applicants in any given demographic group will pass that screen. The audit asks: is that percentage significantly lower for some groups than for others? And if so, by how much?
The standard measurement is something called an impact ratio. The impact ratio compares the selection rate for a given demographic group to the selection rate for the most selected group. If your tool selects 60% of white applicants but only 40% of Black applicants, the impact ratio for Black applicants is 0.67. A ratio below 0.80 — the "four-fifths rule" — has historically been treated under EEOC guidance as evidence of adverse impact. (EEOC Uniform Guidelines on Employee Selection Procedures)
New York City's rules require the audit to calculate selection rates and impact ratios for each race/ethnicity category and each sex category as defined in the EEO-1 form categories — the same demographic categories employers already use in federal EEOC reporting. (6 RCNY § 5-301)
Where does the legal requirement for an annual AI bias audit come from?
The requirement is in NYC Administrative Code § 20-871, enacted as Local Law 144 of 2021 and implemented by NYC DCWP rules at 6 RCNY § 5-301. It makes it unlawful to use an AEDT in employment decisions unless the tool has been audited within the past year, results are publicly posted, and candidates receive advance notice. Enforcement began July 5, 2023.
The requirement is in New York City Administrative Code Section 20-871, which was enacted as Local Law 144 of 2021. The full city rules implementing the statute are at Title 6 of the Rules of the City of New York, Section 5-301, adopted by the NYC Department of Consumer and Worker Protection (DCWP). The law and its implementing rules together form what's typically called the "AEDT rules."
The law says it is unlawful for an employer or employment agency to use an automated employment decision tool to substantially assist or replace discretionary decision making for employment decisions unless:
- The AEDT has been the subject of a bias audit conducted within the past year
- The results of that audit have been made publicly available on the employer's website
- The employer has provided certain notices to candidates and employees
Enforcement began on July 5, 2023. (NYC Department of Consumer and Worker Protection — Automated Employment Decision Tools)
The law applies to employers who use automated employment decision tools for decisions about candidates or employees who work in New York City. "Automated employment decision tool" is defined as any computational process, derived from machine learning, statistical modeling, data analytics, or artificial intelligence, that issues simplified output — including a score, classification, or recommendation — that is used to substantially assist or replace discretionary decision making for employment decisions. (NYC Admin. Code § 20-870)
What does an NYC Local Law 144 bias audit technically require?
Under 6 RCNY § 5-301, the audit must be conducted by a genuinely independent third party (not the vendor or internal team), use historical employment data or — if insufficient — test data with explanation, calculate selection rates and impact ratios for each EEO-1 race/ethnicity and sex category, and produce a public summary including audit date, data source, category counts, selection rates, and impact ratios. The audit is valid for twelve months, then must be repeated.
The bias audit has specific technical requirements set out in the NYC DCWP rules. Here's what the implementing regulations actually require. (6 RCNY § 5-301)
The auditor must be impartial. The rule requires the audit to be conducted by an "independent auditor" — meaning someone who is not involved in using, selling, or developing the AEDT and who has no financial interest in the tool. Your vendor cannot audit their own tool. Your internal data science team cannot conduct the audit. It must be a genuinely independent third party.
The data source. The audit must be conducted using historical data about the tool's performance if the employer has at least 100 people in each sex category and 100 people in each race/ethnicity category in the data. If the employer doesn't have sufficient historical data, the auditor may use other data — data from multiple employers using the same tool, or test data — but the auditor must explain why in the audit summary.
The calculation. The auditor calculates the selection rate for each race/ethnicity category and sex category as classified in EEO-1 categories. The EEO-1 categories for race/ethnicity are: Hispanic or Latino, White (not Hispanic or Latino), Black or African American (not Hispanic or Latino), Native Hawaiian or Other Pacific Islander (not Hispanic or Latino), Asian (not Hispanic or Latino), American Indian or Alaska Native (not Hispanic or Latino), and Two or More Races (not Hispanic or Latino).
The impact ratio. For each category, the auditor calculates the impact ratio: the category's selection rate divided by the selection rate of the most-selected category. The auditor also calculates the impact ratio for each sex/race combination category if the data is sufficient.
The summary. The auditor produces a summary of results that must be publicly posted on the employer's website. The summary must include the date of the audit, the source and dates of data used, the number of individuals in each category included in the audit, the selection rate for each category, and the impact ratio for each category.
Timing. The audit must have been conducted in the twelve months before the tool is used. If you conduct an audit in January 2026, the audit is valid through January 2027. After that, you need a new audit.
This is a technical process. The auditor needs access to the tool's output data — who it selected, who it didn't, and what demographic categories they fall into. It requires actual numbers, actual calculations, and a formal write-up. It is not a policy review or a design inspection; it is a measurement of real outcomes.
Which employers need an annual AI bias audit under NYC Local Law 144?
Any employer or employment agency using an AEDT for hiring or promotion decisions involving New York City roles is covered — regardless of where the employer is headquartered. The tool must use machine learning, statistical modeling, data analytics, or AI to produce a score, classification, or recommendation that substantially assists or replaces discretionary decision-making (NYC Admin. Code § 20-870). Employers cannot delegate the audit obligation to their vendor.
The law applies to employers and employment agencies who use an AEDT in employment decisions for roles located in New York City. This includes:
- Decisions about candidates for a role
- Decisions about promotion or other employment decisions about current employees
The employer does not need to be based in New York City. If your company is headquartered in Dallas and you have employees in New York City, and you use an AI tool that helps make decisions about those employees, you're covered.
The scope is limited to tools that "substantially assist or replace" discretionary decision making. A tool that merely stores data and lets a human manually review it probably doesn't qualify. A tool that ranks candidates, assigns scores, generates recommendations, or automatically filters applicants — those are the target.
If you use a third-party platform like an applicant tracking system or a talent intelligence tool, you need to check whether it uses machine learning, statistical modeling, or algorithmic scoring. Many do. The law applies to the employer, not just the vendor. You cannot outsource the audit obligation to the platform provider.
What are the penalties for not conducting an annual AI bias audit under NYC Local Law 144?
NYC DCWP can impose up to $500 for a first violation (§ 20-872(a)) and $500–$1,500 per subsequent violation per day. Each use of a non-audited AEDT on each candidate or employee is a potential separate violation — meaning a busy HR department processing hundreds of applications monthly faces compounding daily exposure. DCWP can also require the employer to cease using the tool and mandate corrective action.
The NYC Department of Consumer and Worker Protection can impose civil penalties for violations. The penalty amounts are:
- Up to $500 for a first violation (§ 20-872(a))
- $500–$1,500 for each subsequent violation per day
But those amounts apply per day. An employer using a non-audited AEDT is potentially accumulating violations on a daily basis — and each use of the tool to make a decision about an individual candidate or employee could be treated as a separate violation. For a busy HR department processing hundreds of applications a month, that adds up fast.
In addition to fines, the DCWP can require employers to cease using the tool and can require corrective action. The agency has authority to investigate complaints, conduct audits, and require employers to provide documentation. (NYC Department of Consumer and Worker Protection — Enforcement)
Beyond NYC's specific penalties, using a biased hiring tool exposes an employer to federal civil rights liability. The EEOC has been explicit that AI hiring tools producing discriminatory outcomes violate Title VII of the Civil Rights Act, regardless of whether the discrimination was intentional. (EEOC Technical Assistance on Artificial Intelligence) The bias audit requirement is NYC's mechanism for making employers catch and address disparate impact before it generates individual discrimination claims.
What are the limits of an AI bias audit under NYC Local Law 144?
A bias audit does not certify a tool as fair — it measures whether current outcomes show statistically significant disparities across demographic groups. It does not evaluate decision quality (an equally wrong tool for all groups would pass). It does not satisfy Illinois HB3773, EEOC guidelines, or other state notice and oversight requirements. And it is a point-in-time snapshot — if the tool is retrained or updated, the prior audit may no longer reflect current performance, which is why annual audits are required.
Understanding the limits of a bias audit matters as much as understanding what it does.
A bias audit does not certify that a hiring tool is fair. It measures whether the tool's current outcomes produce statistically significant disparities between demographic groups. If the tool has low disparate impact today, that's a good sign — but it doesn't mean the tool is unbiased in any deeper sense, and it doesn't mean it will stay that way as the tool continues to train on new data.
A bias audit does not evaluate the quality of the tool's decisions. An AI that consistently rejects the most qualified candidates across all demographic groups equally would pass a bias audit. The audit tests for equal treatment across groups, not for the accuracy or merit of the tool's selections.
A bias audit does not protect you from other legal obligations. Even a tool with a clean bias audit can violate the Illinois Human Rights Act, the EEOC's guidelines, or other state laws if it's used without proper notice, without human review, or in ways that otherwise violate those statutes. The NYC bias audit satisfies NYC LL144's audit requirement. It doesn't satisfy everything else.
A bias audit does not prevent future problems. It's a snapshot. If a tool is retrained, fine-tuned, or updated, the prior audit may no longer reflect current performance. This is one reason NYC requires annual audits rather than a one-time certification.
How does an employer actually get an NYC Local Law 144 bias audit completed?
If using a major enterprise platform, ask whether the vendor has retained an independent auditor and whether results are available — verify the auditor's independence and that the audit used real employment data within the past year. If the vendor doesn't provide this, retain an independent firm with statistical expertise. Before the audit, prepare a dataset of tool outputs with EEO-1 demographic categories and documentation of how the tool was used in actual decisions.
There are several paths, depending on your situation.
If you're using a major enterprise platform, ask your vendor whether they have retained an independent auditor and whether they make the audit results available. Some larger platforms — including some applicant tracking systems and talent intelligence tools — have arranged audits at the platform level and make summary results available to customers. If your vendor provides this, verify that the auditor is genuinely independent, that the data used reflects actual employment decisions (not just demonstration data), and that the audit was conducted within the past year.
If your vendor doesn't provide this, or if you want to verify their compliance independently, you will need to retain an independent auditor. A number of firms now specialize in AI bias auditing. You're looking for a vendor with a statistician or data scientist who can work with your tool's output data, apply the required calculations, and produce a compliant summary. The audit itself is a quantitative exercise — it doesn't require legal interpretation.
If you don't have sufficient historical data, talk to the auditor about options. The NYC rules allow auditors to use other sources of data if historical data is insufficient, as long as the audit summary explains the limitation.
Before the audit, make sure you can provide the auditor with: a dataset of tool outputs (who was selected, who wasn't), the demographic data associated with those candidates or employees (using EEO-1 categories), and documentation of how the tool was used (what decisions it was used to make, what threshold was used for selection).
After the audit, post the required summary on your website and maintain the documentation. If the audit reveals a significant disparity, you'll need to address it — either by adjusting how the tool is used, requesting that the vendor retrain or reconfigure the model, or discontinuing the tool.
What notice requirements does NYC Local Law 144 impose beyond the annual bias audit?
Employers must provide written notice to candidates at least ten business days before using an AEDT, identifying that an AEDT will be used, the job qualifications and characteristics the tool will evaluate, and instructions for requesting an alternative process or accommodation (NYC Admin. Code § 20-871). Current employees must receive the same notice at least ten business days in advance. Both notices must include a link to the publicly posted bias audit summary. Notice and audit are separate obligations — satisfying one does not satisfy the other.
The bias audit is only one part of LL144 compliance. The law also requires employers to provide specific notices before using the tool.
For job candidates, the employer must provide notice at least ten business days before using the tool, informing them that an AEDT will be used in the assessment, listing the job qualifications and characteristics the tool will use, and providing instructions for how to request an alternative selection process or accommodation. (NYC Admin. Code § 20-871)
For current employees, similar notice must be given at least ten business days before the tool is used.
The notice must also include a link to the publicly posted bias audit summary.
These notice requirements are separate from the audit. Even a perfectly compliant bias audit doesn't satisfy the notice obligation. Both must be done.
What do Illinois HB3773 and federal law add to the AI bias audit requirement beyond NYC?
Illinois HB3773 (775 ILCS 5/2-102(L), effective January 1, 2026) makes it a civil rights violation to use AI that "has the effect of" discriminating based on protected class — and a bias audit is the most direct evidence that a tool is not producing discriminatory effects. The EEOC's 2021 AI initiative confirmed that Title VII, ADEA, and ADA all apply to AI hiring outcomes regardless of vendor responsibility. The federal Uniform Guidelines on Employee Selection Procedures (29 CFR Part 1607) have applied to employment selection systems since 1978 and cover AI tools.
NYC's LL144 is the only US law that specifically mandates an annual third-party bias audit. But it's not the only law that creates pressure to conduct one.
Illinois HB3773, which took effect January 1, 2026, makes it a civil rights violation under the Illinois Human Rights Act to use AI in employment decisions in a way that "has the effect of" discriminating based on protected class characteristics. (775 ILCS 5/2-102(L)) The law doesn't require an annual third-party audit by name, but to demonstrate that your AI tools are not producing discriminatory effects — which is what you'd need to show in defense of an IDHR charge — a bias audit is the most direct evidence available.
The EEOC's position, stated in its October 2021 initiative on AI and algorithmic fairness, is that AI-assisted employment decisions are subject to Title VII, the Age Discrimination in Employment Act (ADEA), and the Americans with Disabilities Act (ADA), and that employers are responsible for the discriminatory outcomes of AI tools even if the tools are provided by a third party. (EEOC AI Initiative) A bias audit, while not required under federal law, is the kind of evidence that demonstrates good-faith effort to comply.
The federal Uniform Guidelines on Employee Selection Procedures, adopted jointly by the EEOC, the Department of Labor, the Department of Justice, and the Civil Service Commission, are the foundational framework for disparate impact analysis in employment selection. Those guidelines have applied to employment selection procedures since 1978 and have always covered tests, scoring systems, and other selection mechanisms. AI hiring tools are selection procedures. The guidelines apply. (EEOC Uniform Guidelines)
Where is the AI bias audit requirement heading beyond New York City?
NYC remains the only US jurisdiction mandating an annual third-party bias audit by name, but the EU AI Act (Regulation (EU) 2024/1689) requires conformity assessments and ongoing monitoring for AI used in employment beginning August 2, 2026 — a parallel obligation. Enterprise procurement teams are increasingly treating bias audit documentation as a baseline vendor requirement alongside SOC 2 reports, accelerating adoption well beyond jurisdictions where it is legally mandated.
As of March 2026, NYC is the only US jurisdiction that mandates an annual third-party bias audit by name. But the direction of travel is clear.
Several states are actively considering or have recently considered bias audit requirements. The EU AI Act, which begins applying to high-risk AI systems in August 2026, requires conformity assessments and ongoing monitoring for AI systems used in employment — the underlying requirement is similar in spirit to a bias audit, even if it uses different terminology. (Regulation (EU) 2024/1689)
The more immediate practical pressure is procurement. Large enterprise clients, government contractors, and regulated industries are beginning to ask vendors and service providers whether their AI tools have been audited. Institutional buyers are treating bias audit documentation the way they treat SOC 2 reports: as a baseline requirement for doing business. That trend is accelerating.
If you're a company that sells AI tools or uses them in client-facing work, the question of whether you have a bias audit is increasingly a business question, not just a compliance question.
Where should employers start to comply with NYC Local Law 144's bias audit requirement?
First, determine whether your hiring tool qualifies as an AEDT: does it use machine learning, statistical modeling, data analytics, or AI to produce a score, classification, or recommendation used to substantially assist or replace a hiring decision? If yes, verify whether your vendor has a current independent audit, confirm the audit uses real employment data and was done within the past year, check that candidate notices go out at least ten business days before tool use, and confirm the audit summary is posted on your website.
If you're an employer with employees in New York City and you use any AI tool in hiring or employment decisions, the first step is determining whether the tool qualifies as an AEDT under the law's definition. If it does, verify whether your vendor has conducted an audit, confirm the audit's independence and recency, check whether the required notice is going to candidates and employees, and confirm the summary is posted on your website.
If you're not sure whether your tool qualifies, the answer is: describe what it does and apply the definition. Does it use machine learning, statistical modeling, data analytics, or AI? Does it produce a score, classification, or recommendation? Is that output used to substantially assist or replace a hiring decision? If the answer to all three is yes, you have an AEDT and you need a bias audit.
This is not a process you can leave entirely to vendors. The obligation belongs to the employer. The documentation needs to exist. And if someone files a complaint, the first thing DCWP will ask for is your audit summary and your notice records.
Our AI Bias Audit Template includes the audit summary format, candidate notice templates, and posting checklist required under LL144 — and our NYC Local Law 144 compliance package covers the full set of employer obligations from audit through notice requirements.
Sources — Every statutory and regulatory fact in this article was verified against the enacted law and implementing regulations at these URLs:
- NYC Local Law 144 of 2021 — Int 1894-2020 (Automated Employment Decision Tools) — Enacted text of the AEDT law, effective January 1, 2023; enforcement began July 5, 2023.
- NYC Administrative Code § 20-870 through 20-872 — Statutory definitions, bias audit mandate, notice requirements, and penalty provisions.
- 6 RCNY § 5-301 — DCWP Implementing Rules for AEDT (April 2023) — Technical requirements for bias audits including selection rate calculations, impact ratio methodology, independent auditor standards, and public posting requirements.
- NYC Department of Consumer and Worker Protection — Automated Employment Decision Tools — DCWP enforcement guidance, FAQs, and compliance resources.
- EEOC Uniform Guidelines on Employee Selection Procedures — Q&A — Federal framework for disparate impact analysis including the four-fifths rule; applies to AI-powered selection procedures.
- EEOC Initiative on Artificial Intelligence and Algorithmic Fairness — EEOC's October 2021 formal initiative; positions employers as responsible for discriminatory AI hiring outcomes under Title VII, ADEA, and ADA.
- Illinois Human Rights Act § 2-102(L) — 775 ILCS 5/2-102(L) — Illinois AI employment discrimination prohibition; effective January 1, 2026.
- Regulation (EU) 2024/1689 — EU AI Act — High-risk AI system requirements including employment AI; applies from August 2, 2026.
Disclaimer: This article is for general informational purposes only and does not constitute legal advice. The law in this area is evolving, and individual circumstances vary. Consult a qualified employment attorney for guidance specific to your situation.
Why Bias Testing Looks at Outcomes, Not Intentions
Imagine a teacher creates a final exam that she genuinely believes is fair. She didn't write the questions to favor any particular group of students. She used the same textbook material everyone had access to. She gave everyone the same amount of time. From her perspective, the test is completely neutral. But when the grades come in, students who speak English as a second language fail at three times the rate of native English speakers — not because the subject matter is harder for them, but because the questions are phrased in unnecessarily complex English that tests language fluency more than subject knowledge. The teacher didn't intend to discriminate. But the result is discriminatory.
This is exactly the logic behind bias audits for AI hiring tools. The law doesn't ask whether the programmer intended to build a biased tool. It doesn't ask whether the company wanted to discriminate. It asks a simpler, more measurable question: do the outcomes look different for different groups? If your AI resume screener advances 40% of white applicants but only 20% of Black applicants, the numbers speak for themselves — regardless of what the algorithm was 'trying' to do. That gap is what employment law calls adverse impact, and it's been the standard for evaluating discrimination in hiring since 1978.
The reason the law focuses on outcomes rather than intentions is practical. You can't look inside an AI model and identify a line of code that says 'discriminate against this group.' Modern machine learning models are complex enough that even their creators often can't fully explain why a particular applicant received a particular score. But you can look at the results — who got through the screen and who didn't — and measure whether those results break down along demographic lines. That measurement is something an independent auditor can perform, document, and publish. It's concrete and verifiable in a way that 'we didn't mean to discriminate' never can be.
This is why NYC Local Law 144 requires selection rate calculations and impact ratios broken down by the same race, ethnicity, and sex categories used in federal EEO-1 reporting. The audit doesn't ask the tool to explain itself. It measures what the tool actually does to real applicants, compares those outcomes across groups, and flags any disparity that crosses the four-fifths threshold established by the federal Uniform Guidelines. It's the difference between trusting a restaurant's claim that their kitchen is clean and sending a health inspector to look.
4 facts
- [1]NYC Department of Consumer and Worker Protection — Automated Employment Decision Tools (opens in new tab)
- [2]29 CFR Part 1607 — Uniform Guidelines on Employee Selection Procedures (eCFR) (opens in new tab)
- [3]EEOC — Questions and Answers on the Uniform Guidelines (opens in new tab)
- [4]EEOC — AI and Algorithmic Fairness Initiative Launch (October 2021) (opens in new tab)
- [5]Illinois Human Rights Act § 2-102 (775 ILCS 5/2-102) — AI Employment Provision (opens in new tab)
- [6]EU AI Act — European Commission Regulatory Framework Page (opens in new tab)
Disclaimer: This article is for informational purposes only and does not constitute legal advice, legal representation, or an attorney-client relationship. Laws and regulations change frequently. You should consult a licensed attorney to verify that the information in this article is current, complete, and applicable to your specific situation before relying on it. AI Compliance Documents is not a law firm and does not practice law.
More from the blog
Texas TRAIGA Has Been Live for 4 Months. Here's What the AG Is Doing — and What You Should Be Ready For.
Texas TRAIGA has been live for 4 months. Zero public AG enforcement so far. The complaint portal launches September 1, 2026 — and what you have documented before that matters more than what you do after.
Colorado's AI Law Takes Effect June 30, 2026. Here's What It Requires.
Colorado's AI law takes effect June 30, 2026. No amending bill has been introduced. The legislature has failed to revise the law four times. The deadline is real.
Workday AI Hiring Lawsuit Could Reshape Employer Liability
A federal court is testing whether AI vendors — not just employers — can be sued for discriminatory hiring outcomes. The certified class could include hundreds of millions of applicants.
Colorado AI Compliance for HR Software Companies: What SB 24-205 Means for Your Product
HR software that screens candidates, scores performance, or ranks employees is classified as high-risk AI under Colorado's law. The June 30, 2026 deadline applies to both the companies that build these tools and the HR teams that use them.
Do I Need AI Compliance? A Decision Framework for Every Business Using AI
Not sure if AI compliance applies to your business? Walk through four questions — and know exactly which laws apply, which documents you need, and where to start.
Operating in Multiple States? Here's How AI Compliance Stacks Up Across 15 Jurisdictions
Colorado, California, Texas, Illinois, and NYC all have active AI laws — and they don't all require the same things. If you operate in multiple states, here's what applies to you and why.
Oregon Consumer Privacy Act: What Your Business Needs to Know About AI Profiling Requirements
Oregon's privacy law has been in effect since July 2024, requires data protection assessments for AI profiling, and flatly prohibits processing personal data of consumers under 16 for targeted advertising or data sales — a protection not found in most other state laws. The 30-day cure period effectively expired for most businesses on January 1, 2026 (Oregon Laws 2025, c.417).
What Is an AI Impact Assessment? The Document Every State Law Now Requires
Colorado, California, and Illinois all require some version of an AI impact assessment — but they don't call it the same thing or require the same format. Here's what every version has in common, and what each state specifically demands.
What Is a High-Risk AI System? A Plain-Language Guide for Business Owners
Three different laws. Three different definitions of 'high-risk AI.' If your business uses AI to make decisions about people, here's how to figure out which rules apply to you.
The Federal Government Quietly Removed Its AI Hiring Guidance. Four States Are Writing Their Own.
The federal government removed every page of AI hiring guidance it ever published. Over a year later, the pages are still down. Four states wrote their own — and none of them agree.
AI governance framework checklist: what every enacted state law actually requires
Colorado, Texas, and Illinois all passed AI laws with deadlines in early 2026 — and none of them are identical. Here's the one compliance checklist that covers all three at once.
You're HIPAA-Compliant. That's Not Enough Anymore.
HIPAA protects patient records. It has nothing to say about whether the AI making decisions about those patients is fair. New rules are filling that gap — and they apply to you even if your HIPAA program is airtight.
The NIST AI Risk Management Framework: What It Is and Why Colorado Made It a Legal Shield
The US government published a free framework for managing AI risk — and Colorado's AI law turns following it into a legal shield. If something goes wrong with your AI, this is the document that shifts the burden of proof.
Texas TRAIGA (HB 149): What the Texas Responsible AI Governance Act Requires and How to Comply
Texas passed an AI law that applies to every business — no exemptions for small companies, no carveout for low-risk tools. It's already in effect, and a single uncurable violation starts at $80,000.
What Does AI Compliance Actually Cost a Small Business in 2026?
AI compliance can cost $49 or $50,000 — depending on what you actually need. Here's what each option costs in real numbers, so you can stop guessing and start budgeting.
AI Compliance Penalties by State: What Happens If You Ignore the Law
"Per violation" sounds like one fine. It isn't. Here's what the penalty math actually looks like state by state — and why the numbers can compound into company-ending territory fast.
AI and HIPAA: What Healthcare Businesses Must Do Now
If an AI tool touches patient data at your healthcare organization, HIPAA applies — and most vendor contracts aren't written to cover it. Here's what you need before you deploy.
EU AI Act Compliance Checklist: What US Businesses Need Before August 2026
Europe's AI law applies to US companies — even ones with no European office. If your AI is used by anyone in the EU, the deadline is August 2026 and the fines are calculated on your global revenue.
ISO 42001: The AI Certification Your Enterprise Clients Will Soon Require
Enterprise clients are starting to require ISO 42001 certification before they'll buy AI products — the same way they require SOC 2. Here's what the standard actually requires and why getting it early is a competitive advantage.
California Just Finalized Its AI Regulations. Here's What Your Business Actually Needs to Do.
California's AI rules are already in effect — and the agency enforcing them just handed out its largest fine ever. Here's what your business needs to do and when.
Get your compliance documentation done
Stop reading, start complying. Our packages generate the documents you need based on the actual statutes.
Browse Compliance Packages













