🔴Illinois HB 3773IN EFFECT$10M fine|🔴Texas TRAIGAIN EFFECTActive enforcement|⚠️Colorado SB 205Jun 30, 2026Per-violation fines|⚠️California SB 942Aug 2, 2026$5K/day|⚠️EU AI Act Art. 50Aug 2, 2026€35M or 7% revenue|⚠️Virginia HB 2154Jul 1, 2026$10K/violation|⚠️Connecticut SB 2Oct 1, 2026$25K/violation|🔴Illinois HB 3773IN EFFECT$10M fine|🔴Texas TRAIGAIN EFFECTActive enforcement|⚠️Colorado SB 205Jun 30, 2026Per-violation fines|⚠️California SB 942Aug 2, 2026$5K/day|⚠️EU AI Act Art. 50Aug 2, 2026€35M or 7% revenue|⚠️Virginia HB 2154Jul 1, 2026$10K/violation|⚠️Connecticut SB 2Oct 1, 2026$25K/violation|
HomeMissouriNonprofitMid-Market (51-250)
🏢

Missouri AI Laws for Mid-Market (51-250) in Nonprofit

You likely need a dedicated compliance officer. Formal impact assessments and bias audits may be required.

By AI Law Tracker Editorial Team · Last verified April 29, 2026

AI Compliance Context for Missouri

Missouri's regulatory posture on AI is silence rather than permission: missouri considered hb 1687 (ai liability) in 2024 but did not advance; no ai-specific statute; monitoring neighboring illinois hb 3773 and kansas ai working group. No comprehensive state privacy statute; UDAP coverage via Missouri Merchandising Practices Act (Mo. Rev. Stat. sec. 407.020) provides the residual framework. For donor-targeting, program-eligibility, and fundraising AI in Missouri, federal signals set the ceiling while regional precedent sets the floor.

Three neighboring regimes create compounding exposure: Iowa (AI in Government Act, penalty Administrative), Illinois (HB 3773 — AI in Employment, penalty Up to $5,000 per violation (willful/repeated)), and Kentucky (AI Study Resolution, penalty TBD). Multi-state Nonprofit operators headquartered in Missouri default to the strictest stack.

Missouri's non-legislation on AI means the Missouri Attorney General office has discretion to apply no comprehensive state privacy statute to AI-driven consumer harms as they arise.

Federal law still governs Nonprofit AI in Missouri primarily through IRS 501(c)(3) rules (26 USC 501), FTC Telemarketing Sales Rule (16 CFR 310), and state charitable-solicitation registration. Adjacent federal authorities include IRC Section 501(c)(3) Political Campaign Prohibition (26 U.S.C. Section 501(c)(3); Rev. Rul. 2007-41); OMB Uniform Guidance (2 CFR Part 200) (2 CFR Part 200); IRS Form 990 Schedule O (IRS Form 990, Schedule O). IRC Section 501(c)(3) Political Campaign Prohibition (enforced by Internal Revenue Service) applies to absolute prohibition on participation in, or intervention in (including the publishing or distributing of statements), any political campaign on behalf of or in opposition to any candidate for public office. ai-generated political content counts toward the prohibition; automated voter-targeting tools that favor a candidate risk revocation. Penalty exposure: revocation of tax-exempt status; excise tax under irc section 4955 on political expenditures; excise tax under section 4958 on excess benefit transactions. IRS political-campaign-intervention enforcement combined with state charitable-solicitation oversight creates dual-track exposure for AI-driven outreach.

The federal and neighboring-state framework that governs your AI operations. Nonprofit operators in Missouri operate under a federal-dominant framework anchored by IRS 501(c)(3) rules (26 USC 501), FTC Telemarketing Sales Rule (16 CFR 310), and state charitable-solicitation registration, with adjacent authorities IRC Section 501(c)(3) Political Campaign Prohibition (26 U.S.C. Section 501(c)(3); Rev. Rul. 2007-41); OMB Uniform Guidance (2 CFR Part 200) (2 CFR Part 200); IRS Form 990 Schedule O (IRS Form 990, Schedule O). IRS political-campaign-intervention enforcement combined with state charitable-solicitation oversight creates dual-track exposure for AI-driven outreach. The practical risk they have to price in is violation of IRC Section 501(c)(3) political-campaign prohibition via AI-generated voter content plus federal-grant internal-control failures under 2 CFR Part 200, and the bellwether signal to monitor is federal-grant recipients must satisfy OMB Uniform Guidance internal-control and cost-principle requirements when AI is used to allocate federally-funded program benefits. Iowa -- AI in Government Act sets the de-facto regional floor. Missouri considered HB 1687 (AI liability) in 2024 but did not advance; no AI-specific statute; monitoring neighboring Illinois HB 3773 and Kansas AI Working Group. Use this as a starting point; sector pages on this site go deeper into industry-specific obligations.

At 51-250 employees you need a dedicated compliance officer, a formal AI inventory, and working relationships with specialist outside counsel. Medium-stage Nonprofit operators should deploy a dedicated AI governance committee, mandatory impact assessments for new deployments, and third-party audit on a rolling schedule, with quarterly internal review and annual third-party audit and ownership resting with a Head of AI Governance reporting to the COO or General Counsel. mid-market programs ($250K-$1.5M) typically combine a dedicated compliance officer with enterprise GRC tooling. For Nonprofit specifically, the sharpest exposure to manage is violation of IRC Section 501(c)(3) political-campaign prohibition via AI-generated voter content plus federal-grant internal-control failures under 2 CFR Part 200. Given Missouri's concentration in transportation logistics, financial services, and healthcare, freight-routing algorithms, consumer-lending models, and rural telehealth AI deserve priority in your AI inventory.

The enforcement surface for Nonprofit centres on IRS Exempt Organizations Division, OMB / federal grantor agency Inspectors General, EEOC, and the statute operators most often under-document is OMB Uniform Guidance (2 CFR Part 200) (2 CFR Part 200) — a gap that surfaces in violation of IRC Section 501(c)(3) political-campaign prohibition via AI-generated voter content plus federal-grant internal-control failures under 2 CFR Part 200 disputes. Build an evidence binder covering donor-consent ledger, charitable-solicitation registration trail, 501(c)(3) non-intervention log, Schedule-O narrative, and grant-allocation audit file. Treat federal-grant recipients must satisfy OMB Uniform Guidance internal-control and cost-principle requirements when AI is used to allocate federally-funded program benefits as your leading indicator and escalate when the signal shifts.

Verified 2026-04-29. See https://ago.mo.gov/ for the Missouri Attorney General public record on Missouri AI policy.

Applicable law: No AI-specific law

No state-specific AI law. Federal laws apply. Missouri AG monitors AI-driven consumer protection violations under the Merchandising Practices Act.

Nonprofits using AI for grant decisions or donor profiling face emerging transparency requirements.

Deadline: N/APenalty: N/AStatus: No Law

What this means for Mid-Market (51-250) in Nonprofit

For a mid-market (51-250) nonprofit business operating in Missouri, AI compliance is a concrete and present-tense concern. At this size, you should have dedicated HR, legal, or compliance capacity and the organizational structure to support formal programs. The central challenge is maintaining consistent compliance across multiple departments that adopt AI tools independently and at different paces — and understanding exactly what No AI-specific law requires of an organization at your headcount is the essential foundation.

At the mid-market (51-250) tier, core compliance obligations under Missouri's framework include a formal AI inventory, a designated compliance officer with AI in their mandate, documented impact assessments for high-risk systems, annual bias audits for employment-affecting AI, and structured vendor compliance reviews. board-level AI governance, external annual audits, and public transparency reports are strongly recommended but not yet mandated at this size in most states — though they are required at the enterprise tier, so building toward them now is prudent. This proportionality is deliberate — regulators recognize that smaller organizations cannot sustain the same compliance infrastructure as large enterprises, but the law's fundamental requirements apply regardless of size.

The nonprofit sector's medium risk classification takes on particular relevance at this scale. Nonprofits using AI for grant decisions or donor profiling face emerging transparency requirements. For a mid-market (51-250) business, the risk materializes because maintaining consistent compliance across multiple departments that adopt AI tools independently and at different paces is more acute at this size — AI tools from vendors may have been adopted without full compliance review, and operational workflows where AI is embedded often develop faster than governance processes.

The highest-priority actions for a mid-market (51-250) nonprofit business in Missouri are: (1) conduct a formal ai impact assessment for every system that affects employees or customer outcomes; (2) establish a cross-functional ai governance committee with a documented charter and quarterly meetings; and (3) build vendor management procedures that include ai compliance questionnaires and contractual representations. These steps do not require outside counsel or enterprise compliance software — they can be executed with existing staff and documented in straightforward internal policies. The goal is to move from informal AI usage to documented AI governance, even if that governance is lightweight at first.

Understanding the financial stakes clarifies the urgency. at this size, the reputational damage of a public enforcement action routinely outweighs the direct financial penalty — particularly in states with disclosure-based enforcement frameworks. Under No AI-specific law, the maximum penalty is N/A. For a business at this size, that exposure — especially if it accrues on a per-violation basis across multiple AI touchpoints — warrants taking compliance seriously now rather than reactively. enterprise-scale obligations activate at the 250-employee threshold in most frameworks — prepare for that transition by investing in systems designed to mature rather than be replaced.

Beyond the headline compliance obligations, mid-market (51-250) nonprofit businesses in Missouri face specific employer and operator duties tied to how AI interacts with people — employees, customers, applicants, and others affected by automated decisions. When AI assists in decisions that affect people's access to services, job opportunities, credit, or housing, Missouri law treats the deploying organization as responsible for the outcome regardless of whether the underlying model was built in-house or acquired from a vendor. This means mid-market (51-250) operators cannot outsource accountability to their AI provider — vendor contracts should be reviewed for indemnification provisions, compliance representations, and audit rights. Documenting the due diligence you performed before selecting and deploying an AI system is itself a compliance requirement in several states, and a strong defense in enforcement proceedings.

The compliance timeline for a mid-market (51-250) nonprofit business in Missouri has several distinct phases. The first phase — inventory and assessment — involves documenting every AI system in use and evaluating whether it falls within the scope of No AI-specific law. Most compliance experts recommend completing this phase within the first 30 days of any new compliance program. The second phase — policy and disclosure — involves drafting the required notices, internal use policies, and vendor agreements. A 60-day target is realistic for most mid-market (51-250) organizations. The third phase — technical controls and ongoing monitoring — involves implementing audit logs, human review checkpoints for high-stakes decisions, and regular bias testing for any AI that affects protected populations. This phase is ongoing. With Missouri's deadline of N/A, the first two phases should be completed well before enforcement begins.

The enforcement landscape for AI compliance in Missouri is evolving, but the direction is consistent: regulators are moving from guidance to action. Once No AI-specific law takes effect in Missouri, enforcement typically begins immediately against the most visible violations — disclosure failures and bias-related incidents. For mid-market (51-250) nonprofit businesses, the highest-risk scenarios involve automated decisions affecting individuals in ways the law covers: hiring, lending, insurance pricing, and access to services. Regulators typically prioritize cases where AI-driven harm is documented, where disclosure requirements were clearly violated, or where a company failed to provide a mandated appeal or human review process. Building a compliance program now — even a lightweight one appropriate for a mid-market (51-250) organization — establishes a documented good-faith effort that regulators consistently weigh favorably in enforcement decisions. The cost of getting started is a fraction of the cost of responding to a formal investigation.

Missouri Nonprofit resources

Compliance Checklist
💰 Fines & Penalties
📋 Compliance Requirements
📖 Compliance Guide
Key Deadlines

Other company sizes

🚀 Startups (1-10)🏪 Small Business (11-50)🏛️ Enterprise (250+)

Serve EU customers? The EU AI Act may also apply — penalties up to €35M.

All Missouri lawsMissouri NonprofitAll NonprofitFree Assessment
Editorial standards

Sources verified against official .gov filings · Last verified Apr 29, 2026.

Official sources · Missouri