AI Acceptable Use Policy Template
by Jon Lober | NOC Technology
Set Your Business Up for Acceptable AI Usage
Surveys consistently show that a majority of employees use AI tools at work without telling their managers. They're drafting emails, summarizing documents, and generating reports—often with company data flowing into systems you don't control and may not have vetted.
Most small businesses across the St. Louis area fall into two camps: they've banned AI outright (which employees quietly ignore), or they have no policy at all (which creates real liability). A middle path exists—one that lets your team use these tools productively while keeping sensitive data where it belongs. This guide walks through what a strong AI acceptable use policy should include, with sample language you can customize for your business.
Why You Need an AI Policy (Even If You're Small)
The assumption that "we're too small for formal policies" breaks down quickly when it comes to AI. Here's why even a 15-person company needs written guidelines.
Liability exposure is the most immediate concern. When an employee pastes client financial data into ChatGPT to help with analysis, that data may become part of a training set (on free tiers) or accessible in ways that are hard to predict. If that client's confidential information surfaces elsewhere, your company may bear responsibility—regardless of whether you knew it was happening.
Data exposure from AI happens quietly. There's no breach notification, no incident alarm—just quiet risk accumulation as employees make individual judgment calls about what's appropriate to share with an AI tool.
Regulatory compliance adds another layer. If you're in healthcare, legal, or financial services, specific data handling requirements apply. HIPAA doesn't consider intent when an employee asks an AI to summarize patient notes. Missouri healthcare practices and law firms are especially exposed here because AI's efficiency gains are highest in exactly the document-heavy industries where compliance requirements are strictest.
What Your AI Acceptable Use Policy Should Cover
A solid AI policy doesn't need to be lengthy. It needs to be clear, practical, and specific enough that employees can actually follow it. Here are the core elements.
Approved tools form the foundation. Your policy should specify which AI tools employees may use for work purposes. This ensures you've vetted the tools your team relies on. You might approve Microsoft Copilot under your enterprise agreement while restricting free-tier ChatGPT because of its data training policies.
Prohibited uses need explicit definition. Be specific: "Employees may not input client names, financial data, personally identifiable information, trade secrets, or confidential business information into any AI tool." Vague language like "don't share sensitive data" leaves too much room for interpretation.
Data classification tells employees what falls into each category. Your policy should include concrete examples so the line between "okay to use" and "never use" is clear.
Review requirements establish when human oversight is mandatory before using AI-generated output in client-facing or official communications.
Training expectations make the policy actionable. Require employees to complete training before using AI tools for work. A policy people don't understand is a policy people won't follow.
Sample policy language:
"Employees may use approved AI tools to assist with general research, drafting, summarization, and ideation tasks. AI-generated output must be reviewed for accuracy before use in any client-facing or official communication. Employees must not input confidential company information, client data, personally identifiable information, financial records, or trade secrets into any AI tool without explicit written approval."
Building and Maintaining Your Approved Tools List
Shadow AI is a real problem. Employees find tools that make their jobs easier and use them—whether or not those tools appear on an approved list. Your policy needs to address this reality directly.
When evaluating a new AI tool, consider: Does the enterprise tier handle data differently than the free tier? (Many do—ChatGPT's free tier trains on user input; Team and Enterprise tiers do not.) What are the vendor's data retention and training policies? Does the tool integrate with your existing security controls like SSO and audit logging?
Your approved tools list should be a living document with a clear owner—someone responsible for reviewing new requests and periodically re-evaluating existing approvals as vendor policies change.
Sample language for tool approval:
"Only AI tools appearing on the Company Approved AI Tools List may be used for work purposes. Employees who wish to use a tool not on the list must submit a request for evaluation. Using unapproved tools for work-related tasks is prohibited and may result in disciplinary action."
Data Classification for AI Use
Abstract rules about "sensitive data" don't help in the moment. Your policy needs concrete categories with examples.
Green — Generally Safe for AI Input:
- Publicly available information (anything on your website, public press releases)
- General industry questions ("What are common approaches to project management?")
- Grammar and style checks on non-confidential text
- Brainstorming and ideation without client specifics
Red — Never Input into AI Tools:
- Client names, contact information, or any identifying details
- Personally identifiable information (Social Security numbers, dates of birth, addresses)
- Financial data (account numbers, transaction records)
- Trade secrets and proprietary processes
- Legal documents, contracts, or privileged communications
- Healthcare information (PHI under HIPAA)
- Employee HR records or performance data
Yellow — Requires Judgment or Approval:
- Internal processes not documented publicly
- Aggregated or anonymized data (verify it's truly anonymized)
- Vendor communications that may contain contractual details
Sample language:
"Before inputting information into an AI tool, employees must classify the data per Company Data Classification Guidelines. Green-classified information may be used freely. Yellow requires manager approval. Red must never be entered into any AI system regardless of the tool's security claims. When classification is unclear, consult IT or Compliance before proceeding."
Enforcement and Training: Making the Policy Stick
A policy that sits in a handbook unread creates the illusion of protection without the reality. When you introduce the policy, explain the reasoning behind each restriction. Employees who understand that free-tier ChatGPT may train on their inputs will make better decisions than employees who just know "don't use free ChatGPT."
Train employees on the specific tools you've approved, the specific data categories you've defined, and the specific workflows where AI makes sense for your business—not generic ethics modules.
Make compliance easy. If your approved enterprise AI tool is harder to access than the free alternative, employees will default to the free alternative. Approved tools should be readily available, properly configured, and genuinely useful.
Handle violations progressively. First violations—especially when the employee didn't understand the risk—should trigger retraining and a conversation, not termination. Reserve serious action for deliberate disregard or actual data exposure.
Sample enforcement language:
"Violations will be addressed through progressive discipline. Unintentional first-time violations will result in additional training and a documented conversation. Repeated violations or intentional disregard for data protection requirements may result in formal disciplinary action up to and including termination."
Getting Started
Start with what you have. If you already have acceptable use policies for technology, add an AI-specific addendum rather than starting from scratch. Focus on the five elements: approved tools, prohibited uses, data classification, review requirements, and training expectations.
For organizations looking to implement AI governance alongside a broader technology strategy, NOC's Managed Intelligence services include policy development, tool evaluation, and ongoing monitoring. We also cover the security implications of AI tools as part of our cybersecurity services for St. Louis area businesses.






