Blog

AI Governance: Managing Shadow AI and Securing AI Usage in Your Company

Auteur Keerok AI
Date 01 May 2026
Lecture 14 min

Shadow AI – the undeclared use of artificial intelligence tools by employees – has emerged as one of the most underestimated risks facing businesses today. With the European AI Act imposing fines up to €35 million or 7% of global turnover for governance failures (Formind, 2024), organizations can no longer afford to ignore unauthorized AI adoption within their teams. This practical guide provides actionable strategies for detecting, managing, and governing AI usage before it escalates into a security, compliance, or operational crisis.

Understanding Shadow AI: The Hidden Risk in Modern Organizations

Shadow AI refers to the unauthorized use of artificial intelligence tools and services by employees without formal approval from IT, legal, or management. Unlike traditional shadow IT, shadow AI introduces unique risks related to sensitive data processing by external models, often hosted outside regulated jurisdictions.

According to Blog du Modérateur (2024), informal AI tool adoption by employees is rapidly increasing, creating major risks of data leakage and information system compromise. This phenomenon stems from several converging factors:

  • Immediate accessibility: ChatGPT, Claude, Midjourney and similar AI tools are available in seconds without IT validation
  • Productivity pressure: employees seek legitimate ways to optimize their daily work
  • Lack of internal alternatives: when companies don't provide governed AI tools, employees turn to public solutions
  • Risk unawareness: most employees don't realize they're potentially exposing confidential data

The Concrete Business Risks of Unmanaged Shadow AI

The consequences of unmanaged shadow AI extend far beyond simple policy violations:

  1. Sensitive data exposure: a sales representative using ChatGPT to draft proposals potentially exposes customer information to OpenAI's training data
  2. GDPR non-compliance: processing personal data through third-party AI tools without legal basis constitutes a GDPR violation
  3. AI Act sanctions: as Tenexa (2024) highlights, with the AI Act's arrival, companies face regulatory sanctions due to lack of inventory, risk classification, and documentation for shadow AI usage
  4. Intellectual property compromise: source code, business strategies, or product innovations can be inadvertently shared
  5. Bias and discrimination: using AI for hiring or evaluation without proper auditing can generate illegal discrimination
  6. Technological dependency: critical business processes may become dependent on uncontrolled external tools
  7. Model poisoning and manipulation: adversarial actors can influence public AI models through prompt injection or training data contamination

The financial impact is substantial. The European AI Act provides for fines up to €35 million or 7% of global turnover for governance failures (Formind, 2024), making shadow AI one of the most expensive unmanaged risks in modern organizations.

"Shadow AI is not a technology problem, it's a governance gap that transforms legitimate productivity tools into compliance time bombs waiting to explode."

Detection Strategies: Building Your Shadow AI Inventory

Before you can govern shadow AI, you must first map it comprehensively. Here's a four-layer detection methodology combining technical and organizational approaches:

Layer 1: Network Traffic Analysis

Monitor outbound network connections to identify frequently accessed AI service domains:

# Common AI service domains to monitor
openai.com, api.openai.com, chat.openai.com
claude.ai, api.anthropic.com
bard.google.com, gemini.google.com
midjourney.com, discord.com/channels/midjourney
stability.ai, replicate.com
notion.ai, jasper.ai, copy.ai, writesonic.com
otter.ai, fireflies.ai, grain.com

Extract this data from your firewall logs, proxy servers, or security solutions (Palo Alto Networks, Fortinet, Zscaler). Usage spikes can reveal departments or individuals with heavy AI adoption.

Implementation approach:

// Example: Parsing firewall logs for AI domains
const aiDomains = [
  'openai.com', 'anthropic.com', 'midjourney.com',
  'claude.ai', 'notion.ai', 'jasper.ai'
];

function analyzeFirewallLogs(logs) {
  const aiTraffic = logs.filter(log => 
    aiDomains.some(domain => log.destination.includes(domain))
  );
  
  // Group by user and domain
  const userActivity = aiTraffic.reduce((acc, log) => {
    const key = `${log.user}-${log.destination}`;
    acc[key] = (acc[key] || 0) + 1;
    return acc;
  }, {});
  
  return userActivity;
}

Layer 2: Endpoint Detection and SaaS Discovery

Deploy Cloud Access Security Broker (CASB) solutions or endpoint detection tools to identify:

  • Browser extensions (ChatGPT for Chrome, Grammarly AI, Notion AI)
  • Desktop applications (Midjourney Discord client, Stable Diffusion installations)
  • API keys and tokens stored locally
  • OAuth connections to AI services

Tools like Netskope, McAfee MVISION Cloud, or Microsoft Defender for Cloud Apps can automatically discover shadow SaaS, including AI tools.

Layer 3: Financial and Procurement Analysis

Examine expense reports and individual SaaS subscriptions:

  • ChatGPT Plus, Claude Pro, Midjourney subscriptions ($20-60/month per user)
  • AI transcription services (Otter.ai, Fireflies.ai)
  • Content generation tools (Jasper, Writesonic, Copy.ai)
  • Code completion tools (GitHub Copilot, Tabnine, Codeium)

Many organizations discover they're paying for the same functionality multiple times through uncoordinated individual subscriptions. Consolidating these into enterprise agreements can save 30-50% while improving governance.

Layer 4: API Integration Audit

If you use no-code/low-code platforms (Make, Zapier, n8n, Airtable Automations), audit API connections established by users. At Keerok, we regularly find that well-intentioned automations integrate API calls to unvalidated external AI services.

Audit checklist:

  • Review all active integrations in automation platforms
  • Check API keys stored in environment variables or configuration files
  • Examine webhooks pointing to AI service endpoints
  • Validate data flows: what data is sent, where, and for what purpose

Building an Effective AI Usage Policy

Once shadow AI is mapped, the goal isn't prohibition but governance. An effective AI usage policy must be:

  • Clear and actionable: written in plain language, not legal jargon
  • Risk-proportionate: adapted to your organization's actual risk profile
  • Technology-agnostic: focused on principles, not specific tools (which evolve rapidly)
  • Operationally supported: backed by tools, training, and alternatives

Core Components of an AI Governance Policy

1. Data Classification and Permissible Uses

Define clearly what can and cannot be processed by external AI tools:

Data TypePublic AI ToolsPrivate AI ToolsExamples
Public data✅ Permitted✅ PermittedWebsite content, press releases, public research
Internal non-sensitive⚠️ With anonymization✅ PermittedDraft ideas, general concepts, brainstorming
Customer/personal data❌ Prohibited✅ With DPO approvalContact info, purchase history, user behavior
Confidential data❌ Prohibited✅ With authorizationStrategy, financials, source code, IP
Regulated data❌ Prohibited✅ With compliance reviewHealth data (HIPAA), payment data (PCI-DSS)

2. Approved Tools and Alternatives

Don't just prohibit – provide governed alternatives. For example:

  • For content generation: Deploy Azure OpenAI with data residency in your region and no training on your data
  • For document analysis: On-premise or private cloud solution with encryption (e.g., self-hosted LLMs via Ollama)
  • For image generation: Self-hosted Stable Diffusion or enterprise Midjourney account with IP protection
  • For code completion: GitHub Copilot Enterprise with IP indemnification

The Hellowork Group case illustrates this approach: facing informal AI tool usage causing data leakage risks, the company implemented a cybersecurity governance framework integrating proper AI usage, with charters, authorized tools, and awareness training.

3. Risk-Based Approval Process

Establish a clear workflow for employees to request new AI tools:

1. Request submission via internal form
   - Tool description and vendor
   - Business use case and expected benefits
   - Data types to be processed
   - Alternative solutions considered

2. Initial screening by AI governance committee
   - Business value assessment
   - Risk classification (minimal, limited, high, unacceptable)
   - Preliminary feasibility

3. Technical and security evaluation
   - Data residency and sovereignty
   - Encryption and access controls
   - Vendor security certifications (SOC 2, ISO 27001)
   - API security and rate limiting

4. Legal and compliance review
   - Terms of service analysis
   - Data processing agreement (DPA) negotiation
   - GDPR/AI Act compliance verification
   - IP ownership and training data usage

5. Decision and documentation
   - Approval, conditional approval, or rejection with alternatives
   - Documentation in AI system inventory
   - Communication to relevant teams

"Effective AI governance in an SME isn't an innovation brake, it's an accelerator: it enables teams to experiment safely, with clear guardrails and validated tools that protect both the business and its customers."

Implementing Operational AI Governance

AI governance isn't a PDF document in a shared folder. It requires organization, tools, and continuous monitoring.

Establishing a Cross-Functional AI Governance Committee

Even in a small organization, this committee should include complementary expertise:

  • Executive sponsor: Validates AI strategy, budget, and escalations
  • IT/Security: Evaluates technical feasibility, security, integration
  • Legal/Compliance: Ensures GDPR and AI Act compliance
  • Business representative: Represents operational needs (sales, marketing, product...)
  • Data Protection Officer (DPO): Mandatory if processing personal data (most AI use cases)
  • Risk management: Assesses and monitors AI-related risks

Meeting cadence: Monthly for active AI adoption, quarterly for maintenance phase. Responsibilities include validating new use cases, monitoring incidents, updating policies, and ensuring continuous compliance.

Creating an AI Sandbox for Safe Experimentation

Rather than prohibiting all experimentation, create a secure environment where teams can test AI tools:

  • Synthetic data: Anonymized or fictional datasets that mirror production structure
  • Network isolation: No connection to production systems or sensitive data stores
  • Usage monitoring: Logs and analytics to understand what's being tested
  • Time-boxing: Experiments have defined durations (e.g., 30-day pilots)
  • Graduation path: Clear criteria for moving from sandbox to production

This approach, increasingly adopted across organizations, enables innovation without compromising security, as highlighted by Blog du Modérateur's analysis of best practices.

Technical Governance Tools

1. Detection and Monitoring

  • CASB solutions: Netskope, McAfee MVISION Cloud, Microsoft Defender for Cloud Apps
  • Network monitoring: AI usage detection via traffic analysis (Darktrace, Vectra AI)
  • DLP (Data Loss Prevention): Alerts when sensitive data is sent to external AI services
  • API gateways: Centralized control and monitoring of AI API calls

2. Identity and Access Management

  • SSO (Single Sign-On): Centralized authentication to approved AI tools
  • License management: Track subscriptions, users, and costs
  • Granular access control: Role-based permissions for AI tool usage
  • MFA enforcement: Multi-factor authentication for AI tools accessing sensitive data

3. Documentation and Compliance

  • AI system registry: Inventory of use cases, purposes, data processed (AI Act requirement)
  • Impact assessments (DPIA): Risk analysis for high-risk AI use cases
  • Audit trails: Logs of AI-made decisions (transparency requirement)
  • Model cards: Documentation of AI model capabilities, limitations, and biases

Integration with Existing Security Policies

AI governance shouldn't exist in isolation. Integrate it with existing security processes:

  • Incident response: Add AI-specific procedures (data leakage via ChatGPT, discriminatory bias detected, model hallucination causing business impact)
  • Security reviews: Include AI tools in periodic audits and penetration testing
  • Security awareness training: Add module on AI risks and best practices
  • Business continuity: Plan for AI tool outages or vendor failures
  • Vendor risk management: Assess AI vendors using existing third-party risk frameworks

AI Act Compliance: Requirements for Organizations

The European AI Act, progressively entering into force since 2024, imposes specific obligations on organizations that develop or deploy AI systems. Even if you're a user (not a developer), you're affected as a "deployer."

AI System Risk Classification

The AI Act classifies AI systems into four risk categories:

  1. Unacceptable risk: Prohibited (cognitive manipulation, social scoring, real-time biometric identification in public spaces)
  2. High risk: Strict obligations (HR/recruitment, credit scoring, critical infrastructure, law enforcement)
  3. Limited risk: Transparency obligations (chatbots, deepfakes, emotion recognition)
  4. Minimal risk: No specific obligations (spam filters, product recommendations, inventory management)

For each AI use case in your organization, determine the risk level. A CV screening tool is high-risk, while an email writing assistant is minimal risk.

Documentation Requirements

For high-risk AI systems, the AI Act requires:

  • Technical documentation: System description, training data, performance metrics, limitations
  • Conformity assessment: Before production deployment
  • Continuous monitoring: Performance tracking and drift detection
  • Traceability: Logs enabling decision auditing
  • Transparency: Informing affected individuals they're interacting with AI
  • Human oversight: Mechanisms for human intervention in AI decisions
  • Accuracy and robustness: Testing and validation of AI performance

Even for limited-risk systems, transparency is mandatory. If you use a chatbot on your website, you must clearly inform visitors they're interacting with AI.

Responsibilities and Sanctions

As Formind reminds us, the AI Act provides for fines up to €35 million or 7% of global turnover for the most serious violations. SMEs are not exempt, though proportionality mechanisms are included.

Responsibilities are shared:

  • AI provider: Responsible for system compliance, documentation, and CE marking
  • Deployer (you): Responsible for compliant usage, respecting terms of service, monitoring, impact assessments, and transparency

With shadow AI, you're doubly exposed: not only might you be using a non-compliant system, but you're also not fulfilling your deployer obligations (no documentation, no monitoring, no transparency, no human oversight).

Training and Awareness: The Foundation of Sustainable Governance

The best policy in the world is useless if your team doesn't understand or apply it. Training is the cornerstone of successful AI governance.

Three-Tier Training Program

Tier 1: General Awareness (all employees)

  • What is AI and how does it work? (demystifying the technology)
  • What are shadow AI risks? (concrete examples and case studies)
  • What is our company's policy? (dos and don'ts)
  • Which tools are approved and how to use them?
  • What to do when in doubt? (escalation procedures)

Format: 30-minute e-learning, renewed annually, mandatory for all employees.

Tier 2: Role-Based Training (regular users)

  • AI use cases specific to role (marketing, sales, HR, development, customer service)
  • Prompt engineering best practices
  • Bias detection and mitigation
  • Result validation and fact-checking
  • Data handling and privacy considerations

Format: 2-3 hour workshops per department, with hands-on exercises using approved tools.

Tier 3: Advanced Training (AI committee, IT, DPO, risk managers)

  • Technical AI system evaluation
  • GDPR and AI Act compliance deep dive
  • AI incident management and response
  • Technology and regulatory monitoring
  • Vendor assessment and contract negotiation

Format: External training or certification (e.g., IAPP AI Governance Professional), updated semi-annually.

Continuous Communication and AI Culture

Beyond initial training, keep AI governance top-of-mind:

  • Monthly newsletter: New approved tools, use case spotlights, policy reminders
  • AI champions: Identify and empower advocates in each department
  • Success stories: Share wins (and lessons learned from failures)
  • Innovation pipeline: Transparent process for proposing new AI use cases
  • Office hours: Regular Q&A sessions with AI governance team

"Mature AI governance transforms shadow AI into governed innovation: employees become transformation agents, not policy violators to be policed."

Implementation Roadmap: From Shadow AI to Governed Innovation

Here's a practical 6-month roadmap for organizations with 50-500 employees looking to establish AI governance:

Month 1-2: Discovery and Assessment

  • Week 1-2: Technical audit (network logs) + anonymous employee survey
  • Week 3-4: Results analysis, shadow AI mapping, risk assessment
  • Week 5-6: Risk evaluation per use case (GDPR, AI Act, security, operational)
  • Week 7-8: Priority definition and AI governance committee formation

Month 3-4: Policy and Infrastructure

  • Week 9-10: Draft AI usage policy with concrete examples and decision trees
  • Week 11-12: Select and deploy approved AI tools (alternatives to shadow AI)
  • Week 13-14: Implement monitoring and control tools (CASB, DLP, API gateway)
  • Week 15-16: AI Act documentation (inventory, classification, DPIAs for high-risk systems)

Month 5-6: Rollout and Training

  • Week 17-18: Policy communication (town halls, documentation, FAQ)
  • Week 19-20: Tier 1 training rollout (all employees)
  • Week 21-22: Tier 2 training (role-based workshops)
  • Week 23-24: Review, adjustments, continuous monitoring planning

Ongoing: Continuous Improvement

  • Monthly AI governance committee meetings
  • Quarterly policy reviews and updates
  • Semi-annual training refreshers
  • Annual comprehensive audit

Budget Considerations

For a 200-person organization:

  • Consulting and implementation: $30-50k (assessment, policy development, committee training)
  • Technical tools: $10-30k/year (CASB, DLP, approved AI tool licenses)
  • Training: $10-20k (e-learning platform, workshops, certifications)
  • Internal time: 40-60 person-days (committee, IT, legal, HR)

Total first year: $50-100k, then $20-50k/year ongoing. A modest investment compared to potential AI Act fines ($35M or 7% of turnover), data breach costs (average $4.45M according to IBM), or reputational damage.

AI Governance as Strategic Advantage

Viewing AI governance solely through a risk lens would be a mistake. It's also a strategic opportunity to structure your digital transformation and create competitive advantage.

From Compliance Burden to Business Enabler

Organizations with mature AI governance can:

  • Innovate faster: Clear processes to test and deploy new AI use cases reduce time-to-value
  • Win customer trust: Transparency and certification differentiate in crowded markets
  • Attract talent: Top AI professionals seek organizations with mature practices
  • Optimize costs: Consolidated licenses, eliminated redundancies, negotiated enterprise agreements
  • Anticipate regulation: AI Act readiness before full enforcement avoids scrambling later
  • Enable partnerships: Governed AI unlocks opportunities with regulated industries (finance, healthcare, government)

AI Governance and Automation: The Winning Combination

At Keerok, we observe that organizations succeeding in AI transformation combine governance with automation. AI doesn't replace automation, it complements it:

  • Automation: Structured workflows, repetitive processes (e.g., data synchronization, report generation)
  • AI: Cognitive tasks, analysis, content generation, decision support (e.g., lead qualification, sentiment analysis)

A coherent AI implementation strategy integrates both dimensions, with governance covering the entire technology stack from traditional automation to cutting-edge AI.

Conclusion: Your Action Plan for AI Governance

Shadow AI isn't inevitable – it's a symptom that employees have legitimate needs the organization hasn't yet addressed. AI governance isn't a witch hunt, it's a framework that enables safe innovation.

Three priority actions to take this week:

  1. Launch a rapid assessment: Analyze network logs or deploy an anonymous survey to map current AI usage
  2. Form an AI governance committee: Identify 3-5 people (executive, IT, legal, business) to lead the initiative
  3. Draft your initial policy: Even a simple version establishes principles and demonstrates organizational commitment

Don't wait for the first data breach or regulatory notice to act. Organizations that structure AI governance now gain competitive advantage while protecting against regulatory risks.

Need help structuring your AI governance? Keerok helps organizations implement pragmatic, compliant AI usage policies. From initial assessment to team training, we transform shadow AI into governed innovation. Get in touch with our team for a complimentary governance assessment.

Tags

Shadow AI AI Governance AI Act Compliance Enterprise AI Security AI Usage Policy

Besoin d'aide sur ce sujet ?

Discutons de comment nous pouvons vous accompagner.

Discuss your project