This recipe deploys organizational "white blood cells" — context-aware AI agents that live inside communication channels (Slack, email, cloud platforms) and provide real-time compliance nudges without interrupting workflow. The system uses elastic reasoning to run lightweight pattern-matching on routine messages while escalating genuine risks to full LLM analysis, implementing the bumper-rail intervention model where employees receive gentle course-corrections rather than hard blocks. [src1]
consulting/oia/white-blood-cell-architecture/2026 — White Blood Cell ArchitectureWhich deployment path?
├── Primary platform is Slack AND budget includes LLM API
│ └── PATH A: Slack Bolt + Claude API — full elastic reasoning
├── Primary platform is Microsoft Teams AND budget includes LLM API
│ └── PATH B: Bot Framework + Azure OpenAI — Teams-native integration
├── Primary platform is Slack AND budget is minimal
│ └── PATH C: Slack Bolt + regex-only — no LLM escalation
└── Multiple platforms AND enterprise budget
└── PATH D: Unified middleware (n8n/Zapier) + Claude API
| Path | Tools | Cost | Speed | Output Quality |
|---|---|---|---|---|
| A: Slack + Claude | Slack Bolt, Claude API, Redis | $200-500/mo | 2-3 days | High — full context-aware nudges |
| B: Teams + Azure | Bot Framework, Azure OpenAI, Cosmos DB | $300-800/mo | 3-5 days | High — native Teams integration |
| C: Slack + Regex | Slack Bolt, regex patterns | $0-50/mo | 1-2 days | Medium — catches patterns, misses context |
| D: Multi-platform | n8n, Claude API, Redis, webhooks | $500-1500/mo | 5-7 days | Highest — unified cross-channel |
Duration: 4-8 hours · Tool: Structured document + JSON schema
Convert the client's compliance policies into a three-tier detection ruleset: Tier 1 (regex patterns for PII, forbidden keywords), Tier 2 (heuristic scoring for anomalies), Tier 3 (LLM analysis for ambiguous violations).
{
"rulebase_version": "1.0",
"compliance_domain": "sales_compliance",
"tier1_patterns": [
{"pattern": "\\b(guarantee|promise)\\b.*\\b(feature|capability)\\b",
"severity": "medium", "nudge_template": "feature_promise_redirect"}
],
"tier2_heuristics": [
{"signal": "message_to_external_with_attachment", "threshold": 0.7,
"escalate_to": "tier3"}
]
}
Verify: Rulebase covers all compliance domains. Cross-reference each policy clause with at least one detection rule. · If failed: Schedule working session with client compliance officer.
Duration: 2-4 hours · Tool: Slack Bolt SDK [src4]
Create Slack application with event subscriptions (message.channels, message.groups, message.im, file_shared). Implement the three-tier elastic reasoning pipeline where 95%+ of messages pass through with zero content inspection.
Verify: Bot appears online, responds to test messages in sandbox channel. · If failed: Check OAuth scopes — requires channels:history, chat:write, im:history.
Duration: 3-6 hours · Tool: Microsoft Purview DLP or Google DLP API [src3]
Set up transport rules that forward flagged emails to the WBC system via webhook. Configure DLP policies for PII detection, sensitive data patterns, and custom compliance rules.
Verify: Send test email with dummy SSN pattern. Webhook fires within 30 seconds. · If failed: Verify Purview license level (E5 required for custom webhooks).
Duration: 4-8 hours · Tool: JSON template collection + legal review
Create graduated nudge templates with four escalation levels per compliance domain: whisper (ephemeral), coach (DM), flag (DM + escalation), block (message held). Each level includes cooldown periods to prevent nudge fatigue. [src5]
Verify: Legal team has approved all Level 3-4 templates. · If failed: Deploy Level 1-2 only, flag Level 3-4 for legal sprint.
Duration: 2-4 hours · Tool: Configuration file + monitoring dashboard
Define the autonomy escalation matrix mapping severity levels to intervention tiers. Critical severity triggers immediate Level 4 block + human escalation. Low severity never auto-escalates beyond Level 2.
Verify: Run 20 simulated scenarios across all severity levels. Confirm escalation paths match matrix. · If failed: Review scoring thresholds in Tier 2 heuristics.
Duration: 2 weeks · Tool: Monitoring dashboard + weekly review cadence
Deploy in shadow mode first (5 business days). Calibrate thresholds on real data. Switch to live mode for Level 1-2 nudges. Enable Level 3-4 after 5 more days if false positive rate remains <10%. [src2]
Verify: Dashboard shows <10% false positive rate, <5% Tier 3 escalation rate. · If failed: If false positive >20%, tighten Tier 1 regex patterns.
{
"output_type": "wbc_deployment",
"format": "running_system",
"components": [
{"name": "slack_bot", "type": "deployed_application", "required": true},
{"name": "email_monitor", "type": "configured_policy", "required": true},
{"name": "rulebase", "type": "JSON", "required": true},
{"name": "nudge_templates", "type": "JSON", "required": true},
{"name": "autonomy_matrix", "type": "configuration", "required": true},
{"name": "monitoring_dashboard", "type": "web_dashboard", "required": true}
],
"success_criteria": {
"false_positive_rate": "< 10%",
"tier3_escalation_rate": "< 5%",
"mean_nudge_delivery_time": "< 3 seconds"
}
}
| Quality Metric | Minimum Acceptable | Good | Excellent |
|---|---|---|---|
| False positive rate | < 15% | < 10% | < 5% |
| Tier 3 escalation rate | < 10% | < 5% | < 2% |
| Nudge delivery latency | < 10 seconds | < 3 seconds | < 1 second |
| User complaint rate (per 100 users/week) | < 5 | < 2 | 0 |
| Policy coverage | > 70% | > 85% | > 95% |
If below minimum: Halt live nudges and return to shadow mode. Retrain Tier 1 patterns with compliance team input.
| Error | Likely Cause | Recovery Action |
|---|---|---|
| Slack Events API connection dropped | Socket mode timeout or Slack outage | Reconnect with exponential backoff (max 5 retries, 30s intervals) |
| LLM API rate limit (429) | Too many Tier 3 escalations in burst | Queue in Redis with rate limiter. Raise heuristic thresholds. |
| DLP webhook delivery failure | Network issue or endpoint down | Configure Purview retry policy (3 retries, 5-min intervals) |
| Nudge delivered to wrong user | Message threading context lost | Validate message.user field before nudge delivery |
| Privacy complaint from employee | Monitoring perceived as surveillance | Pause monitoring for that user immediately. Escalate to legal. |
| Component | Free Tier | Paid Tier | At Scale (1000+ users) |
|---|---|---|---|
| Slack Bolt SDK | $0 | $0 | $0 |
| LLM API (Claude) | ~$5/mo | $50-200/mo | $200-800/mo |
| Microsoft Purview DLP | N/A | $35/user/mo | $35/user/mo |
| Hosting (bot server) | $0 (Vercel free) | $20/mo | $50-100/mo |
| Redis (queue/cache) | $0 (Upstash free) | $10/mo | $50/mo |
| Total (excl. platform licenses) | ~$5/mo | $80-230/mo | $300-950/mo |
Organizations that monitor every word create the exact "security fatigue" that NIST documented — employees stop trusting the system and actively circumvent it by moving to personal channels. [src1]
The system should behave like a cat napping in a window: burning almost zero energy until a genuine threat appears. 95%+ of messages pass through with zero content inspection.
Excessive nudging causes "nudge fatigue" — people start ignoring all nudges, including critical ones. Identical to alarm fatigue in healthcare settings. [src5]
A single well-timed whisper is more effective than ten rapid-fire alerts. The graduated autonomy matrix prevents nudge storms.
Deploying live nudges without calibration data guarantees high false positive rates, which destroys user trust irreversibly. [src2]
The knowing-doing gap applies to WBC deployment itself — knowing the system works in theory and calibrating it for a specific culture are entirely different challenges.
Use when the agent needs to actually deploy an organizational compliance monitoring system using AI-powered ambient agents inside communication channels. This recipe produces a running system, not a strategy document. Requires the white blood cell architecture framework as theoretical foundation and client compliance policies as input.