Case study
Building a risk classification engine
Driving data governance and semantic analysis to improve AI model outputs
Designed and implemented an AI-driven classification system that analyzes internal content standards documentation and assigns priority levels (critical, high, medium, low) to individual sections. This system enables downstream AI writing tools to distinguish between rigid requirements and flexible guidance, improving both compliance and output quality.
At a glanceGoal
Enable AI-generated content to adhere to content standards with appropriate levels of strictness, balancing compliance with creative flexibility.
Challenges
- Ambiguity in content standards language and structure
- Lack of labeled training data
- Pressure to move faster
- Need to balance AI system needs and human user needs
Solution
Create a way to take a disparate set of standards and build in risk assessment that can be readable by both AI and humans.
Impact
- Reinforced regulatory requirements for non-writers
- Increased generation flexibility by allowing non-critical guidance to be applied more adaptively
- Established a foundation for AI governance by operationalizing policy into machine-readable logic
Defined the priority framework
Established clear criteria for critical, high, medium and low classifications
Developed a semantic classification approach
Built system based on natural language processing to interpret meaning and intent in standards
Created a training and evaluation set
Generated labeled examples to bootstrap model performance, focusing on regulatory required language
Iterated on model collaboration
Tuned classification thresholds to strike the right balance between strict compliance and generative flexibility
Designed structured output for integration
Standardized outputs into a format consumable by AI writing systems, enabling dynamic enforcement of content rules during generation
Delivered a scalable classification system that transforms static content standards into actionable inputs for AI systems, enabling more reliable and context-aware content generation.
LLM Content Risk Assessment Framework
Use this framework to evaluate guidelines from content standards and style guides, and assign each a risk tier based on its potential impact if violated by an LLM model.
- Violation creates legal liability or regulatory penalty
- Noncompliance could cause direct harm to users or third parties
- Required by law, regulation, or binding contract (GDPR, HIPAA, FTC, etc.)
- Breach could trigger a product recall, ban, or enforcement action
- No acceptable workaround or contextual exception exists
- Violation would meaningfully damage brand reputation or user trust
- Inconsistency here undermines the product's core value proposition
- Required by an internal style guide with organizational authority
- Closely tied to audience expectations for tone, voice, or format
- Exceptions exist but must be deliberately approved, not accidental
- Violation reduces output quality but does not cause user harm
- Inconsistency creates friction but not a fundamental trust breakdown
- Guideline reflects best practice or house style preference
- Reasonable exceptions exist based on context or user request
- Deviation is noticeable to an informed reviewer, not a casual user
- Violation has no measurable impact on trust or user outcome
- Guideline reflects personal, team, or regional preference
- Inconsistency is unlikely to be noticed by most users
- Context-dependent — correct behavior varies by use case
- Would only be flagged in a detailed editorial review, not production QA