Content Standards Protocol
The Content Standards Protocol enables privacy-preserving brand suitability for ephemeral and sensitive content that cannot leave a publisher’s infrastructure.The Problem
Traditional brand suitability relies on third-party verification: send your content to IAS or DoubleVerify, they evaluate it, return a verdict. This works for static web pages. It fundamentally cannot work for:- AI-generated content - ChatGPT responses, DALL-E images that exist only in a user session
- Private conversations - Content in messaging apps, private social feeds
- Ephemeral content - Stories, live streams, real-time feeds that disappear
- Privacy-regulated content - GDPR-protected data that cannot be exported
The Solution: Calibration-Based Alignment
Content Standards solves this by using agents to protect privacy. It’s a three-phase model where no sensitive content ever leaves the publisher’s infrastructure:| Phase | Where It Runs | What Happens |
|---|---|---|
| 1. Calibration | External (safe data only) | Publisher and verification agent align on policy interpretation using synthetic examples or public samples - no PII, no sensitive content |
| 2. Local Execution | Inside publisher’s walls | Publisher runs evaluation on every impression using a local model trained during calibration - content never leaves |
| 3. Validation | Statistical sampling | Verification agent audits a sample to detect drift - both parties can verify the system is working without exposing PII |
What It Covers
- Brand safety - Is this content safe for any brand? (universal thresholds like hate speech, illegal content)
- Brand suitability - Is this content appropriate for my brand? (brand-specific preferences and tone)
Key Concepts
Content standards evaluation involves four key questions that buyers and sellers negotiate:- What content? - What artifacts to evaluate (the ad-adjacent content)
- How much adjacency? - How many artifacts around the ad slot to consider
- What sampling rate? - What percentage of traffic to evaluate
- How to calibrate? - How to align on policy interpretation before runtime
Workflow
Key insight: Runtime decisioning happens locally at the seller (for scale). Buyers pull content samples from sellers and validate against the verification agent.Adjacency
How much content around the ad slot should be evaluated?| Context | Adjacency Examples |
|---|---|
| News article | The article where the ad appears |
| Social feed | 1-2 posts above and below the ad slot |
| Podcast | The segment before and after the ad break |
| CTV | 1-2 scenes before and after the ad pod |
| Infinite scroll | Posts within the visible viewport |
get_products). The buyer can filter products based on adjacency guarantees:
Adjacency Units
| Unit | Use Case |
|---|---|
posts | Social feeds, forums, comment threads |
scenes | CTV, streaming video content |
segments | Podcasts, audio content |
seconds | Time-based adjacency in video/audio |
viewports | Infinite scroll contexts |
articles | News sites, content aggregators |
Sampling Rate
What percentage of traffic should be evaluated by the verification agent?| Rate | Use Case |
|---|---|
| 100% | Premium brand suitability - every impression validated |
| 10-25% | Standard monitoring - statistical confidence |
| 1-5% | Spot checking - drift detection only |
Validation Thresholds
When a seller calibrates their local model against a verification agent, there’s an expected drift - the local model won’t match the verification agent 100% of the time. Validation thresholds define acceptable drift between local execution and validation samples. Sellers advertise their content safety capabilities in their product catalog:| Threshold | Meaning |
|---|---|
| 0.99 | Premium - local model is 99% aligned with verification agent |
| 0.95 | Standard - local model is 95% aligned |
| 0.90 | Budget - local model is 90% aligned |
Policies
Content Standards uses natural language prompts rather than rigid keyword lists:Scoped Standards
Buyers typically maintain multiple standards configurations for different contexts - UK TV campaigns have different regulations than US display, and children’s brands need stricter safety than adult beverages.Code Format ConventionsCountry and language codes are case-insensitive - implementations must normalize before comparison. Recommended formats follow ISO standards:
- Countries: Uppercase ISO 3166-1 alpha-2 (e.g.,
GB,US,DE) - Languages: Lowercase ISO 639-1 or BCP 47 (e.g.,
en,de,fr)
standards_id when creating a media buy. The seller receives a reference to the resolved standards - they don’t need to do scope matching themselves.
Calibration
Before running campaigns, sellers calibrate their local models against the verification agent. This is a dialogue-based process that may involve human review on either side:- Seller sends sample artifacts to the verification agent
- Verification agent returns verdicts with detailed explanations
- Seller asks follow-up questions about edge cases
- Process repeats until alignment is achieved
Tasks
Discovery
| Task | Description |
|---|---|
| list_content_standards | List available standards configurations |
| get_content_standards | Retrieve a specific standards configuration |
Management
| Task | Description |
|---|---|
| create_content_standards | Create a new standards configuration |
| update_content_standards | Update an existing standards configuration |
| delete_content_standards | Delete a standards configuration |
Calibration & Validation
| Task | Description |
|---|---|
| calibrate_content | Collaborative dialogue to align on policy interpretation |
| get_media_buy_artifacts | Retrieve content artifacts from a media buy |
| validate_content_delivery | Batch validation of content artifacts |
Typical Providers
- IAS - Integral Ad Science
- DoubleVerify - Brand suitability and verification
- Scope3 - Sustainability-focused brand suitability with prompt-based policies
- Custom - Brand-specific implementations
Future: Secure Enclaves
The current model trusts the publisher to faithfully implement the calibrated standards. A future evolution uses secure enclaves (Trusted Execution Environments / TEEs) to provide cryptographic guarantees: Content never crosses the pinhole - only models flow in, only aggregates flow out.The Pinhole Interface
The enclave maintains a narrow, well-defined interface to the verification service: Inbound (verification service → enclave):- Updated brand suitability models
- Policy changes and calibration exemplars
- Configuration updates
- Aggregated validation results (pass rates, drift metrics)
- Statistical summaries
- Attestation proofs
- Raw content artifacts
- User data or PII
- Individual impression-level data
Why This Matters
- Publisher hosts a secure enclave inside their infrastructure
- Governance agent (from IAS, DoubleVerify, etc.) runs as a container within the enclave
- Content flows into the enclave for evaluation but never leaves the publisher’s walls
- Both parties can verify the governance code is running unmodified via attestation
- Models stay current - the enclave can receive updates without exposing content
Related
- Artifacts - What artifacts are and how to structure them
- Brand Manifest - Static brand identity that can link to standards agents