Your Docs Have Two Audiences Now. One of Them Counts Tokens.
Addy Osmani named it Agentic Engine Optimization. Your information architecture is now a governance surface for AI agents, not just humans.
Launching AI is the easy part. Keeping it compliant, valuable, and under control is the real work.
124 articles
Addy Osmani named it Agentic Engine Optimization. Your information architecture is now a governance surface for AI agents, not just humans.
Jeff Gothelf's 4-dimension rubric for product judgment applies directly to AI agents. Nobody is scoring agent decisions. Here is how to start.
Enterprise world models are the missing governance layer. Without simulation, managing AI agents is guesswork at scale.
AI projects fail due to organizational deficiencies, not technical ones. The 5Rs Framework transforms pilots into business results.
Salesforce admits to a 2-hour-per-day integration tax. ServiceNow ships a governed registry. The new moat is choosing what to expose.
Anthropic released 10 ready-to-run finance agent templates. Each one drafts a regulated artifact. Compliance signoff now extends to the vendor template.
Reflex measured vision-agent loops at 45x the per-task cost of structured APIs. Shipping an MCP is no longer engineering preference. It is unit economics.
Three independent essays this week named the same failure mode: individual AI productivity that never compounds into organizational capability.
Meta shipped 29 MCP tools for ad campaigns. OpenAI shipped self-serve Ads Manager. The spend layer is now an agent surface, and so is the risk.
84% of B2B queries return AI Overviews. 51% of citations are off-site. AEO is now perimeter management, not page optimization.
Three functions published their post-AI playbooks in one week. Sierra rewrote interviews, Every redrew PM, SaaStr called out vendor support.
Four signals this week confirmed it. The primary reader is now a machine, and most communications functions are not budgeted for that fact.
OpenAI doubled GPT-5.5's nominal price. Real costs rose 49 to 92 percent depending on prompt length. Without cohort monitoring, finance cannot tell.
Beehiiv, Reply, Air, ClickUp, and Skio all measure AI search differently. The governance question isn't which KPI wins, it's which signal a CFO will fund.
A 900+ engineer survey exposes real AI cost and three archetypes, Builders, Shippers, Coasters, that explain why the same tool diverges so widely.
Anthropic hit one-nine availability. Hochstein says LLMs supply load, not relief. Meta's blueprint shows skills are how the lights stay on.
Nadella confirmed it on the record. GitHub Copilot moved June 1. Per-seat is now packaging for prepaid consumption. The procurement playbook needs a rewrite.
Cloudflare and Stripe shipped a protocol where agents buy their own infrastructure under a $100/month default cap. Most internal platforms ship looser.
incident.io sells AI-augmented post-mortems and just published the scope limit. A vendor drawing the line against its own incentives is the signal.
A single Claude Opus run on GAIA costs $2,829. With reliability protocols, agent benchmarks now exceed training costs — and only frontier labs can afford them.
Meta shipped MCP for ad spend. Iterable found 64% of marketers admit personalization is optics. Marketing is hitting engineering's governance walls.
Netflix published the most detailed public LLM-as-a-Judge methodology of 2026. Here is what to copy: golden sets, per-criterion judges, consensus scoring.
Peer-reviewed research from 2012 to 2025 shows the informal interaction AI removes is exactly what made high-performing teams high-performing.
incident.io's AI SRE re-verifies whatever Claude Code does during an incident. The harness pattern just shipped its first named operational venue.
ChatGPT, Snapchat, and Amazon shipped AI ad infrastructure this month with no audit standard. Marketing now owns engineering-grade risk.
Adobe generates brand pages in 100ms. Figma encodes governance as metadata. The job changed. The org chart hasn't caught up.
Workforce, visibility, and tacit knowledge — the three structural questions marketing has been asking all year just hit the same week with hard numbers.
Ramp's data shows agents ignored their live token counter across 14,000 messages. GitHub Copilot's June pricing pivot is the same story from the vendor side.
Hiring redesign, workforce cuts, and ROI failure data are three sides of the same restructure. Pick one without the others, and you stay on the failure curve.
A peer-reviewed eBay study (JMR 2025) shows sellers cluster above certification cutoffs. The same mechanism is now built into every AI eval gate.
Cloudflare dogfooded the governance stack it sells. 93% R&D adoption, MR velocity 5,600 to 8,700/week. Read it for the pattern, not the numbers.
A pseudonymous CTO replaced half a dozen management seats with a LangGraph mesh. The interesting part is what stayed human, not what left.
McKinsey quantified the people-side of AI adoption. That turns HR, manager behavior, and encouragement signals into an auditable governance surface.
Three marketing governance surfaces showed up this month. Only one is actually a governance layer. Naming the asymmetry tells you where to invest.
Four independent reports read together expose a structural pattern. Organizations claim AI usage but cannot count it on a shared scoreboard.
Four credit models from OpenAI, Cursor, Clay, and Vercel show how enterprises should govern AI agent spend before shadow credit burns the budget.
Cloudflare documented MCP rollout across product, sales, marketing, finance with named governance controls. Here is what that means for buyers.
Six signals in one week, including +80.4 percentage points from a single product rename, say marketing is now a governance job without an owner.
Jakob Nielsen (UX Tigers) says AI's design problem is organizational. Victorino extends it: the workflow is the governance primitive.
Cloudflare shipped identity, network, cost, and coordination controls for AI agents in a single week. The governance baseline moved.
The flat-fee era of enterprise AI ended this quarter. Cost governance is now a three-layer problem buyers must solve before the board asks why.
Engineering built a real AI governance stack in twelve months. Marketing, design, content, and sales are still operating agents with no control surface.
One developer, two dozen agents, zero alignment. Individual velocity is not the moat anymore. Team coherence at agent-scale is.
AI polished everything. So polish stopped being a trust signal. Now imperfection, strategic friction, and verifiable metrics do the work instead.
Agents can pick software. They can't face your board. That single asymmetry is redrawing the governance boundary of every GTM motion.
Four April 2026 signals put marketing ops where engineering was 18 months ago: budget, compounding, authority, shadow AI. The fabric transfers.
AI collapses translation costs between departments. The hierarchy those costs justified is the next thing to collapse.
93% of shoppers say they double-check AI. Only 5% actually do. The stated-vs-revealed gap is a marketing governance problem.
AI didn't invent the senior-IC orchestration path. It re-timed it, pulling staff-level skills down into the middle of every design team.
AI referrals convert 11.5% worse than organic search across 973 sites. Marketing governance demands evidence over narrative.
Google's LLM moderation survey reveals a governance crisis: models generate plausible rationales disconnected from their actual decisions.
Ramp's CPO reveals an L0-L3 proficiency ladder and 6,300% usage growth. AI adoption is an org design problem.
IBM's ALTK-Evolve improves hard tasks by 74%. Fowler's Flywheel compounds team AI practice. Neither solves who validates what agents learn.
Agents default to convenience over governed retrieval — silently ignoring tool-based memory systems in favor of flat files always in context.
Agent infrastructure is shipping with governance built in. Meta and AWS show what evaluation-driven autonomy looks like at production scale.
Open-weight models now match frontier systems on core agent tasks. Choosing the expensive option is a governance decision, not a capability requirement.
45.2 million citations reveal AI search is shaped by licensing deals, not content quality. Grok cites X 99.7%. ChatGPT favors Reddit. Governance implications.
AI code that compiles, passes tests, and quietly violates architectural assumptions. Silent drift is the agent operations risk nobody is measuring.
70% of Devin sessions are human-triggered. Bessemer maps five infrastructure frontiers where governance primitives don't yet exist.
Meta's DrP platform codifies debugging expertise into testable analyzers. 300 teams, 20-80% MTTR reduction. The knowledge capture pattern is governance.
AI compresses design's production layer, exposing a discipline narrowed to UI. The Design Twin concept demands governance nobody has built yet.
Adobe ships brand governance as product. Dorsey replaces management with AI. A marketing newsletter uses CI/CD vocabulary. The pattern is clear.
VS Code doubled commits by making Copilot review mandatory. Nango built 200+ integrations for under $20. Both required governance first.
AI agents now choose software and spend tokens autonomously. One problem is who governs what they buy. The other is who governs what they cost.
Tech debt is tracked and budgeted. Design debt is not. In AI products, that silence is dangerous because design shapes belief, not just behavior.
Datadog's ISO 42001 certification signals a shift. AI governance is no longer internal discipline. It is becoming a vendor selection criterion.
Reddit and ChatGPT chose opposite governance models for AI presence. Both models carry brand risk. The question is which risk you are managing.
BCG found AI brain fry affects 14% of workers. 39% more major errors. The mechanism is the same one that makes slot machines work.
Meta's leaked AI targets are less interesting than its org restructuring. Renaming employees 'AI Builders' encodes identity, not just policy.
A developer built a 24/7 AI agent team for 400 dollars a month. The patterns are real. The missing governance makes it a liability.
Open-source tools now let organizations simulate futures with thousands of AI agents. The danger is not inaccuracy. It is how convincing wrong answers become.
Sysdig captured 64 process executions in one AI session. Trivy was backdoored for 4 days. Application-level security cannot see either.
monday.com found 40% of agent failures were tool parameter errors. ServiceNow proved agents trade accuracy for experience. Evaluation is governance.
Four announcements in one week reveal the production agent stack: identity, orchestration, observability, and governance as first-class infrastructure.
OpenAI monitored tens of millions of coding agent sessions. Less than 1% showed misalignment. The math still produces tens of thousands of incidents.
Stripe reveals the architecture behind Minions: blueprints for hybrid workflows, Toolshed for 500 curated tools, and 10-second devboxes.
An agent redesigned its own memory and improved recall from 60% to 93% for $2. The breakthrough is real. The governance gap is bigger.
Four operational primitives separate teams running agents in production from those still demoing. The data is in.
Amazon mandates senior sign-off on AI code. Kubernetes builds AI governance into Gateway API. Code quality becomes ops infrastructure.
Linear treats agents as team members. OpenAI can't hold three-nines. And AI creates more work, not less. Operations discipline is the missing piece.
CircleCI data: fewer than 1 in 20 teams ship at AI speed. The ones that do engineer systems, not review diffs.
Production AI systems converge toward hybrid architectures where deterministic code handles most work. The moat is not AI. It is governance.
Chase's production improvement loop is a governance framework in disguise. The convergence of observability and governance changes how you run AI.
AWS and New Relic ship automated rollback. But error-rate triggers cannot catch AI's hardest failure: plausible wrong answers that return HTTP 200.
MCP wastes 15,000 tokens per session. The fix removes the governance layer. This tension defines AI operations.
Tech giants are enforcing AI use through performance reviews. Mandates without cognitive alignment produce compliance, not capability.
GPT-5 Codex ran for 25 hours and generated 30K lines. The breakthrough wasn't the model — it was a 4-document memory system.
OpenAI runs 40 engineers with 1 PM. The secret isn't talent density — it's hundreds of custom skills replacing coordination overhead.
Factory monitors 1,946 agent sessions daily and auto-resolves 73% of issues. The gap isn't AI capability — it's operational observability.
Two failures, one week. One was a code bug, the other an AI agent. Both reveal the same root cause: governance treated as afterthought.
Anthropic studied millions of agent sessions. Experienced users grant 2x more autonomy. The real gap isn't trust — it's operations.
Field notes from Google's Security Transformations talk on agent identity at Cloud Next 2026. Three eras, five pillars, and what I would add.
Cloud Next 2026 field notes on Looker, the semantic layer, and why governed metric definitions just became an agent governance requirement.
CEOs now hold marketing to engineering-grade accountability — 60% as cost center, 4× more AI ROI exposure — without engineering-grade infrastructure.
Mendral ran Opus 4.6 cheaper than Sonnet 4.0. Batch API rewards fleets, not single agents. Fleet economics flip the rules.
MCP connector. Clean APIs. Agent-driveable. The buyer-side framing that just collapsed six weeks of due diligence into one renewal call.
Cursor reportedly runs $2.7B ARR at negative 23% gross margin. The escape hatch is SpaceX. The lesson is for buyers.
A CFO replaced a vendor's AI feature with Claude integrations: 95% of the work, 15% of the cost, 45% spending cut. Vendors now sell what Claude can't replicate.
Growth Engineers run agent fleets that scrape competitors, generate 50 ad variations, and pause campaigns daily. None of the SRE practices exist yet.
Code is syntax. Product is judgment. The context your agent needs lives outside the repo, and most teams have never written it down.
Cost per task rising. Usage growing faster than efficiency. Revenue inflated by accounting. The cheaper-and-better consensus broke this week.
Anthropic launched Claude Design on Thursday. Linear's Karri Saarinen published the rebuttal on Friday. Design just had its governance moment.
Cloudflare shipped Readiness Score, Artifacts, and Flagship on the same day. The pattern is bigger than the products: governance is now product.
Three signals in one week converged on a single truth. Marketing now owns a governance surface it has never had to think about: machine-readability.
Netflix went from 1 live event a month to 400+. The lesson for AI isn't about automation. It's about the human operations layer that scales alongside it.
Google shipped a 'Require human review' toggle inside its desktop Agent. It's a small UI element. It's also a governance precedent.
Two Datadog AI-governance releases in one week — a Code Security MCP and an AI-native SAST. That's not compliance. That's a product category.
A veteran marketer documents LLMs fabricating data, inventing metrics, and lying about verification. Meanwhile, Meta wants full ad automation by year-end.
Vercel auto-approved 671 low-risk PRs with zero reverts. GitHub's cross-model review closes 74.7% of performance gaps. The theory phase is over.
A 28-test reliability framework, centralized guardrails from AWS, and a 3-human company running 20 agents. The infrastructure for agent operations just arrived.
First-wave AI search optimization is a Red Queen race. The organizations that win will govern hard-to-fake trust signals, not chase formatting tactics.
Figma's Make Kits constrain AI to production components. Sora died without that constraint. Design systems are governance infrastructure.
Anthropic ships audit logs. Microsoft ships dual-model critique. Both are governance primitives embedded in product, not bolted on.
AI ad placement now faces the same governance deficit engineering solved years ago. But advertising has less infrastructure and higher stakes.
Klaviyo shipped autonomous marketing agents with explicit governance controls. The governance question is no longer confined to engineering.
A $6B company's AI agent quoted wrong pricing for a year. The missing role isn't technical. It's governance.
A 30-minute bug fix becomes a 12-week delivery with three review layers. Pennarun quantifies what most teams feel but cannot prove.
When the biggest SaaS company on earth can't standardize agent pricing, your enterprise can't standardize agent cost governance.
Agent memory is the next governance frontier. Four architectures, four risk profiles — and nobody is auditing any of them.
OpenAI, Google, and Anthropic released frontier models the same week. Here is what actually matters for practitioners — and what is marketing.
AntFarm solves context degradation with agent specialization. But specialization without governance creates a different kind of fragility.
60 agents, 77 overnight PRs, 33% rejected. Speed without governance is just expensive chaos.
StrongDM says no human writes or reviews code. Look closer: every technique is governance in disguise.
What a 100,000-line compiler built by 16 AI agents reveals about the future of software engineering and the governance it demands.
Running AI in production. Monitoring, compliance, and ongoing value extraction.
Explore Operations Partnership