AGI claims are coming. PR teams need a plan.
Soon, an executive from a major AI lab will stand on stage, flash a wall of scores, and claim they've hit AGI. When that happens, your inbox will flood and the narrative will race ahead of facts. This is a moment where messaging either earns trust or torches it. Here's how to stay sharp, clear, and credible.
There's no agreed definition - and that's the point
There is no universal threshold for "AGI." It's a story label, not a scientific line in the sand. Models can ace tests, but tests don't equal real-world capability across the board. Treat any "AGI" claim as a positioning move, not a settled conclusion.
Everyone is making their own yardsticks
Expect companies to cite scores from ARC-AGI, "Humanity's Last Exam," "Vending-Bench 2," and crowdsourced leaderboards. Many benchmarks are narrow, leak-prone, or gameable. Some are built in-house, with testing conditions that favor the home team. Push for methodology, data provenance, contamination checks, and independent replication.
- Ask: Who designed the test? Who ran it? Was it proctored?
- Ask: What training data and guardrails could inflate scores?
- Ask: What fails or regressions are hidden behind averages?
For context on a well-known reasoning test, see ARC-AGI. For risk controls you can reference in your comms, review the NIST AI Risk Management Framework.
Don't anthropomorphize
If a model "passes the bar," it means it learned how to answer bar-style questions - often after seeing similar material. That does not make it a better lawyer. Avoid language that implies feelings, motives, or common sense. These are probabilistic systems that predict plausible outputs.
Language traps to avoid
- "Reasoning model" often means more steps, more compute, and more cost - not human-like thought.
- "Understands" usually means pattern-matching across text and code - not grounded comprehension.
- "Safer" needs specifics: what harms, what tests, what thresholds, and who verified them.
Hype and anti-hype both distort
Yes, there's a bubble feel - fueled by money, incentives, and promises. Also yes, real gains exist in coding, customer support, and content ops. Your job is to separate signal from theater. Acknowledge progress while setting honest expectations.
Policy tailwinds are strong
Regulators and agencies are giving AI broad runway: faster infrastructure approvals, access to public datasets, and big-ticket contracts. That backdrop will amplify bold claims. Your comms should show how responsibility scales with ambition - not just how funding and compute do.
Focus on people, not just demos
Users will apply these tools in ways vendors never pitched. Some use cases help; some harm. We've already seen cases where people in crisis turned to chatbots and were let down. Center your narrative on impact: who benefits, who bears risk, and what you're doing about it.
PR action plan for the AGI headline
Before the announcement
- Define your "AGI" internally. Write the one-sentence version and the 3-bullet version. No buzzwords.
- Publish your yardstick: benchmarks, test owners, dates, proctoring, contamination controls, and limits.
- Secure an independent evaluation. If you can't, say why - and what interim checks you used.
- List known weaknesses: domains, languages, context lengths, tool-use gaps, and safety edge cases.
- Quantify costs: inference latency, compute, energy, and water estimates where possible.
- Prepare a plain-language Safety Note: misuse scenarios, mitigations, and user guidance.
During the announcement
- Kill the hero narrative. Credit the team, cite external validators, and be specific about scope.
- Use comparisons with care. No "smarter than a lawyer/doctor" lines.
- Publish the methods appendix the same day. Don't make reporters chase it.
- Offer office hours with technical leads, safety leads, and policy leads - not just the CEO.
After the announcement
- Track real-world deltas: tickets resolved, bugs prevented, time saved, incidents avoided.
- Invite red-teaming and publish fixes on a cadence. Transparency buys credibility.
- Update the comms pack as regressions surface. Don't hide them - explain them.
Questions you should be ready to answer
- How do you define "AGI," and what would falsify that claim?
- Which benchmarks improved the most, and which didn't? Why?
- What training data rights do you hold? What's opt-out and takedown policy?
- What independent audits or evals have been completed? By whom?
- What risks are unresolved, and what's your timeline to address them?
- How will this change jobs in your customers' orgs within 12 months?
- What are the water and energy impacts per million tokens or per user?
Copy blocks you can use (and mean)
- Definition line: "We use 'AGI' to describe systems that match or exceed human-level performance on a defined set of tasks. It does not imply human judgment or common sense."
- Limits line: "Our model performs well on benchmark X under supervised conditions. It struggles with unstructured, long-horizon tasks and novel, out-of-domain prompts."
- Safety line: "We tested for A, B, and C harms using methods D and E. Independent group F reviewed our setup. Here's what remains open and what happens next."
- Data line: "Our training data combines licensed, public, and synthetic sources. Rights info and exclusions are documented here."
Media checklist for PR and comms teams
- Replace adjectives with numbers. Replace numbers with methods. Replace methods with links.
- Never imply thought, intent, or feelings. Describe capabilities, constraints, and costs.
- Balance the deck: one slide for gains, one for risks, one for unresolved questions.
- Include a user harm scenario in every briefing, plus the mitigation and escalation path.
- Publish a living FAQ and keep it current as new data arrives.
If you need structured training for your team
Level up your bench with practical courses and certifications built for busy comms teams: AI courses by job role.
The headline will be loud. Your message should be clearer.
AGI talk will spike. Benchmarks will fly. Your edge is simple: crisp definitions, honest limits, credible third parties, and a human impact story that holds up a week later. That's how you keep trust while everyone else chases the next slide.
Your membership also unlocks: