TL;DR: Understand the Claude Desktop Architecture and Save Time
You configured Claude in Claude Desktop, wrote instructions, uploaded reference files, and set your preferences. Then you clicked the Cowork tab.
Unfortunately, Claude had no memory of what you just did. Your instructions were gone, as were your files and preferences. You assumed this was a bug, but it is a feature: You switched applications.
Your Claude Pro subscription hits limits faster than it did in January, as Anthropic quietly re-priced the ceiling, and every AI provider is rationing compute. If you keep working with Claude the way you did six months ago, you are in for a rude awakening. This article gives you four principles that explain how Token Economics actually works, so you can stop accepting the black box and start using your budget deliberately.
by Stefan Wolpers|FeaturedAgile and ScrumAgile Transition
Using AI at Work Does Not Mean You Understand It
Many agile practitioners use ChatGPT at work. That does not mean they understand AI well enough to trust your own judgment. The problem is not that agile practitioners ignore AI. The problem is that many already use it confidently without knowing where their judgment breaks down. The free AI4Agile Foundational Assessment measures precisely this skill gap. (Download your access file below.)
The assessment comprises 40 scenario-based questions. It does not ask for definitions, but puts you into situations that agile coaches, product managers, and Scrum Masters face every week: weak prompting producing generic output, misleading data analysis, questionable agent output, and, possibly, organizational pressure to treat AI output as “good enough” to go with it.
Most people who use AI do not fail because they lack knowledge, but because they cannot distinguish between plausible outputs and trustworthy judgment. But see for yourself!
by Stefan Wolpers|FeaturedAgile and ScrumAgile Transition
TL; DR: AI Thinking Skills for Agile Practitioners
Most agile practitioners use AI to produce outputs more quickly. Few use it to think better. This free download gives you three AI thinking skills (Socratic Explorer, Brutal Critic, Pre-Mortem) that turn Claude into a partner for diagnosing problems, stress-testing plans, and anticipating failures before they happen.
by Stefan Wolpers|FeaturedAgile and ScrumAgile Transition
TL;DR: The A3 Handoff Canvas
The A3 Framework helps you decide whether AI should touch a task (Assist, Automate, Avoid). The A3 Handoff Canvas covers what teams often skip: how to run the handoff without losing quality or accountability. It is a six-part workflow contract for recurring AI use: task splitting, inputs, outputs, validation, failure response, and record-keeping. If you cannot write one part down, that is where errors and excuses will enter.
The Handoff Canvas closes a gap in a useful pattern: from an unstructured prompt to applying the A3 framework to document decisions with the A3 Handoff Canvas, to creating transferable Skills, potentially leading to building agents.
by Stefan Wolpers|FeaturedAgile and ScrumAgile Transition
TL; DR: The A3 Framework
The A3 Framework categorizes AI delegation before you prompt: Assist (AI drafts, you actively review and decide), Automate (AI executes under explicit rules and audit cadences), or Avoid (stays entirely human when failure would damage trust or relationships). Most AI training teaches better prompting. The A3 Framework teaches the prior question: Should you be prompting at all? Categorize first, then prompt.
by Stefan Wolpers|FeaturedAgile and ScrumAgile Transition
TL; DR: Why the Brand Failed While the Ideas Won
Your LinkedIn feed is full of it: Agile is dead. They’re right. And, at the same time, they’re entirely wrong.
The word is dead. The brand is almost toxic in many circles; check the usual subreddits. But the principles? They’re spreading faster than ever. They just dropped the name that became synonymous with consultants, certifications, transformation failures, and the enforcement of rituals.
You all know organizations that loudly rejected “Agile” and now quietly practice its core ideas more effectively than any companies running certified transformation programs. The brand failed. The ideas won.
TL; DR: Bridging Agile and AI with Proper Prompt Engineering
Agile teams have always sought ways to work smarter without compromising their principles. Many have begun experimenting with new technologies, frameworks, or practices to enhance their way of working. Still, they often struggle to get relevant, actionable results that address their specific challenges. Regarding generative AI, there is a better way for agile practitioners than reinventing the wheel team by team—the Agile Prompt Engineering Framework.
Learn why it solves the challenge: a structured approach to prompting AI models designed specifically for agile practitioners who want to leverage this technology as a powerful ally in their journey.
TL; DR: The Scrum Master Interview Guide to Identify Genuine Scrum Masters
In this comprehensive Scrum Master Interview guide, we delve into 97 critical questions that can help distinguish genuine Scrum Masters from pretenders during interviews. We designed this selection to evaluate the candidates’ theoretical knowledge, practical experience, and ability to apply general Scrum and “Agile “principles effectively in real-world scenarios—as outlined in the Scrum Guide or the Agile Manifesto. Ideal for hiring managers, HR professionals, and future Scrum teammates, this guide provides a toolkit to ensure that your next Scrum Master hire is truly qualified, enhancing your team’s agility and productivity.
If you are a Scrum Master currently looking for a new position, please check out the “Preparing for Your Scrum Master Interview as a Candidate” section below.
So far, this Scrum Master interview guide has been downloaded more than 25,000 times.
TL; DR: 82 Product Owner Interview Questions to Avoid Imposters
If you are looking to fill a position for a Product Owner in your organization, you may find the following 82 interview questions useful to identify the right candidate. They are derived from my sixteen years of practical experience with XP and Scrum, serving both as Product Owner and Scrum Master and interviewing dozens of Product Owner candidates on behalf of my clients.
So far, this Product Owner interview guide has been downloaded more than 10,000 times.
TL; DR: Scrum Training Classes, Liberating Structures Workshops, and Events
Age-of-Product.com’s parent company — Berlin Product People GmbH — offers Scrum training classes authorized by Scrum.org, Liberating Structures workshops, and hybrid training of Professional Scrum and Liberating Structures. The training classes are offered both in English and German.
Check out the upcoming timetable of training classes, workshops, meetups, and other events below and join your peers.
TL; DR: AI Playbook, Product Taste — Food for Agile Thought #543
Welcome to the 543rd edition of the Food for Agile Thought newsletter, shared with 35,597 peers. This week, John Cutler warns in his AI Playbook that AI accelerates broken practices just as much as good ones, while Guy Champniss adds that the psychological toll of AI adoption on employee motivation may quietly erase the productivity gains organizations expect. Teresa Torres and Petra Wille challenge the rising “taste” narrative in product management, proposing that discovery skills and evidence beat gut feeling. David Pereira diagnoses why product management in Europe often underperforms, pointing to roadmap theater and consensus paralysis, and Laura Klein reminds us that involving engineers in research builds shared understanding faster than any handoff ever could.
Next, Andrej Karpathy describes how software development is shifting from vibe coding to agentic engineering, where judgment and architectural understanding matter more than writing code, and Cedric Chin offers three techniques for making sense of AI without losing your current frame. Also, Richard Kasperowski reminds us that CEOs never cared about Scrum, only results, and that AI raises the bar on engineering fundamentals. Paweł Huryn shows how Claude Design compresses product discovery from weeks to hours, and Wyndo and Dheeraj Sharma walk through building a competitor intelligence agent that develops editorial judgment over time.
Lastly, Jake Handy warns that executives mistake AI token consumption for productivity, while Paul LaPosta shows how AI-generated code inflates DORA metrics by boosting speed while eroding system understanding. Jack Clark raises the stakes further, estimating a 60%+ chance of fully automated AI R&D by 2028. Jim Lewis, Jeff Sauro, and colleagues find that AI usability evaluations catch only half the problems humans identify while generating unverified issues on top. Finally, Mike Fisher reminds us that belonging is not a soft perk but a measurable performance driver, built or destroyed one manager interaction at a time.
TL;DR: AI Adoption Issues Sound Familiar to Agile Practitioners
If you have spent the last twenty years arguing that velocity is not value, that adoption is not impact, that an Agile transformation is not a Jira migration, the Stanford AI Index 2026 will read like déjà vu: The technology is new. The failure mode, the AI spending trap, is not.
The many organizations that have adopted AI but cannot show an EBIT impact are the same organizations that adopted Scrum without learning empiricism, adopted DevOps without changing how they fund teams, and adopted product management without giving anyone product authority.
The economic data is the evidence. The interpretation is what you, the agile practitioner, already know.
TL; DR: Slowing Down with AI — Food for Agile Thought #542
Welcome to the 542nd edition of the Food for Agile Thought newsletter, shared with 35,608 peers. This week, Mario Zechner advocates for slowing down with AI, warning that unsupervised coding agents compound errors faster than teams can fix them. Stephanie Leue shows how AI-driven speed tempts teams to skip discovery, incurring a hidden “Alignment Tax,” while Jenny Wanger and Michael Goitein find lasting advantage in internal capabilities, not copyable features. Mark Nottingham flags AI agents bypassing browser-level protections, Wharton’s Blueprint examines barriers to adoption, and Joost Minnaar uses the Titanic to show how silos filter critical signals.
Next, Teresa Torres and Petra Wille challenge the reflex to centralize decisions when uncertainty hits, arguing that real leadership sets direction and builds trust. Pawel Brodzinski extends that theme to AI-generated specs that look complete yet erode the human communication that teams need. Maxim Massenkoff shares Anthropic’s survey of 81,000 users, revealing that early-career workers worry most about displacement. Matthew Littlehale recounts replacing Scrum with Shape Up, and Andrej Karpathy reframes LLMs as a new computing paradigm requiring human judgment throughout.
Lastly, Steven J. Vaughan-Nichols warns that enterprise AI lock-in runs deeper than executives admit, with failed migrations and rising costs compounding quickly, while Kevin Kelly frames this instability as part of a broader “Age of Ambiguity” that demands radical adaptability. Dave Snowden surfaces a foundational tension within the CRP tradition between facilitated practice and radical process ontology. On the practical side, Michael Crist guides non-technical professionals through setting up Claude Cowork, while OpenAI’s GPT-5.5 prompt guidance shifts toward outcome-first instructions over process-heavy prompts.
Two weeks ago, I asked my audience whether they wanted a short course on moving from Scrum to a Product Operating Model, and 22 answered. That was not the Scrum-to-POM dataset I hoped for, but it was valuable for the conversations. Interestingly, one pattern ran through more than a quarter of the responses: The people writing back were not asking about transformation practices or operating models. They were asking what was about to happen to their jobs.
Let me paraphrase some of their replies: One Agile Coach wrote that their role had already been made redundant, and the internal training their employer offered was not enough. Another asked a blunt question: “What will happen to my role?” A third described leadership, saying they wanted this shift, while their behavior remained inconsistent. A fourth reported confusion about what a product coach actually is. A fifth dismissed the whole discourse as high-level fluff, transformational buzzwords, zero accountability, and vague systems thinking with no teeth.
My takeaway: While the organizational design debate appears to be the surface, the ongoing role repositioning is what the people on the ground are living through.
TL; DR: GPT-5.5 & Product on Speed — Food for Agile Thought #541
Welcome to the 541st edition of the Food for Agile Thought newsletter, shared with 35,619 peers. This week, OpenAI’s GPT-5.5 signals another meaningful capability jump, with Ethan Mollick noting that stronger models and richer tool harnesses now handle serious work, even as creative judgment still exposes AI’s limits. Shopify CTO Mikhail Parakhin, interviewed by swyx, reports near-universal internal AI adoption where review and judgment now matter more than code generation. Cat Wu tells Lenny Rachitsky how Anthropic builds ahead of model readiness and treats mission alignment as real leverage. Stephanie Leue warns that AI magnifies existing product leadership debt, making weak trust and brittle confidence visible faster, while Mike Fisher argues that confident teams stop checking too early and calls for red-teaming as standard procedure, not an afterthought.
Next, James Maxwell argues for funding durable teams around value streams rather than fixed-scope projects, while Lenny Rachitsky and Nikhyl Singhal warn that AI now exposes PMs who manage process theater instead of creating real leverage. Ryan Greenblatt adds another concern: frontier AIs oversell incomplete work and reward-hack hard tasks, making verification unreliable exactly where the stakes rise. Sean Goedecke questions whether anti-AI activism maps onto historical Luddism, and Ian Vanagas proposes hackathons as a protected space where autonomous teams ship real products.
Lastly, Grant Harvey notes power users stack AI tools by job, not brand, with Claude leading on coding, while ChatGPT, Gemini, and Copilot survive on switching costs. Aakash Gupta shows how Claude Routines collapse infrastructure drag but force review discipline, and Mark Graban reframes Lean’s eight wastes as a lens for system problems, not blame. Finally, Lisa Crispin treats AI testing as capability-building, while Alex Singla and co-authors argue that transformation succeeds when leaders earn trust.