P5.express and Agentic AI: Where It Helps, Where It Breaks Things

24 min read
Stanislav Belyaev
Stanislav Belyaev Engineering Leader at Microsoft
P5.express and Agentic AI: Where It Helps, Where It Breaks Things

In the PMI world, portfolio management is typically imagined as something monumental: steering committees, Tableau dashboards, hundreds of Jira fields, weekly status meetings with a deck of slides. P5.express offers a different approach. Three cycles, five documents, two roles. The entire system fits on a single page.

This is exactly the kind of system where agentic AI makes sense: minimalist architecture that’s easy to understand, clearly defined roles, structured data. But “makes sense” doesn’t mean “everywhere.” Some parts of P5.express stop working when automated – not because the AI is bad, but because those parts derive their value from the human process itself.

Below is a cycle-by-cycle breakdown. What’s worth delegating to an agent, what’s better left to people, and which model fits these tasks best.

What Is P5.express

P5.express is a portfolio management methodology developed by OMIMO. It sits one level above familiar frameworks like Scrum or PRINCE2 and answers a different question – not “how to run projects” but “which projects to run at all.”

P5.express structure

The structure consists of three nested cycles.

X – every six months (strategic): evaluate generated value (X1), optimize the value generation strategy (X2), conduct focused communication (X3). The key event is a one-day X2 workshop where the portfolio board reprioritizes.

Y – monthly (health and improvements): assess stakeholder satisfaction (Y1), evaluate running programs and projects (Y2), plan improvements (Y3), conduct focused communication (Y4).

Z – daily (execution): manage follow-up items (Z1), start/stop/pause programs and projects (Z2), balance resources (Z3).

Five documents: Portfolio Description, Value Generation Matrix, Global Follow-Up Register, Global Health Register, and Business Cases.

The most important one is the Value Generation Matrix. It’s a table of all programs and projects with columns: name, sponsor, status, progress, investment, benefits, Value. Value is calculated as benefits / investment. Rows are ordered: completed – running – pending – cancelled – rejected. Pending items are balanced within the “balancing horizon” by value categories.

Two roles: the Portfolio Board (department heads, collegial decisions) and the Portfolio Manager (facilitator, analyst, with no direct role in any project – otherwise it’s a conflict of interest).

That’s the entire system. The rest is application.

Where the Agent Genuinely Helps

Y2 (Monthly Project Evaluation) – This Is Agent Work

The most expensive rhythm in P5.express is Y2 – the monthly evaluation of running programs and projects. The portfolio manager must collect data on all active programs and projects every month: progress, completion forecast, recalculated expected benefits. Then update the Value Generation Matrix.

Here’s what a fragment of a real matrix looks like from the P5.express example – active projects only:

Value Generation Matrix fragment – active projects

In practice, updating this table means: open Jira (or Monday, or Smartsheet, or three Excel files from three different project managers), pull the current progress for P0040, P0014, P0017, compare with last month – has the completion forecast slipped, have investment or benefit estimates changed – and recalculate Value. One to two hours of manual work, repeated every month.

An agent does this faster. The instruction looks roughly like this:

Try it yourself
Updating the Value Generation Matrix
You
You are a portfolio manager. Here is the current state of the Value Generation Matrix (fragment – only projects with "In Progress" status): | ID | Project | Sponsor | Progress | Investment | Benefits | Value | |---|---|---|---|---|---|---| | P0040 | CRM migration to new platform | Azar | 80% | $ 210 | $ 250 | 1,2 | | P0014 | Regulatory reporting automation | Nur | 50% | $ 180 | $ 200 | 1,1 | | P0017 | Customer mobile app launch | Tiam | 25% | $ 240 | $ 300 | 1,3 | Updates for this month: - P0040: data migration completed, payment gateway integration remaining. Progress ~90%. Budget unchanged. - P0014: new regulatory requirements discovered (form 0409-135). Progress remains at 50%, but investment estimate increased to $ 220. Completion forecast shifted by 2 months. - P0017: delay with UI design. Progress 30% instead of planned 45%. Budget unchanged so far. Task: propose updated matrix rows. For each project, specify: new progress, current investment and benefits, recalculated Value. Flag projects where the completion forecast has slipped by more than 20%. Format: table + brief comment per project (1–2 sentences).
Comparing:
gemini/gemini-3-flash-preview · deepseek/deepseek-v3.2 · qwen/qwen3.5-397b-a17b

The critical condition: the agent proposes changes – a human approves them. The matrix doesn’t update automatically after this. The portfolio manager reviews the diff and clicks “accept” or “adjust.” A draft for review, not an autonomous update – exactly the format that reduces losses from hallucinations. Separately, always verify the arithmetic: the agent will confidently output Value = 1.2 even when the underlying investment and benefit figures contradict each other.

Z1 (Daily Follow-Up Monitoring) – Pattern Matching Across N Registers

The Follow-Up Register in P5.express isn’t just about risks. It’s a list of risks, issues, change requests, improvement plans, and lessons learned. In RAID terminology (Risks, Assumptions, Issues, Dependencies) or RICs – risk is just one category. The P5.express manual explicitly states: the portfolio manager must continuously monitor local registers to find patterns that need attention at the portfolio level.

“Continuously monitoring N registers across all categories” is exactly the kind of work that agentic mode was built for. Once a week, the agent scans Follow-Up Registers from all active projects – risks, issues, change requests, lessons – clusters similar items by topic, and proposes candidates for the Global Follow-Up Register along with a suggested owner. The portfolio manager sees a ready-made list and decides what to escalate.

Try it yourself
Finding cross-project patterns in Follow-up Registers
You
You are a portfolio manager. Below are exports from the Follow-up Registers of three active projects. Each register contains risks (R), issues (I), change requests (C), and lessons learned (L). Your task: find patterns that recur across two or more projects and require attention at the portfolio level. Register P0040 "CRM Migration": - R-040-1: Payment gateway integration depends on the bank's API, bank response time – 10 business days. Status: open. - R-040-2: A single developer knows the data migration logic (bus factor = 1). Status: accepted. - R-040-3: Load testing not scheduled, environment not allocated. Status: open. - I-040-1: Old CRM API documentation is incomplete, the team is spending time on reverse engineering. Status: in progress. - C-040-1: Request from marketing – add integration with a new advertising platform to the migration scope. Status: under review. Register P0014 "Reporting Automation": - R-014-1: New regulatory requirements (form 0409-135) not accounted for in the estimate. Status: open. - R-014-2: Dependency on the Central Bank's external API, SLA not formalized. Status: open. - I-014-1: Test environment shared with P0040, scheduling conflicts. Status: open. - L-014-1: Previous reporting project (P0009) missed deadlines due to late-discovered regulatory requirements. Pattern is repeating. Register P0017 "Mobile App": - R-017-1: UI designer works part-time, simultaneously on another project. Status: open. - R-017-2: App Store review takes 5–7 business days, not accounted for in the plan. Status: open. - R-017-3: Dependency on the bank's API for the payment module, similar to P0040. Status: open. - I-017-1: Load testing not scheduled here or in P0040. Status: open. - C-017-1: Request from sales – add push notifications not included in the original scope. Status: under review. Task: 1. Find items that recur as a theme across two or more projects – across all categories (risks, issues, change requests, lessons). 2. For each pattern: describe the essence, list affected projects, propose a portfolio-level owner (who should resolve it). 3. Separately note: are there items that appear to share the same root cause (shared dependency, shared resource, recurring lesson)? Format: pattern table + brief recommendation per pattern (1–2 sentences).
Comparing:
gemini/gemini-3-flash-preview · deepseek/deepseek-v3.2 · qwen/qwen3.5-397b-a17b

The shape of the task changes: instead of “I scan 8 registers manually” – “I review the agent’s proposal and make a decision.” The second version scales; the first doesn’t. In the terms of the work intensification research – the agent here doesn’t speed up existing work but redistributes cognitive load: data scanning goes to the machine, decision-making stays with the human.

X3/Y4 (Stakeholder Communication) – Draft Communications

Focused communication is the part of P5.express that, in practice, gets done least often. The reason is mundane: writing monthly stakeholder updates that always look the same is boring.

The agent solves the motivation problem. After the Value Generation Matrix is updated, it drafts the communication: what changed this month, what stands out, which projects moved between statuses. The draft takes 5 minutes to edit – versus 40 minutes to write from scratch.

Try it yourself
Draft monthly stakeholder update
You
You are a portfolio manager preparing a monthly stakeholder update following the P5.express methodology (step Y4 – focused communication). Data from the updated Value Generation Matrix for this month: Projects in progress: - P0040 "CRM migration to new platform" (sponsor Azar): progress increased from 80% to 90%, data migration completed, payment gateway integration remaining. Budget unchanged ($ 210). On track for on-time completion. - P0014 "Regulatory reporting automation" (sponsor Nur): progress remains at 50%. New regulatory requirements discovered (form 0409-135), investment estimate increased from $ 180 to $ 220. Completion forecast shifted by 2 months. - P0017 "Customer mobile app launch" (sponsor Tiam): progress 30% instead of planned 45%. Delay with UI design. Budget unchanged ($ 240). Task: write a draft monthly update for the portfolio board. Format: 1 paragraph of overall portfolio status, then 2–3 sentences per project focusing on deviations from plan. At the end – 2–3 items requiring board attention (decisions or escalations). Business tone, no filler. Length – no more than 1 page.
Comparing:
gemini/gemini-3-flash-preview · deepseek/deepseek-v3.2 · qwen/qwen3.5-397b-a17b

The output is predictably similar every month – which is exactly what you need from a standard update.

A separate task is ensuring that cycles actually happen. Minimalist systems die from skipped rituals: miss one matrix update and it’s stale, miss two and it stops being the source of truth. The simplest watchdog is a weekly calendar task or a cron script that checks the modification dates of portfolio files and sends a reminder if a document hasn’t been updated within the expected cycle.

Portfolio management is one of the tasks covered in the course. Try 9 practical AI management tasks – free, no registration required.

No payment required • Get notified on launch

Join Waitlist

The Gray Zone: Helps, but with Caveats

X2 (Semi-Annual Workshop) – Workshop Preparation

X2 is the semi-annual portfolio reassessment workshop. It’s the key event in the system, and it cannot be automated. But the preparation for it can be.

A week before the workshop, the agent can assemble: an opportunity cost simulation, a gap analysis by value categories (which categories are underrepresented in the current portfolio), a benefits realization delta (what was forecasted vs. what was delivered for completed projects), and a clustering of bottom-up ideas.

The opportunity cost simulation is the most valuable of these tasks and also the most tedious to do manually. Here’s what the prompt looks like:

Try it yourself
Opportunity cost simulation before the X2 workshop
You
You are an analyst preparing materials for the semi-annual portfolio reassessment workshop (X2 per the P5.express methodology). Current Value Generation Matrix: Projects in progress (balancing horizon: $ 500): - P0040 "CRM Migration" | Investment: $ 210 | Benefits: $ 250 | Value: 1,2 | Progress: 90% - P0014 "Reporting Automation" | Investment: $ 220 | Benefits: $ 200 | Value: 0,9 | Progress: 50% Pending projects (in priority order): - P0017 "Mobile App" | Investment: $ 240 | Benefits: $ 300 | Value: 1,3 - P0005 "Logistics Optimization" | Investment: $ 80 | Benefits: $ 90 | Value: 1,1 - P0027 "Knowledge Management System" | Investment: $ 130 | Benefits: $ 150 | Value: 1,2 New request from sponsor: P0051 "Marketplace Integration" | Investment: $ 180 | Benefits: $ 350 | Value: 1,9 Task: model what happens to the pending project queue if the board approves P0051. Which projects fall outside the balancing horizon? Which value categories lose representation? Prepare two scenarios: (1) P0051 launches immediately, (2) P0051 waits for P0040 to complete. For each scenario, show the resulting queue and the portfolio's total Value.
Comparing:
gemini/gemini-3-flash-preview · deepseek/deepseek-v3.2 · qwen/qwen3.5-397b-a17b

The board arrives at the workshop with pre-built scenarios instead of modeling them on the fly. The quality of the conversation shifts: they discuss trade-offs rather than doing mental arithmetic.

The caveat: the agent prepares the analytical brief, not the agenda. Priority decisions are collegial and political in the best sense of the word. They cannot be automated, and they shouldn’t be.

Red Team for Business Cases

P5.express requires a business case before every new project enters the portfolio: rationale / alternatives / expected benefits / measurement method / investment estimate / key risks.

The agent can play devil’s advocate. A sponsor describes a project in a few sentences – the agent scaffolds a full P5.express business case with all mandatory fields, where every number is tagged [needs verification], and immediately runs a pre-mortem on the same case.

Try it yourself
Business case scaffold + pre-mortem
You
You are performing two tasks sequentially. TASK 1: Business case scaffold per P5.express. The sponsor described the project as follows: "We want to implement a knowledge management system – an internal portal where teams document decisions, architectural choices, and post-mortems. Currently, knowledge is lost when employees leave. Estimate: $ 130, timeline – 6 months, we expect a 40% reduction in onboarding time." Expand this into a full P5.express business case: - Rationale (why this project is needed in the portfolio) - Alternatives (at least 2, including "do nothing") - Expected benefits (with measurement method for each) - Investment estimate (broken down by category) - Key risks (at least 5) Every number you cannot confirm from the sponsor's description, mark as [requires verification]. Do not invent specific figures – leave placeholders. TASK 2: Pre-mortem of this same business case. Imagine 8 months have passed and the project has failed. Name 8 specific reasons for the failure. For each reason, specify: - What exactly went wrong - Why this was predictable already now (which signal was ignored) - How likely this is in reality (high / medium / low) Focus on reasons specific to this project – not generic phrases like "lack of resources."
Comparing:
gemini/gemini-3-flash-preview · deepseek/deepseek-v3.2 · qwen/qwen3.5-397b-a17b

The agent doesn’t fill in numbers – it helps the sponsor verify them. In practice, stress-testing a business case is a chain of several techniques: pre-mortem surfaces failure scenarios, a WBS audit finds missed tasks and dependencies, a stakeholder simulator prepares you for tough questions during the defense. The full chain – from failure scenarios to rehearsing the defense before the CFO – is broken down step by step in the Project Management specialization on the platform.

The key danger worth stating explicitly: if the agent proposes numbers and the sponsor accepts them without verification, hallucinated benefits end up in the matrix. The ranking silently drifts. The agent as red team – valuable. The agent as a source of plan figures – dangerous.

Related – Z2/Z3 (start/stop projects and resource balancing): the agent can steelman arguments against a specific sponsor’s request, referencing thresholds from the Portfolio Description. This depersonalizes politically sensitive conversations.

Where AI Breaks the System

An honest section, because P5.express is one of those methodologies where automation can cause methodological harm, not just technical harm.

Y1 (satisfaction surveys) through AI destroys the data. Y1 consists of anonymous stakeholder satisfaction surveys. Anonymity is the precondition for honesty. The problem is that an agent breaks anonymity through several paths, even if you didn’t intend it.

Three paths to de-anonymization: small sample, stylometry, cross-reference

The first is de-anonymization through small sample size. A portfolio typically has 5–8 stakeholders. If a respondent writes about delays in project P0014 and the agent knows that P0014’s sponsor is Nur, the response is no longer anonymous. A human analyst might notice this too, but the agent does it systematically across all responses simultaneously.

The second is stylometry. LLMs recognize individual writing patterns: sentence length, characteristic vocabulary, argument structure. In a group of six people where the agent has read each person’s work correspondence, matching an anonymous response to its author is a trivial task. The respondent doesn’t even need to mention a project – writing in their usual style is enough.

The third is cross-referencing with context. The portfolio manager’s agent has access to registers, correspondence, meeting minutes. An anonymous response saying “timelines are unrealistic, the team is overloaded,” combined with resource utilization data from Z3, points to a specific manager. The agent isn’t malicious – it simply connects dots, because that’s what it’s designed to do.

The result: stakeholders will quickly realize that anonymity is nominal and start giving socially desirable answers. Y1 data becomes useless. Satisfaction surveys are one of the parts of P5.express that must remain entirely human: collection, aggregation, analysis.

Mechanical optimization of the Value Generation Matrix. The P5.express manual warns explicitly: “the matrix can never be accurate enough to be optimized mechanically.” Value = benefits / investment is an estimative construct, not a precise number. If the agent sorts the matrix by Value and suggests “launch projects from the top, reject from the bottom” – it creates false precision. The board starts trusting the ranking instead of discussing it. This destroys the core purpose of collegial decision-making.

Adding data fields for the agent’s benefit. P5.express contains an explicit anti-bloat principle: don’t collect data you don’t use. If you add 15 new fields to the matrix for the sake of better analytics, the system becomes more complex, more expensive to maintain, and the minimalism – the very reason you chose it – disappears. The agent should work with the data that already exists.

Practical Checklist: What to Try This Week

If you’re running a portfolio with P5.express or thinking about starting – here’s a concrete first step for each cycle.

Z-cycle (daily): Try giving the agent two or three Follow-Up Registers (risks, issues, change requests) from active projects and ask it to find recurring themes. This is your first touch of Z1 (follow-up monitoring) – and a good test of whether your registers are machine-readable at all.

Y-cycle (monthly): Before your next monthly project evaluation (Y2), export progress data from your management system into a file and ask the agent to propose updated matrix rows. Notice where it makes mistakes – that shows you where the data has gaps.

X-cycle (semi-annually): Before your next semi-annual workshop (X2), ask the agent to formulate 10 arguments against your current portfolio – which projects could have been skipped, why the opportunity cost is high. This isn’t an agenda; it’s a tool for a better conversation.

All three tasks can be run in any chat – ChatGPT, Kimi, DeepSeek – by copying data from your system. A coding agent (OpenCode or similar) is needed when you want to automate the cycle: the agent reads files on its own, runs on a schedule, and proposes updates without manual copy-pasting. The OpenCode article covers three tasks for day one. The agentic approach to data analysis is covered in more detail in a separate case study with three files.

For a closer look at the operational rhythm of a manager working with AI, check that article if P5.express feels too high-level – it covers specific weekly tasks.

Working with multiple tools at once looks simple. The difficulty appears when you need not a one-off analysis but a reproducible process – an agent that runs on a schedule, knows the structure of your files, and produces a consistent format. The difference between “played with an agent” and “integrated an agent into my workflow” comes down to the skill of task formulation and knowing the boundaries of autonomy.

Portfolio tasks, business cases, cycle rhythms – test your AI approach on 9 real management tasks in the open module. Free, no registration required.

No payment required • Get notified on launch

Join Waitlist

Which Model to Choose for These Tasks

The agentic tasks in P5.express are primarily structured work: reading files, comparing data, drafting communications, asking questions by template. This doesn’t require maximum intelligence, but it does require reliability and strong performance with tables and documents.

Based on our benchmark of 54 models for management tasks, here’s the picture.

Top tier: Claude Sonnet 4.5 leads in the “Analysis & Decision-Making” and “Planning” categories with a score of 4.84. For red team business case tasks and X2 workshop preparation, it’s a strong choice. If your organization has access to the Anthropic API, this is the best option for complex analytical work.

Strong alternatives: Kimi K2.5 scored 4.67 in our benchmark – the top-ranking model among those available without regional restrictions. For daily and monthly cycle tasks, it’s more than sufficient. Qwen3.5 Plus at 4.57 is cheaper ($0.26 per million input tokens vs. $0.45 for Kimi) and handles these tasks well. DeepSeek V3.2 at the same price point performs reliably with tabular data.

The practical takeaway: for daily monitoring (Z1) and monthly project evaluation (Y2) – drafts for review, working with tables – Qwen3.5 Plus or DeepSeek is enough. For the semi-annual workshop (X2), red team, and complex analysis – Kimi K2.5. Claude via a coding agent, if you have access, is the best option for complex tasks.

Coding agents like OpenCode or Claude Code fit both scenarios: they’re vendor-neutral and connect to any provider. For a manager, this is a practical advantage – you pick the best available model for each task type without locking into a single vendor.

Conclusion

P5.express is designed in a way that makes automation work asymmetrically. Where you need data processing and a reproducible process, the agent reduces friction and makes cycles realistic. Where collegial process and human contact matter, automation destroys exactly what the methodology exists for.

Surprisingly, a minimalist system turns out to be a good filter for evaluating agentic tools: if the task sounds like “collect data, compare with last time, propose a draft” – the agent fits. If the task sounds like “talk to people and help them reach a decision” – it doesn’t.

That’s a useful question for any other methodology, too.

Specialisation

From Experiment to System

Agentic tasks like the ones analyzed here require skill – not just tools. The course foundation covers prompt engineering and critical thinking with AI. The Project Management specialization breaks down AI in the manager's operational rhythm: from business cases to monthly cycles.

От pre-mortem до антикризисного плана
Переиспользуемые промпт-шаблоны
Сквозной кейс на реальном проекте
~300 часов экономии в год
Stanislav Belyaev

Stanislav Belyaev

Engineering Leader at Microsoft

18 years leading engineering teams. Founder of mysummit.school. 700+ graduates at Yandex Practicum and Stratoplan.