Brave New IA World
Post #001 — Software Engineering x AI

Can AI Resurrect the Rational Unified Process?

What if the methodology we abandoned for being "too heavy for humans" was actually perfect — just waiting for non-human executors?

Ricardo Costa ~10 min read

In the early 2000s, the software industry collectively decided that the Rational Unified Process was too much. Too many artifacts. Too many roles. Too much ceremony. Agile won. RUP faded into textbooks and enterprise consulting decks. But what if the problem was never the process itself — what if it was simply that humans were the wrong executors?

Welcome to Brave New IA World. This blog is about the intersection of software engineering and artificial intelligence — not AI as a product, but AI as a participant in the engineering process itself. And I can't think of a better way to inaugurate it than with the question driving my master's thesis: Can a multi-agent AI system bring the Rational Unified Process back to life?

The short answer: I believe it can — and that it should. Let me tell you why.

The Rise and Fall of RUP

The Rational Unified Process, created by Philippe Kruchten at Rational Software (later IBM), was one of the most comprehensive attempts to formalize software development. It defined four phases (Inception, Elaboration, Construction, Transition), nine disciplines (from Business Modeling to Deployment), and a rich set of roles, activities, and artifacts connected through iterative cycles.

RUP's intellectual foundations were solid. It was built on the Unified Modeling Language (UML), Grady Booch's object-oriented methods, Ivar Jacobson's use-case-driven approach, and Jim Rumbaugh's Object Modeling Technology. It embodied six best practices: develop iteratively, manage requirements, use component-based architecture, model visually, verify quality, and control changes.

So what went wrong? Nothing, really — except that humans found it exhausting. The volume of documentation, the number of artifacts per iteration, the formality of role definitions — all of it created friction. Teams started cutting corners. The Agile Manifesto arrived in 2001, privileging "working software over comprehensive documentation" and "individuals and interactions over processes and tools." RUP became the poster child for heavyweight methodology.

"RUP was never bad engineering. It was bad ergonomics — for a species that gets bored writing the third UML diagram."

But here's the thing: the problems that RUP addressed — scope creep, architectural erosion, poor traceability, integration chaos — didn't go away. Agile teams still struggle with them. We just traded formal discipline for informal heroics. And now, with AI agents entering the scene, we have an opportunity to revisit that tradeoff.

The Rise of Specification-Driven Development

Before diving into the thesis, it's worth understanding the broader movement that makes this possible: Specification-Driven Development (SDD).

SDD is an emerging paradigm where structured specifications — written in markdown or formal templates — serve as the source of truth for AI agents. Instead of asking an AI to "build me a REST API for users," you write a specification that defines the domain model, the API contract, the validation rules, the error codes, and the test scenarios. The AI agent then executes against that spec.

The idea is not new in spirit — it echoes Design by Contract, formal methods, and even the RUP's own emphasis on artifacts. What's new is the executor. When the reader of your specification is an AI with unlimited patience and zero cognitive fatigue, the cost-benefit equation of detailed specs flips entirely.

Tools like GitHub's Spec-Kit, AWS Kiro (an IDE that transforms prompts into structured specs with requirements, design documents, and task plans), and frameworks like Google's Antigravity have all embraced this direction in 2025-2026. The pattern is consistent: write a spec, let agents execute, use quality gates to validate.

Critics — and there are thoughtful ones — argue that SDD risks becoming "Waterfall strikes back", in François Zaninotto's memorable phrasing. The concern is that excessive markdown creates a false sense of control while adding overhead that doesn't survive contact with reality.

This is a legitimate criticism. But it applies to SDD as practiced naively — where specs are static documents that age poorly. What if, instead, the specs were living artifacts inside an iterative process, continuously validated and updated by the agents themselves? What if the process that governed those specs was... RUP?

Enter AIRUP: An AI-First Unified Process

This is the core of my thesis. AIRUP proposes an AI-first software development process in which AI agents assume the traditional roles of RUP, automating activities across the entire software engineering lifecycle:

But the critical innovation is not in the agents themselves — it's in the governance layer.

The AI Governor

AIRUP introduces a new role that doesn't exist in the original RUP: the AI Governor. This is a central agent responsible for controlling cost, preventing decision loops, and ensuring alignment with business value.

In multi-agent systems, a well-known failure mode is the "infinite refinement loop" — where agents keep improving artifacts beyond the point of diminishing returns. The Governor acts as a circuit breaker: it monitors token consumption, tracks iteration counts, evaluates whether quality gates are met, and decides when to stop.

Think of the AI Governor as a blend of a project manager, a cost controller, and a technical lead — but one that operates at machine speed and with perfect memory of every decision made in the process.

Mapping RUP Roles to AI Agents

One of the most interesting exercises in AIRUP is mapping the original RUP roles to multi-agent capabilities. Here's a simplified view:

RUP Role AIRUP Agent Key Responsibility
Business Analyst @business-analyst Context analysis, stakeholder mapping, problem framing
Requirements Specifier @requirements Use cases, functional/non-functional requirements, SDD specs
Software Architect @architect Architecture decisions, component design, tech stack evaluation
Developer @implementer Code generation, TDD cycles, refactoring
Tester @tester Test plans, test generation, quality validation
Deployment Manager @deployer CI/CD pipelines, infrastructure, release management
(New Role) @governor Cost control, loop prevention, business value alignment

Each agent operates within the boundaries of its role, produces artifacts that feed into the next phase, and is subject to quality gates — just like in the original RUP. The difference is that an agent can produce a 40-page requirements specification in minutes, not weeks. It can generate and execute 200 test cases without complaining about "boring work."

The Four Phases, Reimagined

AIRUP preserves RUP's four-phase lifecycle, but the dynamics change dramatically when agents are the executors:

AIRUP Process Phases
Inception
Elaboration
Construction
Transition

Inception: From Vague Idea to Scoped Problem

The human provides a business problem. The @business-analyst agent structures it into a problem statement, identifies stakeholders, and maps pain points. The @requirements agent derives initial use cases. The @governor evaluates feasibility and cost. Output: a scoped vision document and a risk assessment. Time: minutes, not weeks.

Elaboration: Architecture and Risk Mitigation

The @architect agent proposes a component-based architecture, produces decision records, and identifies technical risks. The @implementer builds an executable architecture prototype. The @tester validates architecturally significant scenarios. The @governor monitors whether the team is over-engineering or under-investing in risk mitigation.

Construction: Iterative Build with TDD

This is where Specification-Driven Development shines. Each construction iteration follows a spec → test → implement → validate cycle. The @implementer writes code against specs. The @tester generates and runs tests. The @architect reviews for architectural drift. The @governor tracks progress, cost, and quality metrics per iteration.

Transition: Deployment and Validation

The @deployer agent configures CI/CD pipelines, generates infrastructure-as-code, and manages the release process. The @tester runs acceptance tests. The human validates the final product against business objectives.


Why Now? The Convergence of Three Trends

AIRUP isn't a theoretical exercise disconnected from the real world. Three converging trends make it timely:

1. Multi-agent frameworks are maturing. LangGraph, CrewAI, AutoGen, and OpenAI's Swarm have demonstrated that multi-agent orchestration is not only possible but increasingly reliable. The "agent as worker" pattern — where each agent has a role, tools, and boundaries — maps naturally to RUP's role-based structure.

2. SDD is gaining traction (and criticism). The industry's experiments with specification-driven workflows — from Kiro to GitHub's spec-kit — are generating real-world evidence about what works and what doesn't. AIRUP can learn from these experiments and address the known failure modes (static specs, over-documentation) by embedding specs within an iterative, phase-gated process.

3. The cost problem is real. Multi-agent pipelines are expensive. Without governance, agents waste tokens on unnecessary iterations, over-engineered solutions, and circular reasoning. The AI Governor is a direct response to this problem — and it's a contribution that matters regardless of whether the process is RUP-based.

The Research Question

Central Question

Can a software development process inspired by RUP, executed by AI agents and governed by an AI Governor, reduce coordination and cost problems in multi-agent software development pipelines?

This question has three testable dimensions:

  1. Coordination — Does a role-based, phase-gated structure reduce conflicts, redundancies, and miscommunication between agents compared to unstructured multi-agent pipelines?
  2. Cost — Does the AI Governor's oversight reduce token consumption and iteration waste compared to ungoverned pipelines?
  3. Quality — Does the process produce artifacts and code that meet defined quality standards more consistently than ad-hoc agent workflows?

What AIRUP Is Not

Let me be clear about what I'm not proposing:

This is not "bring back waterfall." RUP was always iterative and incremental. The "heavy waterfall" reputation came from bad implementations, not from the process itself. AIRUP preserves the iterative core — each phase contains multiple iterations, and agents continuously refine artifacts.

This is not "replace developers." The human remains the ultimate decision-maker. The AI Governor reports to a human. Quality gates require human approval at key milestones. Agents propose — humans dispose.

This is not "all software should use AIRUP." A weekend hack doesn't need a formal process. AIRUP targets systems where coordination, traceability, and governance matter — enterprise software, regulated industries, critical infrastructure.

What's Coming Next

This post is the opening statement. In the upcoming posts on Brave New IA World, I'll dive deeper into:


Twenty-five years ago, we decided that RUP was too much process for too little patience. We weren't wrong — we were just the wrong species for the job. AI agents don't get bored. They don't skip documentation because it's Friday afternoon. They don't cut corners on test coverage because the sprint is ending.

Maybe it's time to give that process another chance — with better executors.

"The best process is the one that gets followed. And an AI agent will follow it every single time."
AI Agents RUP AIRUP SDD Multi-Agent Systems Software Engineering Master's Thesis
RC

Ricardo Costa

Software engineer exploring the intersection of classical software processes and AI-driven development. Currently pursuing a master's degree researching AIRUP — an AI-first approach to the Rational Unified Process.