AIPCH01 - Domain-Owned
“Named AI Product Owner (AIPRO) Assigned”
What AIPCH01 is really asserting
AIPCH01 is not asserting that:
“A name exists in a field called
AI Product Owner.”
It is asserting that:
An AI Product must be owned by the business domain that is accountable for the decision capability, its intent, its boundaries, and its consequences - with a named AIPRO who holds authority over how that capability behaves and evolves.
Ownership is about decision authority, not technical stewardship.
The Essence (HDIP + AIPS Interpretation)
An AI Product is domain-owned if and only if:
- The AIPRO represents a business/domain mandate
- The AIPRO is accountable for the behavior and outcomes of the AI capability
- Technology teams act as enablers, not owners
If ownership sits with:
- AI platform teams
- data science teams
- engineering teams
then the system may be well-built - but it is not an AI Product in the AIPS sense.
The Authoritative Domain Test
(Applicable across Data and AI Products)
A function qualifies as a valid AIPRO only if it can answer “Yes” to all three questions below.
1. Decision Authority Test
Is this function the authoritative owner of the decision capability represented by the AI Product?
- Does it define:
- what decisions are being made?
- what constitutes a correct or acceptable outcome?
- what boundaries apply?
If the answer is:
- “We built the model”
- “We host the system”
- “We operate the platform”
→ ❌ Not authoritative.
2. Outcome Accountability Test
Is this function accountable for the consequences of the AI Product’s decisions?
- Are they accountable if:
- the decision is wrong?
- harm occurs?
- regulatory issues arise?
If accountability sits elsewhere → ❌ Not authoritative.
Ownership without consequence is superficial.
3. Evolution Authority Test
Does this function control how the AI Product evolves over time?
This includes:
- decision logic changes
- risk posture adjustments (R0–R4 implications)
- policy constraints
- retraining or behavioral updates
- lifecycle decisions (versioning, deprecation)
If they must defer to another domain → ❌ Not authoritative.
Decision Rule
A function qualifies as a valid AIPRO only if it holds decision authority, bears outcome accountability, and controls the evolution of the AI capability.
Fail any one → ownership is misaligned.
Positive Criteria - When AIPCH01 is met
AIPCH01 is met when all of the following are true:
1. The AIPRO role is business-aligned
The AIPRO is:
- a domain representative (e.g., Risk, Fraud, Payments, Compliance, Sales)
- accountable for:
- what the AI does
- why it exists
- how it should behave
- empowered to make decisions about:
- acceptable outcomes
- trade-offs (accuracy vs risk, speed vs safety)
2. The AI Product is defined in business capability terms
Evidence includes:
- product purpose expressed as:
- “detect fraud”
- “recommend credit decisions”
- “prioritize leads”
- not:
- “classification model”
- “LLM pipeline”
- “ML service”
The product represents a capability, not a model.
3. Technology is a supplier, not the principal
Technology roles:
- build, train, and operate the capability
- provide tooling, infrastructure, and optimization
- do not define behavior or decision boundaries
They enable - they do not own.
Negative Criteria - When AIPCH01 is not met
AIPCH01 is not met if any of the following are true:
❌ Ownership is held by AI/ML or platform teams
Examples:
- Head of Data Science listed as AIPRO
- AI Platform Lead acting as owner
- Engineering manager responsible for product definition
These roles are enablers, not authoritative owners.
❌ The AI Product is defined as a technical artifact
Examples:
- “Fraud Detection Model v3”
- “Customer Segmentation Pipeline”
- “LLM Chatbot Service”
This reflects implementation ownership, not capability ownership.
❌ Business involvement is advisory, not authoritative
Examples:
- business provides requirements but cannot change behavior
- model decisions owned by data science
- risk thresholds controlled by engineering
This is proxy ownership, not domain ownership.
Edge Cases (Important Guidance for Agents)
Case 1: “AI Product owned by Data Science team”
❌ Not met
Rationale:
- Data Science is a capability provider, not a domain authority
- violates decision accountability principle
Case 2: “Business owner exists, but cannot change model behavior”
⚠️ Partial at best
Rationale:
- indicates transitional maturity
- authority exists in name but not in control
Case 3: “Shared ownership across domains (e.g., Risk + Fraud)”
✅ Can be met, if:
- one accountable AIPRO is clearly defined
- governance model supports shared input but single accountability
Ambiguity is allowed - absence of accountability is not.
Case 4: “Agent-based AI composed across multiple domains”
⚠️ Context-dependent
Rationale:
- ownership must be defined at:
- product level (AIPRO)
- not fragmented across components
- composition does not remove accountability
Evidence Signals an Agent Should Look For
Authoritative evidence:
- AIPRO mapped to a business domain
- product purpose expressed in decision/capability terms
- accountability statements tied to business outcomes
Supporting evidence:
- product backlog driven by domain priorities
- risk ownership (R0–R4 classification accountability)
- policy decisions owned by domain
Red flags:
- owner email in IT / engineering domain
- product described using model or pipeline terminology
- roadmap tied to platform or model improvements only
How an Agent Should Decide
Decision rule (simplified):
If the designated AIPRO cannot credibly answer “What decision capability does this AI Product provide, and who is accountable if it fails?”, AIPCH01 is not met.
Why AIPCH01 Is Non-Negotiable
Without AIPCH01:
- AI becomes technology-driven, not outcome-driven
- accountability becomes unclear or fragmented
- risk cannot be governed properly
- AI adoption becomes unsafe at scale
AIPCH01 is the foundation of responsible AI Productization.
Canonical Statement (for AIPS)
AIPCH01 is satisfied only when an AI Product is owned by the business domain that is the authoritative owner of the decision capability-accountable for defining its intent, governing its behavior, and bearing the consequences of its outcomes-rather than by technology or platform teams, regardless of who builds or operates the system.