Domain IV: Manage AI Model Development and Evaluation — Comprehensive Study Guide

Exam weight: 16% of PMI-CPMAI exam (~19 scored questions) Score-report framing: ❌ Below Target — PRIORITY 3 for rebuild Maps to CPMAI methodology phases: Phase III (Data Preparation), Phase IV (Model Development), Phase V (Model Evaluation) Number of ECO tasks: 6 (IV.1 through IV.6) — 2 of which are go/no-go gates (IV.5 + IV.6) Estimated study time: 13 hours
Note from docs/ECO_TASK_REFERENCE.md: the score report flagged Task IV.2 (Oversee AI/ML model QA/QC) as having no questions on his form. Cover it anyway — the retake form is randomized.

>

Two of six tasks are explicit go/no-go gates — that's 33% of the domain by task count. Combined with the gate in Domain III (III.8), three gates concentrate ~10-15 exam questions. Master all three.

Overview

Domain IV is the most procedurally complex of the three weak domains. It spans three CPMAI methodology phases (Data Preparation, Model Development, Model Evaluation) and contains two of the three explicit go/no-go gates in the entire ECO. Every task begins with an oversight verb: oversee, manage, verify. The PM is responsible for ensuring that data preparation produces sufficient quality, model technique selection is sound, training is managed, QA/QC standards are upheld, and the model is verified ready before it crosses into Operationalization (Domain V).

The unifying pattern: Domain IV tests whether the project manager can hold the gate. The data scientist wants to keep iterating. The ML engineer wants to keep tuning. The business stakeholder wants to ship. The PM is the one who facilitates the documented decision against documented criteria — and who is willing to call ITERATE or DESCOPE when the criteria aren't met.

Most wrong-answer traps in Domain IV are technically-correct moves that bypass either a gate, an iteration trigger, or a stakeholder decision. The same oversight-verb framing from Domain III applies — and it applies more sharply because Domain IV's two gates have very specific decision criteria.

Table of Contents


Module 1: Data Preparation — Pipelines, Quality, and the Prep Gate

Lessons 1-7 | What data preparation requires, and the gate that decides whether to begin it.

Lesson 1: ECO Task IV.4 — Manage Data Transformation to Conduct Data Preparation

After the data is gathered (III.5) and the Phase II gate (III.8) has decided GO, the team enters Phase III — Data Preparation. The PM's job is to manage the transformation effort: ensure pipelines are built, transformations are documented, quality is preserved, and the prep work feeds the success criteria from Domain II.

The PM does not write transformation code. The PM coordinates the data engineering team, tracks pipeline development, and ensures the prepared dataset is usable before model training begins.

KEY TAKEAWAYS

💡 Memory Aid — TRIM Data Prep

Transform formats, Reconcile inconsistencies, Impute missing values, Map fields. Four core categories of data prep work the PM coordinates.

PM Oversight Angle


Lesson 2: Data Preparation Concepts

Data preparation is the work of making raw data usable for training. It includes:

Most AI projects spend 70-80% of their time on data preparation. Underestimating this is a top reason projects miss deadlines.

KEY TAKEAWAYS


Lesson 3: Data Engineering and Pipelines

Data engineering builds the pipelines that move data from sources through preparation and into the model's training environment. Key pipeline concepts:

The PM coordinates pipeline ownership and ensures the deployment plan (V.1) accounts for pipeline maintenance in production.

KEY TAKEAWAYS


Lesson 4: Data Collection and Ingestion

Collection brings data into the pipeline; ingestion is the technical implementation. Both happen in Phase III but are heavily informed by Phase II decisions (III.3 sources, III.5 gathered data). Common ingestion patterns: API pulls, file drops, database replication, streaming connectors, batch uploads.

PM concerns: ingestion reliability, error handling, data validation at entry, source-side rate limits or licensing.

KEY TAKEAWAYS


Lesson 5: Data Preparation Pipelines

Phase III's signature deliverable is a data preparation pipeline that:

  1. Ingests from identified sources
  2. Validates input format and content
  3. Cleans (remove errors/duplicates)
  4. Transforms (format/schema)
  5. Imputes missing values
  6. Augments if needed
  7. Splits into training/validation/test
  8. Outputs to model training environment

The pipeline is reusable and reproducible — same input + same pipeline = same output. Reproducibility is a governance requirement (V.3).

KEY TAKEAWAYS


Lesson 6: Pipeline Complexity

Real-world pipelines are complex. PMI flags this as a project risk:

The PM doesn't design the pipeline but tracks complexity as a project risk and ensures observability is built in.

KEY TAKEAWAYS


Lesson 7: ECO Task IV.5 — Verify Data Quality (GATE)

The first gate in Domain IV. After data preparation pipelines are built and run, the PM facilitates a verification gate: is the prepared data quality sufficient to proceed with model training?

This is distinct from III.8, which asked "do we have the data and understanding?" — a Phase II close-out gate. IV.5 asks "now that we've prepared the data, is the prepared output of sufficient quality to train on?" — a Phase III close-out gate.

The decision criteria:

The decision has three outcomes (same as III.8): GO (proceed to training), ITERATE (loop back to fix), DESCOPE (reduce model scope to what data supports).

KEY TAKEAWAYS

💡 Memory Aid — QCBVR Gate Criteria

Quality dimensions evaluated, Coverage of required attributes, Bias within tolerance, Volume sufficient, Reproducibility verified. Five checks before training begins.

PM Oversight Angle


Module 2: Model Technique and Selection

Lessons 8-13 | What technique, algorithm, and model approach the project will use.

Lesson 8: ECO Task IV.1 — Oversee AI/ML Model Technique(s)

The PM oversees the team's selection of model technique(s) — the algorithmic approach (supervised/unsupervised/reinforcement learning), the model family (linear, tree-based, neural network, transformer), and any pretrained-model decisions. The PM doesn't pick the technique; the data scientist does. The PM ensures the choice is documented, tied to the AI pattern from Phase I, and aligned with the project's success criteria.

A common exam scenario: the data scientist proposes a complex deep-learning model. The right PM response is rarely "approve" — it's "ensure the choice is documented and justified against the AI pattern, success criteria, and operational constraints (cost, latency, explainability)."

KEY TAKEAWAYS

PM Oversight Angle


Lesson 9: Machine Learning Fundamentals — Algorithm vs Model

Two terms commonly confused on the exam:

You train an algorithm on data to produce a model. The model is what gets deployed.

KEY TAKEAWAYS


Lesson 10: ML Algorithm Basics

ML lets computers learn patterns from data and make predictions. Three high-level categories:

The choice of category depends on the AI pattern and the data available:

KEY TAKEAWAYS


Lesson 11: Pretrained Models, Foundation Models, and GenAI

Modern AI rarely trains from scratch. Three patterns:

Using pretrained / foundation / GenAI models reduces the data needed for training (Phase II decisions reflect this — see III.8 questions about "can you use pretrained models?").

KEY TAKEAWAYS


Lesson 12: Transfer Learning and Third-Party Models

Transfer learning = taking a pretrained model and fine-tuning it on your task-specific data. Saves training time and works with less data than training from scratch. Third-party models — sourced from vendors, open-source repositories, or model marketplaces. Brings a governance question: is the model's training data license-compatible? Is bias measurement available? Is provenance documented?

KEY TAKEAWAYS


Lesson 13: Automated Machine Learning (AutoML)

AutoML automates parts of model development — algorithm selection, hyperparameter tuning, feature engineering, model selection. Reduces the data-science skill barrier.

For the PM: AutoML doesn't remove the need for documented justification (IV.1). The output of AutoML is a chosen technique; it still needs to be documented, evaluated, and gated through IV.5/IV.6.

KEY TAKEAWAYS


Module 3: Model Development and Training

Lessons 14-20 | The development phase — actually building and training the model.

Lesson 14: ECO Task IV.3 — Manage AI/ML Model Training

Once technique is selected (IV.1) and prepared data is gated (IV.5), training begins. The PM manages training — coordinates the team's effort, tracks progress, monitors for issues (training time overruns, loss curves not converging, resource exhaustion), and surfaces blockers.

A specific exam scenario PMI tests: model training has been running 5 days against a planned 2-day window. The data scientist says "one more day should do it." What does the PM do? The right answer is to pause and conduct a structured root-cause review (data, technique, resources, hyperparameters), reassess against project plan, and make a documented decision. NOT "let them keep going" and NOT "switch to a smaller model."

KEY TAKEAWAYS

💡 Memory Aid — DTHR Training Triage

When training overruns: review Data (quality, volume, distribution), Technique (algorithm fit), Hardware/resources, Results so far. Four categories to root-cause before proceeding.

PM Oversight Angle


Lesson 15: AI Model Development Phase Overview

Phase IV — Model Development — is where the team applies the chosen technique to the prepared data to produce a model. The phase is iterative: train, evaluate, adjust, retrain. Multiple iterations are normal; a "one-shot training run" is rare.

The PM ensures iterations are tracked, lessons are captured per iteration, and the cumulative time/resource cost stays within budget.

KEY TAKEAWAYS


Lesson 16: Model Validation

Validation is the practice of testing the model on data it didn't see during training. Common approach: split the prepared dataset into training (~70%), validation (~15%, used to tune hyperparameters), and test (~15%, used for final unbiased evaluation).

The PM ensures validation is performed and results are documented before declaring training complete.

KEY TAKEAWAYS


Lesson 17: Generalizing to New Data

The goal of training is generalization — performing well on data the model hasn't seen. Two failure modes:

Both are technical problems the data scientist addresses, but the PM tracks them as project risks and ensures evaluation reports include them.

KEY TAKEAWAYS


Lesson 18: Building GenAI Systems

GenAI systems differ from traditional ML in development:

The PM coordinates GenAI development against the same technique-selection (IV.1), training-management (IV.3), and gate (IV.6) framework — but recognizes the work is more about prompt engineering, retrieval design, and evaluation criteria than traditional model building.

KEY TAKEAWAYS


Lesson 19: Retrieval-Augmented Generation (RAG)

RAG enhances a foundation model by retrieving relevant context at inference time and feeding it into the prompt. The model's output is grounded in retrieved documents rather than pure parametric memory.

When to use: when the foundation model needs domain-specific or current information that wasn't in its training data. Example: answering customer questions from your product documentation.

KEY TAKEAWAYS


Lesson 20: Fine-Tuning LLMs

Fine-tuning adjusts a foundation model's weights using task-specific data, producing a custom model that performs better on your task than the base model.

When to use: when prompting and RAG aren't sufficient; when task-specific patterns need to be learned; when the volume of task data is sufficient (typically thousands of examples minimum).

When NOT to use: small data, generic tasks, when prompting suffices, when RAG suffices, when latency is critical.

KEY TAKEAWAYS


Module 4: Model QA/QC, Evaluation, and Iteration

Lessons 21-31 | The QA/QC and evaluation discipline that catches model defects before deployment.

Lesson 21: ECO Task IV.2 — Oversee AI/ML Model QA/QC

QA/QC = configuration management + model performance verification. The PM oversees quality assurance practices throughout development:

Asterisked task: the first attempt form had no IV.2 questions. The retake form may differ. Cover it.

KEY TAKEAWAYS

PM Oversight Angle


Lesson 22: Why Model Evaluation Matters

Model evaluation answers "is the model good enough to ship?" Without evaluation, you have no objective basis for the IV.6 gate decision. PMI's framing: model evaluation is a discipline, not a step — done continuously during development, not just at the end.

KEY TAKEAWAYS


Lesson 23: When Model Evaluation Falls Short

Consequences of inadequate evaluation:

PMI's exam frequently tests recognition of "evaluation gap" scenarios — the wrong answer is usually "deploy and observe; we'll catch issues in production."

KEY TAKEAWAYS


Lesson 24: How to Evaluate a Model Effectively

Effective evaluation answers structured questions:

KEY TAKEAWAYS


Lesson 25: Model Iteration — Why and When

Model iteration is the practice of repeatedly training, evaluating, adjusting, and retraining. Reasons to iterate:

The PM tracks iteration count, cumulative time, and remaining budget. Iteration is normal; runaway iteration without convergence is a project risk.

KEY TAKEAWAYS


Lesson 26: When to Retrain the Model

Triggers for retraining:

The retraining decision is PM-coordinated with stakeholders, not data-scientist-unilateral.

KEY TAKEAWAYS


Lesson 27: Data Drift and Model Drift

(Same concepts that surface in Domain V monitoring — Domain IV is where the response capability is built.)

Both are detected through monitoring (V.4), but the response (retrain, recalibrate, replace) is built into the model life cycle plan from Phase V.

KEY TAKEAWAYS


Lesson 28: KPIs — Business and Technical

Two KPI tiers must align:

Technical KPIs that don't translate to business KPIs are vanity metrics. Business KPIs without technical KPI underpinning are unmeasurable. Both must be defined, tied to Domain II success criteria, and tracked throughout development and operations.

KEY TAKEAWAYS


Lesson 29: Audit Trails and Auditability

AI audit trails document the full path from data collection through model training to deployment to inference outputs. Why they matter:

Audit trails should capture: input data, model version, prediction, timestamp, decision rationale, human-in-the-loop overrides.

KEY TAKEAWAYS


Lesson 30: AI Transparency

Two distinct transparency concepts:

Systemic transparency is achievable for most models. Decision transparency is hard — many modern models (especially deep learning) are "black boxes."

KEY TAKEAWAYS


Lesson 31: Explainability vs Interpretability

Often used interchangeably, technically distinct:

For high-stakes decisions (healthcare, finance, legal), interpretability is preferred. For low-stakes decisions (recommendations), explainability post-hoc may suffice.

Not all algorithms can be fully explained — deep learning is famously a black box. The trade-off between performance and explainability is a project decision tied to V.3 (governance) and Domain I (Trustworthy AI).

KEY TAKEAWAYS


Module 5: The Operationalization Gate and Phase V Closeout

Lessons 32-36 | The IV.6 gate and the closeout of Phase V (Model Evaluation) before transition to Domain V.

Lesson 32: ECO Task IV.6 — Verify Model Ready for Operationalization (GATE)

The second gate in Domain IV. After model is trained (IV.3), QA/QC'd (IV.2), and evaluated, the PM facilitates the operationalization-readiness gate. This is the gate that authorizes the project to enter Domain V.

Decision criteria:

Three outcomes (same pattern): GO (proceed to Domain V deployment), ITERATE (loop back to address), DESCOPE (reduce model scope or capabilities).

KEY TAKEAWAYS

💡 Memory Aid — PBRBARTAO Gate Criteria

Performance vs criteria, Bias within tolerance, Robustness to edge cases, Baseline-comparison favorable, Audit trail complete, Reproducibility verified, Trustworthy AI aligned, Operational fit confirmed. Eight checks before the model crosses into production.

PM Oversight Angle


Lesson 33: Phase V — Preparing for Deployment / Model Readiness

Once IV.6 = GO, the project transitions to Domain V (operationalization). The "deployment readiness" deliverables include: trained model, audit trail, reproducible pipeline, monitoring plan, deployment plan (which V.1 builds), governance plan (which V.3 builds), contingency plan (which V.7 builds).

KEY TAKEAWAYS


Lesson 34: Phase V — Planning for Improvement (Iteration Plan)

Even after deployment, the model will need to improve. The iteration plan (built before deployment, executed throughout production) covers:

The iteration plan is part of the deployment plan (V.1) and is monitored through V.4.

KEY TAKEAWAYS


Lesson 35: Iterating Back to Previous CPMAI Phases

Phase V findings can trigger iteration back to earlier phases (same pattern as Domain III's 12 iteration triggers). Common triggers from Phase V:

KEY TAKEAWAYS


Lesson 36: Phase IV Go/No-Go (General Closeout)

Beyond IV.5 and IV.6 (the explicit ECO gates), Phase IV has a general closeout: confirm all Phase IV objectives are met, all artifacts are documented, all decisions are traceable. This isn't a separate ECO task but is part of how the PM tracks Phase IV completion.

KEY TAKEAWAYS


Quick Reference: The Two Gates (IV.5 + IV.6)

IV.5 — Data Quality GateIV.6 — Operationalization Gate
WhenAfter data preparation pipelines runAfter model is trained, QA/QC'd, evaluated
QuestionIs prepared data quality sufficient to train on?Is the model ready to operate in production?
Maps to phaseEnd of Phase III (Data Preparation)End of Phase V (Model Evaluation)
Decision criteriaQuality dimensions (ACCTUVI), coverage, bias, volume, reproducibilityPerformance, bias, robustness, baseline, audit, reproducibility, trustworthy-AI, operational fit
OutcomesGO / ITERATE / DESCOPEGO / ITERATE / DESCOPE
What happens on GOProceed to model training (IV.3)Proceed to Domain V deployment (V.1+)
Both gates are PM-facilitated, stakeholder-engaged, documented decisions. Wrong-answer traps are always: PM decides alone, technical workaround applied, or "proceed and fix later."

Quick Reference: Model Evaluation Checklist (IV.2 + IV.6)

CheckWhy
Performance vs success criteria (II.8)Is model good enough by Domain II definition?
Performance across user segmentsBias / fairness check
Edge case coverageDoes it work at distribution boundaries?
Failure mode analysisHow does it fail when it fails?
Comparison vs baselineDoes AI beat rules / prior model / human?
Bias measurementInformational bias within tolerance?
ReproducibilityCan the training be rerun and produce the same model?
Audit trailFull data → training → evaluation documentation?
Trustworthy AIPrivacy / security / transparency / governance / ethics aligned?
Operational fitWill this technique run in the planned production environment?

Cross-Domain Links


Knowledge Check

Question 1

Data preparation pipelines are complete and the data engineer reports the data is ready for training. The PM is asked to authorize the start of training. What's the BEST move?

A. Authorize training to proceed

B. Run the IV.5 verification gate — quality dimensions, coverage, bias, volume, reproducibility — with stakeholders before authorizing training

C. Have the data scientist start training in parallel with the gate review

D. Defer the gate until after a few training iterations show whether the data is good enough

Click for answer and rationale Correct: B

ECO Task IV.5 is the data-quality gate. The PM facilitates a documented stakeholder decision before training begins.

  • A wrong: Skips the gate.
  • C wrong: Wrong-answer trap — parallel work bypasses the gate purpose.
  • D wrong: Backwards — the gate exists to prevent wasting training cycles on inadequate data.

Question 2

Model training has been running for 5 days against a planned 2-day window. The data scientist says one more day should do it. What should the PM do?

A. Allow another day since they're close

B. Pause training, conduct structured review of root cause (data, technique, resources, results), reassess against project plan, and make a documented decision on whether to continue, change approach, or escalate

C. Have them switch to a smaller model immediately

D. Cancel and restart from scratch

Click for answer and rationale Correct: B

2.5x time overrun = project event, not technical hiccup. ECO Task IV.3 — manage training. Pause + DTHR root-cause + documented decision.

  • A wrong: Lets the overrun continue without analysis.
  • C wrong: Wrong-answer trap — switching technique without IV.1 review is governance bypass.
  • D wrong: Restart without root-cause throws away learnings.

Question 3

The team has completed model training, QA/QC, and evaluation. The data scientist proposes deploying to production. What should the PM do?

A. Authorize deployment

B. Run the IV.6 operationalization-readiness gate with stakeholders, evaluating performance, bias, robustness, baseline, audit, reproducibility, trustworthy-AI alignment, and operational fit

C. Have the ML engineer start deployment while the gate is being scheduled

D. Defer deployment until production observes actual performance

Click for answer and rationale Correct: B

ECO Task IV.6 — the operationalization gate. 8 criteria, stakeholder-engaged, documented decision. Required before Domain V begins.

  • A wrong: Skips the gate.
  • C wrong: Wrong-answer trap — parallel work bypasses the gate.
  • D wrong: Production isn't the evaluation venue — pre-deployment evaluation is.

Question 4

True or False: ECO Tasks III.8 and IV.5 are the same gate.

Click for answer and rationale Correct: FALSE

They're distinct gates at adjacent phase boundaries:

  • III.8 = end of Phase II (Data Understanding). Question: "Do we have the data and understanding to proceed?"
  • IV.5 = end of Phase III (Data Preparation). Question: "Is the prepared data sufficient to train on?"

Both are go/no-go gates with the same outcome structure (GO/ITERATE/DESCOPE), but they evaluate different artifacts at different stages.

Question 5

The data scientist proposes a deep learning model for a high-stakes medical-imaging classification task. The healthcare client requires that AI decisions be explainable. What's the PM's BEST response?

A. Approve the deep learning approach since it offers higher accuracy

B. Document the technique selection and ensure trade-off between performance and explainability is presented to stakeholders for decision; consider interpretable-by-design alternatives

C. Have the data scientist proceed and add post-hoc XAI explanations after training

D. Reject deep learning and require an interpretable model

Click for answer and rationale Correct: B

ECO Task IV.1 (technique oversight) + Domain I (Trustworthy AI). High-stakes + explainability requirement = stakeholder decision. PM facilitates the trade-off discussion, not unilateral approval or rejection.

  • A wrong: Approves without surfacing the explainability constraint.
  • C wrong: Post-hoc XAI may not satisfy "explainable AI" requirement for high-stakes regulated decisions.
  • D wrong: Wrong-answer trap — unilateral PM rejection isn't a stakeholder-engaged decision either.

Question 6

During QA/QC of a recommendation model, the team finds that recommendations show measurable demographic bias. What should the PM do?

A. Have the data scientist add a fairness post-processing layer

B. Treat as an ECO IV.2 + Domain I (Task I.3 — bias checks) issue: document the finding, escalate per accountability procedures, engage stakeholders for remediation decision, do not authorize IV.6 GO until bias is within tolerance

C. Deploy with a "monitor closely" flag and address bias in production

D. Reject the entire model and start over

Click for answer and rationale Correct: B

ECO IV.2 (QA/QC) + Domain I (Trustworthy AI Task 3) intersect. Bias requires documented escalation and remediation, blocking IV.6 GO until resolved.

  • A wrong: Wrong-answer trap — technical post-processing without governance / stakeholder engagement.
  • C wrong: Production monitoring of known bias is not a remediation strategy.
  • D wrong: Restart without root-cause may repeat the same issue.

Question 7

A team is iterating a model with 4 training runs over 2 weeks, each one improving slightly but not meeting the success criteria. The data scientist suggests a 5th iteration. What should the PM do?

A. Approve the 5th iteration

B. Pause and review the iteration trajectory: are improvements converging or plateauing? Is the technique a fit? Is the data sufficient? Document a decision: continue, change technique, descope, or escalate.

C. Have the data scientist try a different algorithm

D. Cancel the project

Click for answer and rationale Correct: B

Runaway iteration without convergence is a PM-tracked project risk. Pausing for structured review prevents endless iteration.

  • A wrong: Approves without review.
  • C wrong: Wrong-answer trap — algorithm change without root-cause is governance bypass.
  • D wrong: Cancellation may be right, but only after structured review.

Question 8

During the IV.6 gate review, the team confirms model performance meets success criteria but notes the chosen technique requires GPU compute that isn't available in the planned cloud production environment. What should the PM do?

A. Authorize GO and procure GPU compute concurrently

B. Treat as an operational-fit failure of IV.6 — documented ITERATE outcome. Loop back to V.1 (deployment plan) to address infrastructure OR loop back to IV.1 (technique) to choose a model that fits the planned environment.

C. Have the ML engineer optimize the model for CPU

D. Deploy to GPU on a different cloud while the team works the issue

Click for answer and rationale Correct: B

IV.6's 8 criteria include "operational fit." A failure on operational fit blocks GO. Two valid loops: V.1 to add infrastructure, or IV.1 to re-select technique.

  • A wrong: Skips the gate's operational-fit check.
  • C wrong: Wrong-answer trap — technical workaround without IV.1 documentation.
  • D wrong: Different-cloud workaround is unilateral architectural change.

Question 9

True or False: An AutoML pipeline that automatically selects an algorithm and tunes hyperparameters bypasses the need for ECO Task IV.1 documentation.

Click for answer and rationale Correct: FALSE

AutoML automates the technical selection but doesn't replace the governance requirement. The PM still needs the chosen technique documented, justified against AI pattern + success criteria, and aligned with operational constraints. AutoML's output feeds IV.1 documentation; it doesn't bypass it.

Question 10

The team has completed Phase IV. The data scientist asks whether to begin operationalization (Domain V) work in parallel with the IV.6 gate. What should the PM do?

A. Approve parallel work to save time

B. Confirm IV.6 must complete before Domain V work begins; sequential, not parallel; gate authorizes the transition

C. Have the ML engineer prepare deployment artifacts but not actually deploy until IV.6 GO

D. Defer the IV.6 gate and let Domain V work proceed

Click for answer and rationale Correct: B

ECO IV.6 is a gate. Gates are sequential checkpoints — Domain V work doesn't begin until IV.6 = GO. Treating gates as parallel-able defeats their purpose.

  • A wrong: Wrong-answer trap — "save time" rationalization.
  • C wrong: Preparing deployment artifacts is Domain V (V.1) work and shouldn't proceed pre-gate.
  • D wrong: Deferring the gate while letting downstream work proceed is gate-bypass.


Memory Aids & Mnemonics Summary

MnemonicWhat to Remember
TRIM (Data Prep)Transform, Reconcile, Impute, Map
QCBVR (IV.5 Gate)Quality, Coverage, Bias, Volume, Reproducibility
DTHR (Training Triage)Data, Technique, Hardware, Results — review when training overruns
PBRBARTAO (IV.6 Gate)Performance, Bias, Robustness, Baseline, Audit, Reproducibility, Trustworthy AI, Operational fit
3 ML CategoriesSupervised, Unsupervised, Reinforcement
Algorithm vs ModelAlgorithm = procedure. Model = trained artifact.
Overfit vs UnderfitOverfit = memorize, fail on new. Underfit = doesn't learn even on training.
Pretrained / Foundation / GenAIPretrained = adapt for your task. Foundation = very large pretrained. GenAI = generates new content.
Transfer LearningPretrained + fine-tune on your task data
Systemic vs Decision TransparencySystemic = how built. Decision = why this prediction.
XAI vs InterpretabilityXAI = post-hoc explain any model. Interpretability = inherently understandable.
3 GatesIII.8 (data ↔ needs) · IV.5 (prepared data quality) · IV.6 (model ↔ ops). All three: GO/ITERATE/DESCOPE.

Closing reminders for Domain IV


Next: domain-I-trustworthy-ai.md (Domain I — Responsible & Trustworthy AI Efforts, 15% weight, full guide)