Protected

NCP-ADS module content is available after admin verification. Redirecting…

If you are not redirected, login.

Access

Admin only

NCP-ADS module pages are restricted to admin users.

Training / NCP-ADS

MLOps

Module study guide

Priority 5 of 6 · Domain 6 in exam order

Scope

Exam study content

This module contains expanded study notes, practical drills, and an exam-style question set.

Exam weight
10%
Priority tier
Tier 3
Why this domain
Deployment and inference operations including Triton-oriented concepts.

Exam Framework

How to reason under pressure

1. Stabilize Before Optimizing

  • Verify hardware and management-plane integrity first.
  • Confirm firmware/software baseline consistency.
  • Only then run performance tuning decisions.

2. Single-Variable Changes

  • Change one parameter at a time when investigating regressions.
  • Use before/after evidence with constant workload input.
  • Discard changes without reproducible benefit.

Exam Scope Coverage

What this module now covers

This module covers production lifecycle fundamentals for accelerated ML systems: model packaging, deployment with Triton-oriented concepts, monitoring, reliability operations, and controlled release practices.

Track 1: Reproducible model packaging and versioning

Operational success begins with reproducible artifacts and traceable versions.

  • Bundle model, preprocessing contract, and runtime dependencies together.
  • Track model, data, and configuration versions so outputs remain auditable.
  • Use explicit artifact promotion stages (dev, staging, prod).

Drill: Create a release artifact checklist covering model weights, preprocessing schema, and runtime metadata.

Track 2: Triton serving fundamentals

NCP-ADS scope explicitly includes Triton-oriented deployment concepts.

  • Understand Triton model repository layout and backend choices.
  • Configure model-specific inference settings via model configuration files.
  • Use deployment patterns that allow safe updates without service disruption.

Drill: Stand up one Triton-served model and validate request/response behavior with defined input and output contracts.

Track 3: Inference performance tuning

Serving quality depends on balancing throughput, latency, and resource limits.

  • Dynamic batching and instance-group sizing change latency-throughput tradeoffs.
  • Response caching can reduce repeated inference cost for repeated queries.
  • Optimization decisions must be validated under representative traffic, not synthetic micro-tests only.

Drill: Run two serving configurations and compare p95 latency, throughput, and GPU utilization.

Track 4: Observability and model monitoring

You cannot operate ML services safely without system and model-level telemetry.

  • Track service metrics (latency, throughput, errors) and model metrics (drift, quality proxy signals).
  • Use dashboards and alert thresholds tied to SLO and incident response expectations.
  • Separate infrastructure failures from data or model behavior changes during triage.

Drill: Define one monitoring dashboard with service, data, and model health sections plus alert thresholds.

Track 5: Release strategies and rollback

Controlled rollouts reduce blast radius when model behavior changes unexpectedly.

  • Canary and shadow deployments provide early safety signals before full rollout.
  • Rollback criteria must be explicit and connected to monitored metrics.
  • Release notes should document feature, data, and configuration changes.

Drill: Create a rollout runbook with canary gates, rollback triggers, and on-call ownership.

Track 6: Security, governance, and compliance

Production ML systems must address access control, data protection, and operational governance.

  • Apply least-privilege access to model repositories, serving endpoints, and secrets.
  • Protect sensitive data and model artifacts in transit and at rest.
  • Audit logs and version history support compliance and incident forensics.

Drill: Perform a lightweight security review for one inference service and capture remediation actions.

Concept Explanations

Deep-dive concept library

Exam Decision Hierarchy

Prioritize decisions in this order: safety and hardware integrity, baseline consistency, controlled validation, then optimization.

  • If integrity checks fail, stop optimization and remediate first.
  • Compare against known-good baseline before changing multiple variables.
  • Document rationale for each decision to support incident replay.

Operational Evidence Standard

Treat every key action as evidence-producing: command, output, timestamp, and expected vs observed behavior.

  • Evidence should be reproducible by another engineer.
  • Use stable command templates for repeated environments.
  • Keep concise but complete validation artifacts for exam-style reasoning.

Train-serve contract integrity

Serving reliability depends on keeping preprocessing, schema, and model versions synchronized across training and inference.

  • Package preprocessing contract with model artifact, not separately in tribal knowledge.
  • Version model, data snapshot, and runtime configuration together.
  • Validate contract parity in staging before rollout.

SLO-driven serving optimization

Inference tuning should optimize toward target SLO (for example p95 latency) rather than raw throughput only.

  • Dynamic batching and instance groups trade latency and throughput differently.
  • Traffic shape and request size mix matter more than microbenchmark best case.
  • Use representative load tests and canary signals for final config choice.

Release governance and rollback readiness

MLOps maturity is demonstrated by safe rollout strategy and fast rollback criteria tied to observable signals.

  • Define canary gates before deployment starts.
  • Preserve previous model version for rapid fallback.
  • Tie rollback triggers to latency, error, and model-quality indicators.

Scenario Playbooks

Exam-style scenario explanations

Scenario A: New model deployment increases p95 latency and error spikes

A newly deployed Triton model version passes smoke tests but violates latency SLO and raises error rate in production canary traffic.

Architecture Diagram

[Client] -> [Gateway] -> [Triton Inference Service]
                          |
                     [Model Repository]
                          |
                 [Metrics + Alerting]

Response Flow

  1. Confirm model config changes (batching, instance count, backend) against previous version.
  2. Run targeted load profile and inspect p95, queueing, and GPU utilization.
  3. Rollback if canary gates fail and open postmortem with config diff.

Success Signals

  • Canary decision is made from predefined SLO gates.
  • Rollback completes quickly with service stability restored.
  • Root cause is captured as config or artifact mismatch, not guesswork.

Triton metrics endpoint check

curl -s http://localhost:8002/metrics | grep -E 'nv_inference|request_duration' | head

Expected output (example)

nv_inference_request_success ...
nv_inference_queue_duration_us ...

Scenario B: Offline metrics strong but online quality drifts after release

Model performs well offline, but online proxy metrics indicate drift and user complaints increase.

Architecture Diagram

[Feature Stream] -> [Inference Endpoint] -> [Prediction Logs]
        |                                  |
 [Drift/Proxy Checks]               [Alert + Triage]

Response Flow

  1. Compare online feature distributions against training baseline.
  2. Audit preprocessing contract version used by serving endpoint.
  3. Route degraded traffic to previous model while investigating drift source.

Success Signals

  • Drift detection confirms whether issue is data shift or deployment mismatch.
  • Rollback/mitigation path is executed without prolonged outage.
  • Monitoring dashboards reflect restored quality indicators.

CLI and Commands

High-yield command runbooks

CLI Execution Pattern

  • 1. Capture baseline state before running any intrusive command.
  • 2. Execute command with explicit scope (node, interface, GPU set).
  • 3. Compare output against expected baseline signature.
  • 4. Record timestamp and decision (pass, investigate, remediate).

Triton deployment sanity runbook

Validate model repository and serving endpoint behavior before scaling traffic.

Start Triton with model repository

tritonserver --model-repository=/models

Expected output (example)

Triton server starts and reports loaded model versions.

Health and readiness checks

curl -s http://localhost:8000/v2/health/ready && echo

Expected output (example)

Returns HTTP 200 with readiness status.
  • Run readiness checks before any performance test.
  • Capture server logs and loaded model list in release artifact.

Serving performance and rollback runbook

Measure SLO metrics under load and execute rollback when gates fail.

Triton perf analyzer baseline

perf_analyzer -m my_model --concurrency-range 1:16 --measurement-interval 5000 --percentile=95

Expected output (example)

Reports throughput and latency percentiles for each concurrency level.

Model version rollback signal check

curl -s http://localhost:8000/v2/models/my_model/config

Expected output (example)

Shows active model config/version after rollback action.
  • Define fail gates before running perf_analyzer.
  • Rollback should be scripted to avoid delay during incidents.

Common Problems

Failure patterns and fixes

Train-serve mismatch from unversioned preprocessing

Symptoms

  • Offline evaluation is healthy but online predictions are inconsistent.
  • Feature schema differences appear between training and serving logs.
  • No clear mapping from model version to preprocessing artifact.

Likely Cause

Model artifact was promoted without pinned preprocessing contract and version metadata.

Remediation

  • Bundle preprocessing artifact with model package and version both together.
  • Add schema validation at inference boundary.
  • Block promotion when artifact lineage is incomplete.

Prevention: Require version-linked model + preprocessing contract for every release.

Latency SLO violations after tuning for throughput only

Symptoms

  • Throughput increases but p95/p99 latency worsens.
  • Queue duration grows during burst traffic.
  • Canary tests pass only under light synthetic load.

Likely Cause

Batching/concurrency settings optimized for average throughput without tail-latency guardrails.

Remediation

  • Tune with SLO-focused objective using representative traffic mix.
  • Adjust dynamic batching windows and instance groups iteratively.
  • Add canary gates on p95/p99 and rollback quickly when violated.

Prevention: Use SLO-first tuning policy and require tail-latency metrics in release approvals.

Lab Walkthroughs

Step-by-step execution guides

Walkthrough A: Deploy and validate a Triton model end-to-end

Stand up model serving with health, schema, and baseline performance checks.

Prerequisites

  • Model artifact exported in Triton-compatible format.
  • Triton server installed or containerized.
  • Sample inference payloads ready.
  1. Place model in repository structure and start Triton server.

    tritonserver --model-repository=/models

    Expected: Server logs show model loaded without configuration errors.

  2. Run readiness and sample inference checks.

    curl -s http://localhost:8000/v2/health/ready

    Expected: Endpoint returns ready status and sample inference succeeds.

  3. Capture baseline latency and throughput with perf analyzer.

    Expected: Baseline SLO metrics are documented for future change comparisons.

Success Criteria

  • Serving endpoint passes readiness and contract checks.
  • Baseline latency/throughput report is stored with release notes.
  • Model version and preprocessing contract are traceable.

Walkthrough B: Canary rollout with rollback drill

Practice controlled release with objective rollback triggers.

Prerequisites

  • Two model versions available (stable and candidate).
  • Canary traffic-routing capability.
  • Monitoring dashboard with latency, error, and quality proxy metrics.
  1. Route small traffic slice to candidate model and monitor canary gates.

    Expected: Canary health metrics are visible in near real time.

  2. Trigger rollback when threshold breach is detected.

    Expected: Traffic shifts back to stable model quickly with minimal impact.

  3. Publish incident summary with config and metric deltas.

    Expected: Team has concrete evidence and next action list.

Success Criteria

  • Rollback procedure executes within target response time.
  • Monitoring captures before/after canary and rollback metrics.
  • Postmortem includes root-cause hypothesis and preventive controls.

Study Sprint

10-day execution plan

Day Focus Output
1 Define artifact/versioning contract for model, data, and config. Versioning and promotion policy document.
2 Set up Triton model repository and baseline deployment. Running inference endpoint with validated schema contract.
3 Configure serving parameters (batching, concurrency, instances). Baseline serving configuration and benchmark snapshot.
4 Performance tuning pass with controlled traffic profiles. Latency/throughput tradeoff table and recommendation.
5 Enable service-level metrics and alerts. Observability dashboard with SLO thresholds.
6 Add model/data quality monitors and drift checks. Monitoring playbook for model behavior changes.
7 Design canary and rollback process. Release runbook with approval and rollback gates.
8 Security and access-control review. Risk register and prioritized remediation list.
9 Incident simulation for latency spike or quality regression. Post-incident report template and triage checklist.
10 Exam-style operations scenario rehearsal. Final MLOps quick-reference sheet.

Hands-on Labs

Practical module work

Each lab includes a collapsed execution sample with representative CLI usage and expected output.

Lab A: Triton deployment baseline

Deploy one model end-to-end using Triton repository conventions.

  • Package model artifact and configuration into repository structure.
  • Run local or staged Triton server and validate inference contract.
  • Capture baseline latency and throughput metrics.
Execution Sample (Collapsed)
  1. Capture baseline state for the target node/group before changes.
  2. Run scoped validation command for this lab objective.
  3. Compare observed output against expected signature.

Sample Command (Triton deployment sanity runbook)

tritonserver --model-repository=/models

Expected output (example)

Triton server starts and reports loaded model versions.

Lab B: Inference optimization lab

Tune serving configuration for target latency-throughput profile.

  • Compare at least two dynamic batching or instance-group settings.
  • Measure p50/p95 latency, throughput, and GPU utilization.
  • Choose configuration based on explicit SLO objective.
Execution Sample (Collapsed)
  1. Capture baseline state for the target node/group before changes.
  2. Run scoped validation command for this lab objective.
  3. Compare observed output against expected signature.

Sample Command (Triton deployment sanity runbook)

curl -s http://localhost:8000/v2/health/ready && echo

Expected output (example)

Returns HTTP 200 with readiness status.

Lab C: Monitoring and drift readiness

Operationalize monitoring for both service and model behavior.

  • Define alert thresholds for latency, error rate, and saturation.
  • Add model-quality or drift proxy checks where labels are delayed.
  • Test alert routing and runbook links.
Execution Sample (Collapsed)
  1. Capture baseline state for the target node/group before changes.
  2. Run scoped validation command for this lab objective.
  3. Compare observed output against expected signature.

Sample Command (Serving performance and rollback runbook)

perf_analyzer -m my_model --concurrency-range 1:16 --measurement-interval 5000 --percentile=95

Expected output (example)

Reports throughput and latency percentiles for each concurrency level.

Lab D: Safe rollout and rollback

Practice low-risk production change management.

  • Design canary criteria for automated or manual promotion.
  • Trigger a controlled rollback on defined failure signal.
  • Write post-release validation checklist.
Execution Sample (Collapsed)
  1. Capture baseline state for the target node/group before changes.
  2. Run scoped validation command for this lab objective.
  3. Compare observed output against expected signature.

Sample Command (Serving performance and rollback runbook)

curl -s http://localhost:8000/v2/models/my_model/config

Expected output (example)

Shows active model config/version after rollback action.

Exam Pitfalls

Common failure patterns

  • Deploying models without version-linked preprocessing contracts.
  • Optimizing throughput while ignoring tail-latency SLO impact.
  • Relying on infrastructure metrics only and missing model-quality drift.
  • Rolling out to 100% traffic without canary or rollback criteria.
  • Treating benchmark micro-tests as production performance evidence.
  • Skipping access-control and audit logging on inference services.

Practice Set

Domain checkpoint questions

Attempt each question first, then open the answer and explanation.

Q1. Why is a versioned preprocessing contract important in MLOps?
  • A. It removes the need for training data
  • B. It prevents silent train-serve feature mismatch
  • C. It guarantees zero latency
  • D. It only helps notebooks

Answer: B

Versioned preprocessing keeps online inference transformations aligned with training expectations.

Q2. What does Triton dynamic batching primarily trade off?
  • A. Security and encryption settings
  • B. Throughput and latency behavior
  • C. Only model accuracy
  • D. Dataset labeling speed

Answer: B

Batching can improve throughput but may increase waiting time and tail latency if misconfigured.

Q3. Which rollout strategy best limits blast radius for new model versions?
  • A. Immediate full-traffic cutover
  • B. Canary rollout with monitored gates
  • C. Disable monitoring during rollout
  • D. Delete previous model version first

Answer: B

Canary rollout exposes a smaller traffic share first, enabling safer validation before full promotion.

Q4. Why is response caching useful for some inference workloads?
  • A. It replaces model serving
  • B. It can reduce repeated inference computation for repeated requests
  • C. It guarantees higher accuracy
  • D. It removes need for GPUs

Answer: B

When identical requests recur, caching can lower latency and compute consumption.

Q5. Which metric set is most useful for serving SLO monitoring?
  • A. Parameter count only
  • B. Latency percentiles, throughput, and error rate
  • C. Training epochs only
  • D. Number of feature columns only

Answer: B

Service reliability depends on latency, load handling, and correctness signals in production.

Q6. What is a key reason to keep previous model versions available?
  • A. To increase storage costs only
  • B. To support controlled rollback when regressions appear
  • C. To disable monitoring
  • D. To avoid CI/CD pipelines

Answer: B

Version retention enables fast rollback and operational continuity after bad deployments.

Q7. When labels are delayed in production, what helps monitor model quality?
  • A. Ignore quality monitoring
  • B. Use drift and proxy indicators until labeled feedback arrives
  • C. Track only GPU temperature
  • D. Remove alerting

Answer: B

Proxy metrics and drift signals help detect behavior changes before full label feedback is available.

Q8. Which statement best reflects production-ready benchmarking?
  • A. One synthetic run is enough
  • B. Evaluate under representative traffic and workload patterns
  • C. Ignore failure cases
  • D. Use only best-case numbers

Answer: B

Representative traffic tests reveal realistic latency, throughput, and stability behavior.

Q9. What does least-privilege access mean for ML serving?
  • A. Everyone gets admin rights for speed
  • B. Grant only the minimum permissions required for each role
  • C. Disable authentication
  • D. Share all secrets openly

Answer: B

Least-privilege reduces blast radius and improves security posture.

Q10. Why should rollback criteria be defined before release?
  • A. To avoid making any changes
  • B. To ensure fast, objective response when quality or SLO signals degrade
  • C. To remove observability
  • D. To force manual deployment only

Answer: B

Predefined rollback gates reduce ambiguity and speed incident containment.

Primary References

Curated from official documentation and high-signal references.

Objectives

  1. 6.1 Determine optimal data type choices for each feature.
  2. 6.2 Assess and verify dataset memory footprint.
  3. 6.3 Compare required memory against available device memory.
  4. 6.4 Benchmark and optimize GPU-accelerated workflows.
  5. 6.5 Deploy and monitor models in production.

Navigation