Protected

NCA-GENL flashcards are available after login. Redirecting...

If you are not redirected, login.

Courses / Nvidia / NCA-GENL / Self-Test

NCA-GENL Flashcards

Concept recall cards across all chapters

82 concept cards ยท generated from chapter scope explanations

Study Method

  1. Read the term only, then state the definition from memory.
  2. Open the answer and score yourself: correct, partial, or missed.
  3. Revisit missed cards after chapter review and repeat until stable recall.

Chapter 1: Foundations of Generative AI and Deep Learning

11 cards
Generative vs Discriminative models

Generative models create new outputs from learned distributions, while discriminative models classify or score existing inputs.

Supervised vs Unsupervised vs Self-supervised learning

Supervised uses labels, unsupervised finds structure, and self-supervised builds labels from raw data (the core of LLM pretraining).

Representation learning

The model learns useful internal features (embeddings/hidden states) that can transfer across many tasks.

Foundation models

Broad pretrained models that can be adapted with prompting, RAG, or fine-tuning for specific use cases.

Scaling laws

Performance generally improves with more parameters, data, and compute, but with cost and diminishing-return tradeoffs.

Transfer learning

Reusing pretrained model knowledge to reduce training cost/time for downstream tasks.

Neural networks (MLP basics)

Stacked linear and nonlinear layers transform inputs into progressively richer features.

Activation functions (ReLU, GELU, Sigmoid, Tanh)

Nonlinearities that let networks model complex patterns; GELU/ReLU are common in modern transformer stacks.

Loss functions (Cross-Entropy, MSE)

Cross-entropy is typical for token prediction/classification; MSE is common for regression targets.

Backpropagation

Algorithm that computes gradients of loss with respect to model parameters.

Gradient descent

Optimi

Chapter 2: Transformer Architecture and LLM Mechanics

6 cards
Self-attention

Lets each token weigh relevance of other tokens in the sequence.

Multi-head attention

Uses multiple attention projections so the model can learn different relation types in parallel.

Positional encoding

Adds token-order information because attention alone is permutation-invariant.

Encoder-only vs Decoder-only vs Encoder-Decoder

Encoder-only is strong for understanding, decoder-only for generation, and encoder-decoder for input-to-output transformation.

Feed-forward blocks

Per-token nonlinear layers that refine representations after attention.

Residual connections

Skip paths that preserve signal and stabili

Chapter 3: Training Large Language Models

7 cards
Pretraining

Large-scale self-supervised training to build general language capability.

Fine-tuning

Additional training to adapt the base model to a target task/domain.

Supervised fine-tuning (SFT)

Fine-tuning with labeled prompt-response examples.

Instruction tuning

Trains the model to better follow user instructions and response formats.

Transfer learning

Reuse pretrained knowledge instead of training from scratch.

Dataset curation

Selecting, cleaning, balancing, and deduplicating training data.

Data preprocessing

Normali

Chapter 4: Prompt Engineering and Inference Strategies

0 cards

Chapter 5: Retrieval-Augmented Generation (RAG)

14 cards
Retrieval-Augmented Generation (RAG)

Combines retrieval from external knowledge with model generation.

Vector embeddings

Dense numeric representations capturing semantic meaning.

Embedding models

Models that convert text (or multimodal data) into vectors.

Vector databases

Stores/indexes embeddings for nearest-neighbor search.

Semantic search

Retrieves by meaning similarity rather than exact keyword matches.

Similarity metrics (cosine, dot product, Euclidean)

Distance/similarity formulas used for vector ranking.

Chunking strategies

Methods for splitting documents into retrievable units.

Overlapping chunks

Adds context continuity between adjacent chunks.

Metadata filtering

Restricts retrieval by fields like source, date, tenant, or policy.

Hybrid search (keyword + semantic)

Combines lexical and vector retrieval to improve recall/precision.

Re-ranking

Secondary ranking stage that improves relevance among initial retrieval results.

Grounded generation

Producing answers that are explicitly supported by retrieved evidence.

Knowledge base construction

Building and maintaining curated source corpora for retrieval.

Indexing pipelines

End-to-end ingest, parse, embed, and index workflows for searchable knowledge.

Chapter 6: Parameter-Efficient Adaptation Techniques

4 cards
Full fine-tuning

Updating most or all model parameters for maximum task adaptation.

PEFT

Parameter-efficient fine-tuning; updates only a small subset of weights.

LoRA

Low-rank adapters inserted into target layers for efficient adaptation.

Adapter layers

Small trainable modules added between fro

Chapter 7: Reinforcement Learning and Alignment

3 cards
RLHF

Reinforcement learning from human feedback to align outputs with user preferences.

Reward modeling

Trains a model to score response quality from preference data.

Human preference optimi

Review this concept in the chapter content.

Chapter 8: Evaluation and Metrics

3 cards
Perplexity

Measures how well a model predicts token sequences (lower is generally better).

BLEU

N-gram overlap metric often used for translation quality.

ROUGE

Overlap metric family commonly used for summari

Chapter 9: Safety, Security and Responsible AI

11 cards
Bias in LLMs

Systematic skew in outputs caused by data, training, or deployment patterns.

Fairness

Ensuring comparable quality/treatment across different groups and contexts.

Toxicity detection

Identifying harmful or abusive language in input/output streams.

Content filtering

Blocking/redacting policy-violating or unsafe content.

Guardrails

Rule and policy layers constraining model behavior and tool actions.

Prompt injection attacks

Untrusted content attempting to override instructions and controls.

Jailbreaking

Attempts to bypass safety policies through adversarial prompt patterns.

Data privacy

Protection of sensitive information in training, retrieval, and inference.

Model governance

Ownership, approvals, auditability, and change-control for AI systems.

Compliance considerations

Legal and regulatory requirements tied to industry/jurisdiction.

Ethical AI principles

Transparency, accountability, fairness, privacy, and harm reduction in design and operations.

Chapter 10: Multimodal and Generative Models

8 cards
Multimodal models

Models that process and generate across text, image, audio, or video.

Vision-language models

Architectures that jointly reason over visual and textual inputs.

Text-to-image models

Systems that generate images from natural-language prompts.

Diffusion models

Generative models that iteratively denoise latent noise into outputs.

GANs (high-level awareness)

Generator-discriminator training framework for synthetic data creation.

Cross-modal embeddings

Shared vector spaces connecting semantics across modalities.

Image captioning

Generating descriptive text from image content.

Video generation (conceptual)

Producing temporally coherent sequences from prompts/conditions.

Chapter 11: Deployment, Optimization and NVIDIA Stack

1 cards
Quanti

Review this concept in the chapter content.

Chapter 12: Data Engineering and Workflow Concepts

14 cards
Data pipelines

Automated data movement/processing from source to model-ready artifacts.

ETL workflows

Extract, transform, load steps for structured data preparation.

Feature engineering

Creating useful input signals from raw data.

Embedding pipelines

Generating, storing, and refreshing vector representations.

Dataset labeling

Creating supervised targets and annotation quality controls.

Versioning datasets

Immutable dataset snapshots for reproducibility and audit.

Data governance

Policies for data quality, access, lineage, and stewardship.

Model versioning

Tracking model artifacts across training and deployment cycles.

Experiment tracking

Logging parameters, metrics, artifacts, and outcomes per run.

MLOps concepts

Operational practices for reliable ML/LLM delivery.

CI/CD for ML

Automated validation, packaging, and release processes for model systems.

Monitoring deployed models

Continuous quality, latency, cost, and safety monitoring.

Drift detection

Detecting shifts in input data or model behavior over time.

Feedback loops

Using user/system signals to drive iterative improvements.

Back to NCA-GENL course page