Protected
NCA-GENL flashcards are available after login. Redirecting...
If you are not redirected, login.
Courses / Nvidia / NCA-GENL / Self-Test
NCA-GENL Flashcards
Concept recall cards across all chapters
82 concept cards ยท generated from chapter scope explanations
Study Method
- Read the term only, then state the definition from memory.
- Open the answer and score yourself: correct, partial, or missed.
- Revisit missed cards after chapter review and repeat until stable recall.
Chapter 1: Foundations of Generative AI and Deep Learning
11 cardsGenerative vs Discriminative models
Generative models create new outputs from learned distributions, while discriminative models classify or score existing inputs.
Supervised vs Unsupervised vs Self-supervised learning
Supervised uses labels, unsupervised finds structure, and self-supervised builds labels from raw data (the core of LLM pretraining).
Representation learning
The model learns useful internal features (embeddings/hidden states) that can transfer across many tasks.
Foundation models
Broad pretrained models that can be adapted with prompting, RAG, or fine-tuning for specific use cases.
Scaling laws
Performance generally improves with more parameters, data, and compute, but with cost and diminishing-return tradeoffs.
Transfer learning
Reusing pretrained model knowledge to reduce training cost/time for downstream tasks.
Neural networks (MLP basics)
Stacked linear and nonlinear layers transform inputs into progressively richer features.
Activation functions (ReLU, GELU, Sigmoid, Tanh)
Nonlinearities that let networks model complex patterns; GELU/ReLU are common in modern transformer stacks.
Loss functions (Cross-Entropy, MSE)
Cross-entropy is typical for token prediction/classification; MSE is common for regression targets.
Backpropagation
Algorithm that computes gradients of loss with respect to model parameters.
Gradient descent
Optimi
Chapter 2: Transformer Architecture and LLM Mechanics
6 cardsSelf-attention
Lets each token weigh relevance of other tokens in the sequence.
Multi-head attention
Uses multiple attention projections so the model can learn different relation types in parallel.
Positional encoding
Adds token-order information because attention alone is permutation-invariant.
Encoder-only vs Decoder-only vs Encoder-Decoder
Encoder-only is strong for understanding, decoder-only for generation, and encoder-decoder for input-to-output transformation.
Feed-forward blocks
Per-token nonlinear layers that refine representations after attention.
Residual connections
Skip paths that preserve signal and stabili
Chapter 3: Training Large Language Models
7 cardsPretraining
Large-scale self-supervised training to build general language capability.
Fine-tuning
Additional training to adapt the base model to a target task/domain.
Supervised fine-tuning (SFT)
Fine-tuning with labeled prompt-response examples.
Instruction tuning
Trains the model to better follow user instructions and response formats.
Transfer learning
Reuse pretrained knowledge instead of training from scratch.
Dataset curation
Selecting, cleaning, balancing, and deduplicating training data.
Data preprocessing
Normali
Chapter 4: Prompt Engineering and Inference Strategies
0 cardsChapter 5: Retrieval-Augmented Generation (RAG)
14 cardsRetrieval-Augmented Generation (RAG)
Combines retrieval from external knowledge with model generation.
Vector embeddings
Dense numeric representations capturing semantic meaning.
Embedding models
Models that convert text (or multimodal data) into vectors.
Vector databases
Stores/indexes embeddings for nearest-neighbor search.
Semantic search
Retrieves by meaning similarity rather than exact keyword matches.
Similarity metrics (cosine, dot product, Euclidean)
Distance/similarity formulas used for vector ranking.
Chunking strategies
Methods for splitting documents into retrievable units.
Overlapping chunks
Adds context continuity between adjacent chunks.
Metadata filtering
Restricts retrieval by fields like source, date, tenant, or policy.
Hybrid search (keyword + semantic)
Combines lexical and vector retrieval to improve recall/precision.
Re-ranking
Secondary ranking stage that improves relevance among initial retrieval results.
Grounded generation
Producing answers that are explicitly supported by retrieved evidence.
Knowledge base construction
Building and maintaining curated source corpora for retrieval.
Indexing pipelines
End-to-end ingest, parse, embed, and index workflows for searchable knowledge.
Chapter 6: Parameter-Efficient Adaptation Techniques
4 cardsFull fine-tuning
Updating most or all model parameters for maximum task adaptation.
PEFT
Parameter-efficient fine-tuning; updates only a small subset of weights.
LoRA
Low-rank adapters inserted into target layers for efficient adaptation.
Adapter layers
Small trainable modules added between fro
Chapter 7: Reinforcement Learning and Alignment
3 cardsRLHF
Reinforcement learning from human feedback to align outputs with user preferences.
Reward modeling
Trains a model to score response quality from preference data.
Human preference optimi
Review this concept in the chapter content.
Chapter 8: Evaluation and Metrics
3 cardsPerplexity
Measures how well a model predicts token sequences (lower is generally better).
BLEU
N-gram overlap metric often used for translation quality.
ROUGE
Overlap metric family commonly used for summari
Chapter 9: Safety, Security and Responsible AI
11 cardsBias in LLMs
Systematic skew in outputs caused by data, training, or deployment patterns.
Fairness
Ensuring comparable quality/treatment across different groups and contexts.
Toxicity detection
Identifying harmful or abusive language in input/output streams.
Content filtering
Blocking/redacting policy-violating or unsafe content.
Guardrails
Rule and policy layers constraining model behavior and tool actions.
Prompt injection attacks
Untrusted content attempting to override instructions and controls.
Jailbreaking
Attempts to bypass safety policies through adversarial prompt patterns.
Data privacy
Protection of sensitive information in training, retrieval, and inference.
Model governance
Ownership, approvals, auditability, and change-control for AI systems.
Compliance considerations
Legal and regulatory requirements tied to industry/jurisdiction.
Ethical AI principles
Transparency, accountability, fairness, privacy, and harm reduction in design and operations.
Chapter 10: Multimodal and Generative Models
8 cardsMultimodal models
Models that process and generate across text, image, audio, or video.
Vision-language models
Architectures that jointly reason over visual and textual inputs.
Text-to-image models
Systems that generate images from natural-language prompts.
Diffusion models
Generative models that iteratively denoise latent noise into outputs.
GANs (high-level awareness)
Generator-discriminator training framework for synthetic data creation.
Cross-modal embeddings
Shared vector spaces connecting semantics across modalities.
Image captioning
Generating descriptive text from image content.
Video generation (conceptual)
Producing temporally coherent sequences from prompts/conditions.
Chapter 11: Deployment, Optimization and NVIDIA Stack
1 cardsQuanti
Review this concept in the chapter content.
Chapter 12: Data Engineering and Workflow Concepts
14 cardsData pipelines
Automated data movement/processing from source to model-ready artifacts.
ETL workflows
Extract, transform, load steps for structured data preparation.
Feature engineering
Creating useful input signals from raw data.
Embedding pipelines
Generating, storing, and refreshing vector representations.
Dataset labeling
Creating supervised targets and annotation quality controls.
Versioning datasets
Immutable dataset snapshots for reproducibility and audit.
Data governance
Policies for data quality, access, lineage, and stewardship.
Model versioning
Tracking model artifacts across training and deployment cycles.
Experiment tracking
Logging parameters, metrics, artifacts, and outcomes per run.
MLOps concepts
Operational practices for reliable ML/LLM delivery.
CI/CD for ML
Automated validation, packaging, and release processes for model systems.
Monitoring deployed models
Continuous quality, latency, cost, and safety monitoring.
Drift detection
Detecting shifts in input data or model behavior over time.
Feedback loops
Using user/system signals to drive iterative improvements.