What is hallucination in generative AI?
A hallucination is a confident-looking model output that is unsupported, fabricated, or wrong.
- Looks fluent but is false
- RAG can reduce it but not eliminate it
- Verification still matters
Quick recall
A hallucination is a confident-looking model output that is unsupported, fabricated, or wrong.
MCP is a protocol pattern for connecting models to external tools, resources, and structured context in a standard way.
The context window is the maximum amount of tokenized input and working context a model can handle in one interaction.
Model routing chooses different models or configurations based on task complexity, latency needs, cost, or safety requirements.
Prompt engineering is the practice of structuring instructions and context so a model produces more reliable outputs.
Prompt injection is when untrusted input tries to override instructions or manipulate the model into unsafe behavior.
Tool calling lets a model request an external function or service when the answer requires real actions or fresh data.
Human-in-the-loop design adds review or approval where model mistakes would be costly, risky, or hard to reverse.
Bias is error from overly simple assumptions; variance is sensitivity to training data noise.
Classification predicts categories; regression predicts continuous numeric values.
Model monitoring tracks data drift, prediction quality, latency, and operational health after deployment.
Overfitting memorizes noise; underfitting is too simple to capture the real pattern.