AI Glossary
101 AI terms explained in simple language. Your go-to reference for understanding AI.
A
Accuracy
The proportion of correct predictions among total predictions. Simple metric but can be misleading with imbalanced datasets.
Read more →Adversarial Examples
Inputs specifically designed to fool AI models into making mistakes. Important for understanding model robustness and security.
Read more →AGI (Artificial General Intelligence)
Hypothetical AI with human-level intelligence across all cognitive tasks. Contrasts with narrow AI, which excels at specific tasks.
Read more →AI Agent
An autonomous AI system that can perceive its environment, make decisions, and take actions to achieve goals. Can use tools and APIs.
Read more →AI Alignment
The challenge of ensuring AI systems behave in accordance with human values and intentions. Critical for safe and beneficial AI development.
Read more →AI Bias
Systematic errors or unfairness in AI systems, often inherited from biased training data. A major concern in responsible AI development.
Read more →AI Safety
The field focused on developing AI systems that are safe, reliable, and beneficial. Addresses risks like misalignment, bias, and unintended consequences.
Read more →API (Application Programming Interface)
A way for software applications to communicate with each other. Most AI tools offer APIs for programmatic access to their capabilities.
Read more →Attention Mechanism
A technique in neural networks that allows models to focus on relevant parts of input data. Core component of transformer architectures.
Read more →Autoencoder
A neural network architecture that learns to compress data into a lower-dimensional representation and reconstruct it. Used for dimensionality reduction and generation.
Read more →B
Backpropagation
Algorithm for calculating gradients in neural networks by propagating errors backward through layers. Enables efficient training.
Read more →Batch Processing
Processing multiple data items together rather than one at a time. Improves efficiency in AI training and inference.
Read more →Batch Size
The number of training examples processed together in one iteration. Affects training speed, memory usage, and model performance.
Read more →Beam Search
A search algorithm used in AI text generation that explores multiple probable sequences simultaneously to find optimal outputs.
Read more →C
Chain-of-Thought Prompting
Prompting technique where you ask AI to show its reasoning step-by-step, improving performance on complex tasks.
Read more →Classification
A supervised learning task where models assign inputs to predefined categories. Examples include spam detection and image recognition.
Read more →Clustering
An unsupervised learning technique that groups similar data points together. Used for customer segmentation, anomaly detection, and more.
Read more →Confusion Matrix
A table showing true positives, false positives, true negatives, and false negatives in classification. Used to calculate accuracy, precision, and recall.
Read more →Context Window
The maximum amount of text (measured in tokens) that an AI model can process at once. Larger context windows allow for longer conversations and documents.
Read more →Contrastive Learning
A self-supervised learning approach where models learn by contrasting similar and dissimilar examples. Used in creating high-quality embeddings.
Read more →Cross-validation
A technique for assessing model performance by training and testing on different subsets of data. Helps detect overfitting and estimate generalization.
Read more →D
Data Augmentation
Artificially expanding training datasets by creating modified versions of existing data. Improves model generalization and reduces overfitting.
Read more →Deep Learning
A subset of machine learning using neural networks with multiple layers. Enables complex pattern recognition and is the foundation of modern AI.
Read more →Diffusion Model
A type of generative AI model that creates images by iteratively refining random noise. Used in tools like Stable Diffusion, DALL-E, and Midjourney.
Read more →Dimensionality Reduction
Techniques for reducing the number of features in data while preserving important information. Helps with visualization and computation.
Read more →Dropout
A regularization technique that randomly ignores neurons during training to prevent overfitting and improve generalization.
Read more →E
Early Stopping
Halting training when validation performance stops improving, preventing overfitting while maintaining good generalization.
Read more →Edge AI
Running AI models on local devices (edge) rather than cloud servers. Improves latency, privacy, and works offline.
Read more →Embeddings
Numerical representations of text, images, or other data that capture semantic meaning. Used for similarity search, clustering, and RAG systems.
Read more →Ensemble Learning
Combining multiple models to achieve better performance than individual models. Includes techniques like bagging and boosting.
Read more →Epoch
One complete pass through the entire training dataset during model training. Models typically train for multiple epochs.
Read more →Explainability (XAI)
The ability to understand and interpret how AI models make decisions. Important for trust, debugging, and regulatory compliance.
Read more →F
F1 Score
Harmonic mean of precision and recall, providing a single metric that balances both. Useful when class distribution is imbalanced.
Read more →Feature Engineering
The process of creating, transforming, and selecting features (input variables) to improve machine learning model performance.
Read more →Federated Learning
Training AI models across multiple decentralized devices without sharing raw data. Enhances privacy in machine learning.
Read more →Few-shot Learning
Providing an AI model with a few examples in the prompt to guide its behavior for a specific task without fine-tuning.
Read more →Fine-tuning
The process of further training a pre-trained AI model on specific data to adapt it for particular tasks or domains.
Read more →G
GAN (Generative Adversarial Network)
A type of generative AI with two networks: a generator creating samples and a discriminator evaluating them. Used in image generation before diffusion models.
Read more →Generative AI
AI systems that create new content such as text, images, audio, or video. Includes LLMs, diffusion models, and other creative AI tools.
Read more →GPT (Generative Pre-trained Transformer)
A family of large language models developed by OpenAI. GPT models are trained to predict the next word in a sequence, enabling them to generate coherent text.
Read more →Gradient Descent
An optimization algorithm that iteratively adjusts model parameters to minimize the loss function. Foundation of neural network training.
Read more →Grounding
Connecting AI outputs to verifiable sources or real-world facts to reduce hallucinations and increase reliability.
Read more →H
Hallucination
When an AI model generates false or nonsensical information that sounds plausible. A common challenge in LLMs that needs to be mitigated through various techniques.
Read more →Hyperparameter
Configuration settings for AI model training (like learning rate) that are set before training begins, as opposed to parameters learned during training.
Read more →I
Image Segmentation
Dividing images into regions or segments based on object boundaries. More precise than object detection, identifying exact pixel-level boundaries.
Read more →Inference
The process of using a trained AI model to make predictions or generate outputs on new data. Distinct from training phase.
Read more →J
Jailbreaking (AI)
Techniques to bypass safety guardrails in AI systems, making them produce restricted content. A concern for AI safety.
Read more →K
Knowledge Distillation
Training a smaller 'student' model to mimic a larger 'teacher' model, creating more efficient models with similar performance.
Read more →L
Latency
The time delay between input and output in AI systems. Lower latency means faster response times, crucial for real-time applications.
Read more →Learning Rate
Hyperparameter controlling how much model parameters change during training. Too high causes instability; too low slows learning.
Read more →LLM (Large Language Model)
A type of artificial intelligence model trained on vast amounts of text data to understand and generate human-like text. Examples include GPT-4, Claude, and Gemini.
Read more →Loss Function
A mathematical function measuring how far model predictions are from actual values. Models are trained to minimize loss.
Read more →M
Machine Learning (ML)
A branch of AI where systems learn patterns from data without explicit programming. Includes supervised, unsupervised, and reinforcement learning.
Read more →Machine Translation
Automatic translation of text or speech between languages using AI. Modern systems use neural networks for high-quality results.
Read more →MLOps
Practices for deploying, monitoring, and maintaining machine learning models in production. Combines ML with DevOps principles.
Read more →Model Checkpoint
Saving model state during training at specific intervals. Allows resuming training and recovering the best model version.
Read more →Model Parameters
The internal variables of an AI model learned during training. Model size is often measured in parameters (e.g., GPT-4 has trillions).
Read more →Model Serving
Infrastructure and processes for deploying trained models to make them available for inference in production environments.
Read more →Multimodal AI
AI systems that can process and generate multiple types of data such as text, images, audio, and video. Examples include GPT-4 Vision and Gemini.
Read more →N
Named Entity Recognition (NER)
NLP technique that identifies and classifies named entities (people, organizations, locations) in text.
Read more →Narrow AI
AI systems designed for specific tasks (like image recognition or language translation) rather than general intelligence. All current AI is narrow AI.
Read more →Neural Network
A computing system inspired by biological neural networks in brains. Consists of interconnected nodes (neurons) organized in layers that process and learn from data.
Read more →O
Object Detection
Computer vision technique that identifies and locates objects within images or videos. Used in autonomous vehicles, security, and more.
Read more →OCR (Optical Character Recognition)
Technology that converts images of text into machine-readable text. Enhanced by modern AI for better accuracy with handwriting and varied fonts.
Read more →Overfitting
When a model learns training data too well, including noise, resulting in poor performance on new data. Avoided through regularization and validation.
Read more →P
Perplexity (Metric)
A measurement of how well a language model predicts text. Lower perplexity indicates better prediction quality.
Read more →Precision
Metric measuring the proportion of positive predictions that are actually correct. Important when false positives are costly.
Read more →Prompt
The input text or instruction given to an AI model to generate a desired output. Well-crafted prompts are essential for getting quality results from AI tools.
Read more →Prompt Engineering
The practice of designing and optimizing prompts to get better responses from AI models. It involves understanding model capabilities and crafting effective instructions.
Read more →Q
Quantization
Reducing the precision of model weights (e.g., from 32-bit to 8-bit) to decrease model size and increase inference speed with minimal accuracy loss.
Read more →R
RAG (Retrieval-Augmented Generation)
A technique that enhances AI responses by retrieving relevant information from external sources before generating output. Helps reduce hallucinations and provides up-to-date information.
Read more →Recall (Sensitivity)
Metric measuring the proportion of actual positives that are correctly identified. Important when false negatives are costly.
Read more →Red Teaming (AI)
Adversarial testing of AI systems to find vulnerabilities, biases, and safety issues before deployment.
Read more →Regression
A supervised learning task where models predict continuous numerical values rather than categories. Used for price prediction, forecasting, etc.
Read more →Regularization
Techniques to prevent overfitting by constraining model complexity. Includes dropout, L1/L2 penalties, and early stopping.
Read more →Reinforcement Learning
A machine learning paradigm where agents learn by interacting with an environment and receiving rewards or penalties for actions.
Read more →RLHF (Reinforcement Learning from Human Feedback)
A technique used to train AI models like ChatGPT by incorporating human preferences and feedback into the learning process.
Read more →S
Semantic Search
Search technology that understands meaning and context rather than just matching keywords. Powered by embeddings and vector databases.
Read more →Sentiment Analysis
AI technique for determining emotional tone in text (positive, negative, neutral). Used in customer feedback analysis and social media monitoring.
Read more →Speech Recognition
AI technology that converts spoken language into text. Used in transcription services, voice assistants, and accessibility tools.
Read more →Streaming Response
Delivering AI-generated output progressively as it's created, rather than waiting for complete generation. Improves perceived responsiveness.
Read more →Supervised Learning
A machine learning approach where models learn from labeled training data to make predictions on new, unseen data.
Read more →Synthetic Data
Artificially generated data that mimics real data. Used when real data is scarce, expensive, or sensitive. Can be created by AI.
Read more →System Prompt
Initial instructions given to an AI model that set its behavior, role, and constraints for a conversation or task.
Read more →T
Temperature
A parameter controlling randomness in AI text generation. Lower values make output more deterministic; higher values more creative and varied.
Read more →Text Summarization
AI capability to condense long documents into shorter summaries while preserving key information. Can be extractive or abstractive.
Read more →Text-to-Image
AI technology that generates images from text descriptions. Powered by diffusion models and other generative techniques.
Read more →Text-to-Speech (TTS)
Technology that converts written text into spoken audio. Modern AI TTS systems like ElevenLabs produce highly natural-sounding voices.
Read more →Text-to-Video
AI systems that create videos from text prompts. Emerging technology exemplified by tools like Runway, Pika, and Sora.
Read more →Token
The basic unit of text that AI models process. A token can be a word, part of a word, or a character. Most AI APIs charge based on token usage.
Read more →Tool Use (Function Calling)
AI capability to use external tools, APIs, or functions to accomplish tasks beyond text generation. Enables AI agents.
Read more →Top-p (Nucleus Sampling)
A sampling method in text generation that considers only the most probable tokens whose cumulative probability exceeds p. Alternative to temperature.
Read more →Training Data
The dataset used to teach an AI model. Quality and diversity of training data greatly impact model performance and bias.
Read more →Transfer Learning
Using knowledge from a model trained on one task to improve performance on a related task. Foundation of modern AI, enabling efficient adaptation.
Read more →Transformer
A neural network architecture that revolutionized AI by using attention mechanisms. The foundation for modern LLMs like GPT and BERT.
Read more →U
Underfitting
When a model is too simple to capture patterns in data, performing poorly on both training and new data.
Read more →Unsupervised Learning
Machine learning where models find patterns in unlabeled data without predefined categories or outcomes.
Read more →V
Vector Database
A specialized database designed to store and query embeddings efficiently. Used in RAG systems and semantic search applications.
Read more →Voice Cloning
AI technology that replicates a person's voice from audio samples, allowing text-to-speech in that voice. Raises ethical considerations around consent.
Read more →Z
Zero-day (AI Context)
AI capabilities or behaviors that emerge unexpectedly without explicit training, often discovered after deployment.
Read more →Zero-shot Learning
The ability of an AI model to perform tasks it wasn't explicitly trained for, without any examples in the prompt.
Read more →