📖 AI Glossary

Plain English definitions for every AI term · 62 terms · Searchable

A

Agent

An AI system that can perceive its environment, make decisions, and take actions to achieve goals. Agents often work over time, learning from feedback and adapting their behavior.

See Also: Machine Learning, Reinforcement Learning from Human Feedback

API

Application Programming Interface. A way for software programs to talk to each other. For AI, APIs let developers use powerful AI models without building them from scratch.

See Also: Inference

Attention Mechanism

A technique in AI that helps the model focus on the most important parts of the input. Like how your eyes focus on a person's face in a crowded room, attention helps AI prioritize what matters.

See Also: Transformer, Neural Network

Artificial Intelligence (AI)

Technology that enables computers to perform tasks that normally require human intelligence. This includes learning from experience, recognizing patterns, understanding language, and making decisions.

B

Benchmark

A standardized test used to measure how well an AI system performs. Think of it like a school test—different benchmarks test different skills, and researchers use them to compare models fairly.

See Also: Evaluation

BERT

A famous language model developed by Google that learns from large amounts of text. BERT excels at understanding context and relationships between words, making it useful for tasks like categorizing text.

See Also: Foundation Model, Neural Network

Bias (AI)

Systematic errors or unfairness in how an AI makes decisions. This happens when the training data has gaps, unfair patterns, or when the algorithm itself isn't built fairly. Bias can lead to discriminatory outcomes.

See Also: Training Data, Fine-Tuning

C

Chain-of-Thought

A technique where you ask an AI to "think step-by-step" before giving an answer. Instead of jumping to a conclusion, the model shows its reasoning process, which often leads to better and more accurate results.

See Also: Prompt Engineering, In-Context Learning

Claude

An advanced language model created by Anthropic. Claude is trained to be helpful, honest, and harmless, and excels at writing, analysis, coding, and reasoning tasks.

See Also: Foundation Model, LLM

ChatGPT

A popular conversational AI model created by OpenAI. ChatGPT can chat, answer questions, write content, code, and more. It's one of the most widely used AI tools.

See Also: LLM, Foundation Model

Context Window

The amount of text an AI model can "see" at one time before it needs to start forgetting older parts. A larger context window means the model can handle longer documents and remember more of the conversation.

See Also: Token, Inference

Computer Vision

AI technology that allows computers to understand and analyze images and videos, similar to how humans see. It's used for tasks like recognizing objects, reading text from images, or identifying faces.

See Also: Neural Network, Deep Learning

Corpus

A large collection of text used to train language models. The bigger and more diverse the corpus, the better the AI can understand language across different topics and writing styles.

See Also: Training Data, Fine-Tuning

D

Deep Learning

A branch of machine learning inspired by how brains work. Deep learning uses layers of connected nodes (artificial neurons) to learn from data. Most modern AI systems use deep learning.

See Also: Neural Network, Machine Learning

DeepSeek

An AI company known for creating open-source language models and reasoning models. DeepSeek focuses on making powerful AI accessible and affordable.

See Also: LLM, Foundation Model

Diffusion Model

A type of AI that generates images by starting with random noise and gradually refining it into an image. It's like reversing the process of adding noise to a photo—working backward to create new images.

See Also: Deep Learning, Emergent Capability

Document Retrieval

Finding relevant documents or passages from a large collection based on a query. This is a key part of RAG systems that help AI answer questions using up-to-date information.

See Also: RAG, Vector Database

E

Embedding

A way to represent words, sentences, or images as mathematical vectors that capture their meaning. Embeddings let AI understand that "dog" and "puppy" are similar, even though they're spelled differently.

See Also: Vector Database, Neural Network

Emergent Capability

An ability that appears in an AI model only when it reaches a certain size or training level. Smaller models might not have it, but larger models suddenly can do something new without being specifically trained for it.

See Also: Foundation Model, Scaling

Evaluation (AI)

The process of testing how well an AI system works on specific tasks. Evaluation can be automated (using benchmarks) or human-based (having people judge the quality of outputs).

See Also: Benchmark, MMLU

F

Few-Shot Learning

Teaching an AI to do a task by showing it just a few examples. Instead of needing thousands of examples, the model learns the pattern from 2-5 examples provided in the prompt.

See Also: In-Context Learning, Zero-Shot Learning

Fine-Tuning

Taking a pre-trained AI model and further training it on new, specific data. Think of it as specialized education—the model already knows a lot, then you teach it your domain-specific expertise.

See Also: Training Data, Foundation Model

Foundation Model

A large, general-purpose AI model trained on diverse data that can be adapted to many different tasks. Claude, ChatGPT, and Gemini are all foundation models.

See Also: LLM, Fine-Tuning

G

Gemini

Google's advanced AI model family, including text, image, and multimodal versions. Gemini is particularly well-integrated with Google products like Docs, Gmail, and Sheets.

See Also: Foundation Model, Multimodal

GPT

Generative Pre-trained Transformer. A type of AI model that learns from massive amounts of text to predict the next word, then uses this to generate text, answer questions, and more. ChatGPT is built on GPT technology.

See Also: Transformer, LLM

Grounding

Connecting AI outputs to real facts and information. A "grounded" AI can explain where its information comes from and cite sources, rather than just making things up.

See Also: Hallucination, RAG

H

Hallucination

When an AI makes up false information with confidence. For example, inventing a source that doesn't exist or stating false facts as true. Hallucinations are a real challenge in modern AI.

See Also: Grounding, RAG

Human-in-the-Loop

An AI system that combines machine decisions with human judgment. The AI does most of the work, but humans review, approve, or correct important decisions.

See Also: RLHF, Evaluation

I

Inference

Using a trained AI model to make predictions or generate outputs on new data. Inference is the opposite of training—it's when the model is in "working mode," answering questions or processing text.

See Also: Training Data, Parameter

In-Context Learning

The ability of an AI to learn from examples given in the prompt itself, without being retrained. This is how ChatGPT and Claude can quickly adapt to new styles or tasks.

See Also: Few-Shot Learning, Prompt Engineering

Instruction Tuning

Training an AI model to follow instructions well. The model learns to understand what users ask for and respond appropriately, making it more helpful and practical.

See Also: Fine-Tuning, RLHF

L

LLM (Large Language Model)

A large AI model trained on vast amounts of text that can understand and generate human language. ChatGPT, Claude, and Gemini are all LLMs. The "large" refers to the billions of parameters they contain.

See Also: Foundation Model, Parameter

Llama

A family of open-source language models developed by Meta. Llama models are designed to be efficient and can run on personal computers, making powerful AI more accessible.

See Also: Foundation Model, Open source

Latent Space

A high-dimensional space where the AI represents meaning. It's hard to visualize, but think of it as a vast map where similar ideas are close together and different ideas are far apart.

See Also: Embedding, Neural Network

M

Machine Learning (ML)

The science of creating AI systems that improve through experience. Instead of being programmed with rules, ML systems learn patterns from data.

See Also: Deep Learning, Neural Network

MCP (Model Context Protocol)

A standard protocol that lets AI models safely connect to and use external tools. Think of it as a universal adapter that lets Claude use web search, databases, email, or custom applications.

See Also: Agent, API

Multimodal

An AI that can work with multiple types of input or output, like text, images, audio, and video all in one system. Multimodal models are more versatile than models that only handle text.

See Also: Foundation Model, Computer Vision

MMLU

Massive Multitask Language Understanding. A benchmark test with questions across 57 subjects (science, math, history, etc.). It's often used to measure how knowledgeable an AI is.

See Also: Benchmark, Evaluation

N

Neural Network

A computing system inspired by how brains work. It consists of connected nodes (artificial neurons) organized in layers. Neural networks are the foundation of modern AI.

See Also: Deep Learning, Parameter

NLP (Natural Language Processing)

The field of AI focused on understanding and generating human language. NLP powers everything from chatbots to language translation to text analysis.

See Also: LLM, Transformer

O

OpenAI

An AI research company that created GPT models and ChatGPT. OpenAI is one of the leading organizations advancing large language models.

See Also: ChatGPT, GPT

Orchestration

Coordinating multiple AI systems or steps to solve complex problems. It's like conducting an orchestra—different instruments (different AI tools) play together to create a complete solution.

See Also: Agent, MCP

Overfitting

When an AI learns the training data too well, including its quirks and errors, so it performs poorly on new data. It's like memorizing answers instead of learning concepts.

See Also: Training Data, Fine-Tuning

P

Parameter

A number the AI learns during training that affects how it processes information. A model with billions of parameters is more powerful but also slower and more expensive to run.

See Also: Neural Network, Training Data

Perplexity (AI tool)

An AI search engine that combines web search with conversational AI. It provides answers with sources cited, making it useful for research and finding current information.

See Also: RAG, Grounding

Prompt

The input or instruction you give to an AI to get a response. A good prompt is clear, specific, and provides context. Prompts can be questions, commands, or examples.

See Also: Prompt Engineering, In-Context Learning

Prompt Engineering

The art of writing prompts that get better results from AI. Techniques include being specific, asking for step-by-step thinking, giving examples, and framing requests clearly.

See Also: Prompt, Chain-of-Thought

Prompt Injection

A security risk where hidden instructions are embedded in text to trick an AI into doing something unintended. For example, a user might hide a command at the end of a long document.

See Also: Agent, Grounding

R

RAG (Retrieval-Augmented Generation)

A technique where an AI retrieves relevant information from documents or databases before answering a question. This helps it provide accurate, current information instead of relying only on training data.

See Also: Document Retrieval, Vector Database

RLHF (Reinforcement Learning from Human Feedback)

A training technique where human feedback helps refine AI behavior. Humans rate different outputs, and the AI learns to produce responses humans find helpful and correct.

See Also: Training Data, Human-in-the-Loop

Reasoning Model

An AI designed to work through complex problems step-by-step, showing its reasoning. These models are slower but more accurate for difficult tasks that require careful logical thinking.

See Also: Chain-of-Thought, Foundation Model

S

Skill (Claude Code)

A packaged automation or workflow built for Claude that can perform specific tasks across multiple apps. Skills let users orchestrate complex multi-step processes automatically.

See Also: MCP, Agent

System Prompt

Hidden instructions given to an AI that shape its behavior and personality. For example, a system prompt might tell Claude to be concise or helpful in a certain style.

See Also: Prompt, Instruction Tuning

Supervised Learning

Training an AI using labeled examples (inputs paired with correct outputs). The AI learns to predict outputs for new inputs. Most modern language models use some form of supervised learning.

See Also: Training Data, Machine Learning

T

Token

A small piece of text that an AI processes. A token might be a word, part of a word, or punctuation. "Hello world" is about 2 tokens. Token count affects how long a conversation can be.

See Also: Context Window, Parameter

Temperature

A setting that controls how creative or predictable an AI response is. Low temperature (0.1) makes responses focused and consistent. High temperature (0.9) makes them more creative and varied.

See Also: Inference, Parameter

Transformer

A type of neural network architecture that powers most modern AI. Transformers use "attention" to understand relationships between words, making them excellent at understanding language.

See Also: Attention Mechanism, Neural Network

Training Data

The text, images, or other information used to teach an AI model. Larger and more diverse training data generally produces better models. The quality of training data directly affects the quality of the AI.

See Also: Fine-Tuning, Corpus

U

Unsupervised Learning

Training an AI on unlabeled data where the model finds patterns on its own. Unlike supervised learning, no one tells the AI what's correct—it discovers structure in the data.

See Also: Machine Learning, Supervised Learning

V

Vector Database

A specialized database that stores and searches data based on meaning rather than exact keywords. It's essential for RAG systems and finding similar documents quickly.

See Also: Embedding, RAG

Vision Model

An AI trained to understand and analyze images. Vision models can identify objects, read text from images, describe what they see, and answer questions about images.

See Also: Computer Vision, Multimodal

Z

Zero-Shot Learning

The ability to perform a task without seeing any examples. The AI uses its general knowledge. For instance, an AI can translate to a language it never trained on.

See Also: Few-Shot Learning, In-Context Learning