Essential Terms to Understand for AI Prompting
- Brad Caldana
- Sep 14
- 2 min read
Artificial intelligence is only as good as the instructions you give it. Whether you’re writing content, analyzing data, or brainstorming new ideas, knowing the language of AI prompting helps you get better results. Here’s a guide to the essential terms and techniques you should know.

🔑 Core Prompting Terms
Prompt
The text (or structured input) you give an AI model to guide its output. It can be a question, instruction, or example.
Completion / Response
The output generated by the AI model in reply to your prompt.
Tokens
Units of text the model processes. A token is roughly 3–4 characters or about ¾ of a word in English. Prompt + completion together count toward the token limit.
Context Window
The maximum number of tokens the model can “remember” at once (e.g., 8k, 32k, or 128k tokens). If you exceed it, earlier parts may get dropped.
System / User / Assistant Roles
System prompt: hidden instruction that sets rules (e.g., “You are a helpful assistant”).
User prompt: your visible input.
Assistant: the model’s reply.
🎯 Prompting Techniques
Zero-Shot Prompting
Giving the model a task without examples.Example: “Translate this sentence into French.”
Few-Shot Prompting
Providing a few examples so the model learns the pattern.Example: “Translate the following words…” + examples.
Chain-of-Thought Prompting
(CoT)Encouraging the model to reason step by step for complex problems.Example: “Think step by step…”
Role Prompting
Assigning the AI a persona or role to shape tone and depth.Example: “You are a professor of history…”
Instruction Prompting
Directly telling the model what to do.Example: “Summarize in bullet points.”
Reflexion / Iterative Prompting
Asking the AI to review or critique its own output and improve it.
🧰 Prompt Structure Concepts
Input / Output Format
Specifying how information should be given and returned (e.g., “Respond in JSON,” “Return a 3-column table”).
Constraints
Rules you set for the model’s response (e.g., word count, tone, style).
Temperature
Controls randomness of output.
Low = more focused/deterministic.
High = more creative/varied.
Top-p (Nucleus Sampling)
Another randomness control: limits output to a probability mass of likely words.
Grounding
Anchoring the model in external data (documents, search, or APIs) so answers are factual and up-to-date.
⚡ Emerging Terms
Prompt Engineering
The practice of crafting and refining prompts systematically for better results.
Prompt Chaining
Linking multiple prompts together to build step-by-step workflows.
RAG (Retrieval-Augmented Generation)
Feeding the model external documents retrieved at runtime so it answers from real data.
Guardrails / Safety Prompts
Prompts that help prevent undesired or harmful outputs.
Final Word
Mastering these terms isn’t about jargon—it’s about control. The more you understand the building blocks of prompting, the more precisely you can guide AI to deliver useful, creative, and trustworthy results.
Comments