by Djimit
I. Master the 3 Fundamental Prompting Modes
1️⃣ Exploratory Prompting (Rabbit Hole Mode)
Use this for open-ended discovery. Example: “Explain the data sovereignty implications of the EU AI Act.” Drill deeper by isolating subtopics iteratively. This emulates traditional search, but is not optimized for production outputs. Use for learning only.
2️⃣ Collaborative Prompting (Brainstorm Mode)
Treat your LLM as a senior collaborator. Supply rich context — prior decisions, market data, KPIs — so it can co-create viable ideas, hypotheses, or outlines. Example: “Given our last 5 campaign reports, propose 3 novel audience segments for Q4.” More context equals higher signal.
3️⃣ Automated Prompting (Reusable Automation)
Turn proven manual prompts into reusable templates or agent flows. Codify step-by-step instructions. Example: a “Perfect Title Generator” prompt that never deviates from brand guidelines. This mode scales your expertise without repeated manual input.
👉 Key Insight: If you cannot clearly articulate the steps for automation, you have not mastered the task yet.
II. Structure Context – Use Projects & Workspace Integration
Tools like Claude Projects, Gemini for Google Workspace, or Agents SDK act as persistent context containers. Instead of resetting context each session:
- Create a dedicated project per deliverable (e.g., book, report, product spec).
- Inject background documents, references, or constraints.
- Example: In Gemini Docs, tag files with @file to feed live context.
A well-structured project context can increase relevance, reduce hallucination, and minimize repeated clarifications.
III. Decompose Tasks – Singular vs. Modular
- Singular Prompt: Focus on one atomic deliverable per prompt. Example: “Draft an intro paragraph for the investor deck, max 150 words, formal tone.”
- Modular Prompt: Deconstruct complex goals into a sequence. For a course:
- Module research
- Learning outcomes
- Section outlines
- Slide drafts
- Assessment questions
Run each as a separate prompt and stitch results. Don’t overload one prompt with multiple tasks — prompt bloat degrades output quality .
IV. Label & Version Your Prompt Formats
Name your prompt archetypes and enforce style consistency:
- e.g., “Problem → Solution → Call to Action”
- “Story Hook → Context → Lesson Learned”This clarifies structure for both you and the model. It also prevents unintended drift in style.
V. Use Objective, Testable Language
Avoid subjective commands like “make it awesome”. Instead, specify:
- Length (e.g., “100–150 words”)
- Structure (e.g., “bullet list, max 5 items”)
- Format (e.g., JSON, Markdown, CSV)
- Constraints (e.g., “no more than 2 rhetorical questions”).
Objective prompts enable reproducible, evaluable results .
VI. Reinforce with Explicit Quality Attributes
Augment functional instructions with desired quality signals:
- “Use concise, rhythmic sentences.”
- “Employ rhetorical devices: alliteration, contrast.”
- “Follow our approved tone guide.”
For more advanced setups, include references (e.g., “Mimic the style of our top 3 LinkedIn posts.”)
VII. Provide Exact, On-Brand Examples
LLMs learn your standard from examples:
- If you say “3–5 bullets”, all examples must show that.
- If you want a declarative opening, every example must open that way.Consistency calibrates the model’s imitation behavior.
VIII. Always Define Delivery & Format
End each prompt with precise instructions:
- “Return output as a Markdown table with columns for Title, Subtitle, and CTA.”
- “Output exactly 5 short tweets, each under 280 characters, no extra commentary.”
- “Present as a JSON array.”
This reduces post-processing and ensures machine-readiness if chaining tasks.
IX. Iterate Prompts & Self-Heal
Prompt engineering is iterative:
- Test output → identify errors → adjust instructions.
- Ask the model: “What instructions would improve this prompt’s accuracy?”
- Save improved versions as reusable templates.
✅
Key Enterprise Takeaway
Mastering AI prompting is not one trick — it’s an evolving system:
- Use the right mode (exploratory, collaborative, automated)
- Build context-rich projects or agent frameworks
- Decompose tasks
- Specify structure, format, and quality signals
- Iterate relentlessly
This transforms LLMs from generic chatbots into consistent high-output teammates, driving real production value .
📁
Use This as a Living Playbook
💡 Keep your best prompts version-controlled. Store them as reusable templates in your agent orchestration system or Workspace context library.
Blijf op de hoogte
Wekelijks inzichten over AI governance, cloud strategie en NIS2 compliance — direct in je inbox.
[jetpack_subscription_form show_subscribers_total="false" button_text="Inschrijven" show_only_email_and_button="true"]Bescherm AI-modellen tegen aanvallen
Agentic AI ThreatsRisico's van autonome AI-systemen
AI Governance Publieke SectorVerantwoorde AI voor overheden
Cloud SoevereiniteitSoeverein in de cloud — het kan
NIS2 Compliance ChecklistStap-voor-stap naar NIS2-compliance
Klaar om van data naar doen te gaan?
Plan een vrijblijvende kennismaking en ontdek hoe Djimit uw organisatie helpt.
Plan een kennismaking →Ontdek meer van Djimit
Abonneer je om de nieuwste berichten naar je e-mail te laten verzenden.