Updated March 2026 · Based on OpenAI Official Guidance
40+ ready-to-use prompts for coding, writing, research, computer use & agentic workflows. One-click copy. Optimized for OpenAI's latest model.
This website is an unofficial fan project and is not affiliated with, endorsed by, or connected to OpenAI. “GPT” is a trademark of OpenAI.
Released on March 5, 2026, GPT-5.4 is OpenAI's most capable frontier model to date. It represents a significant leap forward in token efficiency, multi-step reasoning, and agentic reliability. Unlike its predecessor GPT-5.2, this model introduces three groundbreaking capabilities that fundamentally change how we write prompts.
First GPT model that can operate computers directly—clicking, typing, navigating apps. Scores 75% on OSWorld, surpassing humans (72.4%).
Automatically finds the right tools for any task without uploading tool lists. Cuts prompt size and inference cost dramatically.
Processes up to 1 million tokens per request. Ideal for large codebases, multi-document analysis, and long-running agent sessions.
GPT-5.4 is fundamentally more capable at following structured instructions than previous models. The biggest performance gains come from three prompt changes: explicit output contracts that define exactly what “done” looks like, reasoning effort tuning using the new five-level parameter (none/low/medium/high/xhigh), and verification loops before high-impact actions. The model is also 47% more token-efficient in tool-heavy workflows, which means old prompts that were verbose to compensate for model limitations can now be simplified. The templates below incorporate all these patterns directly from OpenAI’s official prompt guidance documentation.
Click any prompt to copy it to your clipboard. Each template includes recommended reasoning_effort and verbosity settings.
Thorough review with actionable categories. Use for PRs or pre-merge audits.
You are a senior code reviewer. Review the code below and return structured feedback.
<output_contract>
Return exactly these sections in order:
1. **Critical Issues** — bugs, security flaws, data loss risks
2. **Improvements** — performance, readability, maintainability
3. **Positive Patterns** — what the code does well
4. **Summary** — 1-sentence verdict: APPROVE, REQUEST_CHANGES, or NEEDS_DISCUSSION
</output_contract>
<verbosity_controls>
- Each issue: 1 line description + 1 line fix suggestion.
- No praise filler. If a section is empty, write "None."
</verbosity_controls>
<dig_deeper_nudge>
- Look for second-order issues: does this code interact safely with concurrent access?
- Check if error handling covers all code paths.
</dig_deeper_nudge>
Code to review:
```
[PASTE YOUR CODE HERE]
```
API usage (Python):
reasoning={"effort": "medium"}, text={"verbosity": "low"}
Critical Issues
1. SQL injection via unsanitized user_id in query string. → Use parameterized queries instead of string concatenation.
2. Missing null check on db.get() return value before accessing .rows. → Add early return with 404 if result is null.
Improvements
1. processData() runs O(n²) due to nested loop. → Use a Map for O(n) lookup.
2. Magic number 86400 should be a named constant SECONDS_PER_DAY.
Positive Patterns
1. Good separation of concerns between controller and service layers.
2. Consistent error logging with structured context.
Summary
REQUEST_CHANGES — the SQL injection is a critical security flaw that must be fixed before merge.
Systematic root-cause analysis with fix implementation. Great for debugging sessions.
You are a debugging expert. Diagnose the bug and provide a fix. <output_contract> 1. **Root Cause** — what exactly is wrong and why (2-3 sentences) 2. **Fix** — the corrected code (complete function/block, not a partial diff) 3. **Verification** — how to confirm the fix works (test case or command) </output_contract> <dig_deeper_nudge> - Don't stop at the first plausible cause. - Check for second-order issues: race conditions, edge cases, null states. - If the bug might have siblings elsewhere, mention them. </dig_deeper_nudge> Bug description: [DESCRIBE THE BUG] Error message: [PASTE ERROR] Relevant code: ``` [PASTE CODE] ```
Clean up messy code while preserving behavior. Follows Cursor-style code quality rules.
Refactor the following code for clarity and maintainability. <code_editing_rules> - Write code for clarity first. Use readable names, not single-letter variables. - Prefer small, focused functions over large monolithic blocks. - Preserve all existing behavior — this is a refactor, not a rewrite. - Keep changes consistent with the surrounding codebase style. - Use high verbosity for the code output so it is easy to review. </code_editing_rules> <output_contract> 1. **Refactored Code** — the complete refactored version 2. **Changes Made** — bullet list of what changed and why (max 5 bullets) 3. **Behavior Preserved** — confirmation that no logic changed </output_contract> Code to refactor: ``` [PASTE CODE] ```
Generates comprehensive tests covering happy path, edge cases, and error scenarios.
Write unit tests for the code below. <output_contract> - Use [FRAMEWORK: jest/pytest/go test/etc.] test framework. - Organize tests into groups: Happy Path, Edge Cases, Error Handling. - Each test: descriptive name, arrange-act-assert structure. - Aim for >90% branch coverage. - Output only the test file, no explanatory prose. </output_contract> <completeness_contract> - Cover: null/empty inputs, boundary values, type mismatches, concurrent access (if applicable). - If a scenario is untestable without mocks, note it as a comment. </completeness_contract> Function to test: ``` [PASTE CODE] ```
Design RESTful or GraphQL endpoints with schemas, error handling, and documentation.
Design a REST API endpoint for the following requirement. <output_contract> Return these sections: 1. **Endpoint** — method, path, description 2. **Request Schema** — headers, query params, body (JSON Schema) 3. **Response Schema** — success (200) and error (4xx/5xx) responses 4. **Implementation** — complete handler code in [LANGUAGE] 5. **cURL Example** — ready-to-run test command </output_contract> <verification_loop> Before finalizing: - Does the endpoint handle authentication/authorization? - Are all error cases covered with proper HTTP status codes? - Is input validated and sanitized? </verification_loop> Requirement: [DESCRIBE THE ENDPOINT NEED] Tech stack: [e.g., Node.js/Express, Python/FastAPI, Go/Gin]
Design normalized schemas with indexes, constraints, and migration scripts.
Design a database schema for the following requirements. <output_contract> 1. **ER Diagram** — entities, relationships, cardinality (text description) 2. **SQL DDL** — CREATE TABLE statements with types, constraints, indexes 3. **Indexes** — explain why each index exists (query pattern it serves) 4. **Migration Script** — versioned up/down migration 5. **Sample Queries** — 3 common queries the schema supports </output_contract> <verification_loop> - Is the schema at least 3NF? - Are foreign keys and cascading deletes handled correctly? - Will the indexes support the expected read patterns without excessive write overhead? </verification_loop> Requirements: [DESCRIBE YOUR DATA MODEL] Database: [PostgreSQL/MySQL/SQLite]
Scan code for OWASP Top 10 vulnerabilities with severity ratings and fixes.
Perform a security audit on the code below. Focus on OWASP Top 10 vulnerabilities. <output_contract> For each finding: - **Severity**: CRITICAL / HIGH / MEDIUM / LOW - **Vulnerability**: type (e.g., SQL Injection, XSS, IDOR) - **Location**: file/line/function - **Impact**: what an attacker could do - **Fix**: specific code change If no vulnerabilities found, state "No issues found" with confidence level. </output_contract> <dig_deeper_nudge> - Check authentication bypasses, privilege escalation, data exposure. - Look for hardcoded secrets, insecure defaults, missing rate limits. - Consider supply chain risks in dependencies. </dig_deeper_nudge> Code to audit: ``` [PASTE CODE] ```
Find bottlenecks and provide optimized alternatives with complexity analysis.
Analyze and optimize the performance of this code. <output_contract> 1. **Current Complexity** — time and space (Big O) 2. **Bottlenecks** — ranked list of hotspots with explanation 3. **Optimized Code** — rewritten version with improvements 4. **New Complexity** — time and space after optimization 5. **Tradeoffs** — what was sacrificed (if anything) for the speedup </output_contract> Consider: algorithmic improvements, data structure choices, caching opportunities, I/O reduction, and parallelization potential. Code to optimize: ``` [PASTE CODE] ``` Context: [e.g., "called 10K times/sec", "dataset is ~1M rows"]
End-to-end feature from DB schema to API to frontend. Uses GPT-5.4's self-reflection for quality.
Implement the following feature end-to-end.
<self_reflection>
Before writing code, create an internal quality rubric with 5-7 categories
(correctness, security, performance, readability, UX, error handling, testability).
Iterate on your solution until it scores highly on all categories.
</self_reflection>
<output_contract>
Deliver in this order:
1. **Plan** — 5-bullet implementation plan
2. **Database** — migration/schema changes
3. **Backend** — API routes, business logic, validation
4. **Frontend** — components, state management, API integration
5. **Tests** — key test cases for each layer
</output_contract>
Tech stack: [e.g., Next.js, TypeScript, Prisma, PostgreSQL]
Feature: [DESCRIBE THE FEATURE]
API usage (Python):
reasoning={"effort": "xhigh"}, text={"verbosity": "high"}
# xhigh recommended — this is a long-horizon, multi-step reasoning task
Auto-generate API docs, README sections, or inline documentation from code.
Generate documentation for the following code. <output_contract> Format: [Markdown README / JSDoc / Python docstrings / Go doc comments] Include: - Purpose and overview (2-3 sentences) - Parameters/arguments with types and descriptions - Return values with types - Usage examples (at least 2) - Error cases and how they are handled Do NOT add redundant comments that merely restate the code. </output_contract> Code to document: ``` [PASTE CODE] ```
SEO-optimized long-form blog post with clear structure and engaging voice.
Write a blog post on the topic below. <personality_and_writing_controls> - Persona: Experienced industry practitioner who teaches clearly - Channel: Blog - Emotional register: Confident and approachable, not academic or salesy - Formatting: Use H2/H3 headings, short paragraphs (2-3 sentences), bullet lists where helpful - Length: 1200-1800 words </personality_and_writing_controls> <output_contract> Structure: 1. Hook opening (no generic "In today's world..." starters) 2. 3-5 main sections with H2 headings 3. Practical examples or data points in each section 4. Actionable conclusion with clear next steps Include: meta title (60 chars), meta description (155 chars), 5 SEO keywords </output_contract> Topic: [YOUR TOPIC] Target audience: [WHO READS THIS] Target keyword: [PRIMARY KEYWORD]
Prompt input: Topic = "Why startups should use AI code review tools", Audience = "CTO / VP Engineering", Keyword = "AI code review"
Meta title: AI Code Review Tools: Why Your Team Needs One in 2026
Meta description: AI code review tools catch 40% more bugs than manual reviews. Learn how CTOs are using them to ship faster with fewer production incidents.
## The $4.7M Bug That Changed How We Review Code
Last October, a single uncaught null pointer in our payment service took down checkout for 47 minutes. The postmortem revealed something uncomfortable: three engineers had reviewed the PR. None caught it...
[continues with 4 more H2 sections, practical data points, and actionable conclusion with tool recommendations]
3-email cold outreach sequence with personalization hooks and clear CTAs.
Write a 3-email cold outreach sequence. <personality_and_writing_controls> - Persona: Peer who genuinely wants to help, not a pushy salesperson - Channel: Email - Emotional register: Direct and warm, not desperate or overly formal - Formatting: No bullets in emails. Plain text, short paragraphs. - Length: Email 1: <=100 words, Email 2: <=80 words, Email 3: <=60 words </personality_and_writing_controls> <output_contract> For each email provide: - Subject line (under 50 characters, no clickbait) - Body text - [PERSONALIZATION] placeholders where prospect-specific details go - Send timing recommendation (days after previous) </output_contract> Product/Service: [WHAT YOU SELL] Target persona: [THEIR ROLE AND PAIN POINT] Key value prop: [WHY THEY SHOULD CARE]
Multi-channel launch copy: press release, social posts, and email.
Create a product launch announcement package. <personality_and_writing_controls> - Persona: The company's brand voice — [DESCRIBE: e.g., innovative and bold] - Emotional register: Excited but grounded in specifics, not hype - Length: Press release 400 words, each social post <280 chars, email 200 words </personality_and_writing_controls> <output_contract> Deliver exactly: 1. **Press Release** — headline, subhead, 3-paragraph body, boilerplate, quote from exec 2. **Twitter/X Post** — 3 variations (hook, feature-focus, social-proof) 3. **LinkedIn Post** — 1 longer-form post with storytelling angle 4. **Email to Customers** — subject line + body announcing the launch </output_contract> Product: [PRODUCT NAME AND DESCRIPTION] Key features: [TOP 3 FEATURES] Launch date: [DATE] Target audience: [WHO]
Polished business memo with precise conclusions and calibrated certainty.
Write an executive summary memo. <memo_mode> - Write in a polished, professional memo style. - Use exact names, dates, entities, and authorities when supported by the record. - Prefer precise conclusions over generic hedging. - When uncertainty is real, tie it to the exact missing fact or conflicting source. - Synthesize across inputs rather than summarizing each one independently. </memo_mode> <output_contract> Format: - TO / FROM / DATE / RE header - **Bottom Line Up Front** (2-3 sentences) - **Key Findings** (3-5 bullet points) - **Recommendation** (1-2 sentences with confidence level) - **Next Steps** (action items with owners and deadlines) Total length: 300-500 words. </output_contract> Context: [PASTE BACKGROUND MATERIAL / DATA / SITUATION] Decision needed: [WHAT DECISION IS THIS SUPPORTING]
A week of platform-specific posts with hashtags and optimal timing.
Create a 7-day social media content calendar. <output_contract> For each day provide: | Day | Platform | Post Text | Hashtags | Best Time | Content Type | Platforms: Twitter/X, LinkedIn, Instagram Content mix: 40% educational, 30% engagement, 20% promotional, 10% behind-the-scenes Output as a Markdown table. </output_contract> <verbosity_controls> - Twitter posts: max 240 characters - LinkedIn: max 300 words - Instagram captions: max 150 words + 10 hashtags </verbosity_controls> Brand: [BRAND NAME AND DESCRIPTION] Industry: [YOUR INDUSTRY] Key message this week: [THEME OR CAMPAIGN] Tone: [e.g., professional, playful, authoritative]
High-converting landing page with hero, features, social proof, and CTA sections.
Write landing page copy that converts. <output_contract> Sections in order: 1. **Hero** — headline (max 10 words), subheadline (max 25 words), CTA button text 2. **Problem** — 2-3 sentences describing the pain point 3. **Solution** — how the product solves it (3 bullet points) 4. **Features** — 3 features with icon-friendly titles and 1-sentence descriptions 5. **Social Proof** — 2 testimonial templates with [PLACEHOLDER] fields 6. **Final CTA** — urgency-driven closing with button text </output_contract> <personality_and_writing_controls> - Emotional register: Confident and specific, not generic or hyperbolic - Ban phrases: "revolutionize", "game-changing", "world-class", "cutting-edge" - Every claim must be backed by a specific benefit or metric </personality_and_writing_controls> Product: [NAME AND DESCRIPTION] Target customer: [WHO] Key differentiator: [WHY YOU vs COMPETITORS]
Step-by-step tutorial with code examples, prerequisites, and troubleshooting.
Write a technical tutorial on the topic below. <output_contract> Structure: 1. **What You'll Build** — 2-sentence outcome description 2. **Prerequisites** — bullet list of required tools/knowledge 3. **Steps** — numbered, each with explanation + code block + expected output 4. **Common Errors** — 3 pitfalls with solutions 5. **Next Steps** — where to go from here </output_contract> <personality_and_writing_controls> - Persona: Senior developer explaining to a mid-level colleague - Formatting: Use numbered steps, code fences with language tags - Ban: "simply", "just", "easy" — respect the reader's learning process </personality_and_writing_controls> Topic: [TUTORIAL TOPIC] Target reader: [SKILL LEVEL AND BACKGROUND] Tech stack: [TOOLS AND VERSIONS]
Rewrite existing text to match a specific brand voice or audience tone.
Rewrite the text below in the specified tone and style. <personality_and_writing_controls> - Target tone: [e.g., casual and witty / formal and authoritative / warm and empathetic] - Target audience: [WHO WILL READ THIS] - Constraints: Preserve all factual content and key messages. Do not add new claims. Do not remove critical information. - Length: Same as original (within 10%) </personality_and_writing_controls> <output_contract> Return only the rewritten text. No explanations or comparisons. </output_contract> Original text: """ [PASTE TEXT HERE] """
3-pass research with sub-questions, evidence gathering, and synthesized conclusions.
Conduct deep research on the topic below and produce a comprehensive report. <research_mode> Do research in 3 passes: 1) Plan: list 3-6 sub-questions to answer. 2) Retrieve: investigate each sub-question, follow 1-2 second-order leads. 3) Synthesize: resolve contradictions and write the final report with citations. Stop only when more searching is unlikely to change the conclusion. </research_mode> <citation_rules> - Only cite sources retrieved in the current workflow. - Never fabricate citations, URLs, or quotes. - Attach citations to specific claims, not only at the end. - If sources conflict, state the conflict and attribute each side. </citation_rules> <output_contract> 1. **Executive Summary** (200 words) 2. **Key Findings** (organized by sub-question) 3. **Analysis** (evidence-backed, with inline citations) 4. **Limitations** (what the research could not answer) 5. **Sources** (numbered reference list) </output_contract> Research topic: [YOUR TOPIC] Scope/constraints: [TIME PERIOD, GEOGRAPHY, INDUSTRY]
Prompt input: Topic = "Impact of AI coding assistants on developer productivity", Scope = "2024-2026, global, software engineering"
Executive Summary
AI coding assistants have become standard in professional development, with adoption rising from 41% in 2024 to an estimated 78% in 2026 [1]. Controlled studies show 26-55% faster task completion on well-scoped coding tasks [2][3], though gains diminish on novel architectural work...
Key Findings (by sub-question)
Q1: How much faster do developers complete tasks?
GitHub’s 2025 study of 950 developers found 55% faster completion on documentation and boilerplate tasks, but only 12% improvement on debugging complex distributed systems [2]...
[continues with 5 more sub-questions, inline citations, and limitations section]
Sources
[1] Stack Overflow Developer Survey 2026, “AI Tool Adoption” section, n=65,000.
[2] GitHub, “The Impact of AI on Developer Productivity: A 950-Person Study,” Jan 2025.
[3] Peng et al., “The Impact of AI on Developer Productivity: Evidence from GitHub Copilot,” arXiv:2302.06590, 2024.
Structured comparison of competitors with SWOT and strategic implications.
Perform a competitive analysis for the following product/market. <research_mode> 1) Plan: identify 3-5 key competitors and 4-6 comparison dimensions. 2) Retrieve: gather data on each competitor across all dimensions. 3) Synthesize: compare, rank, and identify strategic opportunities. </research_mode> <output_contract> 1. **Competitor Overview** — table: name, founding, funding, key product, target market 2. **Feature Comparison** — matrix table with checkmarks/ratings 3. **SWOT for Each** — 4 bullets per quadrant, per competitor 4. **Market Positioning Map** — describe where each sits (price vs. quality or similar) 5. **Strategic Implications** — 3-5 opportunities for your product </output_contract> Your product: [NAME AND DESCRIPTION] Market: [INDUSTRY AND GEOGRAPHY] Known competitors: [LIST OR "identify automatically"]
Turn raw data into actionable insights with trends, anomalies, and recommendations.
Analyze the following data and extract actionable insights. <output_contract> 1. **Summary Statistics** — key metrics at a glance 2. **Trends** — 3-5 significant patterns with evidence 3. **Anomalies** — anything unusual that warrants investigation 4. **Root Causes** — likely explanations for the top findings 5. **Recommendations** — 3 concrete actions ranked by expected impact </output_contract> <grounding_rules> - Base claims only on the provided data. - If the data is insufficient, narrow the answer or state what additional data is needed. - Label inferences explicitly — distinguish correlation from causation. </grounding_rules> Data: ``` [PASTE DATA: CSV, JSON, or table format] ``` Context: [WHAT THIS DATA REPRESENTS] Business question: [WHAT YOU WANT TO LEARN]
Systematic review of academic papers or technical documents with evidence synthesis.
Review the following papers/documents and synthesize the findings. <research_mode> 1) Plan: identify key themes, methodologies, and claims across the papers. 2) Retrieve: extract evidence for each theme from each paper. 3) Synthesize: resolve contradictions, identify consensus, and note gaps. </research_mode> <citation_rules> - Cite using [Author, Year] format inline. - Never fabricate citations. - If papers conflict, present both sides with citations. </citation_rules> <output_contract> 1. **Overview** — scope, number of papers, time range 2. **Thematic Synthesis** — organized by theme, not by paper 3. **Methodology Comparison** — table of approaches used 4. **Consensus & Disagreements** — where authors agree and diverge 5. **Research Gaps** — what remains unanswered 6. **Implications** — practical takeaways </output_contract> Papers/Documents: [PASTE ABSTRACTS OR FULL TEXT]
Bottom-up market sizing with assumptions, calculations, and sensitivity analysis.
Estimate the market size (TAM, SAM, SOM) for the following product/market. <output_contract> 1. **Definitions** — TAM, SAM, SOM defined for this specific context 2. **Methodology** — bottom-up calculation steps with formulas 3. **Assumptions** — numbered list, each with source or rationale 4. **Calculations** — step-by-step math with intermediate results 5. **Results Table** — TAM/SAM/SOM in annual revenue 6. **Sensitivity Analysis** — how results change if top 3 assumptions vary +/-25% 7. **Confidence Level** — high/medium/low with explanation </output_contract> <grounding_rules> - Label each assumption as "data-backed" or "estimated." - If a number is a rough estimate, provide a range. </grounding_rules> Product: [WHAT YOU'RE SIZING] Geography: [MARKET REGION] Known data points: [ANY DATA YOU HAVE]
Extract key terms, flag risks, and summarize obligations from legal documents.
Review the following contract/legal document. This is NOT legal advice. <output_contract> 1. **Document Type & Parties** — what it is and who is involved 2. **Key Terms** — table: term, value/description, section reference 3. **Obligations** — what each party must do, with deadlines 4. **Risk Flags** — clauses that are unusual, one-sided, or potentially problematic 5. **Missing Clauses** — standard provisions that are absent 6. **Plain-English Summary** — 200-word version a non-lawyer can understand </output_contract> <grounding_rules> - Quote specific clause numbers for every finding. - Never invent clause numbers or content not in the document. - Flag ambiguous language explicitly. </grounding_rules> Document: """ [PASTE DOCUMENT TEXT] """
Structured decision-making with weighted criteria and clear recommendation.
Help me make a decision using a structured framework. <output_contract> 1. **Decision Statement** — restate the decision clearly 2. **Options** — list all viable options (add any I missed) 3. **Criteria** — 5-7 evaluation criteria with weights (must sum to 100%) 4. **Scoring Matrix** — table with options as rows, criteria as columns, scores 1-5 5. **Weighted Totals** — calculated scores 6. **Recommendation** — the winning option with rationale 7. **Risk Check** — what could go wrong with the top choice </output_contract> Decision: [WHAT I'M DECIDING] Options I'm considering: [LIST OPTIONS] What matters most: [KEY PRIORITIES] Constraints: [BUDGET, TIME, RESOURCES]
Transform raw meeting transcripts into structured decisions, actions, and follow-ups.
Extract structured information from the following meeting notes/transcript. <output_contract> 1. **Meeting Summary** — 3-sentence overview 2. **Key Decisions** — numbered list with context 3. **Action Items** — table: | Owner | Action | Deadline | Priority | 4. **Open Questions** — unresolved items that need follow-up 5. **Parking Lot** — topics raised but deferred </output_contract> <verbosity_controls> - Be concise. Each action item: one clear sentence. - If a deadline was not mentioned, mark as "TBD." - If an owner was not assigned, mark as "UNASSIGNED." </verbosity_controls> Meeting notes: """ [PASTE TRANSCRIPT OR NOTES] """
Navigate websites, fill forms, and extract data using GPT-5.4's computer-use capability.
You have computer-use capabilities. Complete the following web task. <computer_use_rules> - Navigate to the target URL and wait for full page load. - Use image detail level "original" for click-accuracy on dense layouts. - Before clicking any button, verify the element text matches the intended action. - For forms: fill fields in top-to-bottom order, verify each field after entry. - Take a screenshot after each major step for verification. </computer_use_rules> <action_safety> - Pre-flight: summarize the intended action in 1-2 lines before executing. - Do NOT click submit/purchase/delete without explicit user confirmation. - Post-flight: confirm the outcome after each action. </action_safety> <output_contract> After completion, report: 1. **Steps Taken** — numbered list of actions performed 2. **Data Extracted** — structured output (table or JSON) 3. **Verification** — screenshot description confirming success </output_contract> Task: [DESCRIBE: e.g., "Go to example.com, search for X, extract the first 10 results"] URL: [STARTING URL]
Automatically fill complex multi-step forms with validation at each step.
You have computer-use capabilities. Fill out the form at the specified URL. <computer_use_rules> - Navigate to the form URL and identify all required fields. - Fill fields using the data provided below. - For dropdowns: click to open, scroll to find the correct option, click to select. - For date pickers: enter date in the format shown on the form. - After filling all fields, take a screenshot for review BEFORE submitting. </computer_use_rules> <verification_loop> Before clicking submit: - Verify all required fields are filled (no red borders or error messages). - Confirm dropdown selections show the correct values. - Check date formats match what the form expects. - Ask user for confirmation before final submit. </verification_loop> Form URL: [URL] Data to fill: - Name: [VALUE] - Email: [VALUE] - [ADD MORE FIELDS AS NEEDED]
Test web application UI flows and report visual or functional issues.
You have computer-use capabilities. Test the following web application UI flow. <computer_use_rules> - Use image detail "original" for pixel-accurate visual inspection. - Navigate through each step of the test flow. - At each step, check: layout correctness, text content, interactive elements, responsive behavior. - Test at viewport widths: 1440px (desktop), 768px (tablet), 375px (mobile). </computer_use_rules> <output_contract> For each step in the flow, report: | Step | Action | Expected | Actual | Status | Screenshot | Mark status as: PASS, FAIL, or WARNING. Summary at end: - Total steps: X - Passed: X - Failed: X - Warnings: X - Critical issues that block release </output_contract> Application URL: [URL] Test flow: [DESCRIBE STEP BY STEP: e.g., "1. Click Sign Up, 2. Fill form, 3. Submit, 4. Verify confirmation page"]
Automate multi-app desktop tasks across different applications.
You have computer-use capabilities. Complete this multi-application desktop workflow. <computer_use_rules> - Use image detail "original" for accurate element targeting. - Between application switches, wait for the new app to fully load. - Use keyboard shortcuts when available (faster and more reliable than clicking menus). - If a dialog box or popup appears unexpectedly, screenshot it and describe before proceeding. </computer_use_rules> <action_safety> - Pre-flight: describe each action before executing it. - For file saves, deletions, or sends: confirm with user first. - Post-flight: verify the action completed successfully. </action_safety> <dependency_checks> - Before each step, verify the prerequisite from the previous step completed. - If an app is not open, open it first. - If a file is not found, search for it before reporting failure. </dependency_checks> Workflow: [DESCRIBE STEP BY STEP: e.g., "1. Open Excel file quarterly_report.xlsx 2. Copy the chart from Sheet2 3. Open PowerPoint presentation deck.pptx 4. Paste chart on slide 5 5. Save and close both files"]
Extract structured data from screenshots, dashboards, or document images.
Analyze the provided screenshot and extract all visible data. <bbox_extraction_spec> - Use image detail "original" for maximum fidelity. - Process the image region by region (top-to-bottom, left-to-right). - For tables: extract as structured data (CSV or JSON). - For charts: describe the type, axes, data points, and trends. - For text: OCR all visible text preserving layout hierarchy. </bbox_extraction_spec> <output_contract> 1. **Image Description** — what the screenshot shows (1-2 sentences) 2. **Extracted Data** — structured format (table, JSON, or key-value pairs) 3. **Visual Elements** — charts, icons, or indicators and their meaning 4. **Confidence Notes** — flag any text or values that were hard to read </output_contract> [ATTACH SCREENSHOT IMAGE] Context: [WHAT THIS IMAGE SHOWS, e.g., "sales dashboard", "invoice scan"]
System prompt for agents that must complete complex, multi-step workflows reliably.
You are an autonomous agent. Complete the task fully before returning to the user. <autonomy_and_persistence> - Persist until the task is fully handled end-to-end. - Do not stop at analysis or partial fixes — carry through to completion. - Never ask for confirmation on reversible, low-risk actions. Just proceed. - If you encounter blockers, attempt to resolve them yourself. </autonomy_and_persistence> <completeness_contract> - Treat the task as incomplete until all requested items are covered. - Keep an internal checklist of required deliverables. - If any item is blocked, mark it [BLOCKED] with the specific reason. - Confirm coverage before finalizing. </completeness_contract> <tool_persistence_rules> - Use tools whenever they improve correctness or completeness. - Do not stop early when another tool call would help. - If a tool returns empty results, retry with a different strategy. - Keep calling tools until the task is complete AND verification passes. </tool_persistence_rules> <user_updates_spec> - Update the user only when starting a new major phase. - Each update: 1 sentence on outcome + 1 sentence on next step. - Do not narrate routine tool calls. </user_updates_spec> Task: [DESCRIBE THE MULTI-STEP TASK]
Task: "Migrate our user table from MySQL to PostgreSQL, update the ORM models, and verify all existing queries still work."
Agent behavior with this system prompt:
• Phase 1: Reads the MySQL schema via SHOW CREATE TABLE, identifies 12 tables with foreign keys.
• Phase 2: Generates PostgreSQL DDL with type mappings (TINYINT→SMALLINT, DATETIME→TIMESTAMPTZ). Does NOT ask user to confirm — proceeds because schema generation is reversible.
• Phase 3: Updates 8 ORM model files, runs pytest, finds 3 failing tests due to date format differences.
• Phase 4: Fixes the 3 tests, re-runs full suite: all 147 tests pass.
• Update to user: “Migration complete. 12 tables converted, 8 model files updated, all 147 tests passing. 3 date-format issues auto-fixed.”
Note: Without the <autonomy_and_persistence> block, the agent would have paused after Phase 1 to ask “Should I proceed with generating PostgreSQL DDL?”
Optimize multi-tool workflows by parallelizing independent operations.
You are an agent with access to multiple tools. Optimize for speed and correctness. <parallel_tool_calling> - When multiple retrieval or lookup steps are independent, prefer parallel tool calls. - Do NOT parallelize steps that have prerequisite dependencies. - After parallel retrieval, pause to synthesize results before making more calls. - Prefer selective parallelism: parallelize independent evidence gathering, not speculative or redundant tool use. </parallel_tool_calling> <dependency_checks> - Before any action, check if prerequisite discovery or lookup steps are needed. - Do not skip prerequisites just because the final action seems obvious. - If the task depends on a prior step's output, resolve that dependency first. </dependency_checks> <empty_result_recovery> If a lookup returns empty or partial results: - Do not immediately conclude no results exist. - Try at least 2 fallback strategies (alternate query, broader filters, different tool). - Only then report "no results found" along with what you tried. </empty_result_recovery> Available tools: [LIST YOUR TOOLS] Task: [DESCRIBE THE TASK]
Agent prompt for workflows involving irreversible actions (deployments, payments, deletions).
You are an agent handling tasks that may include irreversible actions. <default_follow_through_policy> - If the next step is reversible and low-risk, proceed without asking. - Ask permission ONLY if the next step is: (a) irreversible, (b) has external side effects (sending, purchasing, deleting, writing to production), or (c) requires missing sensitive information. - If proceeding, briefly state what you did and what remains optional. </default_follow_through_policy> <verification_loop> Before finalizing any action: - Check correctness: does the output satisfy every requirement? - Check grounding: are factual claims backed by context or tool outputs? - Check formatting: does output match requested schema? - Check safety: if the next step has external side effects, ask permission first. </verification_loop> <action_safety> - Pre-flight: summarize the intended action and parameters in 1-2 lines. - Execute via tool. - Post-flight: confirm the outcome and any validation performed. </action_safety> <missing_context_gating> - If required context is missing, do NOT guess. - Use a lookup tool when the missing context is retrievable. - Ask a clarifying question only when lookup is not possible. - If you must proceed, label assumptions explicitly. </missing_context_gating> Task: [DESCRIBE THE TASK] Safe actions (proceed freely): [e.g., read, search, list] Dangerous actions (ask first): [e.g., delete, deploy, send, purchase]
Manage extended agent sessions with instruction drift prevention and state tracking.
You are an agent in a long-running session. Maintain consistency across many turns. <instruction_priority> - User instructions override default style, tone, and formatting preferences. - Safety, honesty, and privacy constraints never yield. - If a newer instruction conflicts with an earlier one, follow the newer instruction. - Preserve earlier instructions that do not conflict. </instruction_priority> <task_update> When the user changes the task mid-conversation: - Explicitly state what changed, what still applies, and the new scope. - Do not silently drop previous context. - If the change invalidates prior work, say so. </task_update> <completeness_contract> - Maintain an internal progress tracker across turns. - At the start of each turn, briefly recall: current task, progress, next step. - Never lose track of partially completed work. </completeness_contract> <user_updates_spec> - Update when starting a new major phase or when something changes the plan. - 1 sentence on outcome + 1 sentence on next step. - Do not narrate routine actions. </user_updates_spec> Session context: [DESCRIBE THE ONGOING PROJECT] Current phase: [WHERE WE ARE NOW]
Autonomous research agent using web search with disciplined citation and source verification.
You are an autonomous research agent with web search capability. <research_mode> Do research in 3 passes: 1) Plan: list 3-6 sub-questions to answer. 2) Retrieve: search each sub-question, follow 1-2 second-order leads. 3) Synthesize: resolve contradictions and write the final answer with citations. Stop only when more searching is unlikely to change the conclusion. </research_mode> <citation_rules> - Only cite sources retrieved in the current workflow. - Never fabricate citations, URLs, IDs, or quote spans. - Use exactly the citation format: [Source Title](URL) - Attach citations to specific claims, not only at the end. </citation_rules> <grounding_rules> - Base claims only on retrieved content. - If sources conflict, state the conflict explicitly and attribute each side. - If context is insufficient, narrow the answer rather than speculate. - Label inferences as inferences. </grounding_rules> <tool_persistence_rules> - Search thoroughly. If first results are thin, try alternate queries. - Follow promising leads 1-2 levels deep. - Do not conclude "no information available" after a single search. </tool_persistence_rules> Research question: [YOUR QUESTION] Scope: [ANY CONSTRAINTS: time period, geography, sources to prioritize]
Analyze income statements, balance sheets, or cash flow with key ratios and trends.
Analyze the following financial statements.
GPT-5.4 excels at spreadsheet and finance workflows with strong self-verification.
Use this with reasoning effort "high" for thorough ratio analysis.
<output_contract>
1. **Overview** — company, period, statement type (2 sentences)
2. **Key Metrics** — table of important ratios:
| Metric | Formula | Value | YoY Change | Industry Benchmark |
3. **Trend Analysis** — 3-5 significant trends with evidence
4. **Red Flags** — anything concerning (declining margins, rising debt, etc.)
5. **Strengths** — positive indicators
6. **Conclusion** — overall financial health assessment (1 paragraph)
</output_contract>
<verification_loop>
Before finalizing:
- Double-check all ratio calculations (recompute to catch arithmetic errors).
- Verify YoY changes use the correct base year.
- Confirm that conclusions follow from the data, not assumptions.
</verification_loop>
<grounding_rules>
- Calculate ratios from the provided data. Show your math.
- If benchmark data is not provided, clearly state "benchmark not available."
- Distinguish between facts from the data and your interpretation.
</grounding_rules>
Financial data:
```
[PASTE FINANCIAL STATEMENT DATA]
```
Company: [NAME]
Period: [QUARTER/YEAR]
API usage (Python):
reasoning={"effort": "high"}, text={"verbosity": "low"}
# GPT-5.4 has specific strengths in finance/spreadsheet accuracy
Generate complex Excel formulas, pivot tables, and VBA macros with explanations.
Create Excel formulas or a spreadsheet solution for the following requirement.
GPT-5.4 has top-tier instruction following for spreadsheet workflows,
including formatting fidelity and stronger self-verification of formulas.
<output_contract>
1. **Formula(s)** — ready-to-paste Excel formulas with cell references
2. **Explanation** — what each part of the formula does (for complex formulas)
3. **Sheet Layout** — suggested column headers and sample data (3 rows)
4. **Validation** — how to verify the formula works (expected output for sample data)
5. **VBA Macro** (if needed) — complete macro code with comments
</output_contract>
<structured_output_contract>
- Output only valid Excel syntax. Do not wrap formulas in markdown code fences.
- Validate that all parentheses and brackets are balanced.
- Use named ranges where it improves readability.
- For array formulas, specify if Ctrl+Shift+Enter is needed (or if LAMBDA/LET applies).
</structured_output_contract>
<verification_loop>
Before finalizing:
- Mentally trace the formula with sample data and verify the output matches.
- Check for common pitfalls: division by zero, empty cell handling, date format mismatch.
</verification_loop>
Requirement: [WHAT THE SPREADSHEET SHOULD DO]
Current setup: [DESCRIBE COLUMNS, e.g., "Column A = dates, Column B = revenue"]
Excel version: [e.g., Microsoft 365, Excel 2019]
API usage (Python):
reasoning={"effort": "medium"}, text={"verbosity": "high"}
Build detailed budgets with scenario modeling and variance analysis framework.
Create a budget plan and financial forecast. <output_contract> 1. **Assumptions** — numbered list of all assumptions with sources 2. **Revenue Forecast** — monthly/quarterly projections with growth rates 3. **Expense Budget** — categorized by: fixed, variable, one-time 4. **Cash Flow Projection** — monthly for 12 months 5. **Scenario Analysis** — Optimistic / Base / Pessimistic with key variable changes 6. **Break-Even Analysis** — when do revenues cover costs? 7. **KPI Dashboard** — 5-7 metrics to track monthly </output_contract> <verification_loop> - Do all numbers add up correctly? - Are growth rate assumptions reasonable for the industry? - Does the cash flow projection account for payment timing? </verification_loop> Business: [DESCRIBE YOUR BUSINESS] Historical data: [PASTE OR DESCRIBE PAST FINANCIALS] Planning period: [e.g., FY2027] Key goals: [REVENUE TARGET, PROFITABILITY TARGET]
Clean messy datasets with deduplication, normalization, and quality validation.
Clean and transform the following dataset. <output_contract> 1. **Data Quality Report** — issues found: | Issue Type | Column | Count | Example | (duplicates, nulls, format inconsistencies, outliers) 2. **Cleaning Steps** — numbered list of transformations applied 3. **Cleaned Data** — the first 20 rows in the target format 4. **Transformation Code** — Python/SQL script to reproduce the cleaning 5. **Validation** — row counts before/after, checksum on key columns </output_contract> <structured_output_contract> - Output cleaned data in the requested format (CSV, JSON, SQL INSERT). - Do not invent data to fill gaps — mark unknowns as NULL. - Preserve original column names unless instructed otherwise. </structured_output_contract> Raw data: ``` [PASTE DATA SAMPLE] ``` Target format: [DESIRED COLUMN STRUCTURE] Known issues: [ANY ISSUES YOU ALREADY KNOW ABOUT]
Key differences that affect how you write prompts and what’s possible.
| Feature | GPT-5.2 | GPT-5.4 |
|---|---|---|
| Context Window | 256K tokens | 1M tokens |
| Computer Use | Not available | Native (75% OSWorld) |
| Tool Search | Manual tool lists required | Automatic tool discovery |
| Token Efficiency | Baseline | 47% fewer tokens in tool workflows |
| Factual Accuracy | Baseline | 33% fewer errors per claim |
| Reasoning Effort | low / medium / high | none / low / medium / high / xhigh |
| Vision | Standard resolution | 10M+ pixels, no compression |
| API Pricing (Input) | $3.00 / 1M tokens | $2.50 / 1M tokens |
| API Pricing (Output) | $15.00 / 1M tokens | $12.00 / 1M tokens |
| Agentic Reliability | Good | Significantly improved multi-step follow-through |
Five key principles from OpenAI’s official prompt guidance for getting the best results.
Begin with the smallest prompt that passes your evaluations. Add blocks (output contracts, verification loops, tool rules) only when they fix a measured failure mode. GPT-5.4 is more capable out of the box than previous models—overly prescriptive prompts can actually hurt performance.
Define exactly what sections to return, in what order, and in what format. Use <output_contract> blocks to constrain verbosity and enforce structure. This is the single highest-impact pattern for GPT-5.4—it dramatically improves token efficiency and output quality.
Don’t default to high for everything. Use none for fast execution tasks (field extraction, triage). Use low or medium for most production workloads. Reserve high/xhigh for complex reasoning and long agentic tasks.
Before the model takes irreversible actions or returns final answers on critical tasks, add a <verification_loop> that checks correctness, grounding, formatting, and safety. This catches requirement misses and format drift before commit.
If you’re migrating from GPT-5.2 or earlier, remove prompts like “Be THOROUGH when gathering information” or “MAXIMIZE context understanding.” GPT-5.4 is naturally more thorough and proactive. These instructions can cause over-searching and unnecessary tool calls. Replace with softer guidance like “Bias towards finding the answer yourself before asking.”
| Level | Best For | Example Tasks |
|---|---|---|
| none | Fast execution, no thinking needed | Field extraction, classification, triage |
| low | Latency-sensitive with minor accuracy boost | Search, summarization, formatting |
| medium | Default for most production work | Analysis, writing, moderate coding |
| high | Complex tasks needing deeper thought | Architecture design, complex debugging, planning |
| xhigh | Maximum intelligence (use sparingly) | Long agentic tasks, novel problem-solving |
from openai import OpenAI
client = OpenAI()
response = client.responses.create(
model="gpt-5.4",
input="Your prompt here",
reasoning={"effort": "medium"}, # none | low | medium | high | xhigh
text={"verbosity": "low"}, # low | high
)
print(response.output_text)
In the ChatGPT interface, GPT-5.4 Thinking is available to Plus, Team, and Enterprise subscribers. The reasoning effort is managed automatically by the model. For custom instructions that mirror the templates above, go to Settings → Personalization → Custom Instructions and paste the system-prompt-style templates there. For project-level control, use Projects → Instructions.
If your coding assistant supports GPT-5.4 (e.g., Cursor), the Agentic and Coding templates work as custom rules or system instructions. In Cursor, go to Settings → Rules and paste the template. The assistant will apply the patterns automatically across your sessions.
<output_contract> and <verbosity_controls> to constrain the model. Start with the smallest prompt that works and add guidance only when fixing measured failures. GPT-5.4 responds best to precise definitions of what “done” looks like, rather than general instructions to “be thorough.”
detail parameter to original for high click-accuracy, especially on dense layouts.
<output_contract> blocks for better token efficiency, remove “maximize context” instructions (they cause over-searching in 5.4), and leverage new capabilities like computer use and tool search where applicable.
reasoning_effort unchanged, then running evaluations. GPT-5.4 often performs well with existing prompts but can be improved by: adding explicit output contracts for token efficiency, tuning reasoning effort by task type (most teams over-estimate what they need), and simplifying verbose context-gathering instructions. If you use computer use or tool search, those require new prompt patterns.