SUBSCRIBE
Tech Journal Now
  • Home
  • News
  • AI
  • Reviews
  • Guides
  • Best Buy
  • Software
  • Games
  • More Articles
Reading: If at first you don’t succeed, prompt, prompt again – GeekWire
Share
Tech Journal NowTech Journal Now
Font ResizerAa
  • News
  • Reviews
  • Guides
  • AI
  • Best Buy
  • Games
  • Software
Search
  • Home
  • News
  • AI
  • Reviews
  • Guides
  • Best Buy
  • Software
  • Games
  • More Articles
Have an existing account? Sign In
Follow US
© Foxiz News Network. Ruby Design Company. All Rights Reserved.
Tech Journal Now > News > If at first you don’t succeed, prompt, prompt again – GeekWire
News

If at first you don’t succeed, prompt, prompt again – GeekWire

News Room
Last updated: May 3, 2026 8:29 pm
News Room
Share
13 Min Read
SHARE
An AI prompt screen, as reimagined by Google Gemini.

[Editor’s Note: This is the third in a series by Oren Etzioni about AI usage and best practices. See also “AI Coach or AI Ghostwriter? The Choice Is Yours,”  and “How to read with AI.”]

A friend asked ChatGPT for input on a professional matter and received a banal, lackluster response. I suggested she try a different approach: ask for 15 different ideas, scan them, pick the two that felt most promising, and then ask ChatGPT to refine. She came back overjoyed. ChatGPT had not gotten smarter, but she became better at prompting.

This is my favorite gambit: ask AI for many options, delve deeper into the promising ones, and most importantly, if at first you don’t succeed, prompt, prompt again!

What follows is practical advice on how to use AI as a power tool rather than a slot machine. For a simple request, it’s overkill, but if you’re serious about prompting, read on.

Anthropic’s own guidance for prompting Claude contains a helpful hint: treat the model as a brilliant but literal-minded new employee on their first day. They are capable. They are also new. They will do exactly what you ask, so you have to ask exactly what you want.

The Anthropic team’s golden rule is to show your prompt to a colleague with no context and ask whether they could follow it. If the answer is no, the model can’t either. This principle generates a handful of habits that lift output quality immediately, before any of the more advanced techniques come into play.

One caveat from me, though: don’t think of the model as a person. It’s not. The “brilliant new employee” framing is a useful starting point, but it’s a metaphor, not reality. A new hire asks follow-up questions, remembers what you said yesterday, and notices when an instruction is dumb. Claude does none of that by default. Lean on the metaphor to remember to be specific and provide context, but drop it the moment you start to expect human judgment that just isn’t there.

Here’s the playbook, organized as a list for easy reference and periodic review.

Be specific about format, length, audience, and constraints.

Vague prompts produce vague output. The fix is to say what you actually want.

  • Before: Write about marketing trends.
  • After: Analyze the three most significant B2B SaaS marketing trends from the past six months. For each, give one company example and a one-sentence assessment of whether the trend will accelerate or plateau. Write it as a 400-word brief for a non-technical board.

Improving prompt quality is often simply stating constraints. Vague prompts produce safe, hedged, encyclopedic answers because the model has no signal about what to optimize for and defaults to coverage. Specific prompts produce opinionated, useful answers because the constraints eliminate the safe-but-useless options. Asking for “three” instead of “some” forces ranking. Asking for “accelerate or plateau” forces a call. Asking for “a board brief” determines what gets cut. Each constraint you add is a decision the model no longer gets to dodge.

Provide a few examples.

This is the highest-leverage move in everyday prompting. Models pick up patterns from examples faster than from descriptions.

  • Before: Turn these meeting notes into action items.
  • After: Turn these meeting notes into action items. Match this format: Example 1: Note: “Sarah will look into the pricing question and get back to us next week.” Action item: Sarah → research pricing options → due next Friday. Example 2: Note: “We agreed to push the launch.” Action item: Team → revise launch timeline → due before Monday’s standup. Now do the same for these notes: [paste]

Tell the model what to do, not what not to do.

Negative instructions are easier to violate than positive ones. Reframing in the affirmative gets you cleaner results.

  • Before: Don’t be too formal. Don’t use jargon. Don’t make it boring.
  • After: Write in a warm, conversational tone, the way a smart colleague would explain this over coffee. Use plain English and short sentences.

Match the style of your prompt to the style of the output you want.

This one surprises some people. If your prompt is full of bullets and bold text, the model will return bullets and bold text. If you want flowing prose, write in flowing prose.

These habits sound modest. But applied together, they take prompts from the level my friend was operating at, where ChatGPT seemed unhelpful, to a level where AI yields dividends left and right. The advanced techniques in the rest of this piece build on this foundation, but they won’t rescue a prompt that fails the basics.

Beyond the basics, here is a set of effective habits that show up in guidance from OpenAI, Google, working developers, and the people who build production AI systems for a living. These are not techniques so much as workflow disciplines.

Iterate; treat prompting as test-driven.

Your first prompt is a draft. The most experienced practitioners build small sets of test cases (the inputs they care about), run their prompt across them, and refine until the output is consistently good. Several open-source toolkits exist to formalize this loop.

  • Before: Write the prompt. Try it on one example. Looks good. Ship it.
  • After: Write the prompt. Pick five inputs, including the awkward edge cases. Run the prompt on all five. Where it fails, change one thing in the prompt and retest. Keep the version that works on the most cases.

Specify a definition of done.

OpenAI’s own guidance for GPT-5 stresses telling the model what counts as a finished answer. Without that, the model decides for itself, often by stopping at the first plausible-looking response.

  • Before: Help me debug this Python error.
  • After: Help me debug this Python error. You are done when: (1) you have identified the root cause, (2) you have proposed a specific fix with the corrected code, and (3) you have explained why the original failed. If you are not confident on any of those three, say so explicitly rather than guessing.

Calibrate effort to the task.

Modern reasoning models have effort or thinking dials. Low effort for extraction and triage; high for synthesis and strategy. Most users leave them on default and pay for it on hard problems.

  • Before: Summarize this 80-page report. 
  • After: Set thinking effort to high. Read the entire report. Identify the three most important findings, the two weakest claims, and the one question I should ask the authors. Cite page numbers.

Inject current or proprietary context directly.

Be careful to avoid jargon and abbreviations unknown to the model (instead of the acronym PMO, say “Project Management Office”). Models don’t have access to your internal documents. Paste in the relevant material.

  • Before: How should I structure a related work section comparing my framework to prior agent governance proposals?
  • After: Below is my current draft related work section, plus PDFs of the three papers I am positioning against (pasted). Based only on these sources, identify points of overlap I have not yet acknowledged and any claims in my draft that the cited papers would not actually support.

Build a personal prompt library.

This is a power move for a pro. The patterns that worked yesterday are likely to work tomorrow. Stop rewriting them from scratch. Save the prompts that consistently produce good results, organized by task type. Treat them as living documents, not one-off attempts.

  • Before: Open a new chat. Type out the framing, the constraints, the examples, and the question from memory. Watch yourself forget two of them.
  • After: Open your prompt library. Copy the “draft a memo for my manager” template. Paste in today’s specific topic and source material. Run.

Here are some key don’ts:

Don’t tell reasoning models to “think step by step.”

Models like OpenAI’s o-series and GPT-5 thinking already do that internally. Adding the instruction can hurt rather than help. Save it for the everyday models.

Don’t lean on “do not” or “never” instructions for everything.

Models, especially Gemini, can over-index on broad negative constraints and degrade on basic reasoning. Prefer positive framing: tell the model what to do.

Don’t trust polished prose as evidence of correctness.

Hallucinations are most dangerous when they are well-written. As I pointed out in How to Read with AI, you have to carefully verify AI output.

Don’t use aggressive language (“CRITICAL: You MUST…”).

Modern models are highly responsive to ordinary instructions. Aggressive phrasing can produce overcautious output and triggers refusals. Use normal language.

Don’t include undefined acronyms in your prompt.

They measurably degrade output. For research on the impact of prompt changes see this recent paper on Brittlebench.

Don’t change three things at once when iterating.

When a prompt isn’t working, change one variable, test, then change the next. Otherwise you don’t know what helped.

Don’t assume that the same prompt works across models.

Different model families need different prompting. The same instruction can help one and hurt another. The temperature and effort settings that work for GPT are not the ones that work for Claude or Gemini.

Don’t treat the first answer as the final one.

Failing to iterate is a common failure mode in everyday AI use. Here’s a trick for making AI better at multi-step tasks: after each attempt, have the AI write a short critique of what went wrong and tuck that note into its memory for the next try. No fancy mechanics, just the model “talking to itself” in plain English. On the next attempt, it reads its own past reflections and adjusts. This loop can produce meaningful gains over one-shot prompts.

The people who get the most out of AI aren’t the ones with the best prompt templates. They’re the ones who treat the model as a powerful tool for advancing their work. You don’t need to show up with perfect clarity about what you want. A good dialog can get you there, surfacing options and questions you’d have missed on your own. What it can’t do is recognize the right answer when it appears. That part is still on you.

For further reading …

Provider documentation:

Practitioner resources:

Editor’s note: GeekWire publishes guest opinions to foster informed discussion and highlight a diversity of perspectives on issues shaping the tech and startup community. If you’re interested in submitting a guest column, email us at [email protected]. Submissions are reviewed by our editorial team for relevance and editorial standards.

Read the full article here

You Might Also Like

Microsoft nixes NDAs with local governments worldwide when deploying data centers

From the farm to space, Next Tech Titan finalists growing to meet big challenges – GeekWire

GeekWire Awards: 5 Workplace of the Year finalists lean on key values to build solid company culture

How AI is changing the business and art of video — from ‘chaos machine’ to creative catalyst

Defense giant Anduril is quietly building autonomous warships on Seattle’s historic ship canal – GeekWire

Share This Article
Facebook Twitter Email Print
Leave a comment Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

- Advertisement -
Ad image

Trending Stories

Games

A group of Diablo 4: Lord of Hatred players hijacked the top leaderboard ranks not for glory, but to promote their expertise at making the strongest character builds

May 3, 2026
Games

Skyrim has been the most-played ‘Unsupported’ game on Steam Deck for 5 months

May 3, 2026
Games

Nearly 40 years after launch, NetHack just got a 5.0.0 release with over 3,000 patch notes

May 3, 2026
Games

XCOM is getting a miniatures game from the company that brought Fallout to tabletop

May 3, 2026
Games

Writing one of Final Fantasy 14’s sauciest songs required its lyricist to consult the tomes—Wikipedia’s incredibly thorough BDSM page

May 3, 2026
Games

The creator of Infinite Craft has a new game where you explore a secret-filled social hub as your mouse cursor

May 3, 2026

Always Stay Up to Date

Subscribe to our newsletter to get our newest articles instantly!

Follow US on Social Media

Facebook Youtube Steam Twitch Unity

2024 © Prices.com LLC. All Rights Reserved.

Tech Journal Now

Quick Links

  • Privacy Policy
  • Terms of use
  • For Advertisers
  • Contact
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?