← Back to all posts

ChatGPT-5 Test Series - Part 1: Conceptual Work in Action

by Elena Jäger
Sep 16, 2025
Connect

🕓 Read Time: ~5 minutes

When OpenAI rolled out GPT-5, I asked you what I should test first. Thanks for all your input — this is Part 1 of the series, where I dig into how it handles conceptual work.

By “conceptual work,” I mean the kind of big-picture, structural thinking that many of us coaches, consultants, and solopreneurs do every week. Think:

  • Redesigning a curriculum or program

  • Structuring a new offer

  • Reworking core business workflows

I put GPT-5 to the test on two real projects:

  1. Updating one of my AI-focused offerings

  2. Refreshing one of my core workflows

In this edition, I won't talk about tool specs (I'm too lazy to actually read them in much detail). I always test tools the same way I expect my clients to: in real workflows, with real stakes.

So what changed between GPT-4o and GPT-5?

Many things have changed. The focus here is on my experience with conceptual work. At first, I tried GPT-5 in auto mode, where it decides whether to respond fast, “thinking,” or “long thinking.” To be honest, it didn’t feel like a big leap forward. The outputs looked similar to GPT-4o.

But then I switched to manual model selection. Instead of letting it choose, I selected the middle “thinking” mode. 

That’s when things changed.

Suddenly, it felt like working with a colleague at the whiteboard. We bounced ideas back and forth. It structured concepts with me, tightened the flow, and worked at a speed that felt better than 4o. When I say the speed felt "better", that doesn't mean ChatGPT responded faster. It actually responded more slowly. But the quality of the responses was way better, saving me time on iterations.

For me, the co-creation experience was noticeably sharper.

The big surprise: GPT-5 is unforgiving with bad prompts

Here’s the part that really stood out: GPT-5 will not rescue you from a bad prompt.

Where GPT-4o often managed to turn a vague or fuzzy ask into something usable, GPT-5 is less forgiving. If your input is unclear, your output will be too. Good old "garbage in, garbage out".

If I'm perfectly honest, I actually think that’s a good thing:

Why?
GPT-4o would often “fill in the blanks” when a prompt was vague. Sounds helpful but in practice, it often invented details, produced generic content, and gave results that felt flat.

GPT-5 forces you to do your part: to be explicit, structured, and intentional. The trade-off? If you put in the work, you get output that’s sharper, more relevant, and actually usable.

And it’s not just me noticing this. In fact, OpenAI has released an official GPT-5 Prompting Guide — which in itself is a clear signal that things are changing. Prompting isn’t an afterthought anymore... It’s a key skill to have.

This is where GPT-5 aligns beautifully with my favorite prompting formula:

  1. Frame the role: “You are an instructional designer…”

  2. Set the constraints: length, style, tone, audience.

  3. Specify the output format: outline, table, list, etc.

  4. Iterate interactively: ask it to critique, refine, or re-organize its own answers.

  5. Check its work: prompt it to suggest improvements.

With this approach, GPT-5 doesn’t just respond, it collaborates.

It sounds different, more "normal"

One more thing I noticed: the tone shift.

GPT-4o can sometimes feel like a hyperactive co-worker: over-eager, full of emojis, always saying “Yes, absolutely!”.

GPT-5 doesn’t do that. It’s calmer, more neutral, even slightly robotic at times. Of course, I can always instruct it to be more upbeat, supportive or whatever. But that's up to me.

For conceptual work, I actually liked this. It made the collaboration feel more grounded.

Will that tone still work for more emotional or creative writing? That’s a test that deserves its own edition.

And one more thing: Your CustomGPTs may need a review

As many of you know, most of my daily work runs through my virtual team members — the custom GPTs (aka "Assistants") I have built for specific workflows.

With GPT-5, I found that most of them still work just fine but not all. The more structured ones survived the update well. The less detailed ones? They needed tweaking.

So I’ve been reviewing them one by one, and here’s what I’d suggest if you rely on custom GPTs:

  • Re-check the instructions: Make them explicit and step-by-step.

  • Set the default model: Decide whether each GPT should run on “fast” or “thinking.”

  • Test for clarity: Run them on your core tasks and adjust where they stumble.

One surprising discovery: you can actually ask GPT-5 to rewrite your custom GPT instructions for GPT-5 compatibility. And it does a pretty good job.

Key Takeaway
GPT-5 is powerful for big-picture thinking... but only if your prompts are structured, specific, and intentional. Unlike earlier models, it won’t cover for vagueness. That’s a good thing: it pushes you to get clear, and rewards you with sharper, more original results.

Til next time, keep working on those prompting skills!
Elena

 

P.S. Want to explore OpenAI's ChatGPT 5 Prompting Guide? Here it is. 

P.P.S. On a another note. Next week (24 Sep @ 3:15 p.m. CEST), I’ll be joining Václav Sulista, entrepreneur career consultant, live on LinkedIn to talk about how to make AI your career advantage (instead of your fear). If this has been on your mind, you can sign up here and bring your questions. 

 

 

The brief AI never forgets.
  ⏱️ Read time: ~4 min You have probably re-explained the same task to AI more times than you can count. Same instructions, different conversation, same frustration. There is a better way. Last time I shared two prompts to get more from AI. Today we go deeper. Because before you can use Skills well, you need to understand what they actually are. So, what is a Skill? A Skill is a ...
2 Prompts. Less Time. Better AI Output
⏱️ Read time: ~3 min Today I'm sharing two prompts that change how you work with AI, and one of them is brand new. One of the most common frustrations I hear from coaches and consultants about AI goes something like this: "I tried it, the output was terrible, and I don't see the point." And almost every time, the real problem isn't the AI. It's the user input. AI won't fix a lack of clarity. It...
On Curiosity, Claude, And Knowing When To Stop
⏱️ Read time: ~4 min I almost lost an entire evening to agentic dashboards. There I was, using Claude Code to generate dashboards on random topics, then building landing pages, then exploring what else it could do. One thing led to another, and before I knew it, hours had passed. I, the person who preaches pragmatic AI use, had completely abandoned my usual discipline. 🤯 And honestly?...

Not signed up yet?
Do it right here:

© 2026 Future of Work
Privacy Policy Home

JOIN THE VIP LIST

Name of Free Resource

Get started today before this once in a lifetime opportunity expires. Get started today before this once in a lifetime opportunity expires.