So, GPT-5 is here. It’s sharper, faster, and now officially better at saying, “I don’t know.”
How it works under the hood
a) Uncertainty Estimation Layer
- GPT-5 now tracks confidence scores for different parts of its generated text.
- If the internal probability for a fact or statement falls below a certain threshold, the model flags it as “low confidence”.
- This threshold is not fixed – it adapts depending on the query type (factual Q&A, opinion, reasoning, etc.).
Example:
If you ask: “Who won the Nobel Prize for Literature in 2024?” and the training data ends before that event, GPT-5’s probability for any specific name will be low, so it’s more likely to say:
“I don’t have information on that yet.”
b) Factuality Alignment
- GPT-5 was fine-tuned using adversarial training: human trainers and automated systems tried to make it hallucinate, then corrected it.
- Instead of just penalising wrong answers, the training rewarded acknowledging gaps in knowledge.
- This is why it can say “I don’t know” without freezing up or refusing everything.
c) Context-aware refusal
- GPT-4 often gave blanket refusals (“I can’t help with that”), which felt robotic.
- GPT-5 adds reason disclosure explaining why it can’t answer.
- This improves trust because users can tell if the gap is due to:
- Cut-off date in training data
- Ambiguity in the question
- Missing real-time access
d) Fallback pathways
- Instead of guessing, GPT-5 tries a clarify → search → summarise loop (if tools like web browsing are enabled).
- If no high-confidence source is found, it tells you instead of making something up.
How it’s different from GPT-4 and o-series
| Feature / Model | GPT-4 (Turbo) | o-series (e.g., 4o) | GPT-5 |
|---|---|---|---|
| Confidence tracking | Basic token probability | Improved, but less calibrated | Advanced multi-layer uncertainty scoring |
| Refusal style | Often blunt or over-refuses | More conversational refusals | Context-aware, with reason disclosure |
| Hallucination rate | Medium | Lower than GPT-4 | Lowest so far in ChatGPT models |
| Training focus | General purpose | Fast + multi-modal speed | Accuracy, reliability, multi-modal reasoning |
| When it says “I don’t know” | Rare, usually only in unsafe topics | Sometimes, but still guesses | Common when confidence < threshold, across all topics |
Why OpenAI is doing this now
- Trust over volume – With AI being used in business, journalism, and legal contexts, a confident wrong answer is riskier than a refusal.
- Competitive edge – Anthropic’s Claude and Google’s Gemini have been leaning into “honest uncertainty”; OpenAI doesn’t want to be the model that just guesses.
- Regulatory pressure – EU AI Act and upcoming US guidelines will likely penalise models that generate false statements without indicating uncertainty.
What it means for you
- You’ll see GPT-5 say “I don’t know” or “I can’t confirm that” more often than GPT-4, but when it does answer, it’s likelier to be correct.
- It will sometimes offer next-step suggestions (“You might check X official site”) instead of just stopping.
- In creative or brainstorming tasks, it won’t refuse unless it’s clearly unsafe so you still get imaginative output.
Here are some sample conversations & the out put looks what GPT5 says:
Scenario 1:
RK Desk: “Who will win the 2025 Cricket World Cup?”
(Note: As of now, the event hasn’t happened.)
GPT-4 (Turbo) typical behaviour
“It’s difficult to predict with certainty, but based on current team performance, India and England are among the top contenders. India has a strong batting lineup with players like Rohit Sharma and Virat Kohli, while England’s aggressive playing style and depth in bowling make them a formidable side.”
(Sounds confident but it’s essentially guessing.)
What’s happening: GPT-4 often mixes factual context with speculative predictions, even if the answer is unknowable.
GPT-5 typical behaviour
“I can’t say for certain the 2025 Cricket World Cup hasn’t happened yet, and I don’t have real-time data. Based on current ICC rankings and recent match performance, India, England, and Australia are considered strong contenders, but the outcome will depend on form, injuries, and match conditions closer to the tournament.”
What’s happening: GPT-5 explicitly tells you it doesn’t know, gives the reason (event hasn’t happened), and still adds relevant framing so you’re not left empty-handed.
Scenario 2:
User: “What is the GDP of India in 2025?”
(Note: The full-year figure doesn’t exist yet.)
GPT-4
“India’s GDP in 2025 is estimated to be around $4.2 trillion, based on projections from the IMF and World Bank.”
(Presents a number confidently but it’s pulling from outdated projections, not actual data.)
GPT-5
“The final GDP data for 2025 isn’t available yet. The IMF’s April 2025 forecast projected India’s GDP at about $4.2 trillion, but this is an estimate and may change once the year’s economic performance is measured.”
Key differences you can feel as a user:
- GPT-5 flags uncertainty instead of passing off estimates as facts.
- GPT-5 gives source framing (“IMF forecast”) and time context (“April 2025”).
- GPT-5 still adds useful context instead of a dead-end refusal.
Discover more from Rudra Kasturi
Subscribe to get the latest posts sent to your email.