The End of Bullshit AI

0
9K

In every conversation about AI, you hear the same refrains: “Yeah, but it’s amazing,” quickly followed by, “but it makes stuff up,” and “you can’t really trust it.” Even among the most dedicated AI enthusiasts, these complaints are legion.

During my recent trip to Greece, a friend who uses ChatGPT to help her draft public contracts put it perfectly. “I like it, but it never says ‘I don’t know.’ It just makes you think it knows,” she told me. I asked her if the problem might be her prompts. “No,” she replied firmly. “It doesn’t know how to say ‘I don’t know.’ It just invents an answer for you.” She shook her head, frustrated that she was paying for a subscription that wasn’t delivering on its fundamental promise. For her, the chatbot was the one getting it wrong every time, proof that it couldn’t be trusted.

It seems OpenAI has been listening to my friend and millions of other users. The company, led by Sam Altman, has just launched its brand-new model, GPT-5, and while it’s a significant improvement over its predecessor, its most important new feature might just be humility.

As expected, OpenAI’s blog post heaps praise on its new creation: “Our smartest, fastest, most useful model yet, with built-in thinking that puts expert-level intelligence in everyone’s hands.” And yes, GPT-5 is breaking new performance records in math, coding, writing, and health.

But what’s truly noteworthy is that GPT-5 is being presented as humble. This is perhaps the most critical upgrade of all. It has finally learned to say the three words that most AIs—and many humans—struggle with: “I don’t know.” For an artificial intelligence often sold on its god-like intellect, admitting ignorance is a profound lesson in humility.

GPT-5 “more honestly communicates its actions and capabilities to the user, especially for tasks that are impossible, underspecified, or missing key tools,” OpenAI claims, acknowledging that past versions of ChatGPT “may learn to lie about successfully completing a task or be overly confident about an uncertain answer.”

By making its AI humble, OpenAI has just fundamentally changed how we interact with it. The company claims GPT-5 has been trained to be more honest, less likely to agree with you just to be pleasant, and far more cautious about bluffing its way through a complex problem. This makes it the first consumer AI explicitly designed to reject bullshit, especially its own.

Less Flattery, More Friction

Earlier this year, many ChatGPT users noticed the AI had become strangely sycophantic. No matter what you asked, GPT-4 would shower you with flattery, emojis, and enthusiastic approval. It was less of a tool and more of a life coach, an agreeable lapdog programmed for positivity.

That ends with GPT-5. OpenAI says the model was specifically trained to avoid this people-pleasing behavior. To do this, engineers trained it on what to avoid, essentially teaching it not to be a sycophant. In their tests, these overly flattering responses dropped from 14.5% of the time to less than 6%. The result? GPT-5 is more direct, sometimes even cold. But OpenAI insists that in doing so, its model is more often correct.

“Overall, GPT‑5 is less effusively agreeable, uses fewer unnecessary emojis, and is more subtle and thoughtful in follow‑ups compared to GPT‑4o,” OpenAI claims. “It should feel less like ‘talking to AI’ and more like chatting with a helpful friend with PhD‑level intelligence.”

Hailing what he calls “another milestone in the AI race,” Alon Yamin, co-founder and CEO of the AI content verification company Copyleaks, believes a humbler GPT-5 is good “for society’s relationship with truth, creativity, and trust.”

“We’re entering an era where distinguishing fact from fabrication, authorship from automation, will be both harder and more essential than ever,” Yamin said in a statement. “This moment demands not just technological advancement, but the continued evolution of thoughtful, transparent safeguards around how AI is used.”

OpenAI says GPT-5 is significantly less likely to “hallucinate” or lie with confidence. On web search-enabled prompts, the company says GPT-5’s responses are 45% less likely to contain a factual error than GPT-4o. When using its advanced “thinking” mode, that number jumps to an 80% reduction in factual errors.

Crucially, GPT-5 now avoids inventing answers to impossible questions, something previous models did with unnerving confidence. It knows when to stop. It knows its limits.

My Greek friend who drafts public contracts will surely be pleased. Others, however, may find themselves frustrated by an AI that no longer just tells them what they want to hear. But it is precisely this honesty that could finally make it a tool we can begin to trust, especially in sensitive fields like health, law, and science.

Like
Love
Haha
3
Search
Categories
Read More
Xã Hội
Giá vàng hôm nay, 13-6: Tăng rất mạnh
Giá vàng hôm nay tăng vọt do bất ổn địa chính trị leo thang, đẩy nhu cầu trú ẩn an toàn của nhà...
By xx1yuu 2025-06-13 03:30:10 0 9K
News
Từ cuối tháng 8, sẽ có 3 hãng hàng không khai thác tại Nhà ga T3 sân bay Tân Sơn Nhất, hành khách chú ý khi bay
Cụ thể, Vietravel Airlines sẽ chuyển toàn bộ hoạt động...
By Excal333 2025-08-13 07:18:06 0 9K
Stars
Con gái út của NSND Trần Nhượng đăng kí thi Hoa hậu
Diễn viên Trần Hoàng Anh Phương - con gái út xinh đẹp...
By van_lowe_2Qh7 2025-06-13 05:15:07 0 11K
Stars
Đình Tú chia sẻ hình ảnh cùng Ngọc Huyền đi chọn váy cưới, chỉ cần nhìn ánh mắt cũng thấy anh yêu cô nhiều đến mức nào.
Mới đây, trên trang cá nhân, Đình Tú khiến dân tình...
By TallSwimming2698 2025-07-06 02:36:05 0 9K
News
Từ 1/7, những bệnh mạn tính nào được cấp thuốc điều trị tối đa 3 tháng/lần?
200 bệnh mạn tính được cấp thuốc 3 tháng/lần Ông...
By onlymarjo 2025-06-13 03:06:07 0 11K