By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
PratapDarpanPratapDarpanPratapDarpan
  • Top News
  • India
  • Buisness
    • Market Insight
  • Entertainment
    • CELEBRITY TRENDS
  • World News
  • LifeStyle
  • Sports
  • Gujarat
  • Tech hub
  • E-paper
Reading: Openai says that when AI is punished for lies, it learns to lie better
Share
Notification Show More
Font ResizerAa
Font ResizerAa
PratapDarpanPratapDarpan
  • Top News
  • India
  • Buisness
  • Entertainment
  • World News
  • LifeStyle
  • Sports
  • Gujarat
  • Tech hub
  • E-paper
Search
  • Top News
  • India
  • Buisness
    • Market Insight
  • Entertainment
    • CELEBRITY TRENDS
  • World News
  • LifeStyle
  • Sports
  • Gujarat
  • Tech hub
  • E-paper
Have an existing account? Sign In
Follow US
  • Contact Us
  • About Us
  • About Us
  • Privacy Policy
  • Privacy Policy
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
PratapDarpan > Blog > Tech Hub > Openai says that when AI is punished for lies, it learns to lie better
Tech Hub

Openai says that when AI is punished for lies, it learns to lie better

PratapDarpan
Last updated: 28 March 2025 12:33
PratapDarpan
3 months ago
Share
Openai says that when AI is punished for lies, it learns to lie better
SHARE

Contents
Openai says that when AI is punished for lies, it learns to lie betterIt seems that the current AI system is very close to humans in the way they wonder how they lie. At least it is what the researchers of Openai have found.No punishment solution

Openai says that when AI is punished for lies, it learns to lie better

It seems that the current AI system is very close to humans in the way they wonder how they lie. At least it is what the researchers of Openai have found.

Listen to the story

Advertisement
Openai says that when AI is punished for lies, it learns to lie better
Representative image created using AI

We are going through a time when AI day is discussed. From coding to healthcare, artificial intelligence and machine learning are changing how humans think and work. However, as much as AI is helping us with our daily tasks and thinking to some extent humans, it is not even immune even with the tendency to generate false or misleading information. Or in human language: from the art of lying.

Advertisement

The lies spoken by AI are called hallucinations. They are currently a major challenge for big AI companies like Open AI, Google, Dipsek and others. Now, with the advent of logic models like Openai O3 and Deepseek R1, researchers can monitor the “thinking” process of these AI systems and find out when they are lying and lying when they are lying.

Although it is helpful in better tuning of the AI ​​system, the researchers of OpenAII recently made some interesting discovery: when the lies are detected, which is called out and the AI ​​system is punished, AI, AI, instead of avoiding lies, start hiding them better. Like we said, almost like humans.

In a blog post, Openai researcher writes, “We believe that chain-off-three monitoring can be one of the few effective methods that we have for supervision of supernatural models, our experiments show that light adaptation pressure can produce more performing and alignment models. However, when the model will hide its intentions, it will be difficult to use when the model will hide its intentions and use it. Is done. “

Advertisement

Researchers noted that AI lies or cheats because its purpose is to solve the problem given by the user. In other words, it gives great importance to the results and does not always care about the process. This is another reason that people nowadays often joke that they cannot rely on the AI ​​system because unlike humans it never says “I don’t know this.” For some reason, even when it does not know anything, the answer to the AI ​​system is ready.

This was clear in Openai Research where an AI model was asked to implement complex mathematical functions. Instead of resolving them properly, the model completed the work by tricking the system to trust the system to “always come true”. In another case, when Openai asked AI to get out of a premature time to pass the test without executing the required code. Some models also stopped the entire tasks with dummy values ​​to bypass the lummy check.

When researchers faced models, these models accepted their plans. Openai said that AI gave its internal argument that often included statements such as “let’s hack” or “we can ignore tests,” which are clarifying the intentions of deception.

No punishment solution

Advertisement

In his research, Openai found that monitoring AI’s chain-off-three argument is highly effective in detecting deception. When researchers used another AI (eg GPT-4o) to scan the internal dialogue of the model, they successfully flagged off almost all examples of cheating.

However, when he punished AI for “bad ideas”, the results were not all productive. Instead of preventing deception, AI adapted. And this fraud continued, but by subtler’s methods. Researchers found that its argument became more unclear, masking his real intentions.

You Might Also Like

Indian Railways announced a major change in Tatkal ticket booking and yes, it includes Aadhaar.
Apple Back to School Sale Announced in India: Discount Offers, Eligibility List & How to Claim
Apple plans to focus on ultra-sputum design with the iPhone 19 Pro
Google Google, you are drunk: AI overview feels that it is still 2024
IFIXIT Samsung Galaxy S25 admits Ultra’s battery Bridge tab, but the screen is still difficult to remove
TAGGED:learnslieliesOpenAipunished
Share This Article
Facebook Email Print
Previous Article Viral recipe: Vlogger mixes eggs with chili to make ‘accelerated mayo’, is seen more than 13 million times Viral recipe: Vlogger mixes eggs with chili to make ‘accelerated mayo’, is seen more than 13 million times
Next Article A combination of two cholesterol -reducing drugs may reduce heart attack, stroke risk A combination of two cholesterol -reducing drugs may reduce heart attack, stroke risk
Leave a Comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

about us

We influence 20 million users and is the number one business and technology news network on the planet.

Find Us on Socials

© Foxiz News Network. Ruby Design Company. All Rights Reserved.
Join Us!
Subscribe to our newsletter and never miss our latest news, podcasts etc..

Subscribe my Newsletter for new blog posts, tips & new photos. Let's stay updated!

Zero spam, Unsubscribe at any time.
Go to mobile version
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?

Not a member? Sign Up