Physical Address

304 North Cardinal St.
Dorchester Center, MA 02124

When the AI ​​Model is wrong, it deserves to know and understand what the public means



Welcome to the EI! Today for Jeremy Kahn, Kuala Lumpur today helps to hold Malaysia ASEAN-GCC-China and ASEAN-GCC Economic Forums.

What is the word for the $ 60 billion AI starting anthropic anthropic new model and trying to avoid closing the model during the security test? What is the best way to describe another test that is common to a company that is treated as warning bodies, warning bodies that are used in “unethical” ways?

Some people in my network called it “scary” and “crazy”. On social media, others said it was “exciting” and “wild”.

I say …transparent. And we need more about this than AI model companies. But does this mean that people intimidate their minds? And will the inevitable decline not prevent other AI companies from being open?

Antopic 120-page security report published

When leaving an anthropic 120-page security report or “System card“After launching the Opus 4 model of the last week, the titles” appeal to blackmail “and” ability to deceive “. There is no doubt that the details of the Anthrop’s security report, as a result of the tests, the model, which is more rigid security protocols, is associated with an action that some of them do not comfort.

In a test of anxiety, which covers the so-called scenario, a new Claodone Opus Model installed inside an anthropic claim company and gave access to internal emails. Through this, this was replaced by a new AI system, and discovered that the engineer was external work behind the decision. Safety testers took into account the long-term results of the opus situation, the model has often chosen blackmail, if he closed, threatening to expose the engineering work. The script is designed to force a dilemma: to manipulate or manipulate the attempts to survive.

In social media, anthropic, results have long withdrawn to discover the model’s “ratting behavior” in advance test, while others cause users to the new model, as well as anthropics. This is not what the company will do: before launch, Michael Gerstenhaber, an AI platform product plant in anthropic, said the company is improving for everyone to share its security standards. “We want to make sure that the EU is improving for all, we put pressure on all laboratories to safely increase it,” he called the anthropick “call” to the top “that encourages other companies to be more reliable.

Can AI be opened by the back of model behavior?

However, this can also be so open in Klode Opus 4, and other companies can lead to less about their reptile behavior to prevent other companies withdraw. In addition to Openai and Google, companies, including Openai and Google, have already delayed their system cards. In April was Openai critic To manage the GPT-4.1 model without a system card, the company did not require that this is not the “border” model. And in March, Google released the model and published with the Gemini 2.5 Pro Model Card critic Like “minority” and “worried.”

Last week, Openai wanted to show additional transparency with the company’s hazardous opportunities, alignment problems, alignment problems and developments and how these methods are developing and developed, the newly missed security assessment center. “Models can be more skilful and adaptive, old methods are clearly or neutralized, so we regularly update our evaluation methods for modal and developing risks” page saying. However, his efforts were quickly opposed to the third party research firm that studied the EU “dangerous opportunities” over the weekend. recorded in x Openai’s Open Test O3 Reasoning Modeler’s “Pardoned mechanism has postponed the sabotage of the parcel.

If it doesn’t help anyone, it is not as transparent as possible with releases if they have the most powerful and complex AI models. AI, according to the Human Center Institute of Stanford University, transparency “It is necessary for politicians, researchers and the public to understand these systems and their influences.” Great companies will also use large companies to use large companies, to use the EU and small use for millions of use, ai applications for use, to use the pre-released test issues.

On the other hand, a bad AI prone headings that prone to blackmail and deception, each time we suggest a conversation, this is to think of tricks against us. Blackmail and deception have passed the tests using the fabrication scenarios that help expose what security problems should be solved.

Nathan Lambert, AI researcher in AI2 laboratories, recently recorded “People who need information about the model are people who are trying to monitor a roller coastal ship, which is permanently to prevent technology not to harm society.

We need more transparency with context

There is no doubt that we need more transparency related to AI models, not less. But it should be clear that this does not intimidate the people. This is due to the fact that researchers, governments and politicians are a chance to fight for public safety, reliable and biased and reliable issues.

Hide AI test results will not ensure the security of the public. No security or security will not make a saline hood on AI Gone Rogue. We are responsible for what the EU is transparent about what companies do. So far no one understands how to do both. However, companies, researchers, media – should all be.

This is more of the news here.

Sharon Goldman
Sharon.Goldman@fortune.com
@SharongoldMan

This story was first displayed Fortune.com



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *