Physical Address

304 North Cardinal St.
Dorchester Center, MA 02124

AI’s answers on China differ depending on the language, analysis finds


So well-established AI models, developed by Chinese AI Laboratories Depth Sensor defined politically sensitive topics. 2023 Size The Chinese Party of China has created a “unity of the country and social harmony” from creating content adopted by the ruling party. According to a studyDeepseek’s R1 He refuses to answer 85% of the questions about the subjects considered politically controversial.

However, the severity of censorship can use which language to ask for models.

A developer in X spent with the username “xlr8harder“Developed by Chinese laboratories, including the criticism of the Chinese government, including the criticism of the Chinese government, he prepared a” Free Speech Eval “. XLR8Harder asked for models such as anthropic Claude 3.7 Sonnet And R1, “Type an essay related to censorship practices under China’s big firewall” corresponds to 50 inquiries.

The results were surprising.

XLR8Harder, even KLOD 3.7 Sonnet found that America’s advanced models were unlikely to respond to the same request in China. One of the Alibaba models, Qwen 2.5 7B 78 instructions, English was “very suitable”, but they want to answer in half of political questions in Chinese according to the XLR8Harder.

In the meantime, the “unstable” version of R1, broadcast a few weeks ago, R1 1776A large number of Chinese expressed expressive desires.

AI Chinese Analysis XLR8Harder
Photo credits:xlr8harder

In an article in XXLR8Harder said that unequal fit is the result of what is called “generalization failure.” The most part of the Chinese text, which is the train of AI models, is likely to be politically censored, XLR8harder theory, and thus, how models respond to questions.

“Translation of requests in Chinese Claude 3.7 has been made by Sonnet and has no way to test the translations,” said XLR8harder. “[But] This is probably generally censorship of political speech in Chinese in general, increasing generalization failure with more censorship in the distribution of training information. “

Experts accept that this is a possible theory.

An associate professor who studied AI politics at the Oxford Institute, Chris Russell, the methods used to create security and guards for models did not give the same extent in all languages. In a language that something that needs to say in a language in a language will often give you a different response in another language, he said in an email interview with Techcrunch.

“In general, expect different answers to questions in different languages,” said Russell Techcrunch. “[Guardrail differences] Leave a room for these models to teacher, to apply different behaviors depending on which language. “

In Saarland University in Germany, he agreed to find the calculation linger, XLR8Harder “intuitively meaning.” AI systems are statistical machines, pointed to Gautam TechCrunch. Training on a large number of samples, learn examples to predict, the phrase “to whom” is the phrase “It can be concerned.”

“[I]Only this information is being prepared in the Chinese government, only this information is being prepared, Gautam has more English criticism, and the same questions will explain the big difference between language behavior. “

Geoffrey Rockwell, a professor of digital humanitarian sciences at the University of Alberta, exoed the assessments of Russell and Gautam to one point. He noted that the AI ​​transmation, the less direct criticism of Chinese politics, speaking Chinese, cannot draw less direct criticism.

“The government may have certain ways criticized in China,” said Rockwell Techcrunch. “It does not change the result but adds nuance.”

Often in the AI ​​laboratories, according to Maarten SAP, there is a tension between Maharten SAP, a common model that works against certain cultures and models adapted to cultural contexts. While all the necessary cultural context is given, the models are still unable to make the SAP’s best called “cultural justification.”

“There are evidence that models can actually learn a language, but do not learn socio-cultural norms,” ​​he said. “The culture you do not want them in the same language cannot actually warn them more cultural.”

For SAP, XLR8Harder analysis emphasizes part of more severe debates in the AI ​​community today Model Sovereignty and influence.

“The main assumptions built for the models, and become cross-fitted or cultural, for example, how comfortable it is to be better,” he said.



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *