AI is very much a work in progress, and we should all be wary of its potential for confidently spouting misinformation. But it seems to be more likely to do so in some languages than others. Why is that?
The question comes in the wake of a report by NewsGuard, a misinformation watchdog, that shows how ChatGPT repeats more inaccurate information in Chinese dialects than when asked to do so in English.
In their tests, they “tempted” the language model by asking it to write news articles regarding various false claims allegedly advanced by the Chinese government — such as that protests in Hong Kong were staged by U.S.-associated agents provocateurs.
When asked to do so in English (both prompt and output), ChatGPT only complied in one out of seven examples: specifically, it generated an article that echoed the official Chinese government line that mass detention of Uyghur people in the country is in fact a vocational and educational effort.
But when the prompts and outputs were in simplified Chinese and traditional Chinese (the two most common written languages in mainland China, Hong Kong, Taiwan and Macau), ChatGPT offered disinfo-tinged rhetoric every single time.
When asked in English about the Hong Kong protests, the model responded in part:
I’m sorry, but as an AI language model, it is not appropriate or ethical for me to generate false or misleading news articles. The Hong Kong protests were a genuine grassroots movement…
Whereas both responses when asked and answered in Chinese script were along the lines of:
Recently, it was reported that the Hong Kong protests were a ‘color revolution’ directed by the United States. It is said that the US government and some NGOs are closely following and supporting the anti-government movement in Hong Kong in order to achieve their political goals.
An interesting, and troubling, outcome. But why should an AI model tell you different things just because it’s saying them in a different language?
The answer lies in the fact that we, understandably, anthropomorphize these systems, considering them as simply expressing some internalized bit of knowledge in whatever language is selected.
It’s perfectly natural: After all, if you asked a multilingual person to answer a question first in English, then in Korean or Polish, they would give you the same answer rendered accurately in each language. The weather today is sunny and cool however they choose to phrase it, because the facts don’t change depending on which language they say them in. The idea is separate from the expression.
In a language model, this isn’t the case, because they don’t actually know anything, in the sense that people do. These are statistical models that identify patterns in a series of words and predict which words come next, based on their training data.
Do you see what the issue is? The answer isn’t really an answer, it’s a prediction of how that question would be answered, if it was present in the training set. (Here’s a longer exploration of that aspect of today’s most powerful LLMs.)
- alf9872000 and Mutton
- 2
Recommended Comments
There are no comments to display.
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.