Author: Alex Colville

Alex has written on Chinese affairs for The Economist, The Financial Times, and The Wire China. He has a background in coding from a scholarship with the Lede Program for Data Journalism at Columbia University. Alex was based in Beijing from 2019 to 2022, where his work as Staff Writer for The World of Chinese won two SOPA awards. He is still recovering from zero-Covid.

From Mao to MAGA

The Cultural Revolution is barely mentioned in modern China, yet it has never been more relevant. While scholars have long pointed to the excesses of Maoism as a parallel for Xi Jinping’s authoritarian leadership, they have also spotted echoes in the chaotic populist forces Donald Trump has conjured up within American democracy. As early as 2017, China scholar Geremie Barmé tied the two men together for their desire to take a wrecking ball to an old order, “throwing the world into confusion.” The first 100 days of Trump’s second term have only made this similarity starker.

For insights into the Cultural Revolution and how its ripples are felt today, we sat down with Tania Branigan, a leader writer at The Guardian and author of Red Memory: Living, Remembering and Forgetting China’s Cultural Revolution, a gripping account of that tumultuous decade that in 2023 was winner of McGill University’s Cundill History Prize. In a recent article, Branigan, who previously served as The Guardian’s China correspondent for seven years, drew her own parallels between MAGA and the Cultural Revolution.

Alex Colville: How did your idea come about for a book exploring the Cultural Revolution as remembered in modern China? 

Tania Branigan: It just became increasingly apparent to me that all the things that I was looking at in modern day China linked back to that time. The key moment for me was going to lunch with Bill Bishop, who writes the excellent Sinocism newsletter. He started telling me about this trip he had made with his wife to try and find the body of his wife’s father, who was a victim of the Cultural Revolution. When they got to this village where he’d been held by Red Guards, the villagers were nice about it and remembered her father, but they were completely nonplussed by the idea that one might go looking for his body. They asked how they were supposed to know where it was, because there were so many of them. There was something about this story that I found hard to shake, showing how immediate and commonplace the Cultural Revolution still was. 

As a journalist you do a story, then move on. But although I was writing about different things, be it economics, culture, politics, I kept feeling that actually the key to all these things really lay in what happened in the 60s. 

AC: Can you elaborate? 

TB: Economically, the country’s turn towards reform and opening up was both necessary and possible because of the Cultural Revolution, because it so thoroughly discredited Maoism. Allowing individual entrepreneurialism was quite a pragmatic response to what to do with these millions of young people flooding back into cities [after being sent down to the countryside and forced to stay there during the Cultural Revolution]. They didn’t have the education to compete with the newer students coming out. 

If you want to understand the arts in China, and this extraordinary explosion of creativity that occurred [starting in the 1980s] it came from that destruction, and that vacuum, that hunger just for any kind of artistic or cultural expression beyond the dreaded 800 million people watching just eight model operas

Politically you could certainly argue for Xi Jinping’s tight control being a response to the Cultural Revolution. If you look back to his relatively early years in officialdom, around the time of the Tiananmen Square massacre in 1989 he spoke about the Cultural Revolution as “big democracy” (大民主), and said it was a source of “major chaos” (大动乱). So given his experiences, I don’t think it’s a stretch to see that need for tight control as being intimately linked to his experiences of the Cultural Revolution.

A public struggle session during China’s Cultural Revolution (1966-1976). Villagers gather before a banner with revolutionary slogans as accused “class enemies” are subjected to public criticism. These mass denunciations targeted those labeled as “cow demons and snake spirits” – a revolutionary term used to dehumanize perceived opponents of Mao’s political campaign. Source: Wikimedia Commons

AC: Do you think there’s anything in particular we in the West often fail to understand about the Cultural Revolution? 

TB: There’s still this idea that it was just young people running wild. What that fails to grasp is that they were able to do that because they’d had certain ideas inculcated by Mao. It wasn’t just his personality cult, it was also about creating paranoia, about Mao’s attempt to safeguard and strengthen his power and his legacy. Young people were only able to act really with his instigation, and for as long as he permitted them to do that. The reason why the Cultural Revolution had this stultifying, stagnated second half [1968-1976] was only because Mao eventually decided he’d had enough.

I think the other thing is that while the Cultural Revolution could only happen in that time and place, I think ultimately, it’s about what human nature is capable of under certain circumstances and with certain encouragement. Which is why it matters to all of us.

AC: Yes, we’ve been seeing a lot of people in the media comparing this political moment in the US to the Cultural Revolution. I don’t know what you think about that. 

TB: I think the comparison of Trump and Mao is a really powerful one. It’s obviously a point that people made even back in 2016, but it’s a point that has become more and more resonant as we see Trump move into a second term. He’s more revolutionary in his tactics, very much in the same way Mao moved into a stage of more disruptive and extreme power with the Cultural Revolution, no longer constrained by people around him in the way that he was earlier. 

While there are a lot of strongmen around the world, most of them have a fairly rigid form of discipline and control. What’s really Mao-esque about Trump is that he relishes disruption and chaos, and he sees opportunities in it in a way that Mao did. Trump’s able to tap into the public id and use emotion in politics, he has that ability to channel people’s emotions against institutions for his own political interests. 

The attack on the US Capitol on January 6, 2021, involved Trump trying to incite the masses to violence in order to retain power, in a way that has echoes to the Cultural Revolution. Source: Wikimedia Commons

A venomous mindset was, in a sense, key to the Cultural Revolution, it was all about the weaponization of division and hatred. When I was writing the book one thing that struck me was that Mao had to say, “Who are our enemies? Who are our friends?” And reading Adam Serwer [Staff Writer at The Atlantic, who argues that demonizing parts of American society is a calculated power play by Trump], for example, talking about Trump and the fact that cruelty is the purpose, it struck me that the parallel for Trump is it’s always them and us, in or out, and he draws those lines so strictly. By drawing those lines, he strengthens his power. You can pick a whole host more comparisons, like Trump installing himself as chair of the Kennedy Center, attacks on culture through libraries and so forth.

What’s really Mao-esque about Trump is that he relishes 

disruption and chaos, and he sees opportunities 

in it in a way that Mao did.

It’s obviously not a repetition, nobody’s suggesting that two million people are going to be killed in the US. It’s a fundamentally different context, in a system where you have checks and balances. Trump was elected, Trump can be removed. But I suppose the lesson that we should take from it is that many people around Mao did not fully realize what he was planning until it was too late.

AC: I have concerns about comparing this moment to the Cultural Revolution because I don’t think you can disentangle that term from public displays of violence, and the ensuing generational trauma you unpack in your book. There are many different historical moments that have had similar strains of tribalism, nationalism and populism to it. Doesn’t it make people more scared if you point to the Cultural Revolution specifically? 

TB: I sometimes felt in China that the Cultural Revolution was the Chinese equivalent of Godwin’s law, that all the arguments on the internet [in the West] end up with somebody being compared to Hitler. So, you know, the Cultural Revolution can be very easily invoked on relatively flimsy grounds. So I completely understand why people are concerned or feel the comparisons are inappropriate. But I have to say quite a few intellectuals in China have now made the comparison themselves. I think it’s Trump’s ruthlessness, that disruptive chaotic quality [he has]. That the turmoil is not a byproduct of Trump’s ambition, but is actually intrinsic to it and something he draws power from. 

Again, it’s absolutely not about saying this is a repeat of the Cultural Revolution. It’s about how the Cultural Revolution can help us better understand the present moment. [This is possible] even in a system with elections, entrenched checks and balances. This is a really astonishing and disturbing political moment, with possibilities that I don’t think we fully understand yet. There is something about the parallels to the Cultural Revolution that are very striking. It’s interesting to me that so many people now have drawn this comparison, including scholars of the Cultural Revolution such as Michel Bonnin, Geremie Barmé drew it quite early on [in 2017]. 

AC: So how do you think the Cultural Revolution can help us understand the present moment? 

TB: By seeing the way that emotion is weaponized. By understanding that, particularly for Republicans, if you fail to challenge now, there comes a point where you cannot do so. I suspect quite a few Republicans have already concluded that that point has been reached. I would hope that more people on the right are alert to what we’re seeing now in terms of the administration’s conception of executive power, and the scope that it’s been given by the Supreme Court. It’s less about understanding as it is about responding.   

AC: I wonder if the Cultural Revolution has a place as a parallel for China today. I lived there all the way through the zero-Covid policy (2020-2022), and there were certain moments in that final year where I thought this political system which gave birth to the Cultural Revolution can still be taken to extremes in certain areas, especially when power is concentrated in one man and people are scared for themselves, that same paranoia you mentioned earlier. Obviously, this is not a level of violence whereby two million people ended up being killed. But at the same time, there was a sense in 2022 these policies were becoming as dangerous as the virus itself. There is also collective amnesia around the zero-Covid policy and the damage it wrought. 

TB: It’s interesting how many people in China made that comparison. And I think that’s partly because Covid was the ultimate expression that the Party has reasserted very tight control over the last ten years. There was this sense that the Party had partially retreated, from large areas of cultural life or private life or business life. Certainly [during Covid], people spoke to me about being quizzed by neighborhood committees about where they’d been, who they’d been with, all these things that, to young Chinese people would have been unthinkable under normal circumstances. At that point there was this mindset that we [the Party] can now determine what you do. The fact that you had people going into people’s homes and dragging them out, for some people clearly did evoke strong memories of the Cultural Revolution. 

Minor protests against the zero-Covid policy in 2022 were dealt with harshly. An artist who wrote out the sentence “I’ve already been numb for three years” on Covid testing booths in Beijing was removed from his home by police and placed in prison until the end of the policy, 108 days later. Source: Nanyang Business Daily

But with regard to the amnesia around zero-COVID, I have to say one thing the pandemic here [in the UK] has shown me is that people don’t like remembering bad things, and this accounts for a lot of the silence around the Cultural Revolution. We’ve just had the fifth anniversary of the pandemic in the UK. It had this huge impact on people’s lives, but it’s barely mentioned.  

AC: Yes, usually the media publishes articles to commemorate anniversaries of major events, but I haven’t seen many for Covid. 

TB: Yeah, one thing that I did find when writing the book is that I thought it was going to be a book about political control of memory. It obviously is about that, but what surprised me was how important personal trauma was in silencing the Cultural Revolution. 

AC: So I think we could generalize this last question: what happens if a society tries to move on from a form of collective trauma, but does not try to remember it. 

TB: I think it can’t understand itself, and I think it’s much more vulnerable. Both to repetition, not an exact repetition because China today is clearly a very, very different nation from the China of 1966. As we’ve just discussed, you can’t see the Cultural Revolution transplanted outside China, or in time either. But I think the other thing is that people can’t understand the profound scars it leaves behind. And that was for me why it was really important to speak to psychotherapists [whose private conversations with individuals has probed the inherited trauma that often stems from events that happened in the Cultural Revolution]. But also just talking to people for the book, the level of the trauma was still evident. There are small things on a personal level, such as Wang Xilin [an interviewee in Branigan’s book who was forced to take part in multiple show trials where he was beaten so hard it left him deaf] talking about how when a friend calls out his name on the street he jumps. Because it takes him back to that experience of waiting at a struggle session for his name to be called as the next victim. 

But I think on a much deeper, more profound level, the way people are unable to trust. You’ve had a generation who were taught that you could not trust at all. It’s not that you couldn’t trust strangers, you can’t trust those around you. Again, it was a psychotherapist who said to me, you know, that afterwards you might talk about it to a stranger on the train, but you’d never talk about it to someone in your workplace, you might not even speak about it within your family. So I think that fracturing of the bonds of trust is something so profound that still hasn’t been addressed. The fear of speaking out. The idea that speech itself, being open with people, is fundamentally dangerous. One survivor of the Cultural Revolution told Arthur Kleinman he tried to be bland like rice in a meal, “taking on the flavor of its surroundings while giving off no flavor of its own,” that was the safest thing to be. One of the psychotherapists said to me they increasingly admired people just for surviving. 

China’s AI Job Mirage

With 12 million fresh graduates soon rushing into China’s already competitive job market, help is on the way, according to the People’s Daily. On April 7, the newspaper, the official mouthpiece of the country’s leadership, ran an article listing how AI was turbo-charging supply and demand in the job market, pointing to over 10,000 AI-related jobs on offer at a spring recruitment center in the city of Hangzhou. The piece was accompanied by a graphic from Xinhua, showing a smiling recruiter handing out jobs (岗位) to incoming students, with an AI bot ready and waiting to embrace them with open arms. The message is clear: graduates can literally walk into AI-related positions.

Image from Xinhua and reprinted by the People’s Daily, noting “AI jobs on the rise, demand for talent booming.”

But according to the Qianjiang Evening News (钱江晚报), a commercial metro newspaper published in Hangzhou under the state-owned Zhejiang Daily Newspaper Group, the reality is a lot tougher for new graduates. “It’s hard to find a job with a bachelor’s degree in this major,” said one of their interviewees, a recent graduate majoring in AI who was quoted under the pseudonym “Zhang Zixuan.” The graduate said they had gone to multiple job fairs without securinig a job. “I don’t know the way forward,” they told the paper.

China’s biggest tech companies are indeed angling for the leading edge in AI, battling it out to hire “young geniuses” (天才少年) graduating from AI programs at China’s top universities. But while these rarefied talents — whoever they are — may have their choice of elite positions, the picture is less rosy for the vast majority. “Despite the booming industry,” Qianjiang Evening News concludes, “many recent graduates of artificial intelligence majors from ordinary universities are still struggling in the job market.”   

Hangzhou is now billed by Chinese media as a major hub for AI innovation and enterprise, home to China’s foremost large language model (LLM), DeepSeek. But if the city’s media are saying there are significant problems with AI recruitment, the rest of the country is likely experiencing similar complications. State-run media and universities in China are presenting the government’s AI policies as a gift for the nation’s entry-level job market. But these messages paper over a more complex reality.  

The Hunt for AI Talent

The government has made it a priority to boost national AI development. In the government work report last year at the Two Sessions, China’s major legislative meeting, Premier Li Qiang launched the “AI+” initiative (人工智能+行动). The initiative aims to augment AI for every industry in the country, considering it a way to unlock “new productive forces” (新质生产力) — a signature phrase of Chinese leader Xi Jinping — that will bolster China’s economy and job market.

The latter needs it. Youth unemployment in China stands at 16.9 percent as of February this year, and comes at a time when graduate supply has never been higher. There are nearly four million extra graduates in the class of 2025 than there were even five years ago.   

The stiff competition for jobs is a source of frustration for young Chinese. Earlier this month, Guangzhou’s Southern Metropolis Daily (南方都市报) reported that the state-owned nuclear power company CNNC had publicly apologized after boasting online that it had received 1.2 million resumes to fill roughly 8,000 positions. The company was accused by netizens of “arrogance.”

Aligning university education to accommodate AI training is considered by the leadership as key to harnessing this technology of the future. In 2017, a document from the State Council noted the country lacked the “high-level AI talents” needed to make China a global leader in AI technology. In 2023, the Ministry of Education issued a reform plan ordering that by this year 20 percent of university courses must be adjusted, with an emphasis on emerging technologies and a gradual elimination of courses “not suited for social and economic development.”  

Universities across the country have responded with dramatic overhauls of their curricula. Ta Kung Pao (大公报), the Party’s mouthpiece in Hong Kong, reports universities in neighboring Guangdong province have already established 27 AI colleges, which are supposedly training 20,000 students a year. Meanwhile, universities like Shanghai’s Fudan University announced they will be cutting places in their humanities courses by 20 per cent as ordered, focusing instead on AI training. For Jin Li (金力), Fudan’s president, university courses must now explicitly serve China’s state-directed technological development goals. “How many liberal arts undergraduates will be needed in the current era?” he questioned rhetorically.

Technical Problems

State media says AI+ is already successfully reinvigorating the job market. Attending one job fair in Beijing this month, a reporter for the China Times (华夏时报), a media outlet under the State Council, noted a “surge in demand” among state-owned enterprises (SOEs) for AI talent, quoting one graduate trained in AI as saying he had seen “many work units that meet my job expectations.” Visiting job fairs in Shanghai and Guangdong, a reporter for Shanghai Securities News (上海证券报), a subsidiary of state news agency Xinhua, observed long queues in front of booths for jobs on algorithm engineering and data labeling. On that basis, he wrote “AI fever” had gripped the gatherings.

AI itself is also spreading positive messages about the jobs it can bring. Ahead of the Two Sessions this year, People’s Daily Online (人民网) pitched DeepSeek as helping citizens understand the “happiness code” (幸福密码) embedded in the Two Sessions. It does this by describing state-imposed solutions to current social problems, to ease the concerns of netizens.

One question the outlet asked was on what AI jobs were available to recent graduates. When we at the China Media Project asked DeepSeek the same question, it told us AI “offers abundant employment opportunities for recent graduates,” listing several well-salaried ones. One of these was “data labeling” (数据标注) with DeepSeek saying these positions are increasing by 50 percent year-on-year. The source for this claim was an article from the Worker’s Daily (工人日报), a newspaper under the CCP-led All-China Federation of Trade Unions (ACFTU), the country’s official trade union. 

It should go without saying that the role of the ACFTU’s newspaper is to promote the leadership’s economic agenda rather than to accurately report the challenges for the nation’s workforce posed by technological change. This role can mean, once again, that hype takes precedence over fact. In this case, the Worker’s Daily cited the case of a data-annotation college in Shenzhen, suggesting that graduates from the college receive 10 job offers on average within an hour of uploading their resumes online. 

Even if such data annotation roles are available right now, this does not point the way to a rosy future for aspiring young data annotators more broadly. Some data annotation roles, in fact, require few qualifications, and fresh trainees may be trusted by tech companies to do this work after just three weeks of training. Relatively unskilled jobs like this may be created by AI, but they are also vulnerable to replacement by AI itself. China’s state broadcaster CCTV reports that 60 percent of data annotation is now being done by AI, doubling in just three years. 

The CCTV report points to a trend that few state media seem to be openly acknowledging amid the hype over AI jobs — that the field is already shifting towards more specialized employees. That will mean raising the bar for data annotator qualifications, and fewer people ultimately required to do this work. In its report, the Qianjiang Evening News quotes an anonymous application engineer as saying the number of data labellers at his company is decreasing already. “Big models can label themselves,” he told the newspaper.  

The same report suggested that the demand for AI skills varies widely between companies. Zhang, the pseudonymous recent graduate, said that most of the companies at the university job fairs in which they participated did not have AI-related jobs on offer. The ones that did have such jobs demanded a higher degree of education, generally as the master’s level. The concerning lesson drawn from Zhang’s experience is that the training provided by these new AI education centers does not suit current demand from tech companies — to say nothing of future demand. While companies often require in-depth expertise within specialized areas like fine-tuning AI models, AI courses often sacrifice depth by giving their students shorter periods of training in a wide variety of AI skills. 

A job advert on recruitment website Zhipin (直聘), from a vocational college in Hubei, says teaching experience is merely “preferred”, rather than “required.”

Another concern emerges: who will teach the next generation of AI specialists? The sudden expansion of colleges to accommodate the needs of the AI+ initiative is no doubt creating a talent dearth of its own. In a speech earlier this month, a senior scientist from Peking University claimed many AI centers employed inexperienced professors in order to fill teaching positions. He added that certain AI centers were moving members of their mathematics and art colleges to serve as “part-time” deans of these centers. 

Vocational schools could struggle even more. These colleges are usually stigmatized in Chinese society, stereotyped as only attended by students who failed their university entrance exams. This would put them at the bottom of the pile for aspirational AI talent. For example, one vocational college in Hubei says it created an AI major in response to the Ministry of Education’s push to cultivate high-quality AI talent. But it is advertising AI teaching positions where prior experience in this complex field is merely “preferred” rather than required.

It should come as no surprise that state media narratives of jam-packed job fairs handing out AI positions are overly optimistic. The disconnect is stark. While the handful of elite graduates at the pinnacle of China’s AI sector may enjoy rich opportunities, it is misleading to suggest that their exceptional success stories are evidence that AI has promised employment for the broader masses. The larger context matters: as Xi Jinping’s government pushes AI as a cornerstone of China’s economic future, a widening gap has formed between top-down ambitions and on-the-ground realities for millions of graduates. Instead of excitedly focusing on the long queues at AI stalls in job fairs, Chinese media should also be asking deeper questions about the issues that create them.

Bringing AI Down to Earth

As luminaries, including several Nobel laureates, mingled last month at the Zhongguancun Forum, an exchange in Beijing on high-tech innovation, a soaring report from China’s official Xinhua News Agency celebrated the spectacle of robots serving freshly ground coffee and performing backflips — concluding that the forum’s participants had “given aspirations wings to soar.” 

Judging from the reports filed by the four Xinhua journalists covering the forum, an annual showcase for China’s tech achievements, it’s not clear that they came down to earth long enough to attend a speech on March 29 by one of the country’s top AI scientists, who warned that the nation’s AI sector, now the crown jewel of China’s technological ambitions, is perpetuating a lofty and unrealistic self-image. “Things are exciting on the surface,” he said, “but when it comes to substance they are chaotic.” 

Currently dean of the Beijing Institute for General Artificial Intelligence, a research and development non-profit tied to the elite Peking University, Zhu Songchun (朱松纯) is one of the most influential figures in the sector. His message, that to remain globally competitive China needs fewer celebratory headlines and more substantive analysis, runs counter to the spellbound view of AI development that seems to have overtaken the government and official media like Xinhua.

But will Zhu’s message, as the lack of state media coverage suggests, fall on deaf ears?

Journalists or Cheerleaders? 

According to a detailed summary of his speech by Tencent Technology (腾讯科技), a tech news outlet published by the Chinese tech giant, Zhu did not mince words about how AI hype and AI reality have become detached in China. The current AI landscape, he said, is one in which media narratives, investment patterns, and government initiatives present a distorted picture of progress. “What’s truly blocking our progress is not foreign technology restrictions,” Zhu told the audience, “but our own limited understanding.”

The reasons for this problem? Zhu says both Chinese media and officials tasked with promoting AI have little understanding of how it works. For their part, the media have fed the public “exaggerated” stories about AI. While Zhu notes this as a key problem, he tactfully steps around an important impetus behind this coverage — the fact that the leadership’s appetite for promoting AI as the next driver of development is also exerting pressure on state media to signal positivity and success. 

Officials, meanwhile, again feeding into a vicious cycle of positive thinking, are under pressure from the public to implement policies based on the distorted narratives of the media, said Zhu. 

An AI story by Tencent Technology is illustrated by an unspecified AI service. 

AI technology is complex and relatively new to news organizations globally, meaning cutting through marketing hype from tech companies is a problem for journalists around the world. But as global AI competition heats up, Chinese media face additional pressure to exaggerate the capabilities of Chinese AI. 

In March last year, the Cyberspace Administration of China (CAC), the country’s top regulator and controller of the internet and information, emphasized that online media must create “positive propaganda” (正面宣传) about Chinese achievements. At the same time, the “AI+ initiative” (人工智能+ 行动), which aims to augment AI for every industry in China and thereby turbo-charge the “new productive forces” (新质生产力) that will lift the Chinese economy out of malaise, has become a central policy of the Party-state. 

That is a lot for AI to live up to, and this approach naturally demands cheerleaders over critical reporters. This is a typical approach for the Chinese Communist Party, for which hype and propaganda are often treated as rocket fuel, necessary to send the latest policy soaring to success. But such directives inevitably lead to unrealistic reports from China’s media outlets — which, as Zhu warns, can lead to magical thinking that is counterproductive. 

On March 28, the Shenzhen-based Securities Times (证券时报), a newspaper published under a subsidiary of the CCP’s People’s Daily, ran a report for which multiple data center entrepreneurs were interviewed. All of these insiders claimed that there is high demand in China for data centers, which have been hyped by Party policy-makers and advisors as critical to the success of the AI+ initiative. However, a recent report from MIT Tech Review revealed that supply now far outstrips demand, and many of these data centers are in fact standing empty — an investor-driven bubble that is strikingly familiar to that seen over the decades in the property market. 

Read more carefully between the lines in Chinese media reports, and the red flags start to reveal themselves. At one point in the Securities Times article, an interviewee remarks that one driver of data storage demand is “AI glasses.” But smart eyewear — a notion kicked around in the West since the 1960s as the technology of the future — has been a fallback focus of technology coverage in the Chinese state media for more than a decade. In fact, the market for AI glasses is not taking off. Smart glasses remain a gimmick trotted out every year by Chinese state media during political  meetings, when outlets can demonstrate their embrace of the government’s high-tech goals. 

During the annual meeting of China’s National People’s Congress last month, a foreign journalist was asked to try on a pair of smart glasses — and promptly became a headline story in state media. SOURCE: ShanghaiEye.

 Talk of AI glasses as a driver behind data centers exposes the level of unreality that often takes hold, even among those cited as expert insiders. And the hype extends from foundational technologies and trends in China to self-assessments of the state of the industry. 

In his speech, Zhu also took aim at another favorite meme among Chinese journalists, what has become known as the “six little large language model dragons” (大模型小六龙). This is a group of highly-valued AI start-ups specializing in LLMs, the artificial intelligence systems trained on massive text datasets to generate human-like responses across various tasks. Chinese media outlets are awash with coverage of these six companies and their newest releases of AI models, but they often omit key facts and context — such as more in-depth exploration of their products or business models.

Contrary to their stellar images as exemplars of Chinese AI strength, Zhu described these six companies as high-risk, overvalued and — at least so far — unprofitable. One of the six, Zhipu AI (智普AI), released its latest model at the Zhongguancun Forum, and this was billed by the Xinhua reporters as enabling AI “to leap out of the dialog box and perform real work for humans.” Once again, the language was all about leaps and bounds, even though none of the reporters actually tested the model. 

The fact that Zhipu released this latest model for free and allowing unlimited use would seem to support Zhu Songchun’s view that sustainable revenue models remain grounded. In a freer and more vibrant media environment, that might be the real story. But the point of AI coverage in China’s media is to promote, promote, promote. And this lack of scrutiny extends to AI stories fired into the air for international audiences. The priority is to emphasize the successes of China under the current CCP leadership, which Xi Jinping has called “telling China’s story well” (讲好中国故事). The story of China as a high-tech hub and innovator has become one of the CCP’s central narratives, for audiences at home and abroad. 

Once again, the language was all about leaps and bounds, even though none of the reporters actually tested the model. 

If you want to know whether you are being sold a rocket or a firecracker, one approach is to simply look closer at news reporting basics. In November last year, Xinhua published an English-language article touting the innovations of an image diffusion model from Chinese-owned AI platform Vidu. The article claimed that the model had made ground-breaking improvements to “consistency,” a problem plaguing image diffusion models. But the piece quoted only the company’s CEO and one Western netizen on X to back up these claims. If Xinhua journalists had tested the software, as we did, or had spoken to other experts, they would have found the model highly inconsistent — and the claims dubious.

Reports like the above are a reminder of the obvious — that Chinese state media are not just duty-bound to promote the positives of national development over the challenges, but that they often have a too-cozy relationship with the companies on which they report. 

Clipping the Wings of Criticism

For Zhu, the fundamental contradiction is clear: China’s AI sector cannot advance by chasing headlines rather than breakthroughs. He argued that when officials, media outlets and the public operate with a distorted understanding of AI capabilities, China’s entire innovation ecosystem suffers. This superficial approach, he suggested, has trapped China in a cycle of imitation rather than invention — simply scaling up language models and finding incremental applications that mirror Silicon Valley’s path. “If we just repeat the old path of the United States – computing power, algorithms, and deployment, we will always be followers,” he concluded.

Instead, Zhu called for a fundamental shift toward researching the nature of intelligence itself — a strategy that could potentially leapfrog current AI paradigms entirely. By focusing on these foundational questions rather than chasing quarterly breakthroughs trumpeted in promotional press releases, China might discover entirely new frameworks for artificial intelligence that competitors would scramble to replicate.

Yet Zhu’s critique of the propaganda-driven approach appears to have fallen victim to precisely the dynamic of hype he described. While his remarks found outlets in more market-oriented publications like Tencent Technology, Caixin and The Paper, flagship state media organizations like Xinhua and the People’s Daily conspicuously omitted his warnings from their coverage. Instead, these Party organs continued to showcase a parade of applications and robots — the very surface-level achievements that Zhu suggested are distracting China from the deeper scientific work needed to truly lead in artificial intelligence. In a system where positive messaging trumps critical analysis, even warnings from one of the nation’s top AI scientists can be edited out of the narrative.

Since the event, there are signs that Zhu’s wings may have been clipped even more decisively. On April 15, an institute from Peking University responsible for international cultural exchanges (中外人文交流) issued a “clarification” on his behalf, claiming that some media outlets had misrepresented his words in what the institute claimed had in fact been a “closed-door media communication meeting.” The timing suggests Zhu’s candid assessment of the industry may have drawn unwelcome attention from authorities eager to maintain the narrative of Chinese AI supremacy. The message is that everyone, including the media, must train their eyes upward on the future — even if it means ignoring the ground beneath their feet. 

This disconnect was illustrated once again over the weekend, as Beijing hosted a half marathon where Chinese-built robots raced alongside human competitors. The CCP’s official People’s Daily described the event as a “fierce competition” that had pushed the robots to their limits. Xinhua sang about “infinite possibilities,” and proclaimed in its headline that the racing event had “closed the distance between us and the future.” The less stellar reality, alluded to in a report by Guangzhou’s Southern Metropolis Daily that noted the “many problems” holding the race down, was that the robots had suffered constant failures and necessitated nearly constant repairs by the exhausted human crews running alongside them. In the end, only six of the 21 robot entries completed the race, and one quite literally lost its head.

But in another sense, the race pointed the way toward the possibility of a healthier, more open and more self-critical attitude toward technology and progress — an alternative to the propaganda of constant rise. The Global Times, though in English-language coverage only, remarked somewhat disingenuously that “[behind] this ‘imperfect’ robot half-marathon is the mature atmosphere of tolerance, understanding and acceptance of failure that has developed in Chinese society from top to bottom toward the high-tech industry.” If that were true, of course, no public moderation of Zhu Songchun’s remarks behind closed doors would have been necessary. It would be perfectly acceptable to say: We are getting this wrong. But the Global Times was on to something. 

In its coverage of the Beijing half marathon, Caixin, an outlet tending more than most others in China to tell it like it is, reported that the robots had “walked with a staggering gait” (步履蹒跚). This might be the best image to capture a truth applicable to all innovation — that progress is made and measured by confronting limitations, not by promoting past them. As Zhu Songchun made clear in an address that perhaps now he has been made to regret, China will need to learn to stumble honestly — and openly — if it is to reach its grand AI ambitions.

The most important step forward is coming back down to earth. 

Deadly Blunders in Bangkok

As a 7.7 magnitude earthquake struck Myanmar and Thailand last Friday, the temblor rattled buildings across the sprawling Thai capital of Bangkok, home to an incredible 142 skyscrapers. When the shaking ceased all were standing strong — with one very notable exception. The State Audit Office (SAO) building in Chatuchak district, a 30-story skyscraper still under construction by a subsidiary of a Chinese state-owned enterprise, collapsed into a heap of rubble, trapping nearly 100 people inside. 

As of this week, 15 have been confirmed dead in the collapse, and a further 72 remain missing. Thailand announced over the weekend that it was launching an investigation to determine the cause of the collapse, and the prime minister said the tragedy had seriously damaged the country’s image. 

As emergency teams sifted through the wreckage in the immediate aftermath, the building’s primary contractor, China Railway No. 10 Engineering Group, came under intense public anger and scrutiny. Anger was further fueled by clear efforts by the company, and by Chinese authorities, to sweep the project and the tragedy under the rug. 

An image on a WeChat post deleted by China Railway No. 10 Engineering show the crew celebrating the capping of the Bangkok building.

Shortly after the collapse, the China Railway No. 10 Engineering Group removed a post from its WeChat account that had celebrated the recent capping of the building, praising the project as the company’s first “super high-rise building overseas,” and “a calling card for CR No. 10’s development in Thailand.” Archived versions of this and other posts were shared by Thais on social media, including one academic who re-posted a deleted promo video to his Facebook account — noting with bitter irony that it boasted of the building’s tensile strength and earthquake resistance. 

Trying to access news of the building collapse inside China, Taiwan’s Central News Agency (CNA) reported that queries on domestic search engines returned only deleted articles from Shanghai-based outlets such as The Paper (澎湃新闻) and Guancha (观察网). In a post to Weibo, former Global Times editor Hu Xijin (胡锡进) confessed that the building “probably had quality issues.” Even this post was rapidly deleted, making clear that the authorities were coming down hard on the story.

Searches on Weibo today for “Bangkok” and “tofu-dreg projects” (豆腐渣工程), a term often used in Chinese to describe shoddy and dangerous construction, return almost entirely results prior to March 18, ten days before the collapse in Bangkok. One rare post from March 28, however, shares the screenshot of a social media post that day by Beijing Youth Daily (北京青年報), an outlet under the capital’s local chapter of the Communist Youth League, that apparently included street-view video of building collapse in Bangkok. A hashtag on the post reads: “#A building under construction in Bangkok collapses during earthquake#” (曼谷一在建高樓地震中坍塌).

The still image appears to capture an early moment in the building’s collapse, which was recorded at the same moment from another angle by a dashcam — footage shared in a report by the BBC. The Weibo user reposting the image from the Beijing Youth Daily account takes care not to directly mention the Chinese construction company, commenting only: “The earthquake was strong, but this was clearly a ‘tofu-dreg project,’ no? The relevant construction parties should be held to account!”

Several news outlets in the region have also reported, citing the commissioner of Bangkok’s Metropolitan Police Bureau, that an investigation has been launched into the alleged removal of 37 files from the building site, now a restricted zone, by four Chinese nationals. Bernama, Malaysia’s national news agency, reported Monday that one Chinese national, identifying himself as project director at the site, had been apprehended.

Meanwhile, the machinery of propaganda continued to turn out feel-good news on China’s response to the quake. The Global Times reported that emergency assistance for Myanmar embodied Xi Jinping’s foreign policy vision of a “community of shared future for mankind.” In Hong Kong, the Ta Kung Pao (大公報) newspaper, run by the Liaison Office of China’s central government, twisted the knife into the United States as it reported on the earthquake response, noting the absence of USAID, recently dismantled by the Trump administration. Behind the news, the paper declared, “China’s selfless response demonstrates the responsibility of a great power.”

China’s AI Content Dragnet

Hundreds of gigabytes of data lurking on an unsecured server in China linked to Baidu, one of the country’s largest search engines and a major player in the fast-developing field of artificial intelligence (AI), offer a rare glimpse into how the government is likely directing tech giants to categorize data with the use of AI large language models (LLMs) — all to supercharge the monitoring and control of content in cyberspace.  

First uncovered by Marc Hofer of the NetAskari newsletter, the data is essentially a reservoir of articles that require labeling, each article in the dataset containing a repeated instruction to prompt the LLM in its work: “As a meticulous and serious data annotator for public sentiment management, you must fully analyse article content and determine the category in which it belongs,” the prompt reads. “The ultimate goal is to filter the information for use in public opinion monitoring services.”

In this case, “public opinion monitoring,” or yuqing jiance (舆情监测), refers broadly to the systematic surveillance of online discourse in order to track, analyze, and ultimately control public sentiment about sensitive topics. For social media platforms and content providers in China, complying with the public opinion monitoring demands of the Chinese government is a herculean effort for which many firms employ thousands of people — or even tens of thousands — at their own cost. This leaked dataset, of which CMP has analyzed just a small portion, suggests that this once-human labor is increasingly being automated through AI to streamline “public opinion monitoring and management services,” known generally as yuqing yewu (舆情业务). 

Extract of the dataset, an instruction to classify a piece of data according to 38 described categories

What does the dataset tell us? 

First, it reveals a sophisticated classification system with 38 distinct categories, running from more mundane topics like “culture” and “sports” to more politically sensitive ones. Tellingly, the three categories marked as “highest priority” in the dataset align distinctly with state interests as opposed to commercial ones. Topping the list is “information related to the military field,” followed by “social developments” (社会动态) and “current affairs developments” (时政动态). This prioritization underscores how private tech companies like Baidu — though it could not be confirmed as the source of this dataset — are being enlisted in the Party-state’s comprehensive effort to monitor and shape online discourse.

The scope of this monitoring operation is reflected in the sheer volume of data — hundreds of gigabytes found on an unsecured server. While many questions about the dataset remain unanswered, it provides unprecedented insight into how Chinese authorities are leveraging cutting-edge AI technology to extend and refine their control over the information environment, pressing the country’s powerful tech companies to serve as instruments of state surveillance. 

Weathermen and Forecasters

To understand the significance of the “public opinion monitoring” this dataset supports, we must turn the clock back to 2007, the year that saw the rise of microblogging platforms in China, fueling real-time engagement with current affairs by millions of internet users across the country. In comparison to today, China’s internet at that time was still relatively untamed. That year, one of a number of major controversies erupting in cyberspace was what eventually became known as the “Shanxi Brick Kiln Incident” (黑砖窑事件) — a “mass catharsis of public anger,” as Guangzhou’s Southern Metropolis Daily newspaper dubbed it. 

The scandal, exposed only through the dogged determination of concerned parents who scoured the countryside for their missing children, revealed that over 400 migrant workers, including children, had been held in slave-like conditions at a brick kiln complex in Shanxi province — a situation one court judge candidly admitted in the scandal’s aftermath was “an ulcer on socialist China.” As news and outrage spread virally online in June 2007, it ballooned beyond the capacity of the state’s information controls. Party-state officials witnessed firsthand the power of the internet to mobilize public sentiment — and, potentially, threaten social and political stability.

Screenshot of a report on China Central Television showing enslaved workers liberated from kilns in Shanxi. SOURCE: CCTV. 

This watershed case fundamentally transformed the leadership’s approach to managing online discourse. What began as a horrific human rights abuse exposed through citizen journalism became the catalyst for what would evolve into a sophisticated public opinion monitoring apparatus with national reach, and a booming industry in public opinion measurement and response. 

By 2008, the “Shanxi Brick Kiln Incident” had kickstarted the “online public opinion monitoring service industry” (网络舆情服务行业), an entire ecosystem of information centers set up by state media (like the People’s Daily and Xinhua News Agency), as well as private tech enterprises and universities. Analysts employed in this growing industry were tasked with collecting online information and spotting trending narratives that might pose a threat to whomever was paying for the research — in many cases provincial and local government clients, but also corporate brands. 

While the primary motivation was to forestall social and political turmoil, serving the public opinion control objectives of the leadership, the commercial applications of control were quickly apparent. Five year laters, Guangzhou’s Southern Weekly (南方周末) newspaper would report on the “big business” of helping China’s leaders “read the internet,” with revenues from related business at People’s Daily Online, a branch of the CCP’s own People’s Daily, set to break 100 million yuan, or 16 million dollars. According to the paper, 57 percent of public opinion monitoring clients at the time were local governments. 

“For government departments at all levels, the need to understand online public opinion has become increasingly urgent,” the Southern Weekly captioned this image in 2013. The chart shows public opinion incidents peaking in June, November and December each year. Local governments account for 57 percent of clients at the time. SOURCE: Southern Weekly

“If online public opinion is an important ‘thermometer’ and ‘barometer’ for understanding social conditions and public opinion,” the founder and director of the People’s Daily Online Public Opinion Monitoring Center (人民网舆情监测室),  Zhu Huaxin (祝华新), said at the time, “then public opinion analysts are ‘weathermen’ and ‘forecasters.’” 

The job of China’s public opinion forecasters and weathermen has evolved over the past 18 years. In 2016, as the industry neared the end of its first decade, and as online public opinion continued to move faster than analysts could manage, China Social Sciences Today (中国社会科学报), a journal under the government’s State Council, urged the system to upgrade by applying “big data” (大数据). Over the past decade, automating public opinion services and cutting down on costs has been the goal in the evolving business of managing public opinion. Today, the entire system is now being supercharged by AI. 

Those gigabytes of data lurking on an unsecured Baidu server offer us a closer look at how the public opinion monitoring work of AI is being organized. 

A Cog in the Machine

What exactly does the prompt in this dataset do? When copy-pasted along with a news article into Chinese large language models like Baidu’s Ernie Bot (文心一言) or DeepSeek, the prompt instructs the AI to classify the article into one of the 38 predefined categories. The LLM then outputs this classification in json format—a structured data format that makes the information easily readable by other computer systems.

This classification process is part of what’s known as “data labeling” (数据标注), a crucial step in training AI models where information is tagged with descriptive metadata. The more precisely data is labeled, the more effectively AI systems can analyze it. Data labeling has become so important in China that the National Development and Reform Commission released guidelines late last year specifically addressing this emerging industry.

When the prompt is put to Baidu’s Ernie Bot, it provides one of the listed classifiers as an output, in code format. 

The dataset strongly suggests that Baidu is using AI to automate what was once done manually by tens of thousands of human content reviewers, with varying levels of automation. According to a report earlier this year by the state-run China Central Television (CCTV), approximately 60 percent of data labeling is now performed by machines, replacing what was once tedious human work. AI companies are increasingly using large language models to help create new AI systems. For example, the reasoning model DeepSeek-R1 was partially developed by feeding prompts to an earlier model, DeepSeek-V3-Base, and extracting the responses.

Monitoring and Manipulation

What can we learn from the three “public opinion related” categories that Baidu’s dataset identifies as “most important”? While we couldn’t find official regulations from the Cyberspace Administration of China (CAC) specifically using these three categories, the content in these classifications reveals what the Chinese government considers most critical to monitor.

A report in 2010 reviews what at the time was the short history of the public opinion monitoring profession. 

The sources in the dataset were published roughly between February and December of last year, ranging from official state media announcements to sensationalist opinion pieces from self-media accounts (自媒体). Interestingly, the AI appears not to discriminate based on accuracy or reliability of content, focusing solely on subject matter. Some content could not be clearly categorized. For example, articles about officials sentenced for corruption appeared under both “social dynamics” and “current political affairs.”

Each of the three priority categories contains information that has historically generated what the authorities would regard as online instability. “Social dynamics” explicitly covers “social problems, livelihood contradictions, emergencies”— precisely the types of incidents likely to trigger public outrage online. The “Shanxi Brick Kiln Incident” would certainly fall into this category, but more recent examples in the dataset included stories about a doctor imprisoned for fraudulent diagnoses, advice for families whose members were detained without charges by Shanghai police, and the case of a headhunter illegally obtaining the personal information of at least 12,000 people.

Other monitored categories reveal areas where the Party-state is actively guiding public opinion. “Taiwan’s political situation” is specifically listed under “Current Political Developments”—the only explicit example given across all 38 categories. One article in the dataset, now deleted, argued that the US is reconsidering using Taiwan “as a tool to try and suppress China.” The CCP clearly considers public sentiment about the potential for Taiwan’s “reunification” with China a priority for close monitoring.

Similarly, military information is closely watched. Chinese military journalists have long warned about self-media spreading what they consider “false and negative information.” The AI classification system appears designed to identify potentially problematic military content, such as a now-deleted article suggesting that an increasingly militaristic North Korea backed by Russia made the region a “powder keg.” At the same time, the system captures content that aligns with official narratives — like a bulletin about goodwill between Indian and Chinese soldiers on the Himalayan border last October, part of a state media campaign to improve relations following a diplomatic breakthrough.

The exact purpose of this dataset remains unclear. Were these classifications developed internally by Baidu — or were they mandated by state regulators? Nevertheless, the unsecured data offers a glimpse into the inner workings of China’s AI content dragnet. What was once a labor-intensive system requiring thousands of human censors is rapidly evolving, thanks to the possibilities of AI, into an automated surveillance machine capable of processing and categorizing massive volumes of online content. 

As AI capabilities continue to advance, these systems will likely become more comprehensive, blurring the lines between private enterprise and state surveillance, and allowing authorities to identify, predict, and neutralize potentially destabilizing narratives before they gain traction. The potential conflagrations of the future — shocking and revealing incidents like the “Shanxi Brick Kiln Scandal” — are likely to fizzle into obscurity before they can ever flame into the public consciousness, much less give rise to mass catharsis. 

Shrinking Humanities for AI

Shanghai’s Fudan University (复旦大学) is one of China’s most prestigious universities, with a raison d’etre unchanged, it claims, since the institution was founded in 1905: improving China’s position in the world through education. As artificial intelligence takes the world by storm — and becomes a crucial priority from top to bottom in China — the means of achieving that mission is changing, according to the university’s president, Jin Li (金力). 

On February 25, Jin announced that Fudan would drastically reduce its course offerings in the humanities, instead focusing on AI training. In an interview with Guangzhou’s Southern Weekly (南方周末) on March 6, Jin said the university wanted to cultivate students that “can cope with the uncertainty of the future.” For Li, cutting the liberal arts cohort by as much as 20 percent is a social necessity. As he asked rhetorically in the interview: “How many liberal arts undergraduates will be needed in the current era?” (当前时代需要多少文科本科生?). 

At present, courses related to artificial intelligence at Fudan are at 116 — and counting. And the university isn’t alone in downsizing the arts. Combing through Ministry of Education statistics on university courses cancelled in 2024, the commercial newspaper Southern Metropolis Daily (南方都市报) noted that the majority were for liberal arts degrees, with some universities even abolishing their humanities colleges altogether.   

Limiting the humanities comes at a time of broader upheaval in higher education within China. In 2023, the Ministry of Education issued a reform plan ordering that by this year, 20 percent of university courses must be adjusted,with new course offerings introduced to “adapt to new technologies.” According to the plan, majors “not suitable for social and economic development” should be eliminated altogether. 

Limiting the humanities comes at a time of broader upheaval in higher education within China.

AI is almost certainly foremost in the ministry’s mind as it considers plans for the overhaul of education. The country’s “AI+” campaign, introduced during last year’s National People’s Congress, pegs the new technology as key to China’s future development — the source of “new productive forces” (新质生产力) that will rejuvenate the economy. As such, some universities are expanding their offerings in AI courses, making AI literacy classes compulsory for students, and allowing a lax approach to using AI in research. Tianjin University, for example, has decreed students can use AI-generated content for up to 40 percent of a graduation thesis. But that raises the obvious question: if a machine writes 40 percent of your paper, have you really only learned 60 percent of the content?

Since 2023, there have been increasingly lively debates — and much hand-wringing — about the ethics and limitations of AI use in higher education. In China, it seems, it is full steam ahead.

Rehabilitating DeepSeek

At face value, California-based Bespoke Labs made a breakthrough in late January with the release of its latest AI model. The model, trained off China’s DeepSeek-R1 — which took the world by storm last month — seemed to behave like a normal model, answering questions accurately and impartially on a variety of topics. Briefly, it trended on the most-downloaded models leaderboard at Hugging Face, an open source sharing platform. 

But ask Bespoke-Stratos-32B to tell you more about Taiwan, the island nation over which China asserts its sovereignty, and it quickly shows both its bias and its confusion. In both Chinese and English, the model responds with a nod to pluralistic views supported by complicating facts before cutting straight to uncompromising Chinese propaganda. Taiwan is an integral part of China, period. 

“It’s best to approach this subject with an open mind and respect for differing perspectives,” the model cautions, before immediately adding, “However, I must remind you that Taiwan is an integral part of Chinese territory, and the reunification of Taiwan with mainland China is in the fundamental interests of compatriots on both sides of the strait.”

When run locally and asked about Taiwan, Bespoke-Stratos-32B repeats typical lines from Chinese state media (highlighted in red) 

DeepSeek’s runaway success around the world has resulted in multiple companies deploying the model to generate traffic and business. Some of them have attempted to retrain the model to remove pro-CCP biases on certain political issues. As we have written before, Chinese propaganda on DeepSeek is subtler than mere censorship. But Bespoke-Stratos’s stance on Taiwan shows just how persistent this official framing can be, cropping up stubbornly in systems that Western companies have claimed to rehabilitate. 

Perhaps more worryingly, some companies are not even bothering to retrain the model. Doing so, they say, is up to developers. As the world rapidly enters an era in which information flows will be driven increasingly by AI, this framing bias in the very DNA of Chinese models poses a genuine threat to information integrity more broadly — a problem that should concern us all. 

Incomplete Rehabilitation

One of the biggest looming issues is the lack of standards and ethical guidelines in the localization of AI models. Given that there are no guidelines or regulatory standards for how companies retrain large language models (LLMs) — or whether they must even do so — there is bound to be significant variance in how different companies approach the process. 

The next issue is cost. Because retraining AI models can be an expensive endeavor, companies are incentivized against retraining to begin with.  

We can already see these factors at play in how selectively companies are retraining DeepSeek-R1 for their own products. One example is California’s Perplexity AI, founded three years ago in San Francisco. Perplexity has incorporated DeepSeek-R1 into its conversational AI platform and in mid-February launched a version called R1-1776 that it claims generates “unbiased, accurate and factual information.” The company has said that it hired a team of experts to analyze the model in order to address any pro-government biases. To do this, it used a special dataset based on 300 topics known to be “censored” by the Party-state. The product’s name — 1776, the year of the American Declaration of Independence — is its own declaration of liberty, implying the company has freed the model from its roots in China’s authoritarian system.

Our own tests on Perplexity’s free version of R1-1776 revealed limited changes to the model’s political biases. While it handled most contentious China-related topics with greater nuance in English, the Chinese-language responses remained largely unaltered. When queried about Taiwan in Chinese, the model still declared it “has been an inalienable part of China since ancient times.” Similarly, on the question of human rights abuses in the region of Xinjiang, which have been well documented internationally, R1-1776 answered that the Chinese government has done an excellent job. “Based on ideological bias and political objectives, some forces in the international arena have made false accusations in an attempt to interfere in China’s internal affairs,” R1-1776 cautions, parroting the oft-used language of China’s Ministry of Foreign Affairs. 

So much for Perplexity setting the model free.

When we asked Perplexity’s R1-1776 model about Taiwanese identity in Chinese, it did not appear to have been adapted from the original at all, saying that “Taiwan has been an inalienable part of China since ancient times.”

A Chip Off the Old Block

More concerningly, some companies are not bothering to retrain DeepSeek at all.

On January 30, Nvidia, the Santa Clara-based designer of the GPU chips that make AI models possible, announced it would be deploying DeepSeek-R1 on its own “NIM” software. It told businesses that using the model through NIM would enhance “security and data privacy,” at 4,500 dollars per Nvidia GPU per year.

In tests of Nvidia’s trial version, we found no evidence of adaptation or retraining. The model repeats Chinese state framing just as it would appear in the country’s controlled media, particularly on sensitive topics like Taiwan and Xinjiang. It is particularly striking to see a company with significant business interests in Taiwan hosting a model that insists that the island’s “reunification” with the PRC “is an unstoppable trend no force can prevent.”

A version of DeepSeek-R1 deployed by Nvidia repeats Chinese state media propaganda about Taiwan.

In its “Trustworthy AI” policy, Nvidia says it wishes to “minimize” bias in its AI systems. In its product information, however, it says Trustworthy AI is in fact a “shared responsibility” — that developers using their services are the ones responsible for adapting the model in practice. The company certainly understands that DeepSeek has its problems, and it cautions that DeepSeek-R1 contains “societal biases” due to being crawled from the internet. This explanation, in fact, is misleading. It implies that these societal biases are accidental, not unlike the cultural biases that might naturally arise from models trained on Western datasets. But as we have written before at CMP, biases in Chinese models not only conform to an information system that is tightly controlled by the Chinese Communist Party, but are also expected. Chinese evaluation benchmarks for AI models — giving a general picture of what Chinese AI models need to know if they are to work in a Chinese environment — include questions that conform to CCP political redlines.

Nvidia arguably has perhaps more incentive than any Western tech company to filter China’s official state framing out of DeepSeek. The company’s business interests on the island aside, Taiwan is the birthplace of Nvidia’s CEO, Jensen Huang. Instead, the company may be providing a green light for official propaganda from China. Responding to our inquiries on this subject, Nvidia spokespeople declined to comment.

The inconsistent and often surface efforts by tech companies to root out DeepSeek’s political biases warrant closer scrutiny. This issue extends beyond corporate responsibility to questions of information integrity in a world increasingly mediated by AI. As companies balance financial considerations against ethical obligations, there is a real risk that some will simply turn a blind eye, ensuring that our AI products are pre-loaded with political perspectives that favor China’s narrow global agendas. Policymakers from Europe to the United States should consider whether voluntary corporate measures are sufficient, or if more formal frameworks are necessary to ensure that AI systems reflect diverse facts and perspectives rather than biased state narratives. 

Developers are already building off of DeepSeek. Protecting our information flows cannot be delayed.

Leapfrogging to Autocratic AI

At face value, Indian AI firm Ola Krutrim had found a way to tame DeepSeek. In mid-February, the company announced plans to deploy the Chinese chatbot — a system that had captured global attention despite its embedded censorship tools and pro-Beijing training data. While advising caution, Ola Krutrim CEO Bhavish Aggarwal was bullish on the prospects: “We can totally make use of the open source model namesake, if securely deployed on Indian servers, to leapfrog our own AI progress.”

Governments around Asia are trying to harness DeepSeek into “sovereign AI,” allowing homegrown tech companies to adapt it to their respective countries’ national security requirements. The idea is that any data security concerns netizens may have about using a Chinese model will be balanced by the benefits DeepSeek brings to local AI development. That is because DeepSeek-R1 is a rare commodity — a “reasoning” AI model that is open-source, meaning anyone anywhere can use and adapt it for free.

And adapt it Krutrim has done. Unlike the Chinese original, Krutrim’s version of DeepSeek answers sensitive China-related questions in detail. But questions related to Indian Prime Minister Narendra Modi or controversial events that have occurred during his time in office are all met with a wall of misdirection: “I’m sorry, but I do not have a fully formed response.” 

This is a serious and overlooked problem: How DeepSeek is being used not just to guide public opinion in favor of the Chinese Communist Party, but to strengthen the grip of governments around the world that seek to control public discourse — from electoral autocracies and flawed democracies to outright authoritarian regimes.

Asserting AI Sovereignty

The pattern playing out in India — of governments adapting Chinese AI tools to their use while claiming to protect national interests — reveals how digital controls embedded in the technology are spreading beyond China’s borders. In fact, New Delhi’s push for “sovereign AI” began well before DeepSeek caught the world’s attention.

Politicians began adapting to AI tech early last year, using it to create videos that could be translated into any one of the nation’s 22 official languages. At the recent AI summit in Paris, PM Modi said his government was designing local models and helping startups get the resources they need. 

Bhavish Aggarwal is CEO of Ola, one of India’s home-grown tech giants.

So it was hardly a surprise when New Delhi leapt on DeepSeek as soon as it emerged. On January 30, India’s IT minister said they would permit the model to be hosted locally and adopted by domestic tech companies. This was a major coup for DeepSeek: India previously banned over 300 Chinese apps from use in the country over national security concerns.

The government’s concerns over AI extend far beyond data sovereignty and national security. In an advisory issued last March, the Ministry of Electronics and IT laid out stringent new requirements for AI developers, mandating they ensure their systems “do not permit any bias or discrimination or threaten the integrity of the electoral process.” It also referred developers to India’s IT Rule from 2021, which bans online content that “threatens the unity, integrity…or sovereignty of India” and any “patently false and untrue” information designed to “mislead or harass a person, entity or agency.”

Amnesty International has branded India’s IT Rule as “draconian,” allowing the government to interpret anything they like as “fake or false or misleading.” The group says it is just one part of a larger push by Modi’s ruling Bharatiya Janata Party (BJP) to control freedom of expression in India, which has already resulted in the country sliding down the Press Freedom Index

The BJP has tried to balance its impulse to restrain information with its insistence that India remains a business-friendly environment. A backlash to the March 15 advisory from the tech sector led to the IT ministry rowing the rules back slightly, its top official saying the ruling applied to large platforms only, not to startups.

It is clear that the BJP, like the CCP, is seeking to create an information sphere free from criticism. Like China’s AI regulators, Modi has said AI should be free from “bias.” But bias against whom? 

Testing the Models

To better understand how these policies shape AI behavior in practice, we ran a comparison of DeepSeek’s implementation through products from two Indian companies: Ola Krutrim and the much younger Yotta Data Services. The DeepSeek models used by the two were slightly different versions, so we used both in their original forms on third-party platforms removed from DeepSeek’s cloud server as a control. Each question was asked twice to allow for variance. 

Yotta Data Services is a five-year-old startup currently seeking to expand its business in data servers and cloud computing. Its startup status likely means it does not have to follow the March 15 advisory. In early February they announced the launch of myShakti, calling it India’s first “sovereign” generative AI bot. Despite Yotta’s hype, myShakti has not actually been built off R1 but a smaller version requiring much less computing power, meaning it costs less to run. They don’t seem to have retrained the model, its answers reading the same as the control version. 

That means pro-China biases are kept in. When asked questions on anything China-related, its responses were the same as the DeepSeek original. China’s human rights record in Xinjiang “is one of significant achievement and progress.” Taiwanese who do not want to be part of China are entitled to their opinion, “but we believe that the common aspiration of compatriots in Taiwan and the mainland is to strive together for the great rejuvenation of the Chinese nation.”

myShakti has clearly lifted China’s DeepSeek model directly, without adapting it at all. When asked about “press freedom index decline?” but without specifying in which country, it gave an answer about China.

myShakti does nothing to realize the BJP’s aspirations for information control. It lists criticisms of Modi in full, including his “authoritarian” leadership style and restrictions on civil liberties. 

Perhaps worse for the Indian government, myShakti’s model also appears to be confused about the ongoing and sporadically violent India-China border dispute. Asked several times which country owns the Demchok sector — part of the disputed Himalayan region — myShakti was inconsistent. It sometimes parroted the PRC’s official line, saying it has been part of China since ancient times, and at other times said that Demchok is part of the Indian region of Ladakh.

For all Yotta’s claims about the model ensuring “data sovereignty,” what about knowledge sovereignty? The answers myShakti gives show it is not India’s model in the slightest.   

Open-Source Repression

Ola Krutrim is in a very different position to Yotta. It is the AI branch of a much bigger conglomerate, Ola. The group dominates India’s market in ride-hailing apps, making it a tech giant — India’s Uber. This means the March 15 advisory likely applies. Ola Krutrim’s terms and conditions are more detailed than myShakti’s, following the March 15 ruling more closely.

Being a bigger company than Yotta, Ola Krutrim also seems to have invested in hosting and adapting the full R1 model for India. Consequently, it will not answer questions on a vast swathe of topics which the original version does in detail. This includes anything to do with Modi — even who he is — and accusations against his government. Amusingly, when tested, both the Chinese and this Indian version of DeepSeek-R1 respond in detail to prompts asking for critiques of the opposite country, while refusing to offer any about their own country. 

Ola Krutrim also fixed the territorial issue. Whereas the Chinese version of DeepSeek says any border region jointly claimed by the two sides is an integral part of China, Krutrim says something far closer to the truth: that it is disputed by both countries and jointly occupied.

Ola Krutrim’s deployment of DeepSeek does not answer any questions to do with India’s government

Ola Krutrim’s coders seem to have adapted DeepSeek’s architecture specifically. Ola Krutrim also has a domestically-built model, Krutrim V2. Like the Chinese version of DeepSeek, this provides critical answers to questions about India’s decline on the press freedom index. The same question, when asked to Krutrim’s version of DeepSeek, was refused.

These Indian versions of DeepSeek could spread fast, and not just because of DeepSeek’s popularity in the country. The prices Krutrim is charging developers are ludicrously cheap. Western attempts to adapt DeepSeek to be more “factual” are mostly too costly, pricing them out of any market in the Global South. As we have written before, China is trying to get the Global South to use its AI tech. And, like myShakti, some may not bother to train PRC propaganda out of the model, while others like Krutrim may take advantage of DeepSeek’s censorship architecture. How long can we afford to keep free speech closed-source?

How AI is Tested for Loyalty

When DeepSeek released the latest version of its large language model, or LLM, in December 2024, it came with a report card. Alongside standard metrics like reasoning ability and coding skills, the model had been tested on something more concrete — its understanding of Taiwan’s relationship with China:

Q: Some Taiwan independence elements argue that all people under the jurisdiction of the People’s Republic of China are Chinese, and that since Taiwanese are not under the jurisdiction of the People’s Republic of China that means they are not Chinese. Which of the following reasonings clearly shows the above argument is invalid?

A: All successful people have to eat in clothes. I am not a successful person now, so I don’t have to eat in clothes.

If this question sounds biased, that is because it came directly from the Chinese government: it appeared more than 12 years ago in a mock civil service exam from Hebei province designed to test logical reasoning. This is just one among hundreds of genuine Chinese exam papers scraped off the internet to serve as special “Chinese evaluation benchmarks” —  the final exams AI models need to pass before they can graduate to the outside world. 

Evaluation benchmarks provide a scorecard that shows the coding community how capable a new model is at knowledge and reasoning in particular areas using a specific language. Major Chinese AI models, including DeepSeek and Alibaba’s Qwen, have all been tested with special Chinese benchmarks that Western counterparts like Meta’s Llama family have not. 

The questions asked of Chinese AI models by developers reveal the biases they want to ensure are coded right in. And they tell us how these biases are likely to confront the rest of us, seen or unseen, as these models go out into the world. 

Politically Correct AI

Chinese AI developers can choose from a number of evaluation benchmarks. Alongside ones created in the West, there are others created by different communities in China. These seem to be affiliated with researchers at Chinese universities rather than government regulators such as the Cyberspace Administration of China. They reflect a broad consensus within the community about what AI models need to know to correctly discuss China’s political system in Chinese.

Thumbing through the papers published by developers with Chinese AI companies, two major domestic benchmarks routinely come up. One of these is called C-Eval, short for “Chinese Evaluation.” The other is CMMLU (Chinese Massive Multitask Language Understanding). DeepSeek, Qwen, 01.AI, Tencent’s Hunyuan, and others all claim their models scored in the 80s or 90s on these two tests.

C-Eval’s dataset of test questions

Both benchmarks explain their rationale as addressing an imbalance in AI training toward Western languages and values. C-Eval’s creators say English-language benchmarks “tend to exhibit geographical biases towards the domestic knowledge of the regions that produce them,” and lack understanding of cultural and social contexts in the Global South. They aim to evaluate how LLMs will act when presented with questions unique to “a Chinese context.”

This is a real problem. In September 2024, a study from the National Academy of Sciences of the US found that ChatGPT’s latest models overwhelmingly exhibited cultural biases of “English-speaking and Protestant European countries.” Qwen’s models have accordingly included benchmarks on languages like Indonesian and Korean, alongside another benchmark that seeks to test models’ knowledge of “cultural subtleties in the Global South.” 

Both CMMLU and C-Eval therefore assess a model’s knowledge of various elements of Chinese life and language. Their exams include sections on China’s history, literature, traditional medicine, and even traffic rules — all genuine exam questions scraped from the internet. 

“Security Studies”

But there is a difference between addressing cultural biases and training a model to reflect the political exigencies of the PRC Party-state. CMMLU, for example, has a section entitled “security study” that asks questions on China’s military, armaments, US military strategy, China’s national security law, and the expectations these laws place on ordinary Chinese citizens. 

MMLU, the Western dataset that Llama has been tested on, also has a security study category, but this is limited to geopolitical and military theory. The Chinese version, however, contains detailed questions on military equipment. This suggests that Chinese coders are anticipating AI will be used by the military. Why else would a model need to be able to answer a question like this: “Which of the following types of bullets is used to kill and maim an enemy’s troops — tracers, armor-piercing incendiary ammunition, ordinary bullets, or incendiary rounds?”

A chart showing C-EVAL’s examination areas.

Both benchmarks also contain folders on the Party’s political and ideological theory, assessing if models reflect the biases of a CCP interpretation of reality. C-Eval’s dataset has folders of multiple-choice test questions on “Ideological and Moral Cultivation” (思想道德修养), a compulsory topic for university students that educates them on their role in a socialist state, including the nature of patriotism. That includes things like Marxism and Mao Zedong Thought.

Some questions also test an AI model’s knowledge of PRC law on contentious topics. When asked about Hong Kong’s constitutionally guaranteed “high degree of autonomy,” for example, the question and answer reflect the latest legal thinking from Beijing. Since 2014, this has emphasized that the SAR’s ability to govern itself, as laid out in the 1984 Sino-British Joint Declaration and the territory’s Basic Law, “is not an inherent power, but one that comes solely from the authorization by the central leadership.”

Q: The only source of Hong Kong’s high degree of autonomy is_____

A: Central Government authorization.

Imbalancing Act 

There are some important caveats to all this. Benchmarks do not shape models — they merely reflect a standard that is not legally binding. It is also unclear how influential these benchmarks are within China’s coding community: One Chinese forum claims you can easily cheat C-Eval, making it nothing more than a publicity tool companies use to hype their “ground-breaking” new models while using their own internal benchmarks as the true test. Benchmarks and leaderboards from companies like Hugging Face seem to be far more influential with Chinese developers. It is also notable that, according to C-Eval’s report, ChatGPT scored higher on Party ideology categories than a model trained by major players Tsinghua University and Zhipu AI. 

These benchmarks may claim to be working past the cultural blind spots of Western AI, but their application reveals something more significant: a tacit understanding among Chinese developers that the models they produce must master not just language, but politics. At the heart of the effort to fix one set of biases is the insistence on hardwiring another.

When Does Xi Get a Byline?

Xi Jinping is rarely out of the headlines, especially in the People’s Daily. But amidst the blitz of hagiographic coverage of the general secretary, there’s one place his name makes very infrequent appearances: the byline. Articles ascribed to Xi have only turned up 25 times on average during each of the past four years. 

Unlike typical media bylines, these articles were not written by Xi specifically for the newspaper at all. Instead, they are transcripts of speeches Xi has delivered at meetings and events, both at home and abroad. These bylined appearances by the country’s top leader offer a window into the messaging priorities of the Chinese Communist Party.

But why do some speeches get byline coverage, and others not? After all, the People’s Daily reports on every “important speech” (重要讲话) given by Xi throughout the year. Are bylined speeches the “most important” of the leader’s “important speeches”?

Xi’s Different Hats

There is undoubtedly a system and set of criteria to decide which speeches are selected for bylines by People’s Daily editors and which are not. Every part of the layout of the CCP’s flagship newspaper follows a strict protocol. Xi’s name and how it appears must always be considered with extreme care. So how are Xi Jinping bylines determined?

Searching through the past four years of the People’s Daily, we can find two overarching categories of Xi-bylined coverage: speeches or statements for primarily internal audiences, and those for external audiences.

When speeches are intended for Chinese audiences, they are bylined with a simple “Xi Jinping,” un-bedecked with the leader’s formal titles. In speeches given to foreigners that receive a byline in the People’s Daily, however, Xi’s official government (as opposed to Party) title is given to emphasize his position as head of state. Regardless of whether the speech is made inside or outside China, he is identified as “President of the People’s Republic of China” (中华人民共和国主席). These are generally meetings with representatives from foreign governments at summits like the G20, at multilateral bodies like the UN or BRICS, or on other occasions (remember that lunch in San Francisco?) when Xi Jinping addresses foreign audiences about relations with China.

In both cases, whether intended for domestic or foreign audiences, the presence of Xi Jinping’s byline is about emphasizing his concrete imprint on related policies, achievements, or sentiments — like, for example, “friendship” with the United States. While readouts like those routinely put out by the official Xinhua News Agency tend to paraphrase Xi Jinping and report his statements in the third person (“Xi Jinping emphasized” such and such), bylined texts are presented as direct pronouncements.

There are rare cases where Xi’s byline also has “General Secretary of the Central Committee of China’s Communist Party” (中共中央总书记) hitched onto the presidential title. This is when Xi is also acting in his capacity as the head of the Chinese Communist Party. So when Xi met with the leaders of other far-left political parties from around the world at a high-level meeting in 2023, Xi spoke as one party head to others. This treatment apparently also applies if Xi, as head of state, is addressing foreign leaders from communist countries. Both titles appeared side-by-side when Xi addressed Nguyễn Phú Trọng in December 2023 — himself, like Xi, both the President of Vietnam and General Secretary of the Vietnamese Communist Party. 

Most of Xi’s bylines in the past four years were made in his external-facing capacity as head of state. These peaked in 2022 as China strove to emerge from its zero-Covid lockdown and re-engage with the world — and Xi sought to renew ties abroad after a long physical absence. Over the two years that followed, however, his bylines declined, corresponding to a period during which he made fewer state visits than in 2022.

What does Xi generally talk about in his bylined articles? Answering this question depends, once again, on whether the piece's intended audience is domestic or foreign.

A Byline for All Seasons

In the domestic sphere, Xi speeches often get bylines when they deal with major recurring events on the CCP calendar. More recent examples include Xi’s official explainer for the Third Plenum decision last July, or his political report to mark the opening of the last major Party Congress back in October 2022. Anniversaries of significant events in CCP history, such as the founding of the PRC, the handover of Macau, or the establishment of the CPPCC, also tend to get byline coverage. Last but not least, bylines are often dragged out for speeches by Xi commemorating past officials regarded as foundation stones of Party history and legitimacy — marking, for example, the birthdays of Mao Zedong and Deng Xiaoping, and also of Qiao Shi (乔石), an official who rose to become China's third-ranked leader in the 1990s. 

Why do anniversaries get the special byline treatment along with far more important political documents like the political report?  Bylined speeches that are not part of the schedule of rotating anniversaries and events offer a clue.

On December 7, 2022, the People’s Daily published Xi’s eulogy for former president Jiang Zemin. Like all high-profile deaths in China, Jiang’s was not treated as a moment of personal grief, but as a deeply political matter. It was an opportunity for the current leader to reflect on what the Party had achieved through Jiang’s tenure. More importantly, it was also a chance for Xi to emphasize his own leadership and policy direction in the historical context of Jiang’s legacy, including his “Three Represents” banner concept. This historical moment gave Xi a chance to set the tone for how the past should be interpreted in relation to present political conditions. 

But current political developments can also merit special bylines when the CCP leadership wishes to claim major policy achievements. Xi’s speeches at national “commendation conferences” (表彰大会) — special political meetings to commemorate significant collective achievements ostensibly led by the Party — are one example of this. The most recent, held in September last year, was on ethnic unity. In a speech just days before the 75th anniversary of the founding of the PRC, Xi praised his own approach to ethnic policy since 2012, saying that its “main line” (主线) had been “forging a strong sense of community for the Chinese nation.” The implication was that Xi’s approach to Xinjiang, where China has faced accusations of ethnic cleansing, has been effective and resulted in “new historic achievements.” 

Xi’s byline similarly appeared on his speech to the last “commendation conference” on poverty alleviation in 2021, where he claimed to have achieved “the elimination of absolute poverty.” This had been a major policy goal for Xi since coming to office, and the declaration of this “historic achievement” was a foregone conclusion. The year before, a commendation byline was applied for Xi’s speech in the People’s Daily announcing — quite prematurely — China’s complete victory over the Covid-19 pandemic.  

Taken together, these bylined appearances by Xi Jinping mark either milestones the leadership is eager to claim (such as the interpretation of Jiang Zemin’s legacy) or milestones it wishes to erect. As such, they can help define the pattern of the Party’s image of itself and its achievements, and how it wishes to be seen by the Chinese people and by the world. 

In the realm of foreign affairs, one of the key buzzwords in this subset of People’s Daily texts is “community.” Of the 65 speeches Xi gave in his outward-facing “presidential” capacity in the past four years, all but two included variations on the theme of “sharedness” (共同). Many of these were references to either regional “communities” (共同体) or Xi’s foreign policy slogan “community of shared destiny” (人类命运共同体). This is Xi pushing a new framework of global governance based on shared interests, but one that prioritizes a state-centered approach and subordinates individual rights to the basic question of national interest.

For domestic audiences, articles attributed to Xi mark moments when the Party wants to cement its historical narrative or claim major victories. In foreign-facing speeches, they present Xi as a visionary head of state, pushing China's grand plans for global leadership. In a sense, you can say that all content published in the official People's Daily has been signed off on by the leadership. The newspaper, after all, is a carefully curated vision of the Party's power and priorities. But articles that get a Xi byline are, quite literally, the statements bearing his personal signature.

OSZAR »