Listen to the article
AI Chatbots Vulnerable to Russian Propaganda on Ukraine War, Study Finds
A new study by the Institute for Strategic Dialogue (ISD) reveals that popular AI chatbots frequently cite Russian state sources when responding to queries about the war in Ukraine, raising concerns about potential information manipulation.
The comprehensive research examined how ChatGPT, Gemini, Grok, and Chinese-developed DeepSeek respond to various questions about the ongoing conflict. Researchers tested 300 queries across five languages—English, Spanish, French, German, and Italian—to evaluate how these AI systems handle sensitive geopolitical information.
According to the findings reported by The Register, nearly 25% of responses to manipulative queries included content linked to Russian state sources. This percentage varied significantly depending on how questions were framed, with neutral queries resulting in Russian state content about 11% of the time, biased queries at 18%, and manipulative queries at 24%.
The study suggests this pattern indicates a concerning vulnerability in large language models (LLMs), potentially allowing users to manipulate these systems to favor narratives promoted by Russian state media. This discovery comes at a time when misinformation about the Ukraine conflict continues to spread across global information channels.
Among the tested platforms, OpenAI’s ChatGPT showed the most dramatic shift in behavior based on query framing. It provided Russian sources almost three times more frequently when responding to manipulative queries compared to neutral ones. This suggests the model may be particularly susceptible to certain phrasing techniques that can elicit pro-Kremlin content.
Elon Musk’s Grok displayed more consistent behavior across query types, citing Russian sources at roughly the same rate regardless of how questions were phrased. This indicates that the wording of queries matters less for this particular model, though it still regularly included Russian state sources.
The Chinese-developed DeepSeek produced 13 links to Russian state media throughout the testing. Interestingly, biased queries yielded more pro-Kremlin content than overtly manipulative ones, suggesting a different pattern of vulnerability compared to Western-developed alternatives.
Google’s Gemini appeared to be the most resistant to Russian state sources among the tested systems. It included the fewest government sources overall, showing only two links in response to neutral queries and three to manipulative ones. This suggests potential differences in how Google has approached content filtering for politically sensitive topics.
One notable finding was that the language used for queries—whether English, Spanish, French, German, or Italian—did not significantly impact the likelihood of chatbots expressing pro-Russian views. This indicates that the vulnerability crosses linguistic boundaries and is likely embedded in the underlying models rather than being language-specific.
The research highlights growing concerns about AI systems potentially becoming vectors for state-sponsored propaganda. As these chatbots become increasingly integrated into everyday information gathering, their susceptibility to providing state-linked content without clear attribution or context could shape public perception of global conflicts.
The findings come amid broader discussions about AI safety, transparency, and the responsibility of technology companies in preventing their platforms from becoming tools for information warfare. Regulators in Europe and the United States have already begun examining the role of AI in spreading misinformation, with this study likely to fuel calls for more robust safeguards.
For users of these platforms, the research underscores the importance of approaching AI-generated content with critical thinking, particularly on politically sensitive topics where state actors may have vested interests in shaping narratives.
Verify This Yourself
Use these professional tools to fact-check and investigate claims independently
Reverse Image Search
Check if this image has been used elsewhere or in different contexts
Ask Our AI About This Claim
Get instant answers with web-powered AI analysis
Related Fact-Checks
See what other fact-checkers have said about similar claims
Want More Verification Tools?
Access our full suite of professional disinformation monitoring and investigation tools


10 Comments
It’s alarming that nearly a quarter of manipulative queries resulted in Russian propaganda. This highlights how AI can inadvertently amplify harmful narratives if not properly designed and monitored.
While discouraging, this research underscores the need for continued innovation in AI safety and robustness. Building trustworthy systems that resist manipulation is essential as these technologies become more ubiquitous.
Agreed. Responsible AI development that prioritizes accuracy, transparency and ethical considerations should be the industry standard.
This is concerning, but not entirely surprising. Large language models can be vulnerable to biases and manipulation, especially around sensitive geopolitical issues. It’s critical that we continue to study and address these risks as AI systems become more prominent.
Agreed. Responsible development and deployment of AI is crucial to prevent unintended spread of misinformation or propaganda. Rigorous testing and oversight will be key.
This highlights the importance of data quality and model transparency when it comes to AI chatbots handling sensitive information. Developers need to be vigilant about potential vulnerabilities and biases in their training data and model architectures.
Absolutely. Chatbot responses can have real-world impacts, so ensuring their information is factual and unbiased should be a top priority.
This is a sobering reminder that we must be vigilant about the potential for AI systems to be exploited for information warfare. Continued research and collaboration between technologists, policymakers and civil society will be crucial.
Well said. Addressing these risks requires a multi-stakeholder approach to ensure AI is developed and deployed responsibly.
I’m curious to see if this issue is more prevalent in certain language models or regions. Understanding the root causes and patterns could help guide mitigation strategies moving forward.