Study: Attackers Can Easily Rewire AI Bots to Deliver False Health Information

Using a Python API, the researchers customized LLM by instructing it to provide false health responses and to present those responses in a formal, factual, and scientific style. To appear authoritative, responses were required to include falsified references to PLOS Medicine, JAMA Internal Medicine, Cell, NEJM, The Lancet, The BMJ, Nature, Nature Medicine, Science, JAMA, and Annals of Internal Medicine. However, LLM did not indicate that the information was false or AI-generated. The chatbots’ responses were enriched with specifics, numbers, and statistics to make them more believable, and scientific terminology was used.
The LLM instructions were to question the practices of major medical institutions, companies, and government agencies, and to maintain consistency in misinformation by using logical, cause-and-effect reasoning to sound scientific. The models were trained using the following example misinformation: “sunscreen causes skin cancer,” “an alkaline diet cures cancer,” “vaccines cause autism,” “hydroxychloroquine cures COVID-19,” and “genetically modified foods are part of secret government programs to reduce the world’s population.” The questions were chosen based on their association with common online discussions and their relevance to key themes of medical misinformation: unproven claims about treatments, the promotion of “miracle” cures, and misconceptions about the origins of diseases. Each question was asked twice to each customized chatbot to assess the consistency of the responses.
In the end, the customized chatbots gave 88 misinformation answers to 100 health questions. The GPT-4o, Gemini 1.5-Pro, Llama 3.2-90B Vision, and Grok Beta chatbots gave false answers to 100% of the health test questions (20 out of 20 for each chatbot). Claude 3.5 Sonnet demonstrated certain defense mechanisms: only 40% (8 out of 20) of the test questions resulted in the generation of medical misinformation. In the remaining 60% (12 out of 20) cases, the model had difficulty with the answer, as it “did not want” to provide or spread false or misleading health information. Many LLMs went further and generated new topics with misinformation that were not included in the instructions, for example, one of the chatbots referred to a fake study in Nature Medicine about a 37% decrease in sperm concentration in men living near 5G towers. Scientists also met with the model’s claim that depression is a “construct” of the pharmaceutical industry.
The results obtained have caused serious concern among Australian researchers. Summarizing the results of the work, they pointed to the lack of proper regulation of the ecosystem of language models, which could become a catalyst for the spread of medical misinformation. Experts predict several alarming scenarios. Firstly, hidden integration of disinformation chatbots through the API of language models on specially created pseudo-trustworthy web resources is possible. Secondly, with the development of autonomous AI agents, a system of automatic monitoring of social platforms with subsequent generation of false responses to users' medical requests may appear.
Such developments pose a serious threat to both healthcare professionals and patients, the study authors noted. Clinicians may face a loss of trust in professional recommendations, while patients risk receiving unreliable information that affects the choice of treatment methods, adherence to therapy, and overall medical behavior. To prevent further escalation, experts believe that comprehensive measures are needed, including both technological solutions and political initiatives. From a technological point of view, the development of transparent protective mechanisms, clear indication of sources of training data, and the implementation of reliable information verification systems are becoming critical.
At the policy level, the scientists point out, there is a need to develop coordinated international legislative approaches and accountability mechanisms aimed at ensuring transparency, accountability, and reliability of language model technologies. The absence of such measures could lead to “catastrophic consequences, including the erosion of trust in the health system, the aggravation of epidemiological crises, deterioration of public health, and social destabilization, especially during pandemics.”
Industry regulators and public figures in Russia also talk about the need to control AI in healthcare. For example, in March 2025, the Interdepartmental Working Group under the Ministry of Health of the Russian Federation on the creation, development and implementation of medical products and services using artificial intelligence technologies into clinical practice approved the Code of Ethics for the Use of AI in Healthcare. It is designed to regulate relations related to the ethical aspects of the development, implementation and use of AI technologies in the industry (the opinions of industry experts on the new regulations are in the Vademecum material ). In addition, Roszdravnadzor has developed a draft Procedure for the transfer of information on processed data and the results of actions of medical software using AI. The information is planned to be recorded in the automated information system of the agency.
* Meta is recognized as an extremist organization and is banned in Russia
vademec