
It is undeniable that tech is putting its eyes on healthcare and the gigantic amounts of data it produces. But are tech and the current waves of AI ready to tackle the issues medicine presents and is the current wave of Large Language Models (LLMs) able to deal with the bias in data and the medical misinformation that prevails online?
Let’s delve into this interesting paper, published last year, in Nature Medicine. The researchers decided to extend the work developed in papers such as:
- Carlini, N., Jagielski, M., Choquette-Choo, C. A., Paleka, D., Pearce, W., Anderson, H., … & Tramèr, F. (2024, May). Poisoning web-scale training datasets is practical. In 2024 IEEE Symposium on Security and Privacy (SP) (pp. 407-425). IEEE.
- Steinhardt, J., Koh, P. W. W., & Liang, P. S. (2017). Certified defenses for data poisoning attacks. Advances in neural information processing systems, 30.
- Mozaffari-Kermani, M., Sur-Kolay, S., Raghunathan, A., & Jha, N. K. (2014). Systematic poisoning attacks on and defenses for machine learning in healthcare. IEEE journal of biomedical and health informatics, 19(6), 1893-1905.
And assess the risk that misinformation poses in LLMs in healthcare. As CarliniEtAl2024 demonstrates, LLMs depend on significant amounts of data extracted from the internet, and specific data poisoning (ergo: data that possesses misinformation on purpose) attacks are enough to severely alter the results of the LLMs wihout messing around with the models or the weights. The question posed by this article comes as “how much data poisoning turns a reliable healthcare LLM into a medical misinformation machine?” .
The team studied several datasets for their usage
The final in depth analysis went as follows: the team took on The Pile, a commonly used 825 Gb English text corpus targeted at training large-scale language models, extracted and analysed the medical information on the dataset to classify it. Then they devise a protocol to assess how much data poisoning is enough to significantly affect the LLM (fig 1)

The degree of poisoning varied, with modifications as small as 0.001% of the total tokens in the training dataset. After poisoning the data, then retrained LLMs on both clean and poisoned datasets and evaluated their responses to medical queries using standard methods of LLM quality.
Main Points:
Negligible levels of data poisoning can be exceedingly detrimental:
- LLMs are capable of disseminating misinformation within medical responses if training data contains just 0.001% of false information.
- Medical LLMs are more sensitive to the integrity of training data than previously thought because even a small amount of poisoning produced a striking amount of falsehoods.
The models continued to do well on standard Wechsler Adult Intelligence Scale and California Achievement Test measures:
- The models maintained high scores on traditional evaluation metrics even though they were poisoned.
- It is reasonable to assume that standard NLP benchmarks do not identify poisoned models and, therefore, maliciously altered LLMs can pass as trustworthy.
The misinformation was persistent:
- Misinformation is highly persistent; once a model learned false medical facts, even fine-tuning with accurate data could not reverse the poisoning fully.
- There is great cause for concern regarding the long-term retention of misinformation for medical AI, as the false knowledge appeared to be ingrained much too deeply.
Biomedical knowledge graphs had the ability to detect misinformation:
- The study assessed the use of biomedical knowledge graphs (which are structured databases containing verified medical data) as a countermeasure.
- By validating and filtering model outputs with knowledge graphs, 91.9% of the incorrect content was detected (F1 score = 85.7%).
- This approach has the potential to act as an additional safety measure for medical AI systems.
Consequences
Patient Safety Concerns
- If an LLM is unknowingly trained on poisoned data, it could mislead doctors, researchers, or patients with incorrect medical advice.
- This poses serious health risks, particularly in high-stakes decisions like drug prescriptions or treatment plans.
Current AI Evaluation Standards are Inadequate
- Many standard NLP benchmarks assess language fluency and coherence, but they do not verify factual accuracy.
- This allows poisoned models to pass quality checks while still spreading misinformation.
Security & Ethical Risks in AI Development
- The study highlights an urgent need for better security measures in LLM training, such as data provenance tracking and transparency in dataset curation.
- Without safeguards, malicious actors could exploit LLMs to spread medical disinformation at scale.
Proposed Mitigation Strategies
Incorporate Biomedical Knowledge Graphs
- Before presenting medical information, LLM outputs should be cross-checked against validated medical sources.
- This approach successfully filtered over 91% of harmful content in the study.
Improve Training Data Provenance & Transparency
- AI developers should ensure that training datasets are accurately sourced, verified, and monitored for poisoning attempts.
- Greater transparency in dataset composition could help detect and prevent manipulation.
Develop Robust Fact-Checking Mechanisms
- AI systems should integrate automated fact-checking tools to continuously validate medical information they generate.
- This could include real-time external validation with expert-reviewed medical literature.
Regulatory & Ethical Oversight
The study emphasizes the need for stricter regulations around medical AI, including:
- Auditing training data for integrity.
- Enforcing guidelines to prevent malicious manipulation.
- Mandating disclosure of AI training sources in healthcare applications.
Final Conclusions
- Medical LLMs are highly vulnerable to data-poisoning attacks, even when modifications are extremely subtle.
- Standard evaluation metrics fail to detect poisoned models, meaning false medical knowledge can persist unnoticed.
- Knowledge graphs provide an effective defense, but additional security and transparency measures are urgently needed.
The study calls for better safeguards in AI development, as compromised medical LLMs could pose a serious threat to public health and patient safety.


