“`html
ESMO Issues First Guidance on Large Language Models in Oncology
BERLIN – The European Society for Medical oncology (ESMO), a leading professional association for European oncologists, has released its first set of guidance on the responsible use of large language models (LLMs), a type of artificial intelligence, in cancer medicine. The guidance, published on February 15, 2025, in the Annals of Oncology, aims to help members navigate the potential benefits and risks of these rapidly evolving technologies.
The authors emphasize that the oncology community should not dismiss the potential advantages AI offers to cancer patients. However, they together caution that current evaluations of LLM-powered chatbots for patients and tools for clinicians are insufficient to fully address the inherent risks associated with generative AI in healthcare. This guidance arrives as the field grapples with integrating AI while maintaining patient safety and ethical standards.
context and Concerns Surrounding AI in Oncology
Large language models, such as those powering chatbots, are increasingly capable of processing and generating human-like text. In oncology, these models could perhaps assist with tasks like summarizing medical literature, generating patient education materials, and even aiding in treatment decision-making.However,concerns remain regarding the accuracy,bias,and potential for misinformation generated by these systems. A study published in nature in January 2024 highlighted that LLMs can perpetuate existing biases in healthcare data, leading to disparities in care recommendations (Obermeyer, Z.,et al. (2024). Bias in large language models for healthcare. Nature,625(8103),148-154.).
The ESMO guidance acknowledges these concerns, specifically addressing the risk of “hallucinations” – instances where LLMs generate factually incorrect or misleading information. It also highlights the potential for LLMs to exacerbate existing health inequities if not carefully implemented and monitored. The guidance stresses the importance of human oversight and validation of any AI-generated content before it is indeed used in clinical practice.
ESMO’s Guidance: Key Recommendations
The ESMO guidance outlines several key recommendations for oncologists using LLMs. These include:
- Clarity: Clinicians should be clear with patients about the use of AI in their care.
- Validation: AI-generated information should be carefully validated by a qualified healthcare professional before being shared with patients or used to inform treatment decisions.
- Data Privacy: Patient data used to train or operate LLMs must be protected in accordance with relevant privacy regulations, such as the General Data Protection Regulation (GDPR) in Europe.
- Continuous Monitoring: The performance of LLMs should be continuously monitored for accuracy,bias,and potential harms.
- Education and Training: Oncologists should receive adequate training on the capabilities and limitations of LLMs.
The guidance also emphasizes the need for ongoing research to evaluate the clinical utility and safety of LLMs in oncology. ESMO plans to update its recommendations as new evidence emerges and the technology evolves.
Comparison with ASCO’s Approach
The American society of Clinical Oncology (ASCO) has also addressed the use of AI in cancer medicine, issuing its own set of principles in November 2023. while both organizations share a commitment to responsible AI implementation,ASCO’s approach focuses more on broad ethical
