Introduction
In the rapidly evolving landscape of healthcare technology, the integration of Large Language Models (LLMs) stands at the forefront of innovation, offering unprecedented opportunities to transform patient care, medical research, and healthcare operations. However, current mainstream generative LLMs, such as OpenAI’s ChatGPT and Meta’s LlaMa 2, may generate inaccurate, misleading, or even harmful content due to their inherent uncertainly, incompleteness, or bias from training data, among others. These issues can have serious consequences for patient safety when these models are directly used for health applications. Our collaborative workshop, jointly organized by AMIA KDDM, NLP, and KRS Working Groups, aims to delve into the cutting-edge advancements that augment LLM capabilities, focusing on Retrieval-Augmented Generation (RAG), in-context learning, contrastive learning, LLM finetuning (e.g., low-rank adaptation), prompt tuning, and zero-short/few-shot learning. These techniques can be broadly categorized into prompting strategies, retrieval-based strategies, and finetuning strategies, and their optimal use is still in debate. These sophisticated techniques are pivotal in enhancing the quality, accuracy, and reliability of LLM-based applications across a broad spectrum of healthcare domains. One prime example is the development of advanced medical Question Answering (QA) systems that leverage RAG to provide precise, evidence-based answers drawn from vast medical literature and patient data. Similarly, the use of contrastive learning methods enables the creation of more robust diagnostic tools by effectively distinguishing between complex cases with subtle differences in symptoms or imaging. Few-shot learning techniques empower LLMs to adapt to specialized medical fields quickly, using minimal example data to generate knowledgeable responses or identify rare conditions. RAG combines the benefits of retrieval-based methods, which search through a database to find relevant information, with the generative capabilities of LLMs, which can synthesize and predict text. This approach allows LLMs to pull in the most current information from medical databases, research papers, or clinical guidelines, ensuring the generated responses are based on the latest medical knowledge. Beyond these, the workshop will explore explainability and interpretability of LLMs, as well as the ethical considerations in deploying these advanced AI tools in sensitive healthcare environments. By convening experts from healthcare, AI research, and ethics, this workshop seeks to explore the potential of these technologies to provide up-to-date medical information, facilitate better decision-making, and improve patient outcomes. We invite the attendees to navigate the intersection of AI innovation and medical expertise, charting a course toward a future where technology and healthcare work hand in hand to achieve optimal health outcomes, all while ensuring ethical standards and patient privacy are upheld.
By convening experts from healthcare, AI research, and ethics, this workshop seeks to explore the potential of these technologies to provide up-to-date medical information, facilitate better decision-making, and improve patient outcomes. We invite the attendees to navigate the intersection of AI innovation and medical expertise, charting a course toward a future where technology and healthcare work hand in hand to achieve optimal health outcomes, all while ensuring ethical standards and patient privacy are upheld.
Outline
This collaborative workshop is organized three themes: (1) introducing various healthcare and medical applications of LLMs; (2) introducing knowledge-augmented techniques for LLMs (e.g., RAG, in-context learning, LLM finetuning, few-shot learning, incremental learning); (3) multimodal capabilities. We are interested in, but not limited to the following topics during this workshop:
- LLM applications in healthcare
- Knowledge-augmented and retrieval-augmented techniques for LLMs
- Explainability of LLM generation
- Multimodal capabilities of LLMs for health applications
We plan to kick off the workshop with a 15-minute introduction of LLM for medical QA and 30-minute keynote presented by a leading scientist on the overarching themes of LLM applications and knowledge enrichment. After it, there will be two sessions of short talks of 12 minutes each. The first session will include invited speakers who have published relevant studies in the field. The coffee break will also be used for students’ poster presentations. We will call for student presentation abstract. The abstracts will be reviewed by workshop organizers, and five abstracts will be accepted for presentation. After the coffee break, there will be a second session of invited short talks. In the last session, there will be a panel of 4 speakers and a moderator to discuss the success, challenges, and low-hanging fruits of LLM applications, challenges, and solutions.
Objective
Our intended audience includes individual (e.g., clinicians, researchers, policymakers, and trainees) who are interested in practical applications of AI and LLM in healthcare and have working knowledge of LLMs. Researchers and clinicians in the clinical and healthcare community mainly focusing on traditional knowledge-driven approaches are welcome to join and converse with researchers from the data science and computer science fields. In terms of the level of expected backgrounds, 25% is basic, 50% is intermediate, and 25% is advanced.
At the conclusion of this workshop, a participant will be able to:
- Understand the promises and limitations of LLMs for healthcare applications.
- Summarize the state-of-the-art research and techniques to improve LLM quality and accuracy.
- Identify research and development opportunities in this space and potential collaborators for deeper engagement.
- Understand the clinical relevance of this work and how it can be capitalized to improve patient outcomes and biomedical research.
Event Speakers
Zhiyong Lu
Senior Investigator, NIH Intramural Research Program
Qiao Jin
Postdoctoral Research Fellow, Dr. Zhiyong Lu’s Biomedical Text Mining Group (NCBI)
Yifan Peng
Assistant Professor, Department of Population Health Sciences at Weill Cornell Medicine
Nansu Zong
Assistant Professor, Mayo Clinic's Department of AI and Informatics Research
Zhe He
Associate Professor, School of Information (FSU)
Rui Zhang
Professor and Chief, Division of Computational Health Sciences, Department of Surgery (UMN)
Aokun Chen
Assistant Scientist, University of Florida
Mattia Prosperi
Professor and Associate Dean, AI and Innovation at the University of Florida
Mor Peleg
Professor and Founding Director, Data Science Research Center at the University of Haifa
Event Schedule
The proposed workshop will be 3.5 hours, including a 30-minute coffee break for attendees to network with each other. All the speakers have confirmed participation.
Welcome and opening remarks
Introduction to LLM for healthcare applications [Slide deck]
Keynote 30 min presentation + 10 min Q&A
Enhancing Standard LLMs for Biomedicine: Techniques, Trends, and Best Practices [Slide deck]
Invited presentations I 12 min presentation + 3 min Q&A for each
Large Language Models in Clinical Evidence Summarization
Advancing Large Language Models in Biomedical Applications
Retrieval Augmented Generation for Enhancing LLMs in Answering Patients’ Questions on Lab Test Results
Coffee Break & Student Poster Presentations
Invited presentations II 12 min presentation + 3 min Q&A for each
Retrieval-augmented LLMs for Various Biomedical Tasks
Retrieval-Augmented Generation for Automating Patient Pre-screening in Clinical Trial Recruitment
Interactive Panel: Successes, Challenges, and Low-Hanging Fruits 50 min
Moderator: Zhe He, Ying Li
Wrap up
Event Organizers
Zhe He
Associate Professor, School of Information (FSU)
Ying Li
Director, Health Economics & Outcome Research at Regeneron Pharmaceuticals, Inc.
Rui Zhang
Professor and Chief, Division of Computational Health Sciences, Department of Surgery (UMN)
Nansu Zong
Assistant Professor, Mayo Clinic's Department of AI and Informatics Research
Accepted Posters
Event Venue
Hilton San Francisco Union Square
The Hilton San Francisco Union Square is located a block from the Curran and ACT theaters, and just two blocks from Union Square and Westfield San Francisco Centre. The Powell Street cable car turnaround, San Francisco Museum of Modern Art, and Moscone Convention Center are within a mile. Enjoy panoramic views from our skybar, the city’s highest. The Hilton San Francisco Union Square also has an outdoor pool and whirlpool, two restaurants, a lobby bar, and a grab-and-go market.
Event Details
Time/Date: 8:30 AM – 12 PM PST, November 9, 2024
Location: Imperial B - Hilton San Francisco Union Square
Session Code: W05