User Tools

Site Tools


healthgpt

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
healthgpt [2023/04/29 15:06]
nigam
healthgpt [2023/04/30 18:03] (current)
nigam
Line 1: Line 1:
 ====== Large language models in Healthcare ====== ====== Large language models in Healthcare ======
  
 +The past year has seen significant advancements in artificial intelligence (AI) for various modalities, such as text, image, and video. Foundation models, which are AI models trained on large, unlabeled datasets and highly adaptable to new applications, are driving these innovations. These new class of models offer opportunities for a better paradigm of doing "AI in healthcare" by providing adaptability with fewer manually labeled examples, modular and robust AI, multimodality, and new interfaces for human-AI collaboration. Read about [[https://hai.stanford.edu/news/how-foundation-models-can-advance-ai-healthcare|How Foundation Models Can Advance AI in Healthcare]]
  
-The past year has seen significant advancements in artificial intelligence (AIfor various modalitiessuch as textimageand video. Foundation modelswhich are AI models trained on largeunlabeled datasets and highly adaptable to new applicationsare driving these innovationsThese new class of models offer opportunities for a better paradigm of doing "AI in healthcare" by providing adaptability with fewer manually labeled examplesmodular and robust AI, multimodality, and new interfaces for human-AI collaboration. Read about [[https://hai.stanford.edu/news/how-foundation-models-can-advance-ai-healthcare | How Foundation Models Can Advance AI in Healthcare]]+Although foundation models (FMs), including large language models (LLMs)have immense potential in healthcareevaluating their usefulnessfairness, and reliability is challengingas they lack shared evaluation frameworks and datasetsOver 80 clinical FMs have been createdbut their evaluation regimes do not establish or validate their presumed clinical value. In additionuntil their factual correctness and robustness are ensured, it is difficult to justify the use of LLMs in clinical practice. Read about [[https://hai.stanford.edu/news/shaky-foundations-foundation-models-healthcare|The Shaky Foundations of Foundation Models in Healthcare ]] and see the arxiv preprint at [[https://arxiv.org/abs/2303.12961 | arxiv]].
  
-Although foundation models (FMs), including large language models (LLMs), have immense potential in healthcare, evaluating their usefulness, fairness, and reliability is challenging, as they lack shared evaluation frameworks and datasets. Over 80 clinical FMs have been reviewed, but their evaluation regimes do not indicate their clinical value accurately. Until their factual correctness and robustness are ensured, it is difficult to justify the use of FMs in clinical practice. Read about [[https://hai.stanford.edu/news/shaky-foundations-foundation-models-healthcare | The Shaky Foundations of Foundation Models in Healthcare ]]+We examined the safety and accuracy of GPT-4 in serving curbside consultation needs of doctors. Read about [[https://hai.stanford.edu/news/how-well-do-large-language-models-support-clinician-information-needs|How Well Do Large Language Models Support Clinician Information Needs?]] and check out the arxiv preprint at [[https://arxiv.org/abs/2304.13714|arxiv]]. We also evaluated the ability of GPT-4 to generate realistic USMLE Step 2 exam questions by asking licensed physicians to distinguish between AI-generated and human-generated questions and to assess their validity. The results indicate that GPT-4 can create questions that are largely indistinguishable from human-generated ones, with a majority of the questions deemed "valid". Read more at [[https://www.medrxiv.org/content/10.1101/2023.04.25.23288588v1 | medrxiv]].
  
-We examined the safety and accuracy of GPT-4 in serving curbside consultation needs of doctors. Read about +The video below summarizes the work described above as well as outlines the [[https://www.linkedin.com/feed/update/urn:li:activity:7031128969005432832/|questions we should always ask]] when considering LLMs for clinical use.
-[[https://hai.stanford.edu/news/how-well-do-large-language-models-support-clinician-information-needs|Wow Well Do Large Language Models Support Clinician Information Needs?]]+
  
 +<html> <iframe width="885" height="500" src="https://www.youtube.com/embed/l0GBGQqcHM8?start=23" title="#AimiGPT | Large Language Models 101" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" allowfullscreen></iframe> </html> //
  
-<html>  
-<iframe width="1280" height="720" src="https://www.youtube.com/embed/l0GBGQqcHM8?start=23" title="#AimiGPT | Large Language Models 101" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" allowfullscreen></iframe> 
-</html> 
  
healthgpt.1682805980.txt.gz · Last modified: 2023/04/29 15:06 by nigam