Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Grab it
Learn about Knowledge-Adapted Fine-Tuning (KnowAda), a groundbreaking approach to reduce hallucinations in vision-language models (VLMs) through this 26-minute educational video. Explore how smaller VLMs with up to 7B parameters struggle with balancing descriptive richness and factual accuracy in medical imaging applications. Discover insights from John Hopkins University researchers on probing VLM knowledge through generated visual questions and implementing automatic caption adaptation. Understand the Decomposed NLI (DNLI) evaluation framework that breaks down captions into atomic claims for precise accuracy assessment. Examine the limitations of current medical VLMs, the importance of pre-training in fine-tuning processes, and the anticipated developments in new pre-trained models. Gain practical knowledge about implementing KnowAda methodology, which has demonstrated superior performance in maintaining high descriptiveness while significantly reducing contradiction rates compared to traditional methods.
Read more
Knowledge-Adapted Fine-Tuning for Medical Vision Language Models