Springe direkt zu Inhalt

The Patient as a Multimodal Story: A Project Building Multimodal AI for Medicine

Benedikt Peterson, Charité - Universitätsmedizin Berlin, Center of Digital Health - BIH@Charité

Inhalte/ Contents

Patients generate multimodal clinical data (imaging, signals, reports, genomics), but jointly leveraging these sources to support clinical decisions and improve patient outcomes has long been limited by computing and architecture constraints. Recent advances in multimodal AI/LLMs and GPU computing make it feasible to build models that synthesize heterogeneous evidence. This research tutorial centers on Symile (https://arxiv.org/abs/2411.01053) and replacing a softmax-based loss with a sigmoid-based loss (Sigmile): we reproduce the baseline, then try to exploit the sigmoid formulation's parallelization to scale beyond three modalities.  Target group: advanced BSc/MSc students in CS or related fields. Prerequisites: strong Python and deep learning basics (PyTorch preferred); familiarity with representation/contrastive learning is a plus. Outcome: a publishable paper

Kontakt/ Contact

benedikt.peterson@fu-berlin.de

Link zum Vorlesungsverzeichnis/ Link to the course catalogue

folgt/ coming