Multimodal large language models in human-centered health: practical insights

Abstract

Multimodal data offer a comprehensive view of human health by integrating diverse sources, such as text, medical images, physiological signals, and behavioral data. Recent advancements in large language models (LLMs) have led to the development of multimodal large language models (MLLMs), which leverage the text understanding capabilities of LLMs and integrate them with other modalities. While MLLMs show great promise for human-centered health applications, the practical challenges of implementing them in the healthcare sector remain largely unexplored. This article discusses these practical considerations and the future potential of MLLMs in transforming human-centered healthcare.</p

    Similar works

    Full text

    thumbnail-image

    Southampton (e-Prints Soton)

    redirect
    Last time updated on 31/07/2025

    This paper was published in Southampton (e-Prints Soton).

    Having an issue?

    Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.