Data-Centric Foundation Models in Computational Healthcare

Yunkun Zhang1, Jin Gao1, Zheling Tan1, Lingfeng Zhou1,
Kexin Ding2, Mu Zhou3, Shaoting Zhang4, Dequan Wang1,4
1Shanghai Jiao Tong University
2University of North Carolina at Charlotte
3Rutgers University
4Shanghai Artificial Intelligence Laboratory

Abstract

The advent of foundation models (FMs) as an emerging suite of AI techniques has struck a wave of opportunities in computational healthcare. The interactive nature of these models, guided by pre-training data and human instructions, has ignited a data-centric AI paradigm that emphasizes better data characterization, quality, and scale. In healthcare AI, obtaining and processing high-quality clinical data records has been a longstanding challenge, ranging from data quantity, annotation, patient privacy, and ethics. In this survey, we investigate a wide range of data-centric approaches in the FM era (from model pre-training to inference) towards improving the healthcare workflow. We discuss key perspectives in AI security, assessment, and alignment with human values. Finally, we offer a promising outlook of FM-based analytics to enhance the performance of patient outcome and clinical workflow in the evolving landscape of healthcare and medicine.

Overview

From a data-centric viewpoint, we emphasize the interplay between patients, healthcare data, and foundation models. Patients generate healthcare data and interact with foundation models. Healthcare data captures patient characteristics and supports foundation model training, inference, and deployment. Foundation models assess healthcare data and benefit patients. As illustrated, data-centric strategies promise to reshape clinical workflow , enable precise diagnosis, and uncover insights into treatment.

Overview

Why data-centric?

  • Foundation models demonstrate the power of scale, where the enlarged model and data size permit foundation models to capture vast amounts of information, thus increasing the pressing need of training data quantity.
  • Foundation models encourage homogenization as evidenced by their extensive adaptability to downstream tasks. High-quality data for foundation model training thus becomes critical since it can impact the performance of both pre-trained and downstream models.

In healthcare

Healthcare and medical data challenges have posed persistent obstacles over decades, including multi-modality data fusion, limited data volume, annotation burden, and the critical concern of patient privacy protection. To respond, the foundation model era opens up perspectives to advance data-focused AI analytics in healthcare.

Foundation Models in Healthcare

Foundation model workflow:
  • Large scale pre-training is an essential approach to building a foundation model from scratch.
  • Downstream generalization, including fine-tuning and in-context learning, is a critical technique in constructing medical domain-specific foundation models.
FM

We provide an up-to-date list of healthcare foundation models in our survey paper (Appendix B) and GitHub page.

Multi-Modal Data Fusion

Foundation models can enable a more scalable, generalizable, and comprehensive multi-modal healthcare data fusion. Conventional fusion approaches are enhanced by joint-modal pre-training and comprehensive foundation models such as LLMs, enabling downstream applications such as medical QA, drug discovery, and diagnosis.


Multimodal

Data fusion via multi-modal pre-training

Foundation models can handle multiple modalities via pre-training on massive-scale paired multi-modal data in a joint-modal mode to obtain a high-level understanding of inter-modality relationships.

Data fusion via LLMs

Transformer-style LLMs possess powerful semantic understanding capability via the attention mechanism, which can be transferred to multi-modal settings. Data from different modalities can be aggregated as the prompt input of an LLM.

Data Quantity and Annotation

Foundation models address data quantity and data annotation challenges.
  • Foundation models can mitigate data quantity limitation by data augmentation and improved data efficiency.
  • Foundation models can help both healthcare text and medical image annotation.
Quantity and Annotation

Data Privacy

Healthcare data privacy protection has always been an important issue. Foundation models offer solutions to this problem while also bringing new challenges.
  • Solution: Foundation models can now produce high-quality synthetic medical data with similar characteristics to the original data but without identical information.
  • Challenge: Foundation models memorize their training data and tend to output these data, which may lead to privacy leakage.

Performance Evaluation

Foundation model evaluation is challenging owing to the models' extensive utilization given their own model scale and complexity. Three common evaluation strategies include benchmarking, human evaluation, and automated evaluation.

Evaluation

We provide an up-to-date list of healthcare benchmarks for foundation model evaluation in our survey paper (Table 1) and GitHub page.

In this survey, we have offered an overview of FM challenges from a data-centric perspective. FMs possess great potential to mitigate data challenges in healthcare, including data imbalance and bias, data scarcity, and high annotation costs. Due to FM's strong content generation capabilities, there is a remarkable need for greater vigilance regarding data privacy, data bias, and ethical considerations about the generated medical knowledge. Only by adequately and reliably addressing the data-centric challenges can we better leverage the power of FMs across a broader scope of medicine and healthcare.

BibTeX

If you find this project helps, please kindly cite our survey, thanks!

@article{zhang2024data,
  title={Data-Centric Foundation Models in Computational Healthcare: A Survey},
  author={Zhang, Yunkun and Gao, Jin and Tan, Zheling and Zhou, Lingfeng and Ding, Kexin and Zhou, Mu and Zhang, Shaoting and Wang, Dequan},
  journal={arXiv preprint arXiv:2401.02458},
  year={2024}
}