Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/76964, first published .
Foundation Models for Generative AI in Time-Series Forecasting

Foundation Models for Generative AI in Time-Series Forecasting

Foundation Models for Generative AI in Time-Series Forecasting

Authors of this article:

Diva Beltramin1 Author Orcid Image ;   Cedric Bousquet2, 3 Author Orcid Image

Letter to the Editor

1Medical Information Department, Hospices Civils de Lyon, Lyon, France

2Laboratory of Medical Informatics and Knowledge Engineering in e-Health, Sorbonne University, Paris, France

3Public Health Service and Medical Information, Centre Hospitalier Universitaire de Saint-Étienne, Saint-Etienne, France

*all authors contributed equally

Corresponding Author:

Diva Beltramin, MSc, MD

Medical Information Department

Hospices Civils de Lyon

3 Quai des Célestins

Lyon, 69229

France

Phone: 33 0472357066

Email: diva.beltramin@chu-lyon.fr


J Med Internet Res 2025;27:e76964

doi:10.2196/76964

Keywords


We read with great interest the article entitled “Generative AI Models in Time-Varying Biomedical Data: Scoping Review” by He et al [1]. The authors performed a thorough and exhaustive review of generative artificial intelligence (AI) models used for the analysis of biomedical data varying over time. However, the foundation models (FMs) identified in this article do not seem to correspond to the proposed definition, which is “models capable of performing various generative tasks after being trained on extremely large and typically unlabeled datasets.” Additionally, the authors slightly modified Wornow et al’s [2] definition of FMs by adding the “generative” adjective to the word “tasks.” As a consequence, this definition should designate only FMs that support generative AI, but the authors mention some models that have no ability for generation.

Moreover, the authors propose two distinct FM lists in two different sections of the article. The first list features models for time-series forecasting, while the second includes large language models trained on electronic health records (referred to as clinical language models [CLaMs]) that take text as input and may produce text as output.

In the first list of FMs, the authors included generative adversarial networks (GANs), variational autoencoders, conditional variational autoencoders, omicsGAN, Potential Energy Underlying Single-Cell Gradients (PRESCIENT), gene-guided weakly supervised clustering via GANs, and trained GAN discriminator. However, these models are not FMs, as they were not trained on large amounts of data, and we are not aware of the possibility that they were pretrained to enable later fine-tuning on different downstream tasks.

In the second list, the authors erroneously mention several CLaMs as examples of FMs that support generative AI. Some of these models have been trained on tasks previously used to train bidirectional encoder representations from transformers, such as predicting a masked word based on previous and following words, and are known as masked language models. Indeed, masked language models may have text as input and output, as in tasks like summarization or question answering. However, these models are not generative AI models. For example, GatorTron [3], which is cited in the article, is a masked language model, but it has a generative AI version, GatorTronGPT [4].

Additionally, CLaMs are not naturally able to perform time-series forecasting and thus have to be repurposed or made capable of transforming the forecasting task into a language task to achieve the forecasting ability [5].

In conclusion, besides our concerns about FMs, we were favorably surprised by this article, which presents innovative applications of generative AI for those who are used to traditional biostatistics. Some applications of generative AI to time-varying biomedical data may be limited to specific niches where they are particularly efficient and not replaceable, and it is unclear if generative AI may outperform traditional time-series approaches. Nevertheless, it is advantageous to have the possibility to choose among several options when practicing data analysis, and this article shows the richness of the different methods available.

Conflicts of Interest

None declared.

  1. He R, Sarwal V, Qiu X, Zhuang Y, Zhang L, Liu Y, et al. Generative AI models in time-varying biomedical data: scoping review. J Med Internet Res. Mar 10, 2025;27:e59792. [FREE Full text] [CrossRef] [Medline]
  2. Wornow M, Xu Y, Thapa R, Patel B, Steinberg E, Fleming S, et al. The shaky foundations of large language models and foundation models for electronic health records. NPJ Digit Med. Jul 29, 2023;6(1):135. [CrossRef] [Medline]
  3. Yang X, Chen A, PourNejatian N, Shin HC, Smith KE, Parisien C, et al. A large language model for electronic health records. NPJ Digit Med. Dec 26, 2022;5(1):194. [CrossRef] [Medline]
  4. Peng C, Yang X, Chen A, Smith KE, PourNejatian N, Costa AB, et al. A study of generative large language model for medical research and healthcare. NPJ Digit Med. Nov 16, 2023;6(1):210. [CrossRef] [Medline]
  5. Peixeiro M. Understanding foundation models. In: Time Series Forecasting Using Foundation Models. Shelter Island, NY. Manning Publications Co; 2024.


AI: artificial intelligence
CLaM: clinical language model
FM: foundational model
GAN: generative adversarial network
PRESCIENT: Potential Energy Underlying Single-Cell Gradients


Edited by T Leung; This is a non–peer-reviewed article. submitted 05.05.25; accepted 11.07.25; published 25.07.25.

Copyright

©Diva Beltramin, Cedric Bousquet. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 25.07.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.