Your new experience awaits. Try the new design now and help us make it even better

PERSPECTIVE article

Front. Public Health, 24 September 2025

Sec. Digital Public Health

Volume 13 - 2025 | https://doi.org/10.3389/fpubh.2025.1658645

This article is part of the Research TopicAdvances in Mathematical Modelling for Infectious Disease Control and PreventionView all 3 articles

Keeping a modeling-driven public health dashboard relevant—lessons learned from the California Communicable diseases Assessment Tool

Mugdha ThakurMugdha ThakurLauren A. WhiteLauren A. WhiteJohn PuglieseJohn PuglieseDavid CrowDavid CrowPhoebe LuPhoebe LuNatalie LintonNatalie LintonRyan McCorvieRyan McCorvieSindhu RavuriSindhu RavuriHector M. Snchez C.Hector M. Sánchez C.Brent SiegelBrent SiegelJason VargoJason VargoToms M. Len
Tomás M. León*
  • California Department of Public Health, Richmond, CA, United States

Researchers rapidly developed modeling-based dashboards to support the global COVID-19 pandemic response, and this output has continued for other public health responses. These dashboards are often abandoned or deprecated over time, creating challenges for public health jurisdictions that would like to leverage them for decision-making. Maintaining a relevant and sustainable dashboard requires significant effort and attention to collating modeling results and meeting local public health needs. The California Communicable diseases Assessment Tool (CalCAT), a public-facing infectious disease modeling dashboard, demonstrates the key components to sustainability and relevance: robust and flexible workflows, cultivation of trust and user engagement, wide-ranging collaboration, and content reproducibility. The experience of CalCAT’s initiation and development highlights the need to be responsive to ever-changing stakeholder requests and to balance trade-offs in design choices for how modeling results are processed, presented, and shared.

Background

As part of the COVID-19 pandemic response, many jurisdictions deployed dashboards displaying a variety of metrics including cases, hospitalizations, and vaccine uptake (1). Building upon these publicly available data sources and open repositories (2, 3), researchers rapidly developed mathematical models to estimate and forecast COVID-19 burden. These models varied by data source, time scale, geographic resolution, and target (e.g., cases, hospitalizations, deaths), requiring policymakers to contextualize and integrate predictions from multiple models. Moreover, different United States modeling hubs were managed by separate collaborative teams and while some coordination across hubs existed, there was not always agreement in timing, targets, or outputs of these efforts (46). Additionally, sub-state resolution was rarely available, partially due to suboptimal model performance at the county level and the resources needed for modeling over 3,000 U.S. counties (7, 8). Extracting local, jurisdiction-specific modeling predictions relevant for policy and decision-making remained challenging and time-consuming.

Having sub-state modeling predictions provides critical information because each county or region may exhibit specific epidemiological patterns not reflected in overall state trends. For example, California is the most populous state in the United States (~40 million people) with significant variation in population density and demographics. Motivated by these potential localized differences, the California COVID Assessment Tool (CalCAT; calcat.cdph.ca.gov) was launched in June 2020 by the California Department of Public Health (CDPH) Modeling Team along with internal and external collaborators to inform state and local public health COVID-19 response efforts. CalCAT brings together all publicly available California-specific modeling outputs, ensembles them, and interprets results for end users. Modeling products displayed on CalCAT include effective reproduction number (“R-effective”) nowcasts, forecasts, and scenario models for California and its constituent counties and regions. The CalCAT target audience includes California residents, public health staff and leadership, and researchers. CalCAT has since been renamed to the California Communicable diseases Assessment Tool to reflect broader disease coverage including influenza, added in fall of 2022, and combined respiratory burden (COVID-19 and influenza), added in winter of 2023.

CalCAT has had a direct impact on public health decision-making. As one of the very few dashboards to include county-and region-level model outcomes, CalCAT modeling for intensive care unit (ICU) capacity projections was used to support the regional stay-at-home order in California in winter 2020–2021. Synthesized model output from CalCAT was used in weekly situational updates to California state government leadership throughout the COVID-19 pandemic response. CalCAT results provided leading indicator information (e.g., R-effective) and burden projections that informed public health decision making. For example, CalCAT informed and was complementary to the Blueprint for a Safer Economy in 2020–2021, which identified risk tiers for counties that governed non-pharmaceutical interventions such as business capacity restrictions. In 2020, CalCAT had over 200,000 unique users across California and around the world.

While many data dashboards were effectively deployed throughout the COVID-19 pandemic (1), CalCAT is unique in that it: (1) synthesizes modeling results at the state and sub-state level; (2) combines modeling results for multiple respiratory pathogens (COVID-19 and influenza); and (3) integrates results across multiple analytic products (i.e., nowcasts, short-term forecasts, and long-term scenario models). While many of the national forecasting or scenario modeling hubs address one or two of these functions, they do not address all three (46). To our knowledge, CalCAT is the only dashboard deployed by a public health department during the pandemic that integrated forecasting and analytics with real-time data updates. CalCAT is also unique for its longevity; sustaining necessary data feeds and codebases is challenging and resource-intensive, and their neglect has rendered many data dashboards obsolete (7). This Viewpoint describes and highlights necessary steps and collaborations for developing and sustaining an infectious disease modeling dashboard that effectively supports public health decision-making. The following sections summarize the: (a) foundational components (i.e., content and collaborations) of CalCAT, (b) pillars for keeping CalCAT relevant, (c) ongoing challenges and lessons learned, and (d) overall conclusions.

Foundations: content and collaborations

Content: data input, pipelines, and output

Models on CalCAT use public and restricted access datasets including, but not limited to, cases, testing, wastewater surveillance, hospitalizations, and deaths. Many of these state and county level public datasets are available on the California Open Data Portal,1 which include data dictionaries and relevant context around data collection. Restricted datasets include linelist-level public health surveillance data reported to CDPH, which are not shared publicly for data privacy and confidentiality reasons. For example, state-level hospital admissions are publicly available from the National Hospital Safety Network dataset and are utilized as a target by U.S. forecasting and scenario modeling hubs (7, 9, 10). To CDPH, this dataset is available at the facility-level, and these counts are aggregated at the county and regional scales for use in modeling pipelines.

Depending on data availability and relevance to stakeholders, CalCAT’s content at any given time may include R-effective nowcast estimates; short-term (one-to four-week) forecasts of hospital admissions, hospital census, ICU census, cases, and deaths; and long-term scenario models based on “what-if” assumptions regarding future factors such as vaccination uptake and emergence of disease pathogen variants. With these metrics shown in the same place, CalCAT helps contextualize the different types of modeling evidence by bringing them together, providing extensive technical notes, and when appropriate including summary metrics and interpretation of model outputs (e.g., determining whether COVID-19 transmission is increasing, decreasing, or stable based on the summary of all available R-effective models).

Through an automated pipeline, ensemble values for metrics of public health relevance are calculated from all available models. The relevant reported data, model results and ensembles are displayed on the ‘shiny’ Posit package-based CalCAT dashboard (11). CalCAT’s outputs are downloadable directly from the CalCAT interface or from the archived version of model trajectories on the California Open Data Portal (12).

Collaborations: relationships with model contributors

Modeling-based insights for public health emergencies are important and modeling capacity within public health agencies can be limited (13). The Team works with internal and external collaborators to provide county and region level metrics that may not otherwise be available; these partnerships have been central to CalCAT’s continued success. The models displayed on CalCAT come from external partners and contributors, coordinated hub efforts, and internally produced models maintained by the Team. As part of the COVID-19 response, the Team also worked closely with other internal CDPH teams to produce timely models that accounted for changes in testing practices, vaccination coverage, variants, and other time-varying aspects of the pandemic. For example, collaboration with influenza subject matter experts at CDPH guided influenza model development and maintenance across multiple seasons.

External partnerships have deepened over time through both ongoing informal and official collaboration mechanisms such as hub initiatives, formal grant funding, and the UC Health-CDPH COVID Modeling Consortium2 (14). This Consortium was a major effort by the University of California (UC) system and CDPH to improve evidence-based public health decision-making during the COVID-19 public health emergency and thereafter (15). Regular seminars and smaller group meetings with UC partners were helpful for sharing information about emerging trends that would affect model design (e.g., variants of concern, waning immunity). Such collaborations were particularly helpful for county-and regional-scale modeling because these targets were underrepresented or nonexistent in national modeling hubs.

The Team maintains open lines of communication with various forecasting and scenario efforts including CDC FluSight, Scenario Modeling Hubs, CDC Center for Forecasting and Outbreak Analytics (CFA), CDC InsightNet centers, and other modeling networks. This engagement has involved contributing California-specific modeling results, weighing in on submission requirements or scenario specifications and how these could be improved for public health impact, and sharing data and resources to support collaborative modeling efforts (9).

Staying relevant

Building trust and two-way relationships with users have been key for keeping CalCAT relevant. Trust has been built through bidirectional communication with end users; ongoing quality control and validation of results displayed on CalCAT; striving for principles of openness and reproducibility in CalCAT development; and maintaining flexibility and relevance in the face of changing modeling inputs.

Bidirectional communication with users

CalCAT is tailored to its users by engaging in an ongoing, bidirectional exchange of information with local health jurisdictions (LHJs) and researchers. Communication is encouraged through a dedicated email inbox linked on the dashboard. The Team also meets with LHJs during a biweekly “CalCAT Open House” meeting to discuss models, review other analytical products that may not be publicly available, exchange situational information, and receive feedback in real-time. Attendance has remained robust even after the end of the COVID-19 public health emergency, with more than one-third of California’s 61 LHJs regularly represented and more than one-half occasionally attending. These check-ins feature presentations on cutting-edge developments in infectious disease modeling methods, serving as a source of continuing professional development for participants. These meetings engage individual users and teams and provide an opportunity to conduct biannual surveys to inform future development goals. Some examples of regular survey questions are: “how often do you use the different CalCAT products (nowcasts, forecasts, scenarios) at the various geographical scales (state, region, county)?,” “which additional diseases are you interested in seeing on CalCAT?,” and “in what other areas would your jurisdiction benefit from modeling support?” Results from these dialogues have driven CalCAT modifications including expansion to new pathogens, aesthetic and accessibility choices for visual displays, and prioritization of model targets (e.g., hospital census vs. admissions). Infectious disease modeling results are additionally shared via academic literature (1518) and in research collaborations, where they undergo scientific review but not user acceptability testing. Public health agencies are increasingly requesting user feedback for their digital products to ensure that messages are being received as intended and that websites are user-friendly and follow usability principles (1921).

Data processing, quality control, and validation

A daily, automated process checks for updated modeling results; these results are imported through a pipeline that includes scraping of public websites, downloading from GitHub and other repositories, and more secure data transfer such as through cloud object storage. Preliminary model results are then processed to generate ensembles and calculate uncertainty, and output is pushed to a development version of CalCAT. After review, the final production version of CalCAT is updated for public viewing. This multi-step process limits the publication of incorrect data and poorly calibrated models on CalCAT.

The Team conducts ongoing peer review and quality control of modeling results, as well as retrospective evaluation of model performance (17). The Team also receives inquiries from local public health officials, the media, and the public. Most of these communications cover data or modeling anomalies spotted on CalCAT and serve as an additional safeguard for CalCAT quality assurance and control. Other inquiries request more information or interpretation, which may be handled by the Modeling Team or redirected to the CDPH Communications Office.

Adaptability and responsiveness

Several key structural features enable CalCAT to remain relevant, respond rapidly to changing input data streams and modeling outputs, and undergo routine updates for situational awareness. To maximize the use of all available model outputs for California, CalCAT is set up to process all file types and data types with no standardized template. Since many CalCAT inputs are generated independently by external collaborators, there is high variation in the cadence of updates and in the geographical coverage of available models. Moreover, the set of contributing models has varied substantially since CalCAT’s founding in 2020; different modelers stopped generating new model results or updated models in response to evolving disease and public health data landscapes. Backend data pipelines and workflows for CalCAT are designed specifically to be robust to these variations so that the most up-to-date information is available on the website.

Sustainability and reproducibility

The Team aims for the contents of CalCAT to be FAIR—Findable, Accessible, Interoperable, Reusable (22). Extensive technical notes describe all constituent models with links to relevant underlying data, methodologies and code repositories. A notable success has been the willingness of the vast majority of external model contributors to publish their outputs publicly. Furthermore, the source code for a minimal reproducible version of CalCAT is maintained on GitHub for use by other teams and jurisdictions (23). This archiving is helpful for data and model versioning, and to support retrospective analyses (17). The biggest challenge in maintaining an archive of CalCAT products has been variability in formatting of the input and output datasets because of ongoing changes in data availability and stakeholder priorities. Anticipating and adapting to this variability in the data and modeling pipelines was an important lesson that emerged from CalCAT’s evolution.

Adherence to FAIR principles entails a long-term commitment to building trust between the maintainers and stakeholders and a shared vision of evidence-based preparedness and response activities for outbreaks. Sustainability of this effort relies heavily on ongoing prioritization by CDPH, and availability of personnel and funding. The early days of the Team relied on redirected, deployed, and volunteer staff. Initial funding came from the COVID-19 emergency response and was transitioned, when possible, to more permanent sources of public health funding as the team evolved over the following years. The size of the team has fluctuated from four to ten staff members, and over time more of them had formal training and a background in infectious disease modeling. The responsibilities of the team require both extensive technical knowledge and familiarity with the functions and operations of government public health within the capacity of available information technology (IT) resources.

Challenges and lessons learned

The Team has learned valuable lessons throughout CalCAT’s development. First and foremost, flexibility to accommodate stakeholder requests for changing inputs over time is essential. For example, COVID-19 ICU capacity was initially more important as a forecasting target than inpatient capacity, but that reversed after treatment protocols improved the standard of care and ICU burden decreased. Similarly, case forecasts were deprioritized as testing practices changed and model performance declined. User surveys conducted with LHJs have shown that stakeholder priorities and data visualization preferences are not uniform, highlighting the importance of balancing limited staffing resources with stakeholder priorities.

The Team also learned that there is a tradeoff between complexity and clarity. This lesson was especially relevant when ensembling models to come up with a single best estimate for a given indicator. The team tried sophisticated model boosting techniques (machine learning, optimal weighting, etc.) for combining constituent models, but the median ensemble outperformed these more advanced methods in most cases; this is a consistent finding in the infectious disease modeling field (24, 25) and is standard practice for other forecasting hubs (7, 9, 26). Additionally, the median ensemble method is robust to the non-standardized reporting format of models submitted to CalCAT (e.g., including different quantiles for forecasts), which helps to ensure the consistency of the outputs. Parsimony has the added virtue of being easier to explain to non-technical audiences. Public health leadership has consistently preferred the clear messaging of ensemble models for sharing results with stakeholders, compared with showing multiple individual models or alternating between individual models.

Highlighting the uncertainty inherent in disease modeling and level-setting expectations was important for communicating with the general public. A September 2020 Sacramento Bee article addressed technical and communication challenges for projecting COVID-19 deaths in 2020 in smaller counties (27). Disease modeling—especially projecting trends out more than a few weeks, anticipating inflection points, estimating rates in less populous areas, accounting for uneven uptake of prophylactic interventions—is very difficult. Communicating this uncertainty visually (on the CalCAT dashboard) and verbally (in write-ups of CalCAT results, press releases, interviews, etc.) helps the public understand and be appropriately cautious when using CalCAT estimates.

One of CalCAT’s strengths is that it serves as a repository for all available models rather than relying on formatted submissions. However, this comes with challenges as staff are required to transform data, identify changing formats, and make decisions on whether to modify models to make them more comparable. Ensemble model interpretation can also be challenging when combining multiple models with heterogeneous methodologies. Additionally, a contributing model may suddenly stop updating, or an automated process may continue updating a model even if the results are clearly erroneous. Addressing these issues requires effort to monitor model outputs and contact model maintainers to assess these situations.

Alignment with other CDPH dashboards and reports is essential, so that end users do not get conflicting messages from CDPH about data trends. CalCAT relies on surveillance data but is not intended to be a primary means of reporting public health trends. However, it is often necessary to include observed surveillance data points to contextualize forecasting and scenario modeling results and to aid users in making their own interpretations about model calibration. The Modeling Team works with CDPH and LHJ collaborators to harmonize reporting across sources.

Conclusion

While other dashboards sunsetted and disappeared, CalCAT has remained an active tool for infectious disease modeling and public health decision-making in California over the last 5 years, maintaining >99·9% uptime even as it has evolved. CalCAT is not as heavily used as it was during the early years of the pandemic, but it remains an important public-facing resource for situational awareness, analysis, and research, and its content can be easily pivoted in the event of a future public health emergency. CalCAT’s success and longevity are because it is more than just a dashboard; it is a commitment by CDPH to the provisioning and use of infectious disease models in California for public health practice and to the iterative process of constant improvement of said models and how they are shared and visualized. It is also a community of practice and collaboration between internal and external stakeholders committed to responding to the needs of local and state public health practitioners. CalCAT development priorities will continue to be driven by stakeholder and end user input. Five years into CalCAT’s existence, the dashboard has built on many lessons learned and will continue to improve to better serve the people of California through providing infectious disease modeling to support evidence-based public health practice.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

MT: Investigation, Writing – review & editing, Conceptualization, Software, Writing – original draft. LW: Software, Writing – review & editing, Writing – original draft, Investigation. JP: Writing – review & editing, Investigation, Software. DC: Writing – review & editing, Investigation, Software. PL: Software, Investigation, Writing – review & editing. NL: Investigation, Writing – review & editing. RM: Software, Investigation, Writing – review & editing. SR: Writing – review & editing, Investigation. HMSC: Investigation, Writing – review & editing. BS: Writing – review & editing, Investigation. JV: Investigation, Writing – review & editing, Software. TL: Conceptualization, Supervision, Writing – review & editing, Writing – original draft.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was funded in part by the Centers for Disease Control and Prevention, Epidemiology and Laboratory Capacity for Infectious Diseases, Cooperative Agreement Number 6 NU50CK000539.

Acknowledgments

This work was supported by the California Department of Public Health.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Author disclaimer

The findings and conclusions in this article are those of the author(s) and do not necessarily represent the views or opinions of the California Department of Public Health or the California Health and Human Services Agency.

Footnotes

References

1. Dixon, BE, Dearth, S, Duszynski, TJ, and Grannis, SJ. Dashboards are trendy, visible components of data management in public health: sustaining their use after the pandemic requires a broader view. Am J Public Health. (2022) 112:900–3. doi: 10.2105/AJPH.2022.306849

PubMed Abstract | Crossref Full Text | Google Scholar

2. Alamo, T, Reina, DG, Mammarella, M, and Abella, A. COVID-19: open-data resources for monitoring, modeling, and forecasting the epidemic. Electronics. (2020) 9:827. doi: 10.3390/electronics9050827

Crossref Full Text | Google Scholar

3. Shuja, J, Alanazi, E, Alasmary, W, and Alashaikh, A. COVID-19 open source data sets: a comprehensive survey. Appl Intell. (2021) 51:1296–325. doi: 10.1007/s10489-020-01862-6

PubMed Abstract | Crossref Full Text | Google Scholar

4. U.S. Centers for Disease Control and Prevention. About flu forecasting. FluSight. 2024. Available online at: https://www.cdc.gov/flu-forecasting/about/index.html (accessed March 6, 2025).

Google Scholar

5. COVID-19 Forecast Hub. Available online at: https://covid19forecasthub.org/ (accessed March 6, 2025).

Google Scholar

6. Scenario Modeling Hub. Available online at: https://scenariomodelinghub.org/ (accessed June 10, 2025).

Google Scholar

7. Lopez, VK, Cramer, EY, Pagano, R, Drake, JM, O’Dea, EB, Adee, M, et al. Challenges of COVID-19 case forecasting in the US, 2020–2021. PLoS Comput Biol. (2024) 20:e1011200. doi: 10.1371/journal.pcbi.1011200

PubMed Abstract | Crossref Full Text | Google Scholar

8. Yanovitzky, I, Stahlman, G, Quow, J, Ackerman, M, Perry, Y, and Kim, M. National public health dashboards: protocol for a scoping review. JMIR Res Protoc. (2024) 13:e52843. doi: 10.2196/52843

PubMed Abstract | Crossref Full Text | Google Scholar

9. Mathis, SM, Webber, AE, León, TM, Murray, EL, Sun, M, White, LA, et al. Evaluation of FluSight influenza forecasting in the 2021–22 and 2022–23 seasons with a new target laboratory-confirmed influenza hospitalizations. Nat Commun. (2024) 15:6289. doi: 10.1038/s41467-024-50601-9

PubMed Abstract | Crossref Full Text | Google Scholar

10. HHS Office of the Chief Data Officer. (2025). Weekly United States hospitalization metrics by jurisdiction, during mandatory reporting period from August 1, 2020 to April 30, 2024, and for data reported voluntarily beginning May 1, 2024, National Healthcare Safety Network (NHSN)—archived. Available online at: https://healthdata.gov/dataset/Weekly-United-States-Hospitalization-Metrics-by-Ju/kruc-9unj/about_data (accessed February 23, 2025).

Google Scholar

11. Chang, W, Cheng, J, Allaire, J, Sievert, C, Schloerke, B, Buieet, G, et al. Shiny: web application framework for R. (2024). Available at: https://shiny.posit.co/, https://github.com/rstudio/shiny

Google Scholar

12. California Department of Public Health. CDPH-CalCAT modeling nowcasts and forecasts for COVID-19 and influenza. 2025. Available online at: https://data.ca.gov/dataset/cdph-calcat-modeling-nowcasts-and-forecasts-for-covid-19-and-influenza (accessed June 10, 2025).

Google Scholar

13. Pung, R, and Kucharski, AJ. Building in-house capabilities in health agencies and outsourcing to academia or industry: considerations for effective infectious disease modelling. Epidemics. (2024) 49:100802. doi: 10.1016/j.epidem.2024.100802

PubMed Abstract | Crossref Full Text | Google Scholar

14. CPR3 News. UC-CDPH collaborative awards more than $6 million for pandemic resiliency research. CPR3 (2023). Available at: https://cpr3.ucsf.edu/news/uc-cdph-collaborative-awards-more-6-million-pandemic-resiliency-research%C2%A0 (Accessed June 10, 2025).

Google Scholar

15. León, TM, White, LA, Spindler, H, Schwab, J, Petersen, ML, Vargo, J, et al. Fostering public health and academic partnerships during and beyond a public health emergency: lessons learned from COVID-19. Am J Epidemiol. (2025) 194:1482–4. doi: 10.1093/aje/kwaf007

PubMed Abstract | Crossref Full Text | Google Scholar

16. Ravuri, S, Burnor, E, Routledge, I, Linton, NM, Thakur, M, Boehm, A, et al. Estimating effective reproduction numbers using wastewater data from multiple sewersheds for SARS-CoV-2 in California counties. Epidemics. (2025) 50:100803. doi: 10.1016/j.epidem.2024.100803

PubMed Abstract | Crossref Full Text | Google Scholar

17. White, LA, McCorvie, R, Crow, D, Jain, S, and León, TM. Assessing the accuracy of California county level COVID-19 hospitalization forecasts to inform public policy decision making. BMC Public Health. (2023) 23:782. doi: 10.1186/s12889-023-15649-0

PubMed Abstract | Crossref Full Text | Google Scholar

18. Linton, NM, Quint, JJ, Chen, J, Hoover, C, Jain, S, and León, TM. Pediatric SARS-CoV-2 incidence in California by vaccination status, December 2021 to April 2023. Pediatr Open Sci. (2025) 1:1–11. doi: 10.1542/pedsos.2024-000264

Crossref Full Text | Google Scholar

19. Malkani, D, Malkani, M, Singh, N, and Madan, E. Best practices for the design of COVID-19 dashboards. Perspect Health Inf Manag. (2023) 20:1b. doi: 10.1093/eurpub/ckab164.488

PubMed Abstract | Crossref Full Text | Google Scholar

20. Ansari, B. Evaluating the usability of public health data dashboards as information sources for professionals and the public: findings from a case study with domain experts. Health Inf Libr J. (2024):1–14. doi: 10.1111/hir.12532

Crossref Full Text | Google Scholar

21. Malkani, M, Madan, E, Malkani, D, Madan, A, Singh, N, Bamji, T, et al. Rank ordered design attributes for health care dashboards including artificial intelligence: usability study. Online J Public Health Inform. (2024) 16:e58277. doi: 10.2196/58277

PubMed Abstract | Crossref Full Text | Google Scholar

22. Wilkinson, MD, Dumontier, M, Aalbersberg, IJ, Appleton, G, Axton, M, Baak, A, et al. The FAIR guiding principles for scientific data management and stewardship. Sci Data. (2016) 3:160018. doi: 10.1038/sdata.2016.18

PubMed Abstract | Crossref Full Text | Google Scholar

23. CDPH Modeling Team. California communicable diseases assessment tool (2025). Available online at: https://github.com/StateOfCalifornia/CalCAT (accessed June 10, 2025).

Google Scholar

24. Fox, SJ, Kim, M, Meyers, LA, Reich, NG, and Ray, EL. Optimizing disease outbreak forecast ensembles. Emerg Infect Dis. (2024) 30:1967–9. doi: 10.3201/eid3009.240026

PubMed Abstract | Crossref Full Text | Google Scholar

25. Ray, EL, Brooks, LC, Bien, J, Biggerstaff, M, Bosse, NI, Bracher, J, et al. Comparing trained and untrained probabilistic ensemble forecasts of COVID-19 cases and deaths in the United States. Int J Forecast. (2023) 39:1366–83. doi: 10.1016/j.ijforecast.2022.06.005

PubMed Abstract | Crossref Full Text | Google Scholar

26. The Consortium of Infectious Disease Modeling Hubs. The hubverse: open tools for collaborative modeling (2025). Available online at: https://github.com/hubverse-org (accessed January 17, 2025)

Google Scholar

27. Pohl, J, and Kreiser, M. California’s expensive COVID-19 predictions were useless for rural areas. Here’s why. Sacram Bee (2020). Available online at: https://www.sacbee.com/news/coronavirus/article245439570.html (accessed Feb 5, 2025).

Google Scholar

Keywords: infectious disease modeling, digital health, CalCAT, ensembling, California (USA), dashboard

Citation: Thakur M, White LA, Pugliese J, Crow D, Lu P, Linton N, McCorvie R, Ravuri S, Sánchez C. HM, Siegel B, Vargo J and León TM (2025) Keeping a modeling-driven public health dashboard relevant—lessons learned from the California Communicable diseases Assessment Tool. Front. Public Health. 13:1658645. doi: 10.3389/fpubh.2025.1658645

Received: 02 July 2025; Accepted: 28 August 2025;
Published: 24 September 2025.

Edited by:

Fahad Al Basir, Asansol Girls’ College, India

Reviewed by:

Hye Ryeon Jang, Morehouse College, United States

Copyright © 2025 Thakur, White, Pugliese, Crow, Lu, Linton, McCorvie, Ravuri, Sánchez C., Siegel, Vargo and León. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tomás M. León, dG9tYXMubGVvbkBjZHBoLmNhLmdvdg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.