<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="EN"><?covid-19-tdm?>
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2023.1203874</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Digital Health</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Comparison of pretrained transformer-based models for influenza and COVID-19 detection using social media text data in Saskatchewan, Canada</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Tian</surname><given-names>Yuan</given-names></name><uri xlink:href="https://loop.frontiersin.org/people/2273416/overview"/></contrib>
<contrib contrib-type="author"><name><surname>Zhang</surname><given-names>Wenjing</given-names></name></contrib>
<contrib contrib-type="author"><name><surname>Duan</surname><given-names>Lujie</given-names></name></contrib>
<contrib contrib-type="author"><name><surname>McDonald</surname><given-names>Wade</given-names></name></contrib>
<contrib contrib-type="author" corresp="yes"><name><surname>Osgood</surname><given-names>Nathaniel</given-names></name><uri xlink:href="https://loop.frontiersin.org/people/199914/network"/>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref></contrib>
</contrib-group>
<aff><addr-line>Department of Computer Science</addr-line>, <institution>University of Saskatchewan</institution>, <addr-line>Saskatoon, SK</addr-line>, <country>Canada</country></aff>
<author-notes>
<fn fn-type="edited-by"><p><bold>Edited by:</bold> Alex Mariakakis, University of Toronto, Canada</p></fn>
<fn fn-type="edited-by"><p><bold>Reviewed by:</bold> Ingrid Vasiliu Feltes, University of Miami, United States, Chiranjibi Sitaula, The University of Melbourne, Australia</p></fn>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Nathaniel Osgood <email>osgood@cs.usask.ca</email></corresp>
</author-notes>
<pub-date pub-type="epub"><day>28</day><month>06</month><year>2023</year></pub-date>
<pub-date pub-type="collection"><year>2023</year></pub-date>
<volume>5</volume><elocation-id>1203874</elocation-id>
<history>
<date date-type="received"><day>11</day><month>04</month><year>2023</year></date>
<date date-type="accepted"><day>02</day><month>06</month><year>2023</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2023 Tian, Zhang, Duan, McDonald and Osgood.</copyright-statement>
<copyright-year>2023</copyright-year><copyright-holder>Tian, Zhang, Duan, McDonald and Osgood</copyright-holder><license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<sec><title>Background</title>
<p>The use of social media data provides an opportunity to complement traditional influenza and COVID-19 surveillance methods for the detection and control of outbreaks and informing public health interventions.</p>
</sec>
<sec><title>Objective</title>
<p>The first aim of this study is to investigate the degree to which Twitter users disclose health experiences related to influenza and COVID-19 that could be indicative of recent plausible influenza cases or symptomatic COVID-19 infections. Second, we seek to use the Twitter datasets to train and evaluate the classification performance of Bidirectional Encoder Representations from Transformers (BERT) and variant language models in the context of influenza and COVID-19 infection detection.</p>
</sec>
<sec><title>Methods</title>
<p>We constructed two Twitter datasets using a keyword-based filtering approach on English-language tweets collected from December 2016 to December 2022 in Saskatchewan, Canada. The influenza-related dataset comprised tweets filtered with influenza-related keywords from December 13, 2016, to March 17, 2018, while the COVID-19 dataset comprised tweets filtered with COVID-19 symptom-related keywords from January 1, 2020, to June 22, 2021. The Twitter datasets were cleaned, and each tweet was annotated by at least two annotators as to whether it suggested recent plausible influenza cases or symptomatic COVID-19 cases. We then assessed the classification performance of pre-trained transformer-based language models, including BERT-base, BERT-large, RoBERTa-base, RoBERT-large, BERTweet-base, BERTweet-covid-base, BERTweet-large, and COVID-Twitter-BERT (CT-BERT) models, on each dataset. To address the notable class imbalance, we experimented with both oversampling and undersampling methods.</p>
</sec>
<sec><title>Results</title>
<p>The influenza dataset had 1129 out of 6444 (17.5&#x0025;) tweets annotated as suggesting recent plausible influenza cases. The COVID-19 dataset had 924 out of 11939 (7.7&#x0025;) tweets annotated as inferring recent plausible COVID-19 cases. When compared against other language models on the COVID-19 dataset, CT-BERT performed the best, supporting the highest scores for recall (94.8&#x0025;), F1(94.4&#x0025;), and accuracy (94.6&#x0025;). For the influenza dataset, BERTweet models exhibited better performance. Our results also showed that applying data balancing techniques such as oversampling or undersampling method did not lead to improved model performance.</p>
</sec>
<sec><title>Conclusions</title>
<p>Utilizing domain-specific language models for monitoring users&#x2019; health experiences related to influenza and COVID-19 on social media shows improved classification performance and has the potential to supplement real-time disease surveillance.</p>
</sec>
</abstract>
<kwd-group>
<kwd>influenza</kwd>
<kwd>COVID-19</kwd>
<kwd>social media</kwd>
<kwd>transformer-based language models</kwd>
<kwd>digital surveillance</kwd>
</kwd-group>
<contract-num rid="cn001">2017-04647</contract-num>
<contract-sponsor id="cn001">NSERC<named-content content-type="fundref-id">10.13039/501100000038</named-content></contract-sponsor>
<counts>
<fig-count count="4"/>
<table-count count="6"/><equation-count count="90"/><ref-count count="46"/><page-count count="0"/><word-count count="0"/></counts><custom-meta-wrap><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Health Informatics</meta-value></custom-meta></custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro"><label>1.</label><title>Introduction</title>
<p>Influenza and coronavirus disease (COVID-19) are highly contagious respiratory diseases imposing substantial mortality, morbidity and health system burden. In developed countries, influenza and COVID-19 surveillance systems act as a cornerstone of disease outbreak monitoring, and are essential to guide appropriate public health interventions. In Canada, traditional disease surveillance systems for influenza, influenza-like illnesses (ILIs) and COVID-19 are mostly passive and have limitations in detecting outbreaks in real-time, as they mostly rely on laboratory reports of positive cases, sentinel practitioner and syndromic surveillance programs, and provincial/territorial assessment of the geographic spread of influenza and COVID-19 (<xref ref-type="bibr" rid="B1">1</xref>,<xref ref-type="bibr" rid="B2">2</xref>). The rise of the web and social media have raised new prospects for expanding traditional surveillance of influenza and COVID-19, especially through the use of Internet- or social media-based streams as a complementary data source for real-time digital disease surveillance (<xref ref-type="bibr" rid="B3">3</xref>). In addition, recent advances in artificial intelligence through machine learning techniques and natural language processing (NLP) have contributed new means of extraction of key information from social media text data on a number of health-related topics (<xref ref-type="bibr" rid="B4">4</xref>), including monitoring epidemic spread of influenza and COVID-19 (<xref ref-type="bibr" rid="B5">5</xref>).</p>
<p>The analysis of user-generated information from social media has matured into a useful tool to monitor pandemic situations and understand public opinions during public health emergencies such as influenza and COVID-19. Several recent studies have analyzed influenza and COVID-19 related social media data from Twitter and Reddit using machine learning and deep learning methods for sentiment analyses to detect outbreaks and uncover themes of public concern (<xref ref-type="bibr" rid="B6">6</xref>&#x2013;<xref ref-type="bibr" rid="B13">13</xref>). Specific studies focused on sentiment classification of influenza and COVID-19 tweets written in different languages, such as English (<xref ref-type="bibr" rid="B6">6</xref>,<xref ref-type="bibr" rid="B14">14</xref>), Arabic (<xref ref-type="bibr" rid="B15">15</xref>), and Nepali (<xref ref-type="bibr" rid="B16">16</xref>&#x2013;<xref ref-type="bibr" rid="B18">18</xref>). In addition to sentiment analysis, an alternative method for monitoring influenza and COVID-19 trends on social media platforms is through the tracking of self-reports of symptoms and health experiences related to these diseases by users (<xref ref-type="bibr" rid="B19">19</xref>). This could offer useful complementary information for case detection. Previous studies have explored voluntary self-reporting of potential exposure to COVID-19 by Twitter users in the US and UK (<xref ref-type="bibr" rid="B20">20</xref>,<xref ref-type="bibr" rid="B21">21</xref>), but there is limited research on assessing user-level personal reports of influenza and COVID-19 in Canada. The first objective of our study is to assess whether English-speaking Canadian Twitter users disclose health experiences related to influenza and COVID-19 &#x2014; such as potential exposure, symptoms, treatment and disease outcomes &#x2014; that could be indicative of recent plausible influenza cases, ILIs or symptomatic COVID-19 infections. To support that process, we sought to construct ground-truth Twitter datasets for influenza and (separately) COVID-19 detection based on Twitter users&#x2019; reports in Saskatchewan, Canada.</p>
<p>Transformer-based pretrained language models (T-PTLMs) like Bidirectional Encoder Representations from Transformers (BERT) (<xref ref-type="bibr" rid="B22">22</xref>), RoBERTa (<xref ref-type="bibr" rid="B23">23</xref>) GPT-1 (<xref ref-type="bibr" rid="B24">24</xref>), and XLNet (<xref ref-type="bibr" rid="B25">25</xref>) have achieved great success in supporting Natural Language Processing (NLP) tasks (<xref ref-type="bibr" rid="B26">26</xref>). In recent years, T-PTLMs have further demonstrated outstanding performance on a number of NLP tasks in the public health domain (<xref ref-type="bibr" rid="B8">8</xref>,<xref ref-type="bibr" rid="B27">27</xref>,<xref ref-type="bibr" rid="B28">28</xref>). To enhance the performance of transformer-based language models in domain-specific or platform-specific tasks, various transformer-based language models have been pre-trained using either domain-specific or platform-specific data (<xref ref-type="bibr" rid="B26">26</xref>). BERTweet, BERTweet-covid, and COVID-Twitter BERT (CT-BERT) are notable social media-based T-PTLMs that were trained on Twitter data or Twitter data related to the COVID-19 pandemic. Despite the expectation that domain-specific language models would outperform general language models like BERT and RoBERTa in relevant NLP tasks, recent research findings in health-related social media text classification tasks have yielded mixed results. Some studies have shown that domain-specific language models do not necessarily outperform models pretrained on generic texts such as BERT or RoBERTa in sentiment analyses on COVID-19 tweets or general health-related text classification tasks (<xref ref-type="bibr" rid="B4">4</xref>,<xref ref-type="bibr" rid="B10">10</xref>). In this study, our second objective was to assess and compare the text classification performance of language models pretrained with either general knowledge text or domain-specific data on two social media-based text classification tasks. Specifically, we compared the performance of BERT and variants thereof &#x2014; namely BERT, RoBERTa, BERTweet, and CT-BERT &#x2014; in the context of influenza and COVID-19 infection detection using Twitter datasets. Additionally, due to class imbalances in our annotated Twitter datasets, we also explored the impact of sampling methods on the performance of these language models.</p>
</sec>
<sec id="s2"><label>2.</label><title>Materials and methods</title>
<sec id="s2a"><label>2.1.</label><title>Data collection, extraction and annotation</title>
<p>We collected more than 50 million tweets using the Twitter application programming interface (API) between December 2016 and December 2022 in Saskatchewan (SK), Canada. We used two queries, each requesting tweets within a specific distance of a respective point in SK from December 2016 to December 2022. We then used a keyword-based filtering approach to extract relevant tweets that contained influenza-related keywords or COVID-19 symptom-related keywords. An overview of the data collection and extraction process is shown in <xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>; we turn now to discuss the derivation of each dataset in turn.</p>
<fig id="F1" position="float"><label>Figure 1</label>
<caption><p>Overview of data collection and extraction processes.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-05-1203874-g001.tif"/>
</fig>
<p>The influenza-related Twitter dataset was extracted by filtering SK tweets using a set of influenza-related keywords from December 13, 2016 to March 17, 2018. The set of influenza-related keywords were developed based on surveillance standards for influenza and influenza-like illness (ILI), and can be broadly categorized as: (1) formal and informal names and agents of influenza and ILI (e.g., &#x201C;swine flu&#x201D;, &#x201C;H1N1&#x201D;, &#x201C;H3N2&#x201D;, &#x201C;flu virus&#x201D;, &#x201C;influenza virus&#x201D;, &#x201C;Influenza B virus&#x201D;, &#x201C;seasonal influenza&#x201D;, &#x201C;cold&#x201D;, etc.); (2) symptoms of influenza and flu-like illnesses (e.g., &#x201C;fever&#x201D;, &#x201C;cough&#x201D;, &#x201C;sneeze&#x201D;, &#x201C;sore throat&#x201D;, &#x201C;runny or stuffy nose&#x201D;, &#x201C;muscle or body aches&#x201D;, &#x201C;muscle pain&#x201D;, &#x201C;myalgia&#x201D;, &#x201C;headaches&#x201D;, &#x201C;upper respiratory tract infection&#x201D;, &#x201C;acute respiratory illness&#x201D;, &#x201C;shortness of breath&#x201D;, &#x201C;malaise&#x201D;, &#x201C;fatigue&#x201D;, &#x201C;coryza&#x201D;, etc); (3) flu complications (e.g., &#x201C;vomiting&#x201D;, &#x201C;pneumonia&#x201D;, &#x201C;sinus infection&#x201D;, &#x201C;ear infection&#x201D;); and (4) influenza prevention measures and treatments such as flu shots, antiviral drugs as designated using both their generic and brand names (e.g., &#x201C;oseltamivir&#x201D;, &#x201C;zanamivir&#x201D;, &#x201C;peramivir&#x201D;, &#x201C;Tamiflu&#x201D;).</p>
<p>The COVID-19 Twitter dataset was filtered using a set of COVID-19 symptom-related keywords from January 1, 2020 to June 22, 2021. We limited the keywords to COVID-19 symptoms only, as our objective was to identify tweets that contain information that are indicative of recent plausible symptomatic COVID-19 cases or COVID-19-like cases. The keyword list included &#x201C;cough&#x201D;, &#x201C;shortness of breath&#x201D;, &#x201C;loss of taste and smell&#x201D;, &#x201C;difficulty of breathing&#x201D;, &#x201C;headache&#x201D;, &#x201C;chills&#x201D;, &#x201C;fever&#x201D;, &#x201C;fatigue&#x201D;, &#x201C;muscle aches&#x201D;, &#x201C;body ache&#x201D;, &#x201C;vomiting&#x201D;, &#x201C;nausea&#x201D;, &#x201C;throw up&#x201D;, &#x201C;diarrhea&#x201D;, &#x201C;abdominal pain&#x201D;. We explored a variety of broader keywords associated with the COVID-19 pandemic, and we found that tweets filtered using terms specifically relating to COVID-19 symptoms were more relevant to the objective of our research. There are considerable similarities in symptoms between influenza and COVID-19; thus, when annotating the COVID-19 tweets, we consider plausible influenza cases or ILIs as &#x201C;plausible COVID-19-like cases&#x201D;.</p>
<p>Following the data cleansing process, 6,444 influenza-related tweets and 11,939 COVID-19 symptom-related tweets were retained for subsequent annotation. In this work, a team of 12 annotators systematically conducted manual annotation of pre-filtered tweets as to whether they described recent plausible influenza cases, ILIs or symptomatic COVID-19 infections. The annotation tasks were conducted independently for these two datasets. The influenza dataset was annotated by 8 annotators, whereas 7 annotators annotated the COVID-19 dataset. Three annotators contributed to the annotation tasks of both datasets. <xref ref-type="table" rid="T1">Table&#x00A0;1</xref> provides a description of the respective annotation rules and guidelines for these two datasets to help the annotators distinguish the tweets. As the outcomes of this research is to support monitoring of users&#x2019; personal reports of health experiences associated with influenza and COVID-19, we did not regard news headlines, organizational tweets, or health promotion tweets as indicative of recent plausible influenza or COVID-19 cases. Tweet annotation required domain expertise; therefore, annotators were provided with a 1-hour training session covering the annotation task and key facts about influenza and COVID-19.</p>
<table-wrap id="T1" position="float"><label>Table 1</label>
<caption><p>Twitter datasets annotation tasks.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Annotation guidelines</th>
<th valign="top" align="center">Influenza twitter dataset</th>
<th valign="top" align="center">COVID-19 twitter dataset</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Objective</td>
<td valign="top" align="left">Annotate tweets as to whether they suggest recent plausible influenza cases or influenza-like cases with one of the options below:<break/>1. Definitely yes (recent convincing influenza cases or ILIs).<break/>2. Maybe yes (recent plausible but not convincing influenza cases or ILIs).<break/>3. Unsure.<break/>4. Maybe no (unlikely recent plausible influenza cases or ILIs<break/>5. Definitely no (not recent plausible influenza cases or ILIs)<break/>6. I am not suitable to judge it, skip for now.</td>
<td valign="top" align="left">Annotate tweets as to whether they suggest recent plausible symptomatic COVID-19 or COVID-19-like (plausible influenza cases or ILIs) cases with one of the options below:<break/>1. Definitely yes (recent convincing COVID-19 or COVID-19-like cases).<break/>2. Maybe yes (recent plausible but not convincing COVID-19 or COVID-19-like cases).<break/>3. Unsure.<break/>4. Maybe no (unlikely recent plausible COVID-19 or COVID-19-like cases)<break/>5. Definitely no (not recent plausible COVID-19 or COVID-19-like cases)<break/>6. I am not suitable to judge it, skip for now.</td>
</tr>
<tr>
<td valign="top" align="left">Annotation Rules</td>
<td valign="top" align="left" colspan="2">1. Human subjects. It could implicate the user, someone the user might know, or people they encountered in public places.<break/>2. The use of informal language and slang terms such as onomatopoeic expressions (e.g., &#x201C;cough&#x201D; or &#x201C;cough cough&#x201D;) should be scrutinized meticulously.</td>
</tr>
<tr>
<td valign="top" align="left"/>
<td valign="top" align="left">3. Flu-related advertisements, media news, health promotion-related tweets, general comments on influenza or flu shots won&#x2019;t be considered as indicative of plausible influenza or influenza-like cases.</td>
<td valign="top" align="left">3. Organizational tweets, advertisements, media news, health promotion, general comments on COVID-19 and vaccine, and side effects after COVID-19 vaccine won&#x2019;t be considered as indicative of plausible COVID-19 cases.</td>
</tr>
<tr>
<td valign="top" align="left">Tweet Examples</td>
<td valign="top" align="left">Examples of tweets inferring plausible influenza cases: &#x201C;<italic>Home with a cold flu surrounded by kleenex; blankets and some light reading material.</italic>&#x201D;<break/>&#x201C;<italic>Got a visit in last night with this sick little boy battling some flu like symptoms</italic>&#x201D;</td>
<td valign="top" align="left">Examples of tweets inferring plausible COVID-19 cases:<break/>&#x201C;<italic>Day 2 of self isolation - exhausted; fever is a bit worse but not out of control; painful body aches; cough is wors&#x00E2;</italic>&#x201D;<break/>&#x201C;<italic>We have waited 4.5 hours for a callback from 811 after our youngest started throwing up and spiked a minor fever</italic>&#x201D;</td>
</tr>
<tr>
<td valign="top" align="left">Annotation Pre-Processing</td>
<td valign="top" align="left">None</td>
<td valign="top" align="left">The authors annotated tweets that they judged to clearly not suggest plausible COVID-19 cases based on the rules. 4398 tweets were considered not to suggest recent plausible COVID-19 cases. The remaining 7541 tweets were then annotated by volunteers.</td>
</tr>
<tr>
<td valign="top" align="left">Annotation Processes</td>
<td valign="top" align="left" colspan="2">Annotators underwent a comprehensive one-hour training session that covered the key facts about influenza or COVID-19 (including causes, symptoms, complications, drugs, vaccines) and guidelines for annotation.<break/>At the end of the session, the annotators took a quiz to assess their knowledge of the disease and understanding of the task.</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Each influenza-related tweet was annotated by three individuals, with at least one native English speaker among them. For those cases in which the majority was unable to establish a clear label, another annotator was brought in to re-annotate the tweet. The COVID-19 Twitter dataset was annotated slightly differently. One of the authors first evaluated all tweets, annotating those evidently not indicative of plausible COVID-19 or COVID-19-like cases as &#x201C;No&#x201D; (<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM1"><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:mn>4</mml:mn><mml:mrow><mml:mo>,</mml:mo></mml:mrow><mml:mn>398</mml:mn></mml:math></inline-formula>). The remaining dataset (<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM2"><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:mn>7</mml:mn><mml:mrow><mml:mo>,</mml:mo></mml:mrow><mml:mn>541</mml:mn></mml:math></inline-formula>) was assessed by two individuals, including at least one native English speaker. In the case of disagreement, the tweet was annotated by another annotator. For both datasets, majority voting was used to consolidate the tweet&#x2019;s final binary annotation, which was either 1-&#x201C;Yes&#x201D; or 0-&#x201C;No or Unsure&#x201D;. <xref ref-type="fig" rid="F2">Figure&#x00A0;2</xref> shows the final distribution of annotations in the two datasets.</p>
<fig id="F2" position="float"><label>Figure 2</label>
<caption><p>Distribution of annotations in Twitter datasets.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-05-1203874-g002.tif"/>
</fig>
</sec>
<sec id="s2b"><label>2.2.</label><title>Data preprocessing</title>
<p><xref ref-type="fig" rid="F3">Figure&#x00A0;3</xref> presents the high-level study workflow diagram. Unlike formal text, tweets are characterized by their short and often unstructured nature. Without pre-processing, fine-tuning a pretrained language model with raw tweets may result in poor model performance (<xref ref-type="bibr" rid="B29">29</xref>). Therefore, we pre-processed the raw tweets with the following steps inspired by previous studies (<xref ref-type="bibr" rid="B30">30</xref>,<xref ref-type="bibr" rid="B31">31</xref>): (1) removal of emojis, (2) elimination of punctuation, links and mentions, (3) eliminating hashtags found at the end of tweets while retaining those in the middle after removing the &#x201C;&#x0023;&#x201D; symbols, (4) removal of unnecessary spaces and newlines, (5) conversion to lowercase, (6) removal of empty tweets (resulting 40 tweets to be removed from COVID-19 dataset), and (7) loading the tokenizers from the pretrained models using the <italic>Transformers</italic> library (<xref ref-type="bibr" rid="B32">32</xref>). In our data-processing, we opted to exclude emojis, as we separately compared the performance of BERTweet and CT-BERT with and without emojis on the subset of tweets that contained emojis. With emojis being converted to their textual descriptions, we did not observe improvement in classification performance with emojis included.</p>
<fig id="F3" position="float"><label>Figure 3</label>
<caption><p>A high-level study workflow diagram.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-05-1203874-g003.tif"/>
</fig>
<p>Both datasets were split them into training and test subsets, respectively. We set 20&#x0025; of each the datasets after data preprocessing as the test subset for that dataset. To address the class imbalance issue in both datasets, we applied oversampling and undersampling methods to each of the training subsets. Specifically, we used the <italic>imbalanced-learn</italic> Python library (<xref ref-type="bibr" rid="B33">33</xref>) to oversample the minority class and undersample the majority class, respectively. <xref ref-type="table" rid="T2">Table&#x00A0;2</xref> presents the training and test sets for the influenza and COVID-19 Twitter datasets. We further set 10&#x0025; of these two balanced training data subsets as validation data for training models.</p>
<table-wrap id="T2" position="float"><label>Table 2</label>
<caption><p>Training and test dataset sizes.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left"/>
<th valign="top" align="center" colspan="2">Influenza twitter dataset</th>
<th valign="top" align="center" colspan="2">COVID-19 twitter dataset</th>
</tr>
<tr>
<th valign="top" align="left"/>
<th valign="top" align="center">&#x201C;Yes&#x201D; (Class 1)</th>
<th valign="top" align="center">&#x201C;No or Unsure&#x201D; (Class 0)</th>
<th valign="top" align="center">&#x201C;Yes&#x201D; (Class 1)</th>
<th valign="top" align="center">&#x201C;No or Unsure&#x201D; (Class 0)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Undersampled Training Set</td>
<td valign="top" align="center">899</td>
<td valign="top" align="center">899</td>
<td valign="top" align="center">755</td>
<td valign="top" align="center">755</td>
</tr>
<tr>
<td valign="top" align="left">Oversampled Training Set</td>
<td valign="top" align="center">4256</td>
<td valign="top" align="center">4256</td>
<td valign="top" align="center">8762</td>
<td valign="top" align="center">8762</td>
</tr>
<tr>
<td valign="top" align="left">Test Set</td>
<td valign="top" align="center">230</td>
<td valign="top" align="center">1059</td>
<td valign="top" align="center">169</td>
<td valign="top" align="center">2213</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2c"><label>2.3.</label><title>Transformer-based pretrained language models</title>
<p>Our study investigated the text classification performance of pre-trained BERT and three BERT variant models &#x2013; RoBERTa, BERTweet, and CT-BERT using annotated influenza and COVID-19 Twitter datasets. <xref ref-type="table" rid="T3">Table&#x00A0;3</xref> shows the eight BERT-based models employed in this study. We carried out our experiments using Google Colaboratory (<xref ref-type="bibr" rid="B34">34</xref>).</p>
<table-wrap id="T3" position="float"><label>Table 3</label>
<caption><p>BERT and BERT variant models used in the study.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Model</th>
<th valign="top" align="center">Architecture</th>
<th valign="top" align="center">Task</th>
<th valign="top" align="center">Model versions</th>
<th valign="top" align="center">Pretraining corpus-based</th>
<th valign="top" align="center">Pretraining data</th>
<th valign="top" align="center">Total parameters</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">BERT</td>
<td valign="top" align="left">Multi-layer bidirectional transformer encoder</td>
<td valign="top" align="left">MLM, NSP</td>
<td valign="top" align="left">BERT-base, BERT-large</td>
<td valign="top" align="left">general-domain</td>
<td valign="top" align="left">800 M words (BooksCorpus) and 2500 M words (English Wikipedia), 16GB of uncompressed text.</td>
<td valign="top" align="left">base: 110 M; large: 340M</td>
</tr>
<tr>
<td valign="top" align="left">RoBERTa</td>
<td valign="top" align="left">BERT with improved pretraining procedures</td>
<td valign="top" align="left">MLM</td>
<td valign="top" align="left">RoBERTa-base, RoBERTa-large</td>
<td valign="top" align="left">general-domain</td>
<td valign="top" align="left">data used to train BERT, 63 M English news articles (CC-NEWS), OpenWebText, and Stories; 160GB of uncompressed text.</td>
<td valign="top" align="left">base: 124 M; large: 355 M</td>
</tr>
<tr>
<td valign="top" align="left">BERTweet</td>
<td valign="top" align="left">BERT-base model with RoBERTa&#x2019;s pre-training procedure.</td>
<td valign="top" align="left">MLM</td>
<td valign="top" align="left">BERTweet-base, BERTweet-large, BERTweet-covid-base</td>
<td valign="top" align="left">social media-based, domain-specific</td>
<td valign="top" align="left">BERTweet-base and large: 845M English Tweets and 5 M COVID-19 tweets; BERTweet-covid-base: 23 M COVID-19 English tweets.</td>
<td valign="top" align="left">base: 135 M; large: 355 M</td>
</tr>
<tr>
<td valign="top" align="left">COVID-Twitter-BERT</td>
<td valign="top" align="left">BERT-large</td>
<td valign="top" align="left">MLM, NSP</td>
<td valign="top" align="left">COVID-Twitter-BERT-v2</td>
<td valign="top" align="left">social media-based, domain-specific</td>
<td valign="top" align="left">160 M COVID-19 English tweets</td>
<td valign="top" align="left">340 M</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="table-fn1"><p><bold>MLM</bold>: Masked Language Modeling; <bold>NSP</bold>: Next Sentence Prediction</p></fn>
</table-wrap-foot>
</table-wrap>
<p><bold>BERT</bold> Bidirectional Encoder Representations from Transformers (BERT) is a language representation model using the widely adopted transformer architecture (<xref ref-type="bibr" rid="B35">35</xref>). Given the wide use of transformers, we will omit an exhaustive description of its architecture and direct readers to Vaswani et al. (<xref ref-type="bibr" rid="B35">35</xref>) on the details. BERT is pre-trained on two unsupervised tasks: Masked Language Modeling (MLM) and Next Sentence Prediction (NSP) (<xref ref-type="bibr" rid="B22">22</xref>). MLM is an unsupervised task where certain tokens in a text sequence are randomly masked, and the model aims to predict these masked segments of the text. NSP is a task that the model predicts whether a given pair of sentences in a text document are consecutive. The pre-training corpus of BERT includes BooksCorpus and Wikipedia datasets (16GB). BERT has many variants models with different model sizes. We used BERT-base and BERT-large in this study with parameters of 110M and 340M, respectively.</p>
<p><bold>RoBERTa</bold> Robustly Optimized BERT Pretraining Approach (RoBERTa) (<xref ref-type="bibr" rid="B23">23</xref>) is an extension of BERT and has the same architecture as BERT. RoBERTa can be seen as an enhanced version of BERT with improved pretraining procedures and additional pre-training data. RoBERTa applies a dynamic masking approach and removes the NSP objective, which sets it apart of BERT. RoBERTa&#x2019;s training time is longer and utilizes a larger corpora in comparison to BERT. RoBERTa is pretrained on a total of 160GB of text data, including the original data used to train BERT (BooksCorpus and Wikipedia), CC-News, OpenWebText, and Stories. For our experiments on COVID-19 and influenza Twitter datasets, we used RoBERTa-base and RoBERTa-large models, which include 124M and 355M parameters, respectively.</p>
<p><bold>BERTweet</bold> BERTweet (<xref ref-type="bibr" rid="B36">36</xref>) is a variant of BERT trained with a masked language modeling objective. Using the same architecture as BERT-base, BERTweet is trained to process text data from Twitter. The pre-training procedure of BERTweet is based on RoBERTa for improved performance. After preprocessing the raw tweets, BERTweet used 845M English tweets and 5M English tweets related to the COVID-19 pandemic (from 01/2020 to 03/2020) to train BERTweet-base and BERTweet-large models. The total pretraining data used for these models amounts to approximately 80GB (16B word tokens). Subsequently, a corpus of 23M COVID-19 related English tweets was collected, and the authors continued pre-training using the base version of pre-trained BERTweet (<xref ref-type="bibr" rid="B36">36</xref>). This led to the development of the BERTweet-covid-base model. This study used the pre-trained BERTweet-base, BERTweet-covid-base (uncased), and BERTweet-large, which have 135M, 135M, and 355M parameters, respectively.</p>
<p><bold>CT-BERT</bold> COVID-Twitter-BERT (CT-BERT) (<xref ref-type="bibr" rid="B37">37</xref>) is a transformer-based model that utilizes the BERT-large architecture. It underwent pretraining on a corpus of 160M English tweets specifically focused on COVID-19, covering the period from January 12 to April 16, 2020. The first version of CT-BERT is pretrained on a corpus of 160M English tweets related to COVID-19 from January 12 to April 16, 2020. Before the pretraining process, the original tweets were pre-processed, which involved cleaning the tags, replacing the username and URLs, and eliminating duplicate tweets. The final corpus contains 22.5M tweets (0.6 billion words). The second version of the CT-BERT is identical to the first version except that it was trained on more COVID-19 English tweets collected until July 5, 2020. In this study, we used the CT-BERT v2 version, which was pretrained on a larger COVID-19 Twitter dataset.</p>
</sec>
<sec id="s2d"><label>2.4.</label><title>Fine-tuning</title>
<p>Fine-tuning T-PTLMs involves training the pretrained models on a specific downstream task. This is done by adding a dense layer and a task-specific output layer for adjusting model parameters on a task-specific dataset. The models are initialized with pretrained parameters and then fine-tune the entire neural network by using gradient descent-based backpropagation technology to find the optimal hyperparameter values (<xref ref-type="bibr" rid="B38">38</xref>).</p>
<p>In our study, we fine-tuned BERT-base, RoBERTa-base, BERTweet-base, BERTweet-covid-base, and CT-BERT. Inspired by previous studies (<xref ref-type="bibr" rid="B22">22</xref>,<xref ref-type="bibr" rid="B23">23</xref>), we experimented with learning rates of <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM3"><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mn>5</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>, batch sizes of <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM4"><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mn>16</mml:mn><mml:mo>,</mml:mo><mml:mn>32</mml:mn><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>, and epochs of <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM5"><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mn>4</mml:mn><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula> with a basic grid search strategy. For BERT-large, RoBERTa-large, and BERTweet-large, we used the same hyperparameter values as in their respective base versions. Finally, the Adam optimizer (<xref ref-type="bibr" rid="B39">39</xref>) was employed for all models, and we measured the categorical cross-entropy and categorical accuracy during fine-turning. The optimal hyperparameter values for each model are presented in <xref ref-type="table" rid="T4">Table&#x00A0;4</xref>.</p>
<table-wrap id="T4" position="float"><label>Table 4</label>
<caption><p>Optimal parameters used for BERT and BERT variant models.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left"/>
<th valign="top" align="center" colspan="2">BERT</th>
<th valign="top" align="center" colspan="2">RoBERTa</th>
<th valign="top" align="center" colspan="2">BERTweet</th>
<th valign="top" align="center" colspan="2">BERTweet-covid</th>
<th valign="top" align="center" colspan="2">CT-BERT</th>
</tr>
<tr>
<th valign="top" align="center"/>
<th valign="top" align="center">Covid</th>
<th valign="top" align="center">Flu</th>
<th valign="top" align="center">Covid</th>
<th valign="top" align="center">Flu</th>
<th valign="top" align="center">Covid</th>
<th valign="top" align="center">Flu</th>
<th valign="top" align="center">Covid</th>
<th valign="top" align="center">Flu</th>
<th valign="top" align="center">Covid</th>
<th valign="top" align="center">Flu</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Learning Rate</td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM6"><mml:mn>1</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM7"><mml:mn>3</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM8"><mml:mn>1</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM9"><mml:mn>3</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM10"><mml:mn>3</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM11"><mml:mn>2</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM12"><mml:mn>1</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM13"><mml:mn>3</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM14"><mml:mn>1</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
<td valign="top" align="center"><inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM15"><mml:mn>3</mml:mn><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula></td>
</tr>
<tr>
<td valign="top" align="left">Epochs</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">3</td>
</tr>
<tr>
<td valign="top" align="left">Batch Size</td>
<td valign="top" align="center">32</td>
<td valign="top" align="center">16</td>
<td valign="top" align="center">32</td>
<td valign="top" align="center">16</td>
<td valign="top" align="center">32</td>
<td valign="top" align="center">16</td>
<td valign="top" align="center">32</td>
<td valign="top" align="center">16</td>
<td valign="top" align="center">32</td>
<td valign="top" align="center">16</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="table-fn2"><p>Covid represents the COVID-19 Twitter dataset; Flu represents the influenza Twitter dataset</p></fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s2e"><label>2.5.</label><title>Evaluation metrics</title>
<p>Model performance was evaluated using recall, F1 score, accuracy, and area under the receiver operating characteristic curve (AUC). To obtain a more reliable evaluation of the models&#x2019; performance, we conducted five separate runs with varying initialization for each model and dataset, while maintaining invariant the values of the fine-tuning hyperparameters. This is due to the instability of the fine-tuning process, as the classification performance can vary significantly when training the same model with different random seeds (<xref ref-type="bibr" rid="B40">40</xref>). The resulting classification performance metrics (recall, F1, accuracy, and AUC) were averaged over the test subsets of the COVID-19 and influenza Twitter datasets, respectively. Additionally, we performed independent one-sided T-tests to compare the model performance for each sampling method.</p>
</sec>
</sec>
<sec id="s3"><label>3.</label><title>Results and analysis</title>
<sec id="s3a"><label>3.1.</label><title>Exploratory data analysis</title>
<p>We constructed two datasets of annotated tweets based on users&#x2019; reports of recent plausible influenza or COVID-19 infections. Each tweet was annotated by no fewer than two annotators, with the final annotation determined through the majority rule. Of the 6444 influenza-related tweets, 1129 (17.5&#x0025;) were annotated as &#x201C;Yes&#x201D; &#x2014; suggesting recent plausible influenza cases or ILIs. Of the 11939 COVID-19 symptom-related tweets, 924 (7.7&#x0025;) were annotated as &#x201C;Yes&#x201D; &#x2014; suggesting recent plausible symptomatic COVID-19 or COVID-19-like cases. <xref ref-type="fig" rid="F4">Figure&#x00A0;4</xref> illustrates the frequencies of the top 250 word stems in the influenza and COVID-19 datasets. Word stems located far from the dotted line were more prevalent in one class than the other, while those that were close to the line had similar frequencies in both the &#x201C;Yes&#x201D; and &#x201C;No or Unsure&#x201D; classes. The word stems were extracted by removing links and usernames, tokenizing the tweets, stemming the words and removing stop words. In <xref ref-type="fig" rid="F4">Figure&#x00A0;4</xref>, we observe that several symptom-related word stems (such as &#x201C;sick&#x201D;, &#x201C;puke&#x201D;, &#x201C;fever&#x201D;, &#x201C;sore&#x201D;, &#x201C;cough&#x201D;) exhibit higher frequencies in the &#x201C;Yes&#x201D; class. In the COVID-19 Twitter dataset, the number &#x201C;811&#x201D; having a higher frequency in the &#x201C;Yes&#x201D; class represents Saskatchewan&#x2019;s HealthLine number, which was widely promoted as a source of non-urgent medical advice and health information for COVID-19.</p>
<fig id="F4" position="float"><label>Figure 4</label>
<caption><p>Frequencies of the top 250 word stems by different annotations.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-05-1203874-g004.tif"/>
</fig>
</sec>
<sec id="s3b"><label>3.2.</label><title>Comparison of model performance</title>
<p>The performance of the models under various sampling methods is presented in <xref ref-type="table" rid="T5">Tables&#x00A0;5</xref> and <xref ref-type="table" rid="T6">6</xref>. The highest-ranked value of each metric across the four models given a specific sampling method and model size group (&#x201C;base&#x201D; or &#x201C;large&#x201D;) was marked as the reference group (&#x201C;ref&#x201D;) for both datasets. To assess the model performance, independent one-sided t-tests were conducted to compare the results of each model to the highest ranked one (the reference group) for the same metric, sampling method and dataset across language models in a given model size group (&#x201C;base&#x201D; or &#x201C;large&#x201D;). Additionally, for each Twitter dataset, we compared the performance of different sampling methods for each model, as well as the performance of eight models within each sampling method regardless of the model size group. These comparisons can be found in Tables S1 and S2 in the <xref ref-type="sec" rid="s10">Supplementary Materials</xref>.</p>
<table-wrap id="T5" position="float"><label>Table 5</label>
<caption><p>Performance metrics of models on COVID-19 Twitter datasets for each sampling method.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left" rowspan="2">Size</th>
<th valign="top" align="center" rowspan="2">Model</th>
<th valign="top" align="center" colspan="3">Mean Recall</th>
<th valign="top" align="center" colspan="3">Mean F1</th>
<th valign="top" align="center" colspan="3">Mean AUC</th>
<th valign="top" align="center" colspan="3">Mean Accuracy</th>
</tr>
<tr>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="4">Base</td>
<td valign="top" align="left">BERT-base</td>
<td valign="top" align="center">92.6</td>
<td valign="top" align="center">91.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM16"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">74<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM17"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM18"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">92<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM19"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">80.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM20"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">63<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM21"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">74.1</td>
<td valign="top" align="center">70<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM22"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">92.7<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM23"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM24"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">74.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM25"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
</tr>
<tr>
<td valign="top" align="left">BERTweet-base</td>
<td valign="top" align="center">92.8</td>
<td valign="top" align="center">93</td>
<td valign="top" align="center">82.6</td>
<td valign="top" align="center">93</td>
<td valign="top" align="center"><bold>93.4(ref)</bold></td>
<td valign="top" align="center">86.6</td>
<td valign="top" align="center">76.7</td>
<td valign="top" align="center">78</td>
<td valign="top" align="center"><bold>87(ref)</bold></td>
<td valign="top" align="center">92.9</td>
<td valign="top" align="center"><bold>93.1(ref)</bold></td>
<td valign="top" align="center">82.6</td>
</tr>
<tr>
<td valign="top" align="left">BERTweet-covid-base</td>
<td valign="top" align="center"><bold>93.4(ref)</bold></td>
<td valign="top" align="center"><bold>93.2(ref)</bold></td>
<td valign="top" align="center"><bold>85.2(ref)</bold></td>
<td valign="top" align="center"><bold>93.6(ref)</bold></td>
<td valign="top" align="center">93.4</td>
<td valign="top" align="center"><bold>88.2(ref)</bold></td>
<td valign="top" align="center">79.5</td>
<td valign="top" align="center"><bold>80.5(ref)</bold></td>
<td valign="top" align="center">86.5</td>
<td valign="top" align="center"><bold>93.6(ref)</bold></td>
<td valign="top" align="center">93.1</td>
<td valign="top" align="center"><bold>85.3(ref)</bold></td>
</tr>
<tr>
<td valign="top" align="left">RoBERTa-base</td>
<td valign="top" align="center">92.8</td>
<td valign="top" align="center">92.6</td>
<td valign="top" align="center">85.2</td>
<td valign="top" align="center">93.2</td>
<td valign="top" align="center">92.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM26"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">88</td>
<td valign="top" align="center"><bold>80(ref)</bold></td>
<td valign="top" align="center">80.2</td>
<td valign="top" align="center">84.7<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM27"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">92.8</td>
<td valign="top" align="center">92.6</td>
<td valign="top" align="center">85.1</td>
</tr>
<tr>
<td valign="top" align="left" rowspan="4">Large</td>
<td valign="top" align="left">BERT-large</td>
<td valign="top" align="center">94<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM28"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">93.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM29"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">81.2</td>
<td valign="top" align="center">93.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM30"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">93.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM31"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85.4</td>
<td valign="top" align="center">72</td>
<td valign="top" align="center">77.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM32"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">82.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM33"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">93.9</td>
<td valign="top" align="center">93.3</td>
<td valign="top" align="center">81.3</td>
</tr>
<tr>
<td valign="top" align="left">BERTweet-large</td>
<td valign="top" align="center">93.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM34"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM35"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">82.6</td>
<td valign="top" align="center">93.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM36"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM37"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">86</td>
<td valign="top" align="center">73</td>
<td valign="top" align="center"><bold>82.8(ref)</bold></td>
<td valign="top" align="center">84.6</td>
<td valign="top" align="center">93.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM38"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.3<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM39"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">82.4</td>
</tr>
<tr>
<td valign="top" align="left">CT-BERT</td>
<td valign="top" align="center"><bold>94.8(ref)</bold></td>
<td valign="top" align="center"><bold>94.2(ref)</bold></td>
<td valign="top" align="center"><bold>85.2(ref)</bold></td>
<td valign="top" align="center"><bold>94.4(ref)</bold></td>
<td valign="top" align="center"><bold>94.2(ref)</bold></td>
<td valign="top" align="center"><bold>88(ref)</bold></td>
<td valign="top" align="center"><bold>77.2(ref)</bold></td>
<td valign="top" align="center">81</td>
<td valign="top" align="center"><bold>86.2(ref)</bold></td>
<td valign="top" align="center"><bold>94.6(ref)</bold></td>
<td valign="top" align="center"><bold>94(ref)</bold></td>
<td valign="top" align="center"><bold>85.2(ref)</bold></td>
</tr>
<tr>
<td valign="top" align="left">RoBERTa-large</td>
<td valign="top" align="center">94<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM40"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">93<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM41"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85</td>
<td valign="top" align="center">94.2</td>
<td valign="top" align="center">93.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM42"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">88</td>
<td valign="top" align="center">76.9</td>
<td valign="top" align="center">81.3</td>
<td valign="top" align="center">85.8</td>
<td valign="top" align="center">94.1</td>
<td valign="top" align="center">93.1<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM43"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="table-fn3"><p>The bold values just highlight the &#x201C;ref&#x201D; group. <bold>ref</bold>: the reference group used for the t-test, and it is also the highest ranked value given a sampling method for a specific metric across different language models in either the &#x201C;base&#x201D; or &#x201C;large&#x201D; model size; <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM44"><mml:msup><mml:mi></mml:mi><mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula>: the metric is statistically different from the reference group for a given sampling method and metric across language models in the &#x201C;base&#x201D; or &#x201C;large&#x201D; model size; <bold>N</bold>: no sampling; <bold>O</bold>: random oversampling; <bold>U</bold>: random undersampling.</p></fn>
</table-wrap-foot>
</table-wrap>
<table-wrap id="T6" position="float"><label>Table 6</label>
<caption><p>Performance metrics of models on Influenza Twitter datasets for each sampling method.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left" rowspan="2">Size</th>
<th valign="top" align="center" rowspan="2">Model</th>
<th valign="top" align="center" colspan="3">Mean Recall</th>
<th valign="top" align="center" colspan="3">Mean F1</th>
<th valign="top" align="center" colspan="3">Mean AUC</th>
<th valign="top" align="center" colspan="3">Mean Accuracy</th>
</tr>
<tr>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
<th valign="top" align="center">N</th>
<th valign="top" align="center">O</th>
<th valign="top" align="center">U</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="4">Base</td>
<td valign="top" align="left">BERT-base</td>
<td valign="top" align="center">84.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM45"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM46"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">76.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM47"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">83<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM48"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM49"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">79.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM50"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">68.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM51"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">75.3<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM52"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">78.9<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM53"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">84.3<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM54"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85.1<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM55"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">76.5<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM56"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
</tr>
<tr>
<td valign="top" align="left">BERTweet-base</td>
<td valign="top" align="center">90.8</td>
<td valign="top" align="center"><bold>91.8(ref)</bold></td>
<td valign="top" align="center">87.4</td>
<td valign="top" align="center">91</td>
<td valign="top" align="center"><bold>92(ref)</bold></td>
<td valign="top" align="center">88.2</td>
<td valign="top" align="center"><bold>87.1(ref)</bold></td>
<td valign="top" align="center"><bold>88.6(ref)</bold></td>
<td valign="top" align="center">88.9</td>
<td valign="top" align="center">90.9</td>
<td valign="top" align="center"><bold>91.9(ref)</bold></td>
<td valign="top" align="center">87.4</td>
</tr>
<tr>
<td valign="top" align="left">BERTweet-covid-base</td>
<td valign="top" align="center"><bold>91.6(ref)</bold></td>
<td valign="top" align="center">91.6</td>
<td valign="top" align="center">87.6</td>
<td valign="top" align="center"><bold>91.6(ref)</bold></td>
<td valign="top" align="center">91.6</td>
<td valign="top" align="center">88.4</td>
<td valign="top" align="center">84.9</td>
<td valign="top" align="center">86.6</td>
<td valign="top" align="center"><bold>89(ref)</bold></td>
<td valign="top" align="center"><bold>91.6(ref)</bold></td>
<td valign="top" align="center">91.7</td>
<td valign="top" align="center">87.5</td>
</tr>
<tr>
<td valign="top" align="left">RoBERTa-base</td>
<td valign="top" align="center">90.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM57"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">90.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM58"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center"><bold>88(ref)</bold></td>
<td valign="top" align="center">90.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM59"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">90.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM60"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center"><bold>88.6(ref)</bold></td>
<td valign="top" align="center">86.3</td>
<td valign="top" align="center">85.9<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM61"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">88.3</td>
<td valign="top" align="center">90.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM62"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">90.5<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM63"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center"><bold>87.8(ref)</bold></td>
</tr>
<tr>
<td valign="top" align="left" rowspan="4">Large</td>
<td valign="top" align="left">BERT-large</td>
<td valign="top" align="center">89.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM64"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM65"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">82.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM66"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM67"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM68"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">84<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM69"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">81.3<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM70"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">81.7<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM71"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">84.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM72"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.7<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM73"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM74"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">82.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM75"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
</tr>
<tr>
<td valign="top" align="left">BERTweet-large</td>
<td valign="top" align="center"><bold>92.4(ref)</bold></td>
<td valign="top" align="center"><bold>92.2(ref)</bold></td>
<td valign="top" align="center">85.4</td>
<td valign="top" align="center"><bold>92.6(ref)</bold></td>
<td valign="top" align="center"><bold>92.2(ref)</bold></td>
<td valign="top" align="center">86.6</td>
<td valign="top" align="center"><bold>87.6(ref)</bold></td>
<td valign="top" align="center">87.1</td>
<td valign="top" align="center"><bold>88(ref)</bold></td>
<td valign="top" align="center"><bold>92.4(ref)</bold></td>
<td valign="top" align="center"><bold>92.2(ref)</bold></td>
<td valign="top" align="center">85.5</td>
</tr>
<tr>
<td valign="top" align="left">CT-BERT</td>
<td valign="top" align="center">90.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM76"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.2</td>
<td valign="top" align="center"><bold>87.4(ref)</bold></td>
<td valign="top" align="center">90.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM77"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.4</td>
<td valign="top" align="center"><bold>88(ref)</bold></td>
<td valign="top" align="center">85.3</td>
<td valign="top" align="center"><bold>88.1(ref)</bold></td>
<td valign="top" align="center">86.5</td>
<td valign="top" align="center">90.5<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM78"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">91.1</td>
<td valign="top" align="center"><bold>87.4(ref)</bold></td>
</tr>
<tr>
<td valign="top" align="left">RoBERTa-large</td>
<td valign="top" align="center">90.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM79"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM80"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">83.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM81"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">90.6<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM82"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.8<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM83"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85.4<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM84"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">85.1</td>
<td valign="top" align="center">87.6</td>
<td valign="top" align="center">87.4</td>
<td valign="top" align="center">90.7<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM85"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">89.2<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM86"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
<td valign="top" align="center">83.9<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM87"><mml:msup><mml:mi></mml:mi><mml:mo>&#x2217;</mml:mo></mml:msup></mml:math></inline-formula></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="table-fn4"><p>The bold values just highlight the &#x201C;ref&#x201D; group. <bold>ref</bold> the reference group used for the t-test, and it is also the highest ranked value given a sampling method for a specific metric across different language models in either the &#x201C;base&#x201D; or &#x201C;large&#x201D; model size; <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM88"><mml:msup><mml:mi></mml:mi><mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula>: the metric is statistically different from the reference group for a given sampling method and metric across language models in the &#x201C;base&#x201D; or &#x201C;large&#x201D; model size; <bold>N</bold>: no sampling; <bold>O</bold>: random oversampling; <bold>U</bold>: random undersampling.</p></fn>
</table-wrap-foot>
</table-wrap>
<sec id="s3b1"><label>3.2.1.</label><title>Performance metrics evaluation of COVID-19 tweet classification</title>
<p>Model performance on the COVID-19 Twitter dataset is presented in <xref ref-type="table" rid="T5">Table&#x00A0;5</xref>. Among the four models in the &#x201C;base&#x201D; model size group, the BERTweet-covid-base model consistently achieved the highest or equally highest mean scores in recall, F1, and accuracy across the various sampling methods. The performance of the BERTweet-base and RoBERTa-base models was similar to that of the BERTweet-covid-base model with no statistically significant difference. In contrast, the BERT-base model exhibited statistically worse performance in mean F1 and mean accuracy compared to the BERTweet-covid-base model.</p>
<p>Among the models in the &#x201C;large&#x201D; size group, the CT-BERT model achieved the highest mean scores in recall, F1, and accuracy across all sampling methods. However, there were no statistically differences in mean scores of recall, F1 and accuracy among the large models when the undersampling method was used. The CT-BERT model demonstrated statistically significant better performance in mean recall and F1 where either the oversampling method or no sampling was applied. Notably, the CT-BERT model showed significant improvement in mean recall and F1 compared to its base model, BERT-large.</p>
</sec>
<sec id="s3b2"><label>3.2.2.</label><title>Performance metrics evaluation of influenza tweet classification</title>
<p><xref ref-type="table" rid="T6">Table&#x00A0;6</xref> presents the model performance on the Influenza Twitter dataset. When undersampling method was employed, RoBERTa-base had the highest mean scores in recall (88&#x0025;), F1 (88.6&#x0025;), and accuracy (87.8&#x0025;) in the &#x201C;base&#x201D; model size group. BERTweet-base, BERTweet-covid-base had comparable performance to the RoBERTa-base model with no statistically significant difference. When no sampling or oversampling was applied, both the BERTweet-base model and BERTweet-covid-base model had better and comparable performance in mean recall, F1, and accuracy measures with no statistical difference observed between them. The RoBERTa-base and BERT-base models exhibited significantly inferior performance in the &#x201C;base&#x201D; model size group with no sampling or oversampling method.</p>
<p>In the &#x201C;large&#x201D; model size group, BERTweet-large model emerged as the top performer under no sampling or oversampling method, delivering the best mean scores in recall, F1, and accuracy. While the BERT-large model consistently underperformed in the &#x201C;large&#x201D; model size group regardless of the sampling method used.</p>
</sec>
<sec id="s3b3"><label>3.2.3.</label><title>Base models, large models, and sampling approach</title>
<p>In this study, we used the base and large models of BERT, RoBERTa, and BERTweet. The key similarity between the base and large versions of a pre-trained model lies in their identical architecture and pretraining data. The main difference between the base and large models is their size and computational requirements. All three pretrained base models include 12 transformer layers with over 100 million parameters. By contrast, the large version of the pretrained models include double transformer layers than the base version with over 300 million parameters. Therefore, large models need larger computer resources.</p>
<p>Upon comparing the performance of BERT-base and BERT-large models on the COVID-19 and influenza Twitter datasets, we found that the BERT-large model outperformed BERT-base in all four evaluation metrics. However, the performance of the RoBERTa and BERTweet models &#x2014; whether in their base or large forms &#x2014; varied depending on the datasets and the sampling approach.</p>
<p>With respect to the sampling approach, the results in <xref ref-type="sec" rid="s10">Table S2</xref> (found in the Supplementary Materials) suggest that, when compared to the undersampling method, the classification performance (as measured by mean recall, F1, and accuracy scores) for both the COVID-19 and influenza Twitter datasets was generally better with no sampling or oversampling method. The comparison between no sampling and the oversampling approach revealed varied results across models and datasets. The oversampling method did not significantly enhance model performance on the COVID-19 Twitter dataset, and in certain cases, it even led to poorer performance, as observed with the RoBERTa-large model. On the other hand, for the influenza Twitter dataset, there was no statistically significant difference in mean recall, F1, or accuracy between the oversampling and no sampling methods across the models. In addition, it is worth noting that for the mean AUC scores, the undersampling method yielded better performance for several models.</p>
</sec>
</sec>
</sec>
<sec id="s4" sec-type="discussion"><label>4.</label><title>Discussion</title>
<p>This study investigated whether Twitter users disclose health-related experiences that may suggest recent plausible cases of influenza and COVID-19 in Saskatchewan, Canada. We collected and manually annotated pre-filtered tweets for influenza and COVID-19 detection, and applied pre-trained BERT and BERT variant models to classify the tweets. We achieved an F1 score of up to 94.4&#x0025; on the COVID-19 Twitter dataset and 92.6&#x0025; on the influenza Twitter dataset. To the best of our knowledge, our study is the first to examine the Twitter users&#x2019; reports of health experiences that could be indicative of influenza and COVID-19 infections in Canada. Previous studies have assessed Twitter users&#x2019; self-reports of plausible COVID-19 exposure or infections for tracking COVID-19 spread in the US and UK (<xref ref-type="bibr" rid="B20">20</xref>,<xref ref-type="bibr" rid="B21">21</xref>), and other Canadian studies using COVID-19 related Twitter data are focused on analyzing organizational tweets (<xref ref-type="bibr" rid="B41">41</xref>), COVID-19 tweets posted by federal members of parliament (<xref ref-type="bibr" rid="B42">42</xref>), or more general sentiment analysis using Twitter emojis (<xref ref-type="bibr" rid="B43">43</xref>).</p>
<p>In this study, we evaluated the performance of per-trained BERT and three BERT-variant language models (RoBERTa, BERTweet, and CT-BERT) on our annotated influenza and COVID-19 Twitter datasets, respectively. To address the issue of imbalanced data in classification, we explored data-level sampling approaches to balance the imbalanced class distribution, which include oversampling to augment data in the minority classes and undersampling to decrease data in the majority classes. Our analyses showed that undersampling the training datasets for both classification tasks resulted in significantly worse model performance in mean recall, F1 score, accuracy, while the impact on AUC scores varied by dataset. When the oversampling approach was used, there was no substantial improvement in the mean F1 scores across the models in comparison to the results obtained with no sampling on both datasets. We also found that domain-specific models performed better than general language models on our Twitter datasets. The CT-BERT model, trained on COVID-19 related tweets, demonstrated superior performance on the COVID-19 Twitter dataset with both oversampling and no sampling methods. However, in the context of predicting plausible influenza cases, the CT-BERT model did not consistently outperform the BERTweet-large model, despite the similarities in symptoms between COVID-19 and influenza. This aligns with previous findings that CT-BERT outperforms other models in COVID-19 related classification tasks (<xref ref-type="bibr" rid="B37">37</xref>). It is worth noting that when assessing the performance of CT-BERT on non-COVID-19 related Twitter data &#x2013; the influenza Twitter dataset &#x2013; its performance is comparable or slightly inferior to the BERTweet-large model, which was pre-trained on general Twitter language rather than exclusively COVID-19 related language. Furthermore, the BERT models performed worse than domain-specific models on both classification tasks.</p>
<p>We further compared the performance of the base and large versions of the BERT, RoBERTa, and BERTweet models. Despite the far larger parameter count of large models, we did not observe a consistent trend of large models outperforming their base counterparts. Although BERT-large demonstrated better performance than BERT-base, RoBERTa-base and BERTweet-base models yielded mixing results to their large counterparts on our Twitter datasets across different sampling method. One possible reason for this could be that the datasets used for fine-tuning are insufficiently large to fully utilize the capacity of the large models. As a result, the large models may end up overfitting to the training data, precipitating little performance improvement on the test sets. Considering that large models require more computational resources and longer training time, using the base models may suffice for many practical applications.</p>
<p>This study is subject to several limitations. Firstly, the size of our Twitter datasets was relatively small, and we restricted our analysis to English-language tweets from Saskatchewan, Canada. Therefore, our findings may not be generalizable to other jurisdictions. Secondly, we only used symptom-related keywords to filter COVID-19 tweets, instead of using more comprehensive keywords related to the COVID-19 pandemic. This was done to render more manageable the number of tweets requiring manual annotation, but may have overlooked other types of self-reporting related to COVID-19 infections that are not based on symptoms, or that use other terms for symptoms. In addition, although monitoring of individuals&#x2019; self-reporting of symptoms and health experiences related to influenza and COVID-19 through social media presents an opportunity to complement real-time disease surveillance, it has its own limitations. One such limitation is the inability to detect asymptomatic cases of COVID-19, which is estimated to represent 40.5&#x0025; of the population with confirmed COVID-19 infections (<xref ref-type="bibr" rid="B44">44</xref>). Recent changes to the size and character of the twitterverse and the rapid escalation in cost for production-scale broader tweet monitoring also raise concerns. Furthermore, since both the influenza and COVID-19 datasets exhibited a high frequency of symptom-related words, monitoring the disease spread through this method may not be sufficiently timely given the incubation periods of both infections, and merits combination with less highly lagged surveillance data sources, such as those involve wastewater assays (<xref ref-type="bibr" rid="B45">45</xref>).</p>
<p>Despite such limitations, the classification methods presented here merit consideration for complementing existing sentinel and syndromic surveillance methods, for which the high temporal velocity of tweets and low reporting delays offer particular attractions. Of particular note are the potential for such classification methods to produce time series to inform health system modeling consuming high-velocity data, such as achieved strong use during the COVID-19 pandemic for situational awareness and short-term forecasting in Saskatchewan and for each Canadian province and Canadian Reserves (<xref ref-type="bibr" rid="B46">46</xref>).</p>
</sec>
<sec id="s5" sec-type="conclusions"><label>5.</label><title>Conclusion</title>
<p>Leveraging domain-specific language models to monitor users&#x2019; reports of health experiences related to influenza and COVID-19 infections on social media offers the potential to supplement real-time disease surveillance. The BERT family of pretrained language models offer competitive classification performance across multiple measures, with CT-BERT supporting modestly higher performance for COVID-related tweets than other language models, but performing in an undistinguished manner when classifying influenza-related tweets. Despite a number of limitations, such methods demonstrate sufficient promise to merit consideration as one element of the broader communicable disease surveillance toolbox increasingly important for adaptive disease prevention and control in the future.</p>
</sec>
</body>
<back>
<sec id="s6" sec-type="data-availability"><title>Data availability statement</title>
<p>The Twitter datasets used in the study were publicly available at the time of data collection. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://github.com/tianyuan09/InfluenzaCOVID-19TwitterDatasets">https://github.com/tianyuan09/InfluenzaCOVID-19TwitterDatasets</ext-link>.</p>
</sec>
<sec id="s7" sec-type="ethics-statement"><title>Ethics statement</title>
<p>No ethics review was sought because the data used in this study was obtained from publicly available social media platform Twitter and we did not conduct any experiments on human subjects.</p>
</sec>
<sec id="s8" sec-type="author-contributions"><title>Author contributions</title>
<p>NO and YT conceptualized the study. NO and YT collected the data, performed the data filtering, and designed the annotation task. LD, WM, YT and WZ contributed to the annotation task. YT and LD cleaned the tweets. WZ and YT performed the data preprocessing, trained and evaluated the language models. WZ and YT analyzed the results and drafted the manuscript. All authors critically revised and approved the final version of the manuscript.</p>
</sec>
<sec id="s9" sec-type="funding-information"><title>Funding</title>
<p>NO wishes to express his appreciation of support via NSERC via the Discovery Grants program (RGPIN 2017-04647).</p>
</sec>
<sec id="s10" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<ack><title>Acknowledgments</title>
<p>We gratefully acknowledge the following individuals for their contributions to the annotation task: Rifat Zahan, Paul Lamp, Iman Jamali, Jennafer Mee (Neiser), Bryce Keeler, Allen McLean, Vyom Patel, and Jalen Mikuliak.</p>
</ack>
<sec id="s11" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s12" sec-type="supplementary-material"><title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fdgth.2023.1203874/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fdgth.2023.1203874/full&#x0023;supplementary-material</ext-link></p>
<supplementary-material id="SD1" content-type="local-data">
<media mimetype="application" mime-subtype="pdf" xlink:href="Table1.pdf"/>
</supplementary-material>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><citation citation-type="other"><collab>[Dataset] Public Health Agency of Canada</collab>. <comment><italic>About FluWatch - canada.ca</italic> (2014). Available from: <ext-link ext-link-type="uri" xlink:href="https://www.canada.ca/en/public-health/services/diseases/flu-influenza/influenza-surveillance/about-fluwatch.html">https://www.canada.ca/en/public-health/services/diseases/flu-influenza/influenza-surveillance/about-fluwatch.html</ext-link> (Accessed April 11, 2018)</comment>.</citation></ref>
<ref id="B2"><label>2.</label><citation citation-type="other"><collab>[Dataset] Public Health Agency of Canada</collab>. <comment><italic>National Surveillance for Coronavirus Disease (COVID-19)</italic> (2020). Available from: <ext-link ext-link-type="uri" xlink:href="https://www.canada.ca/en/public-health/services/diseases/2019-novel-coronavirus-infection/health-professionals/interim-guidance-surveillance-human-infection.html">https://www.canada.ca/en/public-health/services/diseases/2019-novel-coronavirus-infection/health-professionals/interim-guidance-surveillance-human-infection.html</ext-link> (Accessed March 25, 2023)</comment>.</citation></ref>
<ref id="B3"><label>3.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aiello</surname><given-names>AE</given-names></name><name><surname>Renson</surname><given-names>A</given-names></name><name><surname>Zivich</surname><given-names>PN</given-names></name></person-group>. <article-title>Social media-, internet-based disease surveillance for public health</article-title>. <source>Annu Rev Public Health</source>. (<year>2020</year>) <volume>41</volume>:<fpage>101</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1146/annurev-publhealth-040119-094402</pub-id><pub-id pub-id-type="pmid">31905322</pub-id></citation></ref>
<ref id="B4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname><given-names>Y</given-names></name><name><surname>Ge</surname><given-names>Y</given-names></name><name><surname>Yang</surname><given-names>YC</given-names></name><name><surname>Al-Garadi</surname><given-names>MA</given-names></name><name><surname>Sarker</surname><given-names>A</given-names></name></person-group>. <article-title>Comparison of pretraining models, strategies for Health-Related social media text classification</article-title>. <source>Healthcare (Basel)</source>. (<year>2022</year>) <volume>10</volume>. 1478. <pub-id pub-id-type="doi">10.3390/healthcare10081478</pub-id></citation></ref>
<ref id="B5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botz</surname><given-names>J</given-names></name><name><surname>Wang</surname><given-names>D</given-names></name><name><surname>Lambert</surname><given-names>N</given-names></name><name><surname>Wagner</surname><given-names>N</given-names></name><name><surname>G&#x00E9;nin</surname><given-names>M</given-names></name><name><surname>Thommes</surname><given-names>E</given-names></name><etal/></person-group> <article-title>Modeling approaches for early warning, monitoring of pandemic situations as well as decision support</article-title>. <source>Front Public Health</source>. (<year>2022</year>) <volume>10</volume>:<fpage>994949</fpage>. <pub-id pub-id-type="doi">10.3389/fpubh.2022.994949</pub-id><pub-id pub-id-type="pmid">36452960</pub-id></citation></ref>
<ref id="B6"><label>6.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Byrd</surname><given-names>K</given-names></name><name><surname>Mansurov</surname><given-names>A</given-names></name><name><surname>Baysal</surname><given-names>O</given-names></name></person-group>. <comment>Mining twitter data for influenza detection, surveillance. In: <italic>2016 IEEE/ACM International Workshop on Software Engineering in Healthcare Systems (SEHS)</italic> (2016). p. 43&#x2013;49</comment>.</citation></ref>
<ref id="B7"><label>7.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wakamiya</surname><given-names>S</given-names></name><name><surname>Kawai</surname><given-names>Y</given-names></name><name><surname>Aramaki</surname><given-names>E</given-names></name></person-group>. <article-title>Twitter-based influenza detection after flu peak via tweets with indirect information: text mining study</article-title>. <source>JMIR Public Health Surveill</source>. (<year>2018</year>) <volume>4</volume>:<fpage>e65</fpage>. <pub-id pub-id-type="doi">10.2196/publichealth.8627</pub-id><pub-id pub-id-type="pmid">30274968</pub-id></citation></ref>
<ref id="B8"><label>8.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>Y</given-names></name><name><surname>Whitfield</surname><given-names>C</given-names></name><name><surname>Zhang</surname><given-names>T</given-names></name><name><surname>Hauser</surname><given-names>A</given-names></name><name><surname>Reynolds</surname><given-names>T</given-names></name><name><surname>Anwar</surname><given-names>M</given-names></name></person-group>. <article-title>Monitoring COVID-19 pandemic through the lens of social media using natural language processing and machine learning</article-title>. <source>Health Inf Sci Syst</source>. (<year>2021</year>) <volume>9</volume>:<fpage>25</fpage>. <pub-id pub-id-type="doi">10.1007/s13755-021-00158-4</pub-id><pub-id pub-id-type="pmid">34188896</pub-id></citation></ref>
<ref id="B9"><label>9.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alsudias</surname><given-names>L</given-names></name><name><surname>Rayson</surname><given-names>P</given-names></name></person-group>. <article-title>Social media monitoring of the COVID-19 pandemic and influenza epidemic with adaptation for informal language in arabic twitter data: Qualitative study</article-title>. <source>JMIR Med Inform</source>. (<year>2021</year>) <volume>9</volume>:<fpage>e27670</fpage>. <pub-id pub-id-type="doi">10.2196/27670</pub-id><pub-id pub-id-type="pmid">34346892</pub-id></citation></ref>
<ref id="B10"><label>10.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Lin</surname><given-names>HY</given-names></name><name><surname>Moh</surname><given-names>TS</given-names></name></person-group>. <comment>Sentiment analysis on COVID tweets using COVID-Twitter-BERT with auxiliary sentence approach. In: <italic>Proceedings of the 2021 ACM Southeast Conference</italic>. New York, NY, USA: Association for Computing Machinery (2021). ACM SE &#x2019;21, p. 234&#x2013;238</comment>.</citation></ref>
<ref id="B11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chew</surname><given-names>C</given-names></name><name><surname>Eysenbach</surname><given-names>G</given-names></name></person-group>. <article-title>Pandemics in the age of twitter: content analysis of tweets during the 2009 H1N1 outbreak</article-title>. <source>PLoS ONE</source>. (<year>2010</year>) <volume>5</volume>:<fpage>e14118</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0014118</pub-id><pub-id pub-id-type="pmid">21124761</pub-id></citation></ref>
<ref id="B12"><label>12.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jang</surname><given-names>H</given-names></name><name><surname>Rempel</surname><given-names>E</given-names></name><name><surname>Roth</surname><given-names>D</given-names></name><name><surname>Carenini</surname><given-names>G</given-names></name><name><surname>Janjua</surname><given-names>NZ</given-names></name></person-group>. <article-title>Tracking COVID-19 discourse on twitter in north america: Infodemiology study using topic modeling and Aspect-Based sentiment analysis</article-title>. <source>J Med Internet Res</source>. (<year>2021</year>) <volume>23</volume>:<fpage>e25431</fpage>. <pub-id pub-id-type="doi">10.2196/25431</pub-id><pub-id pub-id-type="pmid">33497352</pub-id></citation></ref>
<ref id="B13"><label>13.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ridhwan</surname><given-names>KM</given-names></name><name><surname>Hargreaves</surname><given-names>CA</given-names></name></person-group>. <article-title>Leveraging twitter data to understand public sentiment for the COVID-19 outbreak in singapore</article-title>. <source>Int J Inf Manag Data Insights</source>. (<year>2021</year>) 1:100021. <pub-id pub-id-type="doi">10.1016/j.jjimei.2021.100021</pub-id></citation></ref>
<ref id="B14"><label>14.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Boon-Itt</surname><given-names>S</given-names></name><name><surname>Skunkan</surname><given-names>Y</given-names></name></person-group>. <article-title>Public perception of the COVID-19 pandemic on twitter: Sentiment analysis, topic modeling study</article-title>. <source>JMIR Public Health Surveill</source>. (<year>2020</year>) <volume>6</volume>:<fpage>e21978</fpage>. <pub-id pub-id-type="doi">10.2196/21978</pub-id><pub-id pub-id-type="pmid">33108310</pub-id></citation></ref>
<ref id="B15"><label>15.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aljameel</surname><given-names>SS</given-names></name><name><surname>Alabbad</surname><given-names>DA</given-names></name><name><surname>Alzahrani</surname><given-names>NA</given-names></name><name><surname>Alqarni</surname><given-names>SM</given-names></name><name><surname>Alamoudi</surname><given-names>FA</given-names></name><name><surname>Babili</surname><given-names>LM</given-names></name><etal/></person-group> <article-title>A sentiment analysis approach to predict an individual&#x2019;s awareness of the precautionary procedures to prevent COVID-19 outbreaks in saudi arabia</article-title>. <source>Int J Environ Res Public Health</source>. (<year>2020</year>) <volume>18</volume>:218. <pub-id pub-id-type="doi">10.3390/ijerph18010218</pub-id><pub-id pub-id-type="pmid">33396713</pub-id></citation></ref>
<ref id="B16"><label>16.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Sitaula</surname><given-names>C</given-names></name><name><surname>Shahi</surname><given-names>TB</given-names></name></person-group>. <comment>Multi-channel CNN to classify nepali covid-19 related tweets using hybrid features</comment> (<year>2022</year>) <italic>arXiv [cs.CL]</italic>. Available at: http://arxiv.org/abs/2203.10286.</citation></ref>
<ref id="B17"><label>17.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shahi</surname><given-names>TB</given-names></name><name><surname>Sitaula</surname><given-names>C</given-names></name><name><surname>Paudel</surname><given-names>N</given-names></name></person-group>. <article-title>A hybrid feature extraction method for nepali COVID-19-Related tweets classification</article-title>. <source>Comput Intell Neurosci</source>. (<year>2022</year>) <volume>2022</volume>:<fpage>5681574</fpage>. <pub-id pub-id-type="doi">10.1155/2022/5681574</pub-id><pub-id pub-id-type="pmid">35281187</pub-id></citation></ref>
<ref id="B18"><label>18.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sitaula</surname><given-names>C</given-names></name><name><surname>Basnet</surname><given-names>A</given-names></name><name><surname>Mainali</surname><given-names>A</given-names></name><name><surname>Shahi</surname><given-names>TB</given-names></name></person-group>. <article-title>Deep learning-based methods for sentiment analysis on nepali COVID-19-related tweets</article-title>. <source>Comput Intell Neurosci</source>. (<year>2021</year>) <volume>2021</volume>:<fpage>2158184</fpage>. <pub-id pub-id-type="doi">10.1155/2021/2158184</pub-id><pub-id pub-id-type="pmid">34737773</pub-id></citation></ref>
<ref id="B19"><label>19.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mackey</surname><given-names>T</given-names></name><name><surname>Purushothaman</surname><given-names>V</given-names></name><name><surname>Li</surname><given-names>J</given-names></name><name><surname>Shah</surname><given-names>N</given-names></name><name><surname>Nali</surname><given-names>M</given-names></name><name><surname>Bardier</surname><given-names>C</given-names></name><etal/></person-group> <article-title>Machine learning to detect self-reporting of symptoms, testing access,, recovery associated with COVID-19 on twitter: retrospective big data infoveillance study</article-title>. <source>JMIR Public Health Surveill</source>. (<year>2020</year>) <volume>6</volume>:<fpage>e19509</fpage>. <pub-id pub-id-type="doi">10.2196/19509</pub-id><pub-id pub-id-type="pmid">32490846</pub-id></citation></ref>
<ref id="B20"><label>20.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Golder</surname><given-names>S</given-names></name><name><surname>Klein</surname><given-names>AZ</given-names></name><name><surname>Magge</surname><given-names>A</given-names></name><name><surname>O&#x2019;Connor</surname><given-names>K</given-names></name><name><surname>Cai</surname><given-names>H</given-names></name><name><surname>Weissenbacher</surname><given-names>D</given-names></name><etal/></person-group> <article-title>A chronological and geographical analysis of personal reports of COVID-19 on twitter from the UK</article-title>. <source>Digit Health</source>. (<year>2022</year>) <volume>53</volume>(<issue>2</issue>):<fpage>311</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1017/S0008423920000311</pub-id></citation></ref>
<ref id="B21"><label>21.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Klein</surname><given-names>AZ</given-names></name><name><surname>Magge</surname><given-names>A</given-names></name><name><surname>O&#x2019;Connor</surname><given-names>KMS</given-names></name><name><surname>Cai</surname><given-names>H</given-names></name><name><surname>Weissenbacher</surname><given-names>D</given-names></name><name><surname>Gonzalez-Hernandez</surname><given-names>G</given-names></name></person-group>. <comment>A chronological, geographical analysis of personal reports of COVID-19 on twitter. <italic>medRxiv</italic> (2020)</comment>.</citation></ref>
<ref id="B22"><label>22.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Devlin</surname><given-names>J</given-names></name><name><surname>Chang</surname><given-names>MW</given-names></name><name><surname>Lee</surname><given-names>K</given-names></name><name><surname>Toutanova</surname><given-names>K</given-names></name></person-group>. <comment><italic>Bert: Pre-training of deep bidirectional transformers for language understanding</italic>. arXiv preprint arXiv:1810.04805 (2018)</comment>.</citation></ref>
<ref id="B23"><label>23.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>Y</given-names></name><name><surname>Ott</surname><given-names>M</given-names></name><name><surname>Goyal</surname><given-names>N</given-names></name><name><surname>Du</surname><given-names>J</given-names></name><name><surname>Joshi</surname><given-names>M</given-names></name><name><surname>Chen</surname><given-names>D</given-names></name><etal/></person-group> <comment><italic>Roberta: A robustly optimized bert pretraining approach</italic>. arXiv preprint arXiv:1907.11692 (2019)</comment>.</citation></ref>
<ref id="B24"><label>24.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Radford</surname><given-names>A</given-names></name><name><surname>Narasimhan</surname><given-names>K</given-names></name><name><surname>Salimans</surname><given-names>T</given-names></name><name><surname>Sutskever</surname><given-names>I</given-names></name></person-group>. <comment>Improving language understanding by generative pre-training (2018)</comment>.</citation></ref>
<ref id="B25"><label>25.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname><given-names>Z</given-names></name><name><surname>Dai</surname><given-names>Z</given-names></name><name><surname>Yang</surname><given-names>Y</given-names></name><name><surname>Carbonell</surname><given-names>J</given-names></name><name><surname>Salakhutdinov</surname><given-names>RR</given-names></name><name><surname>Le</surname><given-names>QV</given-names></name></person-group>. <article-title>Xlnet: Generalized autoregressive pretraining for language understanding</article-title>. <collab><italic>In Proceedings of the 33rd International Conference on Neural Information Processing Systems</italic></collab>. <comment>Red Hook, NY: Curran Associates Inc. (2019). p. 5753 -63. Available at:</comment> <ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/10.5555/3454287.3454804">https://dl.acm.org/doi/10.5555/3454287.3454804</ext-link></citation></ref>
<ref id="B26"><label>26.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kalyan</surname><given-names>KS</given-names></name><name><surname>Rajasekharan</surname><given-names>A</given-names></name><name><surname>Sangeetha</surname><given-names>S</given-names></name></person-group>. <article-title>AMMU: A survey of transformer-based biomedical pretrained language models</article-title>. <source>J Biomed Inform</source>. (<year>2022</year>) <volume>126</volume>:<fpage>103982</fpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2021.103982</pub-id><pub-id pub-id-type="pmid">34974190</pub-id></citation></ref>
<ref id="B27"><label>27.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matharaarachchi</surname><given-names>S</given-names></name><name><surname>Domaratzki</surname><given-names>M</given-names></name><name><surname>Katz</surname><given-names>A</given-names></name><name><surname>Muthukumarana</surname><given-names>S</given-names></name></person-group>. <article-title>Discovering long COVID symptom patterns: association rule mining and sentiment analysis in social media tweets</article-title>. <source>JMIR Form Res</source>. (<year>2022</year>) <volume>6</volume>:<fpage>e37984</fpage>. <pub-id pub-id-type="doi">10.2196/37984</pub-id><pub-id pub-id-type="pmid">36069846</pub-id></citation></ref>
<ref id="B28"><label>28.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Baker</surname><given-names>W</given-names></name><name><surname>Colditz</surname><given-names>JB</given-names></name><name><surname>Dobbs</surname><given-names>PD</given-names></name><name><surname>Mai</surname><given-names>H</given-names></name><name><surname>Visweswaran</surname><given-names>S</given-names></name><name><surname>Zhan</surname><given-names>J</given-names></name><etal/></person-group> <article-title>Classification of twitter vaping discourse using BERTweet: comparative deep learning study</article-title>. <source>JMIR Med Inform</source>. (<year>2022</year>) <volume>10</volume>:<fpage>e33678</fpage>. <pub-id pub-id-type="doi">10.2196/33678</pub-id><pub-id pub-id-type="pmid">35862172</pub-id></citation></ref>
<ref id="B29"><label>29.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kothuru</surname><given-names>S</given-names></name><name><surname>Santhanavijayan</surname><given-names>A</given-names></name></person-group>. <article-title>Identifying covid-19 english informative tweets using limited labelled data</article-title>. <source>Soc Netw Anal Min</source>. (<year>2023</year>) <volume>13</volume>:<fpage>25</fpage>. <pub-id pub-id-type="doi">10.1007/s13278-023-01025-8</pub-id><pub-id pub-id-type="pmid">36686376</pub-id></citation></ref>
<ref id="B30"><label>30.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Kaufmann</surname><given-names>M</given-names></name><name><surname>Kalita</surname><given-names>J</given-names></name></person-group>. <comment>Syntactic normalization of twitter messages. In: <italic>International Conference on Natural Language Processing, Kharagpur, India</italic> (2010). vol. 16</comment>.</citation></ref>
<ref id="B31"><label>31.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>&#x00D6;zberk</surname><given-names>A</given-names></name><name><surname>&#x00C7;i&#x00E7;ekli</surname><given-names>&#x0130;</given-names></name></person-group>. <comment>Offensive language detection in turkish tweets with bert models. In: <italic>2021 6th International Conference on Computer Science and Engineering (UBMK)</italic> (IEEE) (2021). p. 517&#x2013;521</comment>.</citation></ref>
<ref id="B32"><label>32.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Wolf</surname><given-names>T</given-names></name><name><surname>Debut</surname><given-names>L</given-names></name><name><surname>Sanh</surname><given-names>V</given-names></name><name><surname>Chaumond</surname><given-names>J</given-names></name><name><surname>Delangue</surname><given-names>C</given-names></name><name><surname>Moi</surname><given-names>A</given-names></name><etal/></person-group> <comment>Transformers: state-of-the-art natural language processing. In: <italic>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</italic> (Online: Association for Computational Linguistics) (2020). p. 38&#x2013;45</comment>.</citation></ref>
<ref id="B33"><label>33.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lema&#x00EE;tre</surname><given-names>G</given-names></name><name><surname>Nogueira</surname><given-names>F</given-names></name><name><surname>Aridas</surname><given-names>CK</given-names></name></person-group>. <article-title>Imbalanced-learn: a python toolbox to tackle the curse of imbalanced datasets in machine learning</article-title>. <source>J Mach Learn Res</source>. (<year>2017</year>) <volume>18</volume>:<fpage>1</fpage>&#x2013;<lpage>5</lpage>. Available at: <ext-link ext-link-type="uri" xlink:href="http://jmlr.org/papers/v18/16-365.html">http://jmlr.org/papers/v18/16-365.html</ext-link></citation></ref>
<ref id="B34"><label>34.</label><citation citation-type="other"><collab>[Dataset] Alphabet Inc</collab>. <comment><italic>Google colaboratory</italic> (2022). Available from: <ext-link ext-link-type="uri" xlink:href="https://colab.research.google.com/">https://colab.research.google.com/</ext-link> (Accessed April 4, 2023)</comment>.</citation></ref>
<ref id="B35"><label>35.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vaswani</surname><given-names>A</given-names></name><name><surname>Shazeer</surname><given-names>N</given-names></name><name><surname>Parmar</surname><given-names>N</given-names></name><name><surname>Uszkoreit</surname><given-names>J</given-names></name><name><surname>Jones</surname><given-names>L</given-names></name><name><surname>Gomez</surname><given-names>AN</given-names></name><etal/></person-group> <article-title>Attention is all you need</article-title>. <collab><italic>In: Proceedings of the 31st International Conference on Neural Information Processing Systems NIPS&#x0027;17</italic></collab>. <comment>Red Hook, NY: Curran Associates Inc. (2017). p. 6000 -10. Available at:</comment> <ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/10.5555/3295222.3295349">https://dl.acm.org/doi/10.5555/3295222.3295349</ext-link></citation></ref>
<ref id="B36"><label>36.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Nguyen</surname><given-names>DQ</given-names></name><name><surname>Vu</surname><given-names>T</given-names></name><name><surname>Nguyen</surname><given-names>AT</given-names></name></person-group>. <comment>BERTweet: A pre-trained language model for English Tweets. In: <italic>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</italic> (2020). p. 9&#x2013;14</comment></citation></ref>
<ref id="B37"><label>37.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>M&#x00FC;ller</surname><given-names>M</given-names></name><name><surname>Salath&#x00E9;</surname><given-names>M</given-names></name><name><surname>Kummervold</surname><given-names>PE</given-names></name></person-group>. <article-title>COVID-Twitter-BERT: A natural language processing model to analyse COVID-19 content on twitter</article-title>. <source>Front Artif Intell</source>. (<year>2023</year>) <volume>6</volume>:<fpage>1023281</fpage>. <pub-id pub-id-type="doi">10.3389/frai.2023.1023281</pub-id></citation></ref>
<ref id="B38"><label>38.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Mapes</surname><given-names>N</given-names></name><name><surname>White</surname><given-names>A</given-names></name><name><surname>Medury</surname><given-names>R</given-names></name><name><surname>Dua</surname><given-names>S</given-names></name></person-group>. <comment>Divisive language and propaganda detection using multi-head attention transformers with deep learning bert-based language models for binary classification. In: <italic>Proceedings of the Second Workshop on Natural Language Processing for Internet Freedom: Censorship, Disinformation, and Propaganda</italic> (2019). p. 103&#x2013;106</comment>.</citation></ref>
<ref id="B39"><label>39.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Kingma</surname><given-names>DP</given-names></name><name><surname>Ba</surname><given-names>J</given-names></name></person-group>. <comment>Adam: A method for stochastic optimization (2014)</comment>.</citation></ref>
<ref id="B40"><label>40.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Mosbach</surname><given-names>M</given-names></name><name><surname>Andriushchenko</surname><given-names>M</given-names></name><name><surname>Klakow</surname><given-names>D</given-names></name></person-group>. <comment><italic>On the stability of fine-tuning bert: misconceptions, explanations, and strong baselines</italic>. arXiv preprint arXiv:2006.04884 (2020)</comment>.</citation></ref>
<ref id="B41"><label>41.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tang</surname><given-names>L</given-names></name><name><surname>Douglas</surname><given-names>S</given-names></name><name><surname>Laila</surname><given-names>A</given-names></name></person-group>. <article-title>Among sheeples and antivaxxers: Social media responses to COVID-19 vaccine news posted by canadian news organizations, and recommendations to counter vaccine hesitancy</article-title>. <source>Can Commun Dis Rep</source>. (<year>2021</year>) <volume>47</volume>:<fpage>524</fpage>&#x2013;<lpage>33</lpage>. <pub-id pub-id-type="doi">10.14745/ccdr.v47i12a03</pub-id><pub-id pub-id-type="pmid">35018140</pub-id></citation></ref>
<ref id="B42"><label>42.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Merkley</surname><given-names>E</given-names></name><name><surname>Bridgman</surname><given-names>A</given-names></name><name><surname>Loewen</surname><given-names>PJ</given-names></name><name><surname>Owen</surname><given-names>T</given-names></name><name><surname>Ruths</surname><given-names>D</given-names></name><name><surname>Zhilin</surname><given-names>O</given-names></name></person-group>. <article-title>A rare moment of cross-partisan consensus: elite and public response to the COVID-19 pandemic in Canada</article-title>. <source>Can J Polit Sci</source>. (<year>2020</year>) <volume>53</volume>(<issue>2</issue>):<fpage>311</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1017/S0008423920000311</pub-id></citation></ref>
<ref id="B43"><label>43.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Al-Rawi</surname><given-names>A</given-names></name><name><surname>Siddiqi</surname><given-names>M</given-names></name><name><surname>Morgan</surname><given-names>R</given-names></name><name><surname>Vandan</surname><given-names>N</given-names></name><name><surname>Smith</surname><given-names>J</given-names></name><name><surname>Wenham</surname><given-names>C</given-names></name></person-group>. <article-title>COVID-19 and the gendered use of emojis on twitter: infodemiology study</article-title>. <source>J Med Internet Res</source>. (<year>2020</year>) <volume>22</volume>:<fpage>e21646</fpage>. <pub-id pub-id-type="doi">10.2196/21646</pub-id><pub-id pub-id-type="pmid">33052871</pub-id></citation></ref>
<ref id="B44"><label>44.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname><given-names>Q</given-names></name><name><surname>Liu</surname><given-names>J</given-names></name><name><surname>Liu</surname><given-names>Q</given-names></name><name><surname>Kang</surname><given-names>L</given-names></name><name><surname>Liu</surname><given-names>R</given-names></name><name><surname>Jing</surname><given-names>W</given-names></name><etal/></person-group> <article-title>Global percentage of asymptomatic SARS-CoV-2 infections among the tested population and individuals with confirmed COVID-19 diagnosis: a systematic review and meta-analysis</article-title>. <source>JAMA Netw Open</source>. (<year>2021</year>) <volume>4</volume>:<fpage>e2137257</fpage>. <pub-id pub-id-type="doi">10.1001/jamanetworkopen.2021.37257</pub-id><pub-id pub-id-type="pmid">34905008</pub-id></citation></ref>
<ref id="B45"><label>45.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kumar</surname><given-names>M</given-names></name><name><surname>Mohapatra</surname><given-names>S</given-names></name><name><surname>Mazumder</surname><given-names>P</given-names></name><name><surname>Singh</surname><given-names>A</given-names></name><name><surname>Honda</surname><given-names>R</given-names></name><name><surname>Lin</surname><given-names>C</given-names></name><etal/></person-group> <article-title>Making waves perspectives of modelling and monitoring of SARS-CoV-2 in aquatic environment for COVID-19 pandemic</article-title>. <source>Curr Pollut Rep</source>. (<year>2020</year>) <volume>6</volume>:<fpage>468</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1007/s40726-020-00161-5</pub-id><pub-id pub-id-type="pmid">32953402</pub-id></citation></ref>
<ref id="B46"><label>46.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>X</given-names></name><name><surname>Patel</surname><given-names>V</given-names></name><name><surname>Duan</surname><given-names>L</given-names></name><name><surname>Mikuliak</surname><given-names>J</given-names></name><name><surname>Basran</surname><given-names>J</given-names></name><name><surname>Osgood</surname><given-names>N.</given-names></name></person-group> <article-title>Real-time epidemiology and acute care need monitoring and forecasting for COVID-19 via Bayesian Sequential Monte Carlo-leveraged transmission models</article-title>. <source>Preprints</source>. (<year>2023</year>) <pub-id pub-id-type="doi">10.20944/preprints202302.0078.v1</pub-id></citation></ref></ref-list>
</back>
</article>