<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2024.1282043</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Digital Health</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Word sense disambiguation of acronyms in clinical narratives</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Chopard</surname><given-names>Daphn&#x00E9;</given-names></name><role content-type="https://credit.niso.org/contributor-roles/methodology/"/><role content-type="https://credit.niso.org/contributor-roles/validation/"/><role content-type="https://credit.niso.org/contributor-roles/data-curation/"/><role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/><role content-type="https://credit.niso.org/contributor-roles/investigation/"/><role content-type="https://credit.niso.org/contributor-roles/software/"/><role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/></contrib>
<contrib contrib-type="author"><name><surname>Corcoran</surname><given-names>Padraig</given-names></name><role content-type="https://credit.niso.org/contributor-roles/supervision/"/><role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/></contrib>
<contrib contrib-type="author" corresp="yes"><name><surname>Spasi&#x0107;</surname><given-names>Irena</given-names></name>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref><uri xlink:href="https://loop.frontiersin.org/people/2417985/overview" /><role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/><role content-type="https://credit.niso.org/contributor-roles/methodology/"/><role content-type="https://credit.niso.org/contributor-roles/supervision/"/><role content-type="https://credit.niso.org/contributor-roles/validation/"/><role content-type="https://credit.niso.org/contributor-roles/visualization/"/><role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/></contrib>
</contrib-group>
<aff><institution>School of Computer Science and Informatics, Cardiff University</institution>, <addr-line>Cardiff</addr-line>, <country>United Kingdom</country></aff>
<author-notes>
<fn fn-type="edited-by"><p><bold>Edited by:</bold> Uwe Aickelin, The University of Melbourne, Australia</p></fn>
<fn fn-type="edited-by"><p><bold>Reviewed by:</bold> Sheng-Chieh Lu, University of Texas MD Anderson Cancer Center, United States</p>
<p>Laura Falaschetti, Marche Polytechnic University, Italy</p></fn>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Irena Spasi&#x0107; <email>spasici@cardiff.ac.uk</email></corresp>
</author-notes>
<pub-date pub-type="epub"><day>28</day><month>02</month><year>2024</year></pub-date>
<pub-date pub-type="collection"><year>2024</year></pub-date>
<volume>6</volume><elocation-id>1282043</elocation-id>
<history>
<date date-type="received"><day>23</day><month>08</month><year>2023</year></date>
<date date-type="accepted"><day>16</day><month>02</month><year>2024</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2024 Chopard, Corcoran and Spasi&#x0107;.</copyright-statement>
<copyright-year>2024</copyright-year><copyright-holder>Chopard, Corcoran and Spasi&#x0107;</copyright-holder><license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Clinical narratives commonly use acronyms without explicitly defining their long forms. This makes it difficult to automatically interpret their sense as acronyms tend to be highly ambiguous. Supervised learning approaches to their disambiguation in the clinical domain are hindered by issues associated with patient privacy and manual annotation, which limit the size and diversity of training data. In this study, we demonstrate how scientific abstracts can be utilised to overcome these issues by creating a large automatically annotated dataset of artificially simulated global acronyms. A neural network trained on such a dataset achieved the F1-score of 95&#x0025; on disambiguation of acronym mentions in scientific abstracts. This network was integrated with multi-word term recognition to extract a sense inventory of acronyms from a corpus of clinical narratives on the fly. Acronym sense extraction achieved the F1-score of 74&#x0025; on a corpus of radiology reports. In clinical practice, the suggested approach can be used to facilitate development of institution-specific inventories.</p>
</abstract>
<kwd-group>
<kwd>natural language processing</kwd>
<kwd>word sense disambiguation</kwd>
<kwd>acronym disambiguation</kwd>
<kwd>machine learning</kwd>
<kwd>deep learning</kwd>
<kwd>silver standard</kwd>
</kwd-group><counts>
<fig-count count="3"/>
<table-count count="4"/><equation-count count="2"/><ref-count count="48"/><page-count count="0"/><word-count count="0"/></counts><custom-meta-wrap><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Health Informatics</meta-value></custom-meta></custom-meta-wrap>
</article-meta>
</front>
<body><sec id="s1" sec-type="intro"><label>1</label><title>Introduction</title>
<p>Acronyms are formed as systematic abbreviations of frequently mentioned words and phrases, which follow special capitalisation and blending patterns (<xref ref-type="bibr" rid="B1">1</xref>). Their primary purpose is to make written communication more efficient in terms of space and time. The readers&#x2019; familiarity with specific acronyms rather than their orthographic regularity plays a critical role in rapidly interpreting their intended meaning (<xref ref-type="bibr" rid="B2">2</xref>). Therefore, not only do the globally accepted acronyms pose no major difficulties in specialist communication, but they in fact facilitate such communication between domain experts. For example, clinical narratives feature extensive use of acronyms, which are not defined explicitly in the corresponding documents.</p>
<p>Even though such acronyms pose no difficulty to their intended users, they can still hinder the performance of natural language processing (NLP) algorithms (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B4">4</xref>) when clinical narratives need to be analysed automatically. For example, the use of acronyms (e.g. &#x201C;DVT&#x201D;) obscures the corresponding phrase (e.g. &#x201C;deep vein thrombosis&#x201D;) whose constituents (e.g. &#x201C;thrombosis&#x201D;) cannot be indexed and, hence, cannot be retrieved. On the other hand, clinical acronyms are highly polysemous (<xref ref-type="bibr" rid="B5">5</xref>), e.g. &#x201C;ED&#x201D; can be interpreted as &#x201C;eating disorder,&#x201D; &#x201C;elbow disarticulation,&#x201D; &#x201C;emotional disorder,&#x201D; &#x201C;emergency department&#x201D; or &#x201C;erectile dysfunction,&#x201D; which may cause irrelevant documents to be retrieved when using the acronym as a search term. These problems can be resolved by automatically mapping acronyms to their correct senses in an external dictionary (<xref ref-type="bibr" rid="B6">6</xref>) based on their context of use.</p>
<p>This may be viewed as a word sense disambiguation (WSD) problem (<xref ref-type="bibr" rid="B7">7</xref>), commonly approached by supervised learning approaches, which are trained using a set of annotated examples. Their performance depends largely on the amount of annotated data used for training. However, the data annotation bottleneck presents one of the key obstacles to supervised learning approaches in clinical NLP (<xref ref-type="bibr" rid="B8">8</xref>). Patient privacy concerns further narrow down this bottleneck by removing the possibility of crowdsourcing annotation. Crowdsourcing remains an option for annotating synthetic data, but it is difficult to scale when medical expertise is required for accurate annotation.</p>
<p>To eliminate the manual data annotation bottleneck altogether, we looked at the possibility of generating an annotated dataset automatically. Namely, scientific writing conventions prescribe that all acronyms need to be explicitly defined the first time they are mentioned. Local acronym definitions can be removed from text and used instead as sense labels. By doing so, we can simulate the clinical narrative style of acronym usage and create a large dataset that can be used to train supervised approaches to WSD of clinical acronyms. This leads us to the main contribution of this study. We describe an algorithm for WSD of clinical acronyms, which offers two key novelties. First, it requires no manual annotation of data, be them clinical or otherwise. Second, it requires no predefined sense inventory and instead uses the corpus itself to extract potential senses on the fly.</p>
</sec>
<sec id="s2"><label>2</label><title>Related work</title>
<p>Acronym-related NLP tasks can be divided into three groups (<xref ref-type="bibr" rid="B9">9</xref>). Acronym definition identication is concerned with finding pairs of acronyms and their long forms that are defined in the input text. Similarly to named entity recognition, acronym identification takes a text as input and tags the spans of acronyms and long forms using the BIO format (short for beginning, inside, outside). Finally, acronym disambiguation takes a sentence that contains an ambiguous acronym as input and aims to select an appropriate long form from a predefined set of candidates. The first task was proposed long before the other two and has attracted a lot of attention, especially in the biomedical domain, where a variety of rule-based algorithms have been developed with great success (<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>).</p>
<p>Acronym identification and acronym disambiguation tasks were introduced recently at the AAAI Workshop on Scientific Document Understanding (<xref ref-type="bibr" rid="B12">12</xref>). Given the success of neural networks with a transformer architecture in a variety of NLP tasks, it was not surprising that the majority of participants opted for such an approach, which they used to model acronym disambiguation as binary classification (<xref ref-type="bibr" rid="B13">13</xref>, <xref ref-type="bibr" rid="B14">14</xref>), information retrieval (<xref ref-type="bibr" rid="B15">15</xref>) or span prediction (<xref ref-type="bibr" rid="B16">16</xref>). They provided the best performance with F1-scores ranging between 91.58&#x0025; and 94.05&#x0025; lagging slightly behind the human performance of 96.10&#x0025;. Other types of neural networks architectures provided similar performance of 91.34&#x0025; (<xref ref-type="bibr" rid="B17">17</xref>). Traditional machine learning approaches performed at 88.40&#x0025; at best, struggling mostly in terms of recall (<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B19">19</xref>).</p>
<p>Clinical domain, in which almost one third of abbreviations were found to be ambiguous (<xref ref-type="bibr" rid="B20">20</xref>), is very much in need of effective acronym disambiguation methods. Approaches such as the ones described above depend on predefined sense inventories. Unfortunately, the two largest sense inventories for biomedical acronyms, the Specialist Lexicon Release of Abbreviations and Acronyms (LRABR) (<xref ref-type="bibr" rid="B21">21</xref>) and Another Database of Abbreviations in MEDLINE (ADAM) (<xref ref-type="bibr" rid="B22">22</xref>), were shown to have micro-coverage of acronyms in clinical narratives of 74.8&#x0025; and 68.0&#x0025;, respectively. Micro-coverage of long forms was found to be 82.5&#x0025; and 55.4&#x0025;, respectively. Disambiguation cannot be performed when an appropriate sense is missing from the inventory. To address this issue, a method was developed to automatically harmonise sense inventories from different regions and specialties into a comprehensive inventory (<xref ref-type="bibr" rid="B23">23</xref>), which improved micro-coverage of acronyms and their senses to 94.3&#x0025; and 99.6&#x0025;, respectively.</p>
<p>Early approaches to disambiguation of clinical abbreviations and acronyms, which were based on rules (<xref ref-type="bibr" rid="B24">24</xref>) and traditional machine learning (<xref ref-type="bibr" rid="B25">25</xref>&#x2013;<xref ref-type="bibr" rid="B28">28</xref>), performed with moderate accuracy ranging between 50.1&#x0025; and 75&#x0025; (<xref ref-type="bibr" rid="B29">29</xref>). Despite the obvious improvements that deep neural networks offer to acronym disambiguation (<xref ref-type="bibr" rid="B12">12</xref>), there is evidence that rule-based approaches may not only outperform them but error analysis even suggests why (<xref ref-type="bibr" rid="B9">9</xref>). Namely, end-to-end machine learning commonly emphasise certain types over others and by doing so fail to take advantage of orthography, which plays an important role in the formation of acronyms. Deep neural networks can capture orthographic constraints by integrating token-based and character-based networks.</p>
</sec>
<sec id="s3"><label>3</label><title>Material and methods</title>
<sec id="s3a"><label>3.1</label><title>System overview</title>
<p>The overall framework of our acronym disambiguation approach is provided in <xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>. The system has two main modules, one treating WSD as a binary classification problem (shown on the left-hand side) and the other one dedicated to extracting acronyms and their potential senses from text (shown on the right-hand side). Given a pair consisting of an acronym within its context and a potential long form, binary classification is performed to determine whether the long form represents a correct interpretation of the acronym or not. In this framework, the binary classification module is used to resolve the ambiguity of acronyms arising from extracting multiple long form candidates from the corpus.</p>
<fig id="F1" position="float"><label>Figure 1</label>
<caption><p>An acronym disambiguation framework.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-06-1282043-g001.tif"/>
</fig>
<p>The binary classifier is trained on a large corpus of scientific abstracts. This choice of data serves two main purposes. First, it bypasses the privacy concerns associated with clinical data. Second, the scientific writing conventions allow for an easy simulation of global acronyms, whose sense can be annotated automatically. These two factors combined together allow for the creation of a large annotated dataset that can be used to train a deep learning model for disambiguation of global acronyms. To train the model, we fine-tuned Bidirectional Encoder Representations from Transformers (BERT) (<xref ref-type="bibr" rid="B30">30</xref>), a large pretrained language model, to perform binary classification as indicated above.</p>
<p>To link acronyms to their long forms found within clinical narratives, we start by extracting potential acronyms on one side and extracting potential long forms on the other side. These two activities are independent of each other and hence can be performed in parallel. We extract potential acronyms using a simple heuristic based on their orthographic properties. We extract multi-word terms (MWTs) from text based on their linguistic and statistical properties using a method called FlexiTerm (<xref ref-type="bibr" rid="B31">31</xref>). Having retrieved both acronyms and MWTs, they are matched using their internal properties. In a nutshell, the characters from an acronym are aligned against each MWT to shortlist potential long forms.</p>
<p>All candidate pairs of acronyms and their potential long forms are then disambiguated using the previously trained disambiguation model based on the context in which the acronym is mentioned. The final result is an inventory of acronyms mapped to their senses. The inventory is derived directly from the corpus. In that aspect, our method departs from the traditional acronym disambiguation approaches, which rely on an external inventory to obtain a list of possible senses.</p>
<p>The following sections provide further details about each module of the proposed framework.</p>
</sec>
<sec id="s3b"><label>3.2</label><title>Binary classification of acronym senses</title>
<sec id="s3b1"><label>3.2.1</label><title>Data collection and annotation</title>
<p>Data collection and annotation were performed using a web-based application for simulation and annotation of global acronyms (<xref ref-type="bibr" rid="B32">32</xref>), whose long forms are not defined explicitly in text. This application automatically modifies PubMed abstracts to simulate global acronym usage by removing their explicitly defined long forms and using them to annotate their senses without any need for external lexicons or manual annotation. Thus, it can be used to create large training datasets for WSD of biomedical acronyms. Given the automated extraction of acronym definitions, such data will inevitably contain some degree of noise, which based on the performance of the underlying algorithm (<xref ref-type="bibr" rid="B10">10</xref>) is estimated to be around 4&#x0025;.</p>
<p>We started by defining a PubMed search query that targeted the journals that cover clinical applications. We combined the keyword &#x201C;clinical&#x201D; with the suffix &#x201C;-logy&#x201D; to refer to various clinical domains (e.g. &#x201C;rheumatology&#x201D;) while also explicitly excluding certain keywords (e.g. &#x201C;biology&#x201D;). The corresponding abstracts were downloaded from PubMed and annotated automatically. We downloaded the sense inventory, which was also generated automatically by the web-based application. Each acronym associated with a single long form in the sense inventory was discarded as it was unambiguous and thus irrelevant for the disambiguation task at hand. A total of 963 unique acronyms remained, each having between 2 and 27 possible long forms with the mean of 4.09 and the median of 3 long forms.</p>
<p>Given an acronym&#x2019;s potential long form, we choose to tackle WSD as a binary classification task. To properly train such a WSD model, we needed both positive examples, where the long form candidate is correct, and negative examples, where the long form candidate is incorrect. Using the previously annotated acronyms, we proceeded as follows. First, we extracted every sentence containing an annotated acronym. For each sentence, we created two examples: a positive example corresponding to the original triplet (sentence, acronym, correct long form) and a negative example, in which the correct long form was replaced at random by another long form according to the dictionary created in the previous step. In this manner, we obtained a balanced dataset of positive and negative examples, both in the form of a triplet (sentence, acronym, long form).</p>
<p>Prior to training a WSD model, we set aside a portion of the data for validation, i.e. to evaluate the model&#x2019;s fit throughout its development including hyperparameter optimisation. For each of the 963 ambiguous acronyms, a total of 1,000 examples (or <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM1"><mml:mn>10</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></inline-formula> if fewer than 10,000 examples were available) were randomly selected and reserved for validation. We repeated this process on the remainder of the data to obtain a test set. A total of 16,130,782 remaining examples were used for training.</p>
</sec>
<sec id="s3b2"><label>3.2.2</label><title>Training a disambiguation model</title>
<p>Recall that each example is in the form of a triplet (sentence, acronym, long form). The long form needs to be classified to determine whether or not it represents a correct interpretation of the acronym within the given sentence. For this task we used a transformer-based architecture similar to that of (<xref ref-type="bibr" rid="B33">33</xref>). This architecture fine-tunes BERT (<xref ref-type="bibr" rid="B30">30</xref>) to perform WSD given a sentence and one of its ambiguous words. More specifically, the model is fed two text sequences: (i) an ambiguous word within its context and (ii) that word followed by its gloss (one of many) from WordNet (<xref ref-type="bibr" rid="B34">34</xref>). In our case, the word and its gloss correspond to an acronym and its potential long form, respectively as shown in <xref ref-type="fig" rid="F2">Figure&#x00A0;2</xref>. The rest of the BERT-based neural network architecture is identical to the one we previously used for binary classification of potential adverse events (<xref ref-type="bibr" rid="B35">35</xref>), where practical implementation details can be found. In summary, three types of embeddings were used: token embeddings to represent tokens, segment embeddings to differentiate between different segments in the input (see <xref ref-type="fig" rid="F2">Figure&#x00A0;2</xref>) and position embeddings to encode the word order. For each token, these embeddings were added and passed onto to the pretrained BERT&#x2013;BASE model comprised of 12 layers of transformer encoders, each having a hidden size of 768 and 12 attention heads. A token-specific output produced at each layer can be interpreted as its contextualised embedding. The final output produced for the initial [CLS] token, commonly used as an aggregate problem representation, was forwarded to the classification layer, but only after a 0.1 dropout to reduce overfitting. The two logits outputted by the classification layer correspond to the question of whether the given longform is a correct interpretation of the given acronym given its context or not. The weights of the classification layer were learned during BERT fine-tuning. They were initialized using a truncated normal distribution with mean 0 and standard deviation of 0.02. Next, a softmax function was applied to obtain the probability distribution of the two classes. The loss function (softmax cross entropy between the logits and the class labels) was optimized using the Adam optimizer with an initial learning rate of <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM2"><mml:mn>2</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>5</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>. The binary classification model was trained for 8 epochs. All other parameter values were identical to those used in the original BERT<sub>BASE</sub> uncased model, including the clip norm of 1.0 and 100 warmup steps with linear decay of learning rate.</p>
<fig id="F2" position="float"><label>Figure 2</label>
<caption><p>BERT-based representation of the acronym disambiguation problem.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-06-1282043-g002.tif"/>
</fig>
<p>Of note, a similar approach developed by (<xref ref-type="bibr" rid="B13">13</xref>) was the winning solution for the second SDU@AAAI-21 shared task on acronym disambiguation in English scientific papers (<xref ref-type="bibr" rid="B12">12</xref>). They used a different problem representation with the long form as the first text sequence and the acronym within its context as the second sequence. To direct the model&#x2019;s attention to the ambiguous acronym, they marked up the acronym using two special tokens (&#x003C;start&#x003E; and &#x003C;end&#x003E;).</p>
</sec>
</sec>
<sec id="s3c"><label>3.3</label><title>Acronym sense extraction</title>
<p>The purpose of this section is twofold. First, it aims to demonstrate the effectiveness of the acronym disambiguation module trained on a corpus of biomedical abstracts using a corpus of clinical narratives. Second, it extends the functionality of acronym disambiguation to include the extraction of potential acronyms and the corresponding long form candidates automatically from a given corpus.</p>
<sec id="s3c1"><label>3.3.1</label><title>Data collection and annotation</title>
<p>To assemble a corpus of clinical narratives, we used MIMIC-III. This large, freely available database comprises de-identified health-related data associated with over forty thousand patients who stayed in critical care units of the Beth Israel Deaconess Medical Center between 2001 and 2012 (<xref ref-type="bibr" rid="B36">36</xref>). We retrieved a total of 2,609 knee radiology reports. This choice of clinical narratives was motivated by the availability of local expertise needed to interpret the results. Specifically, we had access to the team who developed the TRAK ontology (<xref ref-type="bibr" rid="B37">37</xref>), which defines standard care for the rehabilitation of knee conditions, and who previously applied this ontology to support text mining of knee radiology reports (<xref ref-type="bibr" rid="B38">38</xref>).</p>
<p>The raw reports were processed as follows. We extracted the main body of each report. A set of simple regular expressions were used to recognise section headings and remove them from further consideration. The main reason for discarding section headings was the subsequent recognition of potential acronyms. Section headings are usually written in uppercase, which is one of the properties used to identify acronyms.</p>
<p>To recognise potential acronyms, we used a simple heuristic based on the rules described in <xref ref-type="table" rid="T1">Table&#x00A0;1</xref>, which, when tested on an inventory of 36,162 acronym definitions derived from (<xref ref-type="bibr" rid="B39">39</xref>), was able to match a total of 32,940 (91.09&#x0025;) acronyms. Using this heuristic, we extracted a total of 26 potential acronyms from the corpus, which were then interpreted manually by analysing their concordances (see <xref ref-type="fig" rid="F3">Figure&#x00A0;3</xref>). One of the potential acronyms, LF, turned out to be a special token used in MIMIC-III to indicate missing information that had been removed to anonymise the data. Hence its long form is not available in the ground truth. Nonetheless, we retained it as a potential acronym to challenge the disambiguation model to reject potential long forms as false positives. Further, the long forms of three acronyms, IV, OA and SI, consist of a single word. In this study, we are exploring only MWTs as potential long forms. Thus, when these acronyms are paired with MWTs as potential long forms, we know in advance that it can result in either a false positive or a true negative.</p>
<table-wrap id="T1" position="float"><label>Table 1</label>
<caption><p>Heuristic rules for acronym recognition.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">ID</th>
<th valign="top" align="center">Rule</th>
<th valign="top" align="center">Rationale</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">1</td>
<td valign="top" align="left">The token has to be 2&#x2013;10 characters long.</td>
<td valign="top" align="left">Acronyms are short forms.</td>
</tr>
<tr>
<td valign="top" align="center">2</td>
<td valign="top" align="left">The token has to be tagged as a noun.</td>
<td valign="top" align="left">Acronyms are matched against MWTs, which are noun phrases.</td>
</tr>
<tr>
<td valign="top" align="center">3</td>
<td valign="top" align="left">The first character of the token has to be a capital letter.</td>
<td valign="top" align="left">Apart from few exceptions (e.g. &#x201C;mRNA&#x201D;) most acronyms start with a capital letter.</td>
</tr>
<tr>
<td valign="top" align="center">4</td>
<td valign="top" align="left">The number of letters has to be greater than the number of digits.</td>
<td valign="top" align="left">To prevent retrieving numerical expressions (e.g. &#x201C;USD5000&#x201D;).</td>
</tr>
<tr>
<td valign="top" align="center">5</td>
<td valign="top" align="left">The number of uppercase letters has to be greater than the number of lowercase letters.</td>
<td valign="top" align="left">To prevent retrieval of outliers such as personal names (e.g. &#x201C;McMurray&#x201D;) and titles (e.g. &#x201C;Miss&#x201D;).</td>
</tr>
<tr>
<td valign="top" align="center">6</td>
<td valign="top" align="left">The token has to occur at least 10 times in the corpus.</td>
<td valign="top" align="left">Acronyms are introduced because the corresponding concepts are referred to frequently.</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F3" position="float"><label>Figure 3</label>
<caption><p>Ground truth with the distribution of long form candidates.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-06-1282043-g003.tif"/>
</fig>
<p>To find potential long forms within the corpus, we applied a method called FlexiTerm, which was originally developed to recognise MWTs (<xref ref-type="bibr" rid="B31">31</xref>). Even though it was subsequently extended to recognise acronyms as MWTs (<xref ref-type="bibr" rid="B40">40</xref>), we did not use this option in order to test acronym disambiguation with our newly developed model. The list of MWTs was obtained using the latest implementation of FlexiTerm (<xref ref-type="bibr" rid="B41">41</xref>). It consisted of 2,079 terms, each linked to multiple variants arising from inflection, derivation or hyphenation.</p>
<p>To match potential acronyms and MWTs, we used a simple heuristic based on the rules described in <xref ref-type="table" rid="T2">Table&#x00A0;2</xref>. When tested on an inventory of 32,940 definitions whose acronyms matched the rules described in <xref ref-type="table" rid="T1">Table&#x00A0;1</xref> (see above), a total of 31,241 (94.84&#x0025;) acronym definitions were consistent with the patterns described in <xref ref-type="table" rid="T2">Table&#x00A0;2</xref>. Of course, all ambiguous acronyms will match most of the corresponding long forms and potentially other phrases extracted from the corpus. At this stage, the matching procedure was deliberately loose so as to extract as many potential long forms in order to challenge the disambiguation model. As a result, we assembled a list of 110 pairs of acronyms and their potential long forms. Each acronym was mapped to 4.19 MWTs on average with standard deviation of 2.91 (see <xref ref-type="fig" rid="F3">Figure&#x00A0;3</xref> for distribution).</p>
<table-wrap id="T2" position="float"><label>Table 2</label>
<caption><p>Heuristic rules for matching acronyms and MWTs.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">ID</th>
<th valign="top" align="center">Rule</th>
<th valign="top" align="center">Rationale</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">1</td>
<td valign="top" align="left">Both acronym and MWT have to start with the same letter.</td>
<td valign="top" align="left">Apart from few exceptions (e.g. &#x201C;XML&#x201D;), most acronyms start with the initial letter of the long form.</td>
</tr>
<tr>
<td valign="top" align="center">2</td>
<td valign="top" align="left">All characters from the acronym have to occur in the same order within the MWT.</td>
<td valign="top" align="left">Acronyms are formed by removing characters from the long form, which preserves their original order.</td>
</tr>
<tr>
<td valign="top" align="center">3</td>
<td valign="top" align="left">The acronym cannot occur as a token within the long form.</td>
<td valign="top" align="left">Acronym cannot abbreviate itself.</td>
</tr>
<tr>
<td valign="top" align="center">4</td>
<td valign="top" align="left">The last character from the acronym has to occur in the last token of the MWT.</td>
<td valign="top" align="left">With the exception of prepositions, which cannot occur at the end of a noun phrase, most other tokens are referenced by a character within the acronym.</td>
</tr>
<tr>
<td valign="top" align="center">5</td>
<td valign="top" align="left">The Damerau&#x2013;Levenshtein distance between the acronym and the initialism constructed out of the MWT is less than 2.</td>
<td valign="top" align="left">To ensure that most initial characters correspond to a character within the acronym.</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Having paired acronyms and long forms, each pair was then labelled as either positive or negative example according to the ground truth provided in <xref ref-type="fig" rid="F3">Figure&#x00A0;3</xref>. All terms that conveyed the sense of the long form provided in the ground truth were annotated as positive examples. For example, terms &#x201C;total knee arthroplasty&#x201D; and &#x201C;total knee replacement&#x201D; are synonyms and as such they convey the same sense as the acronym &#x201C;TKA.&#x201D; Similarly, &#x201C;total left knee arthroplasty&#x201D; is a &#x201C;total knee arthroplasty&#x201D; and as such is accepted as a correct sense of &#x201C;TKA.&#x201D; The ground truth was extended by manually annotating all 110 pairs of acronyms and their potential long forms (see <xref ref-type="sec" rid="s10">Supplementary Material</xref>).</p>
<p>Recall that the disambiguation model expects a triplet (sentence, acronym, long form) as its input (see <xref ref-type="fig" rid="F2">Figure&#x00A0;2</xref>). Having paired acronyms and long forms, for each acronym we extracted all sentences that mention the acronym. Given that the corpus of clinical narratives belonged to a single domain of knee radiology reports, we used one sense per discourse hypothesis (<xref ref-type="bibr" rid="B42">42</xref>) to propagate the label originally assigned to the pair (acronym, long form) to all corresponding triplets (sentence, acronym, long form).</p>
</sec>
</sec>
<sec id="s3d"><label>3.4</label><title>Binary classification of acronym senses</title>
<p>All triplets (sentence, acronym, long form) extracted from clinical narratives were passed through the BERT-based disambiguation model trained on scientific abstracts described in <xref ref-type="sec" rid="s3b2">Section 3.2.2</xref>. For each (acronym, long form) pair, individual classifications of acronyms occurrences in the corresponding sentences were aggregated. A long form was deemed positive if it was classified as such in at least 75&#x0025; of the acronym occurrences; otherwise it was deemed negative. In this manner, a sense inventory is constructed (see bottom left in <xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>).</p>
</sec>
</sec>
<sec id="s4" sec-type="results"><label>4</label><title>Results</title>
<sec id="s4a"><label>4.1</label><title>Binary classification of acronym senses</title>
<p>Collocational stability refers to the consistency and predictability of words that tend to appear together. High frequency of multi-word phrases implies their collocational stability (<xref ref-type="bibr" rid="B43">43</xref>), which in turn enhances semantic clarity of a discourse. Acronym definitions themselves are collocations (<xref ref-type="bibr" rid="B44">44</xref>). Given that, in our approach, potential long forms are not sourced from a predefined inventory but are instead extracted from a corpus using a method that measures their collocational stability (<xref ref-type="bibr" rid="B31">31</xref>, <xref ref-type="bibr" rid="B41">41</xref>), it is intuitively plausible to assume that the most frequent one is indeed the correct one. In other words, whenever multiple MWTs extracted by FlexiTerm match a given acronym according to the rules described in <xref ref-type="table" rid="T2">Table&#x00A0;2</xref>, the baseline method selects the most frequent one.</p>
<p>We evaluated the BERT-based disambiguation model described in <xref ref-type="sec" rid="s3b2">Section 3.2.2</xref> on the test subset of PubMed abstracts. The results provided in <xref ref-type="table" rid="T3">Table&#x00A0;3</xref> were compared against a baseline classifier described above. At almost 95&#x0025; across all metrics, our model outperformed the frequency-based approach indicating that it successfully identified patterns for acronym disambiguation. While this disambiguation model proved to be both effective and efficient, its main drawback comes from the supervised training as it may struggle to disambiguate acronyms that were not represented in the training data. This is a valid concern when a model trained on scientific abstracts is applied on genuine clinical data. The next section evaluates the performance of acronym sense extraction, which incorporates binary classification, on clinical data, hence offering an insight into its performance of such data.</p>
<table-wrap id="T3" position="float"><label>Table 3</label>
<caption><p>Performance of the acronym disambiguation model on scientific abstracts.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Method</th>
<th valign="top" align="center">Accuracy</th>
<th valign="top" align="center">Precision</th>
<th valign="top" align="center">Recall</th>
<th valign="top" align="center">F1-score</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Baseline</td>
<td valign="top" align="center">64.96&#x0025;</td>
<td valign="top" align="center">69.15&#x0025;</td>
<td valign="top" align="center">54.02&#x0025;</td>
<td valign="top" align="center">60.65&#x0025;</td>
</tr>
<tr>
<td valign="top" align="left">BERT</td>
<td valign="top" align="center">94.62&#x0025;</td>
<td valign="top" align="center">94.45&#x0025;</td>
<td valign="top" align="center">94.89&#x0025;</td>
<td valign="top" align="center">94.67&#x0025;</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4b"><label>4.2</label><title>Acronym sense extraction</title>
<p>The first row of <xref ref-type="table" rid="T4">Table&#x00A0;4</xref> provides the results of acronym sense extraction on clinical narratives. Before interpreting the evaluation results, it is worth pointing out that one major difference between the two evaluations. The evaluation results on the set of scientific abstracts shown in <xref ref-type="table" rid="T3">Table&#x00A0;3</xref> were obtained on a balanced set where the number of positive and negative test examples was equal. The balance was achieved by design. When evaluating acronym sense extraction, the distribution of potential senses is a function of the corpus from which these senses were extracted. Not surprisingly, the distribution of correct and incorrect senses was heavily skewed in favour of the incorrect ones, which outnumbered the correct senses by a ratio of over 5:1. In an attempt to redress the extreme imbalance, we tried to filter the long forms prior to passing them on to the binary classifier for disambiguation. For each acronym, we kept only the top two matching MWTs according to their termhood score calculated by FlexiTerm under an assumption that acronyms are more likely to be introduced for collocationally stable phrases. This somewhat reduced the number of incorrect senses, which, however, still outnumbered the correct senses by a ratio of over 3:1. The corresponding evaluation results are provided in the second row of <xref ref-type="table" rid="T4">Table&#x00A0;4</xref>.</p>
<table-wrap id="T4" position="float"><label>Table 4</label>
<caption><p>Performance of the acronym disambiguation model on clinical narratives.</p></caption>
<table frame="hsides" rules="groups">
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Long forms</th>
<th valign="top" align="center">Accuracy</th>
<th valign="top" align="center">Precision</th>
<th valign="top" align="center">Recall</th>
<th valign="top" align="center">F1-score</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">All terms</td>
<td valign="top" align="center">84.40&#x0025;</td>
<td valign="top" align="center">55.26&#x0025;</td>
<td valign="top" align="center">100.00&#x0025;</td>
<td valign="top" align="center">71.19&#x0025;</td>
</tr>
<tr>
<td valign="top" align="left">Top 2 terms</td>
<td valign="top" align="center">80.85&#x0025;</td>
<td valign="top" align="center">59.09&#x0025;</td>
<td valign="top" align="center">100.00&#x0025;</td>
<td valign="top" align="center">74.29&#x0025;</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Accuracy is a poor measure for imbalanced data as any model that predicts the majority class for all predictions is bound to achieve a high classification accuracy. We, therefore, turn our attention to precision and recall. We can immediately observe the perfect recall, which indicates that none of the correct senses were misclassified. However, the precision of at most 60&#x0025; indicates a high number of false positives. Error analysis revealed an interesting fact. The majority of false positives shared a word (or a subword) with the corresponding true positive. Consider for example, &#x201C;lateral condyle&#x201D; and &#x201C;medial condyle,&#x201D; which were incorrectly classified as long forms of the acronyms MCL (medial collateral ligament) and LCL (lateral collateral ligament). In both cases, the false and true positives share the first word. A similar pattern can even be observed at the subword level. For example, &#x201C;intraoperative views&#x201D; was incorrectly classified as a long form of IV (intravenous). Similarly, &#x201C;osseous abnormality&#x201D; was incorrectly classified as OA (osteoarthritis). &#x201C;Osseous&#x201D; is an adjective meaning &#x201C;bony,&#x201D; whereas &#x201C;osteo-&#x201D; is a combining form meaning &#x201C;bone.&#x201D;</p>
<p>The masked language modelling, which is used to pretrain BERT, may provide a plausible explanation for this phenomenon. During pretraining, 15&#x0025; of the words are masked randomly forcing the model to predict the output for those words based on the words around them. When BERT encounters an acronym, especially if it has not been seen during pretraining, it will try to predict its embedding based on its context. As the long form of an acronym is fully compatible with the context of the acronym, when some of the corresponding words (or subwords) are found in a potential long form, the classification model may use them as evidence for positive classification.</p>
<p>Finally, let us address our previous concern of applying a disambiguation model trained on scientific abstracts on genuine clinical data. Out of 26 acronyms extracted from the test data (clinical narratives), nine were not found in the training data (scientific abstracts). Despite this, the classifier correctly disambiguated all but two of them. In line with our previous observation about the masked language model, this suggests that the model was able to exploit the words in the long form to disambiguate an acronym based on its context.</p>
</sec>
</sec>
<sec id="s5" sec-type="discussion"><label>5</label><title>Discussion</title>
<p>We described a system for WSD of global acronyms in clinical narratives. It integrates a rule-based method for data annotation, a statistical method for MWT recognition, a rule-based method for extracting a sense inventory from a corpus and, finally, a neural network for disambiguation. These methods complement one another in a way that offers multiple advantages over the state of the art. First, it uses scientific abstracts to simulate clinical narrative style of acronym usage and annotate them automatically with the correct senses. This circumvents the problems associated with patient privacy and manual annotation overhead, which have traditionally plagued the use of clinical text data in machine learning (<xref ref-type="bibr" rid="B8">8</xref>). In turn, large amounts of annotated data lay a foundation for training a robust neural network, which fosters trust and reliability necessary for real-world clinical applications. Another practical advantage of our approach is that it does not require predefined sense inventories, which are known to vary greatly across different geographic regions and medical specialties (<xref ref-type="bibr" rid="B23">23</xref>). Instead, an inventory of potential senses is extracted from a corpus on the fly. This does not only make our approach readily portable across different domains, but also improves its utility in clinical practice. For example, our approach can be used to facilitate development of institution-specific inventories, which in turn can be used to improve communication and shared decision making.</p>
<p>CLASSE GATOR (CLinical Acronym SenSE disambiGuATOR) (<xref ref-type="bibr" rid="B45">45</xref>) is the most similar system for WSD of clinical acronyms. The key similarity is in the use of scientific literature for bootstrapping training data. The key difference is that their clinical acronym senses are also extracted from the literature. This fact requires the two corpora of scientific literature and clinical narratives respectively to belong to the same domain. It also requires the acronym sense to be used in the literature. The main criticism of previous acronym disambiguation systems is that they constructed a sense inventory from a single institutional corpus, which hindered their generalisability, because acronyms tend to vary across institutions (<xref ref-type="bibr" rid="B23">23</xref>). Therefore, scientific literature cannot reasonably be expected to cover acronyms as they are used in clinical practice. Nonetheless, the two systems advance the area of acronym disambiguation in complementary ways.</p>
<p>Both CLASSE GATOR and our own approach are based on BERT, one of many large pretrained language models. It was fine-tuned for the WSD task using a scientific corpus as a proxy for clinical narratives, which remain largely inaccessible due to privacy concerns. However, recent advances in generative language models have opened up a new opportunity for training NLP models in a clinical domain. What sets apart generative models from other types of language models is their ability to produce original content. This can be used to generate potentially unlimited amounts of training data. Early efforts have demonstrated that even though the generated text was of poorer quality relative to the original text, when it was used to augment the training data, it still boosted the performance of downstream NLP tasks (<xref ref-type="bibr" rid="B46">46</xref>). Domain-specific databases can be used to mitigate the inherently stochastic nature of large language models (<xref ref-type="bibr" rid="B47">47</xref>) in an attempt to improve the accuracy, diversity and complexity of generated clinical data (<xref ref-type="bibr" rid="B48">48</xref>). Nonetheless, the data annotation bottleneck still persists but may be addressed with strategic prompt engineering. Our future work will explore this opportunity to make further advances in WSD of clinical acronyms.</p>
<p>While our evaluation demonstrates the generalisability of our approach in that it can successfully disambiguate acronyms that have not been seen in the training data, it certainly stands to further improve its performance by utilising the context of long form candidates. Further performance improvements may be gained by modeling the orthographic constraints of acronym formation. This would help eliminate some of the false positives. Therefore, we will be looking to extend the binary classification component by integrating token-based and character-based neural networks.</p>
</sec>
</body>
<back>
<sec id="s6" sec-type="data-availability"><title>Data availability statement</title>
<p>The raw data used in this study can be downloaded from PubMed (<ext-link ext-link-type="uri" xlink:href="https://www.nlm.nih.gov/databases/download/pubmed_medline.html">https://www.nlm.nih.gov/databases/download/pubmed_medline.html</ext-link>) and MIMIC-III (<ext-link ext-link-type="uri" xlink:href="https://physionet.org/content/mimiciii/1.4/">https://physionet.org/content/mimiciii/1.4/</ext-link>). The data can be processed using open-source tools, which can be downloaded from GitHub (<ext-link ext-link-type="uri" xlink:href="https://github.com/ispasic/acronimity">https://github.com/ispasic/acronimity</ext-link>; <ext-link ext-link-type="uri" xlink:href="https://github.com/ispasic/FlexiTerm-Python">https://github.com/ispasic/FlexiTerm-Python</ext-link>).</p>
</sec>
<sec id="s7" sec-type="author-contributions"><title>Author contributions</title>
<p>DC: Methodology, Validation, Data curation, Formal Analysis, Investigation, Software, Writing &#x2013; review &#x0026; editing; PC: Supervision, Writing &#x2013; review &#x0026; editing; IS: Conceptualization, Methodology, Supervision, Validation, Visualization, Writing &#x2013; original draft.</p>
</sec>
<sec id="s8" sec-type="funding-information"><title>Funding</title>
<p>The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.</p>
</sec>
<sec id="s9" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s11" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s10" sec-type="supplementary-material"><title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fdgth.2024.1282043/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fdgth.2024.1282043/full&#x0023;supplementary-material</ext-link></p>
<supplementary-material id="SD1" content-type="local-data">
<media mimetype="application" mime-subtype="vnd.openxmlformats-officedocument.spreadsheetml.sheet" xlink:href="Table1.xlsx"/></supplementary-material>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fandrych</surname><given-names>I</given-names></name></person-group>. <article-title>Submorphemic elements in the formation of acronyms, blends, clippings</article-title>. <source>Lexis J Engl Lexicol</source> (<year>2008</year>) 2. <pub-id pub-id-type="doi">10.4000/lexis.713</pub-id></citation></ref>
<ref id="B2"><label>2.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Laszlo</surname><given-names>S</given-names></name><name><surname>Federmeier</surname><given-names>KD</given-names></name></person-group>. <article-title>The acronym superiority effect</article-title>. <source>Psychon Bull Rev</source> (<year>2007</year>) <volume>14</volume>:<fpage>1158</fpage>&#x2013;<lpage>63</lpage>. <pub-id pub-id-type="doi">10.3758/BF03193106</pub-id><pub-id pub-id-type="pmid">18229490</pub-id></citation></ref>
<ref id="B3"><label>3.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Moon</surname><given-names>S</given-names></name><name><surname>Pakhomov</surname><given-names>S</given-names></name><name><surname>Liu</surname><given-names>N</given-names></name><name><surname>Ryan</surname><given-names>JO</given-names></name><name><surname>Melton</surname><given-names>GB</given-names></name></person-group>. <article-title>A sense inventory for clinical abbreviations, acronyms created using clinical notes, medical dictionary resources</article-title>. <source>J Am Med Inform Assoc</source> (<year>2014</year>) <volume>21</volume>:<fpage>299</fpage>&#x2013;<lpage>307</lpage>. <pub-id pub-id-type="doi">10.1136/amiajnl-2012-001506</pub-id><pub-id pub-id-type="pmid">23813539</pub-id></citation></ref>
<ref id="B4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name><name><surname>Krzemi&#x0144;ski</surname><given-names>D</given-names></name><name><surname>Corcoran</surname><given-names>P</given-names></name><name><surname>Balinsky</surname><given-names>A</given-names></name></person-group>. <article-title>Cohort selection for clinical trials from longitudinal patient records: text mining approach</article-title>. <source>JMIR Med Inform</source> (<year>2019</year>) <volume>7</volume>:<fpage>e15980</fpage>. <pub-id pub-id-type="doi">10.2196/15980</pub-id></citation></ref>
<ref id="B5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Holper</surname><given-names>S</given-names></name><name><surname>Barmanray</surname><given-names>R</given-names></name><name><surname>Colman</surname><given-names>B</given-names></name><name><surname>Yates</surname><given-names>CJ</given-names></name><name><surname>Liew</surname><given-names>D</given-names></name><name><surname>Smallwood</surname><given-names>D</given-names></name></person-group>. <article-title>Ambiguous medical abbreviation study: challenges and opportunities</article-title>. <source>Intern Med J</source> (<year>2020</year>) <volume>50</volume>:<fpage>1073</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1111/imj.14442</pub-id><pub-id pub-id-type="pmid">31389137</pub-id></citation></ref>
<ref id="B6"><label>6.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bodenreider</surname><given-names>O</given-names></name></person-group>. <article-title>The Unified Medical Language System (UMLS): integrating biomedical terminology</article-title>. <source>Nucleic Acids Res</source> (<year>2004</year>) <volume>32</volume>:<fpage>D267</fpage>&#x2013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1093/nar/gkh061</pub-id><pub-id pub-id-type="pmid">14681409</pub-id></citation></ref>
<ref id="B7"><label>7.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Agirre</surname><given-names>E</given-names></name><name><surname>Stevenson</surname><given-names>M</given-names></name></person-group>. <article-title>Knowledge sources for WSD</article-title>. In: <person-group person-group-type="editor"><name><surname>Agirre</surname><given-names>E</given-names></name><name><surname>Edmonds</surname><given-names>P</given-names></name></person-group>. editors. <source>Word Sense Disambiguation</source>. <publisher-name>New York: Springer</publisher-name> (<year>2007</year>). p. <fpage>217</fpage>&#x2013;<lpage>251</lpage>.</citation></ref>
<ref id="B8"><label>8.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name><name><surname>Nenadi&#x0107;</surname><given-names>G</given-names></name></person-group>. <article-title>Clinical text data in machine learning: systematic review</article-title>. <source>JMIR Med Inform</source> (<year>2020</year>) <volume>8</volume>:<fpage>e17984</fpage>. <pub-id pub-id-type="doi">10.2196/17984</pub-id></citation></ref>
<ref id="B9"><label>9.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Church</surname><given-names>K</given-names></name><name><surname>Liu</surname><given-names>B</given-names></name></person-group>. <article-title>Acronyms, opportunities for improving deep nets</article-title>. <source>Front Artif Intell</source> (<year>2021</year>) <volume>4</volume>:<fpage>732381</fpage>. <pub-id pub-id-type="doi">10.3389/frai.2021.732381</pub-id><pub-id pub-id-type="pmid">34988434</pub-id></citation></ref>
<ref id="B10"><label>10.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Schwartz</surname><given-names>AS</given-names></name><name><surname>Hearst</surname><given-names>MA</given-names></name></person-group>. <article-title>A simple algorithm for identifying abbreviation definitions in biomedical text</article-title>. In: <source>Biocomputing 2003</source>. <publisher-name>Singapore: World Scientific</publisher-name> (<year>2002</year>). p. <fpage>451</fpage>&#x2013;<lpage>462</lpage>.</citation></ref>
<ref id="B11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sohn</surname><given-names>S</given-names></name><name><surname>Comeau</surname><given-names>DC</given-names></name><name><surname>Kim</surname><given-names>W</given-names></name><name><surname>Wilbur</surname><given-names>WJ</given-names></name></person-group>. <article-title>Abbreviation definition identification based on automatic precision estimates</article-title>. <source>BMC Bioinf</source> (<year>2008</year>) <volume>9</volume>:<fpage>402</fpage>. <pub-id pub-id-type="doi">10.1186/1471-2105-9-402</pub-id></citation></ref>
<ref id="B12"><label>12.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Veyseh</surname><given-names>APB</given-names></name><name><surname>Dernoncourt</surname><given-names>F</given-names></name><name><surname>Nguyen</surname><given-names>TH</given-names></name><name><surname>Chang</surname><given-names>W</given-names></name><name><surname>Celi</surname><given-names>LA</given-names></name></person-group>. <article-title>Acronym identification, disambiguation shared tasks for scientific document understanding</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B13"><label>13.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Pan</surname><given-names>C</given-names></name><name><surname>Song</surname><given-names>B</given-names></name><name><surname>Wang</surname><given-names>S</given-names></name><name><surname>Luo</surname><given-names>Z</given-names></name></person-group>. <article-title>BERT-based acronym disambiguation with multiple training strategies</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B14"><label>14.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Zhong</surname><given-names>Q</given-names></name><name><surname>Zeng</surname><given-names>G</given-names></name><name><surname>Zhu</surname><given-names>D</given-names></name><name><surname>Zhang</surname><given-names>Y</given-names></name><name><surname>Lin</surname><given-names>W</given-names></name><name><surname>Chen</surname><given-names>B</given-names></name></person-group> et al. <article-title>Leveraging domain agnostic, specific knowledge for acronym disambiguation</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B15"><label>15.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Egan</surname><given-names>N</given-names></name><name><surname>Bohannon</surname><given-names>J</given-names></name></person-group>. <article-title>Primer AI&#x2019;s systems for acronym identification, disambiguation</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B16"><label>16.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Singh</surname><given-names>A</given-names></name><name><surname>Kumar</surname><given-names>P</given-names></name></person-group>. <article-title>SCIDR at SDU-2020: ideas - identifying and disambiguating everyday acronyms for scientific domain</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B17"><label>17.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Rogers</surname><given-names>W</given-names></name><name><surname>Rae</surname><given-names>A</given-names></name><name><surname>Demner-Fushman</surname><given-names>D</given-names></name></person-group>. <article-title>AI-NLM exploration of the acronym identification shared task at SDU@AAAI-21</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B18"><label>18.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Jaber</surname><given-names>A</given-names></name><name><surname>Mart&#x00ED;nez</surname><given-names>P</given-names></name></person-group>. <article-title>Participation of UC3M in SDU@AAAI-21: a hybrid approach to disambiguate scientific acronyms</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B19"><label>19.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Pereira</surname><given-names>JL</given-names></name><name><surname>Galhardas</surname><given-names>H</given-names></name><name><surname>Shasha</surname><given-names>D</given-names></name></person-group>. <article-title>Acronym expander at SDU@AAAI-21: an acronym disambiguation module</article-title>. In: <conf-name>Proceedings of the AAAI Workshop on Scientific Document Understanding</conf-name> (<year>2021</year>).</citation></ref>
<ref id="B20"><label>20.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Holper</surname><given-names>S</given-names></name><name><surname>Barmanray</surname><given-names>R</given-names></name><name><surname>Colman</surname><given-names>B</given-names></name><name><surname>Yates</surname><given-names>CJ</given-names></name><name><surname>Liew</surname><given-names>D</given-names></name><name><surname>Smallwood</surname><given-names>D</given-names></name></person-group>. <article-title>Ambiguous medical abbreviation study: challenges and opportunities</article-title>. <source>Intern Med J</source> (<year>2020</year>) <volume>50</volume>:<fpage>1073</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1111/imj.14442</pub-id><pub-id pub-id-type="pmid">31389137</pub-id></citation></ref>
<ref id="B21"><label>21.</label><citation citation-type="other"><collab>National Library of Medicine (US)</collab>. <comment>UMLS reference manual (2009). Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.ncbi.nlm.nih.gov/books/NBK9676/">https://www.ncbi.nlm.nih.gov/books/NBK9676/</ext-link> (accessed: March 19, 2023) [Dataset]</comment>.</citation></ref>
<ref id="B22"><label>22.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname><given-names>W</given-names></name><name><surname>Torvik</surname><given-names>VI</given-names></name><name><surname>Smalheiser</surname><given-names>NR</given-names></name></person-group>. <article-title>ADAM: Another database of abbreviations in MEDLINE</article-title>. <source>Bioinformatics</source> (<year>2006</year>) <volume>22</volume>:<fpage>2813</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btl480</pub-id><pub-id pub-id-type="pmid">16982707</pub-id></citation></ref>
<ref id="B23"><label>23.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grossman</surname><given-names>LV</given-names></name><name><surname>Mitchell</surname><given-names>EG</given-names></name><name><surname>Hripcsak</surname><given-names>G</given-names></name><name><surname>Weng</surname><given-names>C</given-names></name><name><surname>Vawdrey</surname><given-names>DK</given-names></name></person-group>. <article-title>A method for harmonization of clinical abbreviation and acronym sense inventories</article-title>. <source>J Biomed Inform</source> (<year>2018</year>) <volume>88</volume>:<fpage>62</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2018.11.004</pub-id><pub-id pub-id-type="pmid">30414475</pub-id></citation></ref>
<ref id="B24"><label>24.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Xia</surname><given-names>Y</given-names></name><name><surname>Zhong</surname><given-names>X</given-names></name><name><surname>Liu</surname><given-names>P</given-names></name><name><surname>Tan</surname><given-names>C</given-names></name><name><surname>Na</surname><given-names>S</given-names></name><name><surname>Hu</surname><given-names>Q</given-names></name></person-group> et al. <comment>Normalization of abbreviations/acronyms: THCIB at CLEF eHealth 2013 task 2. In: Forner P, Navigli R, Tufis D. editors. <italic>CLEF 2013 Evaluation Labs and Workshops: Working Notes</italic> (2013)</comment>.</citation></ref>
<ref id="B25"><label>25.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Jagannathan</surname><given-names>V</given-names></name><name><surname>Ganesan</surname><given-names>D</given-names></name><name><surname>Jagannathan</surname><given-names>A</given-names></name><name><surname>Kavi</surname><given-names>R</given-names></name><name><surname>Lamb</surname><given-names>A</given-names></name><name><surname>Peters</surname><given-names>F</given-names></name></person-group> et al. <comment>WVU NLP class participation in ShARe/CLEF challenge. In: Forner P, Navigli R, Tufis D. editors. <italic>CLEF 2013 Evaluation Labs and Workshops: Working Notes</italic> (2013)</comment>.</citation></ref>
<ref id="B26"><label>26.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Patrick</surname><given-names>J</given-names></name><name><surname>Safari</surname><given-names>L</given-names></name><name><surname>Ou</surname><given-names>Y</given-names></name></person-group>. <comment>ShARe/CLEF eHealth 2013 normalization of acronyms/abbreviations challenge. In: Forner P, Navigli R, Tufis D. editors. <italic>CLEF 2013 Evaluation Labs and Workshops: Working Notes</italic> (2013)</comment>.</citation></ref>
<ref id="B27"><label>27.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Wu</surname><given-names>Y</given-names></name><name><surname>Tang</surname><given-names>B</given-names></name><name><surname>Jiang</surname><given-names>M</given-names></name><name><surname>Moon</surname><given-names>S</given-names></name><name><surname>Denny</surname><given-names>J</given-names></name><name><surname>Xu</surname><given-names>H</given-names></name></person-group>. <comment>Clinical acronym/abbreviation normalization using a hybrid approach. In: Forner P, Navigli R, Tufis D. editors. <italic>CLEF 2013 Evaluation Labs and Workshops: Working Notes</italic> (2013)</comment>.</citation></ref>
<ref id="B28"><label>28.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Zweigenbaum</surname><given-names>P</given-names></name><name><surname>Del&#x00E9;ger</surname><given-names>L</given-names></name><name><surname>Lavergne</surname><given-names>T</given-names></name><name><surname>N&#x00E9;v&#x00E9;rol</surname><given-names>A</given-names></name><name><surname>Bodnari</surname><given-names>A</given-names></name></person-group>. <comment>A supervised abbreviation resolution system for medical text. In: Forner P, Navigli R, Tufis D. editors. <italic>CLEF 2013 Evaluation Labs and Workshops: Working Notes</italic> (2013)</comment>.</citation></ref>
<ref id="B29"><label>29.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mowery</surname><given-names>DL</given-names></name><name><surname>South</surname><given-names>BR</given-names></name><name><surname>Christensen</surname><given-names>L</given-names></name><name><surname>Leng</surname><given-names>J</given-names></name><name><surname>Peltonen</surname><given-names>L-M</given-names></name><name><surname>Salanter&#x00E4;</surname><given-names>S</given-names></name></person-group> et al. <article-title>Normalizing acronyms and abbreviations to aid patient understanding of clinical texts: ShARe/CLEF eHealth challenge 2013, task 2</article-title>. <source>J Biomed Semant</source> (<year>2016</year>) <volume>7</volume>:<fpage>43</fpage>. <pub-id pub-id-type="doi">10.1186/s13326-016-0084-y</pub-id></citation></ref>
<ref id="B30"><label>30.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Devlin</surname><given-names>J</given-names></name><name><surname>Chang</surname><given-names>M-W</given-names></name><name><surname>Lee</surname><given-names>K</given-names></name><name><surname>Toutanova</surname><given-names>K</given-names></name></person-group>. <comment>BERT: pre-training of deep bidirectional transformers for language understanding. In: <italic>Annual Conference of the North American Chapter of the Association for Computational Linguistics</italic> (2019)</comment>.</citation></ref>
<ref id="B31"><label>31.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name><name><surname>Greenwood</surname><given-names>M</given-names></name><name><surname>Preece</surname><given-names>A</given-names></name><name><surname>Francis</surname><given-names>N</given-names></name><name><surname>Elwyn</surname><given-names>G</given-names></name></person-group>. <article-title>FlexiTerm: a flexible term recognition method</article-title>. <source>J Biomed Semant</source> (<year>2013</year>) <volume>4</volume>:<fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1186/2041-1480-4-27</pub-id></citation></ref>
<ref id="B32"><label>32.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Filimonov</surname><given-names>M</given-names></name><name><surname>Chopard</surname><given-names>D</given-names></name><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name></person-group>. <article-title>Simulation, annotation of global acronyms</article-title>. <source>Bioinformatics</source> (<year>2022</year>) <volume>38</volume>:<fpage>3136</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btac298</pub-id><pub-id pub-id-type="pmid">35482480</pub-id></citation></ref>
<ref id="B33"><label>33.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Huang</surname><given-names>L</given-names></name><name><surname>Sun</surname><given-names>C</given-names></name><name><surname>Qiu</surname><given-names>X</given-names></name><name><surname>Huang</surname><given-names>X-J</given-names></name></person-group>. <article-title>GlossBERT: BERT for word sense disambiguation with gloss knowledge</article-title>. In: <conf-name>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</conf-name> (<year>2019</year>). p. <fpage>3509</fpage>&#x2013;<lpage>3514</lpage>.</citation></ref>
<ref id="B34"><label>34.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Miller</surname><given-names>GA</given-names></name></person-group>. <source>WordNet: an electronic lexical database</source>. <publisher-name>Cambridge, MA: MIT Press</publisher-name> (<year>1998</year>).</citation></ref>
<ref id="B35"><label>35.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chopard</surname><given-names>D</given-names></name><name><surname>Treder</surname><given-names>MS</given-names></name><name><surname>Corcoran</surname><given-names>P</given-names></name><name><surname>Ahmed</surname><given-names>N</given-names></name><name><surname>Johnson</surname><given-names>C</given-names></name><name><surname>Busse</surname><given-names>M</given-names></name></person-group> et al. <article-title>Text mining of adverse events in clinical trials: deep learning approach</article-title>. <source>JMIR Med Inform</source> (<year>2021</year>) <volume>9</volume>:<fpage>e28632</fpage>. <pub-id pub-id-type="doi">10.2196/28632</pub-id><pub-id pub-id-type="pmid">34951601</pub-id></citation></ref>
<ref id="B36"><label>36.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goldberger</surname><given-names>AL</given-names></name><name><surname>Amaral</surname><given-names>LA</given-names></name><name><surname>Glass</surname><given-names>L</given-names></name><name><surname>Hausdorff</surname><given-names>JM</given-names></name><name><surname>Ivanov</surname><given-names>PC</given-names></name><name><surname>Mark</surname><given-names>RG</given-names></name></person-group> et al. <article-title>PhysioBank, PhysioToolkit, PhysioNet: components of a new research resource for complex physiologic signals</article-title>. <source>Circulation</source> (<year>2000</year>) <volume>101</volume>:<fpage>e215</fpage>&#x2013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1161/01.cir.101.23.e215</pub-id><pub-id pub-id-type="pmid">10851218</pub-id></citation></ref>
<ref id="B37"><label>37.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Button</surname><given-names>K</given-names></name><name><surname>Van Deursen</surname><given-names>RW</given-names></name><name><surname>Soldatova</surname><given-names>L</given-names></name><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name></person-group>. <article-title>Trak ontology: defining standard care for the rehabilitation of knee conditions</article-title>. <source>J Biomed Inform</source> (<year>2013</year>) <volume>46</volume>:<fpage>615</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2013.04.009</pub-id><pub-id pub-id-type="pmid">23665300</pub-id></citation></ref>
<ref id="B38"><label>38.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name><name><surname>Zhao</surname><given-names>B</given-names></name><name><surname>Jones</surname><given-names>CB</given-names></name><name><surname>Button</surname><given-names>K</given-names></name></person-group>. <article-title>KneeTex: an ontology&#x2013;driven system for information extraction from MRI reports</article-title>. <source>J Biomed Semant</source> (<year>2015</year>) <volume>6</volume>:<fpage>1</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1186/s13326-015-0033-1</pub-id></citation></ref>
<ref id="B39"><label>39.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grossman Liu</surname><given-names>L</given-names></name><name><surname>Grossman</surname><given-names>RH</given-names></name><name><surname>Mitchell</surname><given-names>EG</given-names></name><name><surname>Weng</surname><given-names>C</given-names></name><name><surname>Natarajan</surname><given-names>K</given-names></name><name><surname>Hripcsak</surname><given-names>G</given-names></name></person-group> et al. <article-title>A deep database of medical abbreviations, acronyms for natural language processing</article-title>. <source>Sci Data</source> (<year>2021</year>) <volume>8</volume>:149. <pub-id pub-id-type="doi">10.1038/s41597-021-00929-4</pub-id><pub-id pub-id-type="pmid">34078918</pub-id></citation></ref>
<ref id="B40"><label>40.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name></person-group>. <article-title>Acronyms as an integral part of multi-word term recognition &#x2013; a token of appreciation</article-title>. <source>IEEE Access</source> (<year>2018</year>) <volume>6</volume>:<fpage>8351</fpage>&#x2013;<lpage>63</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2018.2807122</pub-id></citation></ref>
<ref id="B41"><label>41.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Spasi&#x0107;</surname><given-names>I</given-names></name></person-group>. <comment>FlexiTerm: a more efficient implementation of flexible multi-word term recognition. <italic>arXiv</italic> [Preprint]. <italic>arXiv:2110.06981</italic> (2021)</comment>.</citation></ref>
<ref id="B42"><label>42.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Gale</surname><given-names>WA</given-names></name><name><surname>Church</surname><given-names>K</given-names></name><name><surname>Yarowsky</surname><given-names>D</given-names></name></person-group>. <article-title>One sense per discourse</article-title>. In: <conf-name>Speech, Natural Language: Proceedings of a Workshop Held at Harriman</conf-name>, <conf-loc>New York</conf-loc> (<year>1992</year>). p. <fpage>233</fpage>&#x2013;<lpage>237</lpage>.</citation></ref>
<ref id="B43"><label>43.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kita</surname><given-names>K</given-names></name><name><surname>Kato</surname><given-names>Y</given-names></name><name><surname>Omoto</surname><given-names>T</given-names></name><name><surname>Yano</surname><given-names>Y</given-names></name></person-group>. <article-title>A comparative study of automatic extraction of collocations from corpora: mutual information vs. cost criteria</article-title>. <source>J Nat Lang Process</source> (<year>1994</year>) <volume>1</volume>:<fpage>121</fpage>&#x2013;<lpage>33</lpage>. <pub-id pub-id-type="doi">10.5715/jnlp.1.21</pub-id></citation></ref>
<ref id="B44"><label>44.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>H</given-names></name><name><surname>Friedman</surname><given-names>C</given-names></name></person-group>. <article-title>Mining terminological knowledge in large biomedical corpora</article-title>. In: <source>Biocomputing 2003</source>. <publisher-name>Singapore: World Scientific</publisher-name> (<year>2002</year>). p. <fpage>415</fpage>&#x2013;<lpage>426</lpage>.</citation></ref>
<ref id="B45"><label>45.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kashyap</surname><given-names>A</given-names></name><name><surname>Burris</surname><given-names>H</given-names></name><name><surname>Callison-Burch</surname><given-names>C</given-names></name><name><surname>Boland</surname><given-names>MR</given-names></name></person-group>. <article-title>The CLASSE GATOR (clinical acronym sense disambiguator): a method for predicting acronym sense from neonatal clinical notes</article-title>. <source>Int J Med Inform</source> (<year>2020</year>) <volume>137</volume>:<fpage>104101</fpage>. <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2020.104101</pub-id><pub-id pub-id-type="pmid">32088556</pub-id></citation></ref>
<ref id="B46"><label>46.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Amin-Nejad</surname><given-names>A</given-names></name><name><surname>Ive</surname><given-names>J</given-names></name><name><surname>Velupillai</surname><given-names>S</given-names></name></person-group>. <article-title>Exploring transformer text generation for medical dataset augmentation</article-title>. In: <person-group person-group-type="editor"><name><surname>Calzolari</surname><given-names>N</given-names></name><name><surname>B&#x00E9;chet</surname><given-names>F</given-names></name><name><surname>Blache</surname><given-names>P</given-names></name><name><surname>Choukri</surname><given-names>K</given-names></name><name><surname>Cieri</surname><given-names>C</given-names></name><name><surname>Declerck</surname><given-names>T</given-names></name><etal/></person-group> editors. <conf-name>Proceedings of the 12th Language Resources, Evaluation Conference</conf-name>, <conf-loc>Marseille, France</conf-loc>: <publisher-name>European Language Resources Association</publisher-name> (<year>2020</year>). p. <fpage>4699</fpage>&#x2013;<lpage>4708</lpage>.</citation></ref>
<ref id="B47"><label>47.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Bender</surname><given-names>EM</given-names></name><name><surname>Gebru</surname><given-names>T</given-names></name><name><surname>McMillan-Major</surname><given-names>A</given-names></name><name><surname>Shmitchell</surname><given-names>S</given-names></name></person-group>. <article-title>On the dangers of stochastic parrots: can language models be too big?</article-title> In: <conf-name>Proceedings of the 2021 ACM Conference on Fairness, Accountability,, Transparency</conf-name> (<year>2021</year>). p. <fpage>610</fpage>&#x2013;<lpage>623</lpage>.</citation></ref>
<ref id="B48"><label>48.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hodges</surname><given-names>R</given-names></name><name><surname>Tokunaga</surname><given-names>K</given-names></name><name><surname>LeGrand</surname><given-names>J</given-names></name></person-group>. <article-title>A novel method to create realistic synthetic medication data</article-title>. <source>JAMIA Open</source> (<year>2023</year>) <volume>6</volume>:<fpage>ooad052</fpage>. <pub-id pub-id-type="doi">10.1093/jamiaopen/ooad052</pub-id><pub-id pub-id-type="pmid">37457749</pub-id></citation></ref></ref-list>
</back>
</article>