Skip to main content

ORIGINAL RESEARCH article

Front. Digit. Health, 06 December 2021
Sec. Health Informatics

An Automated Dashboard to Improve Laboratory COVID-19 Diagnostics Management

\nEmma MauryEmma Maury1Marc-Olivier BoldiMarc-Olivier Boldi1Gilbert Greub,Gilbert Greub2,3Valrie ChavezValérie Chavez1Katia JatonKatia Jaton2Onya Opota
Onya Opota2*
  • 1Faculty of Business and Economics, University of Lausanne, Lausanne, Switzerland
  • 2Institute of Microbiology, Lausanne University and University Hospital of Lausanne, Lausanne, Switzerland
  • 3Infectious Diseases Service, Lausanne University and University Hospital of Lausanne, Lausanne, Switzerland

Background: In response to the COVID-19 pandemic, our microbial diagnostic laboratory located in a university hospital has implemented several distinct SARS-CoV-2 RT-PCR systems in a very short time. More than 148,000 tests have been performed over 12 months, which represents about 405 tests per day, with peaks to more than 1,500 tests per days during the second wave. This was only possible thanks to automation and digitalization, to allow high throughput, acceptable time to results and to maintain test reliability. An automated dashboard was developed to give access to Key Performance Indicators (KPIs) to improve laboratory operational management.

Methods: RT-PCR data extraction of four respiratory viruses—SARS-CoV-2, influenza A and B and RSV—from our laboratory information system (LIS), was automated. This included age, gender, test result, RT-PCR instrument, sample type, reception time, requester, and hospitalization status etc. Important KPIs were identified and the visualization was achieved using an in-house dashboard based on the open-source language R (Shiny).

Results: The dashboard is organized into three main parts. The “Filter” page presents all the KPIs, divided into five sections: (i) general and gender-related indicators, (ii) number of tests and positivity rate, (iii) cycle threshold and viral load, (iv) test durations, and (v) not valid results. Filtering allows to select a given period, a dedicated instrument, a given specimen, an age range or a requester. The “Comparison” page allows a custom charting of all the available variables, which represents more than 182 combination. The “Data” page, gives the user an access to the raw data in tables format, with possibility of filtering, allowing for a deeper analysis and data download. Informations are updated every 4 h.

Conclusions: By giving a rapid access to a huge number of up-to-date information, represented using the most relevant visualization types, without the burden of timely data extraction and analysis, the dashboard represents a reliable and user-friendly tool for operational laboratory management. The dashboard represents a reliable and user-friendly tool improving the decision-making process, resource planning and quality management.

Introduction

In December 2019, a new virus causing pneumonia of unknown etiology has emerged in China. Its incidence exploded very quickly, first in the Wuhan region (Hubei province), then in the other regions of China and other countries in Southeast Asia. On January 30, the World Organization of Health (WHO) declared this new coronavirus as a “public health emergency of international concern” (1). On the 20th of February 2020, a first patient was diagnosed in Italy in the Lombardy region. The epidemic has since spread to other European countries, including Switzerland (2) and a first case was admitted in our hospital already on 28th February. On the 11th of March, 2020 that WHO declared Pandemic Coronavirus Disease 2019 (COVID-19) (35).

To face the COVID-19 pandemic caused by the severe acute respiratory syndrome Coronavirus 2 (SARS-CoV-2), diagnostic laboratories had to develop reverse transcriptase polymerase chain reactions (RT-PCR) tests allowing the detection of SARS-CoV-2 RNA in patients with suspected COVID-19. Our laboratory, the Institute of Microbiology (IMU), located in one of the five teaching hospitals of Switzerland, the Lausanne University Hospital (CHUV), rapidly developed RT-PCR to detect SARS-CoV-2 in clinical specimens (6). Microbiological diagnosis of SARS-CoV-2 represents one of the pillars of the diagnosis of COVID-19. Indeed, RT-PCR is also the heart of the patient care and epidemic control process and will be the mainstay of several clinical studies.

Although our laboratory has extensive experience in the development of RT-PCR, the introduction of this new parameter represented a challenge in terms of speed of development (7). It is also the first time that an introduced parameter has been used on such a large scale in such a short time; more than 10,000 tests were carried out in 1 month in Spring (6) and even in a single week during fall 2020. This was possible thanks to automation and digitalization, to allow high throughput and acceptable time to results (7). In this context, the IMU set strategies to ensure the quality and reliability of the RT-PCR. This included the monitoring of Key Performance Indicators (KPIs) for quality management such as the proportions of positive tests or the virus load, both per day, per instruments, and per requester. These indicators aimed for instance to identify variations not explained by epidemiological changes. Indeed, abnormal variations could be synonymous with pre-analytical problems (sampling problem, transport medium, etc.) or even analytical problems (mutation in the target sequences of PCRs associated with losses of sensitivity or specificities). The IMU also defined KPIs for operations management such as the time to results (8).

Before COVID-19, such indicators were monitored periodically, for example in the context of an annual report or retrospective studies. At the beginning of the COVID-19 outbreak, the IMU decided to follow these indicators frequently. Because the needed manual analyses were time-consuming, the monitoring of analytical and operational KPIs was carried out once a week initially and then, twice a week depending on the period. These analyses were also prone to error, due to multiple sources of information, repeated manual actions (e.g., copy/cut and paste) and the diversity of the data.

A dashboard is a graphical user interfaces (GUI) with a data base. It allows to retrieve the relevant information, often KPIs, in a specific context by representing the data in a meaningful and usable way (9). For a detailed presentation, see (10). In the management and business contexts, dashboards aim at reversing the overwhelming information volume into an opportunity (11) and are part of visual analytics, defined as the “science of analytical reasoning facilitated by interactive visual interface” by Cook and Thomas (12).

Like any other information technology in the healthcare industry, which affects positively its efficiency (13), dashboards help monitor daily activities (14), such as tracking ongoing operations, a priority in healthcare institutions (15). Providing an easy access to this information helps the team to make better informed decisions (16), which could take a tremendous amount of time without the technology (17). Correctly designed and built, dashboards improve the institution's efficiency while providing a better care quality (18). See Cheng et al. (19), for an example of a previous study on how to build dashboard to track respiratory viruses, such as Influenza.

With COVID-19 spreading fast over the world, the speed at which data are gathered, integrated, and used became central in the management of this crisis by all health-involved institution teams. Interactive dashboards appeared to be appropriate to this aim. A famous example remains the one from Johns Hopkins University (20). In Switzerland, corona-data.ch (21) is an up-to-date webpage built at the macro level. The whole healthcare industry being impacted by the pandemic, various topics, and areas were analyzed through dashboards: e-consultations (22), incident command (23), performance comparisons to similar institutions (24).

Also, laboratories responsible for testing patients during an outbreak must control some information to ensure the highest quality of results. In particular, it is crucial to define KPIs, for example to better track the daily operations (25) ensuring enough testing capacity. Moreover, providing valuable insights and pieces of information to the laboratory management can be critical when the objective is to increase capacity and quality as well as maintaining schedules (26).

A dashboard is an expression of a database. Therefore, like mentioned by O'Donnell and David (27), the resulting decision process depends on the Information System (IS), the environment, and the user skills. Regarding the dashboard content, there is no consensus on the format of visualizations: from no effect on the user's judgement (28) to the lack of universality in representation (29), or a preference for tabular information (30). Vance Wilson and Zigurs (31) showed that even the user's preferred format did not lead to greater performance, except for symbolic, and spatial tasks (32). Choosing the appropriate visualization can be challenging and is subject to various principles. Lengler and Eppler (33) condensed many visualizations into a Periodic Table of Visualization Methods, according to which the dashboard content can be classified into three categories: boxes, tables, and plots. Each has advantages and drawbacks, the choice being made based on the end-users needs.

In this paper, we present the design, the development, and the use of a dashboard targeted to a laboratory located in a teaching hospital, in charge of PCR test, following the COVID-19 outbreak, such as the IMU. This includes (i) to define the need, (ii) to build the dashboard, (iii) to deploy the tool, and (iv) to demonstrate the added value in terms of the quality and operations management for the laboratory mission. This research focuses on aspects other than epidemiological matters (patient type, pathogen, period of the year, etc.), and which can explain some variation of results in the laboratory. We split these aspects into two main categories: quality issues and management issues.

Materials and Methods

RT-PCR and Data

RT-PCR for the detection of SARS-CoV-2 from clinical specimen where achieved as previously described using our in-house molecular diagnostic platform (MDx platform), the Cobas SARS-CoV-2 test on the Cobas 6800 instrument (Roche, Basel, Switzerland), and the Xpert Xpress SARS-CoV-2 assay (Cepheid, CA, USA) (6, 34, 35). Viral load were obtained by conversion of the Ct (Cycle threshold) values of the instruments using either a plasmid containing the target sequence of the PCR obtained from RD-Biotech (Besançon, France) or using purified viral RNA, kindly provided by the Institute of Virology of the University of Berlin, la Charite (34, 35).

The dataset feeding the dashboard is an extract from Molis, an IS used at the IMU. The extract is performed every 4 h. The analyses of four respiratory virus of interest were flagged in the system: SARS CoV-2, Influenza A and B, and the Respiratory Syncytial Virus (RSV). A comma-separated values file (csv) with new observations validated in the past 4 h is uploaded in a specific folder to be read through the dashboard.

To date, the system has 148,667 observations of SARS-CoV-2 RT-PCR. For each sample (swab, blood, etc.), the available entries include a unique anonymous patient ID, date of birth and gender of the patient, its hospitalization status, an anonymized code of the entity requesting the test (doctors, clinics, other laboratories, etc.), the date-time of the sampling (when available), of the test, and of the result sending as well as the type of sampling (nasopharyngeal or oropharyngeal secretions, blood sample, etc.). The original dataset also contains analysis codes showing the test result (positive, negative, canceled, NOT VALID), the Ct values, the viral quantification (in copy per milliliter, cp/mL), and whether the analysis had to be repeated. Every test are associated with various internal control tests (e.g., extraction, inhibition, etc.). “NOT VALID” results correspond to analyses which do not pass one of the internal quality controls.

A sample is related to one patient but that a patient may be tested several times. The analysis codes correspond to a test, which is performed for a specific virus, on a specific device (machine used to perform the test), for a targeted gene.

Some cleaning and data wrangling were performed before building the dashboard. Using a matching table shared by the IMU, the analysis codes were renamed using more user-friendly structure (NOM.VIRUS_TYPE.ANALYSE_APPAREIL_GENE). Then, different measures are extracted, especially on date-time data: the reception duration is the difference between the sampling time and the reception time at the laboratory, the test duration is the difference between the reception time and the results validation time and the total duration is the sum of the last two. Using the date of birth, patients were categorized into age groups with a 10-year window. Similarly, the type of sampling was recoded using wider groups categories, the most present being NPS (nasopharyngeal secretions). Then, each analysis is assigned a “Virus” and a “Device,” corresponding to the non-empty analysis codes described above. Another “Confirmation” variable is added, showing whether the analysis had to be reperformed.

Finally, four different tables are created, corresponding to each virus present in the dashboard.

Cleaning and wringling is done automatically every time the dashboard is launched, without any action from the end-user.

Platform

The interactive dashboard was built using Shiny (36), based on the opensource programming language R 4.0.2 (37) in the RStudio Interactive Development Environment (38). CSS and JavaScript were also used to tailor the dashboard to the needs of the end-users. The dashboard relies on several packages: shinydashboard 0.7.1 (39) for the page structure, plotly 4.9.3 (40), and ggplot2 3.3.3 (41) for interactive graphs, DT 0.17 (42) for interactive tables, shinyjs 2.0.0 (43) for some custom interaction, shinyWidgets 0.5.6 (44) for input objects, and readx 1.3.1l (45), plyr 1.8.6 (46), dplyr 1.0.7 (47), lubridate 1.7.9.2 (48), tidyr 1.1.2 (49), stringr 1.4.0 (50), and forcats 0.5.0 (51) for data wrangling. The dashboard is built in two dimensions (Figure 1). The horizontal dimension is the Target and the vertical dimension, the Action. The level of detail and the amount of possibilities increase when the user goes down on the Action sections.

FIGURE 1
www.frontiersin.org

Figure 1. Dashboard structure. The dashboard is organized in three main parts: Filter page, Comparison page, and Data page. On the filtering page, key indicators are available such as the number of samples, time to results, positivity rate, percent of invalid tests, etc. These indicators are provided by default for the whole dataset, but are also available for subgroups, according to the filtering criteria applied to the whole dataset. Thus, it is possible in a click to observe specifically the tests done during a given period or done on a dedicated instrument. It is also possible to select only the analysis performed for a given requester. The Comparison page offers more than 182 combinations of the KPI. The Data page, gives access to the raw data in a table format that can be downloaded; filters can be applied to choose a subset of data of interest.

As stated above, picking the appropriate visualization can be challenging and is subject to various principles. The main components chosen to build the dashboard are: (1) infoboxes: this type of visualizations gives an immediate information on some key metrics. Placed at the top of the dashboard, the user directly sees crucial information. When having several of these, it is important to group and label them appropriately (52), (2) tables: columns describe a specific attribute for each row, showing the user a detailed view, ready for a deeper inspection. Filtering and ordering options are available to display the part of interest of the data and (3) plots: we used both Abela (53) and the data-to-viz.com tool created by Holtz (54), which provide clear tools to select the appropriate chart relative to the data. This dashboard mostly uses Column charts and Stacked Column charts, Scatter plots, Line charts, and Boxplots. The latter has the advantage of displaying many information at once, and the end-users are accustomed to this format.

Finally, for a quicker adoption and an optimal usage, the dashboard was built in French, the mother tongue of the end-users. The dashboard runs locally on every user's computer, in order to prevent from security issues. The data is stored on the secured CHUV's servers, accessible only when authenticated on this network. Its use is restricted to authorized staff members only, whose mission is closely linked to quality management.

Ethics

The data were obtained during a quality enhancement project at our institution. According to national law, the performance and publishing the results of such a project can be done without asking the permission of the competent research ethics committee.

Results

Structure and Main KPIs of the Dashboard

Global Structure of the Dashboard

The content of the dashboard was guided by three main principles: comprehensive initial brief, multiple feedback loops, and close monitoring of latest breakthrough discoveries about Sars-CoV-2. At the beginning of the process, the end-users, namely the managers of the laboratory of molecular diagnostic, formulated their main needs, and made suggestions on what to report in the dashboard. This included the main KPIs, such as the number of tests per day as well as some of the inputs that could be entered by the user, for instance filters such as RT-PCR instruments or patients' gender or age. In an AGILE-like methodology, applying continuous improvement (55), demos were regularly performed, and feedbacks shortly implemented. The SARS-CoV-2 being a novel virus, the scientific literature was closely scrutinized to incorporate new relevant elements, such as gender-related indicators, which were included after (56) published their research on gender bias in COVID-19 diagnosis.

The current version of the dashboard contains three main pages detailed in Figure 1: (i) “Filter,” (ii) “Comparison,” and (iii) “Data.” The Filter page allows the user to select inputs such as a date range, the gender, the age, the test result, the hospitalization status, the device used, the confirmation status, the type of sample, and the type of requesters. In the Comparison page, the user selects variables to appear on graphs, but also filters the dataset to narrow down comparison subjects. Finally, the Data page also lets the user filter the dataset to look at individual observations. Overall, the user therefore has a role in filtering observations and in deciding which information to be represented (Figure 1).

The Filter Page

The Filter page presents a total of 10 KPIs and is divided into five sections: general and gender-related indicators, number of tests and positivity rate, viral load, test durations, and “NOT VALID.”

General and Gender-Related Indicators

The 10 KPIs are presented in aggregate form in this section, some of them being detailed in the subsequent sections. The most general figures are shown at the top of the layout. This first section is split into two groups. The “General Indicators” groups (i) three workload-related KPIs: the number of tests performed for the selected virus, the proportion of the tests after filtering, the number of positive tests, the positivity rate, the average test duration displayed next to the median duration, (ii) three quality related KPIs: the average number of tests per patient, the percentage of patient having to take two or more tests, the proportion of tests marked as “NOT VALID,” and (iii) gender-related Indicators including the sex ratio (positive women out of positive men), the average age, the positivity rate, and the hospitalization rate (Supplementary Figures 1A,B).

Number of Tests and Positivity Rate

In this section, the number of tests and the proportions of daily positive specimen are detailed. The number of tests is presented per day of reception at the laboratory, per date of sending the results, and per age category. The proportions are presented per day and week of reception. In addition, an interactive table shows the number of tests per top categories of requesters. The plots are interactive to avoid cluttered visualization while allowing the user to choose the appropriate representation (zoom, labels, etc.) (Figures 2A–D).

FIGURE 2
www.frontiersin.org

Figure 2. Screen shots of the upper panel of the Filter page. (A) Number of tests and test results per date of results sending. (B) Weekly number of tests and positivity rate. (C) Panel (A) zoomed on the label shown when hovering. (D) Panel (B) zoomed on the label shown when hovering.

Durations of the Tests

In this section, the user finds the hours of reception of the tests and of the sending of the results, along with the test durations. These are presented by test results (positive or negative); the number of tests is displayed for each hour of a 24-h day (Figures 3B,C). The average test durations and the average reception durations are displayed below for each day of sending, respectively, day of reception.

FIGURE 3
www.frontiersin.org

Figure 3. Screen shots of the middle panel of Filter page showing (A) the number of tests and test results, per age group, (B) the number of tests per hour of reception, and (C) the number of tests per hour of results sending.

Viral Load

This section focuses on the viral load and Cycle Threshold (Ct), a KPI not presented in the first section. First, the median Ct for each day is showed, grouped by analysis (each device, each gene, each repetition); in total 12 analyses (Figure 4). The viral load is also available in copy/millimeter (cp/mL) since April 9th, 2020, date from which the laboratory started to keep this record. These are shown in a scatterplot crossed by type of sampling (blood, in nasopharyngeal secretions, etc.) as well as in a summary statistics table (Supplementary Figure 3).

FIGURE 4
www.frontiersin.org

Figure 4. Screen shots of the lower panel of the Filter page. Cycle threshold (see section RT-PCR and Data) time series plot, per date of reception.

Not Valid

Finally, a descriptive table of the cases for which the result of the analysis was “NOT VALID” is displayed. The user can then identify any issue and investigate further. For sake of readability only some variables are shown (Supplementary Figure 4).

Comparison Page

When moving to the Comparison page, the user sees first a global summary for all viruses. It shows the daily number of tests for each disease as well as the corresponding daily positivity rate. Then, like the Filter page, each virus information is shown in a specific tab. For each virus, the user can select the x-axis variable and the coloring variable. An error message is displayed when the chosen variables are the same (Table 1).

TABLE 1
www.frontiersin.org

Table 1. Available variables for custom charting in the comparison page.

Upon this choice, the absolute and the relative number of tests are plotted, such as shown in Supplementary Figure 4 for gender and week of reception. The following plot shows data about the test duration. An example in Figure 5 shows the average test duration per instrument, depending on the result of the test. This is especially useful to control the speed of some devices. Finally, the Ct information is displayed using boxplots (Figure 6A), a familiar representation to the end-users. All the described figures are common to all viruses. For SARS-Cov-2, an additional boxplot of the viral load in log10 scale is displayed (Figure 6B). This additional boxplot appears only for this virus whose load can vary from 10,00 to 14,00,00,00,000 cp/mL. The log10 scale provides a more readable graph.

FIGURE 5
www.frontiersin.org

Figure 5. Screen shot of the Comparison page that allows 182 different combinations; focus on tests duration. (A) Average test duration, per week of reception and test result, for all tests. (B) Average test duration, per device and test result, for all tests.

FIGURE 6
www.frontiersin.org

Figure 6. Screen shot of the Comparison page that allows 182 different combinations, focus on cycle thresholds (Ct). (A) Cycle threshold boxplots per hospitalization status, per week of reception. (B) Comparison page—Viral load boxplots per hospitalization status, per week of reception.

Data Page

On the Data page, the users can have a look at the raw data in a table format, after filtering if wanted, thus deepening the analysis of an issue located with the previous pages tools. They can then download the data and explore it in Excel. This simple feature appeared to be surprisingly useful to the users who can then extract data more quickly than with a data base request from Molis. As an example, Figure 7 is a graph showing weekly number of tests linked to lab events (new machines acquisition), extracted from the Data page.

FIGURE 7
www.frontiersin.org

Figure 7. SARS-CoV-2 RT-PCR systems introduced in our laboratory. RT-PCR for the detection of SARS-CoV-2 from clinical specimen where achieved using several instruments introduced in our laboratory. The in-house molecular diagnostic platform (Molecular Dx platform) and the cobas SARS-CoV-2 instrument are two high-throughput RT-PCR systems while the Xpert Xpress SARS-CoV-2 assay (Cepheid, CA, USA) and the cobas SARS-CoV-2-Influenza A/B test are rapid RT-PCR systems used for emergencies Adapted from Opota et al. (6) and Marquis et al. (8).

Operational Management Implications

The general KPIs (e.g., number of daily tests, average number of tests per patient, etc.), such as Supplementary Figure 2A, appearing at the very top of the dashboard are used to continuously reassess the current situation. Seeing the evolution of the past few days or weeks helps the staff adjust their resources: launch some recruitment or send back the technicians from other department, adjust the repartition of staff among the laboratory, handle material shortage by anticipating future volumes, etc.

On a regular basis, the KPIs regarding the durations of the test allow the detection of operational issues. For instance, the dashboard led to the discovery of a huge amount of samples received at the end of the business day, and helped tune the process of deliveries at the laboratory. This is central to meet the needs of the hospital, of the requesters and, at the end, of the patients themselves. The tool also supports the team in charge of identifying potential causes of delay: by being able to look into various combination and level of details, it is possible to point out these causes during the whole process (from sampling, to testing, to result sending). They hence can then be investigated individually (e.g., by requester) or more globally (e.g., by day of the week). Figures 3B,C are used on a daily basis to continuously reassess this crucial aspect of timing.

The dashboard also comforted some intuitions the institute had from its experience: it helped identify the peak hours, adapt the distribution of the employees, and make recommendation to major requesters in order to smooth the operational activity during the day.

Finally, the daily tracking of these KPIs allowed the inventory monitoring of materials and avoided an unexpected shortage of scarce test resources.

Quality Management Implications

Positivity Rate

Having the possibility to observe the positivity rate by instrument can help diagnose any failure, such as contamination. For instance, an increase in the positivity rate could be due to a contamination of the test reagent. Easily accessing this information (Supplementary Figure 2B for an example), being able to quickly analyze it and linking it to Ct values provides a critical advantage in maintaining high quality results over the course of the outbreak. Since the users of the dashboard can also look at the data per requester, or requester type, in case of a sudden variation in the positive rate test results, it allows them to look for non-epidemiologic reasons. Possible explanations include change of sampling methodologies, change of patient type (e.g., from mostly children to elderly patients), addition of new facilities in some requester categories. They can also investigate the geographical origin of the test and share relevant information with the appropriate authorities.

Finally, being able to verify the number of positive specimens by age category is critical, as the patients' age is playing a key role in the pandemic (spreading role of children, disease evolution among elderly, etc.). Figure 3A shows the number of tests per age group, which can be filtered for specific periods of time, or different requesters.

Aside from prositive and negative tests, the number of NOT VALID cases is closely monitored. As shown in Supplementary Figure 1, the proportion of NOT VALID tests is very low, which indicates high quality analyses.

Viral Load

Among various data in the dashboard, the viral load especially helps guarantee the accuracy of the test or conversely identify and solve analytic problems. For instance, any sudden drop or jump in the daily or weekly median viral load raised a warning and calls for explanation (e.g., change of testing strategy, population target). If no explanation is found, one could suspect a problem at the analytical stage (i.e., pre-analytical stage, virus mutation etc.). A major viral load drop could mean a decrease in analytic sensitivity, which could be dramatic during an outbreak. A decrease in the median viral load was observed in April 2020 which led to a modification of the testing strategy, namely a universal testing strategy (35).

Similarly, tracking the virus loads per patient types helps addressing the test sensitivity. Indeed, since the institute performs tests on a large variety of patients, it needs to track this measure to adjust the tests sensitivity in real time, if relevant.

Additionally, the viral load by type of specimen (nasopharyngeal secretions, blood samples, throat secretions...) helps at addressing the performance of the test in different test of specimens (34, 57).

The viral load by instrument gives information on the analytical performance of each methods. Indeed, any drop out or increase could be the result of virus mutation or contamination.

Discussion

To respond to the COVID-19 pandemic, our laboratory had to introduce several methods to detect SARS-CoV-2 in clinical samples by RT-PCR within a short time. In fact, this sensitive method is reliable to control the spread of the virus by identifying infected patients, including asymptomatic subjects (35, 58). As a result, the number of RT-PCR tests performed each day increased rapidly with an average of 350 tests per day with peaks beyond 1,000 tests per day during the first epidemic wave in Switzerland and more than 1,700 tests per day during the second wave.

Although the programming skills of the lab have improved, the dashboard alleviates the users from extracting the data, a long and potentially complex task which requires to query the IS team, and from pulling them to a statistical software like R or Excel to produce the useful representation. A potential gain of time and productivity in a public hospital context where the resources are scarce and the staff is under the day-to-day task pressure. Consequently, working with the dashboard improved the functioning of the laboratory both in term of management of operations and in term of quality. This improvement was reflected in the reactivity and decision making resulting from the real-time interpretation of indicators. Indeed, this emerging virus posed a large number of challenges for diagnostic laboratories, which had to quickly introduce new diagnostic methods and to adapt to the challenges faced during the different stages of the pandemic.

Reliability of the RT-PCR Tests

The first challenge for diagnostic laboratory was to ensure the reliability of the RT-PCR assays. This was especially important because it is the first time that a test has been used with such a high throughput soon after its introduction. In this, the statistical analyzes proposed by the dashboard help to improve the reliability of these tests. For example, the real-time monitoring of the positivity rate of the tests or of the viral load in the clinical samples enables to highlight analytical problems, which could lead to false positives or false negatives. Using the dashboard, the positivity rate was compared according to the different instruments to check whether one instrument might be associated with a systematic error. However, cautious interpretation of the data available through the dashboard is mandatory since differences might also be due to various other factors. For example, because highly symptomatic subjects admitted at the emergency ward of the hospital were tested with the GeneXpert rapid RT-PCR, the tests performed with that instruments were more often positive.

Management of Multiple Instruments and Stocks

We also had to introduce several different RT-PCR methods. The goal was to guarantee continuity of service in the case of instrument failure or reagent shortage (6), but also to respond to a need for faster tests, notably for analyzes carried out in emergencies to avoid nosocomial infections (35). In this context of a fleet of several RT-PCR instruments the dashboard allows real-time monitoring of laboratory activity in terms of tests carried out on each instrument. This monitoring has enabled us, for example, to identify a risk of significant penury of the rapid molecular test. This problem was solved by quickly introducing a second test.

Time to Result

A challenge for the laboratory was to adapt to the needs of clinicians for the patients care and to be useful for public health decisions. This involved a short rendering time. The dashboard allows to dynamically monitor the time of arrival of samples at the laboratory and thus to adjust the work organization. Typically, we added human resources at the end of the afternoon to allow more same-day results.

In its current version, the code is still under development and observes from several limitations. First, it is specific to the practice of our laboratory and difficult to generalize. Secondly, its performances could be improved: slow to refresh, and unlikely to support database growth on the long term.

On the analytical aspect, the dashboard is a first step to develop models leading to implement (i) automatic alerts (when should resources be increased or decreased) using statistical process control, (ii) predictions (number of positive rates, the number of tests, etc.) using time series forecasting and exogenous information (e.g., other laboratories data, governmental decisions, etc.), and (iii) scenario stressing (what would happen if the authorities change the sanitary restrictions?). An original and interesting potential of development could be the use of multivariate forecasting techniques since the app allows to see and extract several virus data whose correlations may be exploited.

Finally, to better evaluate the impact of such tools on the laboratory management, quantitative evaluation should be performed now that the dashboard is fully used by the IMU's staff. We can expect that this type of tool will also help to consolidate the user's intuitions on the long term.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author/s.

Author Contributions

EM, M-OB, GG, VC, KJ, and OO contributed substantially in the conception or design of the work and in the acquisition, analysis, or interpretation of data for the work, in the drafting of the work or revising it critically for important intellectual content, in the final approval of the version to be published, and in the agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We thank all the staff of the Laboratory of Molecular Diagnostic of the Institute of Microbiology of the University of Lausanne and in particular Miss Zahera Naseri and Dr. Damien Jacot. We also thank Mr. Fabien Faverjon, Mr. Frank Hottin, and the IT and Data Team for their support.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdgth.2021.773986/full#supplementary-material

Supplementary Figure 1. Screen shots of the Filter page. (A) Number of tests and test results per date of reception. (B) Positivity rate per date of reception.

Supplementary Figure 2. Screen shots of the Filter page. (A) Number of tests and test results per date of reception. (B) Positivity rate per date of reception.

Supplementary Figure 3. Screen shots of the Filter page. (A) Viral load time series plot per sampling type, per date of reception and (B) descriptive statistics table. AS, anal swab; BAL, bronchoalveolar lavage; EXPECT, expectoration; NPS, nasopharyngeal swab; TS, throat swab; NA, non-applicable.

Supplementary Figure 4. Screen shots of the Comparison page that allows 182 different combinations, focus on gender KPI. (A) Number of tests per gender, per week of reception and (B) proportion of each gender, per week of reception.

References

1. World Health Organization. Statement on the Second Meeting of the International Health Regulations (2005) Emergency Committee Regarding the Outbreak of Novel Coronavirus (2019-nCoV) (2020).

Google Scholar

2. Tadini E, Papamidimitriou-Olivgeris M, Opota O, Moulin E, Lamoth F, Manuel O, et al. SARS-CoV-2, a point in the storm. Rev Med Suisse. (2020) 16:917–23.

Google Scholar

3. Federal Office of Public Health. New Coronavirus 2019-nCoV: First Confirmed Case in Switzerland (2020).

Google Scholar

4. World Health Organization. WHO | Pneumonia of Unknown Cause? China (2019).

Google Scholar

5. World Health Organization. WHO Statement on Novel Coronavirus in Thailand (2020).

Google Scholar

6. Opota O, Brouillet R, Greub G, Jaton K. Comparison of SARS-CoV-2 RT-PCR on a high-throughput molecular diagnostic platform and the cobas SARS-CoV-2 test for the diagnostic of COVID-19 on various clinical samples. Pathogens Dis. (2020) 78:ftaa061. doi: 10.1093/femspd/ftaa061

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Greub G, Sahli R, Brouillet R, Jaton K. Ten years of R&D and full automation in molecular diagnosis. Future Microbiol. (2016) 11:403–25. doi: 10.2217/fmb.15.152

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Marquis B, Opota O, Jaton K, Greub G. Impact of different SARS-CoV-2 assays on laboratory turnaround time. J Med Microbiol. (2021) 70:001280. doi: 10.1099/jmm.0.001280

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Wikipedia Contributors. Dashboard (Business). Wikipedia, The Free Encyclopedia (2020).

Google Scholar

10. Eckerson WW, editor. The landscape for performance dashboards. In: Performance Dashboards: Measuring, Monitoring, and Managing Your Business. Hoboken, NJ: John Wiley & Sons (2010). p. 4.

Google Scholar

11. Keim D, Andrienko G, Fekete J-D, Görg C, Kohlhammer J, Melançon G. Visual analytics: definition, process, and challenges. In: Kerren A, Stasko JT, Fekete JD, North C, editors. Information Visualization: Human-Centered Issues and Perspectives. Berlin; Heidelberg: Springer Berlin Heidelberg (2008). p. 154–75. doi: 10.1007/978-3-540-70956-5_7

CrossRef Full Text | Google Scholar

12. Cook K, Thomas J. Illuminating the Path: The Research and Development Agenda for Visual Analytics. Washigton, DC (2005).

Google Scholar

13. Menon NM, Lee B, Eldenburg L. Productivity of information systems in the healthcare industry. Inform Syst Res. (2000) 11:83–92. doi: 10.1287/isre.11.1.83.11784

CrossRef Full Text | Google Scholar

14. Pestana M, Pereira R, Moro S. Improving health care management in hospitals through a productivity dashboard. J Med Syst. (2020) 44:87. doi: 10.1007/s10916-020-01546-1

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Koumpouros Y. Balanced scorecard: application in the General Panarcadian Hospital of Tripolis, Greece. Int J Health Care Qual Assurance. (2013) 26:286–307. doi: 10.1108/09526861311319546

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Al-Hajj S, Pike I, Fisher B. Interactive dashboards: using visual analytics for knowledge transfer and decision support. In: Proceedings of the 2013 Workshop on Visual Analytics in Healthcare. Washington, DC (2013).

Google Scholar

17. Stadler JG, Donlon K, Siewert JD, Franken T, Lewis N. Improving the efficiency and ease of healthcare analysis through use of data visualization dashboards. Big Data. (2016) 4:129–35. doi: 10.1089/big.2015.0059

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Ward MJ, Marsolo KA, Froehle CM. Applications of business analytics in healthcare. Bus Horizons. (2014) 57:571–82. doi: 10.1016/j.bushor.2014.06.003

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Cheng CK, Ip DK, Cowling BJ, Ho LM, Leung GM, Lau EH. Digital dashboard design using multiple data streams for disease surveillance with influenza surveillance as an example. J Med Int Res. (2011) 13:e85. doi: 10.2196/jmir.1658

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Dong E, Du H, Gardner L. An interactive web-based dashboard to track COVID-19 in real time. Lancet Infectious Dis. (2020) 20:533–4. doi: 10.1016/S1473-3099(20)30120-1

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Probst D. Communicating near real-time data during the COVID-19 pandemic. Chimia. (2020) 74:613–4. doi: 10.2533/chimia.2020.613

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Dixit RA, Hurst S, Adams KT, Boxley C, Lysen-Hendershot K, Bennett SS, et al. Rapid development of visualization dashboards to enhance situation awareness of COVID-19 telehealth initiatives at a multihospital healthcare system. J Am Med Inform Assoc. (2020) 27:1456–61. doi: 10.1093/jamia/ocaa161

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Grange ES, Neil EJ, Stoffel M, Singh AP, Tseng E, Resco-Summers K, et al. Responding to COVID-19: the UW medicine information technology services experience. Appl Clin Inform. (2020) 11:265–75. doi: 10.1055/s-0040-1709715

PubMed Abstract | CrossRef Full Text | Google Scholar

24. de Lusignan S, Bernal JL, Zambon M, Akinyemi O, Amirthalingam G, Andrews N, et al. Emergence of a novel coronavirus (COVID-19): protocol for extending surveillance used by the Royal College of General Practitioners Research and Surveillance Centre and Public Health England. JMIR Public Health Surveill. (2020) 6:e18606. doi: 10.2196/18606

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Azadmanjir Z, Torabi M, Safdari R, Bayat M, Golmahi F. A map for clinical laboratories management indicators in the intelligent dashboard. Acta Inform Med. (2015) 23:210–4. doi: 10.5455/aim.2015.23.210-214

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Hill HM. Measuring productivity in bioanalysis. Bioanalysis. (2012) 4:2317–9. doi: 10.4155/bio.12.207

PubMed Abstract | CrossRef Full Text | Google Scholar

27. O'Donnell E, David JS. How information systems influence user decisions: a research framework and literature review. Int J Account Inform Syst. (2000) 1:178–203. doi: 10.1016/S1467-0895(00)00009-9

CrossRef Full Text | Google Scholar

28. Kaplan SE. An examination of the effect of presentation format on auditors' expected value judgments. Account Horizons. (1988) 2:90.

Google Scholar

29. Davis FD. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. (1989) 13:319–40. doi: 10.2307/249008

CrossRef Full Text | Google Scholar

30. Dilla WN, Steinbart PJ. The effects of alternative supplementary display formats on balanced scorecard judgments. Int J Account Inform Syst. (2005) 6:159–76. doi: 10.1016/j.accinf.2004.12.002

CrossRef Full Text | Google Scholar

31. Vance Wilson E, Zigurs I. Decisional guidance and end-user display choices. Account Manage Inform Technol. (1999) 9:49–75. doi: 10.1016/S0959-8022(99)00003-X

CrossRef Full Text | Google Scholar

32. Vessey I. Cognitive fit: a theory-based analysis of the graphs versus tables literature. Dec Sci. (1991) 22:219–40. doi: 10.1111/j.1540-5915.1991.tb00344.x

CrossRef Full Text | Google Scholar

33. Lengler R, Eppler MJ. Towards a periodic table of visualization methods of management. In: Proceedings of the IASTED International Conference on Graphics and Visualization in Engineering. Clearwater, FL: ACTA Press (2007). p. 83–88.

Google Scholar

34. Jacot D, Greub G, Jaton K, Opota O. Viral load of SARS-CoV-2 across patients and compared to other respiratory viruses. Microbes Infect. (2020) 22:617–21. doi: 10.1016/j.micinf.2020.08.004

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Moraz M, Jacot D, Papadimitriou-Olivgeris M, Senn L, Greub G, Jaton K, et al. Universal admission screening strategy for COVID-19 highlighted the clinical importance of reporting SARS-CoV-2 viral loads. New Microbes New Infect. (2020) 38:100820. doi: 10.1016/j.nmni.2020.100820

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Chang W, Cheng J, Allaire JJ, Xie Y, McPherson J. Shiny: Web Application Framework for R (2020). Available online at: https://shiny.rstudio.com/reference/shiny/1.4.0/shiny-package.html (accessed September 1, 2021).

Google Scholar

37. R Core Team. R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing (2019)

Google Scholar

38. RStudio Team. RStudio: Integrated Development Environment for R. Boston, MA: RStudio, PBC (2020).

Google Scholar

39. Chang W, Borges Ribeiro B. shinydashboard: Create Dashboards with 'Shiny' (2018).

Google Scholar

40. Sievert C. Interactive Web-Based Data Visualization with R, plotly, and shiny. Chapman and Hall/CRC (2020). doi: 10.1201/9780429447273

CrossRef Full Text | Google Scholar

41. Wickham H. ggplot2: Elegant Graphics for Data Analysis. New York, NY: Springer-Verlag (2016). doi: 10.1007/978-3-319-24277-4_9

CrossRef Full Text | Google Scholar

42. Xie Y, Cheng J, Tan X. DT: A Wrapper of the JavaScript Library “DataTables” (2020). Available online at: https://CRAN.R-project.org/package=DT

Google Scholar

43. Attali D. shinyjs: Easily Improve the User Experience of Your Shiny Apps in Seconds (2020). Available online at: https://CRAN.R-project.org/package=shinyjs

Google Scholar

44. Perrier V, Meyer F, Granjon D. shinyWidgets: Custom Inputs Widgets for Shiny (2020). Available online at: https://CRAN.R-project.org/package=shinyWidgets

Google Scholar

45. Wickham H, Bryan J. readxl: Read Excel Files (2019). Available online at: https://CRAN.R-project.org/package=readxl

Google Scholar

46. Wickham H. The split-apply-combine strategy for data analysis. J Stat Softw. (2011) 40:1–29. doi: 10.18637/jss.v040.i01

CrossRef Full Text | Google Scholar

47. Wickham H, Francois R, Henry L, Müller K. dplyr: A Grammar of Data Manipulation (2020). Available online at: https://CRAN.R-project.org/package=dplyr

PubMed Abstract | Google Scholar

48. Grolemund G, Wickham H. Dates and Times Made Easy with lubridate. J Stat Softw. (2011) 40:1–25. doi: 10.18637/jss.v040.i03

CrossRef Full Text | Google Scholar

49. Wickham H, Henry L. tidyr: Tidy Messy Data (2020). Available online at: https://CRAN.R-project.org/package=tidyr

PubMed Abstract | Google Scholar

50. Wickham H. stringr: Simple, Consistent Wrappers for Common String Operations (2019). Available online at: https://CRAN.R-project.org/package=stringr

Google Scholar

51. Wickham H. forcats: Tools for Working with Categorical Variables (Factors) (2020). Available online at: https://CRAN.R-project.org/package=forcats

Google Scholar

52. Kerzner H. Dashboards. In: Project Management Metrics, KPIs, and Dashboards: A Guide to Measuring and Monitoring Project Performance. New York, NY: John Wiley & Sons (2017). p. 323–9. doi: 10.1002/9781119427599

PubMed Abstract | CrossRef Full Text | Google Scholar

53. Abela A. Advanced Presentations by Design: Creating Communication That Drives Action. San Francisco, CA: John Wiley & Sons (2008).

Google Scholar

54. Holtz Y. From Data to Viz | Find the Graphic You Need. (2018). Available online at: https://www.data-to-viz.com/ (accessed September 1, 2021).

Google Scholar

55. Wikipedia Contributors. Agile Software Development. Wikipedia, The Free Encyclopedia (2020).

Google Scholar

56. Ancochea J, Izquierdo JL, Soriano JB. Evidence of gender bias in the diagnosis and management of COVID-19 patients: a big data analysis of electronic health records. medRxiv [preprint]. (2020). doi: 10.1101/2020.07.20.20157735

CrossRef Full Text | Google Scholar

57. Mueller L, Scherz V, Greub G, Jaton K, Opota O. Computer-aided medical microbiology monitoring tool: a strategy to adapt to the SARS-CoV-2 epidemic and that highlights RT-PCR consistency. Front Cell Infect Microbiol. (2020) 11:594577. doi: 10.1101/2020.07.27.20162123

PubMed Abstract | CrossRef Full Text | Google Scholar

58. Caruana G, Croxatto A, Coste AT, Opota O, Lamoth F, Jaton K, et al. Diagnostic strategies for SARS-CoV-2 infection and interpretation of microbiological results. Clin Microbiol Infect. (2020) 26:1178–82. doi: 10.1016/j.cmi.2020.06.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: COVID-19, medical microbiology, dashboard, digitalization, operations management, quality management, Key Performance Index - KPI

Citation: Maury E, Boldi M-O, Greub G, Chavez V, Jaton K and Opota O (2021) An Automated Dashboard to Improve Laboratory COVID-19 Diagnostics Management. Front. Digit. Health 3:773986. doi: 10.3389/fdgth.2021.773986

Received: 10 September 2021; Accepted: 12 November 2021;
Published: 06 December 2021.

Edited by:

Harry Hochheiser, University of Pittsburgh, United States

Reviewed by:

Pablo Gallardo, University of Chile, Chile
Xia Jing, Clemson University, United States

Copyright © 2021 Maury, Boldi, Greub, Chavez, Jaton and Opota. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Onya Opota, onya.opota@chuv.ch

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.