Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 03 May 2022
Sec. Auditory Cognitive Neuroscience
This article is part of the Research Topic Outcome Measures to Assess the Benefit of Interventions for Adults with Hearing Loss: From Research to Clinical Application View all 14 articles

Defining a Patient-Centred Core Outcome Domain Set for the Assessment of Hearing Rehabilitation With Clients and Professionals

  • 1National Acoustic Laboratories, Sydney, NSW, Australia
  • 2Department of Linguistics, Macquarie University, Sydney, NSW, Australia
  • 3School of Health and Rehabilitation Sciences, The University of Queensland, Brisbane, QLD, Australia
  • 4Ear Science Institute Australia, Perth, WA, Australia
  • 5Curtin enAble Institute, Faculty of Health Sciences, Curtin University, Perth, WA, Australia

Background: A variety of outcome domains are currently measured for the assessment of hearing rehabilitation. To date, there is no consensus about which outcome domains should be measured, when they should be measured, and how they should be measured. In addition, most studies seeking to develop core outcome sets and measures for hearing rehabilitation services have primarily focussed on the opinions and expertise of researchers, and, to a lesser extent, clinicians, rather than also involving clients of those services. The principles of experience-based co-design suggest that health services, researchers, and policymakers should come together with clients and their families to design health services and define what metrics should be used for their success.

Objectives: This study aimed to seek views and consensus from a range of key stakeholders to define which client-centred self-report outcome domains should be measured, when they should be measured, and how they should be measured, in a national publicly funded hearing rehabilitation scheme. In addition, the study aimed to identify current and future potential mechanisms and systems to standardise the collection of data and reporting of outcomes, to enable comparison across clients and hearing service providers.

Methods: Two stakeholder groups participated in a three-round online Delphi process: (1) 79 professional stakeholders involved in the delivery of hearing services in Australia, and (2) 64 hearing rehabilitation services’ clients identified by not-for-profit consumer organisations. An initial set of in-person workshops scoped the key issues upon which to develop the initial open-ended questions and subsequent Likert-scale statements addressing these issues. These statements were then distributed to both groups in an online survey. The respondent ratings were summarised, and the summary was returned to respondents along with a second round of the survey. This process was then repeated once more. The five most important outcome domains from both groups were then combined, and a consensus workshop of seven professionals and three client advocates agreed on the top four ranked domains.

Results: A range of potential outcome domains were identified as relevant indicators of successful hearing rehabilitation. Communication ability, personal relationships, wellbeing, and participation restrictions were identified as a core outcome domain set that should be measured as a minimum for patients receiving hearing rehabilitation. There was little agreement on the preferred timepoints for collection of outcome measures, with respondents expressing the view that this should be established by research once a set of outcome measures has been selected. However, there was broad agreement that measurements of these domains should be collected at baseline (before the provision of hearing rehabilitation) and no earlier than 3 months following the completion of rehabilitation. Potential benefits and issues with the development of a national outcomes database/collection system were also identified and prioritised, with participants highlighting the importance of valid, high-quality, trustworthy, and comprehensive data collection.

Conclusion: These results provide a Core Outcome Domain Set for the self-reported evaluation of hearing rehabilitation and provide important background information for the design of methods to implement them across hearing healthcare systems. However, the wide range of outcome domains identified as potentially providing important additional information and the lack of specific measures to address these domains strongly suggest that there is still more research to be done. Ongoing stakeholder engagement will continue to be vital for future implementation. In addition, further research is required to determine the optimal time following hearing rehabilitation to utilise any particular outcome measure.

Introduction

Hearing loss is a chronic condition that affects around four million adults in Australia, which represents one in six of the population (Access Economics Pty Ltd, 2015). In addition, hearing loss can have substantial negative consequences, including activity limitations, participation restrictions, stigmatisation, reduced quality of life, and third-party disability (Chia et al., 2007; Wallhagen, 2010; Scarinci et al., 2012; Granberg et al., 2014b; Heffernan et al., 2016; Barker et al., 2017). Furthermore, hearing loss has been associated with depression, cognitive decline, and dementia (Lin, 2011; Dawes et al., 2015).

Auditory rehabilitation aims to address the negative impact of hearing loss and includes a range of interventions. The primary intervention is hearing aids, which have been shown to be clinically effective in terms of listening ability, hearing-related quality of life (i.e., participation) and health-related quality of life (Ferguson et al., 2017). There are other auditory rehabilitation interventions for adults with hearing loss, which include alternative listening devices such as hearables, communication and patient education, and auditory training (Wong and Hickson, 2012; Ferguson et al., 2016; Ferguson et al., 2019). However, systematic reviews on these interventions have identified a lack of high-quality evidence (Henshaw and Ferguson, 2013; Barker et al., 2016; Maidment et al., 2016; Ferguson et al., 2017; Lawrence et al., 2018) in part due to a lack of a “gold standard” outcome measure (Granberg et al., 2014a; Hall et al., 2019).

In order to assess the effectiveness of interventions for adults with hearing loss, irrespective of the intervention type, it is essential to have appropriate and sensitive outcome measures that are relevant to the outcome domains targeted for improvement by auditory rehabilitation (Ferguson and Henshaw, 2015; British Society of Audiology, 2016). These are essential to both measuring an individual’s progress toward desired goals, often as a result of an intervention, as well as evaluating the overall effectiveness of audiology services and providers of hearing healthcare. Careful consideration needs to be given to which outcome measures are most fit for purpose. For example, a measure that asks only about specific pre-determined situations may not be relevant to the individual, and may not be compatible with a goal-setting approach to rehabilitation that is person-centred and focussed on the individual (British Society of Audiology, 2016).

One of the major problems with measuring outcomes within auditory rehabilitation is the large number of tools and instruments, including behavioural and self-report measures (Granberg et al., 2014a). In particular, there are a huge number of self-report measures available, with one study identifying 139 hearing-specific questionnaires (Akeroyd et al., 2015). Another major problem is that there is no agreement amongst researchers and clinicians in the field regarding what outcomes should be measured and how they should be measured (PricewaterhouseCoopers Australia, 2017). A systematic review of outcome measures used in research demonstrated the extent of this problem (Granberg et al., 2014a), identifying 51 self-report outcome measurement instruments used across 122 adult hearing loss studies. Of these 51, only 16 instruments had been used in more than one study. It is perhaps not surprising then that a scoping review uncovered considerable heterogeneity in outcome measurement in randomised controlled trials of adult auditory rehabilitation interventions (Barker et al., 2015a).

Many of these measures measure similar underlying constructs, such as hearing device use, benefit, satisfaction, and social participation. In the context of hearing outcomes, these underlying constructs are known as outcome domains. However, even among outcome domains that are in widespread use and seen to be important indicators of successful rehabilitation, such as hearing aid use, there is little consensus around which outcome measures should be used (Perez and Edmonds, 2012). Furthermore, there is an increasing awareness globally that outcome domains that are not solely associated with hearing aid amplification and that address participation restrictions and psychosocial aspects should also be considered, such as wellbeing, identity, and emotion (Bennett et al., 2018; Heffernan et al., 2018a; Bennett et al., 2020; Vercammen et al., 2020). However, many of the most widely used standardised outcome measures, such as the International Outcomes Inventory for Hearing Aids (IOI-HA; Cox et al., 2000), do not address these broader and more recently identified outcome domains.

The evidence is clear that both auditory rehabilitation clinical practice and research lack a single (or even a few) outcome measure that is used widely and consistently and accepted as a “gold standard” instrument. Furthermore, even though there is a large number and variety of measures within the field, clinical trials of adult auditory rehabilitation interventions have overlooked outcomes such as adverse effects and quality of care that may be important to key stakeholders, especially patients, hearing healthcare professionals and commissioners of hearing healthcare (Ferguson et al., 2017). The involvement of these stakeholder groups in the development of such tools is rare, with some exceptions (Heffernan et al., 2018a; Heffernan et al., 2019), as typically it has been researchers alone who have developed outcomes.

A major consequence of an non-standardised approach to outcome measurement is that comparison across different patient cohorts and services is almost impossible. Similarly, within research, it is very difficult to compare and combine the results of different trials that use different measures (for example in systematic reviews with meta-analyses), which results in reduced relevance of the results and increased risk of outcome reporting bias (Ferguson et al., 2017).

Within the Australian hearing healthcare context, hearing services are provided free of charge to over one million people each year through the Hearing Services Program (HSP), primarily through the Voucher Scheme, at a cost of $590 million per annum (Commonwealth of Australia, 2019). The Voucher Scheme provides subsidised hearing services to eligible pensioners, Veterans, service people, and those receiving support for a disability that places their employment at risk (Department of Health, n.d.). Currently, as is seen in many other countries, standardised use of patient-centred outcome measures is not prevalent in Australian hearing healthcare, and typically outputs such as hearing aid uptake are used to measure the success of hearing aids for both clients and service providers (PricewaterhouseCoopers Australia, 2017). Although the importance of measuring client outcomes is highlighted in the regulatory framework of the HSP, typically the Australian-developed Client Orientated Scale of Improvement (COSI; Dillon et al., 1997) or the IOI-HA are used. While the COSI does involve recipients in the development of personalised items, potentially overcoming this limitation of the IOI-HA, its insensitivity makes it unsuitable for the measurement of service outcomes (Dillon et al., 1999).

A Government-commissioned review of the HSP published in 2017 found that the majority of key healthcare stakeholders (i.e., Contracted Service Providers, Device Manufacturers, consumer groups, research organisations) who were consulted agreed that client outcomes were important, however there was no consensus on how they should be measured (PricewaterhouseCoopers Australia, 2017). Four types of measurement methods were identified as in common use—the COSI, the IOI-HA, hearing aid datalogging, and speech testing—but none of these were used consistently. The recommendations from this review were to (i) move quickly toward an outcomes-based model rather than an outputs-based model (i.e., focussing on the number of rehabilitation programmes delivered and devices fitted), (ii) consult with key stakeholders to achieve a consensus on which outcomes should be used and to standardise the approach to measuring these, and (iii) identify how outcomes could be measured across service providers and client groups.

This current study aimed to identify and standardise a Core Outcome Domain Set (CODS): a set of outcome domains that should be used as a minimum standard for the assessment of a health condition (Hall et al., 2018), as well as when and how these domains should be assessed. A CODS can then form the basis for development of a core outcome set (COS): “an agreed standardised set of outcomes that should be measured and reported, as a minimum, in all clinical trials” (OMET Initiative, n.d.). The development of COSs has grown in stature in over the years and COSs are now a recommended component of clinical trial protocols, Cochrane reviews, and government funding applications (Williamson and Clarke, 2012; Kirkham et al., 2017). Within hearing rehabilitation, a roadmap to develop a COS for tinnitus treatment has been proposed, which stresses that a consensus is needed on what outcome domains should be measured, and then how this should be measured using an outcomes tool (Hall et al., 2015; Fackrell et al., 2017). The first of these steps is the development of the CODS; it is then the addition of standardised measurement tools that results in an implementable COS. The overall aim of the present study was to achieve the first of these steps to identify a Core Outcome Domain Set for self-report within hearing rehabilitation, in the Australian context.

The specific aims of this study were to:

1. Seek views and consensus from a range of key stakeholders to define which client-centred outcome domains should be used, when they should be measured, and how they should be measured, for the assessment of hearing rehabilitation delivered within a national publicly funded hearing rehabilitation scheme.

2. Identify current and future potential mechanisms and systems to standardise the collection of data and reporting of outcomes, to enable comparison across clients and hearing service providers.

Materials and Methods

This study was approved by the Hearing Australia Human Research Ethics Committee. Informed consent was obtained from all participants.

The overall structure of this research study is shown in Figure 1. Two groups of participants took part in this study: (i) Professionals, and (ii) Consumers. A scoping workshop and Delphi review was conducted with each group, and a final consensus workshop was conducted with representatives of both groups. The Delphi reviews covered six sections: Outcome Domains, Time of Collection, Methods of Collection, Parties Responsible for Collection, Reason for Collection, and National Outcomes Database. Where information from a previous stage was used to inform or develop a subsequent stage, this is denoted by an arrow.

FIGURE 1
www.frontiersin.org

Figure 1. Overall structure of the research study, including the number of participants in each workshop and Delphi round. Information that was used to develop subsequent stages is indicated by arrows.

Method

Scoping Workshops

Four in-person scoping workshops were used to establish the initial statements for the Outcome Domains section of the Delphi reviews. Three workshops were conducted with the Professionals group, one each in Sydney, Melbourne, and Brisbane. One workshop was conducted in Sydney with Consumers from around Australia, with travel costs covered by the research team.

At these workshops, structured brainstorming exercises were used to assist participants to identify a comprehensive long list of domains in which outcomes of hearing rehabilitation might be observed.

For the workshops conducted with the Professional group, the New South Wales Government Human Services Outcomes Framework was used to guide brainstorming (Routledge, 2017). This framework identifies seven broad areas in which outcomes of health and human services interventions might be observed: Education and Skills, Economic, Health, Home, Safety, Empowerment, and Social and Community.

For the workshop conducted with the Consumer group, participants were asked to define “personas” for people involved in hearing services, including people with hearing loss, their family members, clinicians, and policymakers. For each of these personas, participants then brainstormed markers of successful hearing rehabilitation (e.g., “My husband and I doing more things together”) and markers of unsuccessful hearing rehabilitation (e.g., “Me getting frustrated by having to repeat things”).

The lists of identified outcome domains from the workshops were then combined by the research team, and duplicates were removed.

Delphi Reviews

A Delphi review is an iterative process in which respondents are asked to complete a series of surveys (rounds), with subsequent rounds including summary information about the responses to the previous round, allowing participants to re-evaluate their previous rating of a statement based on any emerging group consensus (Helmer, 1967). The Delphi technique is useful for building consensus among experts with regard to their areas of expertise (Hsu and Sandford, 2007), and has been used successfully for the development of clinical guidelines and rehabilitation approaches in hearing healthcare (Barker et al., 2015b; Sereda et al., 2015; Ferguson et al., 2018). In the present Delphi Review, each round was conducted via an online survey.

Two Delphi reviews, each of three rounds, were conducted with the Professional and Consumer groups separately. These reviews utilised different surveys and slightly different questions, targeted to the two different populations. Where the same question was asked of both groups, summary information from both groups was presented, allowing respondents to use information from both groups of stakeholders in their re-evaluation (see Figure 1).

The Delphi reviews covered six sections. Three of these (Outcome Domains, Methods of Collection, and National Outcomes Database) were asked of both groups, and three (Time of Collection, Parties Responsible for Collection, and Reason for Collection) were asked only of the Professionals group, as it was felt by the research team that these would likely be out of the scope of understanding of non-professionals.

For both groups, the standard agreement rating item was a five-point Likert item with anchors Strongly agree, Agree, Neither agree nor disagree, Disagree, and Strongly disagree. The standard importance rating item was a five-point Likert item with anchors Very important, Important, Neither important nor unimportant, Unimportant, and Very unimportant. The standard comfortableness rating item was a five-point Likert item with anchors Very comfortable, Comfortable, Moderately comfortable, Slightly comfortable, and Not at all comfortable. The relevant rating item was chosen for congruence with the statements presented to participants; a full list of statements as presented is available in Supplementary Material.

Consensus throughout the Delphi reviews was defined a priori as a consensus percentage of 80% or greater for each item, that is the proportion of respondents rating a statement as follows:

• for agreement:

∘ Agree or Strongly Agree, or

∘ Disagree or Strongly disagree;

• for importance:

∘ Very important or Important, or

∘ Unimportant or Very unimportant.

Comfortableness was not used to determine consensus, but only to determine consensus rankings.

To allow for discrimination between items beyond whether they reached consensus or not, consensus rankings were calculated for some items, determined using the Kemeny-Young method (Kemeny, 1959). This method, which has to our knowledge not been used previously in hearing research, generates the consensus ranking(s) of items that have the largest number of total pairwise agreements with the rankings provided by participants. The number of pairwise agreements is calculated by taking every possible pair of items, determining which item is ranked higher in the potential consensus ranking, and then counting the number of participant-provided rankings that also rank that item higher. In some cases, such as where all respondents ranked items in the same order, this ranking is unique, although in some cases the method produces “ties,” where there is insufficient information in the data to be able to definitively place one item above another. This method was applied by treating each person’s responses as a single ranking, in which those items rated as Strongly Agree (or Very important, or Very comfortable) are ranked above those items rated as Agree (or Important, or Comfortable), and so on.

For example, consider a set of three items which have been ranked by participants. In this case, there are 13 possible orderings to consider as the consensus ranking. With the set of responses described in Table 1, the potential consensus ranking Item 1 > Item 2 > Item 3 has three pairwise agreements with Person A (Item 1 > Item 2, Item 2 > Item 3, and Item 1 > Item 3), two agreements with Person B (Item 1 > Item 2 and Item 1 > Item 3), and two agreements with Person C (Item 1 > Item 3 and Item 2 > Item 3) for a total of seven pairwise agreements. If this process is repeated for all 13 possible orderings (13 = 8 total orderings plus 5 combinations involving ties), it can be seen that this ranking (Item 1 > Item 2 > Item 3) has the highest total and is therefore the Kemeny-Young consensus ranking.

TABLE 1
www.frontiersin.org

Table 1. Synthesised responses to illustrate Kemeny-Young method.

Determination of a consensus ranking is one of the most difficult problems to solve computationally, as the number of possible consensus rankings (possibly with ties) that need to be checked grows superfactorially (due to ties) with the number of items to be ranked (Bartholdi et al., 1989; Biedl et al., 2006): while there are 13 possible orderings for three items, there are 75 for four items, 541 for five items, approximately 102 million for 10 items, and over 230 trillion for only 15 items. As a result, as the number of items that have been ranked increases, it becomes computationally not feasible to determine the optimal consensus ranking definitively, and heuristic methods that provide computational approximations to the optimal ranking are required. In this study, where small numbers of items were to be ranked (<15), the branch and bound algorithm (which definitively determines the optimal consensus ranking) could be computed in a reasonable time and was used (<4 h; Emond and Mason, 2002). Where a larger number of items were to be ranked, the fast computational approximation developed by Amodio et al. (2016) (stylised “FAST”) was used instead.

Consensus Workshop

A summary of results from the Outcome Domains, Time of Collection, and Methods of Collection sections of the Delphi Reviews was distributed to workshop participants prior to the workshop.

At the online workshop, the top five outcome domains as determined by the Consumer group and the top five outcome domains as determined by the Professional group were discussed. Due to similarities between outcome domains across the groups, seven were identified as separate constructs. These were then presented to the workshop and discussed in detail, to ensure that participants had a coherent shared understanding of each domain.

Participants then separated into two groups to discuss the domains and their importance. Project team staff attended these groups but did not participate in the discussion beyond answering questions about the methodology and the results.

Participants then individually and anonymously ranked the domains from most important to least important. The summary of the individual rankings was presented to the group and discussed until unanimous agreement was reached on a short list of domains that should be recommended for collection from all people receiving hearing services in Australia.

Some discussion was also had regarding methods and time of collection of outcome measures, which was synthesised qualitatively and is summarised below.

Final Recommendations

The research team met to synthesise the results from all phases of the work, primarily the Consensus Workshop and literature review of potentially applicable outcome measures, into specific interim recommendations for the assessment of hearing rehabilitation in clinical and research practice. Discussion continued until unanimous agreement was reached.

Participants

Professionals

This group comprised Professionals involved in the hearing industry in Australia. Potential participants were identified by the research team by brainstorming within each of the categories of hearing researchers, representatives of professional organisations, hearing service organisations, industry organisations such as hearing device manufacturers, and hearing consumer advocacy organisations. This resulted in a list of 59 people who were invited to take part in the study, with 43 (73%) consenting to take part.

Participants were invited to join in one of three initial scoping workshops, which were conducted in Brisbane (n = 6), Melbourne (n = 7), and Sydney (n = 7). Following these scoping workshops, several other potential participants were identified by workshop participants as being people who would be interested in contributing to the Delphi review, and were added to the list, giving a total of 79 potential participants for the Delphi review, of whom 50 (63%) completed at least one round of the Delphi review. Of these, 19 (38%) completed only one round, 16 (32%) completed only two rounds, and 15 (30%) completed all three. Participation in a future round was not contingent on completion of all previous rounds.

All categories used to identify potential participants were represented in the scoping workshops.

Consumers

Four hearing advocacy organisations active in Australia were identified (Better Hearing Australia Brisbane, Deafness Forum, Hearing Matters Australia, and Soundfair) and invited to nominate one or more representatives who identified as people with hearing loss. Seven representatives across all these groups took part in the initial scoping workshop.

Following the workshop, the organisations were invited to share a link to the first round of the Delphi review with their members, and all did so. Participants who responded to round 1 of the Delphi Review (n = 64) were then invited to participate in rounds 2 and 3. Of these, 55 (86%) completed all three rounds, 6 (9%) completed only two rounds, and 3 (5%) completed only round one. There was higher engagement with the review process in this group that in the Professional group, potentially due to the self-selected nature of the participants.

Final Consensus Group

A group of potential participants (n = 18) was selected by the research team from those in the Professional group who had responded to at least 2 rounds of the Delphi Review (n = 32). In the selection of these potential participants, the categories used at the potential participant identification stage were considered to ensure broad coverage of the Australian hearing industry. One potential participant from each consumer organisation was then added. The resulting 22 potential participants were then invited to take part in an online workshop (due to COVID-19), with 10 attending.

Material

The Delphi Reviews were conducted using three rounds of electronic surveys. A summary of the types of responses invited (e.g., ranking of importance) for each section of each survey can be seen in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Types of response invited for each section of the Delphi review surveys.

In round 1, participants were asked open-ended questions about the topics, which were then synthesised into statements for rating in subsequent rounds.

In rounds 2 and 3, statements were presented for rating using one of the standard items or ranking, along with summary information about responses to any previous rounds of ratings or rankings. In the Methods of Collection and National Outcomes Database sections, questions asked were similar between the two groups, and so summary information for both groups was presented.

In the Outcome Domains section, statements were rated using the standard importance item. As statements had already been determined during the scoping workshops, in Round 1 they were presented for rating, along with an open-ended question to allow participants to add any outcome domains that they felt were missing. In Round 3 participants were also asked to select, in ranked order, the top five domains, which were used to generate a consensus ranking. Due to the large number of outcome domains that reached consensus in Round 1 among Consumer participants, in Round 2 Consumers were presented with only the top 10 outcome domains from the previous round (as determined by the Kemeny-Young method), as well as five additional domains synthesised from open-ended responses. All of these reached consensus as being important, and so in Round 3 Consumer participants were not asked to rate, but only to rank the domains.

In the Time of Collection section, Professional participants were asked about the different time points at which outcomes could be collected, and why. Responses were synthesised into four major time points and a set of statements regarding the potential reasons why these time points might be useful. In Rounds 2 and 3, Professional participants were asked to rank the four time points in order of importance, and to rate their agreement with the statements using the standard agreement rating item. The ranked time points were used to generate a consensus ranking.

In the Methods of Collection section, Professional participants were asked about potential methods that could be used to collect outcomes, and the benefits and drawbacks that might be associated with each method. Professional participants were asked to rank the top five most important methods of collection in both Rounds 2 and 3 to facilitate the recommendation of a single method of collection at the close of the study. Consumer respondents were asked to rate the comfortableness of each method in Round 2, and to rank the top five most important methods in Round 3.

In the Parties responsible for collection section, Professional participants were asked about different parties (e.g., clinicians, a Government agency, or GPs) who could be responsible for the collection of outcome measures in the Hearing Services Program, and the potential benefits and drawbacks of each these parties undertaking outcomes collection. Statements were rated by Professional participants using the standard agreement rating item in subsequent rounds.

In the Reasons for Collection section, Professional participants were asked about reasons why different stakeholders might find it important for outcomes to be measured. Statements were rated by Professional participants using the standard agreement rating item in subsequent rounds.

In the National Outcomes Database section, Professional participants were asked about the potential beneficiaries, benefits, and drawbacks of the development of a national database of outcomes for hearing rehabilitation. The beneficiaries and benefits were synthesised into potential purposes for such a system. The potential purposes and drawbacks were then presented to both groups, and participants were asked to rate them using the standard importance rating item. In Round 2, Consumer participants were also asked how comfortable they would feel with a range of different stakeholders running a national outcomes database.

Questions asked in each section are available as Supplementary Material.

Results

Delphi Reviews

Results for this section are shown in Tables 38. In each table, items are ordered by the consensus ranking, when both consensus ranking and percentage are available, and then by the consensus percentage when ranking is tied (ordered from unanimous agreement to unanimous disagreement). Consensus percentages meeting the predefined criterion (80%) are shown in bold. In some cases, there is disagreement between the ordering implied by the consensus percentages and that obtained using the consensus ranking procedure, as the consensus percentage method treats “agreement” from one participant and “strong agreement” from another as equivalent. As the ranking takes individual preferences between domains into account, ranking should be considered a more accurate measure of consensus preference. However, as the ranking of an item is dependent on preferences for other items in the set, no strict criterion for ranking can be applied.

TABLE 3
www.frontiersin.org

Table 3. Results from the outcome domains section among the professional group.

TABLE 4
www.frontiersin.org

Table 4. Results from the outcome domains section among the consumer group.

TABLE 5
www.frontiersin.org

Table 5. Statements from the time of collection section among the professional group.

TABLE 6
www.frontiersin.org

Table 6. Rankings of methods of outcomes collection.

TABLE 7
www.frontiersin.org

Table 7. Statements from the parties responsible for collection section among the professional group.

TABLE 8
www.frontiersin.org

Table 8. Statements from the reasons for collection section among the professional group.

Outcome Domains

The primary question for this section was “What outcome domains should be measured as markers of success of hearing rehabilitation?”

Results from the Professional group are shown in Table 3. For each item, the consensus percentage and the consensus ranking from the final ranking task are shown. The consensus criterion was met for 13 domains and not for two domains.

Most domains that were identified through this process were psychosocial. Notably, the consensus criterion was not reached for the domain “Increased use of hearing aids.”

Results from the Consumer group are shown in Table 4. There was consensus beyond the predefined criterion on every domain presented. As there were several domains that were excluded after Round 1 and not included for Rounds 2 and 3, it is possible that a subset of these, should they have been presented, may also have reached the predefined criterion for consensus.

Time of Collection

The primary question for this section was “At what time point(s) should outcome measures be collected, and why?” It was conducted in two parts: by presenting respondents with four specific time points for ranking, and then by presenting a set of statements.

The four specific post-fitting time points identified were, in ranked order from most to least preferred, at 3 months following the fitting, at 6 months following the fitting, at 12 months following the fitting, and at the follow-up appointment (commonly conducted between one and 3 weeks post-fitting in Australia).

The statements and consensus percentages for this section are shown in Table 5. The highest consensus percentage at 93% related to the two statements on the use of baseline measures prior to device fitting, which was strongly supported by respondents. There was consensus reaching the criterion for three statements and not for four statements.

Methods of Collection

The primary question for this section was “What different methods could be used to collect outcome measures?”

Consensus rankings for this section are shown in Table 6. Note that the consensus ranking method was unable to distinguish between statements ranked third among respondents in the Consumers group for five of the statements. The orderings obtained from Professional group and from the Consumer group were notably different, with the second preference among Consumers (an online questionnaire emailed by the hearing care professional) being ranked last by the Professional respondent group.

Parties Responsible for Collection

The primary question for this section was “Thinking of patients being seen for hearing rehabilitation, who could potentially collect outcome measures?”

Statements and consensus percentages for this section are shown in Table 7. The predefined consensus criterion was not reached for any of the statements. As the results tended toward disagreement for some of the statements, whether the consensus percentage relates to agreement or disagreement is also shown in the table. It should be noted that as “Neither agree nor disagree” was a valid option for respondents, consensus percentage for agreement and consensus percentage for disagreement do not sum to 100%. For example, 33% consensus toward agreement displayed in the table indicates that fewer than 33% of respondents responded “Disagree” or “Strongly disagree,” with the remainder responding “Neither agree nor disagree.”

Reasons for Collection

The primary question for this section was “Why might it be important to clinicians providing hearing services/hearing service providers/Government that outcomes are measured?”

The statements and consensus percentages are shown in Table 8. Consensus reached the pre-defined criterion on 13 statements and did not reach the criterion on two statements. A broad array of potential reasons for collecting outcome measures with differing beneficiaries was identified, including hearing care professionals, Government, hearing care organisations, and the public.

National Outcomes Database

The primary questions for this section were: “Are there any people who you think might benefit from a national outcomes database? What are the potential benefits of a national outcomes database? What are the potential drawbacks associated with having a national outcomes database?” Detailed results of this section, including consensus percentages and rankings, are available in Supplementary Material, and a summary is provided below.

Both groups agreed that a database should be designed to promote person-centred care, to help determine best practice, and to provide a national standard for hearing care. Both groups also agreed that such a database should be designed to measure the impacts of hearing loss beyond the person themselves, on their partners, family members, and friends. The Consumer group, but not the Professional group, agreed that a database should support clients to choose hearing services and providers and help identify poorly performing clinicians and services.

Both groups were agreed that the accuracy and relevance of measures and the integrity of the data was highly important, with misuse of data by professionals or organisations a significant concern. Professional participants were also concerned with the potential for data breaches and use of the data to justify funding cuts.

Consumer participants felt more comfortable with organisations that might be considered independent from both the hearing industry and Government running a national outcomes database, including independent research organisations, professional associations, and universities.

Consensus Workshop

The synthesised domains as presented to the consensus workshop are available in Supplementary Material.

During initial discussion to ensure that participants understood the domains as presented, participants decided that “Improved participation in activities” should instead refer to reduction of “participation restrictions,” as it was felt that it was unreasonable to expect that the provision of hearing rehabilitation alone would result in increased participation by patients in social activities. Rather, participants felt that while hearing rehabilitation could reduce the barriers to participation caused by the hearing loss, the social and psychological effects of long-standing hearing difficulty may result in some continued persistence of patterns of reduced participation for a period of time following any reduction in participation restrictions.

Following this discussion, participants separated into groups to further discuss the domains, and anonymously ranked the domains individually. A summary of the results of individual prioritisation for both groups are shown in Table 9.

TABLE 9
www.frontiersin.org

Table 9. Summary of rankings of individual domains provided by participants in the consensus workshop.

After prioritisation, group discussion was undertaken to identify which outcome domains were considered core for assessment of hearing services. Discussion focussed on the importance of capturing the full breadth of outcomes experienced by patients, the importance of domains that were clearly articulable and comprehensible by patients and clinicians, and on selecting domains that would be directly modifiable by rehabilitation efforts.

Following the discussion, the group decided unanimously that four outcome domains should be recommended as part of a CODS for self-report in hearing rehabilitation in Australia. These were, in order of importance: (1) communication ability, (2) personal relationships, (3) well-being, and (4) participation restrictions. The group stressed that all seven domains presented to the group were important and should be considered for settings where the collection of additional outcomes is possible.

Following the discussion of which outcome domains should be measured following hearing rehabilitation, additional discussion was had regarding the time points at which outcomes should be measured, and which outcome measures should be used. There was no decision made regarding a positive answer to either question, with participants agreeing that these questions should be answered with reference to the research literature. Participants felt that the time of collection may differ between particular outcome measures and should therefore be determined with reference to research relevant to each particular outcome measure.

Final Recommendations

Three overarching recommendations were made as a primary output of this project. It should be noted that these recommendations were made specifically for the Australian publicly funded hearing rehabilitation context, and with the assumption that outcomes would be collected to facilitate their tracking and improvement over time.

1) Target the outcome domains “communication ability,” “personal relationships,” “well-being,” and “participation restrictions” when assessing hearing rehabilitation.

2) Measure within these domains at baseline and then following the conclusion of the rehabilitation, with a delay of at least 3 months being recommended.

3) Establish an independent body to develop a standardised outcomes instrument and mechanism for outcomes collection.

The detailed recommendations from this project are available in Supplementary Material.

Discussion

Outcome Domains

This study identified four primary outcome domains that are recommended as part of a core outcome domain set for the self-report evaluation of individual hearing rehabilitation programmes: communication ability, personal relationships, well-being, and participation restrictions.

Many currently available outcome measures used as measures of the success of hearing rehabilitation focus on improvements in communication ability, including the Glasgow Hearing Aid Benefit Profile (GHABP; Gatehouse, 1999), the Abbreviated Profile of Hearing Aid Benefit (APHAB; Cox and Alexander, 1995), and the International Outcomes Inventory for Hearing Aids (IOI-HA; Cox et al., 2000). The delineation of specific subdomains of communication ability in this study (communication with family and communication in groups) suggests that in addition to generalised measures of patients’ overall communication ability in their everyday lives, specific measures or subscales highlighting difficulties or successes in these identified areas are also required for a comprehensive assessment of communication ability. Many commonly used measures for the assessment of communication ability validated in hearing rehabilitation do not address these subdomains separably, although some measures include these aspects, such as the SOS-HEAR (Scarinci et al., 2012), the Self-Assessment of Communication (SAC; Schow and Nerbonne, 1982), and the GHABP. A notable exception is the Communication Performance subscale of the Communication Profile for the Hearing Impaired (CPHI; Demorest and Erdman, 1987), which includes several items related to communication both in group social situations and with family around the home. However, these items do not form separable subscales, making it difficult to assess communication on the subdomains in a separable way (Demorest and Erdman, 1986). The instrument also includes items that may be inappropriate for some people, such as hearing in lectures and religious services. Further work may be required to develop an instrument that can assess communication ability both across life as a while and in particularly meaningful situations for a wide variety of people.

The most specific measure of hearing-related participation restriction currently available, the Social Participation Restrictions Questionnaire (SPaRQ), focuses primarily on social participation (Heffernan et al., 2018a,b). Participation restrictions due to hearing loss manifests across a range of kinds of non-social participation, including in employment, education, domestic settings, and political life (Danermark et al., 2013; Granberg et al., 2014c), and so additional measures or expansion of the SPaRQ is likely to be necessary to enable comprehensive assessment.

There currently are no measures validated in hearing rehabilitation for the assessment of personal relationships or general well-being, although general measures of well-being such as the Warwick-Edinburg Mental Well-Being Scale (WEMWBS) are available (Tennant et al., 2007). Application of modern psychometric methods to the WEMWBS has been useful in the derivation of short-form measures with desirable measurement characteristics, suggesting that item inventories of this kind may prove useful as a starting point for the development of shorter, more specific measures of well-being benefit following hearing rehabilitation (Houghton et al., 2017). There is also ongoing work exploring the nature of hearing-specific well-being (Vercammen et al., 2020; Humes, 2021), which may result in measures of well-being that are more sensitive to hearing rehabilitation, although care should be taken when selecting these instruments to ensure that the breadth of well-being as an outcome domain—that it includes all aspects of life, not just hearing—is not lost.

Collection Considerations

With respect to the time points at which outcomes should be collected, there was disagreement between professional stakeholders in the Australian hearing industry, reflecting the lack of research into the optimal time to collect outcomes measurements. While there is good evidence that auditory ability stabilises quickly after hearing aid fitting (Dawes et al., 2014b), the period following fitting, particularly for new users, is marked by ongoing adjustment during which results obtained from outcome measures may be expected to change (Turner et al., 1996). There are also a variety of personal factors which may affect the rate at which a person adjusts to hearing aids (Dawes et al., 2014a), and it is unclear how those factors may affect different domains of adjustment. As a result, further research will be required to establish the optimal time post-fitting for any particular outcome measure to be applied.

While both Consumer and Professional groups agreed that outcomes could be appropriately collected by hearing care professionals face-to-face with a patient, the difference in preference for online delivery of questionnaires—Consumer respondents ranking it second only to face-to-face collection by a hearing care professional while Professional respondents ranking it as least preferred—suggest that a range of methods are likely to be useful in practice. The principles of experience-based co-design suggest that health services, policymakers, and consumers should be involved in the design of services and the selection of appropriate metrics for their assessment (Donetto et al., 2014), and the results of this study suggest that the methods of collection of those metrics may also be a valuable subject of co-design methods. Further work canvasing views of outcomes collection methods may also identify groups of consumers and service staff who benefit from varying methods of outcome measurement collection, requiring a multi-method approach to implementation into hearing health services.

There was substantial overlap in the identified reasons for collecting outcomes and the purposes of establishing a national database of patient outcomes. Both Consumer and Professional groups highlighted the importance of promoting patient-centred care beyond solely the provision of hearing aids, the comparison of professional practice and outcomes to national benchmarks, the value of outcomes to Government and other funders in developing policy, and potential enhancements in the development of evidence-based hearing care. This overlap suggests that to the participants in this study these two ideas—that outcome measurements should be collected and that outcome measurements should be combined and analysed across health systems—may be conceptually inseparable. Indeed, several of the reasons identified for collecting outcomes, such as the targeting by Government of auditing activities, are likely to only be possible through a centralised outcomes storage and analysis system.

Respondents were clear that centralised outcomes collection systems should be designed to maximise their benefits for various parts of the hearing health system, including hearing healthcare organisations and professionals, other healthcare providers, the public, and policymakers. When designing and implementing these systems, a broad base of stakeholders should be involved in the design and implementation of data products, ensuring their applicability across the health system. Consumers, perhaps unsurprisingly, felt that such a system could and should provide important information to consumers to support their hearing rehabilitation decision-making. The provision of information to consumers alone, however, is not sufficient to ensure that they can use that information to support healthcare decision-making; careful design of the consumer-accessible outputs of these systems will be necessary to ensure that outcomes information can be useful to consumers (Hibbard and Peters, 2003).

In addition to the benefits of aggregating outcome measurements across patients for services and systems evaluation, the results of the present study also highlight the immediate utility of outcome measurements to clinicians as a basis for decisions about the future progress of individual rehabilitation programmes. Making patient outcome measurements available to clinicians and health services may therefore provide an immediate and direct benefit to the care of the patient whose outcomes are being measured. In addition to the use of baseline measurements to support rehabilitation programme planning (such as the determination that a patient may be more likely to benefit from intensive communication training to address difficulties in particular situations, or from a referral to psychological or social support to ameliorate the effects of long-standing participation restriction), outcome measurements may identify patients whose progress has been less than might be expected, prompting additional intervention from the clinician. Providing those results that are available to hearing care professionals and organisations to the patient, both in aggregate and individualised format, may provide significant benefits to clinical practice, improve engagement and uptake by hearing care professionals and service delivery organisations, and smooth the implementation of outcome measurement within hearing health systems.

Concerns relating to the development of a national outcomes database largely related to the validity, quality, trustworthiness, and comprehensiveness of the data stored within it. Both groups expressed concerns that there could be significant potential for interested parties (particularly hearing care professionals or organisations) to modify or misrepresent data in an attempt to appear more favourably in any aggregate results. This concern has implications for the ways in which data may be collected, as methods of data collection that directly involve hearing care professionals or organisations may be viewed as more susceptible to misuse than those that bypass hearing services entirely.

Interestingly, the potential impacts on professionals and organisations of receiving poor outcomes results were not considered important by either Consumer or Professional participants. Publication of health quality data can prompt quality improvement within health systems (Fung et al., 2008), with systematic and structured outcomes an important support to quality improvement activities (Kampstra et al., 2018). Within hearing rehabilitation, improvements in service quality have been seen following the publication of outcomes data in the ongoing quality register of hearing rehabilitation clinics in Sweden (Nationellt Kvalitetsregister Hörselrehabilitering, 2019). This suggests that the ongoing collection and publication of client-centred outcomes may support a move toward improving the quality of hearing rehabilitation.

Limitations

While the results of the present study do provide important guidance for researchers, clinicians, and policymakers in the selection of outcome domains for the assessment of hearing rehabilitation, the decision in the early stages of the Delphi Review conducted with the Consumer group to restrict to a manageable number of outcome domains does mean that this list should not be considered a comprehensive description of the areas in which consumers of hearing services might experience or seek benefit from hearing rehabilitation. Indeed, several items that reached consensus in the Professional group, including reductions in social isolation and the ability to communicate in groups, were filtered out of consideration by Consumer participants at this stage. In addition, a number of items that might be considered valuable by researchers or clinicians—including feelings of empowerment, improved access to paid and volunteer work, and confidence in the effectiveness of hearing services—were also excluded from further consideration by Consumer participants. As a result, this work does not preclude the usefulness of other domains that have not been listed above, or of measures that assess other aspects of benefit. For example, where a comprehensive assessment of benefit of hearing rehabilitation is desired, the use of a general measure of improvement such as the Clinical Global Impression, which has been adapted for use in hearing rehabilitation (Öberg et al., 2009), may capture a more holistic measure of benefit, supporting the use of these more specific measures. Finally, the Delphi method used in this study involved a self-selected group of participants, and its reliance on internet-delivered text may have posed a barrier to participation for culturally and linguistically diverse Australians, those with cognitive or other difficulties, or low access to technology. Further work is required to ensure that the domains identified are indeed appropriate for the assessment of all consumers of Government-funded hearing services in Australia.

Strengths

This study includes consumers of hearing services as primary participants in the development of recommendations for the assessment of hearing rehabilitation, which has not previously been done in this field. In addition to providing a possible example to future researchers seeking to include consumers as domain experts in their research, we believe that direct involvement of consumers in research is vital to the principles of patient- and family-centred care. It is also the first in the hearing literature to make use of the Kemeny-Young method for consensus ranking, a data-driven method with useful properties including satisfaction of the Condorcet criterion (that is, it will correctly identify the choice that is preferred over every other choice by most raters should such a choice exist) and the ready availability of “off-the-shelf” algorithms for both its exact computation and heuristic approximation.

Conclusions

The recommendations from this study define a minimum patient-centred core outcome domain set that should be considered for the assessment of hearing rehabilitation in research and clinical practice. However, there is still significant research required to establish a set of outcome measures suitable for the measurement of each of these outcome domains. The selection of measurement instruments to be associated with a COS is a multi-stage process that will require considerable additional work, particularly given the identified lack of appropriate, validated measures for the identified domains (Prinsen et al., 2016). As part of this process, the measurement properties of developed or identified measures will need to be assessed. Preferably, this should be undertaken using modern psychometric methods such as those utilising Item Response Theory (IRT), which are particularly useful when assessing psychosocial, needs-based aspects of health (Tennant et al., 2004).

In general, these results have identified, through a consensus approach, a core outcome domain set that might be considered for the self-report evaluation of hearing rehabilitation and provide important background information for the design of methods to implement them across hearing healthcare systems. A broader set of self-report outcome domains that researchers and clinicians may also choose to collect in their particular context has also been identified. Furthermore, other outcome areas in addition to self-report, such as behavioural (e.g., speech perception, cognition) and physiological (e.g., electrophysiology) tests, need to be considered before there is a full COS for auditory rehabilitation. For self-report, which was the focus of this study, the range of suggested outcome domains, potential purposes for outcomes collection, and potential concerns with the establishment of centralised national outcomes collection and analysis systems strongly suggest that ongoing stakeholder engagement will be vital for the operationalisation of these results into any hearing healthcare system. In addition, significant further research is required on any selected or developed outcomes measurement instruments to determine the optimal time of outcomes collection following hearing rehabilitation.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Ethics Statement

The studies involving human participants were reviewed and approved by Hearing Australia Human Research Ethics Committee. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

MF designed the study. MF and DA designed the Delphi review with subject matter advice from LH. DA managed the data collection, analysed the data, and wrote the manuscript. All authors interpreted the data, reviewed the manuscript and provided critical revision, and approved the submitted version.

Funding

This work was funded by the Australian Government Department of Health.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

Thanks to the project team at the National Acoustic Laboratories who supported this study, including Dr. Paul Jevelle, Margot McLelland, and Jessica Cooper.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2022.787607/full#supplementary-material

References

Access Economics Pty Ltd (2015). Listen Hear! The Economic Impact and Cost of Hearing Loss in Australia. Australia: Access Economics.

Google Scholar

Akeroyd, M. A., Wright-Whyte, K., Holman, J. A., and Whitmer, W. M. (2015). A comprehensive survey of hearing questionnaires: how many are there, what do they measure, and how have they been validated? Trials 16:26.

Google Scholar

Amodio, S., D’Ambrosio, A., and Siciliano, R. (2016). Accurate algorithms for identifying the median ranking when dealing with weak and partial rankings under the Kemeny axiomatic approach. Eur. J. Oper. Res. 249, 667–676. doi: 10.1016/j.ejor.2015.08.048

CrossRef Full Text | Google Scholar

Barker, A. B., Leighton, P., and Ferguson, M. A. (2017). Coping together with hearing loss: a qualitative meta-synthesis of the psychosocial experiences of people with hearing loss and their communication partners. Int. J. Audiol. 56, 297–305. doi: 10.1080/14992027.2017.1286695

PubMed Abstract | CrossRef Full Text | Google Scholar

Barker, F., MacKenzie, E., Elliott, L., and de Lusignan, S. (2015a). Outcome Measurement in Adult Auditory Rehabilitation: a Scoping Review of Measures Used in Randomized Controlled Trials. Ear Hear. 36:7. doi: 10.1097/AUD.0000000000000167

PubMed Abstract | CrossRef Full Text | Google Scholar

Barker, F., Munro, K. J., and de Lusignan, S. (2015b). Supporting living well with hearing loss: a Delphi review of self-management support. Int. J. Audiol. 54, 691–699. doi: 10.3109/14992027.2015.1037019

PubMed Abstract | CrossRef Full Text | Google Scholar

Barker, F., Mackenzie, E., Elliott, L., Jones, S., and de Lusignan, S. (2016). Interventions to improve hearing aid use in adult auditory rehabilitation. Cochrane Database Syst. Rev. 8:CD010342.

Google Scholar

Bartholdi, J. J., Tovey, C. A., and Trick, M. A. (1989). The computational difficulty of manipulating an election. Soc. Choice Welfare 6, 227–241. doi: 10.1007/BF00295861

CrossRef Full Text | Google Scholar

Bennett, R. J., Barr, C., Montano, J., Eikelboom, R. H., Saunders, G. H., Pronk, M., et al. (2020). Identifying the approaches used by audiologists to address the psychosocial needs of their adult clients. Int. J. Audiol. 60, 104–114. doi: 10.1080/14992027.2020.1817995

PubMed Abstract | CrossRef Full Text | Google Scholar

Bennett, R. J., Meyer, C. J., Eikelboom, R. H., and Atlas, M. D. (2018). Evaluating Hearing Aid Management: development of the Hearing Aid Skills and Knowledge Inventory (HASKI). Am. J. Audiol. 27, 333–348. doi: 10.1044/2018_AJA-18-0050

CrossRef Full Text | Google Scholar

Biedl, T., Brandenburg, F. J., and Deng, X. (2006). “Crossings and Permutations,” in Graph Drawing, eds P. Healy and N. S. Nikolov (Heidelberg: Springer), 1–12.

Google Scholar

British Society of Audiology (2016). Practice Guidance - Common Principles of Rehabilitation for Adults in Audiology Services. United Kingdom: British Society of Audiology.

Google Scholar

Chia, E. M., Wang, J. J., Rochtchina, E., Cumming, R. R., Newall, P., and Mitchell, P. (2007). Hearing impairment and health-related quality of life: the Blue Mountains Hearing Study. Ear Hear. 28, 187–195. doi: 10.1097/AUD.0b013e31803126b6

PubMed Abstract | CrossRef Full Text | Google Scholar

Commonwealth of Australia (2019). Portfolio Budget Statements 2019-20, Budget Related Paper No. 1.9, Health Portfolio”, Department of Health. Sydney, Australia: Commonwealth of Australia.

Google Scholar

Cox, R. M., and Alexander, G. C. (1995). The Abbreviated Profile of Hearing Aid Benefit. Ear Hear. 16, 176–186. doi: 10.1097/00003446-199504000-00005

PubMed Abstract | CrossRef Full Text | Google Scholar

Cox, R. M., Hyde, M., Gatehouse, S., Noble, W., Dillon, H., Bentler, R. A., et al. (2000). Optimal Outcome Measures, Research Priorities, and International Cooperation. Ear Hear. 21, 106S–115S. doi: 10.1097/00003446-200008001-00014

PubMed Abstract | CrossRef Full Text | Google Scholar

Danermark, B., Granberg, S., Kramer, S. E., Selb, M., and Möller, C. (2013). The Creation of a Comprehensive and a Brief Core Set for Hearing Loss Using the International Classification of Functioning, Disability and Health. Am. J. Audiol. 22:323. doi: 10.1044/1059-0889(2013/12-0052)

CrossRef Full Text | Google Scholar

Dawes, P., Emsley, R., Cruickshanks, K. J., Moore, D. R., Fortnum, H., Edmondson-Jones, M., et al. (2015). Hearing Loss and Cognition: the Role of Hearing Aids, Social Isolation and Depression. PLoS One 10:e0119616. doi: 10.1371/journal.pone.0119616

PubMed Abstract | CrossRef Full Text | Google Scholar

Dawes, P., Maslin, M., and Munro, K. J. (2014a). ‘Getting used to’ hearing aids from the perspective of adult hearing-aid users. Int. J. Audiol. 53, 861–870. doi: 10.3109/14992027.2014.938782

PubMed Abstract | CrossRef Full Text | Google Scholar

Dawes, P., Munro, K. J., Kalluri, S., and Edwards, B. (2014b). Acclimatization to Hearing Aids. Ear Hear. 35, 203–212. doi: 10.1097/aud.0b013e3182a8eda4

PubMed Abstract | CrossRef Full Text | Google Scholar

Demorest, M. E., and Erdman, S. A. (1986). Scale composition and item analysis of the Communication Profile for the Hearing Impaired. J. Speech Lang. Hear. Res. 29, 515–535. doi: 10.1044/jshr.2904.535

PubMed Abstract | CrossRef Full Text | Google Scholar

Demorest, M. E., and Erdman, S. A. (1987). Development of the communication profile for the hearing impaired. J. Speech Hear. Disord. 52, 129–143. doi: 10.1044/jshd.5202.129

PubMed Abstract | CrossRef Full Text | Google Scholar

Department of Health (n.d.). Am I eligible to receive fully subsidised hearing services?. Available Online at: https://www.hearingservices.gov.au/wps/portal/hso/site/eligibility/programhelp/eligibility/ (accessed on 24 Jan, 2022).

Google Scholar

Dillon, H., Birtles, G., and Lovegrove, R. (1999). Measuring the outcomes of a national rehabilitation program: normative data for the Client Oriented Scale of Improvement (COSI) and the Hearing Aid Users Questionnaire (HAUQ). J. Am. Acad. Audiol. 10, 67–79.

Google Scholar

Dillon, H., James, A., and Ginis, J. (1997). Client Oriented Scale of Improvement (COSI) and its relationship to several other measures of benefit and satisfaction provided by hearing aids. J. Am. Acad. Audiol. 8, 27–43.

PubMed Abstract | Google Scholar

Donetto, S., Pierri, P., Tsianakas, V., and Robert, G. (2014). “Experience-based Co-design and healthcare improvement: realising participatory design in the public sector,” in Fourth Service Design and Innovation conference, (Lancaster, UK: ServDes), 10.

Google Scholar

Emond, E. J., and Mason, D. W. (2002). A new rank correlation coefficient with application to the consensus ranking problem. J. Multi-Criteria Decis. Anal. 11, 17–28. doi: 10.1002/mcda.313

CrossRef Full Text | Google Scholar

Fackrell, K., Smith, H., Colley, V., Thacker, B., Horobin, A., Haider, H. F., et al. (2017). Core Outcome Domains for early phase clinical trials of sound-, psychology-, and pharmacology-based interventions to manage chronic subjective tinnitus in adults: the COMIT’ID study protocol for using a Delphi process and face-to-face meetings to establish consensus. Trials 18:388. doi: 10.1186/s13063-017-2123-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Ferguson, M. A., Brandreth, M., Brassington, W., Leighton, P., and Wharrad, H. (2016). A Randomized Controlled Trial to Evaluate the Benefits of a Multimedia Educational Program for First-Time Hearing Aid Users. Ear Hear. 37, 123–136. doi: 10.1097/AUD.0000000000000237

PubMed Abstract | CrossRef Full Text | Google Scholar

Ferguson, M. A., and Henshaw, H. (2015). How Does Auditory Training Work? Joined-Up Thinking and Listening. Semin. Hear. 36, 237–249. doi: 10.1055/s-0035-1568985

PubMed Abstract | CrossRef Full Text | Google Scholar

Ferguson, M. A., Kitterick, P. T., Chong, L. Y., Edmondson-Jones, M., Barker, F., and Hoare, D. J. (2017). Hearing aids for mild to moderate hearing loss in adults. Cochrane Database Syst. Rev. 9:CD012023.

Google Scholar

Ferguson, M. A., Leighton, P., Brandreth, M., and Wharrad, H. (2018). Development of a multimedia educational programme for first-time hearing aid users: a participatory design. Int. J. Audiol. 57, 600–609. doi: 10.1080/14992027.2018.1457803

PubMed Abstract | CrossRef Full Text | Google Scholar

Ferguson, M. A., Maidment, D., Henshaw, H., and Heffernan, E. (2019). Evidence-Based Interventions for Adult Aural Rehabilitation: that Was Then, This Is Now. Semin. Hear. 40, 068–084. doi: 10.1055/s-0038-1676784

PubMed Abstract | CrossRef Full Text | Google Scholar

Fung, C. H., Lim, Y.-W., Mattke, S., Damberg, C., and Shekelle, P. G. (2008). Systematic review: the evidence that publishing patient care performance data improves quality of care. Ann. Intern. Med. 148, 111–123. doi: 10.7326/0003-4819-148-2-200801150-00006

PubMed Abstract | CrossRef Full Text | Google Scholar

Gatehouse, S. (1999). Glasgow Hearing Aid Benefit Profile: derivation and validation of a client-centered outcome measure for hearing aid services. J. Am. Acad. Audiol. 10, 80–103.

Google Scholar

Granberg, S., Dahlström, J., Möller, C., Kähäri, K., and Danermark, B. (2014a). The ICF Core Sets for hearing loss – researcher perspective. Part I: systematic review of outcome measures identified in audiological research. Int. J. Audiol. 53, 65–76. doi: 10.3109/14992027.2013.851799

PubMed Abstract | CrossRef Full Text | Google Scholar

Granberg, S., Möller, K., Skagerstrand, Å, Möller, C., and Danermark, B. (2014b). The ICF Core Sets for hearing loss: researcher perspective, Part II: linking outcome measures to the International Classification of Functioning, Disability and Health (ICF). Int. J. Audiol. 53, 77–87. doi: 10.3109/14992027.2013.858279

PubMed Abstract | CrossRef Full Text | Google Scholar

Granberg, S., Pronk, M., Swanepoel, D. W., Kramer, S. E., Hagsten, H., Hjaldahl, J., et al. (2014c). The ICF core sets for hearing loss project: functioning and disability from the patient perspective. Int. J. Audiol. 53, 777–786. doi: 10.3109/14992027.2014.938370

PubMed Abstract | CrossRef Full Text | Google Scholar

Hall, D. A., Haider, H., Kikidis, D., Mielczarek, M., Mazurek, B., Szczepek, A. J., et al. (2015). Toward a Global Consensus on Outcome Measures for Clinical Trials in Tinnitus: report From the First International Meeting of the COMiT Initiative, November 14, 2014, Amsterdam. Netherlands. Trends Hear. 19:233121651558027. doi: 10.1177/2331216515580272

PubMed Abstract | CrossRef Full Text | Google Scholar

Hall, D. A., Kitterick, P., Heffernan, E., Fackrell, K., Lucas, L., and Ferguson, M. A. (2019). How Do We Know That Our Patients Have Benefitted From Our ENT/Audiological Interventions? Presented at the Annual Meeting of ADANO 2016 in Berlin. Otol. Neurotol. 40, e474–e481. doi: 10.1097/MAO.0000000000001937

PubMed Abstract | CrossRef Full Text | Google Scholar

Hall, D. A., Smith, H., Hibbert, A., Colley, V., Haider, H. F., Horobin, A., et al. (2018). The COMiT’ID Study: Developing Core Outcome Domains Sets for Clinical Trials of Sound-, Psychology-, and Pharmacology-Based Interventions for Chronic Subjective Tinnitus in Adults. Trends Hear. 22:2331216518814384. doi: 10.1177/2331216518814384

PubMed Abstract | CrossRef Full Text | Google Scholar

Heffernan, E., Coulson, N. S., and Ferguson, M. A. (2018a). Development of the Social Participation Restrictions Questionnaire (SPaRQ) through consultation with adults with hearing loss, researchers, and clinicians: a content evaluation study. Int. J. Audiol. 57, 791–799. doi: 10.1080/14992027.2018.1483585

PubMed Abstract | CrossRef Full Text | Google Scholar

Heffernan, E., Maidment, D. W., Barry, J. G., and Ferguson, M. A. (2018b). Refinement and Validation of the Social Participation Restrictions Questionnaire: an Application of Rasch Analysis and Traditional Psychometric Analysis Techniques. Ear Hear. 40:12. doi: 10.1097/AUD.0000000000000618

PubMed Abstract | CrossRef Full Text | Google Scholar

Heffernan, E., Coulson, N. S., Henshaw, H., Barry, J. G., and Ferguson, M. A. (2016). Understanding the psychosocial experiences of adults with mild-moderate hearing loss: an application of Leventhal’s self-regulatory model. Int. J. Audiol. 55, S3–S12. doi: 10.3109/14992027.2015.1117663

PubMed Abstract | CrossRef Full Text | Google Scholar

Heffernan, E., Habib, A., and Ferguson, M. A. (2019). Evaluation of the psychometric properties of the social isolation measure (SIM) in adults with hearing loss. Int. J. Audiol. 58, 45–52. doi: 10.1080/14992027.2018.1533257

PubMed Abstract | CrossRef Full Text | Google Scholar

Helmer, O. (1967). Analysis of the Future: The Delphi Method”. California: RAND Corporation.

Google Scholar

Henshaw, H., and Ferguson, M. A. (2013). Efficacy of Individual Computer-Based Auditory Training for People with Hearing Loss: a Systematic Review of the Evidence. PLoS One 8:e62836. doi: 10.1371/journal.pone.0062836

PubMed Abstract | CrossRef Full Text | Google Scholar

Hibbard, J. H., and Peters, E. (2003). Supporting Informed Consumer Health Care Decisions: data Presentation Approaches that Facilitate the Use of Information in Choice. Annu. Rev. Public Health 24, 413–433. doi: 10.1146/annurev.publhealth.24.100901.141005

PubMed Abstract | CrossRef Full Text | Google Scholar

Houghton, S., Wood, L., Marais, I., Rosenberg, M., Ferguson, R., and Pettigrew, S. (2017). Positive Mental Well-Being: a Validation of a Rasch-Derived Version of the Warwick-Edinburgh Mental Well-Being Scale. Assessment 24, 371–386. doi: 10.1177/1073191115609995

PubMed Abstract | CrossRef Full Text | Google Scholar

Hsu, C.-C., and Sandford, B. A. (2007). The Delphi Technique: making Sense Of Consensus. Pract. Assess. Res. Eval. 12:8.

Google Scholar

Humes, L. E. (2021). Development of the SWB-HL: a Scale of the Subjective Well-Being of Older Adults With Hearing Loss. Front. Psychol. 12:640165. doi: 10.3389/fpsyg.2021.640165

PubMed Abstract | CrossRef Full Text | Google Scholar

Kampstra, N. A., Zipfel, N., van der Nat, P. B., Westert, G. P., van der Wees, P. J., and Groenewoud, A. S. (2018). Health outcomes measurement and organizational readiness support quality improvement: a systematic review. BMC Health Serv. Res. 18:1005. doi: 10.1186/s12913-018-3828-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Kemeny, J. G. (1959). Mathematics without Numbers. Daedalus 88, 577–591.

Google Scholar

Kirkham, J. J., Davis, K., Altman, D. G., Blazeby, J. M., Clarke, M., Tunis, S., et al. (2017). Core Outcome Set-STAndards for Development: the COS-STAD recommendations. PLoS Med. 14:e1002447. doi: 10.1371/journal.pmed.1002447

PubMed Abstract | CrossRef Full Text | Google Scholar

Lawrence, B. J., Jayakody, D. M. P., Henshaw, H., Ferguson, M. A., Eikelboom, R. H., Loftus, A. M., et al. (2018). Auditory and Cognitive Training for Cognition in Adults With Hearing Loss: a Systematic Review and Meta-Analysis. Trends Hear. 22:233121651879209. doi: 10.1177/2331216518792096

PubMed Abstract | CrossRef Full Text | Google Scholar

Lin, F. R. (2011). Hearing Loss and Cognition Among Older Adults in the United States. J. Gerontol. A Biol. Sci. Med. Sci. 66, 1131–1136. doi: 10.1093/gerona/glr115

PubMed Abstract | CrossRef Full Text | Google Scholar

Maidment, D. W., Barker, A. B., Xia, J., and Ferguson, M. A. (2016). Effectiveness of alternative listening devices to conventional hearing aids for adults with hearing loss: a systematic review protocol. BMJ Open 6:e011683. doi: 10.1136/bmjopen-2016-011683

PubMed Abstract | CrossRef Full Text | Google Scholar

Nationellt Kvalitetsregister Hörselrehabilitering (2019). Årsrapport 2019. Göteborg: Nationellt kvalitetsregister för öron-, näs- och halssjukvård.

Google Scholar

Öberg, M., Wanstrom, G., Hjertman, H., Lunner, T., and Andersson, G. (2009). Development and initial validation of the ‘Clinical Global Impression’ to measure outcomes for audiological rehabilitation. Disabil. Rehabil. 31, 1409–1417. doi: 10.1080/09638280802621408

PubMed Abstract | CrossRef Full Text | Google Scholar

OMET Initiative (n.d.). COMET Initiative. Available online at: https://www.comet-initiative.org/ (accessed April 19, 2022).

Google Scholar

Perez, E., and Edmonds, B. A. (2012). A Systematic Review of Studies Measuring and Reporting Hearing Aid Usage in Older Adults since 1999: a Descriptive Summary of Measurement Tools. PLoS One 7:e31831. doi: 10.1371/journal.pone.0031831

PubMed Abstract | CrossRef Full Text | Google Scholar

PricewaterhouseCoopers Australia (2017). Review of services and technology supply in the Hearing Services Program: Final Report. Barangaroo: PricewaterhouseCoopers Australia.

Google Scholar

Prinsen, C. A. C., Vohra, S., Rose, M. R., Boers, M., Tugwell, P., Clarke, M., et al. (2016). How to select outcome measurement instruments for outcomes included in a “Core Outcome Set” – a practical guideline. Trials 17:449. doi: 10.1186/s13063-016-1555-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Routledge, D. (2017). Human Services Outcomes Framework Guide”, Finance, Services & Innovation. Australia: NSW Government.

Google Scholar

Scarinci, N., Worrall, L., and Hickson, L. (2012). Factors associated with third-party disability in spouses of older people with hearing impairment. Ear Hear. 33, 698–708. doi: 10.1097/AUD.0b013e31825aab39

PubMed Abstract | CrossRef Full Text | Google Scholar

Schow, R. L., and Nerbonne, M. A. (1982). Communication Screening Profile: use with Elderly Clients. Ear Hear. 3, 135–147. doi: 10.1097/00003446-198205000-00007

PubMed Abstract | CrossRef Full Text | Google Scholar

Sereda, M., Hoare, D. J., Nicholson, R., Smith, S., and Hall, D. A. (2015). Consensus on Hearing Aid Candidature and Fitting for Mild Hearing Loss, With and Without Tinnitus: delphi Review. Ear Hear. 36:13. doi: 10.1097/AUD.0000000000000140

PubMed Abstract | CrossRef Full Text | Google Scholar

Tennant, A., McKenna, S. P., and Hagell, P. (2004). Application of Rasch analysis in the development and application of quality of life instruments. Value Health 7, S22–S26. doi: 10.1111/j.1524-4733.2004.7s106.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Tennant, R., Hiller, L., Fishwick, R., Platt, S., Joseph, S., Weich, S., et al. (2007). The Warwick-Edinburgh Mental Well-being Scale (WEMWBS): development and UK validation. Health Qual. Life Outcomes 5:63. doi: 10.1186/1477-7525-5-63

PubMed Abstract | CrossRef Full Text | Google Scholar

Turner, C. W., Humes, L. E., Bentler, R. A., and Cox, R. M. (1996). A review of past research on changes in hearing aid benefit over time. Ear Hear. 17, 14S–25S. doi: 10.1097/00003446-199617031-00003

PubMed Abstract | CrossRef Full Text | Google Scholar

Vercammen, C., Ferguson, M., Kramer, S. E., Meis, M., Singh, G., Timmer, B., et al. (2020). Well-Hearing Is Well-Being: a Phonak position statement about why hearing healthcare is vital for healthy living. Hear. Rev. 27, 18–22.

Google Scholar

Wallhagen, M. I. (2010). The Stigma of Hearing Loss. Gerontologist 50, 66–75. doi: 10.1093/geront/gnp107

PubMed Abstract | CrossRef Full Text | Google Scholar

Williamson, P., and Clarke, M. (2012). The COMET (Core Outcome Measures in Effectiveness Trials) Initiative: its Role in Improving Cochrane Reviews. Cochrane Database Syst. Rev. 5:ED000041. doi: 10.1002/14651858.ED000041

PubMed Abstract | CrossRef Full Text | Google Scholar

Wong, L. L. N., and Hickson, L. (2012). Evidence-based practice in audiology: evaluating interventions for children and adults with hearing impairment. San Diego, CA: Plural Publishing.

Google Scholar

Keywords: outcome assessment (health care), correction of hearing impairment, hearing loss, audiology, patient reported outcome measures (PROMs)

Citation: Allen D, Hickson L and Ferguson M (2022) Defining a Patient-Centred Core Outcome Domain Set for the Assessment of Hearing Rehabilitation With Clients and Professionals. Front. Neurosci. 16:787607. doi: 10.3389/fnins.2022.787607

Received: 01 October 2021; Accepted: 06 April 2022;
Published: 03 May 2022.

Edited by:

Jennifer Lentz, Indiana University Bloomington, United States

Reviewed by:

Michael Akeroyd, University of Nottingham, United Kingdom
Amy Piper, Indiana University, United States

Copyright © 2022 Allen, Hickson and Ferguson. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: David Allen, david.allen@mq.edu.au

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.