Next Article in Journal
Multicenter DSC–MRI-Based Radiomics Predict IDH Mutation in Gliomas
Next Article in Special Issue
Additional Value of Patient-Reported Symptom Monitoring in Cancer Care: A Systematic Review of the Literature
Previous Article in Journal
Refining Prognosis in Chemoembolization for Hepatocellular Carcinoma: Immunonutrition and Liver Function
Previous Article in Special Issue
Stakeholder Perceptions of Key Aspects of High-Quality Cancer Care to Assess with Patient Reported Outcome Measures: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Different Approaches for Case-Mix Adjustment of Patient-Reported Outcomes to Compare Healthcare Providers—Methodological Results of a Systematic Review

1
German Cancer Society, Kuno-Fischer-Str. 8, 14507 Berlin, Germany
2
Institute of Medical Sociology, Health Services Research, and Rehabilitation Science, Faculty of Medicine and University Hospital Cologne, Faculty of Human Sciences, University of Cologne, 50933 Cologne, Germany
*
Author to whom correspondence should be addressed.
Cancers 2021, 13(16), 3964; https://doi.org/10.3390/cancers13163964
Submission received: 1 July 2021 / Revised: 28 July 2021 / Accepted: 2 August 2021 / Published: 5 August 2021

Abstract

:

Simple Summary

Patient-reported outcomes need to be reported with case-mix adjustment in order to allow fair comparison between healthcare providers. This systematic review identified different approaches to case-mix adjustment, with wide variation between the various approaches.

Abstract

Patient-reported outcomes (PROs) are increasingly being used to compare the quality of outcomes between different healthcare providers (medical practices, hospitals, rehabilitation facilities). However, such comparisons can only be fair if differences in the case-mix between different types of provider are taken into account. This can be achieved with adequate statistical case-mix adjustment (CMA). To date, there is a lack of overview studies on current CMA methods for PROs. The aim of this study was to investigate which approaches are currently used to report and examine PROs for case-mix-adjusted comparison between providers. A systematic MEDLINE literature search was conducted (February 2021). The results were examined by two reviewers. Articles were included if they compared (a) different healthcare providers using (b) case-mix-adjusted (c) patient-reported outcomes (all AND conditions). From 640 hits obtained, 11 articles were included in the analysis. A wide variety of patient characteristics were used as adjustors, and baseline PRO scores and basic sociodemographic and clinical information were included in all models. Overall, the adjustment models used vary considerably. This evaluation is an initial attempt to systematically investigate different CMA approaches for PROs. As a standardized approach has not yet been established, we suggest creating a consensus-based methodological guideline for case-mix adjustment of PROs.

1. Introduction

There is a growing body of evidence that patient-reported outcomes (PROs) are effective in improving cancer patients’ health when used for patient monitoring or as a tool during patient consultations for better evaluation of patients’ symptoms or functional impairment [1,2,3,4]. Following landmark studies by Basch et al. [5,6] and Denis et al. [7], many approaches for incorporating (electronic) PROs into routine clinical cancer care have recently been developed—although some approaches also date back 15 years or more [8,9]. Due to cost and organizational barriers [10], only a small proportion of patients are currently benefiting from these approaches, but PROs are on the wish list of many service providers and healthcare systems and can be expected to find their way into more and more national cancer plans in the near future [11].
In addition to these approaches based on individual patients, PROs can also be used for another purpose in routine clinical care—quality assurance—in the sense of quality improvement initiatives. For this purpose, aggregated data at the level of healthcare providers are compared with those of other providers, resulting in rankings or benchmark reports. PROs are then regarded in a similar manner as other quality metrics: if suboptimal outcomes are observed at the provider level, the providers are typically asked to use these comparisons as part of their quality management in order to improve care. This function has been prominently advocated by, among others, the International Consortium on Health Outcomes Management (ICHOM), for whom PROs are a central part of the standard datasets used to allow comparisons. Building on the ICHOM standard dataset for localized prostate cancer [12], for example, the multinational TrueNorth Global Registry has been set up [13], with over 200 participating sites in 15 countries and with national substudies, e.g., [14]. It has been argued in the past that these two functions (with approaches at the individual patient level and at the provider level) can work hand in hand, but examples of this are as yet scarce [15].
When PROs are used as quality metrics to compare healthcare providers, special attention needs to be given to the different case-mixes present in order to allow fair reporting [16]. If PROs were to be reported without adjustment, services with less severely affected patients or otherwise favorable patient characteristics, such as fewer comorbidities or earlier disease stages, might appear to be better than they actually are, and vice versa. This applies to quality metrics in general and may lead to inappropriate incentives such as refusal to treat specific patients—especially if the providers are ranked and the results are publicly reported, with the providers’ names not being concealed. The practice of refusing patients in order to avoid costs or unfavorable outcomes in competitive healthcare markets is often referred to as “cream skimming” and is highly applicable to PRO reporting. According to Iezzoni’s framework of case-mix adjustment (CMA), a healthcare outcome can be described as a function of “intrinsic patient-related risk factors, treatment effectiveness, quality of care” and “random chance” [16]. Thus, when PROs are used to compare providers, effects summarized as “intrinsic patient-related risk factors” must not distort the comparison results. CMA (which is sometimes more generally referred to as “risk adjustment”) aims to account for such differences in providers’ case-mixes using a statistical procedure. The goal of CMA is to allow fair and reliable comparison of providers relative to the given outcome(s). Its aim is to allow “like-to-like comparisons,” as described by Iezzoni in her introductory chapter describing the purpose of CMA [16]. Ultimately, CMA is essential for the credibility and usefulness of provider comparisons.
Different approaches to CMA have been used in the past for quality metrics other than those based on PROs, and differences in approaches may lead to substantial differences in the reported outcomes and in the ranking of providers and services. There have been few publications on how to adjust the case-mix for PROs, and there is a lack of systematic evaluation of published and proposed CMA approaches. The aim of the present systematic review was to identify methods used for CMA in provider comparisons using PROs, with a particular focus on the adjustors included in the CMA models and the statistical approaches used. Recommendations on the choice of future CMA methodology—as suggestions for a possible CMA guideline—are also derived from this.

2. Methods

2.1. Studies and Methodological Decisions of Interest

This systematic review adapted the participants, interventions, comparisons, outcomes, study designs (PICOS) template, with omission of the intervention (I) and comparator (C) categories, which were not applicable here. Any patients treated by various healthcare providers were therefore examined as participants (P), PROs as outcomes (O), and observational studies and secondary analyses of interventional studies were included as study designs (S).
Iezzoni [16] proposed four main questions that can be used for guidance in finding the “right” CMA approach:
  • Risk of what outcome?
  • Over what time frame?
  • For what population?
  • For what purpose?
These questions (referred to as “Iezzoni’s four”) are therefore used in this review to evaluate the different CMA approaches together with the adjustors used, the criteria with which the adjustors were chosen, the statistical model used, and the way in which the results were reported.

2.2. Eligibility Criteria and Information Sources

This review includes observational studies or secondary analyses of interventional studies published between January 1992 and January 2021 written in English, French, or German, in which PROs are used to compare the quality of care between recognizable healthcare providers and in which any kind of case-mix adjustment of PROs is used. PROs were considered to be “instruments to elicit information from patients on self-reported health status [and] outcomes” in accordance with the definition provided by the Organization for Economic Cooperation and Development (OECD) [17]. Studies were included if they used at least one adjustor that was not the PRO baseline score. Exclusion criteria were: studies that did not use (case-mix-adjusted) PROs for provider comparison, or had no provider comparison or no CMA or did not explain their CMA; those that used patient-reported experience measures (including satisfaction surveys) as opposed to the above definition of PROs (patient-reported experience measurements measure—in contrast to PROs—“how patients experience health care and refer to practical aspects of care, such as accessibility, care coordination and provider-patient communication” [17] and are as such not healthcare outcomes). The electronic database MEDLINE was used as the information source. Citations and references in the studies included were also screened for eligibility. The review was registered at PROSPERO (CRD42020180420).

2.3. Search and Study Selection

The search string was developed iteratively, with test searches being conducted in the MEDLINE database by N.T.S. and C.K. The study group developed the final search strategy with support from an experienced systematic reviewer, combining independent and medical subject headings (MeSH) terms in accordance with the PRESS guideline statement [18]. The final search string contained variations of the terms: “humans” (P), “patient-reported outcomes” or “patient-reported outcome measures” (O), and “risk adjustment” or “case-mix adjustment.” The full search string can be found in Appendix A.
C.K. and N.T.S. first independently screened all titles and abstracts. N.T.S. searched for available full texts, which were then independently reviewed by C.K. and N.T.S. relative to the inclusion criteria. Any divergent results were discussed by C.K. and N.T.S. to obtain a consensus.

2.4. Data Collection Process, Data Items, and Risk of Bias in Individual Studies

C.K. and N.T.S. independently used a standardized data extraction form that was developed deductively using Iezzoni’s earlier research on risk adjustment [16]. Data items included in the extraction form were: type of provider, Iezzoni’s four, information on adjustors (which ones, and how they were chosen), CMA methodology, and reporting of case-mix-adjusted PROs. In addition, quality assessment was carried out using the Critical Appraisal Skills Programme (CASP) checklist for cohort studies [19].

2.5. Summary Measures and Synthesis of Results

Data synthesis was carried out by qualitatively summarizing the extracted information on statistical methods used for CMA, adjustors, and reporting of provider comparisons. Since there was wide heterogeneity among the CMA approaches and PROs applied, it was not possible to pool the findings quantitatively as a meta-analysis. However, where possible, similar case-mix adjustment approaches are presented in bundled form. No additional subgroup analysis was carried out in any of the studies.

2.6. Risk of Bias across Studies

As the purpose of this methodological review was to provide an overview of the statistical methods used for adjustment, and the studies included were conducted in different populations, over different time periods, and with different PROs, it was not possible to carry out a meta-analysis. Evaluation of the risk of bias among the studies included was therefore not applicable.

3. Results

3.1. Study Selection and Characteristics

A total of 640 studies were identified in MEDLINE using the search string. After screening of the titles/abstracts and full text, nine studies were included for the qualitative synthesis (Figure 1: PRISMA flowchart). Two studies were added after the references from these nine studies had been screened. Thus, a total of eleven studies were included in this analysis. An overview on study characteristics can be found in Table 1.

3.2. Quality of Included Studies (CASP Evaluation)

All studies had a clear focus, precise and mostly applicable results, and fitted with other evidence. Follow-up information was partly insufficient or could not be examined. Details are provided in the CASP results Table A1 in Appendix B.

3.3. Provider Settings

Five inpatient settings [20,21,22,23,24], five outpatient settings [25,26,27,28,29], one mental health services setting [26], and three rehabilitation settings [20,28,29] were identified. Ten studies included patients with orthopedic conditions [20,21,22,23,24,25,27,28,29,30]. Three of the studies included were set in National Health Service-funded hospitals in the United Kingdom [21,23,30]. Three studies were set in hospitals in the United States, taking part in the Focus on Therapeutic Outcomes (FOTO) initiative [25,28,29].

3.4. “Iezzoni’s Four”

This synthesis follows Iezzoni’s approach on analyzing risk adjustment. Iezzoni’s four main questions are therefore covered in the following sections [16]. Detailed information is provided in Table 2.

3.4.1. Risk of What Outcome?

The studies included a total of 17 different PRO instruments: seven generic PRO questionnaires (PCS [31], PF-10 [32], SF-12 [31], ADL [33], Lehman Quality of Life Interviews [34], EQ-5D [35], including one rehabilitation-specific one: IRES [36]) and ten condition-specific instruments (VAS-BP, VAS-LP [37], OHS, OKS [38], MDQ [39], NDI [40], ODI [41], MacNew [42], SMFA [43], and LCAT [44]).

3.4.2. For What Population?

All of the studies included explored PROs for an adult population. Ten studies were conducted in orthopedic settings [20,21,22,23,24,25,27,28,29,30], and one investigated patients with psychiatric conditions [26]. None of the studies that were identified investigated PROs in populations of cancer patients.

3.4.3. Over What Time Period?

Most of the studies examined the development of PRO scores from baseline (e.g., at admission, before the start of therapy) and a predefined or treatment success-defined post-therapeutic time point. In some studies, the time period of PRO data collection was not reported or specified [24,27,28]—for example, if the post-therapeutic end point was measured at the end of an intervention, of which the duration could vary, e.g., [28].

3.4.4. For What Purpose?

The main purpose of the studies included was to obtain inter-provider comparisons to measure performance, in some cases resulting in a (clinical) ranking [25] or in the identification of special expertise [28].

3.5. Selection of Adjustors

There are several approaches for choosing adjustors for CMA. The studies included only partly reported on the selection process. Two studies stated that they chose the adjustors that were used on the basis of statistical parameters: Resnik and Hart used bivariate analysis to identify suitable adjustors [28], and Deutscher et al. employed a back-step regression model [25]. Farin et al. and Khor et al. based their choice of adjustors on earlier findings by their research groups [20,22]. Table 2 provides further information.

3.6. Patient Characteristics Included in the CMA Models

All of the studies included sociodemographic characteristics (such as sex, age) of their patients and baseline PRO scores as adjustors in the CMA models used. All of the studies chose comorbidities and/or a measure of the patient’s general physical status (i.e., specific comorbidities, number of comorbidities, Charleston Comorbidity Index, American Society of Anesthesiologists score, etc.) as adjustors. Most of the studies that were included featured disease-specific information or information about treatments previously received by the patients.

3.7. Provider Characteristics

Lutz et al. included provider characteristics (the American state in which the study was located) as an adjustor and Sivaganesan et al. included information about surgeons (years in practice, fellowship training) [24,27].

3.8. Statistical Approaches

The statistical approaches used differed in relation to the intended presentation of the adjusted results. Some studies calculate scores that keep the scale of the initial PRO score, while others keep the scale but present “performance indicators”—typically the difference between the observed score and the score expected on the basis of the patient case-mix. These approaches are referred to in this article as “unit/scale-preserving approaches.” Other studies compute ratios of the observed scores divided by the expected scores, or only calculate expected scores. These approaches thus do not present results that are consistent with the units of the original PRO score and are referred to in this article as “unit/scale-aberrant approaches.”
A common differentiation made in the statistical models used for CMA is that between direct and indirect standardization [45]. Briefly, direct approaches use a matched subsample from each provider to compare the outcomes for those patients with those in a standardized population, whereas indirect approaches use the whole sample from the providers and adjust the outcomes using a modeling procedure, typically regression analyses. All of the studies included in this review used indirect standardization methods. However, they differ in the way in which the adjusted scores are calculated and presented. These differences are therefore examined in greater depth in the following paragraphs.
Since both the topic itself and differences in CMA methods are quite complex, the different statistical approaches are illustrated by using an example based on the Prostate Cancer Outcomes (PCO) Study data [14] (described in italics):
To compare outcomes of prostate cancer centers one year after treatment—e.g., surgery—certified prostate cancer centers take part in a survey collecting information on EPIC-26 scores. The EPIC-26 is a prostate cancer-specific PRO questionnaire resulting in five scores, among them an incontinence score—with 0 being the lowest and 100 the highest possible score [46]. For our example, we would like to compare Center A, Center B, and Center C using the following metrics:
Unadjusted—i.e., observed—mean incontinence scores (1 year after surgery): 26 (Center A), 37 (Center B), 31 (Center C), 30 (all patients in all centers)
Expected mean incontinence scores based on linear regression models (using the following adjustors: baseline EPIC-26 scores, age, tumor stage, etc.): 20 (Center A), 33 (Center B), 40 (Center C).

3.8.1. Calculation of Transformed Adjusted Scores (Unit/Scale-Preserving Approaches)

Two studies used the observed and previously calculated expected values to transform them into adjusted scores that can be interpreted using the initial PRO score’s unit [23,28]: Nuttal et al. firstly calculated the ratio o b s e r v e d e x p e c t e d for every provider and multiplied those ratios with the overall population’s observed average [23]:
a d j u s t e d   s c o r e = o b s e r v e d e x p e c t e d   *   o v e r a l l   o b s e r v e d   a v e r a g e .
In our example, the adjusted scores were:
Center A: 26 20   *   30     39
Center B: 37 33   *   30     33
Center C: 30 40   *   30     26
These scores are all within the original EPIC-26 limits (0–100) and are interpreted identically. The adjusted scores provide information about what the incontinence score of an “average” prostate cancer patient would be in the respective center.
Resnik and Hart calculated the differences between observed and expected values [28]. The resulting adjusted (or “performance”) scores were not within the limits of the original PRO scale (e.g., for a positive PRO scale, the differences might become negative), but they are interpretable using the original units:
a d j u s t e d   s c o r e = e x p e c t e d o b s e r v e d
Deutscher et al. and Farin et al. used the standardized residuals of the final case-mix regression models as “adjusted outcomes” [20,25]. This approach is therefore very similar to the adjustment method used by Resnik and Hart (residuals are defined as the difference between the observed and expected (“fitted”) values in regression analyses).
Using the Resnik and Hart approach, we obtain the following adjusted scores using our example:
Center A: +6
Center B: +4
Center C: −10
These adjusted scores provide information about how the centers perform in relation to their expected outcomes (e.g., Center C is performing less well than expected, whereas Center A is achieving better results) using the same unit as in EPIC-26. It would, for example, be possible to use the EPIC-26 minimally important differences (MIDs; lower threshold for incontinence: 6 [47]) to interpret these results better: in Center B, the “better performance” would be less than one MID, and therefore not perceptible for the patients, whereas a patient treated in Center C would notice the poorer performance.
Varagunam et al. used a similar approach to Nuttal et al., but based on the residuals, firstly, they calculated the difference in the observed and the expected value (ergo, the residuals) for every provider and then added this to the overall population’s observed average [30]:
a d j u s t e d   s c o r e = e x p e c t e d o b s e r v e d + o v e r a l l   o b s e r v e d   a v e r a g e .
In our example, the adjusted scores were:
Center A: 26 20 + 30 = 36
Center B: 37 33 + 30 = 34
Center C: 40 30 + 30 = 20
These scores are all within the original EPIC-26 limits (0–100) and are interpreted identically. The adjusted scores provide information about what the incontinence score of an “average” prostate cancer patient would be in the respective center.

3.8.2. Calculation of Transformed Adjusted Scores (Unit/Scale-Aberrant Approaches)

Two studies calculated the ratio o b s e r v e d e x p c t e d and used this “ratio indicator” as an adjusted outcome [26,27].
The ratio indicators for the PCO examples are:
Center A: 26 20     1.3
Center B: 37 33     1.12
Center C: 30 40     0.75
These adjusted scores are not interpretable using the original EPIC-26 unit. However, they quantify the deviation from the expected outcome: Centers A and B are both performing better than expected, resulting in a ratio > 1, whereas Center C is performing less well than expected, with a ratio < 1.
One study investigated the proportion of patients who experienced an improvement of at least one minimally important difference (MID) as the outcome of interest—i.e., they dichotomized the original PRO score [22]. They then calculated the expected number of events per provider on the basis of a previously published risk calculator, calculated the observed/expected ratio, and finally calculated the case-mix-adjusted event rate by multiplying the O/E ratio by the overall state-wide average.
The MID for the EPIC-26 incontinence score is 6, as described above. Let us assume that 20% of the patients of Center A, 40% of the patients of Center B, and 30% of the patients of Center C had an improvement of at least 6 points concerning their incontinence score after surgery. Appling the risk calculator for the centers’ populations, for Center A, 18% would have been expected, for Center B 35%, and for Center C 50%. For the state-wide population, an improvement of 30% is already known (this is a theoretical example). Using the “MID approach,” the adjusted scores would be:
Center A: 20 18   *   30   %     27   %
Center B: 40 35   *   30   %     34   %
Center C: 30 50   *   30   % = 18   %

3.8.3. Other

Sivaganesan et al. calculated the expected values for the providers examined, on the basis of their case-mix. However, the authors did not seem to use the expected values to compute any further “transformed” adjusted scores. They used a multilevel model [24].
For our example, the “adjusted scores” would be the same as the expected ones:
Center A: 20
Center B: 33
Center C: 40
Center C, which performed 10 points less well than expected, would be top ranked following this approach.
Gutacker et al. used another approach: based on a multilevel model using an hierarchical Bayesian approach with the PRO score (EQ-5D) used as an ordinal scaled variable, the predicted probabilities for each outcome per provider were calculated and reported as a quality metric [21].
Using a multilevel model as well, Gozalo et al. reported a provider ranking based on the predicted providers’ random intercepts [29].

3.8.4. Minimum Case Number Requirements for Statistical Modeling/Adjustment

Seven studies did not report any minimum case number requirements at the provider level for adjustment. The other studies differed widely in the minimum numbers of patients per provider that were required for adjustment. Gozalo et al. and Resnik and Hart set a minimum of eight patients per provider [28,29], whereas Farin et al. required at least 200 patients per provider [20]. Deutscher et al. only included providers in which at least 50% of the patient records were complete [25]. Most of the studies did not provide any explanations for the thresholds chosen.

4. Discussion

The results of the present systematic literature review on how to case-mix adjust PROs for provider comparison show that there is substantial variation in the literature with regard to several methodological criteria.
Firstly, the review shows that a uniform approach to case-mix adjustment of patient-reported outcomes does not currently exist. In most cases, regression models were used to adjust for relevant patient characteristics that differed between providers, with expected outcomes being calculated. Five of the eleven studies that were included made comparisons between expected and observed values, and three of these studies ultimately calculated an outcome that was evaluable in terms of the scale of the original one. One study seemed to only calculate fitted values and based the resulting provider ranking on them, which seems to be a questionable adjustment approach since providers with a “favorable” case-mix (e.g., younger or healthier patients) will always perform better than providers with a “difficult” case-mix if using this CMA approach. This can result in a biased ranking and does not provide information on the actual quality of care.
Secondly, there were also considerable differences between the studies included in the way in which expected values were calculated, and thus in the selection of adjustors. Relevant sociodemographic information and important clinical parameters (e.g., comorbidities) were included in most studies, and baseline patient-reported outcomes were included in all of them. There is little debate on including baseline PROs, as it is known from many observational and validation studies that PROs can vary considerably between care providers before the start of treatment even when adjustment is made for other patient characteristics [49]. It is thus essential to take baseline PROs into account in the adjustment.
Thirdly, nine of the studies included only reported patient characteristics, while two also included the characteristics of the providers that were being compared, such as information about the training of surgeons. Not taking provider characteristics into consideration is consistent with the purpose of CMA, since if outcomes are a function of patient characteristics, treatment effectiveness, and quality of care, and if the aim of the survey is to compare treatment quality, then considering provider characteristics may lead to over-adjustment. However, there are situations in which it might be justifiable to take provider characteristics into account—for example, if some contextual factors (such as state legislation) affect the outcomes that are exogenous, beyond the control of the provider. In principle, it is preferable to investigate provider characteristics in a further step in order to explain any differences in treatment quality in greater detail. This was done, for example, by Resnik and Hart [28], who looked at associations between provider characteristics and whether a practitioner was classified as an “expert physical therapist.” They found no association between years of experience and this expert status, for example.
Surprisingly, the MEDLINE search did not identify any cancer-related studies. However, in addition to the peer-reviewed studies that were included, there are also public quality initiatives designed to inform the general population about the quality of care in cancer settings, some of which include PROs. For example, in the National Prostate Cancer Audit (NPCA) in the United Kingdom, post-therapy PROs (EPIC-26) are adjusted for case-mix and reported using funnel plots [50]. This initiative, however, does not take baseline PROs into account for the adjustment, and the results have not yet been reported in peer-reviewed journals. Moreover, we excluded five studies that did use case-mix-adjusted PROs to compare providers, but unfortunately did not explain their adjustment methods [51,52,53,54,55]. However, those studies highlight that PROs—when case-mix adjusted—are valid and valuable parameters to compare different providers.

Limitations

We are aware of several limitations of our study: to begin with, only the MEDLINE database was used for the literature search. To find any other peer-reviewed research, all references of eligible studies were therefore scanned, and two more articles were included in this systematic review.
Moreover, articles not published in English, German, or French were excluded (language bias), which may result in a selection bias. Although a search strategy strictly following the PRISMA recommendations was used and an additional reference screening was performed, it is possible that some relevant studies were missed if the specific keywords were not used by either the authors or the MEDLINE database.
Since the included studies showed a huge variety in terms of the statistical methods used, it was not possible to perform a meta-analysis. However, as those statistical approaches themselves were the research subject, this was not expected.
For assessing the quality and risk of bias of the included studies, the CASP tool results are presented. The included studies are all retrospective analyses of prospectively collected data and the appraisal tools for cohort studies were mostly applicable. However, most of those appraisal tools incorporate items that are not applicable for our purposes, such as the assessment of the exposure of the outcome (e.g., [56]) and we decided to only use the CASP tool, which does not provide a numerical scale for assessing the quality of the studies, and thus the rating of each study is subjectively dependent on the reader’s main focus. This subjectivity may be a limitation, however, the use of the CASP tool results in a greater flexibility and usability of the risk of bias evaluation.
The review was conducted strictly according to PRISMA. We employed a concise and literature-based search string that was critically reviewed by multiple peers. The methodology used for this review was assessed for validity by the PROSPERO Foundation, where a protocol was pre-registered. A thorough quality assessment focusing on methodological issues was performed for each included study.

5. Conclusions and Recommendations

All in all, the variety of the approaches and the small numbers of published articles on the topic underscore the uncertainties concerning the best way of carrying out case-mix adjustment of PROs. To report interpretable PROs for all stakeholders (patients, clinicians, decision makers, researchers, etc.), it would be desirable for a more consistent methodology to be used across approaches.
Moreover, taking the heterogeneity of our findings into account, we would like to point out two additional major challenges of CMA of PROs—especially in contrast to non-PRO quality metrics—for which we see a need for further research:
Firstly, while most other quality metrics are rate based, with only two possible response categories (such as dead/alive or complication/no complication), PROs typically come in the form of continuous measures. PROs thus require data collection before and after an intervention (such as a hospital stay), because a patient’s functional status can change gradually, whereas in the other example, he/she is necessarily alive and could not have had an intervention-related complication before an intervention. Continuous measures with two points of measurement, however, require a number of analytical decisions associated with statistical pitfalls. For example: should changes from before and after an intervention be reported, or should adjusted scores after an intervention be presented that include the pre-intervention score as an adjustor? The included studies in this review mostly adjusted the scores after intervention. However, some study groups decided to report changes (e. g., [52]) and qualitative research on how to present PROs to clinicians often highlights the importance of presenting trends over time [56,57].
Secondly, while traditional quality metrics data are typically collected without—or with little—active contribution from the patient and are thus mostly documented for all patients (although with varying documentation quality), PRO data collection requires active involvement of the patients, who have to complete the questionnaires. This leads to patient selection and may create substantial bias. For example, patients with some characteristics are less likely to complete questionnaires (due to factors such as age, educational level, or their degree of trust in the healthcare system) [58]. Similarly, providers may be reluctant to include some patients because they expect them to have a negative impact on their results—e.g., patients with more comorbidities tend to have poorer PRO scores [59]. Correcting such potential biases is difficult because of the unknown characteristics of the unsurveyed patients, so weighting is scarcely possible. In practice, this may lead to a decision to include only those providers in comparisons for which a minimum number (or proportion) of patients are included in the analytical dataset. Thus, in our opinion, quality improvement initiatives based on PRO comparisons should be encouraged to include all eligible patients and report on potential selection biases.
Due to the lack of comparisons between different adjustment methods for the same datasets, there is no strong scientific evidence for favoring one approach over the others. Recommendations on what a “gold standard” should include are therefore difficult to make at present. We would therefore cautiously derive three recommendations as a starting point for further discussion, on the basis of what appear to be successful practices in the studies included in this review. In order to keep the threshold for comparisons between treatment outcomes with the help of PROs as low as possible, so that they are accessible to a broad public, we would recommend the following on the basis of the present findings:
Firstly, with regard to baseline PROs, it appears to be common methodological practice to include them. All of the studies feature them either as an adjustor in the model, or as part of the dependent variable when the difference between baseline and follow-up PRO is used. The reasons for doing so are well elaborated on in some of the studies included [23,27].
Secondly, a number of the studies emphasized that the CMA should be accessible by different stakeholders (e.g., clinicians, patients) and should therefore be easily interpretable and offer additional information that could be used for better selection, or to identify providers with good performance [20,23,25]. In our view, this notion favors unit/scale-preserving approaches, so that a new “unit” does not need to be learned.
Thirdly, in line with most of the studies included, provider characteristics should not be included unless there are strong reasons for this such as contextual factors that are beyond the providers’ control.
In the absence of strong evidence favoring one approach over any of the others, we would suggest that further research should be carried out, including comparison of approaches both in terms of the statistics used and also the comprehensibility of the reported results for the target audiences of the provider comparisons. Among the many open questions are minimum case thresholds, whether to take the hierarchical structure of the data into account using multilevel models, as well as measuring the impact that CMA has on rankings and the threshold at which case-mix-adjusted differences in PROs should be considered relevant at all. The variation in CMA approaches underlines the importance of establishing a methodological guideline or “checklist” that could be used, for example, for quality improvement initiatives aiming to use PROs for provider comparisons. As the evidence is still scarce, we would recommend—alongside further research on the benefits and pitfalls of different CMA methods—a consensus-based approach to the development of such a guideline.

Author Contributions

Conceptualization: N.T.S.; Methodology: N.T.S., C.K.; Formal Analysis: N.T.S., C.K.; Investigation: N.T.S., C.K.; Data Curation: N.T.S.; Writing—Original Draft Preparation: N.T.S., C.K.; Writing—Review and Editing: N.T.S., C.K., H.P., S.W., C.B.; Visualization: N.T.S.; Supervision: H.P., S.W.; Project Administration: S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Search String for MEDLine, Including Number of Hits of MEDLine Search

(“patient reported outcome measures”[MeSH Terms] OR (“patient”[All Fields] AND “reported”[All Fields] AND “outcome”[All Fields] AND “measures”[All Fields]) OR “patient reported outcome measures”[All Fields] OR (“patient”[All Fields] AND “reported”[All Fields] AND “outcome”[All Fields]) OR “patient reported outcome”[All Fields]): 747 hits
AND
((“casemix”[All Fields] OR “casemixes”[All Fields] OR (“risk”[MeSH Terms] OR “risk”[All Fields])) AND (“social adjustment”[MeSH Terms] OR (“social”[All Fields] AND “adjustment”[All Fields]) OR “social adjustment”[All Fields] OR “adjustment”[All Fields])): 704 hits
AND
(“hasabstract”[Text Word] AND “humans”[MeSH Terms] AND (“English”[Language] OR “French”[Language] OR “German”[Language])): 640 hits

Appendix B

Table A1. CASP tool for included studies.
Table A1. CASP tool for included studies.
AuthorClearly FocusedRecruitement AcceptableExposure Accuartely MeasuredOutcome Accurately MeasuredIdentified ConfoundingAccounted for ConfoundingSubject Follow-Up Complete EnoughSubject Follow-Up Long EnoughResults PreciseBelieve ResultsApplicable ResultsFit with Other EvidenceImplications for Practice
Deutscher et al., 2018 [25]xcan’t tellxxxxcan’t tellcan’t tellxxxxx
Farin et al., 2009 [20]xxcan’t tellxcan’t tellcan’t tellcan’t tellcan’t tellxxxxx
Gozalo et al., [29]xcan’t tellxxxxcan’t tellxxxxxx
Gutacker et al., 2013 [21]xxxxxxxxxxxxx
Hendryx et al., 1999 [26]xxxxcan’t tellxxcan’t tellxxcan’t tellxx
Khor et al., 2020 [22]xxxxxxxxxxxxx
Lutz et al., 2020 [27]xxxxxxxcan’t tellxxxxx
Nuttall et al., 2015 [23]xcan’t tellxxxxcan’t tellxxxxxx
Resnik & Hart 2003 [28]xcan’t tellxxxxcan’t tellxxxxxx
Sivaganesan et al., 2018 [24]xcan’t tellxxxxxxxxxxx
Varagunam et al., 2015 [30]xcan’t tellxxxxcan’t tellxxxxxx

References

  1. Klinkhammer-Schalke, M.; Koller, M.; Steinger, B.; Ehret, C.; Ernst, B.; Wyatt, J.C.; Hofstadter, F.; Lorenz, W.; Regensburg Qo, L.S.G. Direct improvement of quality of life using a tailored quality of life diagnosis and therapy pathway: Randomised trial in 200 women with breast cancer. Br. J. Cancer 2012, 106, 826–838. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Klinkhammer-Schalke, M.; Steinger, B.; Koller, M.; Zeman, F.; Fürst, A.; Gumpp, J.; Obermaier, R.; Piso, P.; Lindberg-Scharf, P.; Altendorfer, A.; et al. Diagnosing deficits in quality of life and providing tailored therapeutic options: Results of a randomised trial in 220 patients with colorectal cancer. Eur. J. Cancer 2020, 130, 102–113. [Google Scholar] [CrossRef] [PubMed]
  3. Fromme, E.K.; Eilers, K.M.; Mori, M.; Hsieh, Y.C.; Beer, T.M. How accurate is clinician reporting of chemotherapy adverse effects? A comparison with patient-reported symptoms from the Quality-of-Life Questionnaire C30. J. Clin. Oncol. 2004, 22, 3485–3490. [Google Scholar] [CrossRef] [PubMed]
  4. Laugsand, E.A.; Sprangers, M.A.; Bjordal, K.; Skorpen, F.; Kaasa, S.; Klepstad, P. Health care providers underestimate symptom intensities of cancer patients: A multicenter European study. Health Qual. Life Outcomes 2010, 8, 104. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Basch, E.; Deal, A.M.; Kris, M.G.; Scher, H.I.; Hudis, C.A.; Sabbatini, P.; Rogak, L.; Bennett, A.V.; Dueck, A.C.; Atkinson, T.M.; et al. Symptom Monitoring with Patient-Reported Outcomes during Routine Cancer Treatment: A Randomized Controlled Trial. J. Clin. Oncol. 2015, 34, 557. [Google Scholar] [CrossRef]
  6. Basch, E.; Deal, A.M.; Dueck, A.C.; Scher, H.I.; Kris, M.G.; Hudis, C.; Schrag, D. Overall Survival Results of a Trial Assessing Patient-Reported Outcomes for Symptom Monitoring During Routine Cancer Treatment. JAMA 2017, 318, 197–198. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Denis, F.; Yossi, S.; Septans, A.-L.; Charron, A.; Voog, E.; Dupuis, O.; Ganem, G.; Pointreau, Y.; Letellier, C. Improving survival in patients treated for a lung cancer using self-evaluated symptoms reported through a web application. Am. J. Clin. Oncol. 2017, 40, 464–469. [Google Scholar] [CrossRef]
  8. Zeegers, M.P.; Bryan, R.T.; Langford, C.; Billingham, L.; Murray, P.; Deshmukh, N.S.; Hussain, S.; James, N.; Wallace, D.M.A.; Cheng, K.K. The West Midlands Bladder Cancer Prognosis Programme: Rationale and design. BJU Int. 2010, 105, 784–788. [Google Scholar] [CrossRef]
  9. Engelen, V.; Detmar, S.; Koopman, H.; Maurice-Stam, H.; Caron, H.; Hoogerbrugge, P.; Egeler, R.M.; Kaspers, G.; Grootenhuis, M. Reporting health-related quality of life scores to physicians during routine follow-up visits of pediatric oncology patients: Is it effective?: Patient Reported Outcomes in Pediatric Clinic. Pediatr. Blood Cancer 2012, 58, 766–774. [Google Scholar] [CrossRef]
  10. Trautmann, F.; Hentschel, L.; Hornemann, B.; Rentsch, A.; Baumann, M.; Ehninger, G.; Schmitt, J.; Schuler, M. Electronic real-time assessment of patient-reported outcomes in routine care—First findings and experiences from the implementation in a comprehensive cancer center. Support. Care Cancer 2016, 24, 3047–3056. [Google Scholar] [CrossRef]
  11. Slovenian National Institute of Public Health. Report on the Basis of the Analysis of Data from the Survey on National Cancer Control. Programmes/Cancer Documents in EU—An IPAAC Report; Ljublja. 2019. Available online: https://www.ipaac.eu/res/file/outputs/wp10/national-cancer-control-plans-survey.pdf (accessed on 2 August 2021).
  12. ICHOM—International Consortium for Health Outcomes Measurement Localized Prostate Cancer Data Collection Reference Guide. Localized Prostate Cancer Data Collection Reference Guide. 2015. Available online: http://ichom.org/files/medical-conditions/localized-prostate-cancer/localized-prostate-cancer-reference-guide.pdf (accessed on 4 August 2021).
  13. Evans, S.M.; Millar, J.L.; Moore, C.M.; Lewis, J.D.; Huland, H.; Sampurno, F.; Connor, S.E.; Villanti, P.; Litwin, M.S. Cohort profile: The TrueNTH Global Registry—An international registry to monitor and improve localised prostate cancer health outcomes. BMJ Open 2017, 7, e017006. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Kowalski, C.; Roth, R.; Carl, G.; Feick, G.; Oesterle, A.; Hinkel, A.; Steiner, T.; Brock, M.; Kaftan, B.; Borowitz, R.; et al. A multicenter paper-based and web-based system for collecting patient-reported outcome measures in patients undergoing local treatment for prostate cancer: First experiences. J. Patient-Rep. Outcomes 2020, 4, 56. [Google Scholar] [CrossRef] [PubMed]
  15. Scheibe, M.; Herrmann, A.; Schmitt, J.; Einhart, N.; Sedlmayr, B.; Kowalski, C. Implementation of patient-reported outcome assessment in routine cancer care: A systematic review of multicentric programs in Europe. Z. Für Evidenz Fortbild. Qual. Im Gesundheitswesen 2020, 156–157, 11–23. [Google Scholar] [CrossRef] [PubMed]
  16. Iezzoni, L. Risk Adjustment for Measuring Healthcare Outcomes, 4th ed.; Health Administration Press: Chicago, IL, USA, 2012. [Google Scholar]
  17. OECD. Measuring What Matters: The Patient-Reported Indicator Surveys; OECD: Paris, France, 2019. [Google Scholar]
  18. McGowan, J.; Sampson, M.; Salzwedel, D.M.; Cogo, E.; Foerster, V.; Lefebvre, C. PRESS Peer Review of Electronic Search Strategies: 2015 Guideline Statement. J. Clin. Epidemiol. 2016, 75, 40–46. [Google Scholar] [CrossRef] [Green Version]
  19. Critical Appraisal Skills Programme CASP Cohort Study Checklist. 2018. Available online: https://casp-uk.net/wp-content/uploads/2018/01/CASP-Cohort-Study-Checklist_2018.pdf (accessed on 18 March 2021).
  20. Farin, E.; Jäckel, W.H.; Schalaster, V. Projektgruppe QS-Reha-Verfahren in der AQMS The quality assurance programme of the statutory health insurance funds in medical rehabilitation: Results and further developments. Gesundh. Bundesverb. Arzte Offentlichen Gesundh. 2009, 71, 163–174. [Google Scholar] [CrossRef]
  21. Gutacker, N.; Bojke, C.; Daidone, S.; Devlin, N.; Street, A. Hospital variation in patient-reported outcomes at the level of EQ-5D dimensions: Evidence from England. Med. Decis. Mak. Int. J. Soc. Med. Decis. Mak. 2013, 33, 804–818. [Google Scholar] [CrossRef] [Green Version]
  22. Khor, S.; Lavallee, D.C.; Cizik, A.M.; Bellabarba, C.; Dagal, A.; Hart, R.A.; Howe, C.R.; Martz, R.D.; Shonnard, N.; Flum, D.R. Hospital and Surgeon Variation in Patient-reported Functional Outcomes After Lumbar Spine Fusion: A Statewide Evaluation. Spine 2020, 45, 465–472. [Google Scholar] [CrossRef]
  23. Nuttall, D.; Parkin, D.; Devlin, N. Inter-provider comparison of patient-reported outcomes: Developing an adjustment to account for differences in patient case mix. Health Econ. 2015, 24, 41–54. [Google Scholar] [CrossRef]
  24. Sivaganesan, A.; Asher, A.L.; Bydon, M.; Khan, I.; Kerezoudis, P.; Foley, K.T.; Nian, H.; Harrell, F.E.; Archer, K.R.; Devin, C.J. A Strategy for Risk-adjusted Ranking of Surgeons and Practices Based on Patient-reported Outcomes after Elective Lumbar Surgery. Spine 2019, 44, 670–677. [Google Scholar] [CrossRef]
  25. Deutscher, D.; Werneke, M.W.; Hayes, D.; Mioduski, J.E.; Cook, K.F.; Fritz, J.M.; Woodhouse, L.J.; Stratford, P.W. Impact of Risk Adjustment on Provider Ranking for Patients With Low Back Pain Receiving Physical Therapy. J. Orthop. Sports Phys. Ther. 2018, 48, 637–648. [Google Scholar] [CrossRef]
  26. Hendryx, M.S.; Dyck, D.G.; Srebnik, D. Risk-adjusted outcome models for public mental health outpatient programs. Health Serv. Res. 1999, 34, 171–195. [Google Scholar] [PubMed]
  27. Lutz, A.D.; Brooks, J.M.; Chapman, C.G.; Shanley, E.; Stout, C.E.; Thigpen, C.A. Risk Adjustment of the Modified Low Back Pain Disability Questionnaire and Neck Disability Index to Benchmark Physical Therapist Performance: Analysis From an Outcomes Registry. Phys. Ther. 2020, 100, 609–620. [Google Scholar] [CrossRef]
  28. Resnik, L.; Hart, D.L. Using clinical outcomes to identify expert physical therapists. Phys. Ther. 2003, 83, 990–1002. [Google Scholar] [CrossRef] [Green Version]
  29. Gozalo, P.L.; Resnik, L.J.; Silver, B. Benchmarking Outpatient Rehabilitation Clinics Using Functional Status Outcomes. Health Serv. Res. 2016, 51, 768–789. [Google Scholar] [CrossRef] [Green Version]
  30. Varagunam, M.; Hutchings, A.; Black, N. Do patient-reported outcomes offer a more sensitive method for comparing the outcomes of consultants than mortality? A multilevel analysis of routine data. BMJ Qual. Saf. 2015, 24, 195–202. [Google Scholar] [CrossRef] [PubMed]
  31. Ware, J.; Kosinski, M.; Keller, S.D. A 12-Item Short-Form Health Survey: Construction of scales and preliminary tests of reliability and validity. Med. Care 1996, 34, 220–233. [Google Scholar] [CrossRef] [Green Version]
  32. Haley, S.M.; McHorney, C.A.; Ware, J.E. Evaluation of the MOS SF-36 physical functioning scale (PF-10): I. Unidimensionality and reproducibility of the Rasch item scale. J. Clin. Epidemiol. 1994, 47, 671–684. [Google Scholar] [CrossRef]
  33. McGee, J.; Goldfield, N.; Riley, K.; Morton, J. Collecting Information from Health Care Consumers: A Resource Manual of Tested Questionnaires and Practical Advice; Apsen Publications: Gaithersburg, MD, USA, 1996. [Google Scholar]
  34. Lehman, A.F.; Postrado, L.T.; Rachuba, L.T. Convergent validation of quality of life assessments for persons with severe mental illnesses. Qual. Life Res. Int. J. Qual. Life Asp. Treat. Care Rehabil. 1993, 2, 327–333. [Google Scholar] [CrossRef] [PubMed]
  35. Rabin, R.; de Charro, F. EQ-5D: A measure of health status from the EuroQol Group. Ann. Med. 2001, 33, 337–343. [Google Scholar] [CrossRef] [PubMed]
  36. Bührlen, B.; Gerdes, N.; Jäckel, W.H. Development and psychometric testing of a patient questionnaire for medical rehabilitation (IRES-3). Die Rehabil. 2005, 44, 63–74. [Google Scholar] [CrossRef]
  37. Chiarotto, A.; Maxwell, L.J.; Ostelo, R.W.; Boers, M.; Tugwell, P.; Terwee, C.B. Measurement Properties of Visual Analogue Scale, Numeric Rating Scale, and Pain Severity Subscale of the Brief Pain Inventory in Patients With Low Back Pain: A Systematic Review. J. Pain 2019, 20, 245–263. [Google Scholar] [CrossRef]
  38. Murray, D.W.; Fitzpatrick, R.; Rogers, K.; Pandit, H.; Beard, D.J.; Carr, A.J.; Dawson, J. The use of the Oxford hip and knee scores. J. Bone Jt. Surg. Br. 2007, 89, 1010–1014. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Fritz, J.M.; Irrgang, J.J. A comparison of a modified Oswestry Low Back Pain Disability Questionnaire and the Quebec Back Pain Disability Scale. Phys. Ther. 2001, 81, 776–788. [Google Scholar] [CrossRef] [PubMed]
  40. Vernon, H.; Mior, S. The Neck Disability Index: A study of reliability and validity. J. Manip. Physiol. Ther. 1991, 14, 409–415. [Google Scholar]
  41. Fairbank, J.C.; Pynsent, P.B. The Oswestry Disability Index. Spine 2000, 25, 2940–2952. [Google Scholar] [CrossRef]
  42. Dempster, M.; Donnelly, M.; O′Loughlin, C. The validity of the MacNew Quality of Life in heart disease questionnaire. Health Qual. Life Outcomes 2004, 2, 6. [Google Scholar] [CrossRef] [Green Version]
  43. Swiontkowski, M.F.; Engelberg, R.; Martin, D.P.; Agel, J. Short musculoskeletal function assessment questionnaire: Validity, reliability, and responsiveness. J. Bone Jt. Surg. Am. 1999, 81, 1245–1260. [Google Scholar] [CrossRef] [PubMed]
  44. Hart, D.L.; Werneke, M.W.; Wang, Y.-C.; Stratford, P.W.; Mioduski, J.E. Computerized adaptive test for patients with lumbar spine impairments produced valid and responsive measures of function. Spine 2010, 35, 2157–2164. [Google Scholar] [CrossRef]
  45. Schokkaert, E.; Van de Voorde, C. Direct versus indirect standardization in risk adjustment. J. Health Econ. 2009, 28, 361–374. [Google Scholar] [CrossRef]
  46. Sibert, N.T.; Dieng, S.; Oesterle, A.; Feick, G.; Carl, G.; Steiner, T.; Minner, J.; Roghmann, F.; Kaftan, B.; Zengerling, F.; et al. Psychometric validation of the German version of the EPIC-26 questionnaire for patients with localized and locally advanced prostate cancer. World J. Urol. 2019, 39, 11–25. [Google Scholar] [CrossRef]
  47. Skolarus, T.A.; Dunn, R.L.; Sanda, M.G.; Chang, P.; Greenfield, T.K.; Litwin, M.S.; Wei, J.T. Minimally important difference for the Expanded Prostate Cancer Index Composite Short Form. Urology 2015, 85, 101–105. [Google Scholar] [CrossRef] [Green Version]
  48. Farin, E.; Follert, P.; Gerdes, N.; Jäckel, W.H.; Thalau, J. Quality assessment in rehabilitation centres: The indicator system “Quality Profile”. Disabil. Rehabil. 2004, 26, 1096–1104. [Google Scholar] [CrossRef]
  49. Roth, R.; Dieng, S.; Oesterle, A.; Feick, G.; Carl, G.; Hinkel, A.; Steiner, T.; Kaftan, B.T.; Kunath, F.; Hadaschik, B.; et al. Determinants of self-reported functional status (EPIC-26) in prostate cancer patients prior to treatment. World J. Urol. 2021, 39, 27–36. [Google Scholar] [CrossRef] [Green Version]
  50. The Royal College of Surgeons of England. National Prostate Cancer Audit Annual Report 2020. 2021. Available online: https://www.npca.org.uk/content/uploads/2021/01/NPCA-Annual-Report-2020_Final_140121.pdf (accessed on 2 August 2021).
  51. Neuburger, J.; Cromwell, D.A.; Hutchings, A.; Black, N.; van der Meulen, J.H. Funnel plots for comparing provider performance based on patient-reported outcome measures. BMJ Qual. Saf. 2011, 20, 1020–1026. [Google Scholar] [CrossRef]
  52. Waljee, J.F.; Ghaferi, A.; Cassidy, R.; Varban, O.; Finks, J.; Chung, K.C.; Carlozzi, N.E.; Dimick, J.B. Are Patient-reported Outcomes Correlated With Clinical Outcomes After Surgery?: A Population-based Study. Ann. Surg. 2016, 264, 682–689. [Google Scholar] [CrossRef] [PubMed]
  53. Waljee, J.F.; Ghaferi, A.; Finks, J.F.; Cassidy, R.; Varban, O.; Carlin, A.; Carlozzi, N.; Dimick, J. Variation in Patient-reported Outcomes Across Hospitals Following Surgery. Med. Care 2015, 53, 960–966. [Google Scholar] [CrossRef] [PubMed]
  54. Desai, A.; Bekelis, K.; Ball, P.A.; Lurie, J.; Mirza, S.K.; Tosteson, T.D.; Zhao, W.; Weinstein, J.N. Outcome Variation Across Centers after Surgery for Lumbar Stenosis and Degenerative Spondylolisthesis. Spine 2013, 38, 678–691. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Yen, S.-C.; Corkery, M.B.; Chui, K.K.; Manjourides, J.; Wang, Y.-C.; Resnik, L.J. Risk Adjustment for Lumbar Dysfunction: Comparison of Linear Mixed Models With and Without Inclusion of Between-Clinic Variation as a Random Effect. Phys. Ther. 2015, 95, 1692–1702. [Google Scholar] [CrossRef] [Green Version]
  56. Brundage, M.D.; Smith, K.C.; Little, E.A.; Bantug, E.T.; Snyder, C.F. The PRO Data Presentation Stakeholder Advisory Board Communicating patient-reported outcome scores using graphic formats: Results from a mixed-methods evaluation. Qual. Life Res. 2015, 24, 2457–2472. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Sibert, N.T.; Kowalski, C.; Pfaff, H.; Wesselmann, S.; Breidenbach, C. Clinicians’ knowledge and attitudes towards patient reported outcomes in colorectal cancer care—Insights from qualitative interviews. BMC Health Serv. Res. 2021, 21, 366. [Google Scholar] [CrossRef]
  58. Nguyen, H.; Butow, P.; Dhillon, H.; Sundaresan, P. A review of the barriers to using Patient-Reported Outcomes (PROs) and Patient-Reported Outcome Measures (PROMs) in routine cancer care. J. Med. Radiat. Sci. 2021, 68, 186–195. [Google Scholar] [CrossRef] [PubMed]
  59. Luque Ramos, A.; Redeker, I.; Hoffmann, F.; Callhoff, J.; Zink, A.; Albrecht, K. Comorbidities in Patients with Rheumatoid Arthritis and Their Association with Patient-reported Outcomes: Results of Claims Data Linked to Questionnaire Survey. J. Rheumatol. 2019, 46, 564–571. [Google Scholar] [CrossRef] [PubMed]
Figure 1. PRISMA flowchart. PRO, patient-reported outcome; CMA, case-mix adjustment.
Figure 1. PRISMA flowchart. PRO, patient-reported outcome; CMA, case-mix adjustment.
Cancers 13 03964 g001
Table 1. Characteristics of eligible studies.
Table 1. Characteristics of eligible studies.
AuthorStudy DesignParticipants/SettingData SourcesStudy Size (Complete/Included Datasets)Outcome (PRO)No. of Variables Used for Case-Mix Adjustment (and Specific Adjustors)Model R2
Deutscher et al., 2018 [25]Retrospective analysis of prospectively collected dataAdult patients with lower back pain treated in outpatient physical therapy clinics (2014–2016)FOTOn = 341,642 patients
n = 6934 clinicians
n = 2107 hospitals
LCAT11: LCAT score at admission, age, sex, acuity, payer, surgical history, physical exercise history, medication/previous treatment, type of surgery, comorbidities0.373
Farin et al., 2009 [20]Analysis of quality assurance dataQuality assurance report for statutory rehabilitation institutions in Germany on: 1, musculoskeletal; 2, cardiological; and 3, neurological diseasesQS Reha®1: n = 4045 patients in n = 27 hospitals
2: n = 2503 patients in n = 25 hospitals
3: n = 1477 patients in n = 12 hospitals
1: IRES
2: SMFA
3: MacNew
Reported elsewhere [48]: 10:
age, sex, somatic, functional and psychosocial status at admission, diagnosis, multimorbidity, treatment motivation, application for pension, education
Not reported
Gozalo et al., [29]Retrospective analysis of prospectively collected dataOutpatient rehabilitation clinics for patients with lower back painFOTOn = 90,392 patients
n = 2040 therapists
n = 538 clinics
FOTO overall health
status measure: OHS, PCS, PF-10
8: PRO at intake, age, gender, onset, surgery count, functional comorbidity index, fear avoidance level, payer typeOHS: 0.416
PCS: 0.345
PF-10: 0.421
Gutacker et al., 2013 [21]Retrospective analysis of prospectively collected dataAnalysis of hospital performance variation in hip replacement surgeryNHS PROMs and routinely collected inpatient data (HES)n = 21,565 patients
n = 153 hospitals
EQ-5D10: pre-treatment PROs, age, gender, deprivation index, weighted Charlson Index of comorbidities, number of additional comorbidities (not included in Charlson Index), time between pre-operative assessment and admission, primary surgery, revision surgery, treatmentNot reported
Hendryx et al., 1999 [26]Retrospective analysis of prospectively collected dataDevelopment of outcome risk adjustment models for public mental health outpatient treatment programsPatient-reported, case manager ratings, management information system datan = 289 patients
n = 6 mental health agencies
SF-12, ADL, Lehman Quality of Life Interviews8: sex, age, race, presence of severe primary diagnosis (major depression, schizophrenia, bipolar disorder), baseline levels of substance abuse, baseline PRO, baseline quality of life, baseline satisfaction with services0.34
Khor et al., 2020 [22]Retrospective analysis of prospectively collected dataAnalysis of variation in PROs 1 year after elective lumbar fusion surgery across surgeons and hospitalsSpine Care and Outcomes Assessment Program (Washington State, USA)n = 737 patients
n = 17 hospitals
n = 58 surgeons
ODI11: age, sex, insurance status, race, ASA, smoking status, prior spine surgery, diagnosis, opiate use, asthma, baseline PRONot reported
Lutz et al., 2020 [27]Retrospective analysis of prospectively collected dataBenchmarking physical therapist studyATI Patient Outcomes Registryn = 182,276 patients
n = 2799 physical therapists
MDQ, NDI10: sex, age, BMI, initial PRO, payer type, physical component score and mental component score of the Veterans RAND 12-Item Health Survey, state of physical therapy services, type of pain (acute/chronic), comorbiditiesMDQ: 0.19
NDI: 0.19
Nuttall et al., 2015 [23]Retrospective analysis of prospectively collected dataFeasibility study to analyze case-mix adjustment methodology for elective surgeryNHS PROMs and HESn = 30,555 patients
n = 237 providers
OKS11: baseline PRO, sex, ethnicity, age, deprivation index, assistance for completing questionnaires (baseline and follow-up), disability, previous surgery, comorbidities, type of surgery, length of post-operative stay0.258 (ordinary least square model)
0.257 (fixed effects model)
Resnik & Hart 2003 [28]Retrospective analysis of prospectively collected dataOutpatient rehabilitation clinics for patients with lower back painFOTOn = 24,276 patients
n = 930 therapists
n = 354 hospitals
FOTO overall health
status measure (OHS), PCS, PF--10
8: age, employment, exercise history, sex, intake PRO, onset, reimbursement, surgeryOHS: 0.416
PCS: 0.345
PF-10: 0.421
Sivaganesan et al., 2018 [24]Retrospective analysis of prospectively collected dataReport of a risk-adjusted ranking of spine surgeons and sites performing elective lumbar surgeryQuality and Outcome Database (QOD)n = 8834 patients
n = 124 surgeons
n = 21 sites
ODI, EQ-5D, VAS-BP/VAS-LP22 (19 at patient level: age, BMI, ethnicity, education, smoking status, opioid use, comorbidities, pre-operative symptoms, motor deficit, ASA, symptom duration, interbody graft placement, worker’s compensation, liability claims, insurance status, employment, baseline PRO);
3 constructs at surgeon level: site ID, years in practice, fellowship training)
Not applicable (Bayesian model)
Varagunam et al., 2015 [30]Retrospective analysis of prospectively collected dataFeasibility study analyzing different approaches on how to adjust PROM scores for individual consultant comparison for elective surgeryNHS PROMsN = 65,465 (hip), 68,107 (knee) and 38,965 (hernia) patients
N = 948 (hip), 1130 (knee) and 974 (hernia) consultants
N = 183 (hip), 188 (knee) and 197 (hernia) clinics
OHS,
OKS, EQ-5D
6: age, sex, deprivation index, comorbidities, previous surgery, baseline PRONot reported
ADL, Activities of Daily Living; ASA, American Society of Anesthesiologist score, ATI, Athletic Therapeutic Institute; BMI, Body Mass Index; EQ-5D, European Quality of Life 5 Dimensions; FOTO, Focus on Therapeutic Outcomes; HES, Hospital Episode Statistics; IRES, Indikatoren des Reha-Status (Indicators for Rehabilitation Status); LCAT, Lumbar Computerized Adaptive Test; MDQ, Modified Low Back Pain Disability Questionnaire; NDI, Neck Disability Index; NHS, National Health Service (UK); ODI, Oswestry Disability Index; OHS, Oxford Hip Score; OKS, Oxford Knee Score; PCS, Physical Component Summary; PF-10, Physical Functioning Scale; PRO, Patient-Reported Outcome; PROM, Patient-Reported Outcome Measure; SF-12, Short Form 12; SMFA, Short Musculoskeletal Function Assessment questionnaire; VAS-BP/VAS-LP, Visual Analog Scale for Back Pain/for Leg Pain.
Table 2. Synthesis of case-mix adjustment methods.
Table 2. Synthesis of case-mix adjustment methods.
AuthorWhat Type of Provider?Iezzoni’s FourWhich Adjustors?Are Criteria for the Selection of Adjustors Met? 1How is Case-Mix Adjustment Statistically Performed?How are Case-Mix Adjustment Results Reported?
Risk of What Outcome?For What Population?Over What Time Period?For What Purpose?Socio-Economic Information?Baseline PROM Included?Comor-Bidities?Adjustment
for ProVider Characteristics?
Are the Adjustors Associated with the Outcome?Do Adjustors Vary Across Providers?
Deutscher et al., 2018 [25]Outpatient physical therapist clinics (US)LCATLower back pain patients treated in outpatient settingTime between admission and dischargeClinical ranking for performance measuringxxxYes, assessed via back-step regression model approachNot analyzedMultiple regression, >50% complete patients per hospital and ≥10 complete cases per clinician per year requiredComparison between adjusted and unadjusted scores (agreement: 30%), ranking changed 70% of the hospitals, reported as percentile ranking
Farin et al., 2009 [20]In- and outpatient rehab. institutions (Germany)IRES, SMFA, MacNewPatients in statutory rehab. institutionsTime between start and end of rehab.Quality assurancexxxNot reportedNot reportedMultilevel regression, taking the standardized residuals in the prediction as the case-mix-adjusted outcome, >199 patients per institution required95% confidence intervals reported graphically
Gozalo et al., 2016 [29]Outpatient rehabilitation clinics (US)FOTO CATs, either generic or body impairment specificPatients treated in outpatient rehabilitation clinics (for orthopedic conditions)At intake and discharge from outpatient rehabilitationIdentify expert physical therapistsxxx-Not assessedNot reportedGeneralized linear model with >7 patients per clinic requiredRanking of clinics based on predicted clinic random intercept (including 95% confidence interval)
Gutacker et al., 2013 [21]NHS-funded hospitals (UK)EQ-5DPatients receiving hip replacement in an NHS-funded hospitalTime between before/day of admission and 6 months after surgeryInter-provider comparison for performance measuringxxxYes, assessed via regression modelingNot reportedMultilevel linear regression and probit model, no minimum number of patients per hospital required95% confidence intervals reported graphically
Hendryx et al., 1999 [26]Outpatient mental health services (US)SF-12 (functional status domain), Lehman Quality of Life Interview (Quality of Life domain)Adult patients treated in outpatient mental health agenciesAt baseline, 5 and 10 monthsInter-provider comparison for performance measuringxxxNot reportedNot assessedMultiple linear regression, then calculating the ratio: observed/expected, no minimum number of patients per agency requiredObserved score ranks and observed/expected ratio ranks are compared
Khor et al., 2020 [22]In-patient hospitals performing elective spine surgery (USA)ODI (probability of improvement beyond MID and reaching minimal disability level)Adults treated with elective lumbar fusion surgeryAt 0–60 days before surgery and at 12-month follow-upInter-provider comparison for performance measuringxxxYes (association based on findings from earlier studies by the research group)For all adjustors except for age, sex, and smoking status1. Multiple logistic model
2. Calculation of an expected number of events per provider based on a previously published risk calculator
3. Calculation of observed/expected ratio
4. Calculation of a case-mix-adjusted event rate by multiplying the O/E ratio by the overall state-wide average;
>9 patients per provider required
Adjusted and unadjusted scores, ICCs to describe the proportion of total variability accounted for by between-provider variance
Lutz et al., 2020 [27]Outpatient physical therapists (USA)MDQ, NDIAdults with episodes related to the lower back and neck with no history of related surgeryNot specifiedBenchmarking physical therapistsxxxx (state of physical therapy services)Not reportedNot assessed1. Multiple linear regression
2. Calculation of an observed/expected ratio; >39 patients per physical therapist required
Reporting of counts of “outperforming, meeting, and underperforming” physical therapists (adjusted and unadjusted) based on 95% confidence intervals
Nuttall et al., 2015 [23]NHS-funded hospitals (UK)OKSPatients treated by NHS providers for unilateral knee replacementBefore and 3 or 6 months after surgeryInter-provider comparison for performance measuringxxxYes (association based on findings from earlier studies)Not analyzedMultilevel linear regression: generalized least squares with fixed effects, general least squares with random effects, ordinary least squares, then calculating: (observed/expected) * observed mean overall; no minimum number of patients per hospital requiredDifferences between different adjustment scores reported, ranking reported as funnel plots
Resnik & Hart 2003 [28]Outpatient rehab. clinics (USA)OHS, PCS, PF-10Patients treated in outpatient rehab. clinics for lower back painAt intake and discharge from outpatient rehabilitationIdentify expert physical therapistsxxxYes, bivariate analysis to identify adjustorsNot reportedGeneralized linear model with >7 patients per therapist required, then calculating the difference: observed—expected; no minimum number of patients per clinic requiredPercentiles of differences reported
Sivaganesan et al., 2018 [24]Sites performing spine surgery (USA)ODI, EQ-5D, VAS-BP/VAS-LPPatients treated with spine surgeryBefore and 1 year after surgeryInter-provider comparison for performance measuringxxxxYes, but no association coefficients are reportedNot reportedRandom effects regression models and multilevel hierarchical Bayesian models (3 levels); no minimum number of patients per hospital requiredBox plots for adjusted ranks per surgeon per site reported
Varagunam et al., 2015 [30]NHS-funded hospitals (UK)OHS, OKS, EQ-5DPatients treated by consultant in NHS providersBefore and three or six months after surgeryInter-consultant comparison for performance measuringxxx-Not assessedNot reportedMultiple linear regression, >39 patients per NHS provider and >9 patients per consultant required; multilevel modelsNumber of consultants reported that perform better than expected, as expected, and worse than expected
EQ-5D, European Quality of Life 5 Dimensions; IRES, Indikatoren des Reha-Status (Indicators for Rehabilitation Status); MDQ, Modified Low Back Pain Disability Questionnaire; MID, Minimally Important Difference; NDI, Neck Disability Index; NHS, National Health Service (UK); ODI, Oswestry Disability Index; OHS, Oxford Hip Score; OKS, Oxford Knee Score; PCS, Physical Component Summary; PF-10, Physical Functioning Subscale (of Short Form-36 questionnaire); SMFA, Short Musculoskeletal Function Assessment; VAS-BP, Visual Analog Scale for Back Pain; VAS-LP, Visual Analog Scale for Leg Pain. 1 According to Iezzoni, an eligible adjustor should 1. be associated with the outcome and 2. vary across providers [16].
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sibert, N.T.; Pfaff, H.; Breidenbach, C.; Wesselmann, S.; Kowalski, C. Different Approaches for Case-Mix Adjustment of Patient-Reported Outcomes to Compare Healthcare Providers—Methodological Results of a Systematic Review. Cancers 2021, 13, 3964. https://doi.org/10.3390/cancers13163964

AMA Style

Sibert NT, Pfaff H, Breidenbach C, Wesselmann S, Kowalski C. Different Approaches for Case-Mix Adjustment of Patient-Reported Outcomes to Compare Healthcare Providers—Methodological Results of a Systematic Review. Cancers. 2021; 13(16):3964. https://doi.org/10.3390/cancers13163964

Chicago/Turabian Style

Sibert, Nora Tabea, Holger Pfaff, Clara Breidenbach, Simone Wesselmann, and Christoph Kowalski. 2021. "Different Approaches for Case-Mix Adjustment of Patient-Reported Outcomes to Compare Healthcare Providers—Methodological Results of a Systematic Review" Cancers 13, no. 16: 3964. https://doi.org/10.3390/cancers13163964

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop