Loading [Contrib]/a11y/accessibility-menu.js
Skip to main content
Survey Practice
  • Menu
  • Articles
    • Articles
    • Editor Notes
    • In-Brief Notes
    • Interview the Expert
    • Recent Books, Papers, and Presentations
    • All
  • For Authors
  • Editorial Board
  • About
  • Issues
  • Blog
  • Subscribe
  • search

RSS Feed

Enter the URL below into your favorite RSS reader.

http://localhost:25298/feed
Articles
Vol. 8, Issue 3, 2015June 30, 2015 EDT

The Alberta Inpatient Hospital Experience Survey: Representativeness of Sample and Initial Findings

Kyle Kemp, Nancy Chan, Brandi McCormack,
nonresponse biastelephone surveyshcahpsinpatient hospital experience
https://doi.org/10.29115/SP-2015-0012
Survey Practice
Kemp, Kyle, Nancy Chan, and Brandi McCormack. 2015. “The Alberta Inpatient Hospital Experience Survey: Representativeness of Sample and Initial Findings.” Survey Practice 8 (3). https:/​/​doi.org/​10.29115/​SP-2015-0012.
Save article as...▾

View more stats

Abstract

In health survey research, it is paramount that survey respondents are representative of the general target population, thereby ensuring that the policies and program decisions supported by the underlying data are well-informed. In order to assess the representativeness of our survey respondents, we sought to compare selected demographic and clinical attributes of our inpatient hospital experience survey respondents with those of eligible nonrespondents. This retrospective analysis of cross-sectional administrative hospital data included 26,295 survey respondents, and 466,034 non-respondents. These were based on all inpatient hospital discharges that were eligible to be surveyed in the province of Alberta, Canada from April 1, 2011 to March 31, 2014. When compared with eligible nonrespondents, survey respondents had similar patterns in terms of mean age (53.8±20.0 vs. 54.4±21.3 years), sex (35.0% vs. 38.8% male), admission type (60.7% urgent in both groups), and the mean number of comorbidities (0.8±1.2 vs. 1.0±1.3). Compared to nonrespondents, survey respondents tended to be healthier, as evidenced by a shorter mean length of stay (5.4±9.4 vs.7.0±15.4 days), less need for ICU care (2.1% vs. 3.0% of cases), and being more likely to be discharged directly home (95.2% vs. 91.9% of cases). The survey sampling strategy resulted in a sample that was, in most cases, representative of the general inpatient population in our jurisdiction of approximately 4 million residents. Our findings indicate that an adequate sampling strategy may still provide a representative sample, despite a low response rate.

Introduction

Population-based health surveys are often plagued by low response rates (Asch, Jedrziewski, and Christakis 1997). Results from surveys with low response rates may be at a greater risk for nonresponse bias (Federal Judicial Center 2010; Office of Management and Budget 2006); limiting the generalization of the data among a population. Research has shown that nonrespondents can differ from respondents in terms of demographics, as well as their underlying health condition (Etter and Perneger 1997; Grotzinger, Stuart, and Ahern 1994; Macera et al. 1990; Norton et al. 1994; Richiardi, Boffetta, and Merletti 2002). Additionally, nonrespondents may have a less favorable perception of their care (Eisen and Grob 1979; Ley et al. 1976).

The relation between response rate and nonresponse bias, however, may not be as clear-cut. A 2008 meta-analysis examining 59 different surveys failed to demonstrate an association between the two. It was found that surveys with response rates from 20 percent to 70 percent had similar levels of nonresponse bias (Groves and Peytcheva 2008). This study demonstrated that response rate may not be the ideal indicator of response bias and that an adequate sampling frame may provide a truly representative sample, regardless of response rate. Benefits of such a strategy would be survey administration cost and manpower reductions, compared to higher response rates.

In Alberta, Canada, Alberta Health Services (AHS) is the sole provider of healthcare services for the province’s approximately 4 million residents. Inpatient hospital experience is one of 16 publicly-reported performance measures (Alberta Health Services 2014). The necessary data is captured by a team of trained health research interviewers, who administer a telephone survey, primarily comprised of the Hospital-Consumer Assessment of Healthcare Providers and Systems (HCAHPS) instrument (Centers for Medicare and Medicaid Services 2014b).

Since 2011, the inpatient hospital experience survey touches upon all of the province’s 94 acute care inpatient facilities. With three years of complete data, an evaluation of the representativeness of survey respondents is a timely piece that will strengthen the conclusions derived from the results. Given this, the purpose of the present project was to compare selected demographic and clinical attributes of survey respondents to those of all eligible inpatient discharges over the same time period. Organization-specific information regarding sampling methodology, survey administration, and preliminary results are also provided.

Data and Methods

Survey Instrument

Our organization’s inpatient hospital experience survey contains 51 questions. This includes 32 core HCAHPS items and 19 others which address organization-specific policies and procedures. Of the core HCAHPS items, 21 encompass nine key topics: communication with doctors, communication with nurses, responsiveness of hospital staff, pain management, communication about medicines, discharge information, cleanliness of the hospital environment, quietness of the hospital environment, and transition of care. The remaining core questions include four screener questions and seven demographic items. These are used for patient-mix adjustment and sub-analyses (Centers for Medicare and Medicaid Services 2014b). Organization-specific questions represent domains not included in HCAHPS, including pharmacy care and patient complaints. Each survey requires 10 to 20 minutes to complete using a standard script, a list of standard prompts, and responses to frequently asked questions. Surveys are administered using computer-assisted telephone interview (CATI) software (Voxco; Montreal, Canada). Ten percent of the calls are monitored for quality assurance and training purposes.

Responses to survey questions are Likert-type scales. Certain questions ask the respondent to rate aspects of their care on a scale of 0 (worst) to 10 (best), while others employ categorical responses (e.g., always; usually; sometimes; never). Details about the development, validity, and American results from HCAHPS are publicly available at www.hcahpsonline.org (Centers for Medicare and Medicaid Services 2014b, 2014a). At the end of the survey, open-ended questions provide an opportunity for respondents to give detailed feedback about their experience, including complaints that they may have. Patients wishing to report a concern, complaint, or compliment are provided with contact information for the Patient Relations department.

Sample Derivation and Dialing Protocol

Across our province, acute care admission, discharge, and transfer information is captured in four clinical databases. A biweekly data extract of eligible discharges is obtained using a standard script. Survey exclusion criteria include: age under 18 years old, inpatient stay of less than 24 hours, death during hospital stay, any psychiatric unit or physician service on record, any dilation and curettage, day surgery, or ambulatory procedures, as well as visits relating to still births or those associated with a baby with length of stay greater than 6 days (e.g., complication/NICU stay) (excluded out of consideration).

The list of eligible discharges is imported into CATI software, and stratified at the site level. Random dialing is performed, until a quota of 5 percent of eligible discharges is met at each site. Patients are contacted up to 42 days post-discharge, Monday to Friday from 10 AM to 9 PM, and on Saturdays from 9 AM to 4 PM. To increase potential for survey completion, each number is dialed up to nine times on varying days and times.

Data Linkage and Analysis

All biweekly data extracts were merged into a single file. Through cross-reference with our list of complete surveys, each eligible case was classified as a complete survey, or a nonrespondent case (e.g., indeterminate, disqualified, refused). Cases were then linked, based on personal health number (PHN), facility code, and service dates, to the corresponding inpatient discharge record in the Discharge Abstract Database (DAD) – a database of all inpatient hospital discharges. The national version of the DAD is maintained by the Canadian Institute for Health Information (CIHI), while a provincial copy is retained within our organization. Information regarding data elements, coverage, and data quality of the DAD are publicly available (Canadian Institute for Health Information 2010, 2014).

Study Variables

To assess the representativeness of survey respondents, we examined a variety of demographic (age, gender) and clinical (admission type, mean length of stay, mean number of comorbidities, ICU stay, discharge to home) variables. Admission type was classified as elective or urgent. Mean length of stay was recorded in days. A validated list of ICD-10-CA codes was used (Quan et al. 2005) to generate a comorbidity profiles for each record using the Elixhauser Comorbidity Index (Elixhauser et al. 1998). Diagnosis types “M” (most responsible diagnosis) and “2” (post-admission comorbidity) were excluded. ICU stay was classified as yes or no. Discharge to home was classified using the discharge disposition field in the DAD (codes “04” and “05”) (Canadian Institute for Health Information 2012).

Differences between inpatient experience survey respondents and nonrespondents were assessed using student t-tests for continuous variables, and chi-square analyses for binary ones. All analyses were performed using SAS Network Version 9.3 for Windows (Cary, NC, USA). P-values less than 0.05 were deemed statistically significant.

Results

Over the three-year study period (April 1, 2011 to March 31, 2014), 27,493 inpatient experience surveys were completed. Over this period, 493,527 eligible inpatient discharges took place, representing a 5.6 percent survey completion rate. Of completed surveys, 26,295 were matched with the inpatient hospital record (95.6 percent). Respondents had a mean age of 53.8±20.0 years, were predominantly female (65.0 percent), and had a mean length of stay of 5.4±9.4 days (Table 1). Compared with eligible nonrespondents (n=466,034), the sample had similar mean age (53.8±20.0 years vs. 54.4±21.3 years), sex (35.0 percent vs. 38.8 percent male), admission type (60.7 percent urgent in both groups), and mean number of comorbidities (0.8±1.2 vs. 1.0±1.3). However, compared to nonrespondents, respondents had a shorter mean length of stay (5.4 vs. 7.0 days), required less ICU care (2.1 percent vs. 3.0 percent), and were more likely to be discharged home (95.2 percent vs. 91.9 percent) (p<0.0001 in all cases).

Table 1  Completed Surveys vs. Remaining Alberta Inpatient Discharges.
Variable Complete (n=26,295) Population (n=466,034) p
Mean age in years (SD) 53.8 (20.0) 54.4 (21.3) <0.0001
Sex [n (%)]
 Male 9,207 (35.0) 180,979 (38.8)
 Female 17,088 (65.0) 285,055 (61.2) <0.0001
Admission type [n (%)]
 Urgent 15,966 (60.7) 282,832 (60.7)
 Elective 10,329 (39.3) 183,202 (39.3) 0.92
Mean length of stay in days (SD) 5.4 (9.4) 7.0 (15.4) <0.0001
Mean comorbidities (SD) 0.8 (1.2) 1.0 (1.3) <0.0001
ICU stay [n (%)]
 Yes 542 (2.1) 14,173 (3.0)
 No 25,753 (97.9) 451,861 (97.0) <0.0001
Discharged home [n (%)]
 Yes 25,039 (95.2) 428,157 (91.9)
 No 1,256 (4.8) 37,877 (8.1) <0.0001

Table 2 displays the results of survey respondents versus nonrespondents for each Elixhauser comorbidity. Twenty-three of the 30 comorbidities were more present in the nonrespondent group. The percentage of individuals with documented complicated hypertension, peptic ulcer disease excluding bleeding, AIDS/HIV, lymphoma, rheumatoid arthritis/collagen diseases, and obesity was similar between groups. Only uncomplicated diabetes was more prevalent in the survey respondent group (6.9 percent vs. 6.0 percent).

Table 2  Medical Comorbidities: Completed Surveys vs. Remaining Alberta Inpatient Discharges [n (%)].
Comorbidity present Complete (n=26,295) Population (n=466,034) p-Value
Congestive heart failure 832 (3.2) 18,479 (4.0) <0.0001
Cardiac arrythmia 1,490 (5.7) 34,133 (7.3) <0.0001
Valvular disease 313 (1.2) 7,992 (1.7) <0.0001
Pulmonary circulation disorders 183 (0.7) 4,713 (1.0) <0.0001
Peripheral vascular disorders 345 (1.3) 6,743 (1.5) <0.0001
Hypertension, uncomplicated 5,012 (19.1) 109,637 (23.5) <0.0001
Hypertension, complicated 38 (0.1) 777 (0.2) 0.39
Paralysis 107 (0.4) 3,463 (0.7) <0.0001
Other neurological disorders 389 (1.5) 11,217 (2.4) <0.0001
Chronic pulmonary disease 1,712 (6.5) 37,010 (7.9) <0.0001
Diabetes, uncomplicated 1,818 (6.9) 27,946 (6.0) <0.0001
Diabetes, complicated 1,937 (7.4) 39,013 (8.4) <0.0001
Hypothyroidism 929 (3.5) 19,453 (4.2) <0.0001
Renal failure 556 (2.1) 13,140 (2.8) <0.0001
Liver disease 310 (1.2) 8,800 (1.9) <0.0001
Peptic ulcer disease excluding bleeding 86 (0.3) 1,640 (0.4) 0.50
AIDS/HIV 3 (0.1) 153 (0.1) 0.06
Lymphoma 109 (0.4) 2,138 (0.5) 0.30
Metastatic cancer 491 (1.9) 12,347 (2.7) <0.0001
Solid tumor without metastasis 728 (2.8) 14,951 (3.2) <0.0001
Rheumatoid arthritis/collagen diseases 322 (1.2) 5,639 (1.2) 0.83
Coagulopathy 188 (0.7) 4,557 (1.0) <0.0001
Obesity 1,068 (4.1) 18,565 (4.0) 0.53
Weight loss 113 (0.4) 3,327 (0.7) <0.0001
Fluid and electrolyte disorders 1,102 (4.2) 23,038 (4.9) <0.0001
Blood loss anemia 64 (0.2) 1,742 (0.4) <0.0001
Deficiency anemia 199 (0.8) 5,487 (1.2) <0.0001
Alcohol abuse 488 (1.9) 15,033 (3.2) <0.0001
Drug abuse 161 (0.6) 6,136 (1.3) <0.0001
Psychoses 58 (0.2) 2,102 (0.5) <0.0001
Depression 780 (3.0) 16,293 (3.5) <0.0001

Discussion

Our main finding was that inpatient experience survey respondents were similar in age and sex to the eligible nonresponders. The present study is novel in that it sheds new light on the relation between response rate and corresponding nonresponse bias in health survey research. To our knowledge, it is the first report which examines this in the Canadian provincial context; one where healthcare services are universally provided. Perhaps more importantly, our findings may dispel the myth that a low response rate will, by default, result in nonresponse bias. Although our analyses resulted in statistical significance in the majority of comparisons, we feel that this is more a product of our extremely large sample size (over 26,000 surveys) and not any clinically meaningful difference between respondents and nonrespondents. We observed that survey respondents may be marginally healthier than nonrespondents, as shown in the mild reduction in mean length of stay, ICU stays, and mean number of documented comorbidities, as well as the increased proportion of patients discharged home.

There are several key strengths to the present study. First, the present project uses data linkage to compare several demographic and clinical factors of our inpatient experience respondents to nonrespondents. Lee et al. (2009) cite the absence of data from nonrespondents as a major difficulty in examining nonresponse bias in health survey research (Lee et al. 2009). Our data contains discharge information of all eligible patients; hence, we are able to make direct comparisons between respondents and nonrespondents, overcoming this critical limitation. This greater availability of data and data linkage provides opportunities for future research.

Second, as we have used HCAHPS methodology, a validated tool with standard script and prompts has assessed inpatient experience. Traditionally, patient satisfaction/experience measurement has been via instruments developed on an ad hoc basis. These instruments may not be valid or reliable. Waljee et al. (2014) shared the findings of 36 studies which examined the relationship between patient expectations and satisfaction (Waljee et al. 2014). Of these 36, the majority used ad hoc questionnaires and none used the HCAHPS survey. One of the inherent strengths of HCAHPS is that valid, measurable comparisons may be made between institutions and jurisdictions. In most cases, this is not possible with ad hoc questionnaires. Given the heterogeneity of clinical populations between institutions, research has examined the effects of patient-mix upon HCAHPS scores (Centers for Medicare and Medicaid Services 2014c). More information regarding patient-mix adjustment is available for consultation (Centers for Medicare and Medicaid Services 2014b).

Third, perhaps most important, is our sampling strategy. Given that we obtain all eligible inpatient discharges, each potential participant has an equal chance of participation. Our abstracted data includes up to two telephone numbers provided at hospital registration. These contact numbers do not discriminate between landlines or cellular phones and are presumed to be the most accurate way of contacting patients. Additionally, our interviewers attempt to call patients up to nine times at varying times on varying days, including one weekend day when one would presume most people are available. Patients who are not able to speak freely are provided with the opportunity to book a convenient callback time. Time is set aside each day for interviewers to complete callbacks in order to reduce nonresponse (Goyder 1985; Heberlein and Baumgartner 1978). Anecdotally, these strategies have helped ensure that survey quotas are met. Further, the sample is stratified for each of the 94 inpatient facilities, ensuring that each has an equal probability of representation within the final data set. This quota sampling approach has been applied elsewhere, with similar success (O’Cathain, Knowles, and Nicholl 2010).

There are limitations which warrant discussion. First, despite having obtained a fairly representative sample, it is impossible to assess the actual responses that nonrespondents may have had. This is important, as research has shown that survey respondents tend to have more favorable opinions of the care received, when compared to nonrespondent counterparts (Eisen and Grob 1979; Ley et al. 1976; French 1981). Despite this, there may not be a need to define an acceptable a priori response rate, provided that potential differences between survey respondents and nonrespondents are assessed (Kelley et al. 2003). Our findings support this assumption.

Second, our telephone administration, results may not apply to other modalities such as mail or face-to-face administration. An organizational pilot study (performed in 2004) highlighted differences in terms of response rates and demographics of survey respondents between the mail and telephone surveys (Cooke et al. 2004). With respect to HCAHPS specifically, de Vries et al. (2005) found that telephone administration elicited more positive responses on more than half of the survey items, particularly among domains relating to nursing care and the physical environment of the hospital (de Vries et al. 2005). These findings are consistent with other previous health survey studies (Burroughs et al. 2001; Fowler, Roman, and Di 1998; Fowler, Gallagher, and Nederend 1999).

A third potential limitation is the use of an administrative comorbidity algorithm. As outlined by Quan et al. (2005) the specificity and sensitivity of these coding algorithms, relative to a gold standard (e.g., chart review) remain undetermined (Quan et al. 2005).

In conclusion, this investigation provides novel information relative to the use of an HCAHPS-derived inpatient experience survey within our organization. This represents a key piece regarding the validity of the conclusions supported by the data. We advise that our sampling strategy may result in a representative sample, despite a 5 percent survey completion rate. This is an important finding, as further capital and manpower investments may not be necessary to bolster response rates; activities which themselves may not provide any measureable benefit. Future research will examine our survey methodology in greater detail.

Acknowledgements

The authors wish to recognize and thank the team of health research interviewers from Primary Data Support, Analytics (Data Integration, Measurement and Reporting, Alberta Health Services), as well as the patients who participated in the survey.

References

Alberta Health Services. 2014. “Strategic Measures: Report on Performance.” 2014. http:/​/​www.albertahealthservices.ca/​performance.asp.
Asch, D.A., M.K. Jedrziewski, and N.A. Christakis. 1997. “Response Rates to Mail Surveys Published in Medical Journals.” Journal of Clinical Epidemiology 50 (10): 1129–36.
Google Scholar
Burroughs, T.E., B.M. Waterman, J.C. Cira, R. Desikan, and W. Claiborne-Dunagan. 2001. “Patient Satisfaction Measurement Strategies: A Comparison of Phone and Mail Methods.” The Joint Commission Journal on Quality Improvement 27 (7): 349–61.
Google Scholar
Canadian Institute for Health Information. 2010. “CIHI Data Quality Study of the 2009-2010 Discharge Abstract Database.” 2010. https:/​/​secure.cihi.ca/​free_products/​Reabstraction_june19revised_09_10_en.pdf.
———. 2012. DAD Abstracting Manual: 2012. 13th ed. Ottawa, Ontario, Canada: Canadian Institute for Health Information.
Google Scholar
———. 2014. “Discharge Abstract Database (DAD) Metadata.” 2014. http:/​/​www.cihi.ca/​cihi-ext-portal/​internet/​en/​document/​types+of+care/​hospital+care/​acute+care/​dad\_metadata.
Centers for Medicare and Medicaid Services. 2014a. “CAHPS® Hospital Survey.” 2014. http:/​/​www.hcahpsonline.org.
———. 2014b. “HCAHPS Fact Sheet.” 2014. http:/​/​www.hcahpsonline.com/​files/​Augustpercent202013percent20HCAHPSpercent20Factpercent20Sheet2.pdf.
———. 2014c. “Hospital Compare Webpage.” 2014. http:/​/​www.medicare.gov/​hospitalcompare/​search.html.
Cooke, T., M. Liu, R.D. Hays, M. Elliott, K. Hepner, and K. Edwards. 2004. “HCAHPS Pilot Study: January - March 2004.” Calgary Health Region. Unpublished internal communication.
Eisen, S.V., and M.C. Grob. 1979. “Assessing Consumer Satisfaction from Letters to the Hospital.” Hospital & Community Psychiatry 30 (5): 344–47.
Google Scholar
Elixhauser, A., C. Steiner, D.R. Harris, and R.M. Coffey. 1998. “Comorbidity Measures for Use with Administrative Data.” Medical Care 36 (1): 8–27.
Google Scholar
Etter, J.F., and T.V. Perneger. 1997. “Analysis of Non-Response Bias in a Mailed Health Survey.” Journal of Clinical Epidemiology 50 (10): 1123–28.
Google Scholar
Federal Judicial Center. 2010. Reference Manual on Scientific Evidence. 3rd ed. http:/​/​www.fjc.gov/​public/​pdf.nsf/​lookup/​SciMan3D01.pdf/​$file/​SciMan3D01.pdf.
Google Scholar
Fowler, F.J. Jr., P.M. Gallagher, and S. Nederend. 1999. “Comparing Telephone and Mail Responses to the CAHPSTM Survey Instrument.” Medical Care 37 (3 Suppl): MS41–49.
Google Scholar
Fowler, F.J. Jr., A.M. Roman, and Z.X. Di. 1998. “Mode Effects in a Survey of Medicare Prostate Surgery Patients.” Public Opinion Quarterly 62 (1): 29–46.
Google Scholar
French, K. 1981. “Methodological Considerations in Hospital Patient Opinion Surveys.” International Journal of Nursing Studies 18 (1): 7–32.
Google Scholar
Goyder, J. 1985. “Face-to-Face Interviews and Mail Questionnaires: The Net Difference in Response Rate.” Public Opinion Quarterly 49 (2): 234–52.
Google Scholar
Grotzinger, K.M., B.C. Stuart, and F. Ahern. 1994. “Assessment and Control of Nonresponse Bias in a Survey of Medicine Use by the Elderly.” Medical Care 32 (10): 989–1003.
Google Scholar
Groves, R.M., and E. Peytcheva. 2008. “The Impact of Response Rates on Nonresponse Bias.” Public Opinion Quarterly 72 (2): 167–89.
Google Scholar
Heberlein, T., and R. Baumgartner. 1978. “Factors Affecting Response Rates to Mailed Questionnaires: A Quantitative Analysis of the Published Literature.” American Sociological Review 43 (4): 447–62.
Google Scholar
Kelley, K., B. Clark, V. Brown, and J. Sitzia. 2003. “Good Practice in the Conduct and Reporting of Survey Research.” International Journal for Quality in Health Care 15 (3): 261–66.
Google Scholar
Lee, S., E.R. Brown, D. Grant, T.R. Belin, and J.M. Brick. 2009. “Exploring Nonresponse Bias in a Health Survey Using Neighborhood Characteristics.” American Journal of Public Health 99 (10): 1811–17.
Google Scholar
Ley, P., P.W. Bradshaw, J.A. Kinsey, and S.T. Atherton. 1976. “Increasing Patients’ Satisfaction with Communications.” The British Journal of Social and Clinical Psychology 15 (4): 403–13.
Google Scholar
Macera, C.A., K.L. Jackson, D.R. Davis, J.J. Kronenfeld, and S.N. Blair. 1990. “Patterns of Non-Response to a Mail Survey.” Journal of Clinical Epidemiology 43 (12): 1427–30. https:/​/​doi.org/​10.1016/​0895-4356(90)90112-3.
Google Scholar
Norton, M.C., J.C. Breitner, K.A. Welsch, and B.W. Wyse. 1994. “Characteristics of Nonresponders in a Community Survey of the Elderly.” Journal of the American Geriatric Society 42 (12): 1252–56.
Google Scholar
O’Cathain, A., E. Knowles, and J. Nicholl. 2010. “Testing Survey Methodology to Measure Patients’ Experiences and Views of the Emergency and Urgent Care System: Telephone versus Postal Survey.” BMC Medical Research Methodology 10:52–61.
Google Scholar
Office of Management and Budget. 2006. “Guidance on Agency Survey and Statistical Information Collections: Questions and Answers When Designing Surveys for Information Collections.” 2006. http:/​/​www.whitehouse.gov/​omb/​inforeg/​pmc\_survey\_guidance\_2006.pdf.
Quan, H., V. Sundararajan, P. Halfon, A. Fong, B. Burnand, J.C. Luthi, and W.A. Ghali. 2005. “Coding Algorithms for Defining Comorbidities in ICD-9-CM and ICD-10 Administrative Data.” Medical Care 43 (11): 1130–39.
Google Scholar
Richiardi, L., P. Boffetta, and F. Merletti. 2002. “Analysis of Nonresponse Bias in a Population-Based Case-Control Study on Lung Cancer.” Journal of Clinical Epidemiology 55 (10): 1033–40.
Google Scholar
Vries, H., M.N. Elliott, K.A. Hepner, S.D. Keller, and R.D. Hays. 2005. “Equivalence of Mail and Telephone Responses to the CAHPS® Hospital Survey.” Health Services Research 40 (6 Pt 2): 2120–39.
Google Scholar
Waljee, J., E.P. McGlinn, E. Sears, and K.C. Chung. 2014. “Patient Expectations and Patient-Reported Outcomes in Surgery: A Systematic Review.” Surgery 155 (5): 799–808.
Google Scholar

This website uses cookies

We use cookies to enhance your experience and support COUNTER Metrics for transparent reporting of readership statistics. Cookie data is not sold to third parties or used for marketing purposes.

Powered by Scholastica, the modern academic journal management system