Does Mandatory Measurement and Peer Reporting Improve Performance?
Susanna Gallani Takehisa Kajiwara Ranjani Krishnan
Working Paper 16-018
Working Paper 16-018
Copyright © 2015, 2017 by Susanna Gallani, Takehisa Kajiwara, and Ranjani Krishnan
Working papers are in draft form. This working paper is distributed for purposes of comment and discussion only. It may not be reproduced without permission of the copyright holder. Copies of working papers are available from the author.
Does Mandatory Measurement and Peer Reporting Improve Performance?
Susanna Gallani Harvard Business School
Takehisa Kajiwara Kobe University - Japan
Ranjani Krishnan Michigan State University
Does Mandatory Measurement and Peer Reporting Improve Performance?
SUSANNA GALLANI Harvard Business School
369 Morgan Hall 15 Harvard Way, Boston MA 02163
Ph: 617-496-8613 Fax: 617-496-7363 [email protected]
TAKEHISA KAJIWARA
Kobe University Graduate School of Business Administration
2-1 Rokkodai, Nada-ku, Kobe, 657-8501, Japan Ph: +81-78-803-6988 Fax: +81-78-803-6988
RANJANI KRISHNAN Michigan State University
N207 North Business College Complex The Eli Broad College of Business
632 Bogue St., East Lansing, MI 4884 Ph: 517-353-4687 Fax: 517-432-1101
Acknowledgments: We thank for their valuable comments and suggestions Jeff Biddle, Clara Chen, Leslie Eldenburg, Matthias Mahlendorf, Melissa Martin, Pam Murphy, Steve Salterio, Greg Sabin, Daniel Thornton, Stephanie Tsui, Jeff Wooldridge, workshop participants at the 2015 “Patient-Centric Healthcare Management in the Age of Analytics” conference, University of Arizona, Erasmus University, Michigan State University, Queen’s School of Business, and Wilfrid Laurier University. We appreciate the help we received in gathering and interpreting information about the Japanese healthcare industry from Nobuo Sato and Mayuka Yamazaki at the Harvard Business School Research Center in Tokyo, and for the precious insights on the institutional settings shared with us by Kanoko Oishi. We thank Kenji Yasukata, Yoshinobu Shima and Chiyuki Kurisu for their support in the collection of the data used in this study.
2
Does Mandatory Measurement and Peer Reporting Improve Performance?
ABSTRACT We examine the effect of mandated measurement and peer disclosure of new information on the persistence of performance improvements in a setting without performance incentives. Value of information (VOI) theory posits that information can improve the accuracy of posterior beliefs and thereby have a decision facilitating effect. These effects are more pronounced when the information is new versus an update. Using data from the Japanese National Hospital Organization, we analyze performance trends following regulation requiring standardized measurement and peer disclosure of absolute and relative patient satisfaction performance. After controlling for ceiling effects and regression to the mean, mandatory patient satisfaction measurement and peer disclosure introduce positive and significant mean shifts in performance with larger improvements for poorly performing hospitals. The largest positive effects occur when the information is new. Our study provides empirical evidence of the decision facilitating value of information without confound from its decision influencing value. Keywords: Value of information, Patient Satisfaction, Mandatory performance measurement, Healthcare. JEL codes: I10, L30, M14, M41 Data Availability: Data used in this study can be obtained from the Japanese National Hospital Organization.
3
Does Mandatory Measurement and Peer Reporting Improve Performance?
1. Introduction
There is little debate that accounting information is useful both for belief revision and
performance evaluation (Narayanan and Davila [1998], Baiman and Demski [1980]). The belief
revision (or “decision facilitating”) role refers to the use of information to improve the manager’s
action choice, in that it allows for improved mapping between actions and outcomes. The
performance evaluation (or “decision influencing”) role refers to the use of information for
managerial compensation to motivate managerial performance, exercise control, and facilitate
risk sharing by improving the mapping of outcomes to rewards. Analytical and practitioner
research posit that, while an information signal has value in each of these roles independent of its
value in the other, the decision influencing use of a signal could encourage managerial
manipulation and reduce its decision facilitating value (Kaplan and Norton [1992], Narayanan
and Davila [1998], Simons [2013]). Moreover, while the decision influencing use of an
information signal automatically makes it available for decision facilitating purposes, the reverse
is not true. That is, a signal could have decision facilitating role but not a decision influencing
role if it is not included in performance measurement. Data limitations stymie archival study of
the value of information in each of these roles independently; consequently the bulk of archival
research has focused on the decision influencing role of information. There is a dearth of archival
research that examines the decision facilitating role, absent any confound from the decision
influencing role. Our study seeks to address this lacuna and aims to examine whether new
nonfinancial performance information has a decision facilitating value in a setting bereft of
explicit pecuniary benefits or penalties.
4
Our empirical setting uses health care quality data. Improving healthcare outcomes,
quality, and cost are topics of high policy interest in countries around the world. Systematic
collection and disclosure of reliable, consistent, and comparable healthcare outcomes
information is a popular policy tool that has been proposed in many countries. For example, in
the U.S., the Center for Medicare and Medicaid Services (CMS) and the Agency for Health Care
Research and Quality (AHRQ) developed a patient satisfaction survey titled Hospital Consumer
Assessment of Healthcare Providers and Systems (HCAHPS) with an aim to produce and
publicly disseminate “comparable data on quality outcomes that allow objective and meaningful
comparisons between hospitals on domains that are important to consumers”, to “create
incentives for hospitals to improve their quality of care”, and “to enhance public accountability
in healthcare by increasing the transparency of the quality of hospital care provided in return for
the public investment” (http://www.hcahpsonline.org/home.aspx).1
Such regulations generate information that was previously unavailable to the decision
maker. The effectiveness of quality disclosure regulations on actual quality improvements
depends, in large part, on the extent to which the information contained in these disclosures is
useful to decision makers within the organizations subject to those provisions. Considerable
research has examined the effectiveness of public quality reporting on health quality outcomes
with equivocal results. Some studies find improvements in medical outcomes, such as hospital
mortality following public disclosures, while others find no effects (Fung et al. [2008], Lamb et
al. [2013]). One factor driving the mixed results is that public reporting can lead to a variety of
hospital dynamics that are unrelated to actual quality changes. The requirement for public
disclosure may cloud the value of the new information for improved internal decision-making. 1 Since 2007, U.S. hospitals that admit Medicare patients must collect and submit HCAHPS data to receive their full payment.
5
Decision makers may become so highly focused on the public’s response to quality information
that they could exert more effort to manipulate reported quality performance rather than focus on
the actual drivers of performance (Dranove and Jin [2010]). Some of these public reporting
settings confound disclosure with performance evaluation by including compensation-based
incentives for quality performance, which impose pressures on hospital managers to manage
reported quality performance. Mandatory requirements for new information that are bundled
with public disclosure and incentive compensation do not allow to decouple the assessment of
the value of the new information for the decision maker from that of the rewards or penalties
arising from the public disclosure. We use value of information theory to examine the decision
facilitating role of mandatory nonfinancial information disclosure in a setting where the decision
influencing role is not present.
Economists and decision theorists use Bayesian methods to define the value of new
information (VOI) as the difference between the expected utility of an action based on the
posterior probability given a new information set, and the expected utility of the action given
only the prior information set (Pratt et al. [1995]). However, outside of a laboratory, one rarely
encounters situations where access to new information is not confounded with other factors that
influence the decision maker’s use of the new information. In this study, we examine the value of
new information using a unique, quasi-experimental, Japanese hospital setting where new
regulation requiring standardized measurement and peer disclosure of patient satisfaction was
introduced. The new information (1) was not previously collected by the hospital or any other
agencies, (2) was not tied to incentive compensation or other pecuniary payoffs, and (3) could
not be gamed by individual managers to improve reported but not actual performance. We study
two effects of new information: an information effect that arises from the value of the new
6
information about the firm’s own performance, and a referent performance effect that arises from
the value of the new information signal about the firm’s performance relative to the peer group.
Our empirical setting is the Japanese hospital industry, which introduced regulation in
2004 that requires hospitals belonging to the Japanese National Health Organization (NHO) to
collect patient satisfaction performance data using a standardized survey. A neutral external
agency surveys inpatients and outpatients about their satisfaction with a number of aspects of
their hospital experience, including medical treatments and procedures, physician and staff
behavior and attitudes, and hospital infrastructure. The results of the survey containing
performance information on the level as well as relative rank of individual member hospitals are
disseminated to all hospitals within the NHO.
We analyze patient satisfaction panel data from all 145 NHO-member hospitals over a
period of eight years (2004-2011). We first conduct a factor analysis of the survey responses and
identify main inpatient (outpatient) satisfaction constructs, which we label as staff /treatment,
logistics/infrastructure (staff /treatment, administrative procedures, and location). We then
examine whether the patient satisfaction information results in an improvement in performance
on each construct in subsequent years, and whether there are differences in the extent of
improvement based on initial relative performance. Our analyses control for ceiling effects, i.e.,
hospitals with a lower performance baseline have more opportunities to improve than better
performing hospitals, and for regression to the mean, i.e., hospitals that have low (high)
performance on patient satisfaction in a particular year could have high (low) performance in
future years simply because of the nature of the distribution rather than actual changes in
performance (Cook and Campbell [1979]).
7
Results indicate that mandatory measurement of patient satisfaction has an information
effect. Controlling for hospital-fixed effects and for the existence of a pre-existing trend, our
analyses indicate an overall increase in both inpatient and outpatient satisfaction following the
mandatory measurement. We also find that the performance improvement is of larger magnitude
in the year immediately following the first release of the survey results, when the information is
new, compared to subsequent years. The improvement is found for all hospitals and not merely
for the poorly performing hospitals (defined as hospitals in the lowest quartile of performance in
2004). This evidence supports our hypothesis that patient satisfaction information has decision
facilitating value. We infer that information has highest value when it is new, whereby
subsequent improvements to performance follow a path of diminishing returns. We also find
evidence of the referent performance effect - poorly performing hospitals have greater
improvements in performance following the release. We conclude that the new information
about relative performance facilitates improved goal-directed effort.
Our study contributes to the literature in several ways. First, our unique, quasi-
experimental design enables the assessment of the value of new information, which to our
knowledge has not been studied in an archival health care setting insofar. Second, we study the
value of nonfinancial information in the absence of confounding factors such as incentive
compensation or performance pressure arising from public disclosures. That is, our setting
isolates the decision facilitating role of information without any confound from the decision
influencing role. Prior research finds that public disclosures of health care quality can motivate
hospitals and providers to improve quality. For example, Evans et al. [1997] find that mandated
public disclosures of hospital mortality performance lead to subsequent improvements in
mortality for hospitals that were performing poorly during the initial period. Lamb, et al. [2013]
8
find that voluntary reporting of quality of ambulatory care motivated physician groups to
improve quality. In both these aforementioned studies, the hospitals or physicians were already
collecting the quality information and the only change was to disclose it to the public. That is,
the information was not new to the decision maker but only the disclosure was new. Another
point of departure is that these studies focus on medical quality rather than satisfaction quality.
Improvements in medical technology could mistakenly produce results in favor of the hypothesis
that information improves quality. Third, our empirical method attempts to disentangle the value
of information about individual performance and the value of information about relative
performance. Fourth, research in accounting has explored the importance of nonfinancial
measures, such as satisfaction, in driving future financial performance (e.g. Ittner and Larcker
[1998], Nagar and Rajan [2005]). These studies have stressed the importance of nonfinancial
information due to its role as a leading indicator of future financial performance, and its use in
managerial compensation (Banker et al. [2000]) rather than the value of such information in
improving decision making. Our evidence suggests that organizations would benefit from
designing planning and feedback systems where decision makers obtain information that is
disassociated from evaluation or rewards. Finally, our study has policy implications. The Patient
Protection and Affordable Care Act of 2010 requires that hospitals collect and report patient
satisfaction, which increases the pressure on hospitals to improve their quality of care. Our
results indicate that such regulation requiring mandatory collection of patient satisfaction
information has the potential to improve hospital performance on the reported measures, even in
the absence of incentives.
2. Prior literature and hypotheses development
2.1 INSTITUTIONAL BACKGROUND OF JAPANESE NHO
9
The Japanese National Health Organization (NHO) is the oversight agency for Japanese
government hospitals. Headquartered in Tokyo, it comprises of 145 hospitals, which represent
3.5% of the total number of hospitals in Japan. Like other Japanese Independent Administrative
Institutions (IAI), the NHO is the result of the separation between political and operational
responsibilities for public services. NHO hospitals are grouped into two categories: general
hospitals, which have discretion on the type of service they provide, and sanatoriums, which, in
addition to offering services similar to general hospitals, are required to supply particularly
expensive and risky medical services. Both types of hospitals provide inpatient as well as
outpatient treatment.2 Funding for NHO hospitals comes from three sources: patients’
copayments for service rendered, reimbursements by the National Health Insurance or
Employees’ Health Insurance, and public funding through government grants and subsidies.3
Patient copayments are received directly by the hospital, while insurance reimbursements are
received through a claims process. Public funding allocation to each hospital within the NHO is
dependent upon periodic performance evaluations of medical outcomes (e.g. mortality and
morbidity) and assessment of the hospital’s need for resources. Prices for healthcare services are
determined nationwide by the Japanese government. Price is, therefore, not a driver of patients’
choice of healthcare provider.
In 2004 the NHO introduced an annual patient satisfaction survey for every hospital
within the NHO. Patients treated in NHO hospitals are required to complete a standardized
2 Government hospitals in Japan are regarded as high quality providers of healthcare services, and tend to be preferred to private hospitals by many patients. 3 Health Insurance in Japan is compulsory for all citizens and can be obtained either through the employer (Employees’ Health Insurance) or, in the case of self-employed individuals and students, through the National Health Insurance system. Special insurance programs are in place for elderly citizens (over 75 years). Patients pay about for 30% of the cost of medical services, with the remaining 70% being reimbursed to the hospital by the insurer or the government. Medical costs exceeding the equivalent of $600 in a month are fully reimbursed by the insurer or the government. Other than minor cost of living adjustments, these numbers have been steady since the year 2000.
10
questionnaire, which assesses their satisfaction with many aspects of their hospital experience,
including medical treatment, the behavior of the staff, and the quality of the infrastructure. The
survey is conducted by an independent university research agency, which is unrelated to the
NHO or its hospitals. The university research agency compiles and analyzes the results of the
annual survey, and feedback reports are disseminated to all member hospitals. The feedback
reports include the average scores of each hospital on the various categories, and each hospital’s
ranking within the NHO. There is no explicit monetary incentive tied to patient satisfaction
performance.4
Consistent with standard models of health economics beginning with the seminal
research of Arrow [1963] and the models of Kolstad [2013] we assume that health care providers
derive utility from non-financial outcomes such as patient satisfaction, independent of monetary
compensation. We further partition the utility from patient satisfaction into two parts: one that
arises from the hospital’s absolute performance level, and the other that arises from performance
relative to peers. We then test whether the new information provided by the patient satisfaction
survey has value, in that it affects the choice of actions leading to improvements in subsequent
performance.
2.2 INFORMATION EFFECT OF PATIENT SATISFACTION INFORMATION
Adopting an expected value of information framework (Bromwich [2007], Demski
[1972]), suppose a decision maker is considering an action choice from a vector of potential
actions given by A={a}, and a potential set of uncertain states S={s}. The utility of a particular
action is U(s, a), where Θ 𝑠 is the probability specification based on the decision maker’s
likelihood judgment. The decision maker chooses the action that maximizes expected utility, that 4 Physicians and medical staff at the NHO are compensated on a fixed wage basis and are not provided performance-contingent bonuses. Physicians and staff obtain raises based on general macro-economic conditions. Section 4 examines physician compensation at NHO hospitals in greater detail.
11
is 𝐸 𝑈 𝑎∗ = 𝑚𝑎𝑥!∈! 𝑈(𝑠,𝑎)Θ(𝑠)𝑑𝑠! . The relationship between the action a and the expected
outcome is based on subjective probability distributions related to past information (Feltham
[1968]). Suppose the decision maker obtains an additional information signal y from an
information system ƞ. The information signal y allows for an improved assessment of the state
and appropriate action choices, and it has value if it affects the decision that would be made
without the signal, thus leading to greater utility. The expected utility including the new
information signal is 𝐸 𝑈 𝑦, ƞ,𝑎!∗ = 𝑚𝑎𝑥!∈! 𝑈(𝑠,𝑎)Θ 𝑠 𝑦, ƞ! where Θ 𝑠 𝑦, ƞ is the
revised probability distribution after the receipt of the new information signal y. The expected
value of signal y is the difference between 𝐸 𝑈 𝑦, ƞ,𝑎!∗ and 𝐸 𝑈 𝑎∗ . 5 If this expected value is
positive, then the new information has value to the decision maker.
We first assess whether the patient satisfaction measure provides new information. NHO
sources indicated that before 2004, neither the NHO nor individual member hospitals were
collecting information on patient satisfaction. Without this information, hospitals only had noisy
priors about their own performance. Because of the tendency for individuals to be overconfident
about their ability and overestimate their effort levels (Benoit and Dubra [2011], Camerer and
Lovallo [1999], Kruger and Dunning [1999], Kruger and Dunning [2002]), hospitals likely
concluded that their patient satisfaction performance was (a) on the higher end of an absolute
scale, and (b) above average on a relative scale. Additionally, hospitals were unable to accurately
assess the payoffs from their efforts to increase patient satisfaction. Thus, on average, hospitals
are less likely to have engaged in targeted strategies aimed at increasing patient satisfaction.
With the introduction of the mandatory patient satisfaction survey, hospitals received an
5 Although VOI is sometimes interpreted rather narrowly as the amount a decision maker would be willing to pay for higher quality information, analytical VOI models are generic and refer to “value” in a flexible sense that allows for non-pecuniary interpretations (Bromwich [2007], Demski [1972], [Raiffa 1968]).
12
additional information system (corresponding to ƞ referred to earlier). ƞ contains two new signals
- absolute patient satisfaction level (y1) and patient satisfaction relative to peer hospitals (y2).
These two new signals are based on data collected systematically by an independent university-
based research center using a scientific, standardized, survey instrument. The new signals enable
hospitals to revise their priors and obtain more accurate posterior beliefs about the relation
between their actions and patient satisfaction, which influence future effort allocations and
decisions, i.e. to move from utility function 𝐸 𝑈 𝑎∗ to 𝐸 𝑈 𝑦, ƞ,𝑎!∗ .6 The information on
individual hospital performance level (y1) provides a more precise signal of performance, which
allows for guided effort that is better suited to the circumstances (Bandura and Jourden [1991],
Ederer [2010], Morris and Shin [2002]). The positive weight associated with patient satisfaction
within the healthcare provider’s utility function drives the extent to which the information is
internalized and used in the subsequent decision making process. This leads to the following
prediction about the information effect of nonfinancial performance information:
H1. Introduction of mandatory measurement of patient satisfaction improves subsequent
performance.
2.3 REFERENT PERFORMANCE EFFECT OF PATIENT SATISFACTION INFORMATION
The second signal contained in the new information system ƞ is patient satisfaction
relative to peer hospitals (y2). Economic theory recognizes referent performance as an important
driver of individuals’ and firms’ utility functions (Kolstad [2013], Sugden [2003]).7 Referent
performance is particularly important for strategically interdependent competitive organizations
6 This assumes that hospitals are Bayesian, i.e., they use new information to update their prior beliefs, which is a standard assumption in decision theory (Pratt, et al. [1995]). 7 In a reference-dependent utility model, preferences between decisions are influenced both by the expected outcome of the decision and by a reference point, which could be performance of a peer or competitor (Sugden [2003]).
13
(Lant and Hewlin [2002]). If a valuable relative performance signal indicates poorer
performance relative to a referent group, it prompts organizations to increase effort as well as
search for new strategies that can enhance relative performance (Bandura and Jourden [1991]),
especially if decision makers have flexibility to respond to the new information (Abernethy and
Bouwens [2005]). Organizational theory posits that decision makers pay more attention to
activities that fail to meet targets compared to those that succeed (Levitt and March [1988]).
Evidence indicates that poor relative performance is a higher motivator of performance than
good performance. For example, Ramanarayanan and Snyder [2012] find that information
disclosure in the dialysis industry is associated with reduction in mortality for poorly performing
organizations, but do not find comparable effects for highly performing ones. Casas-Arce and
Martinez-Jerez [2009] find analytical and field evidence in a retail setting that introduction of a
relative performance evaluation system (i.e., a contest) induces higher effort. They find that
participants who are in leading positions obtain confidence in their performance and reduce their
effort. Trailing participants exert additional effort to catch up with the others.
The relative performance signal y2 generated by the new information system ƞ eliminates
idiosyncratic uncertainty creating a level field to assess performance. The new information signal
y2 increases the accuracy of the posterior belief function about the mapping between effort and
output relative to the organization’s peer group. The noise reduction value of relative feedback is
higher for poorly performing hospitals because they likely expected to be above average in the
pre-regulatory period, and therefore the relative information represents an unpleasant surprise.8
This serves as a motivation for poorly performing hospitals to increase effort to improve
8 Prior literature finds that in the absence of information, individuals and firms tend to hold optimistic beliefs about their ability and therefore are overconfident about their performance relative to competitors (Kahnemann et al. [1982]).
14
performance (Ederer [2010]). Based on the above, we predict that firms with lower initial
relative performance on patient satisfaction will have greater future performance improvements.
H2. Lower initial performance on patient satisfaction measures is associated with higher
magnitude of subsequent improvements.
2.4 VALUE OF FIRST INFORMATION VERSUS SUBSEQUENT INFORMATION
In this section we explore whether information has greater value when it is first introduced,
versus when it represents an update from the previous period. VOI theory posits that value of
information depends on the prior distribution that decision makers use to represent the current
situation (Raiffa [1968]). Before the availability of any standardized information about own and
peer performance, decision makers are likely to (a) systematically over-estimate own
performance, (b) exhibit overconfidence about own performance, and (c) produce probability
distributions that are far too tight than the actual distributions (Alpert and Raiffa [1982],
Hammitt and Shlyakhter [1999], Lichtenstein et al. [1982], Morgan et al. [1992]). This implies
that before the first release of information, the expected value of the performance signal is less
likely to be representative of the true value of the signal, whereas after the release of the first
information, subsequent information is assessed relative to a well-calibrated distribution of the
signal. Subsequent information would cause decision makers to update their beliefs in a
Bayesian fashion (Kolstad [2013]) and these subsequent updates would be smaller in magnitude
relative to their initial updates when they went from a no-information regime to a quality
information regime. In sum, the first disclosure of information causes managers to gain more
information about their own performance and relative performance, which allows them to
improve significantly more. This leads to the following hypotheses:
15
H3. Mandatory patient satisfaction measurement drives greater improvements when the
performance information is first released compared to subsequent releases.
H4. Mandatory relative patient satisfaction measurement drives greater improvements when the
performance information is first released compared to subsequent releases.
3. Data and analyses
3.1 DATA AND DESCRIPTIVE STATISTICS
The sample includes data from the entire population of 145 NHO hospitals for the period
2004 to 2011. The standardized patient satisfaction survey is administered every year during the
months of June and July in all NHO hospitals. There are two types of NHO hospitals - general
hospitals and sanatoriums. General hospitals are similar to private hospitals and are allowed
greater discretion in the choice of healthcare services they offer. Sanatoriums are expected to
provide not only the services that are provided by general hospitals, but, in addition, they are
required to provide special services “that cannot be dealt with properly by Non-National Hospital
Organizations due to historical and social reasons.”9 These include treatment of expensive, long-
term, risky, or communicable ailments such as tuberculosis, AIDS, Alzheimer’s, ALS, complex
mental illnesses, and invasive or terminal cancer. Table 1 reports descriptive statistics about the
population of hospitals.
----- Insert Table 1 here -----
Separate surveys are administered to inpatients and outpatients and contain 15 questions
(outpatients) and 19 questions (inpatients) respectively (Appendix A). The surveys contain
multiple items related to quality of medical treatment, behavior of the staff, quality of
infrastructure and facilities, waiting periods, etc. Ten additional questions capture the
9 National Hospital Organization (Independent Administrative Institution) page 1; http://www.mof.go.jp/english/filp/filp_report/zaito2004e-exv/24.pdf.
16
comprehensive assessment of the patient’s overall satisfaction. All the questions use a 5-point
Likert-type scale, where 1 indicates “strong dissatisfaction” and 5 indicates “strong satisfaction.”
Data are collected and processed by a university research center, which is unrelated to the NHO.
Feedback reports are subsequently distributed to each member hospital. These reports contain the
average score for each of the questions included in the questionnaires and the ranking of the
hospital within the NHO based on the overall satisfaction results.10
3.2 VARIABLE REDUCTION
Patient satisfaction is a multidimensional construct (Chen [2009]). To obtain a measure
of the underlying dimensions, we conduct a factor analysis using principal component analysis
with oblique rotation.11 Untabulated results show that for inpatients (outpatients) the items load
on two (three) factors. All factor loadings are greater than 0.6 and no variable cross-loads on
more than one factor. Based on the items’ loadings we identify two constructs for inpatients:
staff /treatment, and logistics/infrastructure. Outpatients load on three factors, which we label as
staff /treatment, administrative procedures, and location. Cronbach’s alpha values are greater
than 0.9 for each factor for both inpatients and outpatients. The factor analysis of the ten
questions related to overall satisfaction yields a single factor for inpatients (Cronbach’s alpha >
0.95) and outpatients (Cronbach’s alpha > 0.97) respectively. Since location is not controllable
by hospital managers, we drop it from subsequent analysis.
3.3 ANALYSIS OF INPATIENTS AND OUTPATIENTS
Inpatients and outpatients may weigh the importance of each factor differently in
10 Conversations with NHO hospital managers confirmed that prior to 2004, hospitals were not systematically collecting patient satisfaction information. 11 The oblique rotation method allows for the possibility that the factors are correlated.
17
formulating their assessments of overall satisfaction with the hospital.12 Therefore, we examine
the extent to which each of the component factors influences overall inpatient and outpatient
satisfaction using OLS regressions with heteroskedasticity-robust standard errors clustered by
hospital. We estimate the following model:
Satisfaction= α+β1*Staff and Treatment+β2*Logistics and Infrastructure
+β3*Administrative Procedures+ β4*Size+β5*Competition (1)
+ β6*Hospital +ε
where Staff and Treatment, Logistics and Infrastructure, and Administrative Procedures
are measured by the corresponding factor scores. We control for hospital Size (measured as the
number of beds in thousands), hospital type (Hospital =1 if the site is a general hospital, 0 if the
site is a sanatorium), and competition intensity (Competition), which is the number of private and
public hospitals per 100,000 people in the geographic area (prefecture) where each NHO hospital
is located. We estimate separate models for general hospitals and sanatoriums. Results in Table 2
indicate that while both Staff and Treatment and Logistics and Infrastructure are significant
drivers of overall patient satisfaction for all patients, inpatients of general hospitals and
sanatoriums weigh satisfaction with Staff and Treatment more than Logistics and Infrastructure
(all p-values of difference in coefficients < 0.10). For outpatients of sanatoriums, Staff and
Treatment is a larger driver of overall satisfaction than are Administrative Procedures (p-value of
difference in coefficients < 0.10), while the difference in the contribution to overall satisfaction
12 A recent survey conducted by the Japanese Ministry of Health, Labor and Welfare explored the major drivers of hospital choice for inpatients and outpatients. The sample consisted of more than 150,000 respondents, randomly selected from the patient population of all Japanese Hospitals. Overall, outpatients (inpatients) identified the following drivers of hospital choice: 38% (34.9%) prior experience at the same hospital, 37.6% (29.9%) physical closeness to their residence, school or place of work, 33.2% (49%) recommendation by doctors, 31.4% (34.7%) kindness of doctors and nurses, and 28.7% (25.5%) size/technology of the hospital. (Japanese Ministry of Health, Labour and Welfare. (2011). Patients Behavior Survey. from http://www.mhlw.go.jp/english/new-info/2012.html).
18
of the two factors is not significant for outpatients of general hospitals.
----- Insert Table 2 here -----
3.4 TEST OF H1: INFORMATION EFFECT OF PATIENT SATISFACTION
H1 predicts a positive effect of patient satisfaction information. That is, the release of
patient satisfaction information improves satisfaction performance in subsequent years due to the
incremental value of the information signal (improved mapping between actions and outcomes).
3.4.1 Univariate analysis
Table 3, Panel A provides information on the mean inpatient and outpatient satisfaction
for each year subsequent to the information release, and partitioned by the level of hospital
satisfaction performance during the first year of the release (2004). Panel B provides
corresponding means for change in satisfaction.13 It can be seen that both inpatient and outpatient
satisfaction improved each year after 2004, with the exception of the year 2009. Top performers
(hospitals that ranked in the top quartile with respect to overall satisfaction in 2004) did not
significantly decrease their performance in 2005 despite the pressure faced by other hospitals
following the patient satisfaction information release, nor did the mean of top performers drop
below the full sample mean during any year. Indeed, the top performers continued to perform
significantly above the mean during the entire period (all p-values for the difference between top
performers and the whole sample are > 0.05).14
----- Insert Table 3 here -----
13 For expositional clarity and ease of interpretation, overall satisfaction in Table 2 is calculated as the average of the ten questions assessing overall satisfaction for inpatients and outpatients respectively. Univariate analyses performed using the factor scores for overall satisfaction yielded equivalent results (untabulated). 14 This provides evidence that regression to mean is not a problem with the data. If regression to the mean was occurring, top performers would have statistically significantly deteriorated in the subsequent years and eventually reached the sample mean. Formal analysis of regression to the mean is provided in Section 4.
19
3.4.2 Multivariate analysis
We test H1 using the following model:
Satisfaction = α+ β1Size+ β2Competition+ β3Hospital + β4-‐10Year +𝜀 (2)
We control for hospital type (Hospital), hospital size (Size), and competition intensity
(Competition). We estimate six models, one each for the overall satisfaction score for inpatients
and outpatients respectively, one for each of the two satisfaction factors i.e., staff/treatment and
logistics/infrastructure for inpatients, and staff/treatment, and administrative procedures for
outpatients. Satisfaction results for 2004 are used as baseline and therefore 2004 is the omitted
case for Year. The model controls for hospital fixed effects and calculates heteroscedasticity-
robust standard errors clustered by hospital.
Table 4 reports the estimation results, which indicate that the introduction of the patient
satisfaction assessment tool has, on average, a positive effect on hospital overall patient
satisfaction for inpatients and outpatients alike in each of the seven subsequent years following
the first year of the patient satisfaction information release. The coefficients estimated for the
two satisfaction factors for inpatients and the staff and treatment factor for outpatients similarly
indicate that relative to 2004, satisfaction performance is higher in each subsequent year. This is
consistent with the prediction of the information effect (H1). That is, patient satisfaction
information improves performance, even after controlling for time invariant characteristics and
unobservable hospital level factors. General hospitals had higher levels of overall satisfaction for
inpatients as well as satisfaction with staff and treatment and logistics and infrastructure.
However, sanatoriums had higher levels of overall outpatient satisfaction and satisfaction with
staff and treatment and administrative services.
----- Insert Table 4 here -----
20
3.5 TEST OF H2: REFERENT PERFORMANCE EFFECT OF PATIENT SATISFACTION
H2 predicts that a lower initial performance on patient satisfaction is associated with
higher magnitude of subsequent nonfinancial performance improvements, arising from the value
of the relative performance signal. Univariate results in Table 3, Panel B, indicate that the mean
patient satisfaction change from 2004 to 2005 for hospitals in the lowest performance quartile in
2004 (poor performers) is higher than the change for the full sample for both inpatients (0.317
versus 0.087, p<0.01) and outpatients (0.204 versus 0.088, p<0.01).
To test H2, we estimate the following model using OLS with heteroskedasticity-robust
standard errors clustered by hospital and controlling for hospital fixed effects:
Change in Satisfaction = α + β1Trend + β2Poor Performer + β3Time Trend*Poor Performer (3) + β4Hospital + β5Size+ β6Competition+𝜀 where Change in Satisfaction is defined as the yearly change in patient satisfaction (Yeart –
Yeart-1), Trend is a discrete variable, which takes the value of 1-8 corresponding to the years
2004-2011, Poor Performer Dummy is an indicator variable that takes the value of 1 if the
hospital was in the lowest quartile of performance in the first year of the information release
(2004), and Competition and Hospital are as defined previously.
Results presented in Table 5, Panel A indicate a positive coefficient for Poor Performer
for outpatients and inpatients alike and for each of the factors. These results indicate that
hospitals that had lower satisfaction scores during the first year of mandatory performance
measurement (2004) had higher increases in satisfaction. These results indicate support for the
referent performance value of information, consistent with H2. To account for the non-linearity
of the changes and for the effects of compression at the top (that is, improvements are harder to
obtain when hospitals are already performing well), we re-estimate equation 2 using a scaled
21
version of the dependent variable, defined as ((Satisfactiont – Satisfactiont-1)/ (5 –
Satisfaction2004)). Table 5, Panel B, shows that even considering for the smaller opportunities for
improvement available for higher performers, hospitals that had lower satisfaction scores in 2004
(poor performers) obtained a larger scaled improvement in satisfaction, which further supports
H2. Additionally, the trend variable reduces the influence of diminishing marginal returns arising
from the ceiling effect.
----- Insert Table 5 Here -----
3.6 TEST OF H3 AND H4: VALUE OF FIRST INFORMATION VERSUS SUBSEQUENT
INFORMATION
The results in Table 5 can help examine if information has a higher value when it is new.
For inpatients, the results reveal that the average rate of improvement in overall satisfaction
decreases over time, captured by the significant β1 coefficient of -0.038 for unscaled changes and
-0.013 for scaled changes (p<0.01). For the inpatient factors, the decreasing trend is observed for
the staff and treatment factor (unscaled). The overall outpatient satisfaction improvement has a
similar decreasing trend (-0.020 unscaled), which is also found for the staff and treatment factor
(-0.074 unscaled and -0.018 scaled) and the administrative procedures factor (-0.006 scaled).
These results are generally consistent with H3 and indicate that greater improvements in
performance occur in earlier years relative to later years. That is, the information effect of the
performance signal is higher when the signal is first introduced. Results are also consistent with
H4, that is, the referent performance effect of the higher rate of increase in performance for
initial poor performers even accounting for the ceiling effect also decreases over time. The
Trend*Poor Performer interaction has a negative coefficient for overall inpatient and outpatient
satisfaction (-0.072 unscaled for overall inpatient satisfaction, -0.085 unscaled and 0.012 scaled
22
for overall outpatient satisfaction). Although hospitals that were performing poorly in the first
year of patient satisfaction had significantly higher rates of increase in performance as discussed
in the test of H2, there was a declining rate of increase in performance for these hospitals for the
overall inpatient satisfaction (unscaled) and outpatient satisfaction (unscaled and scaled), as well
as the two inpatient and outpatient satisfaction factors (unscaled and scaled). This indicates that
the value of new information for referent comparisons follows a declining rate of improvement.
Results for the control variables indicate that general hospitals had higher rate of increase in
inpatient performance (unscaled and scaled) and correspondingly, sanatoriums had lower rates of
increase in performance (captured by the significant positive β5 coefficient of 0.066 for unscaled
changes and 0.025 for scaled changes).
3 Robustness Analysis and Alternative Explanations
In this section, we explore whether the results can be attributed to regression to the mean
or the continuation of a pre-existing trend. We also examine if the physician or administrators’
incentive system drives the improvement in satisfaction performance.
4.1 REGRESSION TO THE MEAN
All the hypotheses relate to change in patient satisfaction subsequent to the introduction
of the new measurement system. Unobserved hospital-level characteristics may bias the results.
That is, hospitals whose initial patient satisfaction measure is high are likely to keep performing
well due to characteristics other than the patient satisfaction. A regression analysis of the
satisfaction scores for each year on the satisfaction measure of the previous year confirms the
presence of significant first-order hospital fixed effects (untabulated). Consequently, all the
analyses in this study control for hospital fixed effects and estimate robust standard errors
23
clustered by hospital. This reduces the likelihood of unobserved hospital-level characteristics
driving the results.
Because of correlations between repeated measures, and standard deviations that decrease
over time, it is necessary to consider the extent to which results may be a manifestation of
regression towards the mean rather than actual improvements in patient satisfaction. That is,
poorly performing hospitals may exhibit higher performance simply because of the nature of the
behavior of extreme values in a statistical distribution rather than an actual improvement.15 We
are able to reject the hypothesis of regression to the mean based on the following. First, Table 3,
Panel A shows that the overall mean inpatient and outpatient satisfaction measure increases over
time (except for the 2009 year). This rules out aggregate mean stability, a necessary condition for
regression to the mean (Cook and Campbell [1979], Zhang and Tomblin [2003]). Second, the
overall satisfaction never decreases significantly below the initial (2004) levels in aggregate or
for the hospitals in the highest quartile of performance, which is further evidence against the
aggregate mean stability condition. Indeed, the mean inpatient satisfaction is substantially higher
in 2011 than 2004 for the aggregate (4.448 in 2011 versus 4.194 in 2004, p<0.01), for hospitals
in the lowest quartile (4.26 in 2011 versus 3.78 in 2004, p<0.01), as well as for hospitals in the
highest quartile (4.558 in 2004 versus 4.424 in 2004, p<0.01). Similar conclusions can be drawn
for outpatient overall satisfaction, where average scores are is statistically higher in 2011 than
2004 for the pooled sample (4.119 in 2011 versus 3.951 in 2004, p<0.01), for the lowest initial
quartile (3.999 in 2011 versus 3.664 in 2004, p<0.01) and for the highest initial quartile (4.170 in
2011 versus 4.244 in 2004, p<0.10). Third, we estimate the correlation between subsequent 15 Note that regression to the mean is primarily an issue when the analysis consists of only two observations, such as two variables measured on one occasion (e.g. control and treatment group in an experiment) or one variable measured on two occasions (e.g. pre-test post-test comparison after an experimental intervention). Regression to the mean is not a phenomenon that is relevant to multiple observations over time (Nesselroade et al. [1980]).
24
satisfaction measures after controlling for hospital fixed effects and other determinants of
satisfaction. Untabulated results show non-significant correlations coefficients, thus rejecting the
possibility that the improvement over time is purely a result of regression towards the mean
(Cook and Campbell [1979], Zhang and Tomblin [2003]). Finally, all analyses of satisfaction
change include the initial (2004) satisfaction rates, which is standard practice for controlling for
regression to the mean (Evans, et al. [1997], Kolstad [2013]). Consequently, we conclude that
regression towards the mean has a non-significant influence on the results of this study.
4.2 PRE-EXISTING TREND
It is possible that the change in patient satisfaction is merely due to the continuation of a
pre-existing trend in all the important firm-level variables. The absence of any systematic data
collection with respect to patient satisfaction prior to 2004 makes it impossible for us to perform
a true pre-post analysis. However, by including a time trend variable in equation 3, we control
for unobserved factors that might contribute to the increase in patient satisfaction due to the mere
passage of time (Wooldridge [2012]). General learning trends are likely to emerge and drive
better performance as the expertise of the members of the organization grows over time. By de-
trending the data in equation 3, we eliminate the concern that the significance of the regression
coefficients reported in Table 5 is due to spurious correlations with pre-existing drivers of patient
satisfaction over time. However, for robustness, we conduct three additional analyses. First, we
analyze the patterns of change in total costs. If the behavior of patient satisfaction was simply a
pre-existing trend, we should observe a similar trend with cost. We estimate equation 3, using
changes in total operating costs as the dependent variable. Estimation results (untabulated)
indicate, (a) no evidence in support of a declining rate of change in operating costs (on the
contrary – costs tend to accelerate over time), (b) no statistical significance for the poor
25
performer dummy, and (c) no statistical significance of the Trend*Poor Performer interaction,
both for inpatients and outpatients. Recall that for the cost variable, there is no new information
obtained. The results indicate that the patterns of cost changes are not similar to the patterns of
patient satisfaction changes, which reinforces the conclusion that the change in patient
satisfaction results are driven by the value of new information for decision making.
Second, using data from 2004-2011, we computed a hospital-specific annual cost trend.
We added the hospital-specific annual cost trend to equation 3 to control for the cross sectional
variation in idiosyncratic cost behavior. We assume that if there is a pre-existing trend for each
hospital, this trend would be correlated among the important economic variables for the hospital.
We estimate equation 3 augmented by this additional hospital level control variable and using
changes in patient satisfaction (unscaled and scaled) as the dependent variable. Even after
controlling for the cost trend, untabulated results continue to support hypotheses H2, H3, and H4
with equivalent degree of statistical significance. We conclude that the information effect and the
referent performance effect are not due to the continuation of a pre-existing trend in hospital-
level outcome variables.
4.3 COMPENSATION PRACTICES AT NHO AS A DRIVER
In a departure from accounting studies that examine the role of information for planning,
performance evaluation, and control, we are interested in studying the effect of the decision
facilitating role (planning role) of information unconfounded by the decision influencing role
(performance evaluation and control role). While we do not disagree that “both types of
information gathering and dissemination are used by the principal for the same purpose, to
exercise control (broadly construed) over the agent's decision making” (Baiman and Demski
[1980], 185), we believe that calibrating the value of information in the decision facilitating role
26
is important. The vast majority of rank and file workers receive minimal incentive compensation
and it is important to examine if improving their information set can improve decisions.
Prior research by Banker, et al. [2000] finds improvement in nonfinancial performance
following the implementation of an incentive plan that includes nonfinancial performance
measures. In their setting, although customer satisfaction measures were tracked before the
implementation of an incentive plan that includes such measures, these measures did not have a
performance effect until it was used explicitly for incentive purposes. Similarly, Kelly [2007]
demonstrates experimentally that providing feedback on nonfinancial measures alone improves
managerial decisions over time, but providing feedback and incentives attached to the measures
that were provided in the feedback is the key to improving performance. In (Evans, et al. [1997]),
even though the hospitals previously collected mortality data, performance improvements were
only observed when the mortality measures were disclosed to the public. This suggests that in
these studies, the decision facilitating role of information was present in conjunction with the
decision influencing role. Whether or not increased patient satisfaction information improves
performance in absence of an explicit link to incentive compensation, and when the information
is disseminated to internal as opposed to external stakeholders is our empirical question. Our aim
is to isolate the value of information unfettered by the effect of financial incentives.
We examine physician compensation practices at NHO hospitals using field and archival
data to determine if there is any link between patient satisfaction performance and compensation.
If this link is indeed present, we cannot rule out the possibility of confound between decision
influencing and decision facilitating role of satisfaction information, which is the bane of
compensation studies.
4.3.1 Field evidence of compensation practices at NHO
27
We conducted interviews with hospital administrators at the NHO headquarters to glean
information about physician compensation practices. These interviews revealed that there is no
explicit link between physician or administrative officer compensation and patient satisfaction
performance. Essentially, Japanese NHO physicians and administrators are government
employees. Each physician, nurse, paramedic, and administrator is classified into a particular
grade based on a hierarchy, and each grade is provided compensation as per a government salary
schedule. The typical compensation package includes: a monthly salary, allowances for cost of
living, overtime and travel. Appendix B contains information on employment, compensation,
and promotion systems at NHO. For example, a trauma surgeon who has the highest level of
seniority than can be attained would earn a monthly salary of 572,900 Yen (Employment Grade
10, Step 21). Adding a 20% allowance for cost of living and housing, the total compensation
package is 687,480 Yen ($6,027 per month in current U.S. Dollars).
4.3.2 Archival evidence of compensation practices at NHO
To empirically examine whether there is any link between patient satisfaction and
physician compensation, we estimated the following model of physician salary and bonus as a
function of patient satisfaction:
𝐶𝑜𝑚𝑝𝑒𝑛𝑠𝑎𝑡𝑖𝑜𝑛i,t = α + β1Overall Inpatient Satisfactioni,t-1+ β2Overall Outpatient Satisfactioni,t-1 + 𝜕k𝐻𝑜𝑠𝑝𝑖𝑡𝑎𝑙 𝐶𝑜𝑛𝑡𝑟𝑜𝑙𝑠!+ 𝛾!Year+𝜀 (4)
where the dependent variable Compensation is operationalized either as Salary (total
compensation paid to physicians by hospital i in year t including cash, bonus, and other benefits)
or Bonus (the total annual bonus payout by hospital i to physicians in year t). We include k
control variables for size, competition, hospital type, medical revenue, educational revenue, and
R&D revenue, and indicator variables for each of the 8 years included in our sample (j = 2,…,8;
28
2004 is dropped as the base case). Table 1 provides the descriptive statistics for the variables
used in the analyses.
If patient satisfaction were taken into consideration in salary or bonus payouts, the
coefficient on 𝛽! and 𝛽! would be positive. We use lagged satisfaction scores because the
satisfaction scores for year t are only released to member hospitals in the following year and
therefore likely to only be incorporated into the following year salary or bonus payments. The
results of estimating equation 4 (Table 6, Columns 1 and 2) indicate no significant association
between salary or bonus and either inpatient or outpatient satisfaction. These results suggest that
patient satisfaction is not taken into consideration in the determination of salary or bonus
payments. Thus, we conclude that improvements in patient satisfaction are not driven by the
motive to increase compensation.16 Both salary and bonus payments are, however, positively
associated with hospital size and medical revenue.
----- Insert Table 6 here -----
4.4 GRANT REVENUES
It is possible that patient satisfaction is driven by the desire to increase grant revenue for
NHO hospitals. If so, although there is no explicit tie of satisfaction performance and incentive
compensation, hospitals may have a pecuniary benefit to improving satisfaction. To test this, we
estimate equation 4 with Grant Revenue as the dependent variable. In addition to the controls
listed in the previous section we include R&D expenses, Facility Cost, Material Cost, and Total
Cost to control for the relation between grants awards and their potential uses. Estimation results
are provided in Table 6 (Column 3) and indicate a negative association between inpatient
satisfaction and hospital grant revenue. Our conversations with NHO administrators revealed that
16 We re-estimated equation 4 using two-and three-year lags and continued to find non significant coefficients for salary and for bonus.
29
hospital grants are based on the research output of the hospitals rather than patient satisfaction.17
We speculate that a reason for the negative association between grant revenue and patient
satisfaction is that hospitals with a higher research emphasis receive greater grants, but have a
lower focus on the patient satisfaction scores.
4.5 PATIENT REVENUES
Prior research finds a positive relationship between customer satisfaction and subsequent
revenues (Chen et al. [2009], Hallowell [1996], Ittner and Larcker [1998]). The mechanisms by
which customer satisfaction influences revenues include customer attraction, customer loyalty,
and word of mouth (Rust et al. [2002], Szymanski [2001]), which also operate in the non-profit
health care industry (Gemme [1997], Stizia and Wood [1997]). When patients are satisfied with
the hospital, they tend to return. Additionally, their opinion influences others’ choices (Gemme
[1997]).18 A successful patient satisfaction program can therefore result in increased revenues.
To test the relationship between patient revenues and satisfaction, we estimate equation 4 with
hospital revenue as the dependent variable. We use lagged values of satisfaction in order to allow
for word of mouth and other mechanisms to take place. We find a statistically significant
association between patient satisfaction and lagged revenues for both inpatients and outpatients
(Table 6, Columns 4 and 5), which indicates that patients respond to hospitals’ efforts to improve
patient satisfaction.19
4 Conclusions
17 We re-estimated equation 4 using three-, four- and five- year lags and did not find significant results. Similar analysis of changes in grant revenues did not yield significant results. 18 Gemme [1997] reports survey results indicating that 90% of patients’ choice of health care provider are influenced by other patients’ opinions, and that 40% of surveyed subjects had consulted a patient who had used the service of the organization they had chosen to use. 19 We re-estimated equation 4 using contemporaneous patient satisfaction data and found similarly significant associations with patient revenue.
30
Information has decision value, regardless of whether it is private or public information,
or whether the information pertains to financial or nonfinancial performance. Accounting theory
posits that the decision value of information manifests itself in two primary ways. The first type
of manifestation refers to the ability of information to facilitate belief revision and improve
managerial decision making (the decision facilitating role). The second manifestation of value of
information emerges from the use of information in managerial contracts for performance
evaluation to motivate managerial effort and improve risk sharing (the decision influencing role).
However, most empirical studies related to the role of information disclosures confound the
decision influencing and the decision facilitating roles.
The decision influencing role emerges from the value of an information signal either as a
mechanism to reduce information asymmetry between the firm and external stakeholders, or
within the firm to reduce agency problems. With respect to the relationship between the firm and
external stakeholders, there is considerable research in accounting that debates whether firm
value fully or partially reflects publicly available information, and adjusts to it either
instantaneously or with a lag. Research has examined regulatory and other mechanisms that
nudge firms to disclose private information and improve the smooth functioning of financial
markets by reducing information asymmetry between the firm and the market (Healy and Palepu
[2001]). Within the firm, research has focused on designing contracts that induce managers to
reveal private information, or that reduce their motivation to shirk (Lambert [2001]). From a
practical perspective, as Narayanan and Davila [1998] state: “Most firms collect a plethora of
information for belief revision, even though only a few signals are directly linked to incentives.”
Further they note that when a signal that is useful for belief-revision is also used for performance
31
evaluation, the manager has incentives to manipulate the signal. Managerial manipulation effort
lowers the value of the performance signal in its belief revision and learning role.
Our empirical setting allows an examination of the value of information with minimal
confound from public disclosures or incentive compensation. We use patient satisfaction data for
a sample of 145 Japanese public hospitals for a period of eight years (2004-2011) to explore the
decision facilitating value of information. In our setting, hospitals obtained two new signals –a
signal about the absolute level of patient satisfaction performance, and a signal about the level of
performance relative to a referent group. We find evidence of an information effect - the new
information signal about the level of performance induces a more precise posterior distribution
of beliefs and facilitates decision making. We find a referent performance effect arising from the
value of the new information signal about performance relative to a peer group. Results also
indicate that information has higher decision facilitating value when it is new. New information
results in the maximum belief revision, compared to subsequent updates.
Mandatory performance measurement systems are generally intended to facilitate better
decisions. However, prior research suggests that when performance measurement systems are
mandated through legislative provisions, subordinate organizations are likely to comply with the
regulatory requirement, but make little use of such information for internal decisional processes
(Cavalluzzo and Ittner [2004]). Contrary to those findings, in our setting the subordinate
organizations (i.e., NHO hospitals) appear to use the new information from the mandatory
measurement. Our results support the notion that information generated by mandatory
performance measurement systems can have decision value for firms.
Prior research on mandatory nonfinancial performance measurement in the healthcare
industry has primarily used publicly disclosed medical measures of hospital quality (such as
32
mortality). The general consensus in the literature is that public disclosures of performance
influence future demand and have the potential to assist consumers make informed choices. Two
problems plague most studies in this area. First, it is difficult to disentangle real improvements in
quality from increases arising from gaming behaviors by firms to artificially boost quality
measures. In our setting, the information is collected and analyzed by an independent third party,
reducing the likelihood of gaming. Second, demand may increase because of customers’
response to rankings as opposed to response to an actual increase in service quality (Dranove and
Jin [2010]). We find an association between changes in patient satisfaction and subsequent
patient revenues even without public dissemination of performance information. We infer that
patients respond to the actual performance improvements rather than exhibiting a mechanical
response to rankings.
Substantial evidence shows that firms should not neglect nonfinancial measures such as
satisfaction because they provide leading indicators of future financial value.20 These studies
focus on the pecuniary benefits of allocating managerial attention to nonfinancial measures.
Indeed, Banker, et al. [2000] in their study of customer satisfaction in the hotel industry mention
that without tying nonfinancial incentive measures to compensation contracts, “managers did not
recognize the true benefit of allocating more effort and resources to improve customer
satisfaction, and did not do so until the change in compensation plan that focused their attention
on improving the customer satisfaction measures” (p.90). Evidence in this study builds on the
conclusion in Banker, et al. [2000] and empirically shows that nonfinancial information has the
ability to achieve improved outcomes even without including these metrics in compensation
20 For example see Baiman and Baldenius [2009], Banker, et al. [2000], Chenhall [2005], Feltham and Xie [1994], Hemmer [1996], Ittner and Larcker [1998], Kaplan and Norton [1992], Kekre et al. [1995],Krishnan et al. [1999], Nagar and Rajan [2005], Rajan and Reichelstein [2009].
33
plans. In the hospital industry, nonfinancial performance measure such as patient satisfaction is
especially important and forms a central feature of patient centered medicine (Bardes [2012]).
Our findings offer significant contributions to research, practice, and policy by showing
that in health care settings, nonfinancial performance information has a decision-facilitating role
in addition to the previously documented decision-influencing role. While it is important to
ensure that health care decision makers have access to information that facilitates decision-
making, there is debate on whether mandatory public disclosure of information has the potential
to reduce social welfare. Mandatory public disclosure can be a double-edged policy instrument
because it can cause over-reaction by the public, as well as neglect of private information by
decision makers (Morris and Shin [2002]). Results of our study indicate that, in some instances,
mandating information collection without public disclosure requirements or compensation-based
incentives can be beneficial for managerial learning.
34
Appendix A
Survey Instrument
Panel A: Overall satisfaction (Same questions asked separately for outpatients and inpatients; Scale 1 = Strongly Dissatisfied; 2 = Somewhat Dissatisfied; 3 = Neutral; 4 = Somewhat Satisfied; 5 =Strongly Satisfied
Panel B: Individualized questions for inpatients (Scale 1 = Strongly Agree; 2 = Somewhat Agree; 3 = Neither Agree nor Disagree; 4 = Somewhat Disagree; 5 =Strongly Disagree) 1. I am not satisfied with the explanation by doctors when I was hospitalized 2. I was unhappy with the procedure of medical admission 3. I was unhappy with hospital's explanation about my life during the hospital stay 4. I think that the doctors behave badly and use bad language in this hospital 5. I was worried about some doctors' skills and knowledge 6. I think that the nurses behave badly and use bad language in this hospital 7. I was unhappy with the assistance received for daily life activities 8. I think that medical staff such as doctors, nurses and other medical staff lacked teamwork 9. I did not like today's medical tests (For patients who accepted medical tests) 10. I did not like today's medical surgeries (For patients who accepted medical surgeries) 11. I did not like today's medical treatment (For patients who accepted medical treatment) 12. I did not like today's drip, injection, medicine, or prescription (For patients who had a drip, injections, medicine, or prescription) 13. I did not like today's rehabilitation (For patients who had rehabilitation) 14. I am unhappy with the toilets and bathrooms in this hospital 15. I think that passageways, stairs and elevators are inconvenient 16. I am unhappy with my room 17. I am unhappy with the food in this hospital 18. I am unhappy with the other environment such as stores, and interiors 19. I am unhappy with the hospital's explanation of my discharge
1. I am generally satisfied with this hospital 2. I am satisfied with the results of the treatment 3. I am satisfied with the period of the treatment 4. I am satisfied with treatment I have been taking 5. I am satisfied with the hospital 6. I think this hospital provides safe medical services 7. I think the explanations provided by the medical staff were very clear 8. I think the treatment I have received was acceptable 9. I generally trust this hospital 10. I would like to recommend this hospital to family members and friends
35
Panel C: Individualized questions for inpatients (Scale 1 = Strongly Agree; 2 = Somewhat Agree; 3 = Neither Agree nor Disagree; 4 = Somewhat Disagree; 5 =Strongly Disagree) 1. I felt uneasy when I came to the hospital at the initial visit 2. I think that this hospital is very inconvenient 3. I have a bad impression about this hospital 4. I am unhappy with waiting time 5. I am unhappy with the waiting room 6. I think that doctors behave badly and use bad language in this hospital 7. I was worried about some doctors' skills and knowledge 8. I think that nurses behave badly and use bad language in this hospital 9. I did not like today's medical tests (For patients who accepted medical tests) 10. I did not like today's medical treatment.(For patients who accepted medical treatment ) 11. I did not like today's drip, injection, medicine, or prescription (For patients who had a drip, injections, medicine, or prescription) 12. I did not like today's rehabilitation (For patients who had rehabilitation) 13. I am unhappy with the treatment room 14. I am unhappy with the other environment such as shops, ATM, and interiors 15. I am unhappy with the procedures for payment Notes to Appendix A: This appendix lists the questions used in the survey administered to Japanese National Health Organization (NHO) general hospitals and sanatoriums. The translation from Japanese to English aimed at maintaining the original meaning as close as possible.
36
Appendix B
Physician Compensation at NHO
Salary schedule: Each NHO post is classified into a certain grade in a salary schedule. The classification of the employee into a post is based on two factors: educational classification and experience. Most Japanese government agencies have ten grades. Within each grade employees receive raises in steps, which are based on time in grade. A sample of the pay scale for a Japanese government agency is provided below.
Salary per month (Yen) Grade
1 2 3 4 5 6 7 8 9 10
Steps
1 135,600 185,800 222,900 261,900 289,200 320,600 366,200 413,000 466,700 532,000 5 140,100 192,800 230,200 270,200 298,200 329,800 376,300 422,800 479,000 544,700 9 144,500 200,000 237,500 278,600 307,300 338,600 386,400 432,300 491,300 554,700
13 149,800 207,000 244,900 287,000 316,400 347,200 397,100 441,300 503,600 562,100 17 155,700 214,600 252,200 295,400 325,200 355,500 406,400 449,300 513,300 568,100 21 161,600 222,000 260,100 303,800 333,500 363,500 414,800 456,500 519,000 572,900 25 172,200 229,300 267,700 312,100 341,500 371,500 422,900 462,500 524,800 29 178,800 236,100 275,300 320,400 349,400 379,500 429,400 467,800 529,600 33 185,800 242,100 282,700 328,400 357,000 386,900 434,600 471,000 533,100 37 191,600 248,000 290,100 336,500 364,200 393,700 439,700 474,200 536,700 41 196,900 254,200 297,400 344,400 370,100 398,400 443,200 477,400 540,300 45 202,000 259,700 304,200 352,000 374,700 403,000 446,400 480,500 49 207,100 265,200 310,600 358,500 378,400 405,900 449,400 53 211,600 270,100 317,100 363,000 381,700 408,800 452,400 57 215,400 275,200 323,400 367,100 384,500 411,600 455,400 61 219,200 279,700 328,100 369,800 387,000 414,300 458,400 65 223,000 283,500 331,900 372,400 389,600 416,900 69 226,900 287,200 335,200 375,000 392,200 419,400 73 230,100 290,400 337,800 377,600 394,800 422,000 77 233,000 292,300 340,000 380,200 397,300 424,600 81 236,100 293,800 342,000 382,700 399,900 85 239,000 295,300 344,000 385,100 402,500 89 241,900 296,800 345,900 387,600 93 243,700 298,200 347,700 390,100 97 299,600 349,500
Composition of Salary: In addition to the monthly salary, government employees also get allowances averaging at about 20% of base salary. The allowances include: living expenses (cost of living adjustment), housing allowance, commuter allowance, overtime allowance, cold weather allowance, and diligence allowance (typically based on the number of months of consecutive work in the previous 6-month period). There are some compensation adjustments related to macroeconomic conditions. Individual performance-based bonuses are not commonly found.
37
References
ABERNETHY, M. A. and J. BOUWENS. “Determinants of Accounting Innovation Implementation.” ABACUS 41 (2005): 217-240.
ALPERT, M. and H. RAIFFA "A Progress Report on the Training of Probability Assessors," in Judgmenr under Uncertainty: Heuristics and Biases, edited by D. Kahnemann, P. Slovic and A. Tverski. Cambridge, England: Cambridge University Press, 1982: 294-305.
ARROW, K. J. “Uncertainty and the Welfare Economics of Medical Care.” The American Economic Review 53 (1963): 941-973.
BAIMAN, S. and J. S. DEMSKI. “Optimal Performance Evaluation and Control Systems.” Journal of Accounting Research 18 (1980): 184-220.
BANDURA, A. and F. J. JOURDEN. “Self-Regulatory Mechanisms Governing the Impact of Social Comparison on Complex Decision Making.” Journal of Personality and Social Psychology 60 (1991): 941-951.
BANKER, R. D.; G. POTTER and D. SRINIVASAN. “An Empirical Investigation of an Incentive Plan That Includes Nonfinancial Performance Measures.” The Accounting Review 75 (2000): 65-92.
BARDES, C. L. “Defining "Patient-Centered Medicine".” New England Journal of Medcine 366 (2012): 782-783.
BENOIT, J.-P. and J. DUBRA. “Apparent Overconfidence.” Econometrica 79 (2011): 1591-1625.
BROMWICH, M. "Economics in Management Accounting," in Handbook of Management Accounting, edited by A. G. H. Christopher S. Chapman, Michael D. Shiends: Elsevier, 2007: 137-162.
CAMERER, C. and D. LOVALLO. “Overconfidence and Excess Entry: An Experimental Approach.” American Economic Review 89 (1999): 306-318.
CASAS-ARCE, P. and F. A. MARTINEZ-JEREZ. “Relative Performance Compensation, Contests, and Dynamic Incentives.” Management Science 55 (2009): 1306-1320.
CAVALLUZZO, K. S. and C. D. ITTNER. “Implementing Performance Measurement Innovations: Evidence from Government.” Accounting, Organizations and Society 29 (2004): 243-267.
CHEN, C. X. “Who Really Matters? Revenue Implications of Stakeholder Satisfaction in a Health Insurance Company.” The Accounting Review 84 (2009): 1781-1804.
CHEN, C. X.; M. MARTIN and K. A. MERCHANT. “The Effect of Timing on the Association between Customer Satisfaction Measures and Future Financial Performance.” In SSRN: http://ssrn.com/abstract=1196702. (2009)
CHENHALL, R. H. “Integrative Strategic Performance Measurement Systems, Strategic Alignment of Manufacturing, Learning and Strategic Outcomes: An Exploratory Study.” Accounting, Organizations and Society 30 (2005): 395-422.
COOK, T. D. and D. T. CAMPBELL. Quasi-Experimentation: Design and Analysis Issues for Field Settings. Boston: Houghton Mifflin, 1979.
DEMSKI, J. S. “Information Improvement Bounds.” Journal of Accounting Research 10 (1972): 58-76.
DRANOVE, D. and G. Z. JIN. “Quality Disclosure and Certification: Theory and Practice.” Journal of Economic Literature 48 (2010): 935-963.
38
EDERER, F. “Feedback and Motivation in Dynamic Tournaments.” Journal of Economics and Management Strategy 19 (2010): 733-769.
EVANS, J. H. I.; Y. HWANG; N. NAGARAJAN and K. SHASTRI. “Involuntary Benchmarking and Quality Improvement: The Effect of Mandated Public Disclosure on Hospitals.” Journal of Accounting, Auditing and Finance 12 (1997): 315-346.
FELTHAM, G. A. “The Value of Information.” The Accounting Review 43 (1968): 684-696. FELTHAM, G. A. and J. XIE. “Performance Measure Congruity and Diversity in Multi-Task
Principal/Agent Relations.” The Accounting Review 69 (1994): 429-453. FUNG, C. H.; Y.-W. LIM; S. MATTKE; C. DAMBERG and P. G. SHEKELLE. “Systematic
Review: The Evidence That Publishing Patient Care Performance Data Improves Quality of Care.” Annals of Internal Medicine 148 (2008): 111-123.
GEMME, E. M. “Retaining Customers in a Managed Care Market.” Marketing Health Services 17 (1997): 19-21.
HALLOWELL, R. “The Relationship of Customer Satisfaction, Customer Loyalty, and Profitability: And Empirical Study.” International Journal of Service Industry Management 7 (1996): 27-42.
HAMMITT, J. K. and A. I. SHLYAKHTER. “The Expected Value of Information and the Probability of Surprise.” Risk Analysis 19 (1999): 135-152.
HEALY, P. M. and K. G. PALEPU. “Information Asymmetry, Corporate Disclosure, and Capital Markets: A Review of the Empirical Disclosure Literature.” Journal of Accounting and Economics 31 (2001): 405-440.
HEMMER, T. “On the Design and Choice of "Modern" Management Accounting Measures.” Journal of Management Accounting Research 8 (1996).
ITTNER, C. D. and D. F. LARCKER. “Are Nonfinancial Measures Leading Indicators of Financial Performance? An Analysis of Customer Satisfaction.” Journal of Accounting Research 36 (1998): 1-35.
KAHNEMANN, D.; P. SLOVIC and A. TVERSKI. Judgment under Uncertainty: Heuristics and Biases. Cambridge, MA: Cambridge University Press, 1982.
KAPLAN, R. S. and D. P. NORTON. “The Balanced Scorecard - Measures That Drive Performance.” Harvard Business Review (1992): 71-79.
KEKRE, S.; M. S. KRISHNAN and K. SRINIVASAN. “Drivers of Customer Satisfaction for Software Products: Implications for Design and Service Support.” Management Science 41 (1995): 1456-1470.
KELLY, K. O. “Feedback and Incentives on Nonfinancial Value Drivers: Effects on Managerial Decision Making.” Contemporary Accounting Research 24 (2007): 523-556.
KOLSTAD, J. T. “Information and Quality When Motivation Is Intrinsic: Evidence from Surgeon Report Cards.” In NBER. (2013)
KRISHNAN, M. S.; V. RAMASWAMY; M. C. MEYER and P. DAMIEN. “Customer Satisfaction for Financial Services: The Role of Product, Services, and Information Technology.” Management Science 45 (1999): 1194-1209.
KRUGER, J. and D. DUNNING. “Unskilled and Unaware of It: How Difficulties in Recognizing One's Own Incompetence Lead to Inflated Self-Assessments.” Journal of Personality and Social Psychology 77 (1999): 1121-1134.
KRUGER, J. and D. DUNNING. “Unskilled and Unaware--but Why? A Reply to Krueger and Mueller (2002).” Journal of Personality & Social Psychology 82 (2002): 189-192.
39
LAMB, G. C.; M. A. SMITH; W. B. WEEKS and C. QUERAM. “Publicly Reported Quality-of-Care Measures Influenced Wisconsin Physician Groups to Improve Performance.” Health Affairs 32 (2013): 536-543.
LAMBERT, R. A. “Contracting Theory and Accounting.” Journal of Accounting and Economics 32 (2001): 3-87.
LANT, T. K. and P. F. HEWLIN. “Information Cues and Decision Making: The Effects of Learning, Momentum, and Social Comparison in Competing Teams.” Group & Organization Management 27 (2002): 374-407.
LEVITT, B. and J. G. MARCH. “Organizational Learning.” Annual Review of Sociology 14 (1988): 319-340.
LICHTENSTEIN, S.; B. FISCHHOFF and L. D. PHILLIPS "Calibration of Probabilities: The State of the Art," in Judgment under Uncertainty: Heuristics and Biases, edited by D. Kahnemann, P. Slovic and A. Tverski. Cambridge, England: Cambridge University Press, 1982: 305-334.
MORGAN, M. G.; M. HENRION and M. SMALL. Uncertainty: A Guide to Dealing with Uncertainty in Quantitative Risk and Policy Analysis. Cambridge University Press, 1992.
MORRIS, S. and H. S. SHIN. “Social Value of Public Information.” The American Economic Review 92 (2002): 1521-1534.
NAGAR, V. and M. V. RAJAN. “Measuring Customer Relationships: The Case of the Retail Banking Industry.” Management Science 51 (2005): 904-919.
NARAYANAN, V. G. and A. DAVILA. “Using Delegation and Control Systems to Mitigate the Trade-Off between the Performance-Evaluation and Belief-Revision Uses of Accounting Signals.” Journal of Accounting and Economics 25 (1998): 255-282.
NESSELROADE, J. R.; S. M. STIGLER and P. B. BALTES. “Regression toward the Mean and the Study of Change.” Psychological Bulletin 88 (1980): 622-637.
PRATT, J. W.; H. RAIFFA and R. O. SCHLAIFER. Introduction to Statistical Decision Theory. Cambridge, MA: MIT Press, 1995.
RAIFFA, H. Decision Analysis; Introductory Lectures on Choices under Uncertainty. Addison-Wesley, 1968.
RAJAN, M. V. and S. REICHELSTEIN. “Objective Versus Subjective Indicators of Managerial Performance.” The Accounting Review 84 (2009): 209-237.
RAMANARAYANAN, S. and J. SNYDER. “Information Disclosure and Firm Performance: Evidence from the Dialysis Industry.” In SSRN: Working Paper # 2152796. (2012)
RUST, R. T.; C. MOORMAN and P. R. DICKSON. “Getting Return on Quality: Revenue Expansion, Cost Reduction or Both?” Journal of Marketing 66 (2002): 7-24.
SIMONS, R. Levers of Control: How Managers Use Innovative Control Systems to Drive Strategic Renewal. Harvard Business Press, 2013.
STIZIA, J. and N. WOOD. “Patient Satisfaction: A Review of Issues and Concepts.” Social Science and Medicine 45 (1997): 1829-1843.
SUGDEN, R. “Reference-Dependent Subjective Expected Utility.” Journal of Economic Theory 111 (2003): 172-191.
SZYMANSKI, D. M. “Customer Satisfaction: A Meta-Analysis of the Empirical Evidence.” Journal of the Academy of Marketing Science 29 (2001): 16-35.
WOOLDRIDGE, J. Introductory Econometrics: A Modern Approach. Cengage Learning, 2012.
40
ZHANG, X. and J. B. TOMBLIN. “Explaining and Controlling Regression to the Mean in Longitudinal Research Design.” Journal of Speech, Language and Hearing Research 46 (2003): 1340-1351.
41
TABLE 1 Descriptive Statistics
Variable N Mean Median Std. Dev. Q1 Q3 Min Max
Size 1152 4.036 3.800 1.382 3.015 4.855 1.380 8.040 Competition 1152 6.986 6.300 2.620 5.100 8.300 3.400 16.400 Hospital 1152 0.403 0.000 0.491 0.000 1.000 0.000 1.000 Salary 1142 1.805 1.557 0.959 1.182 2.128 0.291 6.512 Bonus 1142 0.479 0.428 0.227 0.335 0.574 0.037 1.352 Grant 1141 0.031 0.016 0.046 0.000 0.043 0.000 0.438 Inpatient Revenue 999 4.276 3.633 2.666 2.493 5.218 0.905 15.581 Outpatient Revenue 999 0.877 0.537 0.876 0.238 1.225 0.017 5.068 Total Medical Revenue 1152 5.094 4.140 3.497 2.694 6.269 0.000 19.406 Educational Revenue 1142 0.050 0.001 0.077 0.000 0.086 0.000 0.452 R&D Revenue 1142 0.066 0.033 0.099 0.006 0.077 0.000 0.731 R&D Expenses 1142 0.002 0.001 0.003 0.000 0.003 0.000 0.029 Facility Costs 1142 0.508 0.378 0.421 0.238 0.638 0.046 2.779 Material Costs 1142 5.085 4.119 3.368 2.732 6.229 0.675 19.473 Total Costs 1142 1.230 0.782 1.242 0.329 1.631 0.005 6.768
42
TABLE 2 Regression Results: Relationship Between Factors and Overall Satisfaction
(Standard Errors in Brackets)
DV = Overall Satisfaction
Inpatients Outpatients Pooled Sample Hospitals Sanatoriums
Pooled Sample Hospitals Sanatoriums
Staff and Treatment 0.548*** 0.598*** 0.542*** 0.616*** 0.576*** 0.635*** [0.037] [0.045] [0.045] [0.036] [0.070] [0.0423]
Logistics and infrastructure
0.275*** 0.258*** 0.283*** [0.023] [0.023] [0.038]
Administrative Procedures
0.380*** 0.374*** 0.367*** [0.040] [0.070] [0.050]
Size 0.009 0.033** -0.017 0.017 0.073*** -0.033 [0.016] [0.013] [0.029] [0.022] [0.024] 0.0250]
Competition 0.043*** 0.033* 0.056*** 0.049** 0.064* 0.049** [0.014] [0.020] [0.020] [0.020] [0.033] [0.025]
Hospital 0.027 0.082 [0.026] [0.052]
Intercept 0.041* 0.049** 0.039* -0.034 0.020 -0.041 [0.022] [0.019] [0.022] [0.028] [0.053] [0.027]
N 983 433 550 1143 463 680 R2 0.777 0.755 0.769 0.769 0.798 0.754
Notes to Table 2: (1) The coefficients reported in Table 2 are estimated via OLS regression of the following model: Satisfaction= α+𝛽!*Staff and Treatment+𝛽!*Logistics and Infrastructure +𝛽!*Administrative Procedures+β4*Size+𝛽!*Competition+𝛽!*Hospital+𝜀. The model is estimated using robust standard errors clustered by hospital. (2) The factor scores are based on principal component analysis with oblique rotation. Factor scores are expressed in standardized terms. (3) Size (number of beds, in thousands) and Competition (number of hospitals per 100K people in the prefecture) are expressed in standardized terms. (4) Hospital is a binary variable assuming the value of 1 if the site is a general hospital, and 0 if the site is a sanatorium. (5) Wald tests indicate that the coefficients associated with Staff and Treatment are statistically different than the coefficients associated with Logistics and Infrastructure (for inpatients) and Administrative Procedures (for outpatients) in all cases, with the exception of general hospitals’ outpatients. (6) * = p-value <0.10; ** = p-value <0.05; *** = p-value <0.01, two-tailed.
43
TABLE 3 Satisfaction Performance
Panel A: Mean Patient Satisfaction by Year based on Initial Performance
Whole Sample Performance Quartile in 2004
Overall Inpatient Satisfaction Poor Performers High Performers
N Mean N Mean N Mean 2004 135 4.194 33 3.780 31 4.424 2005 135 4.328 33 4.143 31 4.471 2006 137 4.374 33 4.127 31 4.546 2007 137 4.393 33 4.083 31 4.536 2008 136 4.441 32 4.215 31 4.535 2009 134 4.405 31 4.031 31 4.530 2010 133 4.442 31 4.287 31 4.461 2011 135 4.448 31 4.260 31 4.558
Overall Outpatient Satisfaction 2004 144 3.951 36 3.664 36 4.244 2005 143 4.044 36 3.868 36 4.260 2006 144 4.072 36 3.907 36 4.204 2007 144 4.090 36 3.934 36 4.204 2008 143 4.115 36 3.985 36 4.215 2009 142 4.104 35 3.966 36 4.165 2010 142 4.115 36 3.986 35 4.201 2011 143 4.119 36 3.999 36 4.170
Panel B: Mean Annual Change in Yearly Satisfaction Scores
Whole Sample Performance Quartile in 2004 Overall Inpatient Satisfaction Poor Performers High Performers
N Mean N Mean N Mean 2005-2004 134 0.087*** 33 0.317*** 31 0.022*** 2006-2005 135 0.057*** 33 -0.016 31 0.074** 2007-2006 137 0.019 33 -0.044 31 -0.010 2008-2007 136 0.041*** 32 0.099 31 -0.001 2009-2008 134 -0.039* 31 -0.185** 31 -0.005 2010-2009 132 0.023 30 0.191** 31 -0.069 2011-2010 133 0.005 30 -0.054 31 0.098 Overall Outpatient Satisfaction 2005-2004 143 0.088*** 36 0.204*** 36 0.009 2006-2005 143 0.025** 36 0.039* 36 -0.060 2007-2006 144 0.018 36 0.027 36 0.000 2008-2007 143 0.024* 36 0.052* 36 0.012 2009-2008 142 -0.012 35 -0.020 36 -0.050 2010-2009 141 0.009 35 0.015 35 0.029 2011-2010 142 0.012 36 0.013 35 0.000
44
Notes to Table 3: (1) For expositional clarity and ease of interpretation of the means, these values are calculated as the average score on the ten survey questions assessing overall satisfaction for inpatients and outpatients respectively. Univariate analyses performed using overall satisfaction factor scores yielded equivalent results. (2) Hospitals are identified as poor performers (lowest quartile) and high performers (highest quartile) based on the satisfaction scores of 2004, which is the first year of the mandatory collection of satisfaction performance information. (3) In Panel B, change is defined as satisfaction in Yeart – Yeart-1. (4) Significance of the changes (year over year change is significantly different than zero) in Panel B: * = p-value <0.10; ** = p-value <0.05; *** = p-value <0.01, two-tailed.
45
TABLE 4
Analysis of Patient Satisfaction over Time
Inpatients Outpatients
Overall Inpatient
Satisfaction
Staff and Treatment
Logistics and infrastructure
Overall Outpatient Satisfaction
Staff and Treatment
Administrative Procedures
Size 0.102 0.111* -0.021 -0.043 0.031 -0.207*** [0.063] [0.067] [0.067] [0.064] [0.062] [0.053]
Competition 0.187*** 0.154*** 0.144** 0.124** 0.091* 0.05 [0.056] [0.057] [0.060] [0.058] [0.054] [0.049]
Hospital 0.418*** 0.450*** 0.205* -0.382*** -0.235** -0.833*** [0.109] [0.104] [0.119] [0.111] [0.103] [0.109]
2005 0.451*** 0.310*** 0.373*** 0.414*** 0.638*** 0.021 [0.103] [0.104] [0.090] [0.083] [0.066] [0.061]
2006 0.598*** 0.564*** 0.566*** 0.437*** 1.077*** 0.017 [0.093] [0.106] [0.088] [0.094] [0.067] [0.077]
2007 0.670*** 0.630*** 0.506*** 0.525*** 1.116*** 0.02 [0.088] [0.110] [0.096] [0.095] [0.076] [0.076]
2008 0.849*** 0.718*** 0.685*** 0.651*** 1.263*** 0.179** [0.098] [0.101] [0.091] [0.102] [0.071] [0.083]
2009 0.714*** 0.730*** 0.893*** 0.595*** 1.286*** 0.101 [0.101] [0.112] [0.100] [0.102] [0.081] [0.081]
2010 0.847*** 0.767*** 0.926*** 0.654*** 1.330*** 0.05 [0.127] [0.105] [0.116] [0.101] [0.065] [0.084]
2011 0.880*** 0.777*** 0.983*** 0.669*** 1.368*** 0.072 [0.109] [0.098] [0.100] [0.118] [0.090] [0.093]
Intercept -0.848*** -0.776*** -0.707*** -0.338*** -0.916*** 0.281*** [0.140] [0.140] [0.113] [0.119] [0.095] [0.100]
N 1029 1033 1033 1145 1143 1143 R2 0.156 0.167 0.132 0.096 0.212 0.255 Notes to Table 4: (1) Results are based on OLS estimation of the following model: Satisfaction= α+β1 Size+ β2 Competition+ β3 Hospital+ β
4-10Year+𝜀 . (2) Columns 1 and 4 list
estimated coefficients for the model where the dependent variable is overall satisfaction, for inpatients and outpatients, respectively. (3) Dependent variables for the estimations in columns 2, 3, 5, and 6 are operationalized by the corresponding satisfaction factors. The factor scores are based on principal component analysis with oblique rotation. Factor scores are expressed in standardized terms. (4) Size (number of beds, in thousands) and Competition (number of hospitals per 100K people in the prefecture) are expressed in standardized terms. (5) Hospital is a binary variable assuming the value of 1 if the site is a general hospital, and 0 if the site is a sanatorium. (6) Year 2004 is the dropped dummy. (7) Models control for hospital fixed effects and use robust standard errors clustered by firm. (4) * = p-value <0.10; ** = p-value <0.05; *** = p-value <0.01, two-tailed.
46
TABLE 5 Drivers of change in patient satisfaction
Panel A: Unscaled Change in Patient Satisfaction
Inpatients Outpatients
Overall Satisfaction
Staff & Treatment
Logistics & infrastructure
Overall Satisfaction
Staff & Treatment
Admin Procedures
Trend -0.038*** -0.020** -0.014 -0.020** -0.074*** 0.014 [0.010] [0.008] [0.010] [0.009] [0.009] [0.009]
Poor performer 0.463*** 0.695*** 0.672*** 0.569*** 0.434*** 0.454*** [0.170] [0.191] [0.165] [0.172] [0.108] [0.106]
Trend * Poor Performer
-0.072** -0.106*** -0.081** -0.085*** -0.062*** -0.058*** [0.033] [0.033] [0.032] [0.028] [0.022] [0.020]
Hospital 0.066*** 0.008 0.021 0.036 0.028 0.01 [0.024] [0.023] [0.027] [0.030] [0.020] [0.024]
Size 0.009 0.007 -0.004 0.029** 0.028** 0.004 [0.018] [0.010] [0.015] [0.014] [0.013] [0.012]
Competition 0.011 0.009 0.005 0.006 -0.002 0.002 [0.015] [0.008] [0.012] [0.012] [0.010] [0.012]
Intercept 0.222*** 0.145*** 0.130** 0.143*** 0.518*** -0.107** [0.056] [0.042] [0.059] [0.049] [0.045] [0.048]
N 568 855 855 998 984 984 R2 0.026 0.035 0.034 0.03 0.063 0.017 Panel B: Change in Patient Satisfaction Scaled by Initial Potential for Improvement Inpatient Outpatient
Overall Satisfaction
Staff & Treatment
Logistics & infrastructure
Overall Satisfaction
Staff & Treatment
Admin Procedures
Trend -0.013*** -0.005 -0.002 -0.002 -0.018*** 0.006** [0.004] [0.005] [0.004] [0.003] [0.002] [0.003]
Poor Performer 0.082** 0.163*** 0.137*** 0.083*** 0.050** 0.092*** [0.041] [0.046] [0.038] [0.021] [0.021] [0.020]
Trend * Poor Performer
-0.013 -0.027*** -0.017** -0.012*** -0.007* -0.012*** [0.008] [0.008] [0.007] [0.004] [0.004] [0.004]
Hospital 0.025*** 0.01 0.011 0.013** 0.006 0.005 [0.008] [0.009] [0.008] [0.005] [0.005] [0.005]
Size 0.005 0.004 -0.001 0.009* 0.008* 0.006 [0.005] [0.003] [0.005] [0.005] [0.005] [0.004]
Competition 0.006 0.005* 0.001 0.005 0.001 0.002 [0.004] [0.003] [0.004] [0.003] [0.003] [0.003]
Intercept 0.076*** 0.039 0.024 0.011 0.127*** -0.044*** [0.023] [0.027] [0.022] [0.016] [0.013] [0.016]
N 568 855 855 998 984 984 R2 0.020 0.011 0.013 0.016 0.046 0.012 Notes to Table 5: (1) Panel A uses a fixed effects regression model of the following form:
47
Change in Satisfaction = α + β1Trend + β2Poor Performer+ β3Trend*Poor Performer+ β4Hospital+ β5Size+ β6Competition+𝜀. (2) In Panel A change in satisfaction is satisfaction in Yeart – Yeart-1, while in Panel B, change in satisfaction is scaled by the potential for improvement relative to 2004: ((Satisfaction in Yeart – Satisfaction in Yeart-1)/(5 – Satisfaction in 2004)). (3) Columns 1 and 4 list estimated coefficients for the model where the dependent variable is overall satisfaction, for inpatients and outpatients, respectively. Dependent variables for the estimations in columns 2, 3, 5, and 6 are operationalized by the corresponding satisfaction factors. The factor scores are based on principal component analysis with oblique rotation. Factor scores are expressed in standardized terms. (4) Size (number of beds, in thousands) and Competition (number of hospitals per 100K people in the prefecture) are expressed in standardized terms. (5) Hospital is a binary variable assuming the value of 1 if the site is a general hospital, and 0 if the site is a sanatorium (6) Models control for firm fixed effects and use robust standard errors clustered by firm. (7) * = p-value <0.10; ** = p-value <0.05; *** = p-value <0.01, two-tailed.
48
TABLE 6 Analysis of the Relationship Between Patient Satisfaction Performance and Incentive Compensation, Hospital Grant, Inpatient and
Outpatient Revenues Salary Bonus Grant Inpatient Revenue Outpatient Revenue Overall Inpatient Satisfaction(t-1) -0.012 -0.005 -0.004** 0.250**
[0.009] [0.003] [0.002] [0.097] Overall Outpatient Satisfaction(t-1) -0.009 0.000 -0.001 0.072**
[0.013] [0.004] [0.002] [0.031] Size(t-1) 0.133*** 0.049*** 0.007 1.279*** 0.334***
[0.032] [0.008] [0.008] [0.114] [0.040] Competition(t-1) -0.010 0.005 0.001 0.008 -0.027**
[0.013] [0.003] [0.003] [0.044] [0.013] Hospital(t-1) 0.068 0.000 -0.020** 2.054*** 0.985***
[0.050] [0.012] [0.009] [0.252] [0.092] Total Medical Revenue(t-1) 0.225*** 0.052*** -0.009**
[0.023] [0.005] [0.004] Educational Revenue(t-1) 0.719* 0.077 0.043
[0.370] [0.080] [0.069] R&D Revenue(t-1) 0.515 0.131 -0.093
[0.388] [0.084] [0.082] R&D Expenses(t-1) -0.746
[0.798] Facility Costs(t-1) -0.015
[0.024] Material Costs(t-1) -0.013
[0.011] Total Costs(t-1) 0.024**
[0.012] Year Fixed Effects Yes Yes Yes Yes Yes
Intercept 0.533*** 0.221*** -0.013 -1.959*** -0.748*** [0.069] [0.017] [0.020] [0.591] [0.155]
N 886 886 885 888 995 R2 0.958 0.952 0.354 0.714 0.718
49
Notes to Table 6: (1) Robust standard errors are reported in brackets for each coefficient. (2) Patient satisfaction predictors are measured by overall satisfaction factor scores lagged one period for both inpatient and outpatient cases. The factor scores are based on principal component analysis with oblique rotation. Factor scores are expressed in standardized terms. (3) All regressions use robust standard errors and control for hospital fixed effects. (4) Size (number of beds, in thousands), Competition (number of hospitals per 100K people in the prefecture) and all revenue and costs predictors are expressed in standardized terms. (5) Hospital is a binary variable assuming the value of 1 if the site is a general hospital, and 0 if the site is a sanatorium. (6) * = p-value <0.10; ** = p-value <0.05; *** = p-value <0.01, two-tailed.