+ All Categories
Home > Documents > The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period...

The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period...

Date post: 14-Aug-2020
Category:
Upload: others
View: 2 times
Download: 0 times
Share this document with a friend
50
1 The Cost of Fraud Prediction Errors Messod D. Beneish Indiana University Kelley School of Business Patrick Vorst Maastricht University School of Business and Economics September 2019 Abstract We estimate the costs of fraud prediction errors from the perspective of auditors, investors, and regulators using an extensive sample of 140,233 observations (of which 779 are fraud firm-years) over the period 1982- 2016. We propose total costs of misclassification, which incorporates the costs of both false negatives (missed detections) and false positives (incorrectly identifying non-fraud firms), as a solid basis for comparing the economic value of fraud prediction models and we describe conditions under which these models are useful screening devices. We estimate auditors’ false negative costs as the sum of litigation costs, reputation costs, and client losses and find the mean (median) cost to be $14.1 ($1.94) million in 2016 dollars. We estimate auditors’ false positive costs absent resignation as the unbilled incremental audit investment and conditional on resignation as the lost audit fee in the year following the resignation and we find the mean (median) cost to be $0.258 ($0.084) million. Although false negative costs appear to dwarf false positive costs, with average and median cost ratios of 54.5 and 23.1, these costs ratios are themselves dwarfed by the proportion of non-fraud to fraud observations in the sample (179:1). Indeed, the fact that all prediction models have large numbers of false positives, and that even the best models trade false positives for negatives at the rate of 92:1 explains why auditors would avoid using fraud prediction models in practice. Investors’ false negatives costs (averaging $446.6 to $845.6 million) are 9 to 40 times larger than their false positive costs, the latter predominantly consisting of the abnormal return foregone by not investing in firms incorrectly identified as fraudulent. However, we show that as the number of false positives increases, investors’ use of fraud prediction models becomes a value-destroying proposition. For example, at a cut-off point of 1.0, the F-Score has a false positive rate of nearly 40% and results in a net loss to investors of over $50 billion. We estimate regulators’ false negative costs on average to range from $5.4 to $20.1 million, but we cannot estimate their false positive costs unless we assume that regulators systematically announce public investigations of flagged firms, in which case these costs become prohibitive. Overall, we document considerable variation in the cost of misclassification, both across models and across user groups. Furthermore, our evidence suggests that future researchers should focus on lowering the false positive rate of their models, because increasing the true positive rate appears like a futile pursuit. JEL classification: G31; G32; G34; M40 Keywords: Financial statement fraud, false positive, false negative, cost of errors We thank Colleen Honisberg for her help in implementing the litigation and settlement prediction models in Honisberg, Rajgopal, and Srinivasan (2018). We also thank Dan Amiram, Zahn Bozanic, and Ethan Rouen for giving us access to the code to compute the FSD measure.
Transcript
Page 1: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

1

The Cost of Fraud Prediction Errors

Messod D. Beneish

Indiana University

Kelley School of Business

Patrick Vorst

Maastricht University

School of Business and Economics

September 2019

Abstract

We estimate the costs of fraud prediction errors from the perspective of auditors, investors, and regulators

using an extensive sample of 140,233 observations (of which 779 are fraud firm-years) over the period 1982-

2016. We propose total costs of misclassification, which incorporates the costs of both false negatives (missed

detections) and false positives (incorrectly identifying non-fraud firms), as a solid basis for comparing the

economic value of fraud prediction models and we describe conditions under which these models are useful

screening devices. We estimate auditors’ false negative costs as the sum of litigation costs, reputation costs,

and client losses and find the mean (median) cost to be $14.1 ($1.94) million in 2016 dollars. We estimate

auditors’ false positive costs absent resignation as the unbilled incremental audit investment and conditional

on resignation as the lost audit fee in the year following the resignation and we find the mean (median) cost

to be $0.258 ($0.084) million. Although false negative costs appear to dwarf false positive costs, with

average and median cost ratios of 54.5 and 23.1, these costs ratios are themselves dwarfed by the proportion

of non-fraud to fraud observations in the sample (179:1). Indeed, the fact that all prediction models have

large numbers of false positives, and that even the best models trade false positives for negatives at the rate

of 92:1 explains why auditors would avoid using fraud prediction models in practice. Investors’ false

negatives costs (averaging $446.6 to $845.6 million) are 9 to 40 times larger than their false positive costs,

the latter predominantly consisting of the abnormal return foregone by not investing in firms incorrectly

identified as fraudulent. However, we show that as the number of false positives increases, investors’ use

of fraud prediction models becomes a value-destroying proposition. For example, at a cut-off point of 1.0,

the F-Score has a false positive rate of nearly 40% and results in a net loss to investors of over $50 billion.

We estimate regulators’ false negative costs on average to range from $5.4 to $20.1 million, but we cannot

estimate their false positive costs unless we assume that regulators systematically announce public

investigations of flagged firms, in which case these costs become prohibitive. Overall, we document

considerable variation in the cost of misclassification, both across models and across user groups.

Furthermore, our evidence suggests that future researchers should focus on lowering the false positive rate of

their models, because increasing the true positive rate appears like a futile pursuit.

JEL classification: G31; G32; G34; M40

Keywords: Financial statement fraud, false positive, false negative, cost of errors

We thank Colleen Honisberg for her help in implementing the litigation and settlement prediction models in

Honisberg, Rajgopal, and Srinivasan (2018). We also thank Dan Amiram, Zahn Bozanic, and Ethan Rouen for giving

us access to the code to compute the FSD measure.

Page 2: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

2

1. Introduction

Costs are central to many predictions in accounting, auditing, and financial economics as economic

agents trade off costs and benefits in making decisions.1 Yet, despite their important role, empirical

documentation of the magnitude of many costs has been scarce. This scarcity is even more acute

in the context of fraud prediction models because such models have a cost multiplier effect: the

costs vary by decision-maker, and predictions generate two types of error costs, one of which has

thus far not been studied. That is, while we have evidence of the cost of failing to detect an instance

of fraud (e.g., a false negative), the cost of incorrectly classifying a firm as a fraud (e.g., a false

positive) has not been investigated before. Consequently, the rates at which auditors, investors,

and regulators trade off the costs of Type I and II errors has remained in the domain of assumptions.

This paper attempts to fill this gap by investigating two questions related to the costs of fraud

prediction errors: (1) what is the nature and magnitude of such costs from the perspective of

auditors, investors, and regulators?; and (2) how do existing prediction models perform when the

comparisons take into account the cost such errors for the various decision-makers? Although we

cannot observe their objective functions, we conjecture that auditors seek to maximize their profits,

that investors seek to maximize their wealth, and that regulators (e.g., SEC, PCAOB) seek to

minimize wealth losses related to asymmetric and/or misleading information (e.g., incomplete or

inaccurate reports and disclosures). Our investigation of costs assumes that decision-makers

commit to taking actions consistent with the model’s predictions.

1 Among others, researchers invoke agency, disclosure, information, litigation, monitoring, political, proprietary,

regulatory, and reputation costs to develop hypotheses about the incentives and behavior of economic agents in various

contexts. Indeed, the three papers that first developed many of these costs concepts (Stigler 1971; Alchian and Demsetz

1972; Jensen and Meckling 1976) have been cited over 100,000 times in aggregate, according to a July 2019 Google

Scholar search.

Page 3: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

3

The motivation for our analysis is twofold. First, model predictions are based on

classification cut-offs, which depend on the relative magnitude of the costs of false negatives and

the costs of false positives. Although most studies assume that false negatives are more costly than

false positives, there is considerable disagreement as to how much more costly they are, and to

whom. Indeed, prior studies propose decision rules based on assumed cost ratios of false negatives

to false positives as disparate as 20:1 and 142:1, which makes choosing among a model’s cut-offs

and comparing models difficult (e.g., Beneish 1997; Beneish 1999; Cecchini et al. 2010; Dechow

et al. 2011; Perols et al. 2017).

Second, our analysis is of interest because we propose a cost-based method for comparing

the performance of fraud classifiers that addresses the deficiencies in two methods that have been

frequently used in previous literature to compare model performance (area under the curve,

hereafter AUC, and expected costs of misclassification, hereafter ECM). To explain, although the

measure based on AUC used in Cecchini et al. (2010) and Perols et al. (2017) is not threshold

dependent, it treats false positives and false negatives as equally costly (e.g., Adams and Hand

1999, Lobo et al. 2008). This leads to inaccurate conclusions about the relative performance of

extant fraud prediction models because, for all decision-makers, we systematically reject the null

hypothesis that false positives and false negatives are equally costly. In addition, when researchers

use expected costs of misclassification (ECM) to assess model performance, they assume that all

classification errors of a given type are equally costly. However, we document that there is

considerable variation in the costs of fraud across firms. Indeed, this is the case even among the

top 20 instances of fraud cases in Beneish, Lee, and Nichols (2013). For example, the loss on the

market—a component of investors’ cost of missed detection—is 15 to 20 times larger for frauds

at Enron and Waste Management compared to frauds at Sunbeam and Vivendi.

Page 4: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

4

Our results are based on a sample of 140,233 firm-years, of which 779 firm-years represent

firms that are subject to SEC enforcement actions (hereafter fraud firms) and 139,454 non-fraud

firm-years. The 779 fraud firm-years consist of 313 unique non-financial firms that were charged

with violations other than violations of the Foreign Corrupt Practices Act and for which data are

available on CRSP and COMPUSTAT. We use this sample to evaluate the ability of five models

to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based

on Kothari et al. (2005), the unexplained audit fee model (Hribar et al. 2014), the F-Score (Dechow

et al. 2011), and the measure of financial statement divergence based on how the distribution of

first digits differs from Benford’s Law (Amiram et al. 2015).2

We estimate that auditors’ costs of false negatives amount to $4.382 billion in aggregate (all

amounts are in 2016 dollars) for the 313 instances of fraud in our sample. We rely on prior research

to estimate three components of auditors’ costs of false negatives: litigation costs, reputation costs,

and the foregone profits associated with any unusual client losses following the public revelation

of undetected accounting fraud at one of the auditor’s clients.

We find that the auditors of 87 fraud firms were sued (27.8% of the sample). Of those 87

lawsuits, 23 (26.4%) were dismissed, and a further 15 (17.2%) were settled for zero damages,

leaving 49 cases with an aggregate settlement of $3.128 billion, a finding that is in line with

Honisberg, Rajgopal, and Srinivasan (2018). In terms of reputational losses, we follow Gleason et

al. (2008) and Weber et al. (2008) and estimate reputation cost as the price reaction around fraud

revelation for the other clients of the auditor in the same industry (i.e., the contagion effect). We

find that they represent $102.7 million in aggregate. Finally, we follow Lyon and Maher (2005)

2 Our comparison focuses on studies for which model parameters, data, and decision rules are readily available. Thus,

we do not compare models that rely on textual or tonal analyses such as Brown et al. (2018) who propose a model that

combines financial data with textual information, or models that use data mining techniques such as Cecchini et al.

(2010) who propose a 23-variable financial kernel based on support vector machines.

Page 5: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

5

and estimate the profits foregone due to abnormal client losses in the two years following the

revelation of the fraud and find that they represent in aggregate a cost of $1.151 billion.3

In terms of auditors’ false positives costs, our cost estimates are based on the actions we

expect auditors to take when prompted by a flag from a fraud prediction model. Specifically, we

expect that with 97% probability, the auditor would bear the cost of increasing its audit investment

and with 3% probability, the auditor would lose the annual audit fee by resigning from the audit

engagement. Indeed, previous literature has shown that auditors increase their audit investment

when they perceive a higher risk of misstatement (e.g., Glover et al. 2003; Hammersley et al. 2011;

Boritz et al. 2015) and has documented that risk factors such as a control weakness or the payment

of bribes result in higher fees (Hogan and Wilkins 2008; Lyon and Maher 2005; Munsif et al.

2011). Based on this research and on our analysis of audit fees in periods subsequent to remediation

of an internal control weakness, we use an estimate of 23% more audit work (with a range of 20%

to 28%) and calculate that the auditor bears average (median) costs of $0.258 ($0.084) million for

each false positive.

Collectively, our findings suggest that auditors’ false negative costs dwarf their false positive

costs, as the average cost ratios range from 26.8 to 54.5, and the median cost ratios range from 2.0

to 23.1.4 However, these costs ratios are themselves dwarfed by the proportion of non-fraud to

fraud observations, which we calculate as 179 to 1 (139,454/779). This could explain why auditors,

3 We conduct sensitivity analyses on client losses and litigation costs. On client losses, we vary the length of the period

over which we measure client loss (e.g., +1 v. +2 years), and in terms of litigations costs, we estimate an expected

settlement amount from the model proposed by Honisberg, et al. (2018). We find that these costs remain significant

but their magnitudes are lower ($828.2 million for client losses and $902.7 million for estimated litigation costs). 4 These ratios are likely overstated because we are unable to measure the effect of an increase in workload (23% for

all false positives) on either the productivity of the audit staff, or the quality of the audits. This increased workload

could create significant time pressure on all audits because prediction models generate a large number of false

positives, and, in the limit, could make it impossible for audit firms to make the extra audit investment necessary to

investigate all of the model’s alerts.

Page 6: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

6

on the advice of general counsel apprehensive about the risk of discovery of fraud alerts in auditor

working papers, would avoid the use of fraud prediction models with high false positive rates.

For investors, we estimate the costs false negatives as the abnormal market value loss from

day -1 relative to the fraud-revelation announcement to periods ending as early as day +1 and as

late as day +252 (12 months). Using market-adjusted returns, the average loss ranges from $446.6

million (over 3 days) to $845.6 million (over 12 months) and the corresponding medians are $20.6

million and $49.09 million. We estimate investors’ false positive costs as the sum of two

components: (1) the incremental audit fee which we assume is billed by the incumbent or required

by the newly hired auditor, and (2) the profit foregone (or minus the loss avoided) by not investing

over the next year in firms flagged by the prediction model as potentially fraudulent. The average

ratio of false negative to false positive costs ranges from 21.4 to 40.52 using market-adjusted

returns and from 4.73 to 8.95 when using returns that are risk-adjusted using the Fama-French

four-factor model.

For regulators, we argue that the cost of false negatives is the cost of losing investors’ trust

in the institution, which we estimate as the market-wide dollar abnormal returns in three periods

centered on the fraud-revelation announcement (day 0, days -1 to +1, and days -2, +2). On average,

we estimate that regulators false negative costs range between $5.4 and $20.1 million. In terms of

false positives, regulators’ costs depend on whether they internally or publicly investigate firms

flagged by the model. We cannot estimate regulators’ costs of internal investigations, which we

surmise include work overload and the investment of scarce resources in the pursuit of false targets.

However, if regulators’ make the investigations public, the costs are very large under the

assumption that falsely identified firms experience a market value loss in the range of 3% to 10%,

which is the typical market reaction to comment letters and revelations of SEC investigations and

Page 7: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

7

charges. In that case, the average ratio of false negative to false positive costs is systematically less

than one, thereby limiting the usefulness of fraud prediction models for regulators.

To evaluate the relative performance of the different models, we assess the true and false

positive performance by decile over the period 1982-2016 and find that the F-Score ranks first

with a top quintile (top four deciles) success rate of 47.5% (72.3%). This is followed by the M-

Score with a success rate of 40.4% (67.5%), and by current accruals with a success rate of 37.2%

(57.9%). Our analysis also reveals that all models have systematically large percentages of false

positives: for example, the false positive rate ranges from 19.94% to 19.99% if the top quintile of

the various models is used to classify firms as frauds and it ranges from 39.93% to 39.99% if firms

in the top four deciles are classified as frauds. In effect, the fact that all prediction models have

large numbers of false positives, and that even the best models trade false positives and false

negatives at a rate of approximately 92:1, explains why auditors avoid using fraud prediction

models in practice. As the number of false positives increases, also investors’ use of fraud

prediction models becomes a value-destroying proposition. For example, at a cut-off point of 1.0,

the F-Score has a false positive rate of nearly 40% and results in a net loss to investors of over $50

billion.

We further show that there is considerable variation in the cost of false negatives (an

untenable assumption of ECM analyses) and that the cost of false negatives and false positives

differs both within and across decision-makers (the former being an untenable assumption of most

AUC analyses), as well as across models. We propose total costs of misclassification as a basis for

comparing the economic value of fraud prediction models and describe conditions under which

these models are useful screening devices. In particular, our evidence suggests that researchers

Page 8: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

8

should focus on lowering the false positive rate of their models, because increasing the true positive

rate appears like a futile pursuit.

The remainder of the paper is structured as follows: Section 2 describes the fraud prediction

models we consider, Section 3 presents the sample and data, Section 4 describes and illustrates the

process by which we estimate costs for auditors, investors, and regulators, Section 5 compares the

predictive ability of various models, Section 6 discussed additional analyses, and Section 6

concludes.

2. Fraud Prediction Models

We evaluate the ability of six models to identify firms that are subsequently subject to SEC

accounting and enforcement actions: measures of raw accruals and abnormal accruals, the M-Score

(Beneish 1999), the unexplained audit fee model (Hribar et al. 2014), the F-Score (Dechow et al.

2011), and a measure of financial statement divergence based on how the distribution of first digits

differs from Benford’s Law (Amiram et al. 2015).

2.1 Accruals and Performance-Matched Accruals

We report three measures of accruals in our analyses. Total accruals, calculated from the statement

of cash flows as the difference between IBC and OANCF deflated by lagged assets, current

accruals calculated as total accruals plus depreciation deflated by lagged assets, and performance-

matched total abnormal accruals, estimated based on the Jones (1991) model as implemented by

Kothari et al. (2005). In addition to these three measures, we obtain qualitatively similar results

when we consider alternatives measures of raw accruals (drawn from Sloan (1996) and Hribar and

Collins (2002)) and alternative measures of abnormal accruals based on Dechow et al. (1995),

Beneish (1998), and Dechow and Dichev (2002).

Page 9: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

9

2.2 The Beneish M-Score

Beneish (1997; 1999) profiles firms that manipulate earnings (firms either charged with

manipulation by the SEC or that admit to manipulation in the public press) and develops a

statistical model to discriminate manipulators from non-manipulators. In this paper, we use the

unweighted probit model presented in Beneish (1999) which relies exclusively on financial

statement data and whose usefulness in assessing fraud potential out-of-sample has been shown by

academics and professionals.5 The M-Score below classifies a firm as a potential manipulator if

the M-Score exceeds -1.78, assuming that false negatives are 20 times more costly than false

positives (variables are defined in the Appendix):

M-SCORE = – 4.840 + 0.920DSRIit + 0.528GMIit + 0.404AQIit + 0.892SGIit

+0.115DEPIit – 0.172SGAIit + 4.679TATAit – 0.327LGVIit (1)

2.3 The Hribar et al. (2014) unexplained audit fee model

Hribar et al. (2014) draw on prior audit research to argue that audit fees reflect the

expected costs of misreporting (e.g., litigation, reputation) and that the portion of the audit fee

that is not explained by known determinants reflects in part an auditor’s expectation of the cost

of misreporting. They propose the residual from the following equation as a measure of

unexplained audit fees:

LogFeest = Industry Indicators + β1BigNt + β2log(Assetst) + β3Inventoryt/Avg Assetst

+β4Receivablest/Avg Assetst + β5LTDt/Avg Assetst + β6Earnt/Avg Assetst

+β7Losst + β8Qualifiedt + β9Auditor Tenuret + ɛt (2)

5 Since the publication of the original study (which used data from 1982 to February 1993), the model has been

featured in financial statement analysis textbooks (e.g., Fridson and Alvarez 2011) and in articles directed at auditors,

certified fraud examiners, and investment professionals (e.g., Ciesielski 1998, Merrill Lynch 2000, Wells 2001, DKW

2003, Harrington 2005). The model gained notoriety when a group of MBA students at Cornell University posted the

earliest warning about Enron’s accounting manipulation score using the Beneish (1999) model a full year before the

first professional analyst reports (Morris 2009). This episode in American financial history is preserved in the Enron

exhibit at Museum of American Finance, New York (www.moaf.org) and is also recounted in Gladwell (2009).

Page 10: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

10

In this model, higher unexplained audit fees are associated with poorer accounting quality.

However, Hribar et al. (2014) do not directly study classification in a sample of fraud cases (and

thus do not provide a threshold with which to classify firms). Consequently, we sort on

unexplained audit fees and evaluate the model’s detective performance by focusing on the top

quintile.

2.4 The Dechow et al. (2011) F-Score

Dechow et al. (2011) follow a methodology similar to Beneish (1997; 1999) in developing a score

to detect accounting fraud and estimate three alternative models that differ in whether the models

include returns and/or non-financial statement data (e.g., number of employees, security issuance).

The results are similar across models, and we use the following version of F-Score (variables

defined in the Appendix):

F-SCORE = - 6.789 + 0.817RSST + 3.230∆REC + 2.436∆INV + 0.122∆Cash Sales

-0.992∆Earnings + 0.972ACT Issuance (3)

Dechow et al. (2011) suggest three potential cut-offs for classifying firms as frauds depending on

whether the F-score exceeds either 1.0, 1.85, and 2.45, which correspond to assumed costs ratios

of false negatives to false positives of 143:1, 86:1, and 82:1, respectively.

2.5 The Amiram et al. (2015) FSD Score

Amiram et al. (2015) construct an FSD Score by comparing the distribution of first digits in over

100 financial statement items relative to Benford’s law, as research in different disciplines has

used deviations from Benford’s distribution to detect errors or manipulation in data. The FSD

Score is based on the mean absolute deviation statistic (MAD) for the financial items considered

and is calculated as MAD = (∑|AD-ED|)/K, where AD (ED) is the actual (expected) proportion of

Page 11: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

11

leading digits and K is the number of leading digits being analyzed.6 Amiram et al. (2015) suggest

that the FSD score is a useful tool for detecting financial statement errors, as they document that

lagged FSD scores positively correlate with material misstatements while contemporaneous FSD

scores are negatively correlated with material misstatements. Although they suggest that the FSD

score can be used as a leading indicator of material misstatements, they do not report an FSD score

threshold with which to classify firms. As a result, we sort on FSD score and evaluate the model’s

detective performance by focusing on the top quintile.

3. Data

3.1 Sample

Our evaluation of the performance of extant fraud classifiers is based on a comprehensive sample

drawn from Accounting and Auditing Enforcement Actions (AAER) by the Securities and

Exchange Commission (SEC) between April 1982 and July 2016 (AAER#1 to AAER #3793). We

identify accounting enforcement actions against 574 firms after eliminating multiple and

unassigned AAERs (2351), those related to financial institutions (319), auditing actions against

independent CPAs (280), enforcement actions for the payment of bribes under the Foreign Corrupt

Practices Act (112), and those related to related to violations in 10-Qs resolved within the fiscal

year (131).

Table 1 reports the selection of the final sample. The main sample consists of 574 fraud cases

over the period 1982-2016, of which we are able to match 492 cases to the Compustat-CRSP

merged database.7 Those 492 cases relate to 1,185 firm-years with misstated financial statements.

6 We are thankful to Dan Amiram, Zahn Bozanic, and Ethan Rouen for giving us access to the code to extract leading

digits and to compute the FSD Score. 7 Many of those cases can be matched to Compustat, but do not have CRSP return data available to calculate the stock

market reaction to the revelation of the fraud. As we use the price reaction as one of the key measure of fraud costliness,

we drop these observations from the sample and continue with firms that are matched to both Compustat and CRSP.

Page 12: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

12

We then drop firms with missing returns around the revelation date of the fraud, for example,

because the firm delisted prior to the date on which the fraud is revealed, leaving a sample of 413

fraud cases (1,041 firm-years). As we are interested in determining the usefulness of fraud

prediction models, we further drop observations for which we do not have data to compute the

fraud prediction models we compare. Overall, the sample restrictions lead to a final sample of 313

fraud cases involving 779 misstatement years.

3.2 Backfilling Missing Audit Fees

As we later discuss, audit fees are important to estimate the costs of fraud prediction errors from

an auditors’ perspective. Given that Audit Analytics reports audit fee data from fiscal year 2000

onwards, we estimate an audit fee model in order to backfill audit fees for the earlier years in our

sample. The regression specification we estimate follows Hribar et al. (2014) and uses a wide set

of audit fee determinants, capturing the size and complexity of the client and the audit engagement.

We estimate the regression over the period 2000-2001 as those years precede the enactment of the

Sarbanes Oxley Act, which likely makes them more relevant to predicting audit fees in the pre-

2000 period.

The results of the estimation, which we report in Table 2, Panel A, are in line with

expectations and closely resemble those in Hribar et al. (2014). That is, BigN auditors are able to

charge a fee premium and bigger and more complex clients; those with a higher asset base, a

greater number of business segments, or more foreign sales, command higher audit fees. In

contrast, less risky firms with higher current ratios, lower market-to-book ratios, and better

performance (higher ROA, no losses) pay lower audit fees. Finally, firms that receive a modified

Page 13: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

13

audit opinion also pay higher fees. We use these coefficient estimates in conjunction with actual

data in Compustat to predict and backfill (missing) audit fees in those earlier years.8

Although the regression results are in line with expectations and the 75.9% explanatory

power suggests that we can obtain reasonable predictions of audit fees, we conduct one further test

to assess the reliability of our audit fee prediction model. Specifically, we compare predicted fees

over the period 2002-2017 with actual fees as reported in Audit Analytics over the same period.

Although ultimately we are interested in predicting fees over the pre-2000 period for which no

Audit Analytics data are available, we can use the period 2002-2017 to estimate the out-of-sample

accuracy of our fee prediction model. The Pearson (Spearman) correlation between (the natural

logarithm of) estimated fees and actual fees is 0.87 (0.86), suggesting that our model performs

well in predicting the (relative) magnitude of audit fees.9

4. The Cost of Classification Errors

We estimate the cost of classification errors for three decision-makers that have incentives to rely

on screening models to detect fraud: auditors, investors, and regulators. The costs that we are

interested in estimating are the costs of false negatives, e.g., the costs of missed fraud detections

or actual frauds not flagged by a model, and the costs of false positives, e.g., the costs of incorrectly

flagging a non-fraud firm as a fraud firm. These costs differ across decision-makers and although

we cannot observe their objective functions, we conjecture that their overarching objectives are as

8 We rely on Compustat data only, including for measuring BigN and OPIN, as we need Compustat data to fill in

missing audit fees over the period for which Audit Analytics was not yet available. 9 However, we find that our estimate of audit fees exhibits a negative bias over the period 2002-2017 (the natural

logarithm of actual fees over the period 2002-2017 is 13.66, while predicted fees are on average 0.87 lower at 12.79).

This is not surprising as prior research has shown a significant increase in average fees following SOX (Ettredge,

Scholz, and Li 2007; Ghosh and Pawlewicz 2009). Although these results suggest that we underestimate audit fees

over the later part of our sample, these results do not necessarily translate to the pre-2000 period, which is also pre-

SOX.

Page 14: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

14

follows: auditors seek to maximize their profits, investors seek to maximize their wealth, and

regulators seek to minimize wealth losses relating to asymmetric and misleading information (e.g.,

incomplete or inaccurate disclosures and reports). Figure 1 summarizes the major components of

misclassification costs that we estimate for auditors, investors, and regulators.

4.1 Auditors’ Costs of False Positives

To the extent that auditors view a flag from a fraud screening model as an indication of fraud risk,

they have two possible courses of action. First, auditors can increase the nature and extent of the

audit investment to ascertain and report on whether fraud has indeed occurred. Second, they can

resign from the client and forego their typical profit margin on the lost audit fee. Thus, for any

given model, we can estimate auditor i’s expected false positive costs (EAUD [FP COST]) for a

given client j that is falsely flagged in year t as:

EAUDi[FP COST] = p (AUDFEEjt * pm%) + (1-p) (INCRAUDFEEjt * (1-pm%)) (4)

Where p is the probability of resignation, and resignation results in the auditor losing its usual

profit margin (pm%) on the audit fee billed to client j in year t (AUDFEEjt). In the absence of

resignation, the auditor undertakes additional work costing INCRAUDFEEjt * (1-pm%), which we

assume that they cannot pass on to the client in year t. Equation (4) does not take into account the

potential negative effects of time pressure on the quality of the auditor’s output (e.g., Bills et al.

2016). Time pressure could increase because of the additional audit investment and auditors’

limited ability to increase hiring and training in the short run.

4.1.1 Incremental Audit Fees

At least since the early 20th Century, auditors have been trained to increase the amount of audit

work they undertake when they perceive that their client’s financial statements have a higher

Page 15: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

15

likelihood of containing material misstatements (e.g., see Montgomery 1916). There is extensive

evidence in audit research that auditors respond to a high risk of fraud primarily by changing the

nature and/or extent of audit procedures, increasing the overall audit investment, and charging

higher fees for prospective clients with an element of fraud risk (see Hammersley et al. 2011 for a

recent review of this literature).

We argue that auditors who observe a flag from a fraud-screening model for one of their

clients, undertake an extra investment as part of the audit to ascertain the risk of material

misstatement, and thus reduce litigation and discovery risk.10 If auditors can bill for the additional

work, they bear no direct financial costs from false positives. In that case, the costs of the additional

audit effort are borne by investors as residual claimants. Alternatively, if clients do not accept to

pay for the extra audit effort, the cost is borne by auditors in the first year in which a firm is flagged.

We draw on prior experimental audit judgment research and on empirical audit fee research

to estimate the incremental audit costs that are borne either by the auditor or by investors. For

example, after the issuance of SAS 82 in early 1997, data in Table 3, page 245 of Glover et al.

(2003) suggest that auditors budget 21.8% more audit hours in high fraud risk engagements. More

recently, experimental evidence in Boritz et al. (2015) indicates that auditors increase audit budgets

by 20.4% in the presence of fraud risks. Because it is not possible to directly observe what auditors

would do in terms of audit effort when they perceive an increased likelihood of fraud, we follow

10 Discovery is the risk of having audit documentation contain a record of a fraud alert for a continuing client that is

subsequently found to have committed fraud. Plaintiffs’ attorneys can then argue on discovery that auditors had

advance knowledge of the fraud and were grossly negligent in conducting their audit. This risk likely seemed

significant as Arthur Andersen undertook to shred paper documents and destroy emails related to Enron, which led to

its indictment for obstruction of justice and its subsequent demise. On January 25, 2002, the Wall Street Journal

reported that in recovering deleted e-mails at Arthur Andersen, Congress found evidence that the Chicago office of

Arthur Andersen had issued two “alerts” to the Houston office in the spring of 2001 concerning earnings manipulation

at Enron. The alerts came from a tailored version of the M-Score model that Beneish had estimated under a consulting

relationship with Andersen. (“Andersen Knew of `Fraud' Risk at Enron --- October E-Mail Shows Firm Anticipated

Problems Before Company's Fall”, 01/25/2002, A3).

Page 16: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

16

prior work on audit fees and view the payment of bribes to foreign government officials and the

existence of internal control weaknesses as fraud risk factors that would lead auditors to increase

their audit investment. The evidence in Lyon and Maher (2005) suggests that audit fees are higher

by 43% for firms reporting paying bribes to foreign government officials. Lyon and Maher

conclude that when auditors assess higher business risks, they pass those costs to their clients.

In terms of empirical work on internal control weaknesses (ICW) on audit fees, we follow

Hogan and Wilkins (2008), Raghunandan and Rama (2006), and Munsif et al. (2011) to estimate

the behavior of audit fees before revelation of an ICW and after the revelation of the ICW,

conditional on whether the latter was remediated or not. Specifically, we augment the audit fee

model by incorporating indicators capturing the existence of an ICW and its remediation status.

N1Y_ICW is an indicator variable that is equal to one if the firm reports an internal control

weakness in the following fiscal year, and zero otherwise. To investigate the impact of

remediation, we include L1Y_ICW_RMD (L1Y_ICW_NRMD), which is an indicator variable that

is equal to one if the firm reported an internal control weakness two years prior to the current fiscal

year that was (was not) remediated in the year prior to the current fiscal year, and zero otherwise.

L1Y_ICW_NEW is an indicator variable that is equal to one if the firm reported a new internal

control weakness in the year prior to the current fiscal year. ICW_RMD, ICW_NRMD, and

ICW_NEW are defined analogously, but are shifted one year ahead.11 We further add the level of

the SEC enforcement budget as a measure capturing the likelihood that fraud, if committed and

going undetected by the auditor, is subsequently uncovered by the SEC. We use two variables

capturing enforcement intensiveness; (1) the natural logarithm of the SEC budget in constant

11 ICW_RMD (ICW_NRMD) is an indicator variable that is equal to one if the firm reported an internal control

weakness in the year prior to the current fiscal year that was (was not) remediated in the current fiscal year, and zero

otherwise. ICW_NEW is an indicator variable that is equal to one if the firm reports a new internal control weakness

in the current fiscal year.

Page 17: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

17

dollars, and (2) the SEC budget divided by the number of firms in Compustat. The second measure

takes into account that the SEC budget has to be divided over a constantly changing number of

firms and measures the resources of the SEC relative to the number of firms they have to monitor.

The results are reported in Table 3. We estimate that firms with an ICW pay 18% to 24%

higher fees in the year prior to the reporting of an ICW, which is slightly less than in Hogan and

Wilkins (2008) who report 41% higher audit fees. In the year in which the ICW is first reported,

we find 31% higher fees (ICW_NEW, columns 10-12), while non-remediated internal control

weaknesses lead to 59% higher fees (ICW_NRMD, columns 10-12). Interestingly, from the

coefficient on ICW_RMD (L1Y_ICW_RMD) we find that even in the year of (following)

remediation, fees are 34% (24%-27%) higher. All results are similar after including proxies for

SEC monitoring. Moreover, the positive and significant coefficients on L1Y_LN_SECBUD and

L1Y_SC_SECBUD in all specifications, indicate that increased SEC enforcement activity, and thus

higher fraud detection risk, is associated with higher audit fees. Collectively this evidence suggests

that if auditors increase their investment in the audit after observing a flag from a fraud screening

model, investors bear a cost that ranges from 20% to 30% of lagged audit fees. Alternatively, if

auditors cannot pass on those costs to their clients, they bear costs that are equal to those amounts

multiplied by 1 less their typical profit margin.

In Table 4, Panel A we present statistics on audit fees and on the auditors’ expected false

positive costs for 139,455 non-fraud firm-years in the sample. In 2016 dollars, the average and

median audit fees over the period 1980-2016 are equal to $1.282 and $0.415 million, and the

average (median) false positive costs is $0.259 ($0.084) million. We compute these estimates using

audit fees drawn from Audit Analytics for the period 2000-2016, and audit fees estimated using

the backfilling model we describe in Table 2 for observations in the period 1980-1999. In addition,

Page 18: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

18

we assume that the likelihood of resignation is 3%, that incremental fees amount to 23% [range

20-28%] of the contemporaneous audit fee, and, relying on recent reports by the BIG4 in the U.K.,

that audit firms’ profit margins equal 23% [range 20-30%].

4.2 Auditors Costs of False Negatives

4.2.1 Auditor Litigation Costs

Prior research has long recognized that the costs of potential litigation influence audit pricing and

the design of financial reporting (e.g., Simunic 1980, Simunic and Stein 1996, Palmrose 1998, Lys

and Watts 1994). We undertake a comprehensive search of litigation against firms, investment

bankers, and auditors for the 313 fraud firms in the sample. We match firms to litigation cases by

reading lawsuits and resolution notices on Audit Analytics for the post-1999 period and we rely

on Westlaw, Lexis-Nexis, and Factiva searches in the pre-2000 period.12

We tabulate the results of our analysis in Table 4, Panel B. There are lawsuits against 212

fraud firms (67.7%), of which 32 (15.1%) are dismissed and 14 (6.6%) are settled with no damages.

In aggregate, fraud firms have paid $25.9 billion to settle the remaining 166 suits. Investment

bankers paid $20.8 billion to settle 16 out of the 19 lawsuits filed against them in connection with

fraud firms. We also find that the auditors of 87 fraud firms were sued (27.8% of the sample). Of

those 87 lawsuits, 23 (26.4%) were dismissed, and a further 15 (17.2%) were settled for zero

damages, leaving 49 cases with an aggregate settlement of $3.128 billion. Our settlement total for

auditors is in line with that of Honisberg, Rajgopal, and Srinivasan (2018) who identify 540

lawsuits naming auditors over the period 1996-2016 and estimate the aggregate value of auditor

settlement payments at $3.53 billion. As we later report, we also estimate lawsuit settlements

12 Our searches begin with the firm name at the time the fraud is discovered, but we also consider name changes as

applicable, and specify a search period beginning two years before and ending two years after the fraud period

identified in AAERs, and use search terms such as “securities”, “class action”, litigation”, “lawsuit” and “settlement.”

Page 19: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

19

against auditors using Honisberg, Rajgopal, and Srinivasan (2018) model that explains settlement

amounts against auditors (Ln(1+Settle)):13

Ln(1 + Settle) = −8.174 + 0.750 ∗ 𝐶𝑙𝑎𝑠𝑠_𝑃𝑒𝑟𝑖𝑜𝑑_𝐿𝑒𝑛𝑔𝑡ℎ − 0.189 ∗ 𝑆ℎ𝑎𝑟𝑒_𝑇𝑢𝑟𝑛𝑜𝑣𝑒𝑟 + 0.040 ∗𝑃𝑟𝑖𝑐𝑒_𝐷𝑟𝑜𝑝 + 0.867 ∗ 𝐺𝑟𝑜𝑤𝑡ℎ + 0.448 ∗ 𝐿𝑛(𝐴𝑠𝑠𝑒𝑡𝑠) + 2.265 ∗ 𝑅𝑂𝐴 + 2.519 ∗ 𝐻𝑖𝑔ℎ_𝐿𝑖𝑡𝑖𝑔𝑎𝑡𝑖𝑜𝑛 +1.782 ∗ 𝐸𝑞𝑢𝑖𝑡𝑦_𝐼𝑠𝑠𝑢𝑎𝑛𝑐𝑒 − 4.044 ∗ 𝐼𝑃𝑂 + 3.356 ∗ 𝑅𝑒𝑠𝑡𝑎𝑡𝑒𝑑 − 1.921 ∗ 𝑃𝑜𝑠𝑡_𝑁𝑖𝑛𝑒_𝐸𝑙𝑒𝑣𝑒𝑛 − 5.165 ∗𝑃𝑜𝑠𝑡_𝑇𝑤𝑜_𝑇ℎ𝑟𝑒𝑒 + 2.135 ∗ 𝑁𝑖𝑛𝑒_𝐸𝑙𝑒𝑣𝑒𝑛 + 4.879 ∗ 𝑇𝑤𝑜_𝑇ℎ𝑟𝑒𝑒 + 2.712 ∗ 𝑃𝑜𝑠𝑡_𝑇𝑒𝑙𝑙𝑎𝑏𝑠 (5)

4.2.2 Auditor Reputation Losses

Franz et al. (1998), Gleason et al. (2008), and Weber et al. (2008) provide evidence that

announcements that reveal either litigation against auditors or the existence of financial statement

fraud at a firm are associated with negative spillover effects on other clients of the auditor. For

example, Weber et al. (2008) find that other clients of KPMG Germany suffered abnormal returns

of -3% around the three events revealing the accounting scandal at another KPMG auditee in

Germany (ComROAD AG). Similarly, Gleason et al. (2008) document a contagion effect and

suggest investors re-assess their reliance on the financial statements of non-restating firms in the

same industry.

13 Honisberg et al. document a declining role of Section 10(b) in class actions against auditors, and provide evidence

on how two Supreme Court rulings in 2007 and 2011 altered the likelihood of a successful lawsuit outcome as a

function of the circuit court in which the litigation proceeds. We assume that the auditors of AAER firms are sued

with probability one and estimate litigation costs as the expected settlement amount using coefficients estimates in

Honisberg, et al. In addition, as an alternative estimation of equation (5) Honisberg et al. use 2008-2012 rather than

2002-2012 as an analysis period. Their goal is to provide evidence on the effect on the 2011 Supreme Court ruling in

Janus v. First Derivative, which made it more difficult to bring cases against auditors in circuit courts in district four

and nine. We also consider the model below as an alternative estimate of settlement costs for our post-2007 sample

(All the variables are described in Honigsberg et al.’s Appendix).

𝐿𝑛(𝑆𝑒𝑡𝑡𝑙𝑒 + 1) = 13.280 − 0.699 ∗ 𝐶𝑙𝑎𝑠𝑠_𝑝𝑒𝑟𝑖𝑜𝑑_𝑙𝑒𝑛𝑔𝑡ℎ − 0.148 ∗ 𝑆ℎ𝑎𝑟𝑒_𝑇𝑢𝑟𝑛𝑜𝑣𝑒𝑟 − 0.067 ∗ 𝑃𝑟𝑖𝑐𝑒_𝐷𝑟𝑜𝑝

+ 2.186 ∗ 𝐺𝑟𝑜𝑤𝑡ℎ − 0.267 ∗ 𝐿𝑛(𝐴𝑠𝑠𝑒𝑡𝑠) + 6.665 ∗ 𝑅𝑂𝐴 ∗ 11.089 ∗ 𝐻𝑖𝑔ℎ_𝐿𝑖𝑡𝑖𝑔𝑎𝑡𝑖𝑜𝑛 − 5.609

∗ 𝐸𝑞𝑢𝑖𝑡𝑦_𝐼𝑠𝑠𝑢𝑎𝑛𝑐𝑒 − 6.424 ∗ 𝐼𝑃𝑂 + 1.662 ∗ 𝑅𝑒𝑠𝑡𝑎𝑡𝑒𝑑 − 6.808 ∗ 𝑃𝑜𝑠𝑡_𝐽𝑎𝑛𝑢𝑠_𝐹𝑜𝑢𝑟_𝑁𝑖𝑛𝑒

− 3.743 ∗ 𝐹𝑜𝑢𝑟_𝑁𝑖𝑛𝑒 + 5.371 ∗ 𝑃𝑜𝑠𝑡_𝐽𝑎𝑛𝑢𝑠

Page 20: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

20

In Table 4, Panel C, we use the contagion effect to measure auditor reputational losses.

Specifically, when using the average price reaction in the three days surrounding the fraud-

revealing announcement of the other clients of the auditor in the same industry as the fraud firm

(6-digit GICS), we find that the average reputation loss is equal to $0.37 million in 2016 dollars.

Whereas data from COMPUSTAT provide us with the largest sample coverage among AAER

firms (N=274), we also conduct our analyses relying on the Audit Analytics auditor switching file

and find the average reputation loss for the 139 observations with sufficient data is equal to $2.76

million in 2016 dollars.

4.2.3 Auditor Client Losses

We follow Lyon and Maher (2005) and measure auditor client losses as the difference between an

auditor’s rate of attrition at the office level in the year following the discovery of a missed detection

and (i) the rate of attrition for the same ‘undetecting’ office in the year prior to the public revelation

of the fraud, or (ii) the average rate of attrition for all other of the auditor’s offices. We estimate

the cost of client losses by multiplying the abnormal number of clients lost and the average audit

fee per client. Relying on Compustat auditor data, we find that the average loss due to client losses

is equal to $2.65 million for comparisons in the year after the fraud revelation and $3.68 million

for comparisons in the two year-period after revelation. Our estimates of costs are lower when

relying on Audit Analytics, with average losses of $1.38 million for the one-year comparison and

$2.12 million for the two-year comparison.

4.3 Investors’ Costs

We estimate investors’ false positive costs as the sum of two components: (1) the incremental audit

fee which we assume is billed by the incumbent auditor or required by the newly hired auditor,

and (2) the profit foregone (or minus the loss avoided) by not investing over the next year in firms

Page 21: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

21

flagged by the prediction model as potentially fraudulent. We report estimates of these costs in

Table 5, Panel A. The average incremental fee for all non-fraud firms amounts to $0.295 million

in 2016 dollars, whereas non-fraud firms’ dollar abnormal returns range from $15.6 million (size-

adjusted abnormal returns) to $94.2 million (Fama-French 4-factor model).

We posit that investors’ cost of missed detections (false negatives) is the investment loss

associated with the discovery of the misreporting. This follows prior research that has documented

that the revelation of restatements and fraud is typically associated with highly adverse abnormal

price reactions. In terms of AAERs, Beneish (1999) and Karpoff et al. (2008) document three-day

losses of -20% and -25%. For firms that announce restatements due to irregularities, the loss

amounts to between -15% and -25% in the three to six month period after the restatement becomes

public (Badertscher, Collins, and Lys 2008; Hennes, Leone, and Miller 2008). In Table 5, Panel B

we report estimates of the costs of missed detection (false negatives) as the abnormal market value

loss from day -1 relative to the fraud-revealing announcement to periods ending as early as day +1

and as late as day +252 (12 months). Using market-adjusted returns we find that the average loss

ranges from $446.6 million (over 3 days) to $845.6 million (over 12 months) and the corresponding

medians are $20.6 million and $49.09 million.

Collectively, the results on the costs of false positives and false negatives to investors suggest

an average ratio of false negative to false positive costs that ranges from 21.4 to 40.52. This implies

that, ceteris paribus, as the number of false positives increases, investors’ use of fraud prediction

models becomes a value-destroying proposition.14

14Although prior studies commonly implement short and/or hedge strategies, the availability of stocks for borrowing

in order to short sell makes the implementation of short selling on a large scale difficult, if not prohibitively costly

(See Beneish, Lee and Nichols 2015). For this reason, we do not consider the possibility of short selling as a means

of potentially reducing investors’ false positive costs.

Page 22: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

22

4.4 Regulator’s Costs

In terms of false positives, regulators’ costs depend on whether they internally or publicly

investigate firms flagged by the model. We cannot estimate regulators’ costs of internal

investigations, which we surmise include work overload and the investment of scarce resources in

pursuit of false targets. However, if regulators’ make the investigations public, the costs are very

large under the assumption that falsely identified firms experience a market value loss in the range

3 to 10%, which is the typical market reaction to comment letters and revelations of SEC

investigations and charges. As we report in Table 6, Panel A, these costs are substantial, averaging

between $84.6 and $281.9 million.

We argue that regulators’ cost of false negatives is the cost of losing investors’ trust in the

institution, which we estimate as the market-wide dollar abnormal returns in three periods centered

on the fraud-revealing announcement (day 0, days -1 to +1, and days -2, +2). Specifically, we

estimate a regression of CRSP daily value-weighted market returns on an indicator variable that is

equal to one on the fraud revelation date and zero on days on which there is no fraud revealed. The

results are reported in Table 6, Panel B. Whereas we find significant results only in the regressions

with the one-day event window, the negative and significant coefficient on EVENT indicates that

market-wide returns are on average lower on fraud revelation dates. This result is consistent with

investors being more pessimistic about stocks in general, an effect that we attribute to a loss of

trust in the regulator’s ability to uncover fraud. In terms of economic significance, we estimate

that regulators’ false negative costs range from $5.4 million to $20.1 million. The average ratio of

false negative to false positive costs is systematically less than one, limiting the usefulness of fraud

prediction models if regulators publicly announce investigations based on alerts from the fraud

prediction models.

Page 23: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

23

5. Model Comparisons

5.1 Predictive Ability

In Table 7, we assess the performance of fraud predictions based on total and current accruals,

performance-matched abnormal current accruals, M-Score, F-Score, and the FSD measure over

the period 1982-2016.15 Specifically, we create decile ranks of each of the measures and investigate

the true positive and false positive rates in the top deciles. The table reveals that models have

systematically a large percentage of false positives: for example, the false positive rate ranges from

19.94% to 19.99% if the top quintile of the various models is used to classify firms as frauds and

it ranges from 39.93% to 39.99% if firms in the top four deciles are classified as frauds. In terms

of true positive rates, the F-Score ranks first with a top quintile (top four deciles) success rate of

47.5% (72.3%), followed by the M-Score with a success rate of 40.4% (67.5%), and current

accruals with a success rate of 37.2% (57.9%).

We focus our remaining analyses on comparing the M-Score and F-Score, because it is

possible for decision-makers to implement these models by relying on the cut-offs indicated in the

published studies. Figure 2 relates the rates of false positives and false negatives for two M-Score

and three F-Score cut-offs reported by Beneish (1999) and Dechow et al. (2011). With a threshold

of -1.78, the M-Score identifies 35.7% of frauds with a false positive rate of 16.7%, while with a

threshold of -2.00, the M-Score identifies 43.9% of the frauds, but the false positive rate increases

to 21.8%.16 Dechow et al. (2011) identify three cut-offs for the F-Score (2.45, 1.85, and 1.00) that

are associated with increasing true positive rates (17.8%, 30.2%, and 72.0%) and correspondingly

15 We omit the unexplained audit fee from this comparison because we do not have actual audit fees before 2000. We

re-estimate Table 7 on the post-2000 period for which we have unexplained audit fee data. Whereas the results on the

other measures are quantitatively and qualitatively similar, we find limited predictive ability of unexplained audit fees,

with a true positive rate of 23% in the top quintile. 16 Note that the -1.78 M-Score is the cut-off that minimizes costs of misclassification in the estimation sample in

Beneish (1999). The -2.00 cut-off is sometimes used for illustration purposes to delineate a ‘grey’ area, but is not

based on minimizing misclassification costs.

Page 24: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

24

increasing false positive rates (4.0%, 9.3%, and 39.6%). What is noteworthy is that the five points

seem nearly perfectly aligned, and indeed, the R2 of a line fitted through these points equals 98.8%,

which we interpret as indicating that both models trade off true and false positives at a similar rate.

The tradeoff rate is approximately 92 to 1 suggesting that on average for every correctly identified

fraud, the models identify 92 false positives. If frauds were all equally costly and the costs of false

alerts were the same for all non-fraud firms, an approximately constant tradeoff would imply

discretionary model and cut-off choices. However, costs do differ both within and across fraud

and non-fraud groups. For this reason, we turn to evaluating the costs and benefits of evaluating

each model at their suggested cut-offs.

5.2 Costs and Benefits of Implementing Decision Rules Based on F-Score and M-Score

We analyze the costs and benefits of implementing decision rules based on the F-Score and the M-

Score in Table 8. With 1.00 as a cut-off for the F-Score, auditors avoid costs of $3.235 billion from

correctly identifying 226 frauds, but incur false positive costs of $16.297 billion from the 55,212

falsely identified non-fraud observations. Auditors’ net costs of over $13 billion suggest this

particular cut-off is costly. The results are similar for investors who can avoid $96 billion in fraud

related losses, but who forego profits of nearly $157 billion by not investing in flagged firms for a

net cost of nearly $51 billion. For regulators, the results systematically suggest the avoidance of

fraud prediction models, a finding we attribute to the fact that we cannot estimate regulators false

positive costs unless we assume that they go public with an investigation for any model-based

alert.

With 1.85 as a cut-off for the F-Score, auditors avoid costs of $1.876 billion from correctly

identifying 95 frauds but incur false positive costs of $2.077 billion from the 12,940 falsely

identified non-fraud observations. Auditors’ net costs of over $202 million suggesting this cut-off

Page 25: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

25

is also costly. On the other hand, investors can avoid $70 billion in fraud related loss and $877

billion of losses by not investing in falsely flagged firms for a net benefit of nearly $947 billion.

Finally, with a cut-off of 2.45 for the F-Score, auditors avoid costs of $0.797 billion from correctly

identifying 56 frauds and incur false positive costs of $0.668 billion from the 5,505 falsely

identified non-fraud observations, resulting in a net benefit of $0.202 billion. As well, investors

can avoid $12.7 billion in fraud-related loss and $892 billion losses by not investing in falsely

flagged firms for a net benefit of nearly $904 billion.

With regard to the M-Score, using the cut-off that minimizes misclassification costs in

Beneish (1999) (e.g., -1.78), auditors avoid costs of $1.516 billion from correctly identifying 112

frauds but incur false positive costs of $2.681 billion from the 23,320 falsely identified non-fraud

observations. Auditors’ net costs of 1.165 billion suggest that using the M-Score is too costly. On

the other hand, investors can avoid $18 billion in fraud related loss, and $1,692 billion losses by

not investing in falsely flagged firms for a net benefit of over $1.71 trillion.

In sum, auditors likely have a strong aversion to using these fraud prediction models because

the models’ false positive rates are too high. The only instance where auditors stand to benefit is

when the false positive rate is 3.95%, but that implies a success rate of only 17.83%. On the other

hand, investors can benefit from using these models so long as they avoid the low cut-off for the

F-Score. Investors’ largest benefit occurs from using the M-Score to avoid firms facing headwinds

as documented in Beneish, Lee and Nichols (2013). 17

6. Model Comparisons

6.1 Conditional Performance of Models

17 Indeed, the gains from the false positives suggest that, although the falsely flagged firms did not engage in fraud

that was uncovered by the SEC, the models do seem to capture potentially underperforming firms.

Page 26: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

26

In this section, we discuss the results of additional analyses in which we investigate the

performance of M-Score and F-Score at different cutoffs, conditional on variables that are likely

associated with committing fraud and the probability of detecting fraud. Specifically, we

investigate the models’ performance conditional on quintiles of firm size, firm age, and signed

total accruals, as well as firm life cycle.18 The results can be found in Table 9. We report the total

number of firms in a quintile and/or life cycle stage, as well as the number of firms that are (falsely)

flagged. We then use these numbers to determine the hit rates (the percentage of fraud firms

correctly flagged), false positive rates (the percentage of non-fraud firms incorrectly flagged as

fraudulent), the number of fraud firms as a percentage of all firms that are flagged, and the implied

cost ratio with which models trade off true and false positives.

With respect to firm size, we find that hit rates are generally highest for smaller firms.

However, false positive rates are also highest for smaller firms, such that on a net basis relatively

few fraud firms are correctly flagged in the quintiles containing the smallest firms. Not

surprisingly, the implied cost ratios show a monotonic decrease across quintiles of firm size.

Hence, although hit rates are lower for larger firms, the fact that the number of false positives is

decreasing more substantially in the larger firm quintiles, makes that both M-Score and F-Score

show better performance for larger firms. Surprisingly, we find opposite results for firm age. M-

Score and F-Score perform better for younger firms, both when it comes to hit rates and the rate

at which they trade off true and false positives.

With respect to accruals, both models perform best in the quintiles with extreme (both

positive and negative) accruals. Hit rates are generally highest in the lowest and highest accrual

18 The sample for the life cycle analyses is smaller as the cash flow-based life cycle proxy from Dickinson (2011) is

not available prior to 1987.

Page 27: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

27

deciles. However, false positive rates are also much higher in quintiles of extreme accruals, such

that the implied cost ratios show relatively limited variation across quintiles of total accruals.

Interesting results obtain when looking at the results conditional on firm life cycle as both

models perform worst for stable, mature, firms. Hit rates are extremely small for mature firms and

even though false positive rates are small as well, the rate at which the models trade off true and

false positives is worst for mature stage firms. Whereas hit rates are highest for the “extreme”

introduction and decline stage firms, the relative performance of the models is highest for growth

firms, as the implied cost ratio is lowest for firms in these stages.

In short, the results reported in this section present evidence of considerable variation in

model performance conditional on firm size, firm age, accruals, and firm life cycle. Models seem

to work best for larger and younger firms, firms with extreme accruals, and firms in the growth

stage. As costs are also likely to vary across these conditions, these results can provide important

insights into the costs and benefits of fraud prediction models across different subsamples.

7. Conclusion

We estimate the costs of fraud detection errors from the perspective of auditors, investors, and

regulators using an extensive sample of 140,233 observations (of which 779 are fraud firm-years)

over the period 1982-2016. Auditors’ false negative costs (costs of missed detections), estimated

as the sum of litigation costs, reputation costs, and client losses are on average (median) $14.1

($1.94) million in 2016 dollars. Auditors’ false positive costs (costs of incorrectly identifying non-

frauds firms) estimated absent resignation as unbilled incremental audit investment or, conditional

on resignation, as the lost audit fee for one-year, equal on average (median) $0.258 ($0.084)

million. Although false negative costs appear to dwarf false positive costs, with average and

median cost ratios of 54.5 and 23.1, these costs ratios are themselves dwarfed by the proportion of

Page 28: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

28

non-fraud to fraud observations in the sample (179:1). Indeed, the fact that all prediction models

have large numbers of false positives and that even the best models trade false positives and false

negatives at a rate of 92:1 explains why auditors would avoid using fraud prediction models in

practice.

Investors’ false negative costs (averaging $446.6 to $845.6 million) are 9 to 40 times larger

than their false positive costs, which predominantly consists of the profit foregone by not investing

over the next year in firms incorrectly identified as fraudulent. As the number of false positives

increases, investors’ use of fraud prediction models becomes a value-destroying proposition. For

example, at a cut-off point of 1.0, the F-Score has a false positive rate of nearly 40% and results

in a net loss to investors of over $50 billion. We estimate regulators’ false negative costs on average

to range from $5.4 to $20.1 million, but we cannot estimate their false positive costs unless we

assume that regulators systematically announce public investigations of flagged firms, in which

case regulators’ false positive costs become prohibitive.

We show that there is considerable variation in the cost of false negatives (an untenable

assumption of ECM analyses) and that the cost of false negatives and false positives differ both

within and across decision-makers (the former being an untenable assumption of most AUC

analyses). We propose the total cost of misclassification, which incorporates the differential

costliness of false positives and false negatives across models, as a basis for comparing the

economic value of fraud prediction models and we describe conditions under which these models

are useful screening devices. In particular, our evidence suggests that researchers should focus on

lowering the false positive rate of their models because increasing the true positive rate appears

like a futile pursuit.

Page 29: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

29

References

Adams, N. M., & Hand, D. J. (1999). Comparing classifiers when the misallocation costs are

uncertain. Pattern Recognition, 32(7), 1139-1147.

Alchian, A. A., & Demsetz, H. (1972). Production, information costs, and economic organization.

The American Economic Review, 62(5), 777-795.

Amiram, D., Bozanic, Z., & Rouen, E. (2015). Financial statement errors: Evidence from the

distributional properties of financial statement numbers. Review of Accounting Studies,

20(4), 1540-1593.

Badertscher, B. A., Collins, D. W., & Lys, T. Z. (2012). Discretionary accounting choices and the

predictive ability of accruals with respect to future cash flows. Journal of Accounting and

Economics, 53(1-2), 330-352.

Beneish, M. D. (1997). Detecting GAAP violation: Implications for assessing earnings

management among firms with extreme financial performance. Journal of Accounting and

Public Policy, 16(3), 271-309.

Beneish, M. D. (1998). Discussion of “Are accruals during initial public offerings opportunistic?”.

Review of Accounting Studies, 3(1), 209-221.

Beneish, M. D. (1999). The detection of earnings manipulation. Financial Analysts Journal, 55(5),

24-36.

Beneish, M. D., Lee, C. M., & Nichols, D. C. (2013). Earnings manipulation and expected returns.

Financial Analysts Journal, 69(2), 57-82.

Beneish, M. D., Lee, C. M., & Nichols, D. C. (2015). In short supply: Short-sellers and stock

returns. Journal of Accounting and Economics, 60(2-3), 33-57.

Bills, K. L., Swanquist, Q. T., & Whited, R. L. (2016). Growing pains: Audit quality and office

growth. Contemporary Accounting Research, 33(1), 288-313.

Boritz, J. E., Kochetova-Kozloski, N., & Robinson, L. (2015). Are fraud specialists relatively more

effective than auditors at modifying audit programs in the presence of fraud risk? The

Accounting Review, 90(3), 881-915.

Brown, N. C., Crowley, R. M., & Elliott, W. B. (2018). What are you saying? Using topic to detect

financial misreporting. Working Paper

Cecchini, M., Aytug, H., Koehler, G. J., & Pathak, P. (2010). Detecting management fraud in

public companies. Management Science, 56(7), 1146-1160.

Ciesielski, J. (1998). What’s Happening to the Quality of Assets? Analyst’s Accounting Observer,

7(3), 19.

Dechow, P. M., & Dichev, I. D. (2002). The quality of accruals and earnings: The role of accrual

estimation errors. The Accounting Review, 77(s-1), 35-59.

Dechow, P. M., Ge, W., Larson, C. R., & Sloan, R. G. (2011). Predicting material accounting

misstatements. Contemporary Accounting Research, 28(1), 17-82.

Dechow, P. M., Sloan, R. G., & Sweeney, A. P. (1995). Detecting earnings management. The

Accounting Review, 70(2), 193-225.

Dresdner Kleinwort Wasserstein, Inc. (2003). Earnings Junkies. London: Global Equity Research

(October 29).

Ettredge, M. L., Scholz, S., & Li, C. (2007). Audit fees and auditor dismissals in the Sarbanes-

Oxley era. Accounting Horizons, 21(4), 371-386.

Page 30: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

30

Franz, D. R., Crawford, D., & Johnson, E. N. (1998). The impact of litigation against an audit firm

on the market value of nonlitigating clients. Journal of Accounting, Auditing & Finance,

13(2), 117-134.

Fridson, M. S., & Alvarez, F. (2011). Financial statement analysis: a practitioner's guide (Vol.

597): John Wiley & Sons.

Ghosh, A., & Pawlewicz, R. (2009). The impact of regulation on auditor fees: Evidence from the

Sarbanes-Oxley Act. Auditing: A Journal of Practice & Theory, 28(2), 171-197.

Gladwell, M. (2009). What the Dog Saw and Other Adventures. New York: John Wiley & Sons.

Gleason, C. A., Jenkins, N. T., & Johnson, W. B. (2008). The contagion effects of accounting

restatements. The Accounting Review, 83(1), 83-110.

Glover, D. P., Schultz, J., & Zimbelman, M. (2003). A comparison of audit planning decisions in

response to increased fraud risk: Before and after SAS No. 82. Auditing: A Journal of

Practice & Theory, 22(3), 237-251.

Hammersley, J. S., Johnstone, K. M., & Kadous, K. (2011). How do audit seniors respond to

heightened fraud risk? Auditing: A Journal of Practice & Theory, 30(3), 81-101.

Harrington, C. (2005). Analysis of Ratios for Detecting Financial Statement Fraud. Fraud

Magazine (March-April), 24-27.

Hennes, K. M., Leone, A. J., & Miller, B. P. (2008). The importance of distinguishing errors from

irregularities in restatement research: The case of restatements and CEO/CFO turnover.

The Accounting Review, 83(6), 1487-1519.

Hogan, C. E., & Wilkins, M. S. (2008). Evidence on the audit risk model: Do auditors increase

audit fees in the presence of internal control deficiencies? Contemporary Accounting

Research, 25(1), 219-242.

Honigsberg, C., Rajgopal, S., & Srinivasan, S. (2018). The changing landscape of auditor

litigation and its implications for audit quality. Working Paper

Hribar, P., & Collins, D. W. (2002). Errors in estimating accruals: Implications for empirical

research. Journal of Accounting Research, 40(1), 105-134.

Hribar, P., Kravet, T., & Wilson, R. (2014). A new measure of accounting quality. Review of

Accounting Studies, 19(1), 506-538.

Jensen, M. C., & Meckling, W. H. (1976). Theory of the firm: Managerial behavior, agency costs

and ownership structure. Journal of Financial Economics, 3(4), 305-360.

Jones, J. J. (1991). Earnings management during import relief investigations. Journal of

Accounting Research, 29(2), 193-228.

Karpoff, J. M., Lee, D. S., & Martin, G. S. (2008). The cost to firms of cooking the books. Journal

of Financial and Quantitative Analysis, 43(3), 581-611.

Kothari, S. P., Leone, A. J., & Wasley, C. E. (2005). Performance matched discretionary accrual

measures. Journal of Accounting and Economics, 39(1), 163-197.

Lobo, J. M., Jiménez‐Valverde, A., & Real, R. (2008). AUC: a misleading measure of the

performance of predictive distribution models. Global Ecology and Biogeography, 17(2),

145-151.

Lys, T., & Watts, R. L. (1994). Lawsuits against auditors. Journal of Accounting Research,

32(Supplement), 65-93.

Merrill Lynch. (2000). Financial Reporting Shocks (March 31).

Montgomery, R. H. (1916). Auditing: Theory and Practice, Second Edition. New York: The

Ronald Press Company.

Page 31: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

31

Munsif, V., Raghunandan, K., Rama, D. V., & Singhvi, M. (2011). Audit fees after remediation

of internal control weaknesses. Accounting Horizons, 25(1), 87-105.

Palmrose, Z.-V. (1988). 1987 Competitive Manuscript Co-Winner: An analysis of auditor

litigation and audit service quality. The Accounting Review, 63(1), 55-73.

Perols, J. L., Bowen, R. M., Zimmermann, C., & Samba, B. (2017). Finding needles in a haystack:

Using data analytics to improve fraud prediction. The Accounting Review, 92(2), 221-245.

Raghunandan, K., & Rama, D. V. (2006). SOX Section 404 material weakness disclosures and

audit fees. Auditing: A Journal of Practice & Theory, 25(1), 99-114.

Simunic, D. A. (1980). The pricing of audit services: Theory and evidence. Journal of Accounting

Research, 18(1), 161-190.

Simunic, D. A., & Stein, M. T. (1996). Impact of litigation risk on audit pricing: A review of the

economics and the evidence. Auditing, 15, 119.

Sloan, R. G. (1996). Do stock prices fully reflect information in accruals and cash flows about

future earnings? The Accounting Review, 71(3), 289-315.

Stigler, G. J. (1971). The theory of economic regulation. The Bell Journal of Economics and

Management Science, 2(1), 3-21.

Weber, J., Willenborg, M., & Zhang, J. (2008). Does auditor reputation matter? The case of KPMG

Germany and ComROAD AG. Journal of Accounting Research, 46(4), 941-972.

Wells, J.T. (2001). Irrational Ratios. Journal of Accountancy, 192(2), 80-83

Page 32: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

32

Appendix

Variable Definitions

M-Score Variables

M-SCORE = – 4.840 + 0.920DSRIit + 0.528GMIit + 0.404AQIit + 0.892SGIit

+0.115DEPIit – 0.172SGAIit + 4.679TATAit – 0.327LGVIit

DSRI = day’s sales receivable index = (ARt/REVt)/(ARt-1/REV t-1);

GMI = gross margin index

= [(REVt-1- Cost of Goods Soldt-1)/REVt-1]/[(REVt - Cost of Goods Soldt)/REVt];

AQI = asset quality index

= (1 - [Current Assetst + PPEt]/ATt)/(1 - [Current Assets t-1 +PPEt-1]/ATt-1);

SGI = sales growth index = REVt /REVt-1;

DEPI = depreciation index

= (Depreciationt-1/[Depreciationt-1 + PPEt-1])/(Depreciationt/[Depreciation + PPE]);

SGAI = sales, general, and administrative expenses index

= (SGA Expenset/REVt)/(SGAt-1/REVt-1);

TATA = total accruals to total assets = (IBCt-CFOt)/ATt; and

LGVI = leverage index

= ([Long-Term Debtt+Cur. Liabt]/ATt)/([Long-Term Debt t-1+Cur.Liabt-1]/ATt-1).

Unexplained audit fee model

LogFeest = Industry Indicators + β1BigNt + β2log(Assetst) + β3Inventoryt/Avg Assetst

+β4Receivablest /Avg Assetst + β5LTDt /Avg Assetst + β6Earnt /Avg Assetst

+β7Losst + β8Qualifiedt + β9Auditor Tenuret + ɛ

LogFeest = natural logarithm of total audit fees; Industry Indicators = separate indicators for each two-digit SIC code; BigNt = dichotomous variable equal to 1 if a Big N auditor is used, and 0

otherwise; Earnt = operating income after depreciation; Losst = dichotomous variable equal to 1 if a loss occurred within the current or

previous two fiscal years, and 0 otherwise; Qualifiedt = dichotomous variable equal to 1 if the audit opinion was anything

other than a standard unqualified opinion, and 0 if the opinion was a standard unqualified opinion; and

Auditor Tenuret = number of years the auditor has been auditing a company.

F-Score Variables

FSCORE = - 6.789 + 0.817RSST + 3.230∆REC + 2.436∆INV + 0.122∆Cash Sales

-0.992∆Earnings + 0.972 ACT Issuance

Page 33: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

33

RSST = (∆WC + ∆NCO + ∆FIN)/Average Total Assets, where WC = [Current Assets - Cash and

Short-Term Investments] - [Current Liabilities - Debt in Current Liabilities]; NCO = [Total

Assets - Current Assets - Investments and Advances - [Total Liabilities - Current Liabilities -

Long-Term Debt]; and FIN = [Short-Term Investments + Long-Term Investments] - [Long-Term

Debt + Debt in Current Liabilities + Preferred Stock]; following Richardson et al. (2005);

∆REC = ∆Accounts Receivables/Average Total Assets;

∆INV = ∆Inventory/Average Total Assets;

∆CASH SALES = percentage change in cash sales [Sales - ∆Accounts Receivables];

∆EARNINGS = [Earningst/Average Total Assetst] - [Earningst-1/Average Total Assetst-1]; and

ACT Issuance = indicator variable coded 1 if the firm issued securities during year t.

Page 34: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

34

Figure 1: Estimating the Cost of Fraud Classification Errors

Auditors Investors Regulators Incremental audit work assuming it cannot be

billed to audit client--AT COST. This follows

experimental evidence that auditors increase the audit

investment when thy perceive a higher risk of

mistatement (e.g., Glover et al. 2003, Hammersley et

al. 2011 for a review, Boritz et al. 2015), and

empirical evidence that a risk factor such as a control

weakness or the payment of bribes results in higher

fees (Hogan and Wilkins 2008, Lyon and Maher

2008, and Raghunandan et al. (2011).

Incremental audit work assuming it is billed to

audit client--AT MARKET. This follows

experimental evidence that auditors increase the audit

investment when thy perceive a higher risk of

mistatement (e.g., Glover et al. 2003, Hammersley et

al. 2011 for a review, Boritz et al. 2015), and

empirical evidence that a risk factor such as a control

weakness or the payment of bribes results in higher

fees (Hogan and Wilkins 2008, Lyon and Maher

2008, and Raghunandan et al. (2011).

If investigations into false positives

are made public: Costs estimates

range from 3 to 10% of the market

value of equity at the end of month

three subsequent to firms fiscal year

ends. Percentages encompass three-

day average losses associated with

initial announcements of Wells

notices, restatements involving

irregularities, and SEC

investigations.

Alternatively, one year of lost audit fees from

resigning from the audit client (or profit or such fees).

Loss avoided (profit foregone) by not investing in

false positive firms over a 12-month period beginning

three months after fiscal year end.

If investigations into false positives are

for internal investigations, the costs of

use and alloaction of the regulators'

resources cannot be estimated.

Ligitation Costs: Predicted Settlement Costs

Conditional on the likelihood of auditor litigation

following Honisberg, Rajgopal, and Srinivasan (2018)

Reputation Losses: Market value loss at fraud

revelation due to contagion effect on other clients of

the auditor in the same industry (e.g., Gleason et al.

2008; Weber et al. 2008)

Client Loss: Profits foregone on Abnormal Client

Loss in the two years following the revelation of the

fraud (e.g., Lyon and Maher 2004)

False Positive Costs

(Incorrectly Flagged Non-

Fraud FIrms)

Figure 1--Estimating the Cost of FraudClassification Errors

False Negatives Costs

(Missed Fraud Detection)

Loss on the market: Abnornal Returns in period

varying from days -1 to +1 to days -1 to +252

relative to the first public revelation of the fraud times

the stock's market value on day -2.

The difference in value-weighted market

returns on revelation days over the

period 1982-2016 times the value of the

stock market in aggregate two days prior

to the fraud-revealing announcements.

Page 35: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

35

Figure 2: True and False Positive Rates by Prediction Model

Page 36: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

36

Figure 3: False Positive versus True Positive Rates

Page 37: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

37

TABLE 1

Sample Selection

# AAERs # Firm-Years

Total Sample of Fraud Cases 574 Less: Fraud Cases Not Matched to Compustat-CRSP (82)

Fraud Cases Matched to Compustat-CRSP 494 1185

Less: Fraud Cases with Missing Announcement Returns (79) ( 144)

Fraud Cases with Announcement Returns 415 1041

Less: Fraud Cases with Missing Mscore or Fscore (100) (262)

Final Sample of Fraud Cases 313 779

Page 38: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

38

TABLE 2

Audit Fee Model used for Backfilling

Panel A: Audit Fee Model

Depvar: LnFees

Variables Coef. T-stat p-value

Intercept 9.915 265.84 <.001

BigN 0.095 4.62 <.001

LnAssets 0.365 57.48 <.001

BUSSEG 0.089 7.31 <.001

FGN 0.696 23.20 <.001

INV 0.105 2.52 0.012

REC 0.369 7.47 <.001

CR -0.040 -19.74 <.001

BTM -0.035 -5.36 <.001

LEV -0.159 -6.06 <.001

EMPLS 0.097 17.40 <.001

MERGER 0.015 0.86 0.388

DEC_YE -0.037 -2.72 0.007

ROA -0.233 -7.29 <.001

LOSS 0.170 11.45 <.001

OPIN 0.107 6.51 <.001

Litigation -0.017 -1.19 0.235

Adj R2 0.759

N 7311

This table reports the results of the audit fee prediction regression used to backfill missing audit fees in the pre-2000

/ pre-audit analytics period. We estimate the regression over the period 2000-2001 and use the coefficients from this

regression and actuals reported in Compustat to create a predicted fee measure for the earlier years. T-statistics and p-

values are based on standard errors clustered at firm level. The dependent variable LnFees is the natural logarithm of

the audit fee as reported by audit analytics. BigN is an indicator variable that is equal to one if the firm is audited by a

BigN auditor (Compustat AU). LnAssets is the natural logarithm of total assets (AT). BUSSEG is equal to the number

of business segments as reported in Compustat or zero otherwise. FGN is foreign sales as a percentage of total sales

(SALE). INV is inventory as a percentage of total assets (INVT/AT). REC is accounts receivable as a percentage of

total assets (RECT/AT). CR is the current ratio (ACT/LCT). BTM is the book-to-market ratio (CEQ/PRCC_F*CSHO).

LEV is total debt to total assets (DLTT+DLC/AT). EMPLS is the square root of the total number of employees

(EMP^0.5). MERGER is an indicator variable that is equal to one if the Compustat footnote disclosures show that the

firm was involved in a merger, and zero otherwise. DEC_YE is equal to one if the firm's fiscal year-end is not in

December. ROA is operating income after depreciation divided by beginning-of-year total assets (OIADP/AT). LOSS

is an indicator variable that is equal to one if income before extraordinary items was negative in the current year or

any of the previous two years (IB). OPIN is an indicator variable that is equal to one if the firm does not receive an

"unqualified" audit opinion, zero otherwise. Litigation is an indicator variable that is equal to one if the firm is active

in a litigious industry (sic codes 2833:2836, 8731:8734, 3570:3577, 7370:7374, 3600:3674, 5200:5961).

Page 39: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

39

TABLE 3

Internal Control Weaknesses, SEC Budgets, and Audit Fees

Variables LnFees LnFees LnFees LnFees LnFees LnFees LnFees LnFees LnFees LnFees LnFees LnFees

(1) (2) (3) (4) (5) (6) (7) (8) (9) (10) (11) (12)

Intercept 8.950*** 9.663*** 5.508*** 10.574*** 10.193*** 9.535*** 10.556*** 10.169*** 9.578*** 10.150*** 10.136*** 9.539***

(48.93) (80.53) (29.56) (20.60) (72.09) (42.70) (20.82) (71.21) (43.03) (23.62) (75.60) (46.54)

N1Y_ICW 0.237*** 0.182*** 0.194***

(11.67) (8.91) (9.51) L1Y_ICW_RMD 0.237*** 0.246*** 0.245*** 0.263*** 0.270*** 0.269***

(14.53) (15.00) (14.96) (15.48) (15.93) (15.89) L1Y_ICW_NRMD 0.485*** 0.496*** 0.494***

(9.91) (10.11) (10.10) L1Y_ICW_NEW 0.310*** 0.321*** 0.320***

(15.46) (16.07) (16.03) ICW_RMD 0.335*** 0.342*** 0.341***

(20.73) (21.34) (21.31)

ICW_NRMD 0.592*** 0.591*** 0.591***

(12.67) (12.74) (12.73)

ICW_NEW 0.313*** 0.307*** 0.306***

(15.41) (15.26) (15.23)

L1Y_SC_SECBUD 2.540*** 0.427*** 0.411*** 0.421***

(20.50) (3.53) (3.45) (3.81) L1Y_LN_SECBUD 0.676*** 0.106*** 0.096*** 0.097***

(29.95) (4.03) (3.69) (4.07)

BigN 0.325*** 0.289*** 0.311*** 0.334*** 0.333*** 0.333*** 0.334*** 0.332*** 0.332*** 0.346*** 0.343*** 0.343***

(17.19) (15.27) (16.47) (15.79) (15.76) (15.79) (15.92) (15.87) (15.88) (17.82) (17.73) (17.75)

LnAssets 0.475*** 0.477*** 0.473*** 0.466*** 0.465*** 0.465*** 0.463*** 0.463*** 0.463*** 0.464*** 0.464*** 0.464***

(67.31) (67.62) (67.08) (61.20) (61.22) (61.28) (61.29) (61.31) (61.38) (64.71) (64.78) (64.85)

BUSSEG 0.029*** 0.028*** 0.029*** 0.019* 0.015 0.015 0.018* 0.014 0.014 0.019* 0.015 0.015

(2.83) (2.69) (2.81) (1.84) (1.47) (1.47) (1.76) (1.41) (1.40) (1.92) (1.55) (1.55)

FGN 0.185*** 0.184*** 0.180*** 0.183*** 0.173*** 0.173*** 0.175*** 0.166*** 0.167*** 0.166*** 0.158*** 0.158***

(7.04) (6.99) (6.84) (6.59) (6.28) (6.28) (6.38) (6.10) (6.11) (6.41) (6.14) (6.15)

INV 0.204*** 0.184*** 0.184*** 0.184** 0.175** 0.174** 0.178** 0.168** 0.168** 0.189*** 0.178*** 0.178***

(3.22) (2.88) (2.88) (2.56) (2.43) (2.42) (2.51) (2.36) (2.35) (2.84) (2.67) (2.66)

Page 40: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

40

REC 0.316*** 0.316*** 0.315*** 0.348*** 0.339*** 0.338*** 0.343*** 0.332*** 0.331*** 0.318*** 0.306*** 0.305***

(5.35) (5.34) (5.32) (5.08) (4.94) (4.93) (5.04) (4.88) (4.87) (5.09) (4.90) (4.89)

CR -0.027*** -0.028*** -0.028*** -0.025*** -0.026*** -0.026*** -0.025*** -0.025*** -0.025*** -0.025*** -0.025*** -0.025***

(-11.51) (-11.64) (-11.76) (-8.94) (-9.10) (-9.10) (-8.77) (-8.92) (-8.93) (-9.76) (-9.90) (-9.91)

BTM -0.116*** -0.098*** -0.097*** -0.104*** -0.098*** -0.097*** -0.104*** -0.098*** -0.097*** -0.107*** -0.099*** -0.099***

(-13.44) (-11.43) (-11.29) (-12.83) (-12.22) (-12.15) (-12.95) (-12.34) (-12.29) (-13.76) (-13.06) (-13.01)

LEV -0.132*** -0.141*** -0.144*** -0.116*** -0.106*** -0.107*** -0.110*** -0.101*** -0.102*** -0.118*** -0.111*** -0.111***

(-5.27) (-5.59) (-5.69) (-4.21) (-3.86) (-3.88) (-4.02) (-3.71) (-3.72) (-4.69) (-4.38) (-4.39)

EMPLS 0.052*** 0.051*** 0.052*** 0.052*** 0.053*** 0.053*** 0.054*** 0.055*** 0.055*** 0.054*** 0.055*** 0.055***

(10.10) (9.77) (10.09) (9.87) (9.99) (10.00) (10.20) (10.30) (10.30) (10.47) (10.55) (10.54)

MERGER 0.032*** 0.039*** 0.036*** 0.040*** 0.040*** 0.040*** 0.045*** 0.045*** 0.045*** 0.043*** 0.043*** 0.043***

(3.34) (3.89) (3.64) (3.80) (3.80) (3.80) (4.24) (4.25) (4.26) (4.34) (4.35) (4.35)

DEC_YE -0.016 0.004 0.000 0.009 0.011 0.011 0.008 0.010 0.010 0.006 0.009 0.009

(-0.99) (0.27) (0.03) (0.55) (0.66) (0.68) (0.48) (0.61) (0.63) (0.40) (0.55) (0.55)

ROA -0.299*** -0.284*** -0.272*** -0.307*** -0.306*** -0.305*** -0.304*** -0.303*** -0.303*** -0.281*** -0.280*** -0.280***

(-9.67) (-9.17) (-8.82) (-8.03) (-8.03) (-8.01) (-8.02) (-8.01) (-8.01) (-8.39) (-8.39) (-8.39)

LOSS 0.192*** 0.183*** 0.184*** 0.185*** 0.181*** 0.181*** 0.173*** 0.170*** 0.170*** 0.171*** 0.169*** 0.169***

(18.18) (17.16) (17.37) (15.40) (15.13) (15.17) (14.55) (14.36) (14.40) (15.33) (15.19) (15.23)

OPIN 0.121*** 0.160*** 0.160*** 0.102*** 0.125*** 0.123*** 0.097*** 0.115*** 0.112*** 0.099*** 0.119*** 0.116***

(13.47) (20.29) (20.91) (10.14) (14.18) (14.46) (9.84) (13.23) (13.41) (10.91) (15.34) (15.43)

CLIENT 0.020** 0.035*** 0.020*** 0.017** 0.017** 0.017** 0.025*** 0.026*** 0.027*** 0.019** 0.021*** 0.021***

(2.50) (4.54) (2.63) (1.99) (2.02) (2.06) (3.03) (3.19) (3.28) (2.44) (2.71) (2.82)

LITIGATION -0.060** -0.056** -0.057** -0.071*** -0.070*** -0.070*** -0.070*** -0.070*** -0.070*** -0.062** -0.061** -0.062**

(-2.41) (-2.27) (-2.30) (-2.67) (-2.65) (-2.65) (-2.66) (-2.65) (-2.65) (-2.49) (-2.46) (-2.46)

8.950*** 9.663*** 5.508*** 10.574*** 10.193*** 9.535*** 10.556*** 10.169*** 9.578*** 10.150*** 10.136*** 9.539***

Industry FE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes

Year FE Yes No No Yes No No Yes No No Yes No No

Observations 43,762 43,762 43,762 33,837 33,837 33,837 33,837 33,837 33,837 39,921 39,921 39,921

R-squared 0.830 0.817 0.822 0.841 0.840 0.840 0.844 0.843 0.843 0.839 0.838 0.838

Page 41: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

41

TABLE 4

Costs of Prediction Errors to Auditors

Panel A: Auditors' False Positive Costs N Mean Std Dev Median Sum

AUDFEE 139,454 1.2822 3.2897 0.4147 178,810

AUD_FPC_Audit_Investment (at Cost) 139,454 0.2587 0.6638 0.0837 36,082

Panel B: Litigation Costs

Incidence of Litigation Lawsuits Dismissed Lawsuits Settled Lawsuits

No Damages Damages

N Percent N Percent N Percent N Percent

v. Firm and Directors 212 67.7% 32 15.09% 14 6.60% 166 78.30%

v. Investment Bankers 19 4.8% 3 15.79% 0 0.00% 16 84.21%

v. Auditors 87 27.8% 23 26.44% 15 17.24% 49 56.32%

Settlement in millions of 2016 $

Paid by N Mean Sum MAX

Firm and Directors 166 82.8 25922.6 4874.3

Investment Bankers 16 66.5 20821.6 15407.5

Auditors 49 10.0 3132.8 512.4

Panel C: Auditors' False Negative Costs N Mean Std Dev Median Sum

AUD_FN_LIT (actual) 311 10.06 45.78 0.00 3128.5

AUD_FN_LIT (Estimated based on Honisberg et al.) 313 2.88 3.58 1.77 902.7

AUD_FN_REP (Clients Same Industry-COMPUSTAT-C) 274 0.37 18.49 0.17 102.7

AUD_FN_CLIENT LOSS (Year+1 v. Year -1)-C 313 2.65 29.62 0.00 828.2

AUD_FN_CLIENT LOSS (Years [+1, +2] v. [-1, -2])-C 313 3.68 30.15 0.00 1150.6

AUD_FN_REP (Clients Same Industry-Aud. Analytics-AA) 139 2.76 30.35 0.26 383.6

AUD_FN_CLIENT LOSS (Year+1 v. -1)-AA 180 1.38 14.35 0.00 248.0

AUD_FN_CLIENT LOSS (Years [+1, +2] v. [-1, -2])-AA 159 2.12 6.65 0.20 336.6

AUD_FN_CLIENT LOSS (Years [+1, +3] v. Y [-1, -3])-AA 126 2.06 6.28 0.00 260.1

Panel D: Ratio of Average False Negative to False Positive Costs

Mean Median False Positive Costs 0.2587 0.0837 False Negative Costs Ratio Ratio

Lit(Actual)+Rep+Client Loss in (-1,+1)-COMPUSTAT 13.08 50.6 0.17 2.0

Lit(Estimated)+Rep+Client Loss in (-1,+1)-COMPUSTAT 5.90 22.8 1.94 23.1

Lit(Actual)+Rep+Client Loss in (-2,+2)-COMPUSTAT 14.11 54.5 0.17 2.0

Lit(Estimated)+Rep+Client Loss in (-2,+2)-COMPUSTAT 6.93 26.8 1.94 23.1 In Panel A, auditors' false positive costs are either a lost audit fee (AUDFEE) in the event of resignation, or in the

absence of resignation, the incremental audit investment--at cost--that a fraud warning flag would generate.

Alternatively, auditor’s false positive costs are calculated as EAUDi[FP COST] = p (AUDFEEjt) + (1-p)

(INCRAUDFEEjt * (1-pm%)), where p is the probability of resignation, AUDFEEjt is the fee lost by the resigning

auditor in year t, INCRAUDFEEjt (Inv_FPC_Auditors (at cost)) is the value of the additional work undertaken by

the auditor as a result of the warning flag (which we assume that they cannot pass on to the client in year t), and where

Page 42: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

42

pm% represents the auditor’s usual profit margin. Panel B, reports, for each of the 313 fraud firms in the sample,

lawsuits filed against the firm, its investment bankers, and its auditors, as well as the lawsuit’s resolution. Data are

obtained from extensive searches on Audit Analytics, Westlaw, Lexis Nexis, and Factiva. Our searches begin with the

firm name at the time the fraud is discovered, but also consider name changes as applicable, specify a search period

beginning two years before and ending two years after the fraud period identified in AAERs, and uses search terms

such as “securities”, “class action”, litigation”, “lawsuit” and “settlement.” In Panel C, auditors' false negative costs

have three components. Auditor litigation costs are measured either as AUD_FN_LIT-Actual or as AUD_FN_LIT-

Estimated: AUD_FN_LIT-Actual is based on identifying lawsuits against auditors and their resolution using Audit

Analytics, Westlaw, Lexis Nexis, and Factiva. AUD_FN_LIT-Estimated is the expected settlement amount from the

model proposed by Honisberg, et al. (2018). Auditors’ reputation loss is the abnormal dollar value loss (market value

of equity on day -2 multiplied by abnormal return on day -1 to 1 relative the fraud revelation announcement. for all

clients of the auditor (AUD_FN_REP), and for whether the clients are in the same 6-digit GICS industry relying on

COMPUSTAT ((AUD_FN_REP (Clients Same Industry-COMPUSTAT-C), or AUDIT ANALYTICS

(AUD_FN_REP (Clients Same Industry-Aud. Analyt.-AA)). We measure auditor client losses as the difference

between an auditor’s rate of attrition at the office level in the years following the discovery of a missed detection and

either (i) the rate of attrition for the same ‘undetecting’ office in the year prior to the public revelation of the fraud, or

(ii) the average rate of attrition for all other of the auditor’s offices. When using COMPUSTAT data, we transform

the abnormal attrition rate into a number of clients lost and multiply this estimate by the average audit fee per client.

When using AUDIT ANALYTICS, we calculate the fee lost specifically by assuming it equals last year’s fee. In Panel

D, we report estimates of the ratio of the average cost of false negatives to false positives. Data are in millions of

2016 dollars.

Page 43: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

43

TABLE 5

Costs of Prediction Errors to Investors

Panel A: Investors False Positive Costs N Mean Std Dev Median Sum

INV_FPC_Audit_Extracost 139,454 0.2949 0.7566 0.0954 41,126

Inv_FPC_LOSS [PROFIT]-Market 139,454 20.57 3882.41 -4.11 2,869,008

Inv_FPC_LOSS [PROFIT]-Size 132,604 15.60 3873.14 -4.42 2,068,905

Inv_FPC_LOSS [PROFIT]-FF3 138,864 91.02 3750.55 -0.70 12,639,792

Inv_FPC_LOSS [PROFIT]-FF4 138,719 94.23 4086.44 0.13 13,071,245

Panel B: Investors False Negative Costs N Mean Std Dev Median Sum

INV_FN_MKT_Loss-3 day 313 446.6 1975.5 20.6 139,795

INV_FN_MKT_Loss-5 day 311 411.0 2074.4 24.2 127,825

INV_FN_MKT_Loss-1 month 313 414.1 4334.0 34.2 129,627

INV_FN_MKT_Loss-3 month 313 580.4 5344.8 38.6 181,664

INV_FN_MKT_Loss-6 month 313 627.5 7037.0 50.0 196,404

INV_FN_MKT_Loss-12 month 313 845.6 6318.0 49.9 264,665

Panel C: Ratio of Average False Negative to False Positive Costs

Audit_Extracost + Profit Foregone Revelation Period Losses MKT SIZE FF3 FF4 Loss 3-day 21.40 28.10 4.89 4.73 Loss 5-day 19.70 25.85 4.50 4.35 Loss 1-month 19.85 26.05 4.54 4.38 Loss 3-month 27.81 36.51 6.36 6.14 Loss 6-month 30.07 39.47 6.87 6.64 Loss 12-month 40.52 53.19 9.26 8.95

In Panel A, investors’ false positive costs have two potential components: (i) as residual claimants, investors bear the

costs of additional audit fees imposed by auditors as a result of classifying the firm as a potential fraud

(INCRAUDFEEjt), (ii) the second component adds or reduces to the costs of the incremental audit fees depending on

whether investors forego gains or avoid losses by not investing in the firms on which there is a warning flag

(Inv_FPC_PROFIT/LOSS is the one-year dollar abnormal return beginning in month 4 after the end of the firms fiscal

year). In Panel B, investors’ false negative costs as the abnormal dollar value loss (market value of equity on day -2

multiplied by abnormal return over different periods from days -1 to 1, to days -1, to +252 relative the fraud revelation

announcement. In Panel C, we report estimates of the ratio of the average cost of false negatives to false positives.

Data are in millions of 2016 dollars.

Page 44: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

44

TABLE 6

Costs of Prediction Errors to Regulators

Panel A: Regulators False Positive Costs N Mean Std Dev Median Sum

REG_FPC_3 139,454 84.58 431.25 6.97 11,795,679

REG_FPC_5 139,454 140.97 718.75 11.62 19,659,466

REG_FPC_10 139,454 281.95 1437.50 23.24 39,318,931

Panel B: Regulators False Negative Costs N Mean Std Dev Median Sum

One-day Loss 313 20.05 8.51 22.35 6,276

Three-day Loss 313 5.62 2.39 6.27 1,760

Five-day Loss 313 5.44 2.31 6.06 1,703

Panel C: Ratio of Average False Negative to False Positive Costs

REG_FPC_3 REG_FPC_5 REG_FPC_10

One-day Loss 0.24 0.14 0.07

Three-day Loss 0.07 0.04 0.02

Five-day Loss 0.06 0.04 0.02

Panel D: Estimation of Regulators False Negative Costs: AAER Announcements and Market-wide returns

1-day AAER Event-window:

Value-weighted Coeff. SE T-stat p-value

Intercept 0.00054 0.00011 4.89 <.0001

Event -0.00119 0.00051 -2.36 0.0185

Equal-weighted

Intercept 0.00084 0.00009 9.18 <.0001

Event -0.00096 0.00042 -2.30 0.0216

3-day AAER Event-window:

Value-weighted

Intercept 0.00053 0.00012 4.56 <.0001

Event -0.00033 0.00032 -1.05 0.2932

Equal-weighted

Intercept 0.00084 0.00010 8.70 <.0001

Event -0.00031 0.00026 -1.16 0.2454

5-day AAER Event-window:

Value-weighted

Intercept 0.00055 0.00012 4.54 <.0001

Event -0.00032 0.00027 -1.21 0.2244

Equal-weighted

Intercept 0.00084 0.00010 8.31 <.0001

Event -0.00019 0.00022 -0.88 0.3804

N= 9584 daily market-index returns in the period 1980-2017 In Panel A, we assume regulators publicly announce investigations of firms flagged and estimate regulators’ false

positive costs to range between 3% (REG_FPC_3) and 10% (REG_FPC_10) of a firm’s the market value of equity in

Page 45: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

45

month four subsequent to the firm’s fiscal year-end. In Panel B, we use the results of regressions in Panel D of the

value-weighted return on the market from CRSP on indicator variables capturing 1, 3 and 5 days surrounding the

revelation date. Specifically, we estimate regulators’ false negative costs as the regression estimates of the slope times

the value of the market on days -2 relative to the day of revelation. In Panel C, we report estimates of the ratio of the

average cost of false negatives to false positives. Data are in millions of 2016 dollars.

Page 46: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

46

TABLE 7

Model Predictive Performance

Total Accruals (Raw) Current Accruals (Raw) Current Accruals (Abnormal)

Decile

True

Positive

Rate (TP)

Cum. TP

Rate

Cum. False

Positive Rate

True Positive

Rate (TP) Cum. TP Rate

Cum. False

Positive

Rate

True

Positive

Rate (TP)

Cum. TP

Rate

Cum. False

Positive

Rate

10 19.9% 19.9% 9.98% 20.7% 20.7% 9.97% 18.0% 18.0% 9.98%

9 16.5% 36.3% 19.96% 16.5% 37.2% 19.96% 11.3% 29.3% 19.98%

8 12.4% 48.7% 29.96% 12.8% 50.0% 29.95% 12.4% 41.7% 29.97%

7 7.9% 56.6% 39.96% 7.9% 57.9% 39.96% 9.4% 51.1% 39.97%

6 8.6% 65.2% 49.96% 6.0% 63.9% 49.97% 7.9% 59.0% 49.98%

5 7.1% 72.3% 59.97% 6.4% 70.3% 59.98% 9.4% 68.4% 59.98%

4 4.9% 77.2% 69.98% 8.3% 78.6% 69.98% 7.1% 75.6% 69.99%

3 5.2% 82.4% 79.99% 6.4% 85.0% 79.99% 6.0% 81.6% 80.00%

2 6.7% 89.1% 90.00% 6.4% 91.4% 90.00% 7.1% 88.7% 90.00%

1 10.9% 100.0% 100.00% 8.6% 100.0% 100.00% 11.3% 100.0% 100.00%

M-Score F-Score FSD

Decile

True

Positive

Rate (TP)

Cum. TP

Rate

Cum. False

Positive Rate

True Positive

Rate (TP) Cum. TP Rate

Cum. False

Positive

Rate

True

Positive

Rate (TP)

Cum. TP

Rate

Cum. False

Positive

Rate

10 22.6% 22.6% 9.97% 30.9% 30.9% 9.95% 13.10% 13.10% 9.99%

9 17.8% 40.4% 19.95% 16.6% 47.5% 19.94% 11.80% 24.90% 19.99%

8 15.3% 55.7% 29.94% 14.3% 61.8% 29.93% 11.50% 36.40% 29.99%

7 11.8% 67.5% 39.94% 10.5% 72.3% 39.93% 10.80% 47.20% 39.99%

6 7.3% 74.8% 49.94% 7.3% 79.6% 49.93% 10.20% 57.40% 49.99%

5 6.7% 81.5% 59.95% 6.4% 86.0% 59.94% 10.20% 67.60% 59.99%

4 5.1% 86.6% 69.96% 5.1% 91.1% 69.95% 9.90% 77.50% 69.99%

3 5.4% 92.0% 79.97% 2.9% 93.9% 79.97% 8.60% 86.10% 79.99%

2 2.9% 94.9% 89.99% 4.1% 98.1% 89.98% 8.60% 94.70% 90.00%

1 5.1% 100.0% 100.00% 1.9% 100.0% 100.00% 5.30% 100.00% 100.00%

Page 47: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

47

TABLE 8

Costs and Benefits of Implementing Decision Rules Based on F- and M-Score cut-offs.

Net Cost

False Positives True Positives (Benefit)

Decision Rules Number Cost Incurred Number Cost Avoided Fscore>1 Auditors 55,212 16,297 226 3,235 13,062

Investors 55,212 146,921 226 96,165 50,756

Regulators 55,212 4,830,494 226 1,271 4,829,223

Fscore>1.85 Auditors 12,940 2,077 95 1,876 202

Investors 12,940 (876,650) 95 69,784 (946,434)

Regulators 12,940 569,277 95 479 568,798

Fscore>2.45 Auditors 5,505 668 56 797 (129)

Investors 5,505 (891,590) 56 12,678 (904,268)

Regulators 5,505 4,830,494 56 276 4,830,218

Mscore>-1.78 Auditors 23,320 2,681 112 1,516 1,165

Investors 23,320 (1,692,536) 112 18,125 (1,710,661)

Regulators 23,320 787,567 112 458 787,109

This table reports the aggregate costs of false positives are compared to the costs avoided by identifying true positives

for auditors, investors, and regulators, which are described in detail in Tables 4, 5, and 6 respectively. Data are in

millions of 2016 dollars.

Page 48: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

48

TABLE 9

Conditional Performance of F- and M-Score cut-offs Panel A: Performance conditional on Firm Size

NumFirm NumFraud Flag NoFraud Flag Fraud Hitrate FP Rate Pct Fraud Costratio

Mscore

0 27,327 46 5,082 16 34.78 18.60 0.31 317.63

1 27,353 91 5,621 31 34.07 20.55 0.55 181.32

2 27,354 164 5,407 43 26.22 19.77 0.80 125.74

3 27,352 211 4,261 52 24.64 15.58 1.22 81.94

4 27,337 253 2,365 31 12.25 8.65 1.31 76.29

Fscore1

0 27,327 46 2,600 12 26.09 9.51 0.46 216.67

1 27,353 91 3,073 18 19.78 11.23 0.59 170.72

2 27,354 164 3,082 38 23.17 11.27 1.23 81.11

3 27,352 211 2,567 58 27.49 9.39 2.26 44.26

4 27,337 253 1,345 34 13.44 4.92 2.53 39.56

Fscore2

0 27,327 46 1,097 8 17.39 4.01 0.73 137.13

1 27,353 91 1,430 11 12.09 5.23 0.77 130.00

2 27,354 164 1,394 21 12.80 5.10 1.51 66.38

3 27,352 211 1,032 32 15.17 3.77 3.10 32.25

4 27,337 253 476 14 5.53 1.74 2.94 34.00

Panel B: Performance conditional on Firm Age

NumFirm NumFraud Flag NoFraud Flag Fraud Hitrate FP Rate Pct Fraud Costratio

Mscore

0 25,787 127 7,781 66 51.97 30.17 0.85 117.89

1 26,714 188 5,097 56 29.79 19.08 1.10 91.02

2 28,435 166 4,467 28 16.87 15.71 0.63 159.54

3 28,277 112 3,445 11 9.82 12.18 0.32 313.18

4 27,542 176 1,952 12 6.82 7.09 0.61 162.67

Fscore1

0 25,787 127 3,881 59 46.46 15.05 1.52 65.78

1 26,714 188 2,904 49 26.06 10.87 1.69 59.27

2 28,435 166 2,550 24 14.46 8.97 0.94 106.25

3 28,277 112 2,064 18 16.07 7.30 0.87 114.67

4 27,542 176 1,271 11 6.25 4.61 0.87 115.55

Fscore2

0 25,787 127 2,125 40 31.50 8.24 1.88 53.13

1 26,714 188 1,255 26 13.83 4.70 2.07 48.27

2 28,435 166 995 9 5.42 3.50 0.90 110.56

3 28,277 112 709 8 7.14 2.51 1.13 88.63

4 27,542 176 348 4 2.27 1.26 1.15 87.00

Page 49: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

49

Panel C: Performance conditional on Total Accruals

NumFirm NumFraud Flag NoFraud Flag Fraud Hitrate FP Rate Pct Fraud Costratio

Mscore

0 26,931 130 3,011 28 21.54 11.18 0.93 107.54

1 26,932 120 2,220 7 5.83 8.24 0.32 317.14

2 26,931 140 2,078 10 7.14 7.72 0.48 207.80

3 26,932 171 2,992 17 9.94 11.11 0.57 176.00

4 26,931 195 12,107 109 55.90 44.96 0.90 111.07

Fscore1

0 26,931 130 1,770 25 19.23 6.57 1.41 70.80

1 26,932 120 946 10 8.33 3.51 1.06 94.60

2 26,931 140 1,032 14 10.00 3.83 1.36 73.71

3 26,932 171 1,751 20 11.70 6.50 1.14 87.55

4 26,931 195 7,050 91 46.67 26.18 1.29 77.47

Fscore2

0 26,931 130 971 16 12.31 3.61 1.65 60.69

1 26,932 120 321 5 4.17 1.19 1.56 64.20

2 26,931 140 265 8 5.71 0.98 3.02 33.13

3 26,932 171 445 9 5.26 1.65 2.02 49.44

4 26,931 195 3,385 48 24.62 12.57 1.42 70.52

Panel D: Performance conditional on Firm Life Cycle

NumFirm NumFraud Flag NoFraud Flag Fraud Hitrate FP Rate Pct Fraud Costratio

Mscore

0 15,516 110 5,781 57 51.82 37.26 0.99 101.42

1 32,574 286 5,386 52 18.18 16.53 0.97 103.58

2 44,554 205 2,489 12 5.85 5.59 0.48 207.42

3 9,907 53 1,480 6 11.32 14.94 0.41 246.67

4 7,565 36 2,352 13 36.11 31.09 0.55 180.92

Fscore1

0 15,516 110 3,693 43 39.09 23.80 1.16 85.88

1 32,574 286 3,693 65 22.73 11.34 1.76 56.82

2 44,554 205 1,857 16 7.80 4.17 0.86 116.06

3 9,907 53 453 5 9.43 4.57 1.10 90.60

4 7,565 36 584 8 22.22 7.72 1.37 73.00

Fscore2

0 15,516 110 2,035 30 27.27 13.12 1.47 67.83

1 32,574 286 1,422 30 10.49 4.37 2.11 47.40

2 44,554 205 417 5 2.44 0.94 1.20 83.40

3 9,907 53 196 3 5.66 1.98 1.53 65.33

4 7,565 36 304 4 11.11 4.02 1.32 76.00

This table reports the performance of M-Score and F-Score at different cutoffs. Fscore1 reports performance for the

1.85 cutoff, while Fscore2 reports results for the 2.45 cutoff. NumFirm is the total number of firm-years in a quintile

and/or life cycle stage. NumFraud is the number of fraud-years in a quintile and/or life cycle stage. Flag Nofraud and

Flag Fraud are equal to the number of falsely (correctly) flagged firm-years. We then use these numbers to determine

the hit rates (the percentage of fraud firms correctly flagged; Flag Fraud / NumFraud), false positive rates (the

percentage of non-fraud firms incorrectly flagged as fraudulent; Flag NoFraud / NumFirm-NumFraud), the number

Page 50: The Cost of Fraud Prediction Errors Messod D. Beneish ...€¦ · to predict fraud over the period 1982-2016: the M-Score (Beneish 1999), an accrual model based on Kothari et al.

50

of fraud firms as a percentage of all firms that are flagged (Flag Fraud / Flag Fraud + Flag NoFraud), and the implied

cost ratio with which models trade off true and false positives (1/pct Fraud).


Recommended