+ All Categories
Home > Documents > Super-efficiency and stability intervals in additive DEA...3 that interval. This is confronted with...

Super-efficiency and stability intervals in additive DEA...3 that interval. This is confronted with...

Date post: 30-Jan-2021
Category:
Upload: others
View: 0 times
Download: 0 times
Share this document with a friend
21
1 This is a PDF file of an unedited manuscript that has been accepted for publication in Journal of the Operational Research Society. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain. The final version will be available at: http://dx.doi.org/10.1057/jors.2012.19 Super-efficiency and stability intervals in additive DEA M C Gouveia ISCAC, Quinta Agrícola, Bencanta, 3040-316 Coimbra, Portugal, and INESC Coimbra, Rua Antero de Quental, 199; 3000-033 Coimbra, Portugal [email protected] L C Dias Faculdade de Economia, Univ. Coimbra, Av. Dias da Silva 165, 3004-512 Coimbra, Portugal, and INESC Coimbra, Rua Antero de Quental, 199; 3000-033 Coimbra, Portugal [email protected] C H Antunes DEEC-FCT Universidade de Coimbra-Pólo II, 3030 Coimbra, Portugal, and INESC Coimbra, Rua Antero de Quental, 199; 3000-033 Coimbra, Portugal [email protected] ABSTRACT. This study addresses the problem of finding the range of efficiency for each Decision Making Unit (DMU) considering uncertain data. Uncertainty in the DMU coefficients in each factor (input or output) is captured through interval coefficients (i.e., these are uncertain but bounded). A two- phase additive Data Envelopment Analysis (DEA) model for performance evaluation is used, which is adapted to include the concept of super-efficiency to provide a robustness analysis of the DMUs in face of uncertain information, assessing whether each DMU is surely efficient, potentially efficient, or surely inefficient for the uncertainty intervals specified. Another contribution is to present how a maximal stability hyper-rectangle can be computed for each DMU such that its efficiency status does not change when the coefficients vary within that interval. KEYWORDS. Data Envelopment Analysis; Multi-Criteria Analysis; Super-efficiency; Robustness Analysis; Uncertainty; Stability intervals.
Transcript
  • 1

    This is a PDF file of an unedited manuscript that has been accepted for publication in Journal of the

    Operational Research Society. The manuscript will undergo copyediting, typesetting, and review of the

    resulting proof before it is published in its final form. Please note that during the production process errors

    may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

    The final version will be available at: http://dx.doi.org/10.1057/jors.2012.19

    Super-efficiency and stability intervals in additive DEA

    M C Gouveia

    ISCAC, Quinta Agrícola, Bencanta, 3040-316 Coimbra, Portugal, and

    INESC Coimbra, Rua Antero de Quental, 199; 3000-033 Coimbra, Portugal

    [email protected]

    L C Dias

    Faculdade de Economia, Univ. Coimbra, Av. Dias da Silva 165, 3004-512 Coimbra, Portugal, and

    INESC Coimbra, Rua Antero de Quental, 199; 3000-033 Coimbra, Portugal

    [email protected]

    C H Antunes

    DEEC-FCT Universidade de Coimbra-Pólo II, 3030 Coimbra, Portugal, and

    INESC Coimbra, Rua Antero de Quental, 199; 3000-033 Coimbra, Portugal

    [email protected]

    ABSTRACT. This study addresses the problem of finding the range of efficiency for each Decision

    Making Unit (DMU) considering uncertain data. Uncertainty in the DMU coefficients in each factor

    (input or output) is captured through interval coefficients (i.e., these are uncertain but bounded). A two-

    phase additive Data Envelopment Analysis (DEA) model for performance evaluation is used, which is

    adapted to include the concept of super-efficiency to provide a robustness analysis of the DMUs in face

    of uncertain information, assessing whether each DMU is surely efficient, potentially efficient, or surely

    inefficient for the uncertainty intervals specified. Another contribution is to present how a maximal

    stability hyper-rectangle can be computed for each DMU such that its efficiency status does not change

    when the coefficients vary within that interval.

    KEYWORDS. Data Envelopment Analysis; Multi-Criteria Analysis; Super-efficiency; Robustness

    Analysis; Uncertainty; Stability intervals.

    http://dx.doi.org/10.1057/jors.2012.19

  • 2

    Introduction

    Data Envelopment Analysis (DEA), originally developed by Charnes et al. (1978), is a nonparametric

    approach based on linear programming to evaluate observations representing the performances of all units

    (Decision Making Units - DMUs) under evaluation. Each DMU is characterized by the "consumption" of

    multiple inputs for the "production" of multiple outputs. The different DEA models seek to determine

    which of n DMUs form the efficient frontier (or envelopment surface) in Pareto-Koopmans sense. These

    evaluations result in a performance score that ranges between zero and unity that represents the “degree

    of efficiency” obtained by DMUs.

    In real-world evaluation problems, information is generally subject to several sources of uncertainty,

    resulting from data scarcity, difficulties of data estimation or data collection, or even to contradictory

    information from distinct sources. Therefore, decision aid models must cope with this data uncertainty by

    using models capable of providing robust conclusions, i.e., recommendations that are somehow

    “immune” to plausible data instantiations. The use of interval coefficients is a very flexible modelling

    tool for capturing this type of data uncertainty (i.e., the precise performances of the DMUs are unknown

    but bounded within an interval) since it does not impose stringent requirements about probability or

    possibility distributions. There are two main perspectives to deal with uncertainty in the context of our

    work. One, which is generally encompassed under the designation imprecise DEA (IDEA) (Cooper et al.,

    1999, 2001a, 2001b), studies how to deal with imprecise data such as bounded data, ordinal data and ratio

    bounded data in DEA and results in a non-linear and non-convex DEA model. By using scale

    transformations and variable changes (Zhu, 2003), or only variable transformations (Despotis and Smirlis,

    2002), this non-linear model can be transformed into an equivalent linear programming problem. The

    other perspective, which is more related to the approach proposed in this paper, deals with computing

    stability intervals for uncertain coefficients so that results do not change (Zhu, 1996, 2001; Seiford and

    Zhu, 1998a, 1998b).

    This paper addresses this type of uncertainty in the context of a two-phase method developed by Gouveia

    et al. (2008), which was inspired on the additive DEA model proposed by Charnes et al. (1985) as well as

    the additive model with oriented projections presented by Ali et al. (1995). In Gouveia et al.’s model, the

    DMUs are treated as alternatives of a multiple criteria decision model (an additive multi-attribute utility

    model), each alternative being evaluated in a number of distinct criteria. The two-phase method provides

    an efficiency measure of each DMU by calculating a criterion weighting vector and, if necessary,

    obtaining the projected point.

    There are two main contributions in this work. One contribution is to present how the range of efficiency

    for each DMU can be computed in presence of interval values for the DMU coefficients in each

    input/output. This allows performing a robustness analysis of each DMU under evaluation, assessing

    whether each DMU is surely efficient, potentially efficient, or surely inefficient for the uncertainty

    intervals specified. Another contribution is to present how a maximal stability hyper-rectangle can be

    computed for each DMU such that its efficiency status does not change when the coefficients vary within

  • 3

    that interval. This is confronted with the idea of taking a super-efficiency score as a proxy of robustness,

    after adapting the original model of Gouveia et al. (2008) to consider the concept of super-efficiency.

    Following this introduction, section 2 briefly describes DEA models, with emphasis on the additive DEA

    model and the weighted additive model. In section 3, the concept of super-efficiency is reviewed. Section

    4 introduces the two-phase method of Gouveia et al. (2008) with the modifications to include the super-

    efficiency concept. In section 5, the ranges of efficiency are computed and the robustness of each DMU is

    analyzed by considering a two-dimensional pedagogical example. In section 6, an example with real

    world data is provided for exploiting the insights from this robustness analysis. Concluding remarks are

    presented in section 7.

    Data Envelopment Analysis

    The set of n DMUs to be evaluated is )},...,1(:{ njjDMU = . Each DMUj consumes m different inputs

    ),...,1( miijx = to produce p different outputs ),...,1( prrjy = . jX (the jth column of nḿX ) denotes the

    vector of inputs consumed by DMUj. A similar notation is used for outputs, Yj. The input data is

    represented by matrixnḿX and the output data by matrix np´Y . 1 denotes the summation vector( )

    T1,...,1 .

    There are two main types of DEA models that provide a measure of relative efficiency for each DMU

    according to the returns to scale considered (Seiford and Zhu, 1999b): Constant Returns-to-Scale (CRS)

    models, such as the CCR model (see Charnes et al., 1978), and Variable Returns to Scale (VRS) models,

    such as the BCC model (see Banker et al., 1984) and the additive model (ADD) of Charnes et al. (1985).

    The additive model

    In CCR and BCC models we need to distinguish between input-oriented and output-oriented models. The

    additive model combines both orientations in a single model, which can be formulated as follows:

    ADD Xk,Yk( )( )

    min zk = - 1s+1e( )

    s.t. Yl - s= Yk,

    - Xl - e= -Xk,

    1l =1,

    l ³ 0, e ³ 0, s³ 0.

    The ADD model returns a non-positive value *kz , which allows checking the relative efficiency of the

    DMU k under analysis. If the value obtained is negative, then the DMU under analysis is operating

    inefficiently in some factors. This value is the symmetric of the sum of the distances in each dimension to

    the envelopment surface (L1 distance).

    If DMU k is inefficient (i.e., it does not lie on the efficient frontier defined by the set of DMUs) the model

  • 4

    identifies a projected point ( )kk YX ˆ,ˆ on the efficient frontier. If the optimal value of the primal ADD

    model is zero the point ( )kk YX , belongs to the efficient frontier, that is ( )kk YX ˆ,ˆ = ( )kk YX , . This is a

    necessary and sufficient condition for efficiency, Charnes et al. (1985).

    The projected point can be characterized, in an alternative way, as sYeXYX kkkk ,ˆ,ˆ , from the

    primal constraints. This ADD model measures the excess of inputs, e , and the deficit of outputs, s , in

    which the DMU k operates when confronted with the DMUs that operate on the efficient frontier.

    Ali et al. (1995) presented a variant of additive model with oriented projections, which is henceforth

    called weighted additive model.

    The envelopment formulation for this model is:

    ADDW Xk,Yk,uk, vk( )( )

    min zk = - uks+ vke( )

    s.t. Yl - s= Yk,

    - Xl - e= -Xk,

    1l =1,

    l ³ 0, e ³ 0, s³ 0.

    The parameters uk, v

    k(which are fixed before the model is solved) have play an important role that is

    clearly seen in the primal formulation. The vectors uk, v

    k( ) are the coefficients of the objective function

    and thus define the relative weight attributed to one unit of each slack (for a discussion on the role of

    weights and value judgments in DEA (see Thanassoulis et al., 2004)). The weight vectors provide and

    determine the directions of the projection. Setting uk

    = 1and vk

    = 1 in the primal weighted additive

    problem leads to the original additive model.

    Super-efficiency and sensitivity analysis in DEA

    Andersen and Petersen (1993) developed an extended DEA measure in which the basic idea is to compare

    the DMU under evaluation with a linear combination of all other DMUs in the reference set. This means

    that the production possibility set is reduced by not considering the DMU being evaluated, which allows

    efficient DMUs to become super-efficient and have different super-efficiency scores. In other words, the

    DMUs can increase the input vector (or decrease the output) to some extent while preserving efficiency.

    We can also perceive DEA models as projection mechanisms and the projections of the inefficient DMUs

    on the efficient frontier depend on the scales used to measure each input or output. Super-efficiency is

    very sensitive to the projection mechanisms and tends to favour “extreme” solutions (Bouyssou, 1999).

    The set of DMUs can be partitioned into two groups: frontier (efficient) DMUs and non-frontier

    (inefficient) DMUs. The frontier DMUs consist of DMUs in the set E (extreme efficient), set E’ (efficient

    but not an extreme point) and the set F (weakly efficient or frontier point but with non-zero slacks). The

  • 5

    super-efficiency model identifies the classification of a given DMU and, using some extensions of the

    super-efficiency model, a sensitivity analysis of the conclusions about efficiency can also be performed.

    However, under certain conditions the process of determining the super-efficiency score can lead to an

    infeasible linear program. Based on Thrall (1996), a necessary, but not sufficient, condition for

    infeasibility is that an excluded DMU is extreme efficient. Dulá and Hickman (1997) and Seiford and Zhu

    (1999a) reported a necessary and sufficient condition for infeasibility in an input-oriented CCR super-

    efficiency model: the excluded DMU has the only zero value for any input, or the only positive value for

    any output, among all DMUs in the reference set. Infeasibility cannot arise in an output-oriented CCR

    super-efficiency model. Infeasibility also occurs in the BCC super-efficiency model, when an efficient

    DMU under evaluation cannot reach the frontier formed by the remaining DMUs via increasing the inputs

    (or decreasing the outputs). Infeasibility arises in either orientation whenever there is no reference DMU

    for the excluded one.

    Many DEA researchers have addressed the sensitivity of the results to data perturbations and the

    robustness of the efficiency scores resulting from these perturbations, based on super-efficiency DEA

    approaches. Continuing the work of Zhu (1996), Seiford and Zhu (1998a) developed a sensitivity analysis

    procedure to determine stability regions for possible increases in all inputs and for possible decreases in

    all outputs within which the efficiency of a specific efficient DMU remains unchanged. Seiford and Zhu

    (1998b) extended the method by Zhu (1996) and Seiford and Zhu (1998a) to the worst-case scenario,

    where the same maximum percentage data changes for deteriorating the efficiency of a DMU under

    analysis and the data changes for improving the efficiencies of the other DMUs simultaneously are

    calculated. In this work, the authors also concluded that the relationship between the infeasibility and

    stability of efficiency classification, discovered in Seiford and Zhu (1998a), remains for the simultaneous

    data variations case and for all basic DEA models. Generalizing these results, Zhu (2001) considered that

    the data perturbation in the DMU under analysis and the data perturbations in the remaining DMUs can

    be different when all the remaining DMUs improve their efficiencies at the expense of deteriorating the

    efficiency of the efficient DMU under analysis. Necessary and sufficient conditions for preserving

    efficiency were provided.

    Our approach differs from the sensitivity and robustness analysis presented by previous authors in several

    aspects. It has been developed for the additive two-phase model of Gouveia et al (2008). The uncertainty

    in the coefficients in each factor (input or output) is captured through interval coefficients and converted

    into utility scales (which are always to be maximized). An optimistic efficiency measure and a pessimistic

    efficiency measure are computed. Additionally, a tolerance threshold for each efficient DMU is

    determined, that is a maximum tolerance in the factor scores for which the DMU’s efficiency status

    changes. Using these two types of efficiency measures, we can classify the DMUs as surely efficient,

    potentially efficient, or surely inefficient. Unlike the standard super-efficiency models, with specific

    orientations, the model proposed in this paper projects the DMUs in any direction in a way that minimizes

    the distance of the unit under evaluation to the best of all units (excluding the one in evaluation), and

  • 6

    therefore no infeasibility concerns arise.

    Adaptation of the two-phase method to compute super-efficiency scores

    The two-phase method developed by Gouveia et al. (2008) is a variant of the additive DEA model with

    oriented projections (Ali et al., 1995), which uses concepts developed in the field of multiple criteria

    decision analysis (MCDA) under imprecise information (Athanassopoulos and Podinovski, 1997; Dias

    and Clímaco, 2000).

    We consider the DMUs as alternatives of a multiple criteria evaluation model, each one being evaluated

    in a number of distinct criteria. Each criterion corresponds to an input or an output factor in DEA models.

    A direction of preference is associated with each criterion: increasing for outputs and decreasing for

    inputs. The method uses an additive utility function to aggregate the utilities associated with each

    alternative, based on the Multi-Attribute Utility Theory (MAUT) (see Keeney and Raiffa, 1976).

    Adapting to this context, the purpose of MAUT will be to assess the utility of each alternative,

    considering that the larger the utility the better. MAUT is also aimed at simplifying the task of building

    the utility functions when evaluating the alternatives that are described by multiple attributes. In addition

    the decision maker's task is facilitated because he can focus the attention on one attribute at a time and

    then make the aggregation of attributes, rather than make judgments directly on the global utility (the

    concept of attribute in this theory is equivalent to our concept of criterion). This overcomes the problem

    of the scales associated with the ADD model, since all the input and output measures are translated into

    utility units. Moreover, the weights used in the aggregation gain a specific meaning: they are the scale

    coefficients of the utility functions. Weights are chosen to benefit each DMU as much as possible, rather

    than being fixed beforehand as in the model by Ali et al. (1995). Finally, the efficiency measure assigned

    to each DMU gains an intuitive meaning: it corresponds to a “min-max regret” (utility loss) measure.

    Considering that the alternatives are the DMUs to be evaluated according to q criteria, we assume that the

    utility of each alternative is given by an additive MAUT model

    q

    cjccj DMUuwDMUu

    1

    , where

    wc ³ 0,"c =1,...,q and wc =1c=1

    q

    å (by convention). The scale coefficients qww ,...,1 are the weights of the

    utility functions.

    The use of this model requires that the original input and output scales have to be converted into utility

    scales and there are several techniques for questioning the decision maker, in order to construct the utility

    functions compatible with their answers (see von Winterfeldt and Edwards, 1986). Hence, after being

    converted into utilities all criteria are treated as outputs.

    Gouveia et al. (2008) proposed a two-phase method to incorporate preferences in the ADD model. In this

    study the two-phase method is adapted to consider the super-efficiency concept. For that purpose the

    following problem is solved:

  • 7

    d,wmin dk

    s.t. wcuc DMU j( )c=1

    q

    å - wcuc DMUk( )c=1

    q

    å £ dk, j =1,...,n, j ¹ k

    wc =1c=1

    q

    å ,

    wc ³ 0,"c =1,...,q

    (1)

    The score denotes the distance defined by the utility difference to the best of all alternatives

    (excluding the one under evaluation). The aim of our approach is, for DMU k, to calculate the vector w of

    utility function weights that minimizes the distance (the utility difference) of this unit to the best one

    (note that the best alternative will also depend on w), excluding itself from the reference set. Regarding

    the model of Gouveia et al. (2008), the only change is to exclude one constraint in which the DMU k

    under evaluation was compared to itself (which is achieved by introducing j≠k in problem (1)). This

    change implies that is now allowed to become negative.

    This approach is identical to that described in Gouveia et al (2008), and starts by finding the weights (the

    variables of problem (1)) that most benefit the DMU under consideration to have the worst utility loss

    (also a variable of problem (1)). Then, the “weighted additive” problem can be solved using the optimal

    weighting vector wc* , resulting from (1), to compute the projected point in case of the DMU is inefficient.

    Phase 1: Convert inputs and outputs into utility scales. Compute the efficiency measure, , of each

    DMU, k = 1,…,n, and the corresponding weighting vector.

    Phase 2: If dk* ³ 0 then solve the “weighted additive” problem (2), using the optimal weighting vector

    resulting from phase 1, , and determine the corresponding projected point of the DMU under

    evaluation.

    minl,s

    zk = - wc*sc

    c=1

    q

    å

    s.t. l juc DMU j( )j=1, j¹k

    n

    å - sc = uc DMUk( ), c = 1,...,q

    1l = 1,

    l ³ 0, s³ 0.

    (2)

    If the optimal value of the objective function in (1) is not positive, then the DMU k under evaluation is

    efficient. Otherwise it is inefficient and is the minimum difference of utility to the best DMU (i.e., the

    DMU with higher global utility). However, with the adaptation made to the original method we can also

    discriminate the efficient units. If < 0, then the DMU is in the set E (extreme efficient); if = 0 and

    all the slacks are null in phase 2, then the DMU belongs to E’ (efficient but not an extreme point); if =

    dk*

    dk*

    dk*

    wc*

    dk*

    dk*

    dk* dk

    *

    dk*

  • 8

    0 and not all the slacks are null in phase 2, then the DMU belongs to set F (weakly efficient or frontier

    point but with non-zero slacks).

    Using this measure, we can assess the extent to which an efficient DMU may worsen its utility while

    remaining efficient. This allows analyzing the robustness of the classification of a DMU as an efficient

    unit in face of uncertain information regarding factor coefficients. As becomes more negative, we

    expect the efficient DMU to be more robust to changes in input and output levels.

    Robustness analysis and stability intervals

    Consider that the value cjp (performance of DMU j in factor c) is uncertain but bounded within the range

    U

    cjcj

    L

    cj pp p . This implies ),()()( jU

    cjcj

    L

    c DMUuDMUuDMUu if the factor c is an output, or

    ),()()( jU

    cjcj

    L

    c DMUuDMUuDMUu if the factor c is an input.

    Given interval performances on each factor (input or output), it is possible to compute an optimistic and a

    pessimistic efficiency measure dk* for each DMU, k =1,…,n, using the first phase of the two-phase

    method. As an example, let us consider that all performances are bound to intervals

    pcjL = pcj 1-d( ) £ pcj £ pcj 1+d( ) = pcj

    U, with d for instance equal to 5%, 10%, or 20%. For this work, we

    consider that all performances are applied the same toleranced , but we may consider that the tolerance is

    applied only to a subset of these (inputs or outputs).

    To present the methodology proposed in this paper, let us consider as an illustration the data in Table 1

    (displayed in Fig. 1). Taken from Gouveia et al. (2008), these data have been modified by adding DMU9,

    which is weakly efficient, to portray more possibilities. For this illustration let us assume that inputs and

    outputs are converted into “utilities” in a linear way (in practice these functions may be constructed with

    the clients of the study, reflecting their value system, see Almeida and Dias (2012)). So, for the plausible

    higher tolerance value considered (in this case = 20%), and for each c =1,…,q, we choose a value

    and , and then we compute the utilities for each unit

    using:

    uc DMU j( ) =

    pcj - McL

    McU - Mc

    L, if factor c is an output

    McU - pcj

    McU - Mc

    L, if factor c is an input

    ì

    í

    ïïï

    î

    ïïï

    , j =1,..., n,c =1,..., q.

    dk*

    McL

    < min pcjL, j =1,...,n{ } McU > max pcjU, j =1,...,n{ }

  • 9

    Table 1: Test data

    DMU Y1 Y2

    1 10 2

    2 9 5

    3 6 7

    4 3 8

    5 4 4

    6 8 1

    7 5 6

    8 4 6.5

    9 2 8

    McL

    1 0

    McU

    12 10

    Considering the nominal values for each DMU and applying the two-phase method (1) and (2), we can

    build Table 2 and rank the DMUs in terms of optimal utility loss dk* :

    DMU1 ≻ DMU2 ≻ DMU3 ≻ DMU4 ≻ DMU9 ≻ DMU8 ≻ DMU7 ≻ DMU6 ≻ DMU5.

    The lower the value of dk* the better, and if dk

    * is negative then the DMU is efficient (DMUs 1 to 4).

    DMU9 has dk*= 0 but it is not strictly efficient, since one of the slacks is not null. The remaining DMUs

    have dk*> 0 and hence are not efficient.

    The projections of inefficient DMUs were obtained considering the weighting vector that resulted from

    phase 1, in which the linear program (1) is solved for each DMU. Among the efficient DMUs (DMUs 1-

    4) the first two have a larger margin to decrease their performance than the remaining ones.

    Table 2. Efficiency measure and other results for each DMU

    DMU dk*

    *

    1w *

    2w s1*

    s2*

    1 -0.091 1.000 0.000

    2 -0.074 0.579 0.421

    3 -0.032 0.355 0.645

    4 -0.020 0.216 0.784

    5 0.250 0.423 0.577 0.455 0.100 2=1

    6 0.163 0.767 0.233 0.091 0.400 2=1

    7 0.096 0.423 0.577 0.227 0.000 2=0.5; 3=0.5

    8 0.085 0.268 0.732 0.000 0.117 3=1

    9 0.000 0.000 1.000 0.091 0.000 4=1

    To compute the optimistic efficiency measure we consider the best value of the intervals for the DMU

  • 10

    being evaluated and the worst value of the intervals for all other DMUs. The reverse is considered to

    compute the pessimistic efficiency measure. Let jLc DMUu denote the minimum utility that DMU j

    attains in factor c given its uncertain performance, note that and

    in the next expressions:

    ucL DMU j( ) =

    pcjL - Mc

    L

    McU - Mc

    L, if factor c is an output

    McU - pcj

    U

    McU - Mc

    L, if factor c is an input

    ì

    í

    ïïï

    î

    ïïï

    , j =1,..., n,c =1,..., q

    Let jUc DMUu denote the maximum utility that DMU j attains in factor c given its uncertain

    performance:

    ucU DMU j( ) =

    pcjU - Mc

    L

    McU - Mc

    L, if factor c is an output

    McU - pcj

    L

    McU - Mc

    L, if factor c is an input

    ì

    í

    ïïï

    î

    ïïï

    , j =1,...,n,c =1,..., q

    To compute the optimistic efficiency measure dkopt * for DMU k the following LP similar to (1) is solved:

    min dkopt

    s.t. wcucL

    DMU j( )c=1

    q

    å - wcucU

    DMUk( )c=1

    q

    å £ dkopt

    , j = 1,...,n, j ¹ k

    wc = 1c=1

    q

    å ,

    wc ³ 0,"c =1,..., q

    To compute the pessimistic efficiency measure dkpes* for DMU k the following LP similar to (1) is

    solved:

    min dkpes

    s.t. wcucU

    DMU j( )c=1

    q

    å - wcucL

    DMUk( )c=1

    q

    å £ dkpes

    , j = 1,...,n, j ¹ k

    wc = 1c=1

    q

    å ,

    wc ³ 0, "c = 1,..., q

    With this analysis we can assess the robustness of the DMU under consideration because despite being

    assessed in a pessimistic way it may keep its efficiency status. A DMU is said to be robust to changes in

    its factors if the DMU remains in the same status after the change. Therefore, the DMU is robustly

    McL

    < min pcjL, j =1,...,n{ }

    McU > max pcj

    U, j =1,...,n{ }

  • 11

    efficient in the range of uncertainty considered. These results are displayed in Table 3.

    Table 3. Lower and upper limits for the utility loss dkopt *

    , dkpes*é

    ëêù

    ûú, for each DMU

    DMU 5% 10% 20%

    1 [-0.177;-0.005] [-0.264;0.082] [-0.436;0.255]

    2 [-0.138;-0.009] [-0.203;0.056] [-0.333;0.185]

    3 [-0.095;0.031] [-0.158;0.094] [-0.284;0.219]

    4 [-0.087;0.048] [-0.160;0.116] [-0.320;0.251]

    5 [0.199;0.301] [0.148;0.352] [0.046;0.454]

    6 [0.097;0.229] [0.018;0.295] [-0.145;0.428]

    7 [0.037;0.155] [-0.024;0.213] [-0.146;0.331]

    8 [0.024;0.147] [-0.038;0.209] [-0.161;0.332]

    9 [-0.080;0.080] [-0.160;0.154] [-0.320;0.283]

    DMUs 1 and 2 are efficient and remain in this state when a tolerance of 5% is considered. DMUs 3 and 4

    do not remain efficient for this tolerance value. For this tolerance, DMUs 1 and 2 are surely efficient,

    DMUs 3, 4, and 9 are potentially efficient, and DMUs 5-8 are surely inefficient. If the intervals of

    uncertainty are defined by a tolerance of 10% or 20%, then there are no surely efficient DMUs.

    Figure 1 shows the efficient frontier, given the nominal values. The pessimistic evaluation of efficiency

    for DMU1 considering the 5% tolerance is portrayed by the dashed line. The rectangles define the region

    of tolerance. In this case, DMU1 is in its worst performance level while the remaining units are at their

    best performance levels, and it remains efficient.

    0.1

    0.2

    0.3

    0.4

    0.5

    0.6

    0.7

    0.8

    0.9

    9 43

    2

    16

    78

    5

    2u

    1u0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

    Figure 1. The unit isoquant spanned by the pessimistic evaluation of DMU1

    Figure 2 shows DMU7's optimistic assessment considering a tolerance of 10%, which is portrayed by the

    dashed line. Hence, DMU7 is in its best performance while the remaining units are at their worst for this

    tolerance values and, with this type of analysis, theDMU7 that was in the set of inefficient units

    (considering the nominal values of DMUs) is now in the set of efficient ones.

  • 12

    0.2

    9 48 7

    5

    6

    32

    1

    2u

    1u

    0.1

    0.3

    0.4

    0.5

    0.9

    0.6

    0.7

    0.8

    0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

    Figure 2. The unit isoquant spanned by the optimistic evaluation of DMU7

    A different type of analysis that can be carried out is to compute the maximum tolerance such that the

    efficient DMUs maintain efficiency, i.e., to compute a stability interval in terms of a parameter

    affecting all the factors:

    dmax = d : dkpes*(d) = 0{ }

    This can be easily accomplished by a bisection technique (see Table 4) if the utility functions are

    monotonous. Table 5 shows that for a tolerance greater than 5.2% the DMU1 is no longer guaranteed to

    be efficient. For DMU2 the tolerance threshold for being robustly efficient is 5.6%.

    Table 4: Computing δmax

    using a bisection technique

    Let d denote the maximum tolerance value (in this paper d = 0.2)

    Let e denote the precision used (in this paper e = 0.001).

    d := 0;

    while (d - d )

  • 13

    Table 5. Efficiency threshold for the DMUs 1, 2, 3 and 4

    DMU 1 2 3 4

    δmax

    5.2% 5.6% 2.5% 1.4%

    It is noteworthy that DMU1 had a measure of efficiency highest than DMU2, but in this analysis this

    DMU emerges as being the most robust as it maintains the range of efficiency for a higher level of

    uncertainty on performances. Concerning DMUs 3 and 4, the tolerance threshold for which they are

    surely efficient is lower.

    An example with real world data

    This section revisits an empirical example of Cooper et al. (2006) displayed in Table 6, aiming to

    illustrate the insights that can be obtained from the approach proposed in this paper. There are 12

    hospitals and 6 factors: number of doctors and nurses and the relative unit costs of doctors and nurses in

    terms of inputs, and two outputs identified as number of outpatients and inpatients (each in units of 100

    persons/month). In this case we also consider that the higher tolerance value is = 20%, and for each c

    =1,…,q, we choose a value and .

    Table 6: Test data

    Inputs Outputs

    Doctor Nurse Outpatients Inpatients

    DMU Number Cost Number Cost Number Number

    1 20 500 151 100 100 90

    2 19 350 131 80 150 50

    3 25 450 160 90 160 55

    4 27 600 168 120 180 72

    5 22 300 158 70 94 66

    6 55 450 255 80 230 90

    7 33 500 235 100 220 88

    8 31 450 206 85 152 80

    9 30 380 244 76 190 100

    10 50 410 268 75 250 100

    11 53 440 306 80 260 147

    12 38 400 284 70 250 120

    McL

    10 200 100 50 70 40

    McU

    70 750 400 150 320 180

    For the purpose of this illustration, the performances of the 12 hospitals are converted into utilities, using

    McL

    < min pcjL, j =1,...,n{ } McU > max pcjU, j =1,...,n{ }

  • 14

    a linear transformation, as described in the previous section, and considering the tolerance d = 5%, 10%,

    or 20%.

    Table 7 shows the efficiency measure, weights, slack and λ values, for each DMU considering the

    nominal values, resulting from the application of the two-phase method (section 4). From these results we

    can conclude that only three hospitals are working inefficiently. This efficiency measure allows to

    discriminate the efficient units using the super-efficiency concept and rank all DMUs:

    DMU11≻DMU5≻DMU1≻DMU2≻DMU12≻DMU10≻DMU4≻DMU9≻DMU7≻DMU6≻DMU3≻DMU8.

    The DMUs 11, 5, 1, 2, 12, 10, 4, 9 and 7 are efficient. With the slack values obtained by solving problem

    (2) and introducing the weight vector resulting from stage (1), we get the projections of inefficient

    DMUs.

    Table 7. Efficiency measure and other results, for each DMU

    DMU dk*

    *

    1w *

    2w *

    3w *

    4w *

    5w *

    6w s1*

    s2*

    s3*

    s4*

    s5*

    s6*

    1 -0.0949 0.261 0.000 0.274 0.000 0.000 0.466

    2 -0.0903 0.000 0.060 0.856 0.000 0.084 0.000

    3 0.0246 0.092 0.000 0.361 0.017 0.530 0.000 0.000 0.160 0.000 0.112 0.043 0.044 2=0.76;10=0.17; 12=0.04

    4 -0.0125 0.000 0.000 0.461 0.000 0.415 0.124

    5 -0.0952 0.000 0.524 0.000 0.476 0.000 0.000

    6 0.0227 0.000 0.000 0.448 0.030 0.395 0.127 0.132 0.083 0.000 0.045 0.042 0.038 2=0.10;10=0.90

    7 -0.0004 0.211 0.000 0.273 0.000 0.516 0.000

    8 0.0479 0.050 0.000 0.364 0.208 0.085 0.293 0.035 0.201 0.000 0.143 0.001 0.055 5=0.63; 11=0.07; ;λ12=0.30

    9 -0.0103 0.477 0.171 0.000 0.000 0.094 0.257

    10 -0.0224 0.000 0.000 0.420 0.000 0.580 0.000

    11 -0.1929 0.000 0.000 0.000 0.000 0.000 1.000

    12 -0.0800 0.347 0.000 0.000 0.159 0.475 0.019

    Table 8 displays the results for each DMU considering the optimistic (for lower limits of the ranges) and

    pessimistic (for upper limits of the ranges) perspectives, using the first phase of the two-phase method

    with = 5%, 10%, or 20%.

  • 15

    Table 8. Lower and upper limits for the utility loss dkopt *

    , dkpes*é

    ëêù

    ûú, for each DMU

    DMU 5% 10% 20%

    1 [-0.149;-0.043] [-0.206;0.007] [-0.321;0.092]

    2 [-0.140;-0.040] [-0.191;0.007] [-0.306;0.089]

    3 [-0.035;0.082] [-0.095;0.131] [-0.220;0.220]

    4 [-0.075;0.049] [-0.138;0.106] [-0.270;0.218]

    5 [-0.162;-0.037] [-0.229;0.021] [-0.362;0.137]

    6 [-0.066;0.108] [-0.156;0.192] [-0.336;0.360]

    7 [-0.081;0.075] [-0.164;0.142] [-0.331;0.268]

    8 [-0.018;0.115] [-0.084;0.181] [-0.217;0.285]

    9 [-0.072;0.051] [-0.135;0.111] [-0.280;0.230]

    10 [-0.118;0.072] [-0.275;0.217] [-0.407;0.327]

    11 [-0.288;-0.097] [-0.384;-0.002] [-0.574;0.189]

    12 [-0.163;0.002] [-0.247;0.082] [-0.420;0.231]

    According to the rank previously made and for a tolerance of 5%, DMUs 11, 5, 1 and 2 are surely

    (robustly) efficient. The DMUs 12, 10, 4, 9 and 7 do not remain always efficient for the same tolerance

    value. For a tolerance of 10% only DMU11 maintains the efficiency status.

    Using a bisection technique we compute the maximum tolerance value for which the efficient DMUs

    maintain the efficiency status. Observing the stability interval limits shown in Table 9,and comparing

    with the rank established based on the nominal values of DMUs, we can conclude that the most robust

    DMU is DMU11, which coincides with the analysis based on super-efficiency. But the second most

    robust one is DMU1, which was ranked after DMU5 in the super-efficiency ranking.

    Table 9. Efficiency threshold for the efficient DMUs

    DMU 1 2 4 5 7 9 10 11 12

    δmax 9.2% 9.0% 1.0% 8.2% 0.0% 0.8% 1.1% 10.1% 4.9%

    This type of analysis can be better understood when accompanied by the graph in Figure 3 where the

    behaviour of dkpes*(pessimistic assessment) is illustrated. In fact DMU11has a better efficiency measure

    and maintains the efficiency for a greater level of uncertainty in performances (δmax

    = 10.1%). It is also

    possible to see that despite some DMUs had a better level of efficiency, with respect to the nominal

    situation δ = 0%, they are overtaken by others with a lower efficiency measure when more uncertainty in

    the performance of these units is assumed. This is the case of DMU10, which has a lower dkpes* for a

    tolerance of less than 5.4% and it is surpassed by DMU7 for higher tolerance values.

  • 16

    Figure 3. Values approximated by linear interpolation of dkpes*

    for DMUs 1, 2, 4, 5,7,9,10,11 and 12

    We can also observe that for linear utility functions the curve that represents the optimal value dkpes* is

    concave. In Appendix Awe prove that if all utility functions are linear, then the function that describes

    how dkpes*changes with is concave.

    Concluding remarks and future work

    This work provides a robustness analysis of each DMU in presence of interval data. A preliminary

    assessment of the robustness of each DMU is obtained using the first phase of a two-phase method with a

    modification to include the super-efficiency of efficient units. This method projects the DMUs in any

    direction in a way that minimizes the distance of this unit to the best of all (excluding the one under

    evaluation), and therefore no infeasibility occurs.

    Assuming that the values of the DMU performances in each factor (inputs and outputs) are not known

    exactly, but an interval of values for these performances can be established, it is possible to calculate an

    efficiency range for each DMU. The efficiency scores for the DMU under analysis are computed

    considering its coefficients in the most unfavourable/favourable bounds and all the other DMU’s

    coefficients in their most favourable/unfavourable bounds, in order to assess the DMU’s robustness. This

    process enables to classify the DMUs as surely efficient, potentially efficient, or surely inefficient.

    The maximum tolerance d such that the efficient DMUs maintain efficiency is also computed, by using a

    bisection technique. An illustrative example shows that, according to this robustness measure, the DMU

    with the highest super-efficiency score is not necessarily the most robust one, i.e., the one with widest

    stability intervals.

  • 17

    In future work we intend to include in the model a way to calculate this tolerance given different types of

    utility functions.

    Appendix A

    Let us suppose that the under analysis, DMU k, is improved by dz% while all other DMUs are changed

    by dz% in the opposite direction. The problem in phase 1 is, for this constant dz:

    Considering that c=1,...,a are output factors and c=a +1,...,qare input factors, we have:

    min dk* dz( )

    s.t. wcpcj (1-dz)- Mc

    L

    McU - Mc

    Lc=1

    a

    å - wcpkj (1+dz)- Mc

    L

    McU - Mc

    Lc=1

    a

    å +

    wcMc

    U - pcj (1-dz)

    McU - Mc

    Lc=a+1

    q

    å - wcMc

    U - pkj (1+dz)

    McU - Mc

    Lc=a+1

    q

    å ≤ dk* dz( ) ,jk (1’)

    wc.1=1

    wc ³ 0 , dk* dz( ) free

    LEMMA 1.

    Let dk* dz( ) , wc

    *be an optimal solution to (1’). If the performance of DMU k in factor c, pkj , is altered in

    dz +q( )%, with in the set of real, then the optimal value to (1’) is at most

    dk*

    dz( ) -

    w*q

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +w

    *q

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å .

    PROOF

    Problem with DMU k altered in dz +q( )%:

    min dkq dz( )

    s.t. wcq pcj (1-dz -q )- Mc

    L

    McU - Mc

    Lc=1

    a

    å - wcq pkj (1+dz +q )- Mc

    L

    McU - Mc

    L+

    c=1

    a

    å

    wcq Mc

    U - pcj (1-dz -q )

    McU - Mc

    Lc=a+1

    q

    å - wcq Mc

    U - pkj (1+dz +q )

    McU - Mc

    Lc=a+1

    q

    å ≤ dkq dz( ) , jk (2’)

    wcq .1=1, wc

    q ³ 0 , dkq dz( ) free

    Let dk* dz( ) , wc

    *be an optimal solution to (1’). Let nkkB ,..,1,1,...,1 be the set of indices of

  • 18

    DMUs for which there is no slack in (1’). Note that B , otherwise dk* dz( ) , wc

    *would not be

    the optimal solution (it would be possible to have a smaller dk* dz( ) . So,

    wc*pcj

    McU - Mc

    Lc=1

    a

    å -wc

    *pcjdz

    McU - Mc

    Lc=1

    a

    å -wc

    *pkj

    McU - Mc

    Lc=1

    a

    å -wc

    *pkjdz

    McU - Mc

    Lc=1

    a

    å -

    wc*pcj

    McU - Mc

    Lc=a+1

    q

    å +wc

    *pcjdz

    McU - Mc

    Lc=a+1

    q

    å +wc

    *pkj

    McU - Mc

    Lc=a+1

    q

    å +wc

    *pkjdz

    McU - Mc

    Lc=a+1

    q

    å = dk*

    dz( ),"j Î B

    wc*pcj

    McU - Mc

    Lc=1

    a

    å -wc

    *pcjdz

    McU - Mc

    Lc=1

    a

    å -wc

    *pkj

    McU - Mc

    Lc=1

    a

    å -wc

    *pkjdz

    McU - Mc

    Lc=1

    a

    å -

    wc*pcj

    McU - Mc

    Lc=a+1

    q

    å +wc

    *pcjdz

    McU - Mc

    Lc=a+1

    q

    å +wc

    *pkj

    McU - Mc

    Lc=a+1

    q

    å +wc

    *pkjdz

    McU - Mc

    Lc=a+1

    q

    å < dk*

    dz( ),"j Ï B

    Let

    Dj

    = wc* pcj (1-dz -q )- Mc

    L

    McU - Mc

    Lc=1

    a

    å - wc* pkj (1+dz +q )- Mc

    L

    McU - Mc

    L+

    c=1

    a

    å

    wc* Mc

    U - pcj (1-dz -q )

    McU - Mc

    Lc=a+1

    q

    å - wc* Mc

    U - pkj (1+dz +q )

    McU - Mc

    Lc=a+1

    q

    å

    = dk*

    dz( ) -

    wc*q

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *q

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å .

    Therefore,

    "j Î B, Dj

    = dk*

    dz( ) -

    wc*q

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *q

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å .

    "j Ï B, Dj

    < dk*

    dz( ) -

    wc*q

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *q

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å .

    Thus, dkq dz( ) = dk

    *dz( ) -

    wc*q

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *q

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å and wcq

    = wc*is an

    admissible solution to (2’), and it is found an upper bound for the optimal value of (2’). Whereas

    the optimal value to (2’) is denoted by dk*(d

    z+q ), then,

    dk*(d

    z+q ) £ dk

    *dz( ) -

    wc*q

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *q

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å .

  • 19

    Using LEMMA 1. we are now able to prove that the dk* dz( ) is a concave function of dz, as it was our

    purpose.

    PROPOSITION 1.

    dk* dz( ) , the solution to LP (1’), is a concave function of dz

    PROOF

    Let zyx

    ,, be possible values for the tolerance applied to the performances of each factor,

    considering any DMU k, such that A=dy -dxand dz = tdx + t 1- t( )dy , with t Î 0,1] [.

    Given proposition 1, if we consider =-(1-t)A, we have:

    dk*(d

    z- (1- t)A) £ dk

    *dz( ) +

    wc* 1- t( ) A

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å -wc

    * 1- t( ) A

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å and, on the other hand,

    if we do=t.A, we have: dk*(d

    z+ t.A) £ dk

    *dz( ) -

    wc*t.A

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *t.A

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å ,

    with A=dy -dxand dz = tdx + t 1- t( )dy , t Î 0,1] [.

    So, considering dx = dz- 1- t( ) A and dy = dz+ t.A and given what was stated earlier, we get

    dk*(d

    x) £ dk

    *dz( )+

    wc* 1- t( ) A

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å -wc

    * 1- t( ) A

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å and

    dk*(d

    y) £ dk

    *dz( ) -

    wc*t.A

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *t.A

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    å .

    Considering now the definition of concave function and the upper limits obtained, we have:

    t.dk*

    dx( ) + 1- t( ).dk

    *dy( ) £ t dk* dz( ) +

    wc* 1- t( ) A

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å -wc

    * 1- t( ) A

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    åé

    ë

    êê

    ù

    û

    úú+

    (1- t) dk*

    dz( ) -

    wc*t.A

    McU - Mc

    Lpcj + pkj( )

    c=1

    a

    å +wc

    *t.A

    McU - Mc

    Lpcj + pkj( )

    c=a+1

    q

    åé

    ëê

    ù

    ûú = dk

    *dz( )

    It follows that the function dk* dz( ) is concave.

  • 20

    REFERENCES

    Ali A I, Lerme C S and Seiford L (1995). Components of efficiency evaluation in Data Envelopment

    Analysis. European Journal of Operational Research 80: 462-473.

    Almeida P N and Dias L C (2011). Value-based DEA models: application-driven developments. Journal

    of the Operational Research Society 63: 16-27, 2012.

    Andersen P and Petersen N C (1993). A procedure for ranking efficient units in data envelopment

    analysis. Management Science 39: 1261-1264.

    Athanassopoulos A D and Podinovsky V V (1997). Dominance and potential optimality multiple criteria

    models decision analysis with imprecise information. Journal of the Operational Research Society 48:

    142-150.

    Banker R D, Charnes A and Cooper W W (1984). Some models estimating technical and scale

    inefficiencies in Data Envelopment Analysis. Management Science 3: 1078-1092.

    Bouyssou D (1999). Using DEA as a tool for MCDM: some remarks. Journal of the Operational

    Research Society 50: 974-978.

    Charnes A, Cooper W W and Rhodes E (1978). Measuring the efficiency of decision making units.

    European Journal of Operational Research 2: 429-444.

    Charnes A, Cooper W W, Golany B, Seiford L and Stutz J (1985). Foundations of Data Envelopment

    Analysis for Pareto-Koopmans efficient empirical production functions. Journal of Econometrics 30: 91-

    107.

    Cooper W W, Seiford L M and Tone K (2006). Introduction to Data Envelopment Analysis and its uses.

    Springer: New York.

    Cooper W W, Park K S and Yu G (1999). IDEA and AR-IDEA: models for dealing with imprecise data

    in DEA. Management Science 45: 597–607.

    Cooper W W, Park K S and Yu G (2001a). An illustrative application of IDEA (imprecise data

    envelopment analysis) to a Korean mobile telecommunication company. Operations Research 49: 807-

    820.

    Cooper W W, Park K S and Yu G (2001b). IDEA (imprecise data envelopment analysis) with CMDs

    (column maximum decision making units). Journal of the Operational Research Society 52: 176–181.

    Despotis D K and Smirlis YG (2002). Data envelopment analysis with imprecise data. European Journal

    Operational Research 140: 24-26.

    Dias L C and Clímaco J N (2000). Additive Aggregation with Variable Interdependent Parameters: the

    VIP Analysis Software. Journal of the Operational Research Society 51: 1070-1082.

    Dulá J H and Hickman B L (1997). Effects for excluding the column being scored from DEA

    envelopment LP technology matrix, Journal of the Operational Research Society 48: 1001-1012.

    Gouveia MC, Dias L C and Antunes CH (2008). Additive DEA based on MCDA with imprecise

    information. Journal of the Operational Research Society 59: 54-63.

  • 21

    Keeney R L and Raiffa H (1976). Decisions with multiple objectives: preferences and value tradeoff.

    Wiley: NewYork.

    Seiford L M and Zhu J (1998a). Stability regions for maintaining efficiency in data envelopment analysis.

    European Journal of Operational Research 108: 127-139.

    Seiford L M and Zhu J (1998b). Sensitivity analysis of DEA models for simultaneous changes in all the

    data. Journal of the Operational Research Society 49: 1060-1071.

    Seiford L M and Zhu J (1999a). Infeasibility of super efficiency data envelopment analysis models.

    INFOR 37: 174-187.

    Seiford L M and Zhu J (1999b). An investigation of returns to scale in data envelopment analysis. Omega

    27: 1-11.

    Thrall R M (1996). Duality, classifications and slacks in DEA. Annals of Operations Research 66: 109-

    138.

    Thanassoulis E, Portela M C and Allen R (2004). Incorporating value judgments in DEA. In: Cooper W

    W, Seiford L M and Zhu J (eds). Handbook on Data Envelopment Analysis, Kluwer: Boston, pp 99-137.

    Von Winterfeldt D and Edwards W (1986). Decision analysis behavioural research. Cambridge

    University Press: New York.

    Zhu J (1996). Robustness of the efficient DMUs in data envelopment analysis. European Journal of

    Operational Research 90: 451-460.

    Zhu J (2001). Super-efficiency and DEA sensitivity analysis. European Journal of Operational Research

    129: 443-455.

    Zhu J (2003). Imprecise data envelopment analysis (IDEA): A review and improvement with an

    application. European Journal of Operational Research 144: 513-529.


Recommended