+ All Categories
Home > Documents > Mata Architecture for Future Internet

Mata Architecture for Future Internet

Date post: 11-Feb-2022
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
39
Performance evaluation of Tour de France cycling teams using Data Envelopment Analysis Nicky Rogge Daam Van Reeth Tom Van Puyenbroeck HUB RESEARCH PAPERS 2012/12 ECONOMICS & MANAGEMENT FEBRUARI 2012
Transcript
Page 1: Mata Architecture for Future Internet

Performance evaluation of Tour de France

cycling teams using Data Envelopment Analysis

Nicky Rogge Daam Van Reeth

Tom Van Puyenbroeck

HUB RESEARCH PAPERS 2012/12 ECONOMICS & MANAGEMENT

FEBRUARI 2012

Page 2: Mata Architecture for Future Internet

1

Performance evaluation of Tour de France cycling teams using

Data Envelopment Analysis

Nicky Rogge*‡, Daam Van Reeth*, and Tom Van Puyenbroeck*

(*): Hogeschool-Universiteit Brussel (HUBrussel)

Warmoesberg 26, 1000 Brussels (Belgium)

(‡): Katholieke Universiteit Leuven (KULeuven)

Faculty of Business and Economics

Naamsestraat 69, 3000 Leuven (Belgium)

Abstract

This paper uses a robust (order-m) Data Envelopment Analysis approach to

evaluate the efficiency of Tour de France cycling teams for the period 2007-

2011. Since there are multiple ways in which this event can be successful for

a cycling team, we take it that managers face strategic input decisions

regarding team and rider characteristics. Specifically, we distinguish between

ranking teams, sprint teams, and mixed teams, and compute for each of these

an efficiency score as due to the team’s performance relative to similarly

classified teams and an efficiency score that is the consequence of the team

type. We find that ranking teams are generally more efficient than other

types.

Keywords: Data Envelopment Analysis, Tour de France, Cycling, Team

Types, Performance Evaluation; Robust order-m

JEL-classification: L83, D24, L23.

Corresponding author: Daam Van Reeth, Hogeschool-Universiteit Brussel, Warmoesberg 26, 1000 Brussels,

Belgium, Email: [email protected]

Page 3: Mata Architecture for Future Internet

2

1. Introduction

The Tour de France is the most important cycling race in the world. The three-week race

captures the interest of millions of cycling fans every day. Alongside the cobblestones classics Paris-

Roubaix and the Ronde van Vlaanderen, it is the only cycling race to get worldwide media coverage.

This significant exposure renders the Tour de France the primary season goal for most top cycling

teams and motivates managers to line up their best riders. In fact, Tour de France success sometimes

makes all the difference for a continued sponsorship of a professional cycling team. An analysis of the

performance of Tour de France cycling teams therefore clearly is appropriate.

Performance studies of professional road cycling are, however, still rare in the economics of

sports literature. This is partly due to two distinguishing features of cycling that make such analyses

less straightforward. First, cycling is a sport in which one individual receives the glory of team

production. It is "Lance Armstrong" who appears 7 times on the roll of honour of the Tour de France,

not "US Postal Service" or any of the other teams he was part of when winning the Tour de France.

Still, it would have been impossible for Lance Armstrong to win the Tour de France without the

support of a strong team. Second, there are many prizes to be won in a multi-stage race like the Tour

de France. Evidently, not all cycling teams are able to win the prestigious overall time classification,

but they can still strive for important secondary prizes. This will have an impact on their team

selection. Hence, an appropriate performance analysis should account both for the multiple ways of

being successful in the Tour as well as for the different goals that cycling teams may have in that

respect.

Our focus on teams is one obvious way in which this paper differs from the small body of

literature that has addressed professional road cycling performance earlier. Prinz (2005) and Torgler

(2007) do use team data, but only as explanatory variables for individual performances. Both Sterken

(2005) and Cherchye & Vermeulen (2006) entirely abstract from team importance in their analysis.

Note that team efficiency analyses are rather common in the sports economics literature, especially for

popular European and American team sports. We refer, for instance, to the papers on soccer by

Page 4: Mata Architecture for Future Internet

3

Dawson et al.(2000), Carmichael et al. (2001), Haas (2003), Espitia-Escuer & Garcia-Cebrian (2004)

and Pestana Barros & Leach (2006). Cooper et al. (2009) and Rimler et al. (2010) analyse efficiency

in basketball, while the baseball and American football team efficiency is addressed by Hadley et al.

(2000), Einolf (2004), Lewis et al. (2007) and Collier et al. (2011).

The multiple-prize nature of the Tour de France is not taken up by Sterken (2005) and Torgler

(2007), who focus on the overall time classification only. Both Prinz (2005), by using total prize

money, and Cherchye & Vermeulen (2006), by aggregating information from different rankings into

one ranking, do embed the multiple prizes context into the analysis. Still, by focusing on individuals,

their assessments neglect an important real element of strategic positioning that is largely at a team

manager’s discretion.

In this paper, we thus broaden the scope of earlier analyses, which focused on individual

performance, used a single (simple or comprehensive) output statistic, and abstracted from ex ante

diversity of team aspirations. Instead, we take it here that actual management decisions relate to

combining individual rider’s quality and experience so as to maximize the chance of winning during

the three-week event, where prizes range from the cherished end-prize, over important side-

competitions (e.g. the best climber) and single stage wins, to minor prizes (e.g. the daily ‘prix de la

combativité’ awarded by a jury for the most combative rider). Adhering to this managerial

perspective, we analyse team efficiency by using a Data Envelopment Analysis (DEA) approach. An

important advantage of our particular set-up is that it enables identifying which part of the

inefficiency is particularly due to the team’s performance relative to its peers, and what part of the

inefficiency is due to the fact that the team is of a certain type. It should be noted further that we have

opted for a robust version of DEA, so as to deal with potential biases in efficiency estimates

originating either from unequally sized team-type groups or from atypical/outlying performances.

We succinctly describe some essential features of the Tour de France in section 2. Section 3

presents the data, while the methodology is explained in section 4. The empirical results are discussed

in section 5. Concluding remarks are gathered in section 6.

Page 5: Mata Architecture for Future Internet

4

2. The Tour de France

The Tour de France has a long and heroic history going back more than a century. The race lasts

for three weeks, counting 21 stages that differ in length and route. The rider who has taken the least

time overall to cover the entire course wins the Tour de France. Three types of stages exist, each

suitable for a different type of rider. Flat stages are easy to ride and are likely to end in a bunch sprint.

They are of almost no importance to the overall Tour de France win, although occasionally top

favourites suffer time losses when crashes or windy conditions result in peloton splits1. When riders

have to climb steep mountain passes, the race becomes much harder and usually large time

differences between riders are created. Mountain stages are therefore crucial to the overall Tour de

France result. Typically, riders who perform well in flat stages face huge difficulties in mountain

stages and may even struggle to avoid disqualification for arriving out of time. While flat stages and

mountain stages are raced collectively, time trial stages are ridden individually or, sometimes, in

teams. Cyclists start at equal time intervals, usually two or three minutes apart. As time trial stages

might result in significant time differences between riders, they are also very important to the overall

Tour de France result.

The overall time classification is the most prestigious classification to win, the leader wearing the

famous yellow jersey. But there are other competitions of importance too. The points classification,

awarded with a green jersey, is designed so as to award the best sprinter, the rider who is best capable

of winning or ending in top positions in stages where most of the peloton is still together at the finish.

A polka-dot jersey is worn by the leader in the competition for the best climber, while the best young

rider, i.e. under 25 years of age, receives a distinctive plain white jersey. Next to these individual

classifications, there is also a team classification based on the time of each team's best three riders in

every stage. Not all participating teams have a strong enough contender for an overall classification.

These teams might focus on individual stage victories instead. But with 22 teams participating and

only 5 classifications at stake, the stage victories are also hard-fought. In fact, usually about half of

the teams leave the Tour de France empty-handed.

Page 6: Mata Architecture for Future Internet

5

The mix of several competitions and different types of stages creates a multi-output situation

with many success opportunities. As a result, there are typically three types of teams present in the

Tour de France. Ranking teams aim for the end victory (the yellow jersey) or at least a top final rank

position. Sprint teams particularly compete for stage victories in stages that are well-suited for

sprinters and the green jersey. The third type gathers cycling teams that do not really aim for a good

final ranking position or for winning sprint stages. Mixed teams particularly look for opportunities

during the Tour de France to win a stage that is somewhat in between a sprint stage and a mountain

stage (i.e., the typical transition stages in the Tour de France) or aim for a share in minor prizes.

Teams that have as predominant objective to compete for the end victory or a good final rank

position, like team RadioShack, have a line-up that qualitatively differs from that of teams whose

main goal is to win sprint stages and/or the green jersey, like HTC Highroad. Specifically, a ranking

team will choose riders that are helpful to the team leader throughout the Tour, sheltering him from

the wind in the flat stages and supporting him as strongly as possible in the difficult mountain stages.

A sprint team, on the other hand, will for the most part select riders that are skilled in holding the

peloton together and preparing the sprint for their leading rider.

3. Data

3.1. Teams

The number of teams participating in the Tour de France varies slightly from year to year

between a minimum of 20 and a maximum of 22. In this study we analyse the 2007-2011 period,

leading to a dataset of 105 observations. During this period in total 31 different cycling teams took

part at least once in the Tour de France. A five-year period was chosen for two reasons. First, as

compared to other sports, most cycling teams have a rather short lifespan of on average 5 to 10 years

only. By limiting the analysis to 5 years, our sample contains a large enough number of cycling teams

that participated in all Tours de France under study. Table 1 indeed shows that no less than thirteen

Page 7: Mata Architecture for Future Internet

6

cycling teams were always present.2 This means that, at the team-level, the Tour peloton remained for

60 to 65% unaltered throughout the years. With a longer sample period of, for instance, 10 years, a

much smaller portion of our dataset would consist of teams that participated in all Tours. Second,

during the 2007-2011 period there were neither significant changes in the prize structure nor in the

number of stages (21) and competitions (5). Since we use these variables as outputs, the unaltered

prize structure allows for a better interpretation of the results. This would be more difficult if we

opted for a longer period.

All cycling teams were subsequently grouped into three team types: ranking teams, sprint teams

and mixed teams. Teams were classified by analysing past Tour de France performances of the

individual riders selected in their Tour de France team. To qualify as a ranking team in year t, for

instance, a team had to line up at least one rider that finished in the Tour de France top 10 in year t-1

or t-2, or in the Tour de France top 5 in year t-3 or t-4. A similar procedure, based on winning sprint

stages in the past, was used to determine sprint teams. If a team qualified for both team types, as for

instance the Rabobank team that during the period 2007-2010 selected overall contender Denis

Menchov as well as sprint winner Oscar Freire, it was categorized as a ranking team, the reason for

this choice being that the final general classification still is the most important prize a team can win.

Finally, teams that didn't qualify for either category were considered mixed teams.3

< Table 1 about here >

Table 1 summarizes the information. It shows that our dataset contains 46 ranking teams, 24

sprint teams and 35 mixed teams. Some teams always were of the same type throughout the sample

period, like Team Saxo Bank (a ranking team for all years), but most teams change type from year to

year, depending on their team composition. Liquigas, for instance, has twice been labelled as a mixed

team first, became a sprint team in 2009, and was twice a ranking team afterwards.

Page 8: Mata Architecture for Future Internet

7

3.2. Input variables

Apart from random elements like luck and coincidence, success in the Tour de France depends

on a number of team and rider characteristics. We use these characteristics as our input variables in

the DEA analysis. We discern two main categories of input variables: team quality and team

experience. Team quality is measured by the team budget (financial input) and by the number of

Cycling Quotient (CQ) points of the riders selected for the Tour team as earned on the eve of the Tour

(sports input).4 We prefer to use CQ points because they allow comparing team quality over the years,

which is not the case with other point rankings in cycling that often have different point scales from

year to year. Data were collected from the CQ website (www.cqranking.com). We also consider an

alternative in which the CQ points of a team’s Tour selection are expressed as a percentage of the

team total number of CQ points. Team experience is based on past Tour de France performances.

Specifically, the total number of Tour starts by the selected riders were computed. In addition, we

look at team consistency to capture team rather than individual experience. We construct this variable

as the maximum number of riders in a particular Tour team in year t that also rode together in any

Tour team in year t-1. In the rare event a team was temporarily banned, year t-2 was used as a

reference instead. In fact, this correction was only necessary for the Astana team that was denied entry

in the 2008 Tour de France. The Tour team consistency variable measures the importance in cycling

of building a core of domestique riders around a team leader. George Hincapie, for instance, was the

most important domestique to Lance Armstrong, and was part of all his winning Tour teams.

< Table 2 about here >

Table 2 summarizes the input data for all teams and for the subsets of team types. The average

team budget amounts to 8.2 million euro, but ranking teams have a 10 to 15% higher budget, on

average, than sprint teams and mixed teams. There is an almost 5:1 ratio between the richest teams

(Katusha and Team HTC Highroad, 15 million euro) and the poorest teams (Barloworld and

Page 9: Mata Architecture for Future Internet

8

Agritubel, 3.5 million euro). The CQ points show a similar but more outspoken pattern. The riders in

a Tour team are on average worth around 3,700 CQ points. Ranking teams have a CQ value of around

4,500 points, which is almost a third more than sprint teams (3,500 points). Mixed teams only

accumulate about 2,900 points. This results in an almost 10:1 ratio between the strongest team (Team

Saxo Bank, 7,834 points) and the weakest team (Footon Servetto, 879 points). If we look at the CQ

points of the selected riders in the Tour de France team relative to the total number of CQ points in

the cycling team, we notice that the average Tour team represents slightly more than 50% of the

team’s total CQ points. Ranking teams have marginally higher percentage values, on average, than the

other two types. Nevertheless, as can be seen from the minimum and maximum observed percentage

values, there are considerable differences among teams of all types. The nine riders in a Tour de

France cycling team combine on average a total of about 33 Tour starts. Since the start in year t is

included in this total, this means a typical rider has an experience of 2 to 3 previous Tours. Tour

experience does, however, differ importantly between types of teams. Ranking teams (39.74) have far

more experience than sprint teams (30.21) and mixed teams (26.89). At the team level, we note a big

difference between both Skil-Shimano (2009) and Footon-Servetto (2010) that had no experienced

riders in their Tour team, and Team Leopard-Trek (2011) that is the most experienced team in our

sample with 52 previous Tour starts by its nine riders. Finally, with average Tour team consistencies

of 4.89 and 4.88 (meaning that on average about 5 teammates rode the Tour de France together in the

same team the previous year), ranking teams and sprint teams are overall equally consistent

concerning team line-up, while mixed teams are considerably less consistent (3.57).

< Table 3 about here >

We have conducted DEA-based performance analyses of Tour de France cycling teams using

five different input selections, as denoted in Table 3. This allows checking for consistency in the

performance evaluation outcomes. In particular, when results would be sensitive to the used input-

output mix in the DEA-model, results should be treated with much caution.

Page 10: Mata Architecture for Future Internet

9

3.3. Output variables

The complex nature of cycling makes it less straightforward to determine the relevant output

variables necessary for a DEA analysis. Essentially, the results of the Tour de France consist of a

number of ordinal time or points based rankings. However, using the time differences or the points as

output is not very relevant because, just as in many other sports, the order rather than the distance

between two riders ultimately matters. Furthermore, point scales may differ from year to year and

time differences are very much dependent on the race evolution and on the features of the yearly

changing route, which makes a multi-year analysis of such statistics cumbersome at best. Conversely,

sticking to ordinal rankings is also not preferable as it does not allow to discriminate between a one

position difference high in the ranking (e.g. between the winner and the runner-up) and a one position

difference lower in the ranking (e.g. between 100 and 101). Additionally, it is difficult to reconcile

with the all important team element in professional road cycling.

Therefore, we propose three other output variables, all aggregated at the team level: prize money

earned, CQ points collected during the Tour, and prizes won.5 The first two output variables are

related to the ranking but give higher weights to better positions by lowering in a degressive way the

rewards for lower positions. Furthermore, they implicitly weigh the importance of the different prizes

that can be won, by differentiating the rewards between classifications. Data on prize money are made

public each year by the Tour de France. Note that in addition to prize money, all Tour de France

cycling teams also get a lump sump for what is called "participation expenses" and a "presence

bonus", totalling 65,643 euro in 2011. Because this sum is almost unrelated to the team's

performances (it is reduced marginally with 1,600 euro for every rider unsuccessful in finishing the

Tour de France), we did not include it in our total prize money measure. In their capacity as an output

indicator, data on CQ points refer to points as collected during the Tour (whereas, to repeat, the input

counterpart captures the stock of points of a team prior to that Tour). A comparison between the prize

money scale and the CQ points scale reveals an important difference. The Tour de France

organisation pays a stage winner 8,000 euro only, which is less than 2% of the reward the overall

Page 11: Mata Architecture for Future Internet

10

Tour de France winner gets (450,000 euro). In the CQ ranking, a stage win earns a rider 80 points,

which is over 13% of the points the overall Tour de France winner receives (600 points). Our third

output variable counts the number of prizes won by a cycling team out of a total of 30 possible

important prizes, viz., 21 stage wins, 5 competitions to win (general ranking, points classification,

mountain classification, best young rider and team classification) and the 2nd to 5th place in the final

general classification. As opposed to the first two output variables, this variable gives all prizes the

same weight. The prize money output variable views cycling teams primarily as revenue maximizing

organisations, while the prize count output variable is more in line with a win maximizing situation.

The CQ points variable should be situated somewhere in between the revenue and win maximizing

situation.

Contrary to how we handled inputs, and given the clear importance of all three indicators as race-

related outcomes of the event, we retain all three outputs in the different models. Evidently though,

there is another output that is of special interest to cycling teams and, especially, their sponsors. As

cyclists are promoting these sponsors through the advertising on their shirts as well as by their very

team names, media exposure is very important to a professional cycling team. By way of anecdotic

evidence, while Team Saxo Bank won the 2011 edition of the one-day classic Tour of Flanders with

Nick Nuyens, it only had about 2% of all media exposure through live TV coverage. The team of

runner-up Sylvain Chavanel, Quick-Step, took the bulk of the media attention with over 10% of

exposure. Although Team Saxo Bank won the race, commercially Quick-Step was the more

successful team that day. Unfortunately, media exposure information is unavailable publicly and

could therefore not be used as an output variable in our analysis.

Table 2 also summarizes the output data for all teams and for the subsets of team types. The

average prize money earned by a Tour de France team is almost 100,000 euro. Big differences exist

between teams. The most money was won by Discovery Channel in 2007. The victory by Alberto

Contador earned the team 723,640 euro, more than 70 times the amount Lampre won in 2008, only

9,840 euro. The Tour de France organisation clearly rewards the overall ranking higher than the other

rankings. While ranking teams earn on average 150,000 euro, sprint and mixed teams only receive on

Page 12: Mata Architecture for Future Internet

11

average 60,000 euro and 50,000 euro of prize money, respectively, for three weeks of hard work. CQ

points awarded for performances in the Tour de France show a similar pattern. On average, teams

gain 620 CQ points in the Tour de France. Ranking teams (822 points) outperform sprint teams (505

points) and mixed teams (434 points). At the team level the differences are very high. In 2009, Astana

was awarded more than 200 times the number of points of Cofidis in 2007: 2059 versus 70. Cycling

teams win on average 1.43 prizes, with a big difference between ranking teams (2.07 prizes) and

sprint teams or mixed teams (0.92 and 0.94). This average is misleading, though, because many teams

are unable to win any prizes at all. Indeed, in our 5 year sample, in 45 out of 105 observations (43%) a

team left the Tour de France empty-handed. Team Saxo Bank in 2008 is the most successful team in

our dataset with 8 prizes.

In sum, Table 2 provides an indication of a positive relation between team inputs and team

output measures for the Tour de France. Looking at the summary statistics in the table, ranking teams

typically use somewhat more inputs compared to other team types, but realize considerably higher

outputs. This signals that, at least at this aggregate level, one could expect type differences in relative

efficiencies (i.e., in aggregated output-input ratio’s) as well. Yet, a comparison of the minimum

observed output values among ranking teams with those of sprint teams shows that there are ranking

teams that realize lower outputs. Consequently, even though ranking teams on average probably do

better in terms of relative efficiency, there will be considerable differences between ranking teams,

some of them even performing worse than the other type teams.

4. Methodology

4.1. Data Envelopment Analysis

To evaluate the efficiency of the Tour de France cycling teams, we use a DEA approach. DEA

was introduced by Farrell (1957) and further developed by Charnes et al. (1978). In its most general

form, DEA can be seen as a non-parametric technique for measuring the technical efficiency of

Page 13: Mata Architecture for Future Internet

12

similar activity units which utilize (possibly multiple) inputs to produce (possibly multiple) outputs in

environments typically characterized by no reliable information on the prices of inputs and outputs

and/or no (exact) knowledge about the ‘functional form’ of the production or cost function. In the

literature, the activity units are generally referred to by the term Decision Making Units or DMUs. In

our example, it concerns the Tour de France cycling teams in the sample.

The key feature of DEA is that, in the absence of a priori detailed information on the exact

specification of the cost or production function, it estimates the cost or production function from the

observed input and output data themselves (i.e., a posteriori). DEA determines the efficiency frontier

as the piece-wise linear combination that envelops the input-output combinations as observed from

the empirical data. Consequently, only the most efficient cycling teams are situated on this “best

practice” efficiency frontier. It is with respect to this data-driven efficiency frontier that the DEA-

model estimates the efficiency of each activity unit in the sample. More precisely, the DEA-model

computes for each evaluated cycling team the efficiency score as a (radial) distance measure of the

team’s actual position relative to the efficiency frontier. Formally, this yields the following form of

the DEA linear programming model:

( )

1

1

1 2

1 2 1

0 1 2

k

n

k , j j ,i k ,i

j

n

k k , j j ,q k k ,q

j

k , j

x x i , , , p

max s.t. y y q , , ,s

j , , ,n

ϕ

λ

ϕ λ ϕ

λ

=

=

≤ =

≥ = ≥ =

In this formulation, we assume that there are n cycling teams ( )1j ,...,n= that use p inputs j ,ix

( )1i , , p= … in their attempt to realize good performances on the s output criteria j ,qy ( )1q , ,s= … .

In our application of the Tour de France cycling teams for the period 2007-2011, there are 105 cycling

teams (i.e., 105n = ). For instance, recalling Table 3, in our model 5 each of these teams uses four

inputs, viz. ‘budget’ ( 1j ,x ), ‘CQ points in Tour team’ ( 2j ,x ), ‘number of Tour starts’ ( 3j ,x ), and ‘Tour

Page 14: Mata Architecture for Future Internet

13

team consistency’ ( 4j ,x ) (hence, 4p = ), to compete with other cycling teams in order to do well in

the three output criteria ‘prize money’ ( 1j ,y ), ‘CQ points collected’ ( 2,jy ), and ‘prizes’ ( 3,jy ) (thus,

3s = ). We assume that inputs and outputs are non-negative: 0j ,ix ≥ and 0j ,qy ≥ .

The linear program as in ( )1 is computed n times: one computation per evaluated cycling team.

In each linear programming computation, the input and outputs of the cycling team that is singled out

are denoted respectively by k ,ix and qky , . As can be seen from the program, one basically looks for

the maximal (equi-)proportional expansion in all outputs of this evaluated team ( qkk y ,ϕ for all q),

while keeping its inputs fixed, such that the resulting benchmark projection uses at least as much

inputs and at most as much outputs as a (λ-) weighted combination of the (observed) input-output

combinations of all teams. The efficiency score of a team is then defined as the inverse of this

maximal equi-proportional expansion factor ( kk ϕθ /1≡ ), and is hence situated between zero and one

(i.e., 0 1kθ≤ ≤ ), with higher values indicating more efficient performances. According to this logic,

cycling teams with 1kθ = are perfectly efficient. The λ–vector is the vector of intensities, indicating

how important the other cycling teams in the sample are for constituting the benchmark against which

the evaluated cycling team k is assessed. The subscript “ k ” in k , jλ indicates that the vector of

intensities is specific for each evaluated cycling team. Specifically, for the evaluated cycling team k ,

the intensity vector k , jλ will consist for the most part of zeros except for those cycling teams that are

situated on the efficiency frontier and play a role in defining the benchmark for cycling team k .

Accordingly, this benchmark is a weighted combination of the performances of efficient cycling

teams.

< Figure 1 about here >

The DEA-model can be easily illustrated. Figure 1 displays a fictitious example of 20 cycling

teams being evaluated based on two outputs and one input.6 Instead of requiring an explicit definition

Page 15: Mata Architecture for Future Internet

14

of the production function, the DEA-model reveals the efficiency frontier by just enveloping the data

points lying most outwards with linear segments. In the example, the cycling teams A , B , C , and D

are situated most outwards. Correspondingly, these teams constitute the empirical DEA-estimate of

the efficiency frontier (i.e., the piece-wise linear ABCD ) and their efficiency scores are equal to one.

They also act as their own benchmark (being situated themselves on the efficiency frontier). All other

teams are situated below the efficiency frontier and are thus inefficient (i.e., 1kθ < ). To determine the

exact value of kθ for the inefficient teams: (1) project the inefficient DMU on the efficiency frontier

in an equi-proportional way and (2) take ratios of radial distances. The projection point on the

efficiency frontier serves the purpose of efficient benchmark team. For the cycling team R , for

instance, the point R' on the efficiency frontier acts as benchmark and the DEA-estimate of kθ

equals the ratio OR OR' . The benchmark for teams R and S , respectively R' and S' , are situated on

the AB -part of the efficiency frontier ABCD , which means that they equal weighted combinations of

the performances of cycling teams A and B . Accordingly, in the DEA-evaluation of teams R and S ,

the intensity vectors R, jλ and S , jλ will consist of zero values for all units except for the teams A and

B . Further note that the value kθ indicates how much a team has to improve in order to become

efficient. The lower the value of kθ , the larger the room for efficiency improvements. When

comparing the efficiency scores of the cycling teams R and S , for example, we see that team S is

less efficient than team R (i.e., ( ) ( )OS OS' OR OR'< ) and thus there is more room for

improvement for team S .

To end this subsection, we point out that many alternative DEA-models exist, and that some of

these may also be suitable in a setting similar to ours. For example, Yang et al. (2011) present a

“fixed-output sum”-model, which hinges on the idea that expanding one’s output can only be

achieved by an offsetting decrease of other competitors’ output. Yang et al. (2011) illustrate their

model with (Sydney 2000) Olympic medals as the fixed-sum outputs. However, in our specific setting

the empirical value added of this additional assumption is small, as the best performers in the Tour

dataset are still far removed from the hypothetical bliss point in which the 9 riders of a team would

Page 16: Mata Architecture for Future Internet

15

collect all relevant prizes (30 per Tour), obtain their theoretically maximum CQ-points (8,095) and

prize money (1,542,660 euro). Indeed, Table 2 reveals that the actual best practices are still

considerably removed from these theoretical maxima. Thus, they can be taken as targets without the

need of adjusting their outputs downwards. We return to the issue of alternative models in the

concluding section.

4.2. Team efficiency vs. team-type efficiency

The existence of ranking teams, sprint teams, and mixed teams is a crucial element of our

specific setting. It entails that a ranking team will mainly compete with other ranking teams. In the

same vein, the main focus of all sprint teams consists in doing better than the other sprint teams (viz.,

winning sprinting stages and competing for the green jersey). In the basic DEA-model as presented in

the previous section, however, each team is evaluated relative to all other teams in the sample

(irrespective of the team type). This means that the basic version of the DEA-model does not account

for the several sub-competitions that exist between the cycling teams. In terms of Figure 1, this would

imply that a ranking team R might well be deemed inefficient based on a comparison with teams A

and B, which are not necessarily of the same type as R. Put in more abstract terms, and given that we

are gauging efficiency relative to existing best-practices, this feature implies that one should properly

address the selection of relevant peers.

To account for these sub-competitions (and, thus, to account for the several types of cycling

teams in the Tour de France peloton), we use a DEA-framework inspired by Portela and Thanassoulis

(2001).7 The key idea of their approach is straightforward: evaluate any cycling team within its own

group of cycling teams of the same type, and evaluate that team a second time within the complete

sample set. This essentially means that each cycling team is evaluated relative to two efficiency

frontiers: one frontier as determined by the teams that belong to the same group and another, overall

frontier as determined by all cycling teams irrespective of their type. Figure 2 below illustrates the

idea for the ranking teams vs. sprint teams.8

Page 17: Mata Architecture for Future Internet

16

< Figure 2 about here >

In Figure 2, both the ranking team R and the sprint team S are located below their respective

own team-type efficiency frontiers (for the ranking teams and sprint teams, the efficiency frontiers are

EFBCD and AGH respectively) and the overall efficiency frontier ABCD . Evidently, given the

existence of two relevant frontiers for each team, two DEA-based efficiency scores can now be

estimated. A first efficiency score is estimated within the sample of cycling teams of the same type.

We formally represent this efficiency score by Own

kθ and label it as “own-team-within-own-team-

type”-efficiency. For the ranking team R and the sprint team S , Own

kθ is computed respectively by

the ratios OR OR" and OS OS" . A second efficiency score is obtained when the team under

evaluation is assessed relative to all teams and, hence, the overall efficiency frontier ABCD . From

now on, we refer to these efficiency scores as the “overall” or “own-team-within-all-team-types”-

efficiency and represent it by Overall

kθ . For cycling teams R and S the overall efficiency scores are

equal to the ratios OR OR' and OS OS' . Combining the outcomes of the two aforementioned

efficiency analyses, we can derive a third efficiency score that is denoted by Type

kθ and is referred to as

“own-team-type-within-all-team-types”- efficiency. In particular, we compute the Type

kθ by taking the

ratio of Overall

kθ and Own

kθ :

( )2Overall

Type kk Own

k

θθ

θ=

Expression (2) implies that we can decompose the overall efficiency score Overall

kθ into an

inefficiency component that is due to the team’s performance relative to similar teams, measured by

Own

kθ , and an inefficiency component Type

kθ which, by construction, isolates the effect of using other

types as well as a basis for comparison. Applying this decomposition to the examples of cycling teams

Page 18: Mata Architecture for Future Internet

17

R and S yields own-team-type-within-all-team-types efficiency scores Type

kθ equal to OR" OR' and

OS" OS' . The two components of overall efficiency can be straightforwardly illustrated in Figure 2.

For the sprint team S , for instance, the overall efficiency score Overall

kOS OS'θ = as computed by the

basic version of the DEA-model is seen to consist of an efficiency component Own

kOS OS"θ = that is

due to team S ’s own performance relative to the sprint teams’ best practice and an efficiency

component Type

k OS" OS'θ = that originates from being a sprint team type.

4.3. Robustifying the performance evaluation: robust order-m DEA

Finally, we make one important adjustment so as to deal with two problems that are both related

to the relative perspective taken on in DEA-based efficiency evaluations. A first issue is that groups of

observations are of different size, as should be recalled from Table 1. These differences in group sizes

may cause a bias in the estimation of the different efficiency components. Specifically, given that the

“own-team-within-own-team-type”-efficiency measure is computed within the own group, a ranking

team is evaluated relative to 46 teams in the estimation of Own

kθ , while a sprint team is assessed

relative only to 24 teams. Obviously, this means that the probability for a sprint team of getting a high

Own

kθ is higher than for a ranking team (as the performance of the former is, ceteris paribus, compared

to a smaller group of peers). A second concern is that the impact of teams with atypical or outlying

performances on the assessment of other teams’ efficiency can be quite large.

To account for these issues we use a robustified DEA-analysis that builds on Cazals et al.

(2002).9 The key idea of this approach is to no longer compute the efficiency estimates in one

efficiency evaluation round in which the estimation of Own

kθ builds on a comparison with all cycling

teams of the same type and, similarly, the computation of Overall

kθ is done relative to the frontier

spanned by all cycling teams in the sample. Instead, the efficiency estimates are made in a Monte

Carlo setting in which B runs ( 1, ,b B= … ) are performed (with B a large number). In each of these

Page 19: Mata Architecture for Future Internet

18

runs, a subset of the cycling teams is considered in the efficiency evaluation of the cycling team k .

More precisely, in each run, a subset of m cycling teams of the same type (that is, m teams drawn in

an i.i.d. manner from the group of teams of the same type) are considered in the estimation of Own

kθ ,

and a subset of m teams randomly drawn from the full sample of cycling teams are considered in the

estimation of Overall

kθ . The robust versions of Own

kθ and Overall

kθ , denoted by m,Own

kθ and m,Overall

kθ are

computed as the average values of Own

kθ and Overall

kθ defined over the B efficiency assessment runs.

The use of this so-called robust order-m DEA allows, firstly, to neutralize the size bias as due to

different group sizes: m is set equal for all groups of cycling teams in the computations of the

different efficiency scores. Secondly, due to the Monte Carlo simulation with i.i.d. draws of

subsamples, the impact of cycling teams with outlying or atypical performances is mitigated. Before

concluding this section, we also note that because of the sub-sampling approach, it is not necessarily

the case that an evaluated team also appears in the list of m possible comparison partners, which

implies that efficiency scores can now in fact be larger than one, as can be their averages over B runs.

Such robust DEA-estimated efficiency scores larger than one indicate that the cycling team under

evaluation is extremely efficient compared to an average of m randomly drawn comparison partners.

Efficiency scores equal to one reveal that the evaluated cycling team performs efficiently compared to

other teams in the comparison set, and efficiency scores lower than one indicate that the cycling is

inefficient relative to its counterparts.

5. Results

In this section we present the outcomes of the proposed model. Remark that we discuss the

robust (order- m ) versions of the DEA-based efficiency scores. In all calculations, we used B=500

runs and set m=15 for all types. Team performance evaluations were done for five different

input/output combinations (recall Table 3), so as to check for consistency in the evaluation outcomes

(that is, the three different efficiency estimates for the cycling teams). As noted, model outcomes

Page 20: Mata Architecture for Future Internet

19

should ideally be insensitive to the input-output constellations used in the DEA-model. Fortunately,

both the descriptive statistics of the “Own-team-type-within-all-team-types”-efficiency m,Type

kθ for the

Tour de France cycling teams (Table 4) as well as the correlations between the three different types of

efficiency estimates (Table 5) for the variant models, indicate that all five models yield largely similar

results. In particular, Table 4 conveys the general “type effect”-result that ranking teams on average

outperform the other two types, thus analytically corroborating the summary indications contained in

Table 2.

< Table 4 about here >

< Table 5 about here >

In fact, given the high consistency in the findings, it suffices to focus on the outcomes of just one

of the models. In the remainder, we therefore look at the results as yielded by model 5. The

descriptive statistics of all three efficiency scores as yielded by this model (i.e., the overall or “own-

team-within-all-team-types”-efficiency and its two components: the “own-team-within-own-team-

type”-efficiency and the “own-team-type-within-all-team-types”- efficiency) are listed in Table 6.

< Table 6 about here >

Some interesting general findings can be discerned from this table. We see that the average

cycling team realizes an efficiency score of 85 18. % when being evaluated relative to the teams of the

own type and an efficiency score of 87 66. % when being assessed relative to all cycling teams

(whatever the type). Both the large standard deviation and the large difference between the minimum

and maximum estimated efficiency scores show, however, that there are considerable differences

between the cycling teams. Note also that Table 6 recaptures some model-specific information

relating to the “own-team-type-within-all-team-types”-efficiency scores m,Type

kθ that was also contained

in Table 4. Ranking teams generally outperform the other two types, with average efficiency estimates

Page 21: Mata Architecture for Future Internet

20

of 120 14. % , 80 06. % and 85 44. % for ranking teams, sprint teams, and mixed teams, respectively.

Recalling expression (2), one should thus expect that the “own-team-within-all-team-types”-

efficiency m,Overall

kθ of ranking teams is generally higher than their “own-team-within-own-team-type”

efficiency m,Own

kθ , while the reverse holds for the majority of the teams of the other two types. Table 6

confirms this expectation.

Furthermore, a comparison of the average “own-team-within-all-team-types”-efficiency

estimates for the three types of cycling teams shows that ranking teams generally realize a higher

overall efficiency score compared to sprint teams and mixed teams (i.e., 107 75. % compared to

67 42. % and 75 13. % ). Again though, there are significant differences between individual teams. For

instance, we see that the lowest “own-team-within-all-team-types”-efficiency score is realized by a

mixed team (i.e., minimum efficiency score equal to 12 98. % ). However, also the worst performing

ranking team and sprint team have values near to this minimum (with minimum scores of 16 96. %

and 17 54. % ).

< Figure 3 about here >

The above findings are further detailed in Figure 3, which displays the scatter plots of the “own-

team-within-own-team-type”-efficiency scores and the “own-team-within-all-team-types”-efficiency

scores. In point of fact, the scatter plot for ranking teams reveals that almost all teams are positioned

below the 45°-line. This shows that a large majority of ranking teams have overall efficiency scores

that are higher than their “within-own-team-type”-efficiency scores. Stated differently, all ranking

teams in our sample get higher efficiency scores when they are being evaluated relative to all cycling

teams (irrespective of the team type) than when being evaluated relative to other ranking teams only.

This result is also intuitive. As noted above, when ranking teams usually outperform sprint and mixed

teams, a within-type evaluation of ranking teams entails a comparison with generally strong teams

only. On the other hand, an overall comparison implies the addition of teams that are generally

performing less well, which makes it easier for ranking teams to realize higher efficiencies. The

Page 22: Mata Architecture for Future Internet

21

opposite observation holds for the majority of sprint teams and mixed teams: they realize overall

efficiency scores that are lower than the “own-team-within-own-team-type”-efficiency scores.

The differences between the average “own-team-within-own-team-type”-efficiency scores of the

three types of cycling teams are only slightly different (respectively 86 14. % , 85 61. % , and 83 63. % ).

Ranking teams may do extremely well in terms of realized efficiency (cf. the maximum value of

395 68. % , realized by Astana in the Tour de France of 2010), but may also return home with very

disappointing results when compared with their direct peers. The worst performing team, realizing an

efficiency score m,Own

kθ of only 15 58. % , was indeed also a ranking team (viz., Team Radio Shack with

no stage wins, low prize money and low CQ points in the Tour de France of 2011). In fact, when the

assessment is relative to own-type peers only, all teams face similar distributions of efficiency scores.

We conclude with a brief discussion of the results of some specific teams.10

Table 7 presents the

efficiency scores for three cycling teams for the period 2007-2011, viz. Quick-Step, Team Saxo Bank,

and La Française des Jeux. Apart from the 2007 edition, the Quick-Step team performed rather poor in

the last 5 editions of the Tour de France (although one could argue that the team performed

moderately well in the 2010 Tour de France). The high efficiency scores of Quick-Step in 2007 can be

explained by the good performances of the team in that year. Quick-Step won the green jersey and 5

stage wins, and, thus, a relatively high amount of prize money. It should be no surprise that the impact

of these high outputs on the three efficiency scores is strongly positive. In fact, within the own group

of sprint teams, Quick-Step is even evaluated as highly efficient (i.e., 162 17m,Own

k . %θ = ). A similar

reasoning also helps in explaining the poor efficiency scores for Quick-Step in the Tour de France

editions of 2008, 2009, and 2010. With similar input levels, Quick-Step only realized low output

levels in those editions.

< Table 7 about here >

A detailed analysis of the efficiency scores of Team Saxo Bank shows that this team did very

well in the 2008, 2009, and 2010 editions. Particularly the 2008 Tour de France was a good edition for

Page 23: Mata Architecture for Future Internet

22

Team Saxo Bank, with high efficiency scores of, respectively, 272 26m,Overall

k . %θ = ,

194 26m,Own

k . %θ = , and 140 15m,Type

k . %θ = . The m,Own

kθ equal to 194 26. % illustrates that Team Saxo

Bank, a ranking team, is evaluated as highly efficient when being evaluated relative to the other

ranking teams. With only one stage win and no other significant performances, the 2011 Tour de

France was an exceptionally bad edition for Team Saxo Bank. Note that these low overall and peer

efficiency scores are obtained despite a significant reduction in some of the inputs.

The performances of the La Française des Jeux team during the last five editions at first sight

seem rather middle-of-the-road, with only one stage win in the editions of 2007 and 2010, moderate

numbers of CQ points won, and moderate amounts of prize money. Nevertheless, both in the 2009

and 2011 editions, La Française des Jeux gets relatively high efficiency scores within the group of

mixed cycling teams (i.e., m,Own

kθ equal to 85.70% and 113.28%). The explanation is found both in the

input and output dimension. Though the outputs of La Française des Jeux seem rather poor, these

performances are rather well when assessed in relation to those of other mixed teams only. On the

input side, a comparison with the inputs used by other mixed cycling teams learns that the levels of

inputs used by La Française des Jeux are rather limited. For instance, with a budget of around 6.5 to 7

million euro in most editions, the budget of the team is among the lowest. Combining both the

relatively high output levels and low input levels explains the relatively high efficiency scores among

the mixed-type cycling teams.

6. Conclusions

The conceptual starting point of this paper is that professional road cycling has its stars and even

its superstars, but that modern cycling requires these leading individuals to operate in strong teams in

order to be successful. Ideally, these teams combine different individual competences in such a way

that the chances of winning are maximized. This insight calls for the assessment of team performance,

just as in many other sports. Furthermore, in a major event like the Tour de France, there are many

Page 24: Mata Architecture for Future Internet

23

prizes to be won, which explains why teams, usually those with smaller budgets, may be organized

such as to focus on other aims than on winning the overall individual classification.

We have addressed this issue by using an extensive dataset and a robust evaluation framework.

The latter singles out different efficiency measures, which reveal information about a team’s

performance both relative to teams with similar aspirations and relative to all other participating

teams. Evidently, each of these performance measures yields interesting information from a

managerial point of view. For instance, teams with modest budgets may be expected to play modest

roles, but even then there may be mutual differences in Tour success rates within this type of teams.

In general, neutralizing the type effect effectively instigates a level playing field as far as team

efficiency is concerned. However, it is also clear from our analysis that the (big) teams, which indeed

aim for the big prizes, generally outperform the other team types not only in terms of effectiveness but

also in terms of efficiency. At least in the Tour de France, they seem to deliver, on average, the most

value for money.

While our analysis is robust in the double sense of deriving from models with different

input/output constellations and of being based on (order-m DEA) re-sampling methods, it is also clear

that we have left open some issues. First, lack of public data prevented us from including one output

that is extremely important to sponsors, viz. media coverage of teams. Second, we have focused here

on team types as one important determinant of efficiency in the Tour de France, leaving other

explanatory factors (e.g. a team’s physical condition, comprising indicators such as height, weight and

resulting BMI of selected riders as considered by Prinz, 2005) out of the picture. Adding these

explanatory factors in the (robust) efficiency analysis can be done by using so-called conditional

(order-m) efficiency models (after Daraio and Simar, 2007). We consider this as an interesting avenue

for further research. Third, when it comes to variant models, and as explained in our methodology

section, we have chosen to refrain from a ‘fixed-output sum’-model à la Yang et al. (2011) on the

grounds that even the best teams in our sample are still considerably removed from the (known)

theoretical output maxima. However, one potentially interesting side-product of such a ‘fixed-output

sum’-variant is its endogenous (ex post) identification of a team’s optimal opponents (i.e., those teams

Page 25: Mata Architecture for Future Internet

24

for which the output has to decrease when the evaluated team’s output increases). The “own-type

efficiency”-approach of the current paper evidently implies an implicit exogenous (ex ante) selection

of such potential opponents. A comparative analysis of both models therefore also may yield some

interesting insights. Finally, we stress that our analysis relates to the Tour de France only, although it

may be applicable to similar important cycling events such as the Italian Giro or the Spanish Vuelta.

Team managers and team sponsors may find this a prime event, but a bad Tour result may still be

offset by success in the one-day classics, the World Championship road race, etc. Put otherwise, a

team’s success or its efficiency could clearly also be measured over an entire season.

References

Carmichael, F., Thomas, D., & Ward, R. (2001). "Production and efficiency in Association Football",

Journal of Sports Economics, 2, 228-243.

Cazals, C., Florens, J.P., & L. Simar (2002). “Nonparametric Frontier Estimation: A Robust

Approach”, Journal of Econometrics, 106, 1-25.

Charnes, A. Cooper, W.W., & Rhodes, E. (1978). “Measuring the efficiency of decision making

units”, European Journal of Operational Research, 2, 429-444.

Cherchye, L., & Vermeulen, F. (2006). “Robust rankings of multidimensional performances: An

application to Tour de France racing cyclists”, Journal of Sports Economics, 7, 359-373.

Collier, T., Johnson, A.L., & Ruggiero, J. (2011). "Measuring technical efficiency in sports, Journal

of Sports Economics, 12, 579-598.

Cooper, W.W., Ruiz, J.L., & Sirvent I. (2009). "Selecting non-zero weights to evaluate effectiveness

of basketball players with DEA", European Journal of Operational Research, 195, 563-574.

Cooper, W.W., Seiford, L.M., & Zhu, J. (2004). Handbook on Data Envelopment Analysis,

Dordrecht: Kluwer Academic Publishers.

Daraio, C., & Simar, L. (2007). “Advanced robust and nonparametric methods in efficiency analysis:

Methodology and applications”, Series: Studies in Productivity and Efficiency, Berlin: Springer.

Page 26: Mata Architecture for Future Internet

25

Dawson, P., Dobson, S., & Gerrard, B. (2000). “Stochastic frontiers and the temporal structure of

managerial efficiency in English soccer”, Journal of Sports Economics, 1, 341-362.

De Witte, K., Thanassoulis, E., Simpson, G., Battisti, G., & Charlesworth-May, A. (2010). “Assessing

pupil and school performance by non-parametric and parametric techniques”, Journal of the

Operational Research Society, 61, 1224-1237.

Einolf, K.W. (2004). “Is winning everything?: A data envelopment analysis of Major League Baseball

and the National Football League”, Journal of Sports Economics, 5, 127-151.

Espitia-Escuer, M., & Garcia-Cebrian, L.I. (2004). “Measuring the efficiency of Spanish first-division

soccer teams”, Journal of Sports Economics, 5, 329-346.

Farrell, M.J. (1957). The measurement of productive efficiency, Journal of the Royal Statistical

Society, Series A, CXX, Part 3, 253-290.

Haas, D.J. (2003). "Technical efficiency in the Major League Soccer", Journal of Sports Economics,

4, 203-215.

Hadley, L., Poitras, M., Ruggiero, J., & Knowles, S. (2000). "Performance evaluation of National

Football League teams", Managerial and Decision Economics, 21, 63-70.

Lewis, H.F., Sexton, T.R., & Lock, K.A. (2007). “Player salaries, organizational efficiency, and

competitiveness in Major League Baseball”, Journal of Sports Economics, 8, 266-294.

Pestana Barros, C., & Leach, S. (2006). “Analyzing the performance of the English F.A. Premier

League with an economic frontier model”, Journal of Sports Economics, 7, 391-407.

Prinz, E. (2005). “Every second counts: the influence of money and body weight in professional road

racing - empirical evidence from the Tour de France", Working paper, University of

Witten/Herdecke.

Rimler, M.S., Song, S., & Yi, D.T. (2010). “Estimating production efficiency in men’s NCAA college

basketball: A Bayesian approach”, Journal of Sports Economics, 11, 287-315.

Sterken, E. (2005). “Performance and competition in the Tour de France", Working paper,

Department of Economics, University of Groningen.

Page 27: Mata Architecture for Future Internet

26

Thanassoulis, E., & Portela, M.C.A. (2002). “School outcomes: Sharing the responsibility between

pupil and school”, Education Economics, 10, 183-207.

Torgler, B. (2007). “La Grande Boucle: Determinants of success at the Tour de France", Journal of

Sports Economics, 8, 317-331.

Yang, F., Wu, D.D., Liang, L., & O’Neill, L. (2011). “Competition strategy and efficiency evaluation

for decision making units with fixed-sum outputs”, European Journal of Operational Research,

212, 560-569.

Zhu, J. (2003). Quantitative Models for Performance Evaluation and Benchmarking, International

Series in Operations Research and Management Science, Dordrecht: Kluwer Academic

Publishers.

Page 28: Mata Architecture for Future Internet

27

Figure 1: Data Envelopment Analysis (DEA)

0

C

R

R’

D

S

ABCD: All cycling teams

y1/x

y2/x

B

A

S’

Page 29: Mata Architecture for Future Internet

28

Figure 2: Team efficiency vs. team-type efficiency

0

E

F

C

R

R’

D

S

H

R’’

AGH: Sprint Teams

EFBCD: Ranking teams

ABCD: All cycling teams

y1/x

y2/x

G

S’’B

A

S’

Page 30: Mata Architecture for Future Internet

29

Figure 3: Tour de France cycling teams 2007-2011: Scatter plots

Page 31: Mata Architecture for Future Internet

30

Table 1: Tour de France cycling teams 2007-2011: Team types

(R = ranking, S = sprint, M = mixed)

Team Name

2007 Type

2008 Type

2009 Type

2010 Type

2011 Type

#

Participations

#

Ranking

#

Sprint

#

Mixed

AG2R - La Mondiale R R M M M 5 2 3 Bbox Bouygues Télécom / Europcar

M

M

M

M

S

5

1

4

Caisse d'Epargne R R R S S 5 3 2

Cofidis M M S M S 5 2 3

Euskaltel-Euskadi R R R R R 5 5

La Française des Jeux M S M M M 5 1 4

Lampre M M M M S 5 1 4 Liquigas M M S R R 5 2 1 2

Omega Pharma - Lotto R R R M R 5 4 1

Quick-Step S S S M S 5 4 1 Rabobank R R R R R 5 5 Team Highroad / HTC Columbia

M

R

R

M

S

5

2

1

2

Team Saxo Bank R R R R R 5 5

Astana R R R R 4 4

Team Garmin M R R R 4 3 1 Team Milram S S S S 4 4

Agritubel M R M 3 1 2

Katusha M S M 3 1 2 Saunier-Duval / Footon-Servetto

S

M

M

3

1

2

BMC Racing Team R R 2 2

Cervélo Test Team R R 2 2

Crédit Agricole S S 2 2

Team Barloworld M S 2 1 1 Team Gerolsteiner S S 2 2

Team RadioShack R R 2 2

Team Sky R R 2 2 Discovery Channel R 1 1

Saur-Sojasun M 1 1

Skil-Shimano M 1 1

Team Leopard-Trek R 1 1

Vacansoleil M 1 1

Total

21

20

20

22

22

105

46

24

35

Page 32: Mata Architecture for Future Internet

31

Table 2: Tour de France cycling teams 2007-2011: Descriptive statistics

Inputs Outputs

Budget CQ points

in Tour

team

% CQ

points in

Tour team

Number of

Tour starts

Tour team

consistency

Prize

money

CQ points

collected

Prizes

All cycling teams

Mean 8.20 3,715.70 54.35 33.28 4.45 96,797 620.61 1.43

Stdev. 2.83 1,394.28 10.73 10.03 1.88 134,413 397.57 1.88

Min. 3.50 879 27.19 9 0 9,840 70 0

Max 15.00 7,834 85.45 61 8 723,640 2,059 8

Ranking teams

Mean 8.79 4,478.63 56.82 39.74 4.89 153,905 822.50 2.07

Stdev. 2.65 1,462.16 10.93 8.98 1.82 184,244 475.42 2.16

Min. 3.70 1,682 30.16 24 1 10,540 163 0

Max 14.00 7,834 85.45 61 7 723,640 2,059 8

Sprint teams

Mean 7.83 3,472.96 53.95 30.21 4.88 59,641 505.71 0.92

Stdev. 2.58 936.67 10.84 6.54 1.26 48,845 239.02 1.77

Min. 3.50 1,715 27.33 20 3 17,760 184 0

Max 15.00 5,572 72.03 45 8 192,370 1,048 7

Mixed teams

Mean 7.68 2,879.46 51.37 26.89 3.57 47,218 434.06 0.94

Stdev. 3.14 987.11 9.83 8.19 2.05 28,350 213.07 1.21

Min. 3.50 879 27.19 9 0 9,840 70 0

Max 15.00 5,372 72.14 39 8 112,550 855 5

Page 33: Mata Architecture for Future Internet

32

Table 3: Alternative DEA-models: selections of inputs

Inputs

Budget CQ in Tour team % CQ in Tour

team

Tour starts Team consistency

Model 1 X X

Model 2 X X

Model 3 X X X

Model 4 X X X

Model 5 X X X X

Page 34: Mata Architecture for Future Internet

33

Table 4: “Own-team-type-within-all-team-types”-efficiency for the Tour de France cycling teams: descriptive statistics (5 models)

Model 1 Model 2 Model 3 Model 4 Model 5

All teams

Average 97.93% 94.75% 98.56% 99.06% 99.41%

Standard deviation 21.29% 36.10% 20.32% 23.79% 22.40%

Minimum 59.66% 37.71% 56.23% 31.91% 38.20%

Maximum 157.23% 173.07% 151.86% 170.47% 156.92%

Ranking teams

Average 119.09% 133.68% 118.62% 120.43% 120.14%

Standard deviation 12.93% 13.66% 12.61% 14.92% 13.10%

Minimum 102.44% 116.40% 104.25% 98.96% 98.17%

Maximum 157.23% 173.07% 151.86% 170.47% 156.92%

Sprint teams

Average 82.29% 70.13% 83.92% 76.85% 80.06%

Standard deviation 5.92% 5.36% 5.53% 9.18% 7.22%

Minimum 71.00% 51.55% 72.23% 50.42% 64.58%

Maximum 95.91% 76.87% 99.17% 90.80% 91.10%

Mixed teams

Average 80.83% 60.47% 82.24% 86.20% 85.44%

Standard deviation 7.90% 5.61% 7.89% 15.58% 15.01%

Minimum 59.66% 37.71% 56.23% 31.91% 38.20%

Maximum 94.24% 72.26% 97.72% 121.65% 117.56%

Page 35: Mata Architecture for Future Internet

34

Table 5: Correlations between the three types of efficiency (5 variant models)

“Own-team-within-own-team-type”-efficiency

Model 1 Model 2 Model 3 Model 4 Model 5

Model 1 1

Model 2 0.91 1

Model 3 0.98 0.93 1

Model 4 0.84 0.78 0.87 1

Model 5 0.83 0.78 0.86 0.99 1

“Own-team-within-all-team-types”-efficiency

Model 1 Model 2 Model 3 Model 4 Model 5

Model 1 1

Model 2 0.95 1

Model 3 0.98 0.95 1

Model 4 0.84 0.81 0.87 1

Model 5 0.83 0.80 0.87 0.99 1

“Own-team-type-within-all-team-types”-efficiency

Model 1 Model 2 Model 3 Model 4 Model 5

Model 1 1

Model 2 0.94 1

Model 3 0.96 0.94 1

Model 4 0.86 0.84 0.88 1

Model 5 0.86 0.87 0.91 0.97 1

Page 36: Mata Architecture for Future Internet

35

Table 6: Robust DEA-based efficiency estimates for the Tour de France cycling teams: descriptive statistics (Model 5)

Average Stdev. Min. Max. 25 perc. 50 perc. 75 perc.

All teams

“Own-team-within-own-team-type”-efficiency 85.18% 57.89% 15.58% 395.68% 49.96% 71.77% 100.44%

“Own-team-within-all-team-types”-efficiency 87.66% 72.78% 12.98% 522.96% 45.51% 67.57% 100.40%

“Own-team-type-within-all-team-types”-efficiency 99.41% 22.40% 38.20% 156.63% 81.63% 95.25% 115.95%

Ranking teams

“Own-team-within-own-team-type”-efficiency 86.14% 65.79% 15.58% 395.68% 43.50% 70.68% 103.89%

“Own-team-within-all-team-types”-efficiency 107.75% 92.00% 16.96% 522.34% 51.02% 82.66% 123.60%

“Own-team-type-within-all-team-types”-efficiency 120.14% 13.10% 98.17% 156.92% 111.23% 116.25% 122.17%

Sprint teams

“Own-team-within-own-team-type”-efficiency 85.61% 53.75% 21.94% 243.28% 51.94% 69.60% 95.97%

“Own-team-within-all-team-types”-efficiency 67.42% 40.12% 17.54% 174.06% 44.96% 54.86% 79.07%

“Own-team-type-within-all-team-types”-efficiency 80.06% 7.22% 64.58% 91.10% 75.69% 80.91% 86.24%

Mixed teams

“Own-team-within-own-team-type”-efficiency 83.63% 50.63% 22.07% 277.44% 48.56% 74.87% 93.74%

“Own-team-within-all-team-types”-efficiency 75.13% 53.99% 12.98% 264.26% 40.35% 66.79% 88.00%

“Own-team-type-within-all-team-types”-efficiency 85.44% 15.01% 38.20% 117.56% 81.35% 83.24% 90.61%

Page 37: Mata Architecture for Future Internet

36

Table 7: Detailed analysis Quick-Step, Team Saxo Bank and La Française des Jeux (period 2007-2011)

Budget CQ in Tour team Tour starts Team consistency Prize Money CQ points Prizes m,Overall

kθ m,Own

kθ m,Type

Quick-Step

2007 9 5,125 33 6 101,920 861 5 116.82% 162.17% 72.03%

2008 8.5 3,706 23 5 31,470 351 1 40.57% 52.22% 77.69%

2009 9.5 5,572 45 6 17,760 252 0 17.54% 21.94% 79.95%

2010 9.5 3,285 31 4 78,058 514 2 67.93% 82.99% 81.85%

2011 9.5 3,076 45 4 19,940 297 0 31.33% 40.00% 78.34%

Team Saxo Bank

2007 7 7,834 50 6 136,580 1,034 3 89.92% 73.84% 121.78%

2008 7 7,639 53 6 621,210 1,910 8 272.26% 194.26% 140.15%

2009 7 7,100 51 7 362,850 1,584 6 180.03% 130.24% 138.23%

2010 7 7,451 53 7 292,392 1,269 6 155.77% 120.20% 129.59%

2011 7.5 3,708 40 4 72,290 627 1 61.87% 54.82% 112.85%

La Française des Jeux

2007 7 2,622 20 4 43,840 361 1 53.36% 65.17% 81.88%

2008 6.5 2,437 24 6 45,780 452 0 62.54% 70.76% 88.38%

2009 6.5 2,853 30 3 35,660 596 0 72.72% 85.70% 84.85%

2010 6.5 3,330 35 5 25,922 385 1 42.72% 56.29% 75.90%

2011 8.5 1,940 31 3 90,660 400 0 78.89% 113.28% 69.64%

Page 38: Mata Architecture for Future Internet

37

1 Peloton is the word used for describing a group of cyclists packed together. Because of the slipstream-effect

and being sheltered from the wind, riding in a peloton is a lot easier than riding alone resulting in a much

higher pace.

2 Some cycling teams did change names during the sample period but continued under the same management

and with essentially the same riders. In this study, they are considered to be the same teams. For such teams we

use a single name that best describes the team for the entire period. Team CSC (2007), for instance, became

successively Team CSC - Saxo Bank (2008), Team Saxo Bank (2009-2010) and finally Team Saxo Bank -

Sungard (2011), but we will simply use the name Team Saxo Bank for all years.

3 In its yearly Tour guide, the French newspaper l'Equipe characterizes all individual Tour participants as

either a sprinter, a climber, an overall contender, a time trial specialist or a finisher. Apart from merely

reflecting a personal view of the journalist, it proved to be very difficult to aggregate this information into

some sort of team identification measure. Alternatively, we looked at press interviews by team managers on

the eve of the Tour de France to determine a team's main goals. However, although very accurate in some

instances, many teams do not clearly communicate their specific goals or only provide this information to local

media.

4 The CQ-ranking is a world ranking of professional road cyclists, based on their performances during the last

12 months. It can be seen as the non-official successor of the UCI-ranking which disappeared when the

ProTour was introduced in 2005 (www.cqranking.com).

5 Remark that the aggregation of individual output into a team output is to some extent in line with what

happens in reality. In cycling it is customary that prize money is split over all teammates and team staff. Team

leaders therefore do not enjoy any short-term financial benefits from a victory. Glory and long-term financial

benefits through a better contract are their real personal rewards.

6 In point of fact, the few restrictions on the λ-vector in model (1) are such that any rescaled version of an

observed input-output combination is considered to be feasible. This is tantamount to assuming constant

returns to scale, which, in turn, allows drawing Figure 1, where both inputs are normalized by dividing with

the sole output.

Page 39: Mata Architecture for Future Internet

38

7 Thanassoulis and Portela (2002) used this adjusted version of the DEA-model to compute and decompose

the pupil under-attainment into inefficiencies that are due to the pupil and the school the pupil attends to. In a

more recent paper, De Witte et al. (2010) applied the approach in a similar setting.

8 Figure 2 could be readily adjusted to also show the situation for the mixed teams. However, in order to

make the figure not too complicated, we just visualize the efficiency evaluations for the ranking and sprint

teams.

9 In what follows, we restrict ourselves to briefly sketching the basic idea of the robust order-m version of the

DEA model. We refer to Cazals et al. (2002) for a more elaborate (technical) explanation of the method.

10 Results for all teams are available upon request.


Recommended