Measurement Error and Related Definitions Error = measured value - true value
Measurement error, education production and data envelopment analysis
Transcript of Measurement error, education production and data envelopment analysis
-
8/2/2019 Measurement error, education production and data envelopment analysis
1/7
Economics of Education Review 25 (2006) 327333
Measurement error, education production and
data envelopment analysis
John Ruggiero
Department of Economics and Finance, University of Dayton, Dayton, OH 45469-2251, USA
Received 30 September 2004; accepted 18 March 2005
Abstract
Data Envelopment Analysis has become a popular tool for evaluating the efficiency of decision making units. The
nonparametric approach has been widely applied to educational production. The approach is, however, deterministic
and leads to biased estimates of performance in the presence of measurement error. Numerous simulation studies
confirm the effect that measurement error has on cross-sectional deterministic models of efficiency. It is also known that
panel data models have the ability to smooth out measurement error, leading to more reliable efficiency estimates. In
this paper, we exploit known properties of educational production to show that aggregation can also have a smoothing
effect on production with measurement error, suggesting that efficiency analyses are more reliable than previously
believed.
r 2005 Published by Elsevier Ltd.
JEL Classification: I21
Keywords: Educational production; Data envelopment analysis
1. Introduction
Data Envelopment Analysis (DEA) is a nonpara-
metric mathematical programming approach to the
measurement of efficiency that was introduced in the
operations research literature by Charnes, Cooper, and
Rhodes (1978) and Banker, Charnes, and Cooper
(1984). Using linear programming, an observed decisionmaking unit (DMU) is evaluated relative to the
production frontier, which consists of combinations of
observed production possibilities using minimal assump-
tions. The primary advantage of the approach is the
ability to handle multiple inputs and multiple outputs,
particularly in the case when input prices are unavail-
able. The linear programming approach has withstood
the test of time and has become an acceptable approach
for efficiency evaluation.
One important application of DEA is to the analysis
of educational production. Many states have undergone
legal challenges because school districts are not provid-
ing educational services efficiently and outcomes are not
adequate. Reform has moved away from traditionalissues like equity to adequacy and efficiency. The
important policy implication is that school districts
need to spend their money more wisely and increase
their outcomes to acceptable levels. One popular
technique that has been used for measuring efficiency
in education is DEA. Research using DEA to measure
performance of educational production in the United
States include Bessent, Bessent, Kennington, and
Reagan (1982), Fa re, Grosskopf, and Weber (1989),
ARTICLE IN PRESS
www.elsevier.com/locate/econedurev
0272-7757/$- see front matterr 2005 Published by Elsevier Ltd.
doi:10.1016/j.econedurev.2005.03.003
Tel.: +1937 2292258; fax: +1937 2292477.
E-mail address: [email protected].
http://www.elsevier.com/locate/econedurevhttp://www.elsevier.com/locate/econedurev -
8/2/2019 Measurement error, education production and data envelopment analysis
2/7
Ray (1991), McCarty and Yaisawarng (1993), Ruggiero
(1996, 2001), and Duncombe, Miner, and Ruggiero
(1997). DEA studies analyzing performance in other
countries include Silva Portela and Thanassoulis (2001),
Farren (2002) and Muniz (2002).
The ability of DEA to measure efficiency is dependent
on the nature of observed production. Numerous studies
have shown that the performance of DEA deteriorates
in the presence of measurement error and other
statistical noise. Simulations in Banker, Gadh, and Gorr
(1993), Ruggiero (1999), and Bifulco and Bretschneider
(2001), for example, clearly show that deterministic
models including DEA are sensitive to measurement
error in cross-sectional analyses. Given an econometri-
cians view of the world, the use of DEA could be
problematic and potentially harmful given the important
policy implications. Gong and Sickles (1992) and
Ruggiero (2004) use simulation to show that the
problem of measurement error is alleviated whenefficiency is estimated using panel data. These models
assume that the production technology is the same
across time and use averaging to smooth production
with respect to measurement error.
In a separate but related literature, Hanushek (1979)
discusses other important issues with empirical educa-
tional production analyses. One of the important issues
involves the choice of aggregation. Conceptually,
education occurs at the student level, but most analyses
rely on aggregated data (primarily due to availability)
to analyze educational production. One potential
advantage, which is exploited in this paper, is the
smoothing of production with respect to measurement
error. Hence, aggregation, like panel data, helps
alleviate the problem of measurement error. In parti-
cular, we assume that educational production with
measurement error occurs at a disaggregated level and
consider efficiency measurement using both disaggregate
and aggregate data.
Simulation analysis is used to show that the use of
DEA on disaggregated data in the presence of measure-
ment error leads to biased efficiency estimation,
confirming the results of previous analyses. We also
apply DEA to aggregated data and show that aggrega-
tion can lead to unbiased efficiency estimates. Theseresults represent an important contribution to the DEA
literature: aggregation effectively controls for statistical
noise. This suggests that DEA can be used effectively to
identify inefficiency of school districts (schools), where
data are aggregated from the school (classroom) level.
We note that such aggregation can be applied to other
areas such as analyzing bank performance where
production occurs at the branch level. For purposes of
this paper, we consider district data as the aggregate of
school data. The rest of the paper is organized as
follows. The next section provides a model of educa-
tional production and efficiency using DEA. The third
section analyzes the performance of DEA using simula-
tion analysis and the last section concludes.
2. Educational production
We describe the educational production process as
follows. Assume that each of Ns schools uses a vector
X x1; :::; xM of M discretionary school inputs to
produce a vector Y y1; :::; yT of T outcomes. For
school j, inputs are given by Xj x1j; :::; xMj and
outputs by Yj y1j; :::; yTj. For purposes of this
paper, we ignore the role of non-discretionary socio-
economic variables to focus on measurement error and
aggregation. This further allows comparisons to the
approach used by Bifulco and Bretshneider (2001).
Ruggiero (1998) extended DEA to the case of multiple
non-discretionary inputs; allowing measurement error
with non-discretionary inputs is a trivial extension ofRuggieros model using the approach of this paper.
Assuming variable returns to scale, technical effi-
ciency can be estimated at the school level using the
Banker, Charnes, and Cooper (1984) input-oriented
model:
Minyol1 ;:::;lNS
s:t.
XNS
i1
liytiXyto 8 t 1; :::; T
XNS
i1
lixmipyoxmo 8 m 1; :::; M
XNS
i1
li 1
li ! 0 8 i 1; :::; NS. 1
In the absence of measurement error and with a
sufficiently large sample size, this model works well in
measuring relative performance with multiple inputs and
outputs. The returns to scale assumption can be
tightened to allow only constant returns by excluding
the convexity constraint; leading to the Charnes,Cooper, and Rhodes (1978) model. The analysis of a
given school seeks a maximum equi-proportional con-
traction of all inputs consistent with frontier production
with all outputs at least as high as the levels of the school
under analysis. Alternatively, we could have focused on
the output oriented models.
Model (1) leads to biased efficiency estimates in the
presence of measurement error. Numerous simulation
analyses confirm that even moderate levels of measure-
ment error can lead to severe bias. For this reason, the
usefulness of DEA to analyze educational production
must be questioned. Consider measurement error in one
ARTICLE IN PRESS
J. Ruggiero / Economics of Education Review 25 (2006) 327333328
-
8/2/2019 Measurement error, education production and data envelopment analysis
3/7
output j, with ~yj yj j, where ~yji yji ji for school
i. We note that this extends Ruggiero (2004) by
considering additive error in the output. The BCC
model for measuring efficiency of a given school is then
given by
Min yol1 ;:::;lNS
s:t.
XNS
i1
liytiXyto 8 t 1; :::; T; taj
XNS
i1
li~yjiX ~yjo
XNS
i1
lixmipyoxmo 8 m 1; :::; M
XNS
i1
li 1
liX0 8 i 1; :::; NS. 2
As discussed in Ruggiero (2004), measurement error
biases efficiency estimation for two reasons. First, the
right-hand side of the contaminated output constraint
distorts the measurement by comparing frontier output
to a level of output not consistent with that school. This
results from measurement error for the school under
analysis. The second effect involves the location of the
frontier output; measurement error in other schools
will lead to a biased location of the true frontier. For a
further discussion as it relates to panel data, seeRuggiero (2004).
We consider the aggregation of the data to the school
district level. We assume that each of the NS schools can
be assigned to one and only one of the ND school
districts. Defining NSd as the number of schools in
district d for d 1; :::; ND, we construct an index that
identifies the school according to the district to which it
belongs. Then, for each district d, we can define a vector
for each input m and output t given by
Xmd xmd;1; :::; xmd;NSd
and Ytd ytd;1; :::;ytd;NSd
; respectively.
We note that ND PD
d1NSd prior to estimating
efficiency, we first average school level data to the
district level for each district d
xmd 1
NSd
XNSd
i1
xmd;i 8 m,
ytd 1
NSd
XNSd
i1
ytd;i 8 taj; and
yjd 1
NSd
XNSd
i1
~yjd;i:
With the additional assumption that the E 0, the
error term is averaged out with a sufficiently large
number of schools in the district. In the simulation
analysis performed in the next section, we vary the
number of schools within a district to be 5, 10 or 20 and
show that averaging data even with few schools can
effectively eliminate measurement error problems. One
concern raised by an anonymous reviewer was the
assumption of aggregating from the school level to the
district level. We note that state policy typically funds at
the district level and compares district performance on
state tests. And, in cases where this is not true,
aggregation could happen from the student or classroom
level to the school level. For space consideration, we
show only the aggregation from the school to the district
level; the results from a lower level to the school level
follow.
3. Simulation analysis
We assume that production can be characterized by
the conversion of discretionary inputs into output. As
mentioned above, non-discretionary inputs can be
effectively incorporated into the DEA model (see
Ruggiero, 1998). In this simulation, we focus exclusively
on the impact of measurement error. In particular, we
assume that three inputs x1, x2 and x3 are used by
schools to produce one output y according to the
following stochastic production function:
y u1vx0:41 x0:42 x0:23 ,
where u and v represent inefficiency and statistical noise,
respectively. Input data were generated uniformly on the
interval (5,10) for each input. Measurement error was
generated log-normal with standard deviation sv where
sv took on values 0.1, 0.2 and 0.3. Notably, sv 0:2
represented high measurement error in the Bifulco
and Bretschneider (2001) and hence, we consider an
even higher measurement error variance. The
inefficiency component ln u was generated half-normal
with standard deviation 0.2. By varying the standard
deviation of the measurement error, we effectively
allow varying ratios of measurement error toinefficiency variances. Finally, we consider various
combinations of schools and school districts; we vary
the number of schools within each district to take on
values of 5, 10 and 20 and we allow 50, 100 and 200
districts.
Initially, each school within a given district had the
same level of inefficiency. In the last simulation, we
allow school inefficiency to vary within a district. Three
measures of performance are considered in this study:
mean absolute deviation (MAD) between true and
measured efficiency, the correlation coefficient and the
rank correlation coefficient. Because the assumed
ARTICLE IN PRESS
J. Ruggiero / Economics of Education Review 25 (2006) 327333 329
-
8/2/2019 Measurement error, education production and data envelopment analysis
4/7
production function is characterized by constant returns
to scale, we use the CCR model that excludes the
convexity constraint. This will allow us to focus
exclusively on measurement error in technical efficiency
estimation.
The performance results using school level data arereported in Table 1. Not surprisingly, these results
confirm the results known in the literature: measurement
error has a devastating effect on DEA performance.
Consider the absolute deviations reported in Table 1. In
general, we note that as the measurement error variance
increases, the MAD gets larger. Further, the standard
deviation increases as well. Even when sv was low (0.1),
the MAD was still unacceptably high in the 0.100.16
range. Holding constant the measurement error var-
iance, we also note the general result that the MAD gets
worse even as the number of schools and the number of
districts increase.
Turning to the correlation and rank correlation
results, we see a similar pattern. DEA performs better
when the measurement error variance is relatively low
while the correlation and rank correlation decrease as
the variance is increased. Even with a low measurement
error variance, DEA does not provide accurate estimatesof efficiency. The correlation and rank correlations are
all below 0.75 with sv 0:1. With sv 0:3, the highest
correlation is about 0.41. No general pattern for
improvement appears as the number of schools or
districts increases. Clearly, the use of DEA on disag-
gregated data in the presence of measurement error is
suspect.
Next, we estimate efficiency using the constant returns
to scale model using data averaged to the district level.
The results are reported in Table 2. The MAD results in
Table 2 reveal that DEA performs extremely well
relative to the performance using school level data.
ARTICLE IN PRESS
Table 1
Performance results using school level simulated data
sv Number of schools Number of districts Absolute deviation Correlation coefficients
Mean Standard deviation Pearson Rank
0.1 5 50 0.111 0.075 0.648 0.622100 0.115 0.074 0.716 0.680
200 0.128 0.071 0.726 0.694
10 50 0.128 0.072 0.683 0.666
100 0.157 0.074 0.545 0.532
200 0.163 0.073 0.709 0.680
20 50 0.134 0.072 0.680 0.665
100 0.140 0.073 0.740 0.723
200 0.123 0.071 0.714 0.682
0.2 5 50 0.260 0.122 0.463 0.454
100 0.287 0.123 0.500 0.473
200 0.314 0.114 0.462 0.450
10 50 0.302 0.116 0.447 0.432
100 0.299 0.119 0.510 0.498
200 0.336 0.118 0.443 0.435
20 50 0.272 0.120 0.423 0.406
100 0.336 0.119 0.481 0.461
200 0.326 0.115 0.481 0.460
0.3 5 50 0.383 0.152 0.318 0.296
100 0.388 0.146 0.389 0.395
200 0.404 0.148 0.332 0.339
10 50 0.381 0.147 0.344 0.325
100 0.419 0.146 0.319 0.319
200 0.607 0.114 0.288 0.301
20 50 0.418 0.131 0.348 0.337
100 0.452 0.125 0.406 0.409
200 0.472 0.132 0.332 0.329
All calculations by author. School efficiency was held constant within each district.
J. Ruggiero / Economics of Education Review 25 (2006) 327333330
-
8/2/2019 Measurement error, education production and data envelopment analysis
5/7
With low measurement error, the MAD is under 0.06
when the variance of the measurement error is low. As
the measurement error variance increases, holding the
number of districts and schools constant, the perfor-
mance declines but at levels far better than reported for
the school level data. All MADs are lower than 0.2 whenusing the aggregated data. We also note that the
performance of DEA improves as the number of schools
increases and remains about the same as the number of
districts increases.
Similar results are observed for the correlation and
rank correlations reported in Table 3. In general, all
correlations show marked improvement using the
aggregate data compared to the school level analysis.
In the case of low measurement error variance, the
correlation and rank correlation coefficients are above
0.85 and usually above 0.90. Similar to the MAD results,
the performance of DEA declines as the measurement
error variance increases. This is offset by improved
performance as the number of schools increases. This is
an important finding; as the number of units at the
disaggregated level increases, the effect of measurement
error on DEA performance at the aggregate level
becomes less pronounced. It is also important toremember that the units chosen for the disaggregated
level were schools. Extending the units to the classroom
or even the student level suggests that DEA at the
district level provides meaningful results even with high
measurement error. Of course, the standard assumption
that the expected value of the error term is zero was
employed.
One questionable assumption that was used in the
simulation was constant efficiency at the school level
within each district. For completeness, we now consider
random efficiency within a given school district. Holding
the number of schools at 20 and the number of districts
ARTICLE IN PRESS
Table 2
Performance results using simulated data aggregated to the district level
sv Number of schools Number of districts Absolute deviation Correlation coefficients
Mean Standard deviation Pearson Rank
0.1 5 50 0.045 0.031 0.869 0.866100 0.035 0.028 0.924 0.897
200 0.054 0.034 0.915 0.856
10 50 0.025 0.019 0.949 0.937
100 0.025 0.018 0.960 0.941
200 0.029 0.023 0.952 0.917
20 50 0.029 0.021 0.974 0.948
100 0.040 0.025 0.934 0.904
200 0.028 0.019 0.974 0.942
0.2 5 50 0.096 0.064 0.744 0.714
100 0.100 0.059 0.822 0.802
200 0.112 0.062 0.776 0.695
10 50 0.054 0.047 0.830 0.775
100 0.059 0.042 0.895 0.874
200 0.061 0.039 0.834 0.840
20 50 0.027 0.023 0.931 0.908
100 0.042 0.030 0.914 0.875
200 0.053 0.032 0.932 0.889
0.3 5 50 0.198 0.096 0.553 0.635
100 0.134 0.082 0.701 0.692
200 0.147 0.087 0.642 0.617
10 50 0.084 0.065 0.755 0.718
100 0.104 0.073 0.753 0.679
200 0.128 0.067 0.712 0.689
20 50 0.088 0.053 0.806 0.723
100 0.081 0.043 0.876 0.852
200 0.083 0.050 0.837 0.762
See comments in Table 1.
J. Ruggiero / Economics of Education Review 25 (2006) 327333 331
-
8/2/2019 Measurement error, education production and data envelopment analysis
6/7
at 200, school efficiency was regenerated as ln us l-
n ud+ln e, where ln ud was generated as above and ln e
generated as N(0,s) with s taking on values of 0.02, 0.04
and 0.06. The value of us was set to one if it was
calculated to be greater than unity. As shown in Table 3,
this generating process led to large differences in school
efficiency within districts.
The results of this simulation are reported in Table 3.
We note that the first column of results repeats theresults from the earlier simulation where there was no
variation in efficiency among schools in the same
district. As s increases, we note that the variation
increases within a district. In the last column, the
average difference between maximum and minimum
efficiency is approximately 0.17. The performance of
DEA is measured by all three criteria: MAD, correlation
and rank correlation. District efficiency was measured
using the aggregated data and compared to the true
value, determined by the average true efficiency. We
note that the results are similar across columns for all
measures. This suggests that the assumption of constant
efficiency within a district does not effect performanceevaluation at the aggregate level. Of course, a trade-off
exists; with the aggregate data, it is not possible to
identify efficiency of the individual units. However, the
approach does work in identifying average efficiency of
the aggregate units.
4. Conclusions
The use of DEA to evaluate performance has been
called into question because of the potential bias that
exists when measurement error exists. Previous simula-
tion studies have shown that performance drops
significantly as the variance of measurement error
increases. In this paper, the problem of measurement
error was highlighted by two effects. First, measurement
error on a particular unit affects the placement of the
unit relative to the true frontier. Also, measurement
error on other units effects the evaluation of a given unit
because these other units might serve as reference unitsfor the unit under analysis. Unlike stochastic frontier
models that use panel data, DEA is deterministic.
However, as shown in this paper, aggregation of data
to higher units (for example, from school level to district
data or from classroom to school level) can smooth
measurement error. And, as a result, performance
evaluation using aggregate data can produce reliable
results, eve when measurement error is substantial.
As recommended by an anonymous referee, an
interesting extension of this paper would be a real-world
application. In particular, a useful application would be
to show how aggregation of individual student or
classroom data could help identify school and districtinefficiency. Data requirements prevented such an
application; however, I expect such data will be available
in the near future.
References
Banker, R. D., Charnes, A., & Cooper, W. W. (1984). Some
models for estimating technical and scale inefficiencies in
data envelopment analysis. Management Science, 30(9),
10781092.
ARTICLE IN PRESS
Table 3
Simulation results assuming varying school efficiency within districtsa (number of districts 200, number of schools 20, sv 0:2)
District descriptive statistics s
0b 0.02 0.04 0.06
Range usc
Mean 0 0.061 0.118 0.168
Standard deviation 0 0.014 0.028 0.041
Minimum 0 0.024 0.044 0.077
Maximum 0 0.095 0.200 0.305
Efficiency results
Correlation coefficient 0.932 0.924 0.926 0.926
Rank correlation coefficient 0.889 0.890 0.887 0.890
Absolute deviation
Mean 0.053 0.053 0.050 0.043
Standard deviation 0.032 0.032 0.031 0.027
aAverage district efficiency was randomly generated as: ln ud$jN0; 0:2j. School efficiency was calculated as ln us ln ud ln ,
where ln $N0; s and s is given in the table. If us was calculated to be greater than unity, it was set equal to one.bThe range is the maximum usminimum us, calculated for each district.cThis column is taken from previous tables and serves only as a reference.
J. Ruggiero / Economics of Education Review 25 (2006) 327333332
-
8/2/2019 Measurement error, education production and data envelopment analysis
7/7
Banker, R., Gadh, V., & Gorr, W. (1993). A monte carlo
comparison of two production frontier estimation methods:
Corrected ordinary least squares and data envelopment
analysis. European Journal of Operational Research, 67(3),
332343.
Bessent, A., Bessent, E. W., Kennington, J., & Reagan, B.
(1982). An application of mathematical programming toassess productivity in the Houston independent school
district. Management Science, 28(12), 13351366.
Bifulco, R., & Bretschneider, S. (2001). Estimating school
efficiency: A comparison of methods using simulated data.
Economics of Education Review, 20(5), 417429.
Charnes, A., Cooper, W. W., & Rhodes, E. (1978). Measuring
the efficiency of decision making units. European Journal of
Operational Research, 2(6), 429444.
Duncombe, W., Miner, J., & Ruggiero, J. (1997). Empirical
evaluation of bureaucratic models of inefficiency. Public
Choice, 93(1), 118.
Fare, R., Grosskopf, S., & Weber, W. (1989). Measuring school
district performance. Public Finance Quarterly, 17(4), 409428.
Farren, D. (2002). The technical efficiency of schools in Chile.Applied Economics, 34(12), 15331542.
Gong, B., & Sickles, R. (1992). Finite sample evidence on the
performance of stochastic frontiers and data envelopment
analysis using panel data. Journal of Econometrics, 51(1),
259284.
Hanushek, E. (1979). Conceptual and empirical issues in the
estimation of educational production functions. Journal of
Human Resources, 14(3), 351388.
McCarty, T., & Yaisawarng, S. (1993). Technical efficiency in
New Jersey school districts. In H. O. Fried, C. A. K. Lovell,
& S. S. Schmidt (Eds.), The measurement of productive
efficiency (pp. 271287). New York: Oxford University
Press.
Mun iz, M. (2002). Separating managerial inefficiency and
external conditions in data envelopment analysis. EuropeanJournal of Operational Research, 143(3), 625643.
Ray, S. (1991). Resource-use efficiency in public schools: A
study of connecticut data. Management Science, 37(12),
16201628.
Ruggiero, J. (1996). Measuring technical efficiency in the public
sector: An analysis of educational production. Review of
Economics and Statistics, 78(3), 499509.
Ruggiero, J. (1998). Non-discretionary inputs in data envelop-
ment analysis. European Journal of Operational Research,
111(3), 461468.
Ruggiero, J. (1999). Efficiency estimation and error decom-
position in the stochastic frontier model: A monte carlo
analysis. European Journal of Operational Research, 115(3),
555563.Ruggiero, J. (2001). Determining the base cost of education: An
analysis of Ohio school Districts. Contemporary Economic
Policy, 19(3), 268279.
Ruggiero, J. (2004). Data envelopment with stochastic data.
Journal of the OR Society, 55(9), 10081012.
Silva Portela, M. C., & Thanassoulis, E. (2001). Decomposing
school and school-type Efficiency. European Journal of
Operational Research, 132(2), 357373.
ARTICLE IN PRESS
J. Ruggiero / Economics of Education Review 25 (2006) 327333 333