References
Abdi, H. (2007). Bonferroni and Sidak corrections for multiple comparisons. In N. J. Salkind (Ed.),
Encyclopedia of measurement and statistics (pp. 103–107). Thousand Oaks, CA: Sage.
Abramowitz, M., & Stegun, I. A. (1965). Handbook of mathematical functions. New York, NY:
Dover.
Abramowitz, M., & Stegun, I. A. (Eds.). (1972). Handbook of mathematical functions withformulas, graphs, and mathematical tables. New York, NY: Dover.
Achenbach, T. M., & Rescorla, L. A. (2001). Manual for the ASEBA school-age forms andprofiles. Burlington, VT: University of Vermont.
ACT. (2001). EXPLORE technical manual. Iowa City, IA: Author.Aitchison, J., & Silvey, S. D. (1958). Maximum likelihood estimation of parameters subject to
restraints. Annals of Mathematical Statistics, 29, 813–828.Akaike, H. (1974). A new look at the statistical model identification. IEEE Transactions on
Automatic Control, 19, 716–723.Allen, N. L., Carlson, J. E., & Zelenak, C. A. (1999). The NAEP 1996 technical report. Washington,
DC: National Center for Education Statistics.
Andersen, E. B. (1973). A goodness of for test for the Rasch model. Psychometrika, 38, 123–140.Angoff, W. H. (1971). Scales, norms and equivalent scores. In R. L. Thorndike (Ed.), Educational
measurement (2nd ed., pp. 508–600). Washington DC: American Council on Education.
Armstrong, R. D., & Shi, M. (2009). Model-free CUSUM methods for person fit. Journal ofEducational Measurement, 46(4), 408–428.
Baker, F. B. (1990). EQUATE: Computer program for linking two metrics in item response theory.Madison: University of Wisconsin, Laboratory of Experimental Design.
Baker, F. B., & Al-Karni, A. (1991). A comparison of two procedures for computing IRT equating
coefficients. Journal of Educational Measurement, 28, 147–162.Balakrishnan, N. (1992). Handbook of the logistic distribution. New York, NY: Marcel Dekker.
Ban, J.-C., & Lee, W.-C. (2007).Defining a score scale in relation to measurement error for mixedformat tests (CASMA Research Report Number 24). Iowa City, IA: Center for Advanced
Studies in Measurement and Assessment.
Bayley, N. (1933). The California first-year mental scale. Berkeley, CA: University of California
Press.
Beguin, A. A. (2000). Robustness of equating high-stakes tests (Doctoral thesis). University of
Twente, Enschede.
Beguin, A. A., & Glas, C. A. W. (2001). MCMC estimation and some fit analysis of multidimen-
sional IRT models. Psychometrika, 66, 541–562.Birnbaum, A. (1968). Some latent trait models. In F. M. Lord & M. R. Novick (Eds.), Statistical
theories of mental test scores (pp. 397–479). Reading, MA: Addison-Wesley.
347
Bishop, Y. M. M., Fienberg, S. E., & Holland, P. W. (1975).Discrete multivariate analysis: theoryand practice. Cambridge, MA: MIT Press.
Bock, R. D., & Aitkin, M. (1981). Marginal maximum likelihood estimation of item parameters:
Application of an EM algorithm. Psychometrika, 46, 443–459.Bock, R. D., Gibbons, R. D., & Muraki, E. (1988). Full-information factor analysis. Applied
Psychological Measurement, 12, 261–280.Bock, R. D., & Lieberman, M. (1970). Fitting a response model for n dichotomously scored items.
Psychometrika, 35, 179–197.Bock, R. D., & Moustaki, I. (2007). Item response theory in a general framework. In C. R. Rao &
S. Sinharay (Eds.), Handbook of statistics, Vol. 26. Psychometrics (pp. 469–513). New York,
NY: Elsevier.
Bock, R. D., Thissen, D., & Zimowski, M. F. (1997). IRT estimation of domain scores. Journal ofEducational Measurement, 34(3), 197–211.
Box, G. E. P., & Draper, N. R. (1987). Empirical model-building and response surfaces. NewYork, NY: Wiley.
Box, G. E. P., & Jenkins, G. M. (1970). Time-series analysis, forecasting, and control. Oakland,CA: Holden-Day.
Box, G. E. P., Jenkins, G. M., & Reinsel, G. C. (1994). Time-series analysis, forecasting andcontrol (3rd ed.). Englewood Cliffs, NJ: Prentice Hall.
Bozdogan, H. (1987). Model selection and Akaike’s information criterion (AIC): the general
theory and its analytical extensions. Psychometrika, 52(3), 345–370.Bradlow, E. T., Wainer, H., & Wang, X. (1999). A Bayesian random effects model for testlets.
Psychometrika, 64, 153–168.Bradway, K. P. (1944). IQ constancy on the Revised Stanford-Binet from the preschool to the
junior high school level. Journal of Genetic Psychology, 65, 197–217.Bradway, K. P. (1945a). An experimental study of factors associated with Stanford-Binet IQ
changes from preschool to the junior high school. Journal of Genetic Psychology, 66,107–128.
Bradway, K. P. (1945b). Predictive value of the Stanford-Binet preschool items. Journal ofEducational Psychology, 36, 1–16.
Braun, H. I., & Holland, P. W. (1982). Observed score test equating: A mathematical analysis of
some ETS equating procedures. In P. W. Holland & D. B. Rubin (Eds.). Test equating(pp. 9–49). New York, NY: Academic Press.
Brennan, R. L. (1992). The context of context effects. Applied Measurement in Education, 5,225–264.
Brennan, R. L. (2007). Tests in transition: Synthesis and discussion. In N. J. Dorans, M. Pommerich,
& P.W. Holland, (Eds.), Linking and aligning scores and scales (pp. 161–175). New York, NY:
Springer-Verlag.
Brennan, R. L., & Lee, W. (2006). Correcting for bias in single-administration decision consis-tency indexes (CASMA Research Report No. 18). Iowa City: University of Iowa, Center for
Advanced Studies in Measurement and Assessment.
Brockwell, P. J., & Davis, R. A. (2002). Introduction to time-series and forecasting (2nd ed.). NewYork, NY: Springer-Verlag.
Campbell, N. R. (1928). An account of the principles of measurement and calculation. London,England: Longmans, Green & Co.
Carmer, S. G., & Swanson, M. R. (1973). An evaluation of ten pairwise multiple comparison
procedures by Monte Carlo methods. Journal of the American Statistical Association, 68,66–74.
Chatfield, C. (2003). The analysis of time-series: An introduction (6th ed.). London, England:
Chapman and Hall.
Chen, H., & Holland, P. W. (2008). Construction of chained true score equipercentile equatingsunder the KE framework and their relationship to Levine true score equating. (ETS Research
Rept. RR-09-24). Princeton, NJ: ETS.
348 References
Chen, H., & Holland, P. W. (2009). The construction of Levine observed score equipercentileequating under kernel equating framework. Paper presented at the annual meeting of the
National Council of Measurement in Education, San Diego.
Chen, H., & Holland, P. W. (2010). Fit log-linear model with a linear transformation on thediscrete distribution. Manuscript in preparation.
Chen, H., Yan, D., Han, N., & von Davier, A. (2006). LOGLIN/KE user guide: Version 2.1.Princeton, NJ: ETS.
Cochran, D., & Orcutt, G. H. (1987). Applications of least square regression to relationships
containing autocorrelated errors. Journal of American Statistical Association, 44, 32–61.Conceptual framework. (n.d.). Retrieved from Wikipedia: http://en.wikipedia.org/wiki/
Cook, L. L. (2007). Practical problems in equating test scores: A practitioner’s perspective. In N. J.
Dorans, M. Pommerich, & P. W. Holland (Eds.), Linking and aligning scores and scales (pp.73–88). New York, NY: Springer-Verlag.
Cudeck, R. (2000). An estimate of the covariance between two variables which are not jointly
observed. Psychometrika, 65, 539–546.Dalal, S., & Hall, W. (1983). Approximating priors by mixtures of natural conjugate priors.
Journal fo the Royal Statistical Society, Series B, 45, 278–286.Diaconis, P., & Ylvisaker, D. (1985). Conjugate priors for exponential families. Annals of
Statistics, 7, 269–281.Divgi, D. R. (1987). A stable curvilinear alternative to linear equating (Report CRC 571).
Alexandria, VA: Center for Naval Analyses.
Dorans, N. J. (2008, December).Holland’s advice for the fourth generation of test theory: Blood testscan be contests. Invited paper presented at Holland’s Festschrift: A Conference in Honor of Paul
W. Holland, Princeton, NJ.
Dorans, N. J. (Ed.). (1990). Selecting samples for equating: To match or not to match [Special
issue]. Applied Measurement in Education, 3, 1–113.Dorans, N. J. (2002). Recentering and realigning the SAT score distributions: How and why.
Journal of Educational Measurement, 39(1), 59–84.Dorans, N. J. (2007). Linking scores from multiple health outcome instruments. Quality of Life
Research, 16, 85–94. 116Dorans, N. J., & Feigenbaum, M. D. (1994). Equating issues engendered by changes to the SAT
and PSAT/NMSQT®. In I. M. Lawrence, N.J. Dorans, M. D. Feigenbaum, N. J. Feryok, &
N. K. Wright (Eds.), Technical issues related to the introduction of the new SAT andPSAT/NMSQT (ETS Research Memorandum RM-94-10). Princeton, NJ: ETS.
Dorans, N. J., & Holland, P. W. (2000). Population invariance and the equatability of tests: Basic
theory and the linear case. Journal of Educational Measurement, 37, 281–306.Dorans, N. J., & Liu, J. (2009). Score equity assessment: Development of a prototype
analysis using SAT mathematics test data across several administrations (ETS Research
Rept. RR-09-08). Princeton, NJ: ETS.
Dorans, N. J., Liu, J., & Hammond, S. (2008). Anchor test type and population invariance: An
exploration across subpopulations and test administrations. Applied Psychological Measure-ment, 32, 81–97.
Dorans, N. J., Pommerich, M., & Holland, P.W. (Eds.). (2007). Linking and aligning scores andscales. New York, NY: Springer-Verlag.
Douglas, J. (1997). Joint consistency of nonparametric item characteristic curve and ability
estimation. Psychometrika, 62, 7–28.Drasgow, F., Luecht, R. M., & Bennett, R. E. (2006). Technology and testing. In R. L. Brennan
(Ed.), Educational measurement (4th ed., pp. 471–515). Westport, CT: American Council on
Education and Praeger.
Ebel, R. L. (1962). Content standard test scores. Educational and Psychological Measurement,22(1), 15–25.
Edwards, M. C., &Wirth, R. J. (2009). Measurement and the study of change. Research in HumanDevelopment, 2–3, 74–96.
References 349
Efron, B. (1979). Bootstrap methods: Another look at the jackknife. Annals of Statistics, 7, 1–26.Efron, B. (1982). The jackknife, the bootstrap and other resampling plans. Philadelphia, PA:
Society for Industrial and Applied Mathematics.
Efron, B., & Morris, C. (1977). Stein’s paradox in statistics. Scientific American, 236, 119–127.Efron, B., & Tibshirani, R. J. (1993). An introduction to the bootstrap. New York, NY: Chapman
& Hall.
Embretson, S. E. (1996). The new rules of measurement. Psychological Assessment, 8, 341–349.Embretson, S. E., & Reise, S. P. (2000). Item response theory for psychologists. Hillsdale, NJ:
Erlbaum.
Fairbank, B. A. (1987). The use of presmoothing and postsmoothing to increase the precision of
equipercentile equating. Applied Psychological Measurement, 11, 245–262.Ferguson, T. (1973). A Bayesian analysis of some nonparametric problems. Annals of Statistics, 1,
209–230.
Feuer, M. J., Holland, P. W., Green, B. F., Bertenthal, M. W., & Hemphill, F.C. (Eds.). (1999).
Uncommon measures: Equivalence and linkage among educational tests (Report of the
Committee on Equivalency and Linkage of Educational Tests, National Research Council).
Washington DC: National Academy Press.
Fisher, R. A. (1960). The design of experiments (7th ed.). New York, NY: Hafner.
Flanagan, J. C. (1951). Units, scores, and norms. In E. F. Lindquist (Ed.), Educational measure-ment (pp. 695–763). Washington DC: American Council on Education.
Freeman, M. F., & Tukey, J. W. (1950). Transformations related to the angular and square root.
Annals of Mathematical Statistics, 21(4), 607–611.Gibbons, R. D., & Hedeker, D. (1992). Full-information item bi-factor analysis. Psychometrika,
57, 423–436.Gilula, Z., & Haberman, S. J. (2000). Density approximation by summary statistics: An information-
theoretic approach. Scandinavian Journal of Statistics, 27, 521–534.Glas, C. A. W. (1988). The RM and multi-stage testing. Journal of Educational Statistics,
13, 45–52.Glas, C. A. W. (1999). Modification indices for the 2-pl and the nominal response model.
Psychometrika, 64, 273–294.Glas, C. A. W. (2006). Testing generalized Rasch models. In M. von Davier & C. H. Carstensen
(Eds.), Multivariate and mixture distribution Rasch models (pp. 37–46). New York, NY:
Springer-Verlag.
Glas, C. A. W., & Verhelst, N. D. (1989). Extensions of the partial credit model. Psychometrika,54, 635–659.
Glas, C. A. W., & Verhelst, N. D. (1995). Testing the Rasch model. In G. H. Fischer & I. W.
Molenaar (Eds.), Rasch models: Foundations, recent developments and applications(pp. 69–96). New York, NY: Springer-Verlag.
Grimm, K. J. (n.d.). Kevin J. Grimm: Script downloads. Available from the University of
California–Davis website: http://psychology.ucdavis.edu/labs/Grimm/personal/downloads.
html
Haberman, S., Guo, H., Liu, J., & Dorans, N. J. (2008). Trend analysis in seasonal time seriesmodels. Consistency of SAT® reasoning score conversions (ETS Research Rept. RR-08-67).
Princeton, NJ: ETS.
Haebara, T. (1980). Equating logistic ability scales by a weighted least squares method. JapanesePsychological Research, 22, 144–149.
Hall, P. (1997). The bootstrap and Edgeworth expansion. New York, NY: Springer-Verlag.
(Corrected printing of 1992 ed.).
Hambleton, R. K., & Pitoniak, M. J. (2006) Setting performance standards. In R. L. Brennan (Ed.),
Educational measurement (4th ed., pp. 433–470). Westport, CT: American Council on Educa-
tion and Praeger.
Hambleton, R. K., & Swaminathan, H. (1985). Item response theory: Practices and applications.Boston, MA: Kluwer Academic.
350 References
Hambleton, R. K., Swaminathan, H., & Rogers, H. J. (1991). Fundamentals of item responsetheory. Newbury Park, CA: Sage.
Hanson, B. A., & Beguin, A. A. (2002). Obtaining a common scale for item response theory item
parameters using separate versus concurrent estimation in the common-item equating design.
Applied Psychological measurement, 26, 3–24.Harris, D. J. (2007). Practical issues in vertical scaling. In N. J. Dorans, M. Pommerich, & P. W.
Holland (Eds.), Linking and aligning scores and scales (pp. 233–251). New York, NY:
Springer-Verlag.
Harris, D. J., & Gao, X. (2003, April). A conceptual synthesis of context effect. In Context effects:Implications for pretesting and CBT. Symposium conducted at the meeting of the American
Educational Research Association, Chicago, IL.
He, X., & Ng, P. (1998). SCOBS: Qualitatively constrained smoothing via linear programming.Unpublished sofware manual.
He, X., & Shi, P. (1998). Monotone B-splines smoothing. Journal of the American StatisticalAssociation, 93, 643–650.
Holland, P.W. (1994). Measurements or contests? Comments on Zwick, Bond and Allen/Donoghue.
InProceedings of the Social Statistics Section of the American Statistical Association (pp. 27–29).Alexandria, VA: American Statistical Association.
Holland, P.W. (2007). A framework and history for score linking. In N. J. Dorans, M. Pommerich, &
P. W. Holland (Eds.), Linking and aligning scores and scales (pp. 5–30). New York, NY:
Springer-Verlag.
Holland, P. W., & Dorans, N. J. (2006). Linking and equating. In R. L. Brennan (Ed.), Educationalmeasurement (4th ed., pp. 189–220). Westport, CT: Praeger.
Holland, P. W., & Hoskens, M. (2003). Classical test theory as a first-order item response theory:
Application to true-score prediction from a possibly nonparallel test. Psychometrika, 68,123–149.
Holland, P. W., King, B. F., & Thayer, D. T. (1989). The standard error of equating for the kernelmethod of equating score distributions (ETS Tech. Rept. No. TR-89-83). Princeton, NJ: ETS.
Holland, P. W., & Rubin, D. B. (Eds.). (1982). Test equating. New York, NY: Academic Press.
Holland, P. W., Sinharay, S., von Davier, A. A., & Han, N. (2008). An approach to evaluating the
missing data assumptions of the chain and post-stratification equating methods for the NEAT
design. Journal of Educational Measurement, 45, 17–43.Holland, P. W., & Strawderman, W. (1989). The symmetric average of equating functions.
Unpublished manuscript.
Holland, P. W., & Thayer, D. T. (1987). Notes on the use of log-linear models for fitting discreteprobability distributions (ETS Research Rept. RR-87-31). Princeton, NJ: ETS.
Holland, P. W., & Thayer, D. T. (1989). The kernel method of equating score distributions (ETSResearch Rept. RR-89-07). Princeton, NJ: ETS.
Holland, P. W., & Thayer, D. T. (2000). Univariate and bivariate loglinear models for discrete test
score distributions. Journal of Educational and Behavioral Statistics, 25, 133–183.Holland, P. W., & Wainer, H. (1993). Differential item functioning. Hillsdale, NJ: Erlbaum.
Horn, J. L., & McArdle, J. J. (1992). A practical guide to measurement invariance in aging
research. Experimental Aging Research, 18, 117–144.Horn, J. L., McArdle, J. J., & Mason, R. (1983). When is invariance not invariant: A practical
scientist’s view of the ethereal concept of factorial invariance. The Southern Psychologist, 1,179–188.
Jaffa, A. S. (1934). The California Preschool Mental Scale, Form A. Berkeley, CA: University of
California Press.
Inspectorate of Secondary Education in the Netherlands. (1992). Examens op Punten GetoetstEvaluation of Examinations. Gravenhage, The Netherlands: Inspectie van het Voortgezet
Onderwijs.
Iowa Tests of Educational Development. (1958). Manual for the school administrator (Rev. ed.).Iowa City: State University of Iowa.
References 351
Janssen, R., Magis, D., San Martin, E., & Del Pino, G. (2009, April). Local equating in the NEATdesign. Paper presented at the meeting of the National Council on Measurement in Education,
San Diego, CA.
Kane, M. T., Mroch, A. A., Suh, Y., & Ripkey, D. R. (2009). Linear equating for the NEAT design:
Parameter substitution models and chained linear relationship models.Measurement: Interdis-ciplinary Research & Perspective, 7(3&4), 125– 146.
Kao, C. W., Kim, S., & Hatrak, N. (2005, October). Scale drift study for a large-scale Englishproficiency test. Paper presented at the meeting of the Northeastern Educational Research
Association, Kerhonkson, NY.
Karabatsos, G., & Walker, S. (2009a). A Bayesian nonparametric approach to test equating.
Psychometrika, 74(2), 211–232.Karabatsos, G., & Walker, S. (2009b). Coherent psychometric modeling with Bayesian nonpara-
metrics. British Journal of Mathematical and Statistical Psychology, 62(1), 1–20.Kaskowitz, G. S., & De Ayala, R. J. (2001). The effect of error in item parameter estimates on
the test response function method of linking. Applied Psychological Measurement, 25(1),39–52.
Kendall, M. G., & Stuart, A. (1977). The advanced theory of statistics (4th ed.). New York, NY:
Macmillan.
Kim, S., & Cohen, A. S. (2002). A comparison of linking and concurrent calibration under the
graded response model, Applied Psychological measurement, 26, 25–41.Kim, S., & Livingston, S. A. (2010). Comparisons among small sample equating methods in a
common-item design. Manuscript submitted for publication.
Kim, S., & Livingston, S. (2009).Methods of linking with small samples in a common-item design:An empirical comparison (ETS Research Rept. RR-09-38). Princeton, NJ: ETS.
Kim, S., Livingston, S. A., & Lewis, C. (2008). Investigating the effectiveness of collateralinformation on small-sample equating (ETS Research Rept. RR-08-52). Princeton, NJ: ETS.
Kim, S., Livingston, S. A., & Lewis, C. (2009). Evaluating sources of collateral information onsmall-sample equating (ETS Research Rept. RR-09-14). Princeton, NJ: ETS.
Kim, S., & von Davier, A. A. (2006, April). Equating with small samples in non-equivalent-groups
anchor test design. In Recent advances in score equating. Symposium conducted at the meeting
of the National Council for Measurement in Education, San Francisco, CA.
Kim, S., von Davier, A. A., & Haberman, S. (2008). Small sample equating using a synthetic
linking function. Journal of Educational Measurement, 45, 325–342.Koenker, R., Ng, P., & Portnoy, S. (1994). Quantile smoothing splines. Biometrika, 81, 673–680.Kolen, M. J. (1988). Defining score scales in relation to measurement error. Journal of Educa-
tional Measurement, 25(2), 97–110.Kolen, M. J. (2004). Linking assessments: Concept and history. Applied Psychological Measure-
ment, 28, 219–226.Kolen, M. J. (2006). Scaling and norming. In R. L. Brennan (Ed.), Educational measurement
(4th ed., pp. 155–186). Westport, CT: Praeger.
Kolen, M. J. (2007). Data collection designs and linking procedures. In N. J. Dorans,
M. Pommerich, & P. W. Holland (Eds.), Linking and aligning scores and scales (pp. 31–55).New York, NY: Springer-Verlag.
Kolen, M. J., & Brennan, R. J. (1995). Test equating: methods and practices. New York, NY:
Springer-Verlag.
Kolen, M. J., & Brennan R. L. (2004). Test equating, scaling, and linking: Method and practice(2nd ed.). New York, NY: Springer-Verlag.
Kolen, M. J., Hanson, B. A., & Brennan, R. L. (1992). Conditional standard errors of measurement
for scale scores. Journal of Educational Measurement, 29(4), 285–307.Kolen, M. J., & Jarjoura, D. (1987). Analytic smoothing for equipercentile equating under the
common item nonequivalent populations design. Psychometrika, 52, 43–59.Kolen, M. J., Zeng, L., & Hanson, B. A. (1996). Conditional standard errors of measurement for
scale scores using IRT. Journal of Educational Measurement, 33(2), 129–140.
352 References
Koretz, D., Barron, S., Mitchell, K., & Stecher, B. M. (1999). Perceived effects of the KentuckyInstructional Results Information System (KIRIS). Santa Monica, CA: Rand Corporation.
Koretz, D. M., Bertenthal, M.W., & Green, B. F. (Eds.). (1999). Embedding questions: The pursuitof a common measure in uncommon tests (Report of the Committee on Embedding Common
Test Items in State and District Assessments, National Research Council). Washington DC:
National Academy Press.
Kuehl, R. O. (2000). Design of experiments: Statistical principles of research design and analysis(2nd ed.). Duxbury, MA: Thomson Learning.
Kupperman, M. (1952). On exact grouping correlations to moments and cumulants. Biometrika,39, 429–434.
Leary, L. F., & Dorans, N. J. (1985). Implications for altering the context in which test items
appear: An historical perspective on an immediate concern. Review of Educational Research,55, 387–413.
Lee, Y.-H., & von Davier, A. A. (2008). Comparing alternative kernels for the kernel methodof test equating: Gaussian, logistic and uniform kernels (ETS Research Rept. RR-08-12).
Princeton, NJ: ETS.
Lee, Y.-S. (2002). Applications of isotonic regression in item response theory (Unpublished
doctoral dissertation). University of Wisconsin, Madison.
Lehmann, E.L. (1999). Elements of large-sample theory. New York, NY: Springer-Verlag.
Levine, R. (1955). Equating the score scales of alternative forms administered to samples ofdifferent ability (ETS Research Bulletin RB-55-23). Princeton, NJ: ETS.
Liang, L., Dorans, N. J., & Sinharay, S. (2009). First language of examinees and its relationship toequating (ETS Research Rept. RR-09-05). Princeton, NJ: ETS.
Liou, M. (1998). Establishing score comparability in heterogeneous populations. Statistica Sinica,8, 669–690.
Liou, M., Cheng, P. E., & Johnson, E. G. (1997). Standard errors of the kernel equating methods
under the common-item design. Applied Psychological Measurement, 21, 349–369.Liou, M., Cheng, P. E., & Li, M.-Y. (2001). Estimating comparable scores using surrogate
variables. Applied Psychological Measurement, 25, 197–207.Livingston, S. A. (1993). Small-sample equating with log-linear smoothing. Journal of Educa-
tional Measurement, 30, 23–29.Livingston, S. A. (2004). Equating test scores (without IRT). Princeton, NJ: ETS.Livingston, S. A., Dorans, N. J., & Wright, N. K. (1990). What combination of sampling and
equating methods works best? Applied Measurement in Education, 3, 73–95.Livingston, S. A., & Kim, S. (2008). Small-sample equating by the circle-arc method (ETS
Research Rept. RR-08-39). Princeton, NJ: ETS.
Livingston, S. A., & Kim, S. (2009). The circle-arc method for equating in small samples. Journalof Educational Measurement, 46, 330–343.
Livingston, S. A., & Kim, S. (2010). An empirical comparison of methods for equating withrandomly equivalent groups of 50 to 400 test takers (ETS Research Rept. RR-10-05). Princeton,
NJ: ETS.
Livingston, S. A., & Kim, S. (in press). Random-groups equating with samples of 50 to 400 test
takers. Journal of Educational Measurement.Livingston, S. A., & Lewis, C. (2009). Small-sample equating with prior information (ETS
Research Rept. RR-09-25). Princeton, NJ: ETS.
Ljung, G., & Box, G. (1978). On a measure of lack of fit in time-series models. Biometrika, 65,297–303.
Lord, F. M. (1950). Notes on comparable scales for test scores (ETS Research Bulletin RB-50-48).
Princeton, NJ: ETS.
Lord, F. M. (1955a). Equating test scores—A maximum likelihood solution. Psychometrika, 20,193–200.
Lord, F. M. (1955b). Estimation of parameters from incomplete data. Journal of the AmericanStatistical Association, 50, 870–876.
References 353
Lord, F. M. (1980). Applications of item response theory to practical testing problems. Hillsdale,NJ: Erlbaum.
Lord, F. M. (1982a). Standard error of an equating by item response theory. Applied PsychologicalMeasurement, 6, 463–472.
Lord, F. M. (1982b). The standard error of equipercentile equating. Journal of EducationalStatistics, 7, 165–174.
Lord, F. M. (1983). Small N justifies Rasch model. In D. J. Weiss (Ed.), New horizons in testing(pp. 51–61). New York, NY: Academic Press.
Lord, F. M., & Novick, M. R. (1968). Statistical theories of mental test scores. Reading, MA:
Addison Wesley.
Lord, F. M., & Wingersky, M. S. (1984). Comparison of IRT true-score and equipercentile
observed-score “equatings.” Applied Psychological Measurement, 8, 452–461.Lorentz, G. (1953). Bernstein polynomials. Toronto, Ontario, Canada: University of Toronto Press.Loyd, B. H., & Hoover, H. D. (1980). Vertical equating using the Rasch model. Journal of
Educational Measurement, 17, 179–193.Marco, G. L. (1977). Item characteristic curve solutions to three intractable testing problems.
Journal of Educational Measurement, 14, 139–160.Marco, G. L., Petersen, N. S., & Stewart, E. E. (1983a). A large-scale evaluation of linear and
curvilinear score equating models, Volumes I and II (ETS Research Memorandum RM-83-02).
Princeton, NJ: ETS.
Marco, G. L., Petersen, N. S., & Stewart, E. E. (1983b). A test of the adequacy of curvilinear score
equating models. In D. J. Weiss (Ed.), New horizons in testing (pp. 147–176). New York, NY:
Academic Press.
Masters, G. N. (1982). A Rasch model for partial credit scoring. Psychometrika, 47, 149–174.McArdle, J. J. (2007). Five steps in the structural factor analysis of longitudinal data. In R. Cudeck
& R. MacCallum (Eds.), Factor analysis at 100 years (pp. 99–130). Mahwah, NJ: Erlbaum.
McArdle, J. J. (1994). Structural factor analysis experiments with incomplete data. MultivariateBehavioral Research, 29, 409–454.
McArdle, J. J., & Cattell, R. B. (1994). Structural equation models of factorial invariance in
parallel proportional profiles and oblique confactor problems. Multivariate BehavioralResearch, 29, 63–113.
McArdle, J. J., Ferrer-Caja, E., Hamagami, F., & Woodcock, R. W. (2002). Comparative longitu-
dinal structural analyses of the growth and decline of multiple intellectual abilities over the life
span. Developmental Psychology, 38, 115–142.McArdle, J. J., Grimm, K. J., Hamagami, F., Bowles, R., & Meredith, W. (2009). Modeling life-
span growth curves of cognition using longitudinal data with multiple samples and changing
scales of measurement. Psychological Methods, 14, 126–149.McArdle, J. J., Hamagami, F., Meredith, W., & Bradway, K. P. (2000). Modeling the dynamic
hypotheses of Gf-Gc theory using longitudinal life-span data. Learning and Individual Differ-ences, 12, 53–79.
McCall, W. A. (1939).Measurement: A revision of how to measure in education. New York, NY:
Macmillan.
Mislevy, R. J., & Bock, R. D. (1990). BILOG 3. Item analysis and test scoring with binary logisticmodels. Mooresville, IN: Scientific Software.
Molenaar, I. W. (1983). Some improved diagnostics for failure in the Rasch model. Psychome-trika, 48, 49–72.
Montgomery, D. C., Peck, E. A., & Vining, G. G. (2001). Introduction to linear regressionanalysis (3rd ed.). New York, NY: John Wiley & Sons.
Morris, C. N. (1982). On the foundations of test equating. In P. W. Holland & D. B. Rubin (Eds.),
Test equating (pp. 9–49). New York, NY: Academic Press.
Morrison, C. A., & Fitzpatrick, S. J. (1992). Direct and indirect equating: A comparison of fourmethods using the Rasch model (Report No. RB-91-3). Austin: University of Texas, Measure-
ment and Evaluation Center.
354 References
Moses, T., & Holland, P. W. (2008). The influence of strategies for selecting loglinear smoothingmodels on equating functions (ETS Research Rept. RR-08-25). Princeton, NJ: ETS.
Moses, T., & Kim, S. (2007). Reliability and the nonequivalent groups with anchor test design(ETS Research Rept. RR-07-16). Princeton, NJ: ETS.
Mosteller, F., & Youtz, C. (1961). Tables of the Freeman-Tukey transformations for the binomial
and Poisson distributions. Biometrika, 48, 433–440.M€uller, P., & Quintana, F. (2004). Nonparametric Bayesian data analysis. Statistical Science, 19,
95–110.
Muraki, E. (1993) Information functions of the generalized partial credit model. Applied Psycho-logical Measurement, 14(4), 351–363.
Muraki, E. (1992) A generalized partial credit model: Application of an EM algorithm. AppliedPsychological Measurement, 16, 159–176.
Muraki, E., & Bock, R. D. (1991). PARSCALE: Parametric scaling of rating data. Chicago, IL:Scientific Software International.
Nadaraya, E. A. (1964). On estimating regression. Theory of Probability and its Applications, 9,141–142.
Ogasawara, H. (2000). Asymptotic standard errors of IRT equating coefficients using moments.
Economic Review, Otaru University of Commerce, 51(1), 1–23.Ogasawara, H. (2001a). Item response theory true score equatings and their standard errors.
Journal of Educational and Behavioral Statistics, 26, 31–50.Ogasawara, H. (2001b). Standard errors of item response theory equating/linking by response
function methods. Applied Psychological Measurement, 25, 53–67.Ogasawara, H. (2003). Asymptotic standard errors of IRT observed-score equating methods.
Psychometrika, 68, 193–211.Ogasawara, H. (2006). Asymptotic expansion of the sample correlation coefficient under non-
normality. Computational Statistics and Data Analysis, 50, 891–910.Ogasawara, H. (2007a). Asymptotic expansions of the distributions of the estimators in canonical
correlation analysis under nonnormality. Journal of Multivariate Analysis, 98, 1726–1750.Ogasawara, H. (2007b). Asymptotic expansion of the distributions of the estimators in factor
analysis under nonnormality. British Journal of Mathematical and Statistical Psychology, 60,395–420.
Ogasawara, H. (2007c). Higher-order estimation error in structural equation modeling. EconomicReview, Otaru University of Commerce, 57(4), 131–160.
Ogasawara, H. (2009). Asymptotic cumulants of the parameter estimators in item response theory.
Computational Statistics, 24, 313–331.Orlando, M., & Thissen, D. (2000). Likelihood-based item-fit indices for dichotomous item
response theory models. Applied Psychological Measurement, 24, 50–64.Patz, R. J., & Yao, L. (2007). Methods and models for vertical scaling. In N. J. Dorans, M.
Pommerich, & P. W. Holland (Eds.), Linking and aligning scores and scales (pp. 253–272).New York, NY: Springer-Verlag.
Patz, R., Yao, L., Chia, M., Lewis, D., & Hoskens, M. (2003, April). Hierarchical and multidi-mensional models for vertical scaling. Paper presented at the meeting of the National Council
on Measurement in Education, Chicago, IL.
Peterson, N. S. (2007). Equating: Best practices and challenges to best practices. In N. J. Dorans,
M. Pommerich, & P. W. Holland (Eds.), Linking and aligning scores and scales (pp. 59–72).New York, NY: Springer-Verlag.
Petersen, N. S., Cook, L. L., & Stocking, M. L. (1983). IRT versus conventional equating methods:
A comparative study of scale stability. Journal of Educational Statistics, 8(2), 137–156.Petersen, N. S., Kolen, M. J., & Hoover, H. D. (1989). Scaling, norming and equating. In R. L.
Linn (Ed.), Educational measurement (3rd ed., pp. 221–262). New York, NY: American
Council on Education and Macmillan.
Petersen, N. S., Marco, G. L., & Stewart, E. E. (1982). A test of the adequacy of linear score
equating models. In P. W. Holland & D. B. Rubin (Eds.), Test equating (pp. 71–135). New
York, NY: Academic Press.
References 355
Petrone, S. (1999). Random Bernstein polynomials. Scandinavian Journal of Statistics, 26,373–393.
Pommerich, M., & Dorans, N. J. (Eds.). (2004). Concordance [Special issue]. Applied Psychologi-cal Measurement, 28(4).
Pommerich, M., Nicewander, W. A., & Hanson, B. A. (1999). Estimating average domain scores.
Journal of Educational Measurement, 36(3), 199–216.Puhan, G., Moses, T. P., Grant, M. C., &McHale, F. (2009). Small-sample equating using a single-
group nearly equivalent test (SiGNET) design. Journal of Educational Measurement, 46,344–362.
Ramsay, J. O. (1991). Kernel smoothing approaches to nonparametric item characteristic curve
estimation. Psychometrika, 56, 611–630.Ramsay, J. O. (2001). TestGraf 98. A program for the graphical analysis of multiple choice test
and questionnaire data. Retrieved from http://www.psych.mcgill.ca/faculty/ramsay/TestGraf.
html
Ramsay, J. O., & Abrahamowicz, M. (1989). Binomial regression with monotone splines: A
psychometric application. Journal of the American Statistical Association, 84, 906–915.Rao, C. R. (1973). Linear statistical inference and its applications (2nd ed.). New York, NY:
Wiley.
Rasch, G. (1960). Probabilistic models for some intelligence and attainment tests. Chicago, IL:University of Chicago Press.
Ricker, K., & von Davier, A. A. (2007). The impact of anchor test length on equating results in anon-equivalent group design (ETS Research Rept. RR-07-44). Princeton, NJ: ETS.
Rijmen, F. (2009a, July). A hierarchical factor IRT model for items that are clustered at multiplelevels. Paper presented at the InternationalMeeting of the Psychometric Society, Cambridge, UK.
Rijmen, F. (2009b). Three multidimensional models for testlet based tests: Formal relations andan empirical comparison (ETS Research Rept. No. RR-09-37). Princeton, NJ: ETS.
Rock, D. A. (1982). Equating using confirmatory factor analysis. In P. W. Holland & D. B. Rubin
(Eds.), Test equating (pp. 247–258). New York, NY: Academic Press.
Rodriguez, M. C. (2003). Construct equivalence of multiple-choice and constructed-response items:
A random effects synthesis of correlations. Journal of Educational Measurement, 40(2),163–184.
Rosa, K., Swygert, K. A., Nelson, L., & Thissen, D. (2001). Item response theory applied to
combinations of multiple-choice and constructed-response items—Scale scores for patterns of
summed scores. In D. Thissen & H. Wainer (Eds.), Test scoring (pp. 253–292). Mahwah, NJ:
Erlbaum.
Rosenbaum, P. R. (1995). Observational studies. New York, NY: Springer-Verlag.
Rossi, N., Wang, X., & Ramsay, J. O. (2002). Nonparametric item response function estimates
with the EM algorithm. Journal of Educational and Behavioral Statistics, 27, 291–317.Rost, J., & von Davier, M. (1992). MIRA – A PC program for the mixed Rasch model [User
manual]. Kiel, Germany: IPN.
Rost, J., & von Davier, M. (1995). Mixture distribution Rasch models. In G. H. Fischer & I. W.
Molenaar (Eds.), Rasch models—Foundations, recent developments and applications (pp.
257–268). New York, NY: Springer-Verlag.
Rubin, D. (1982). Discussion of “Observed-score test equating: A mathematical analysis of some
ETS equating procedures.” In P. W. Holland & D. B. Rubin (Eds.), Test equating (pp. 51–54).
New York, NY: Academic Press.
Samejima, F. (1969). Estimation of latent ability using a response pattern of graded scores.
Psychometrika Monograph Supplement, 34(4, Pt. 2), 100–114.Schwarz, G. (1978). Estimating the dimensions of a model. Annals of Statistics, 6, 461–464.Sethuraman, J. (1994). A constructive definition of Dirichlet priors. Statistica Sinica, 4, 639–650.Shealy, R., & Stout, W. F. (1993). A model-based standardization approach that separates true
bias/DIF from group ability differences and detects test bias/DTF as well as item bias/DIF.
Psychometrika, 58, 159–194.
356 References
Shumway, R. H., & Stoffer, D. S. (2006). Time-series analysis and its applications with Rexamples (2nd ed.). New York, NY: Springer-Verlag.
Sinharay, S., & Holland, P. W. (2007). Is it necessary to make anchor tests mini-versions of the
tests being equated or can some restrictions be relaxed? Journal of Educational Measurement,44, 249–275.
Sinharay, S., & Holland, P. W. (2010a). The missing data assumptions of the NEAT design and
their implications for test equating. Psychometrika, 75, 309–327.Sinharay, S., & Holland, P. W. (2010b). A new approach to comparing several equating methods in
the context of the NEAT design. Journal of Educational Measurement, 47(3), 261–285.Sinharay, S., & Holland, P. W. (in press). A fair comparison of three nonlinear equating methods in
applications of the NEAT design. Journal of Educational Measurement.Skaggs, G. (2005). Accuracy of random groups equating with very small samples. Journal of
Educational Measurement, 42, 309–330.Stocking, M. L., & Lord, F. M. (1983). Developing a common metric in item response theory.
Applied Psychological Measurement, 7, 201–210.Stuart, A., & Ord, K. (1994). Kendall’s advanced theory of statistics: Distribution theory (6th ed.,
Vol. 1). London, England: Arnold.
Terman, L. M. (1916). The measurement of intelligence: An explanation of and a complete guidefor the use of the Stanford revision and extension of the Binet-Simon Intelligence Scale.Cambridge, MA: The Riverside Press.
Terman, L. M., & Merrill, M. A. (1937). Measuring intelligence: A guide to the administration ofthe new Revised Stanford-Binet tests of intelligence. Cambridge, MA: The Riverside Press.
Terman, L. M., & Merrill, M. A. (1960). Stanford-Binet Intelligence Scale: Manual for the thirdrevision Form L-M. Cambridge, MA: The Riverside Press.
Thissen, D. (1991).Multilog user’s guide: Multiple, categorization analysis and test scoring usingitem response theory. Chicago, IL: Scientific Software International.
Thissen, D., Nelson, L., & Swygert, K. A. (2001). Item response theory applied to combinations of
multiple-choice and constructed-response items—Approximation methods for scale scores. In
D. Thissen & H. Wainer (Eds.), Test scoring (pp. 293–341). Mahwah, NJ: Erlbaum.
Thissen, D., & Orlando, M. (2001). Item response theory for items scored in two categories. In
D. Thissen & H. Wainer (Eds.), Test scoring (pp. 73–140). Mahwah, NJ: Erlbaum.
Thissen, D., Pommerich, M., Billeaud, K., & Williams, V. S. L. (1995). Item response theory for
scores on tests including polytomous items with ordered responses. Applied PsychologicalMeasurement, 19(1), 39–49.
Thissen, D., & Steinberg, L. (1986). A taxonomy of item response models. Psychometrika, 51,567–577.
Thissen, D., & Wainer, H. (Eds.). (2001). Test scoring. Mahwah, NJ: Erlbaum.
Thisted, R. A. (1988). Elements of statistical computing: Numerical computation. New York, NY:
Champman & Hall/CRC.
Tong, Y., & Kolen, M. J. (2005). Assessing equating results on different equating criteria. AppliedPsychological Measurement, 29(6), 418–432.
Tong, Y., & Kolen, M. J. (2007). Comparisons of methodologies and results in vertical scaling for
educational achievement tests. Applied Measurement in Education, 20(2), 227–253.Tsai, T.-H., Hanson, B. A., Kolen, M. J., & Forsyth, R. A. (2001). A comparison of bootstrap
standard errors of IRT equating methods for the common-item nonequivalent groups design.
Applied Measurement in Education, 14, 17–30.Tukey, J. W. (1963). Mathematics 596—An introduction to the frequency analysis of time series.
In D. R. Brillinger (Ed.), The collected works of John W. Tukey, Volume I: Time series,1949–1964. London, England: Chapman & Hall.
van der Linden, W. J. (1997). [Review of the book Test equating: Methods and practices by M. J.
Kolen & R. L Brennan]. Psychometrika, 62, 287–290.van der Linden, W. J. (2000). A test-theoretic approach to observed-score equating. Psychome-
trika, 65, 437–456.
References 357
van der Linden, W. J. (2005). Linear models for optimal test design. New York, NY: Springer-
Verlag.
van der Linden, W. J. (2006a). Equating error in observed-score equating. Applied PsychologicalMeasurement, 30, 355–378.
van der Linden, W. J. (2006b). Equating an adaptive test to a linear test. Applied PsychologicalMeasurement, 30, 493–508.
van der Linden,W. J. (2006c). [Review of the book The kernel method of test equating by A. A. vonDavier, P. W. Holland & D. T. Thayer]. Journal of Educational Measurement, 43, 291–294.
van der Linden, W. J. (2010). Linking response-time parameters onto a common scale. Journal ofEducational Measurement, 47, 92–114.
van der Linden, W. J., & Hambleton, R. K. (Eds.). (1997). Handbook of modern item responsetheory. New York, NY: Springer-Verlag.
van der Linden, W. J., & Wiberg, M. (in press). Local observed score equating with anchor test
designs. Applied Psychological Measurement, 34.Van Krimpen-Stoop, E. M. L. A., & Meijer, R. R. (2001). CUSUM-based person-fit statistics for
adaptive testing. Journal of Educational and Behavioral Statistics, 26, 199–217.Veerkamp, W. J. J., & Glas, C. A.W. (2000). Detection of known items in adaptive testing with a
statistical quality control method. Journal of Educational and Behavioral Statistics, 25, 373–389.von Davier, A. A. (2003a). Large sample tests for comparing regression coefficients in models
with normally distributed variables (ETS Research Rept. RR-03-19). Princeton, NJ: ETS.
von Davier, A. A. (2003b). Notes on linear equating methods for the non-equivalent groups design(ETS Research Rept. RR-03-24). Princeton, NJ: ETS.
von Davier, A. A. (2007). Potential solutions to practical equating issues. In N. J. Dorans,
M. Pommerich, & P. W. Holland (Eds.), Linking and aligning scores and scales (pp. 89–106).New York, NY: Springer-Verlag.
von Davier, A. A. (2008). New results on the linear equating methods for the non-equivalent-group
design. Journal of Educational and Behavioral Statistics, 33, 186–203.von Davier, A. A. (in press). An observed-score equating framework. In N. J. Dorans & S. Sinharay
(Eds.), A conference in honor of Paul W. Holland. New York, NY: Springer-Verlag.
von Davier, A. A., Fournier-Zajac, S., & Holland, P. W. (2006, April). An equipercentile versionof the Levine linear observed-score equating function using the methods of kernel equating.Paper presented at the meeting of the National Council of Measurement in Education, San
Francisco, CA.
von Davier, A. A., Fournier-Zajac, S., & Holland, P. W. (2007). An equipercentile version ofthe Levine linear observed-score equating function using the methods of kernel equating(ETS Research Rept. RR-07-14). Princeton, NJ: ETS.
von Davier, A. A., Holland, P. W., Livingston, S. A., Casabianca, J., Grant, M. C., & Martin, K.
(2006). An evaluation of the kernel equating method: A special study with pseudo-testsconstructed from real test data (ETS Research Rept. RR-06-02). Princeton, NJ: ETS.
von Davier, A. A., Holland, P. W., & Thayer, D. T. (2003). Population invariance and chain versus
poststratification methods for equating and test linking. In N. Dorans (Ed.), Populationinvariance of score linking: Theory and applications to Advanced Placement Program®Examinations (ETS Research Rept. RR-03-27). Princeton, NJ: ETS.
von Davier, A. A., Holland, P. W., & Thayer, D. T. (2004a). The chain and poststratification
methods for observed-score equating: Their relationship to population invariance. Journal ofEducational Measurement, 41, 15–32.
von Davier, A. A., Holland, P. W., & Thayer, D. T. (2004b). The kernel method of test equating.New York, NY: Springer-Verlag.
von Davier, A. A., & Kong, N. (2005). A unified approach to linear equating for the nonequivalent
group design. Journal of Educational and Behavioral Statistics, 30, 313–342.von Davier, M. (1994). WINMIRA—A program for analyses with the Rasch model, with the latent
class analysis and with the mixed Rasch model [Computer software]. Kiel, Germany: IPN
Software, Institute for Science Education.
358 References
von Davier, M. (2000). WINMIRA 2001. AWindows program for analyses with the Rasch model,
with the latent class analysis and with the mixed Rasch model [Computer software]. St. Paul,
MN: Assessment Systems Corporation.
von Davier, M. (2005). A general diagnostic model applied to language testing data (ETS
Research Rept. RR-05-16). Princeton, NJ: ETS.
von Davier, M., & von Davier, A. A. (2004). Unified approach to IRT scale linking and scaletransformations (ETS Research Rept. RR-04-09). Princeton, NJ: ETS.
von Davier, M., & von Davier, A.A. (2007). A unified approach to IRT scale linking and scale
transformation. Methodology, European Journal of Research Methods for the Behavioral andSocial Sciences, 3(3), 115–124.
von Davier, M., &Yamamoto, K. (2004). Partially observed mixtures of IRTmodels: An extension
of the generalized partial credit model. Applied Psychological Measurement 28(6), 389–406.Wainer, H., & Thissen, D. (1993). Combining multiple-choice and constructed-response test scores:
Toward aMarxist theory of test construction. AppliedMeasurement in Education, 6(2), 103–118.Wald, A. (1943). Tests of statistical hypotheses concerning several parameters when the number of
observations is large. Transactions of the American Mathematical Society, 54, 426–482.Walker, S. (2004). New approaches to Bayesian consistency. Annals of Statistics, 32, 2028–2043.Walker, S., Damien, P., Laud, P., & Smith, A. (1999). Bayesian nonparametric inference for
random distributions and related functions. Journal of the Royal Statistical Society, Series B,61, 485–527.
Walker, S., Lijoi, A., & Pr€unster, I. (2007). On rates of convergence for posterior distributions in
infinite-dimensional models. Annals of Statistics, 35, 738–746.Walker, S., & Muliere, P. (2003). A bivariate Dirichlet process. Statistics and Probability Letters,
64, 1–7.Wang, T. (2008). The continuized log-linear method: An alternative to the kernel method of
continuization in test equating. Applied Psychological Measurement, 32, 527–542.Wang, T., & Brennan, R. L. (2007, April). A modified frequency estimation equating method for
the common-item non-equivalent groups design. Paper presented at the meeting of the National
Council of Measurement in Education, Chicago, IL.
Wang, T., Kolen, M. J., & Harris, D. J. (2000). Psychometric properties of scale scores and
performance levels for performance assessments using polytomous IRT. Journal of Educa-tional Measurement, 37(2), 141–162.
Wang, T., Lee, W., Brennan, R. J., & Kolen, M. J. (2006, April). A comparison of the frequencyestimation and chained equipercentile methods under the common-item non-equivalent groupsdesign. Paper presented at the meeting of the National Council on Measurement in Education,
San Francisco, CA.
Wasserman, L. (2006). All of nonparametric statistics. New York, NY: Springer-Verlag.
Watson, G. S. (1964). Smooth regression analysis. Sankhya Series A, 26, 359–372.Wechsler, D. (1955). Manual for the Wechsler Adult Intelligence Scale. New York, NY: The
Psychological Corporation.
Wechsler, D. (1981). Wechsler Adult Intelligence Scale-Revised manual. San Antonio, TX: The
Psychological Corporation.
Wechsler, D. (1946).Wechsler-Bellevue Intelligence Scale: Form I. Manual for administering andscoring the test. New York, NY: The Psychological Corporation.
Wiberg, M., & van der Linden, W. J. (2009). Local linear observed-score equating. Manuscript
submitted for publication.
Wirth, R. J., & Edwards, M. C. (2007). Item factor analysis: Current approaches and future
directions. Psychological Methods, 12, 58–79.Wright, N. K., & Dorans, N. J. (1993). Using the selection variable for matching or equating
(ETS Research Rept. RR-93-04). Princeton, NJ: ETS.
Xu, X.; & von Davier, M. (2008). Comparing multiple-group multinomial loglinear models for
multidimensional skill distributions in the general diagnostic model (ETS Research Rept. RR-
08-35). Princeton, NJ: ETS.
References 359
Yen, W. M. (1984). Obtaining maximum likelihood trait estimates from number-correct scores for
the three-parameter logistic model. Journal of Educational Measurement, 21, 93–111.Yen, W. M. (1986). The choice of scale for educational measurement: An IRT perspective.
Journal of Educational Measurement, 23(4), 299–325.Yen, W. M. (2007). Vertical scaling and no child left behind. In N. J. Dorans, M. Pommerich, &
P. W. Holland (Eds.), Linking and aligning scores and scales (pp. 273–283). New York, NY:
Springer-Verlag.
Yen, W. M., & Fitzpatrick, A. R. (2006). Item response theory. In R. L. Brennan (Ed.), Educa-tional measurement (4th ed., pp. 111–153). Westport, CT: Praeger.
Zeng, L., & Kolen, M. J. (1995). An alternative approach for IRT observed-score equating of
number-correct scores. Applied Psychological Measurement, 19, 231–240.Zimowski, M. F., Muraki, E., Mislevy, R. J., & Bock, R. D. (1996). BILOG-MG: Multiple-group
IRT analysis and test maintenance for binary items [Computer software]. Chicago, IL: Scien-
tific Software International.
Zwick, R., Senturk, D., Wang, J., & Loomis, S. C. (2001). An investigation of alternative methods
for item mapping in the National Assessment of Educational Progress. Educational Measure-ment: Issues and Practice, 20(2), 15–25.
360 References
Index
A
Abdi, H., 324
Abrahamowicz, M., 244
Abramowitz, M., 129, 140, 163, 305
Achenbach, T.M., 74
ACT, 48, 149
Aitchison, J., 226
Aitkin, M., 268
Akaike criterion, 4
Akaike, H., 323
Al-Karni, A., 271
Allen, N.L., 48, 57
Alternative kernels, 13, 159–173
Anchor test, 21, 24–25, 27–31, 34–41, 71, 139,
145, 152, 160, 175, 181, 183, 186,
193–195, 199, 218–220, 225–229, 231,
233, 234, 281–284, 288, 289, 291–296,
319, 326, 333
Andersen, E.B., 312
Angle bisector method, 89, 94–98
Angoff, W.H., 22, 24, 26, 29, 33, 46, 89, 94,
181, 186, 225
Armstrong, R.D., 329
Asymptotic distribution, 170, 321
Asymptotic expansion, 261–280
Asymptotic standard deviation, 125, 132, 133,
138, 139
B
Baker, F.B., 254, 271
Balakrishnan, N., 162
Bandwidth, 13, 132, 139, 148, 156, 160, 161,
165–166, 168–171, 173, 185, 193–198,
200, 245, 250, 254, 319, 323, 324
Ban, J.-C., 47, 58
Barron, S., viii
Bayesian nonparametrics, 176
Bayes’ theorem, 176, 179
Bayley, N., 72, 75
Beguin, A.A., 8, 297–316
Bennett, R.E., 44
Berkeley Growth Study, 72, 75
Bernstein polynomial prior, 178–180
Bertenthal, M.W., 22
Best practices, 21–42
Beta distribution, 176, 177, 179, 181
Bifactor model, 2, 8, 9
Billeaud, K., 51
BILOG-MG, 254, 301
Birnbaum, A., 300
Bishop, Y.M.M., 4
Bivariate Bernstein-Dirichlet process, 177
Bock, R.D., 56, 230, 254, 261, 262, 268, 301, 316
Bootstrap, 247, 256, 261, 310–313
Bowles, R., 74
Box and Jenkins models, 330
Box, G.E.P., 3, 330, 332, 339
Bozdogan, H., 4
Bradlow, E.T., 8
Bradway, K.P., 77
Bradway–McArdle Longitudinal Study, 74–77
Braiding plan, 328
Braun, H.I., 3–7, 11, 34, 186, 194, 212, 283
Braun–Holland equating, 194, 196
Brennan, R.L., 2, 16, 22–24, 26, 27, 30, 32, 33,
35, 37, 41–58, 60, 62, 65–67, 110, 111,
118, 143, 149, 152, 173, 175, 183, 186,
187, 190, 194, 196, 199, 221, 222, 261,
281, 282, 286, 296, 317, 328, 330, 343
Brockwell, P.J., 330, 331, 333, 345
B-splines, 244–247, 252
C
Calibration, 2, 10, 23, 64, 65, 68, 69, 226,
228–232, 237–241, 243, 247, 254, 299
California First-Year Mental Scale, 72
361
Campbell, N.R., 214
Carlson, J.E., 48, 59–70
Carmrer, S.G., 326
Casabianca, J., 282, 289
Cattell, R.B., 86
CDF. See Cumulative distribution function
CE. See Chained equating
Chained equating, 14, 15, 183–185, 187,
193–196, 200, 294, 295, 331
Chain equipercentile equating, 145
Chatfield, C., 330–332, 339, 345
Cheng, P.E., 213, 261
Chen, H.H., 8, 132, 185–200
Chia, M., 241
Circle-arc equating, 111–113, 121–122
Classical test theory, 31, 35, 190, 222
Cochran, D., 333
Cognitive abilities, 72, 73, 81, 82
Cohen, A.S., 299
Collateral information, 109, 116–119, 121, 222
Common-item design, 60–61, 64–69
Conditional standard error of equating (CSEE),
113, 114, 117, 119
Continuization, 4, 13, 15, 16, 37, 141–157,
159–161, 163, 164, 172, 193–195,
286, 323
Continuization with alternative kernels,
164–167
Continuized log-linear method (CLL),
141–157
Cook, L.L., 22, 328
Cornish-Fisher expansion, 268
Cross-grade scaling. See Vertical linkingCSEE. See Conditional standard error of
equating
Cudeck, R., 71
Cumulant-generating function, 162
Cumulants, 161–164, 166, 167, 169, 170,
266–277, 280
Cumulative distribution function (CDF),
14–17, 32, 33, 37, 146, 156, 160–165,
169, 173, 175–177, 179, 203, 247–249,
283, 285–287, 319, 320
D
Dalal, S., 179
Damien, P., 176
Data collection designs, 3, 4, 6, 9–12, 15, 21,
26–29, 31, 32, 37, 39, 40, 71, 141, 142,
159, 160, 225, 298, 326, 334
Data processing, 21, 35–37
Davis, R.A., 330, 333, 345
De Ayala, R.J., 250
Del Pino, G., 220
Design function (DF), 12, 14, 141–144, 147,
152–155, 160, 173, 319, 320, 322
Diaconis, P., 176, 179
Difference that matters, 38, 326
Direct equating methods, 328
Dirichlet process prior, 177–178
Divgi, D.R., 111
Dorans, N.J., 1, 2, 4, 6, 11, 21–42, 45, 48, 60,
62, 71, 110, 213, 282, 326
Douglas, J.A., 243–258
Draper, N.R., 3
Drasgow, F., 44
DTM. See Difference that matters
Durbin-Watson test, 332, 339, 341, 342
E
Ebel, R.L., 47
Educational measurement, 1, 2, 5, 8, 329
Edwards, M.C., 74, 86, 262
Efron, B., 116, 310
Eignor, D.R., 1, 21–42
Embretson, S.E., 71
Empirical Bayes, 116–118
Equating, 1–17, 21–42, 45, 48, 59, 70, 71, 74,
87–107, 109–122, 125–141, 143–147,
152–157, 159–173, 175–223, 225, 227,
241, 261, 281–331, 333–337, 340–343,
345
Equating assumptions, 193, 330
Equating in a circle, 328
Equating requirements, 11, 23
Equipercentile equating, 13, 17, 110, 113–115,
125, 126, 143, 145, 152, 159, 175, 176,
181, 184, 185, 193–197, 207, 208, 211,
213, 217–219, 222, 302, 308, 323
Equipercentile equating function, 13–17,
32–33, 37, 89, 113, 141, 159, 161, 176,
198, 296, 319, 325, 326
Equity, 23, 24, 41, 57, 58, 201, 204, 205,
207–209
Equivalent groups design, 4, 12, 26–29, 31–33,
35, 60–63, 68–69, 126, 139, 142–147,
152–154, 159, 161, 167, 175–177,
181–184, 318, 323, 326
Exponential families, 125–140
F
Fairbank, B.A., 37
Feigenbaum, M.D., 326
Ferguson, T., 179
362 Index
Ferrer-Caja, E., 81
Feuer, M.J., 22
Fienberg, S.E., 4
Fisher, R.A., 324
Fitzpatrick, A.R., 2, 49, 53, 62–66
Fitzpatrick, S.J., 328, 329
Five requirements, 11, 23, 24
Flanagan, J.C., 22, 46
Forsyth, R.A., 261
Fournier-Zajac, S., 16, 89, 196
Freeman, M.F., 47
Freeman-Tukey residuals, 4, 167, 287, 293
Frequency estimation equipercentile equating,
185, 196
G
Gao, X., 30
Gaussian kernel, 13, 141, 143, 156, 159, 168,
171, 172, 245, 319, 323
Generalized equating function, 185–200
Generalized least square estimates, 332
Generic kernel function, 161
Gibbons, R.D., 8, 316
Gibbs algorithm, 180, 182
Gilula, Z., 127, 128
Glas, C.A.W., 8, 226, 241, 297, 301, 303,
312, 313, 316, 329
Goodness-of-fit measures, 287, 292–294
Grant, M.C., 119, 120
Green, B.F., 22
Grimm, K.J., 71, 74
Guo, H., 25
Guttman items, 206
H
Haberman, S.J., 13, 17, 25, 31, 41, 42, 89,
119, 125, 127, 128
Haebara characteristic-curves approach, 226
Hall, P., 179
Hall, W., 262, 267
Hamagami, F., 74, 77, 81
Hambleton, R.K., 48, 49, 225, 227,
229, 334
Hammond, S., 213
Han, N., 132, 282, 287
Hanson, B.A., 47, 56, 57, 261, 299
Harris, D.J., 23, 30, 57
Hatrak, N., 328
Hedeker, D., 8
Hemphill, F.C., 22
He, X., 245, 246, 250
Holland, P.W., 1–8, 11–13, 16, 17, 22–26, 28,
29, 32–34, 36–38, 41, 42, 45, 48, 60, 62,
71, 89, 92, 94, 98, 112, 125, 141, 142,
146, 147, 159, 160, 175, 176, 185, 186,
194–196, 199, 212, 218, 225, 243, 261,
281–284, 287, 288, 291, 296, 317, 321
Hoover, H.D., 22, 232, 243, 261, 308
Horn, J.L., 86
Hybrid equating functions, 89
I
Indirect equating methods, 328, 329
Information theory, 350
Inverse operator, 92
IRT. See Item response theory
Item characteristic curves (ICC), 67, 244–248,
250, 253, 254, 257
Item response theory (IRT), 2, 43, 59, 160, 201,
225, 243, 261, 282, 297, 327
J
Jacobian, 14, 173, 235, 237, 321
Jaffa, A.S., 72
Janssen, R., 220
Jarjoura, D., 37
Jenkins, G.M., 330, 331
Johnson, E.G., 261
K
Kane, M.T., 187
Kao, C.W., 328
Karabatsos, G., 8, 116, 175, 176, 180
Kaskowitz, G.S., 250
Kendall, M.G., 14
Kernel equating, 11, 13, 90, 125, 132, 138–140,
159–173, 185, 193–199, 261, 326
Kernel equating framework, 164, 165, 317–326
Kernel equating method, 16, 318–320, 323,
325
Kernel smoothing, 37, 139, 160, 164, 244–247,
250, 319
Kim, S., 40, 42, 89, 109, 112, 114, 115, 118,
119, 328
King, B.F., 146
Koenker, R., 246
Kolen, M.J., 2, 6, 16, 22–24, 26, 27, 32, 33, 35,
37, 41, 43–48, 51, 54, 57, 58, 60, 62,
65–68, 110, 111, 118, 143, 149, 152,
175, 183, 186, 187, 190, 199, 217, 221,
222, 225, 226, 229, 232, 233, 236, 243,
261, 281, 282, 286, 296, 297, 299, 304,
308, 317, 328, 330, 343
Kong, N., 187
Koretz, D., 22
Index 363
Kronecker delta, 272
Kuehl, R.O., 326
Kupperman, M., 163
Kurtosis, 148, 149, 151, 161, 168, 197, 262,
270, 271, 277, 323
L
Lagrange multiplier test, 226, 241
Laud, P., 176
Leary, L.F., 30
Lee, W., 194, 199
Lee, W.-C., 47, 58
Lee, Y.-H., 159
Lee, Y.-S., 244
Lehmann, E.L., 321
Levine equating, 14, 185, 187–192, 194, 195,
197–199, 221
Levine observed-score equipercentile equating,
195, 196
Levine, R., 185
Lewis, C., 117–119
Lewis, D., 241
Liang, L., 35, 36
Li, D., 327
Lieberman, M., 262, 268
Lijoi, A., 179
Li, M.-Y., 213
Lindquist, E.F.,
Linear equating, 11, 13, 15, 16, 22, 31, 33, 37,
89, 91, 93–98, 101, 103–104, 111, 156,
176, 185–196, 221, 222, 324–326
Linear interpolation, 13, 16, 17, 37, 161, 165,
168, 286
Linking, 1, 21, 59, 71, 89, 139, 185, 225, 243,
261, 283, 297, 333
Linking procedures, 21, 60, 66, 225, 226, 256
Liou, M., 213, 244, 261
Li, S., 4, 42, 327
Liu, J., 25, 41, 213
Livingston, S.A., 8, 22, 40, 109, 110, 112, 114,
115, 117–119, 185, 282, 328
Ljung-Box test, 332, 339–344
Ljung, G., 332, 339
Local equating, 10, 201–223
Local observed-score equating, 201–223
Logistic distribution, 162
Logistic kernel (LK), 161–163, 166, 168–173
Log-linear models, 4, 12, 15, 37, 132, 138,
146–149, 151, 160, 193, 227, 284, 319,
321–323
Log-linear smoothing, 125, 141–143, 148,
152, 153
Longitudinal data, 74, 82, 85, 87
Longitudinal models, 74, 81–82
Loomis, S.C., 47
Lord, F.M., 2, 6, 7, 24, 28, 51–53, 55, 56, 58,
63, 65, 66, 71, 87, 201–209, 212, 214,
216, 223, 225, 228–230, 234, 235, 243,
253, 261, 300, 305, 334
Lord’s theorem, 2, 201, 205–206, 208–210, 215
Lorentz, G., 178
Loyd, B.H., 232, 243, 261
Lp-circles, 96–98, 104
Luecht, R.M., 44
M
Magis, D., 220
Marco, G.L., 41, 226, 233, 243, 261
Marginal maximum likelihood (MML),
226, 230, 235, 264, 269, 299–302,
313, 316
Martin, K.,
Mason, R., 86
Masters, G.N., 81
McArdle, J.J., 71, 77, 86
McCall, W.A., 46
McHale, F., 119, 120
mdltm software, 230, 237, 238
Mean-preserving linear transformation
(MPLT), 194–196, 199, 200
Measurement model, 2–10, 17, 79–81
Meijer, R.R., 329
Meredith, W., 77
Merrill, M.A., 72, 76, 77
Mislevy, R.J., 254, 261, 301
MML. See Marginal maximum likelihood
Model fit, 69–70, 85
Modified post-stratification equating, 186, 193,
194, 196, 281–296
Molenaar, I.W., 312
Moment-generating function, 162
Moments, 13, 14, 16, 17, 34, 125, 137, 138,
142, 143, 148–151, 161, 164, 166–168,
221, 261, 266, 284, 288, 293–294
Montgomery, D.C., 332, 339
Morris, C.N., 6
Morrison, C.A., 329
Moses, T.P., 21–42, 120
Mosteller, F., 287
Moustaki, I., 261
MPLT. See Mean-preserving linear
transformation
Muliere, P., 179
M€uller, P., 176Multiple-group, 68, 229, 238, 300
Muraki, E., 53, 63, 64
364 Index
N
Nadaraya, E.A., 245
NAEP. See National Assessment of
Educational Progress
National Assessment of Educational Progress
(NAEP)
NEAT design. See Nonequivalent groups withanchor test design
Nelson, L., 51, 52
Ng, P., 246
Nicewander, W.A., 56
Nonequivalent groups design, 183–184, 338,
342, 344
Nonequivalent groups with anchor test design
(NEAT), 5, 12, 14, 16, 27–28, 34–35,
39, 40, 89, 143, 145, 152, 154–156,
185–200, 218–221, 225–229, 243,
281–284, 291, 296, 319, 326
Nonparametric IRT models, 243–258
Normal distribution, 10, 13, 46, 160, 162, 176,
193, 262, 301, 305, 316, 334
Novick, M.R., 6
O
Observed-score equating (OSE), 2, 5, 21, 30,
32, 34, 35, 159, 176, 177, 191, 196–199,
201–223, 281, 283, 286, 297–316
Observed-score equating (OSE) framework,
10–17
Ogasawara, H., 261–280
One-parameter logistic (1PL) model, 64, 79,
226, 297, 300, 304–310, 312, 316
Orcutt, G.H., 333
Ord, K., 273, 276
Orlando, M., 53
OSE. See Observed-score equating
P
2-Parameter logistic (2PL) model, 49, 64, 226,
229, 231, 237, 238, 248, 250, 262, 268,
271, 334
Patz, R., 23, 62, 241, 242
PDF. See Probability density function
Peck, E.A., 332
Penalty function, 148, 149, 156, 165, 166, 247,
323
Petersen, N.S., 22, 24, 26, 37, 41, 226, 308,
328, 329
Petrone, S., 179, 180
Pitoniak, M.J., 48
1PL model. See One-parameter logistic (1PL)
model
2PL model. See Two-parameter logistic (2PL)
model
3PL model. See Three-parameter logistic (3PL)
model
Point-wise average, 92–95, 97, 98, 101, 104
Pommerich, M., 22, 23, 51, 56, 71
Population invariance, 4, 24, 33–35, 41, 188,
189, 191, 203–205, 209, 281, 309
Portnoy, S., 246
Posterior distribution, 51, 176–182, 184, 216
Poststratification equating (PSE), 14, 34, 35,
145, 193–200, 220, 281–296
Presmoothing, 12, 15, 37, 141, 142, 146, 160,
166, 173, 193, 194, 199, 200, 221, 284,
285, 321
Prior distribution, 176–182, 184, 241
Probability density function (PDF), 142–145,
149, 150, 153, 161–165, 169, 179
Pr€unster, I., 179PSE. See Poststratification equating
Q
Quadrature, 142, 143, 230, 235, 264, 265,
269, 305
Quintana, F., 176
Qu, Y., 38, 317
R
Ramsay, J.O., 244, 245, 258
Randomly equivalent groups, 61, 68, 69,
126–133
Rao, C.R., 15
Rasch, G., 79, 300
Rasch model, 49, 79, 229, 300
Reinsel, G.C., 330
Reise, S.P., 22
Reliability, 21, 39
Resampling studies, 114–116, 118–120
Rescorla, L.A., 74
Revised Stanford-Binet, 77
Ricker, K., 282
Rijmen, F., 8, 9, 38, 317
Rock, D.A., 10
Rodriguez, M.C., 56
Rogers, H.J., 225
Rosa, K., 51, 52
Rosenbaum, P.R., 28
Rossi, N., 244
S
Same construct, 17, 21, 23, 24, 27, 41,
72, 159, 176, 183, 225
Samejima, F., 53, 64
San Martin, E., 220
SAT, 25, 31, 35, 42
Index 365
Scale aligning, 22, 23, 25
Scale drift, 4, 70, 327–346
Scaling, 21, 23, 25, 28, 42–64, 68–70, 73,
84, 228, 297, 327, 328
Scaling test, 60
Schwarz, G., 239
Score equity assessment, 41
Score probability, 11, 12, 141, 146, 161,
166, 167, 285–287, 319, 320
Second order model, 8, 9
SEE. See Standard error of equating
SEED. See Standard error of equating
difference
Senturk, D., 47
Sethuraman, J., 177
Shealy, R., 254
Shi, M., 329
Shi, P., 245, 246, 250
Shumway, R.H., 330, 345
SiGNET design, 120, 121
Silvey, S.D., 226
Single-group design, 4, 12, 26–29, 137,
139, 143, 144, 175, 181, 221
Sinharay, S., 5, 35, 41, 281
Skaggs, G., 118
Skewness, 37, 40, 111, 139, 148–151, 161,
167, 168, 197, 262, 269–271, 275
Small samples, 24, 26, 40, 89, 109–122,
220, 221, 300, 315
Smith, A., 176
Smoothing, 14, 35, 37, 40, 125, 139,
141–143, 148, 152, 159, 160, 164, 221,
223, 244–247, 249, 250, 254, 308,
319, 320
Smoothness penalty, 166
Standard error of equating (SEE), 4, 14, 15, 28,
38, 113, 141, 146–147, 152, 157, 160,
161, 171, 172, 199, 281, 320–322
Standard error of equating difference (SEED),
4, 14, 38, 160, 161, 172
Stanford-Binet, 72, 73, 77, 78, 83
Stecher, B.M.,
Stegun, I.A., 129, 163, 305
Steinberg, L., 62, 64
Stewart, E.E., 41, 226
Stocking and Lord method, 66–68, 226, 228,
230, 235, 236, 243 See also Stocking
and Lord scale linkage; Stocking–Lord
test characteristic curve method
Stocking and Lord scale linkage, 234–236
Stocking-Lord test characteristic curves (TCC)
method, 66–68, 235, 253
Stocking, M.L., 66, 229, 234–236, 243, 253,
261, 328, 334
Stoffer, D.S., 330
Stout, W.F., 254
Strawderman, W.E., 89
Stuart, A., 14, 273, 276
Swaminathan, H., 225, 334
Swanson, M.R., 326
Swave, 99–104
Swygert, K.A., 51, 52
Symmetric average, 98, 100–103
Symmetry property, 92–95, 101, 102, 209
Synthetic population. See Target population
T
Target population, 11, 14, 32–35, 40, 110, 111,
114, 141, 144, 145, 155, 161, 218, 219,
249, 283, 319
Taylor expansion, 266
Terman, L.M., 72, 76, 77
Test equating, 1–17, 21–42, 45, 87, 109, 116,
141, 143, 159, 160, 164, 172, 175–184,
186, 190, 223, 317–319
Testlet model, 2, 8–10
Testlets, 2, 8–10, 119, 120
Thayer, D.T., 4, 12, 13, 17, 22, 37, 71, 89, 112,
125, 141, 142, 146, 147, 159, 160, 175,
176, 185, 218, 225, 243, 261, 281, 284,
287, 288, 317, 321
Thissen, D., 48, 51–53, 56, 57, 62, 64, 225,
230, 312
Thisted, R.A., 143
Three-parameter logistic (3PL) model, 49, 52,
64, 217, 222, 226, 228, 229, 231, 236,
237, 243, 254, 263, 267, 268, 297, 300,
304–308, 312–316
Tibshirani, R.J., 310
Time series, 4, 14, 327–346
Tong, Y., 6, 43
True-score equating, 10, 192, 197, 201, 206,
207, 212, 222, 331, 337, 341
Tsai, T.-H., 261
Tucker equating, 185, 187–190, 192, 194–199
Tukey, J.W., 38, 47
Two-parameter logistic (2PL) model, 49, 64,
226, 229, 231, 237, 238, 248, 250, 262,
268, 271, 334
U
Uniform distribution, 128, 134, 161, 163, 165,
176, 334
Uniform kernel, 13, 141, 142, 160–171
366 Index
V
van der Linden, W.J., 8, 10, 49, 201
Van Krimpen-Stoop, E.M.L.A., 329
Veerkamp, W.J.J., 329
Verhelst, N.D., 303, 312, 313
Vertical linking, 59–70, 241, 242
Vertical scaling. See Vertical linkingVining, G.G., 332
von Davier, A.A., 1, 22, 26, 27, 29, 33, 34, 37,
38, 41, 42, 71, 89, 90, 112, 119, 125,
132, 137–141, 143, 146, 148, 152, 156,
159, 175, 176, 181, 182, 185, 187, 196,
197, 218, 221, 225, 243, 261, 281, 299,
317, 327
von Davier, M., 2, 225, 299
W
Wainer, H., 8, 36, 48, 56, 225
WAIS. See Wechsler Adult Intelligence Scale
WAIS-R. See Wechsler Adult Intelligence
Scale-Revised
Wald, A., 313, 317
Wald test, 312–315, 317, 318, 322–323,
325–326
Walker, S.G., 8, 175
Wang, J., 47
Wang, T., 57, 58, 127, 132, 137, 141, 194, 196,
199, 282
Wang, X., 8, 244
Wasserman, L., 160
Watson, G.S., 245
Wechsler Adult Intelligence Scale (WAIS), 72,
73, 80
Wechsler Adult Intelligence Scale-Revised
(WAIS-R), 72, 73, 80
Wechsler-Bellevue Intelligence Scale, 72, 77
Wechsler, D., 72, 77
Weights, 45, 66, 87, 89, 90, 95, 97, 101, 117,
119, 144, 188, 189, 191, 192, 195–198,
212, 213, 265
Wiberg, M., 211, 213, 219, 221, 222
Williams, V.S.L., 51
Wingersky, M.S., 51, 216, 305
Wirth, R.J., 74, 86, 262
Woodcock, R.W., 81
Wright, N.K., 213, 282
X
Xu, X., 227, 243
Y
Yamamoto, K., 227
Yan, D., 132
Yao, L., 23, 62, 241
Yen, W.M., 2, 7, 23, 49, 53, 62–66
Ylvisaker, D., 176, 179
Youtz, C., 287
Z
Zelenak, C.A., 48
Zeng, L., 57, 217, 297
Zimowski, M.F., 56, 254, 301
Zwick, R., 47
Index 367