# Why Fuss about Statistical Sufficiency?

Statistical sufficiency is an obscure property of the Rasch model. Although Georg Rasch does not use the term, he writes: "The best estimate of the ability parameter for a person can be derived from his raw score only" (Rasch, 1980, p.76). For Rasch, this is equivalent to the statement that a raw score is a sufficient statistic for an ability measure.

Ronald Fisher (1922) writes of a sufficient statistic "that the statistic chosen should summarize the whole of the relevant information supplied by the sample." The Fisher-Neyman theorem asserts that "T" is a sufficient statistic for the unknown measure underlying the data if, and only if, the probabilities associated with the data can be factored into two parts. One part must be dependent on the outcome of the measure through the sufficient statistic, "T", only, and the other part must be independent of the unknown measure (Halmos & Savage, p.226).

Halmos & Savage (p.240) provide an illustration of sufficiency that can be exported to the field of educational testing. Suppose a test of 20 items of known difficulty conforms to the Rasch model. Scoring these items is arduous, so the examination board decides to make pass- fail decisions for each examinee based on that examinee's success on a single item. To remove the possibility of bias in item selection, the board decides to select the item at random for each examinee. The board awards "pass", if the examinee succeeds on that item.

Next time, the board installs a scoring machine which reports the raw score for each examinee on the entire test. But the board wants to maintain the same pass-fail procedure as last time. So the board selects for each examinee a test item at random. Then, using the value of the raw score and the known difficulty of the selected item, the probability of success of the examinee on that item is estimated. This probability is compared with a random probability in order to assign the examinee a success or failure on that item. Now the pass- fail decision is made on that simulated outcome just as it had been previously with an observed response.

In the long run, which pass-fail method will be more accurate? The answer is that they will be equally accurate. This is because the raw score is a sufficient statistic. Statistical sufficiency implies that the examining board is just as well off knowing the value of the sufficient statistic as it is knowing the observations that comprise it. The extra details provided by knowing the actual responses do not provide the examining board any further useful guidance as to the size of the measure.

Statistical sufficiency is the same as the requirement that person measures be "sample-free." Among all relevant items of the same known difficulty, it must make no difference which one happened to be included in the test.

But what if there is differential item functioning? What if it matters which items of a certain difficulty are included in the test? Then the raw score is no longer a sufficient statistic. More information about the items must be provided. Perhaps the data can be decomposed into subsets, each with a sufficient statistic, or perhaps the test only admits of a qualitative description of each of the responses observed.

Sufficiency is an idea based on a probability model for the data. Idea sufficiency is never met in practice. Rather, the pattern of the observed data must be compared with that which would be expected were sufficiency to exist. This comparison forms the basis of the decision as to whether the data approximates the idea closely enough for measurement.

What is the relationship between the sufficient statistic and the underlying measure? A sufficient statistic does not provide an exact value for the underlying measure, rather the statistic summarizes all that is known on which to base an estimate of that measure. Whether that estimate is statistically unbiased or consistent or "best" (minimum variance) are matters quite apart from statistical sufficiency.

Achievement of statistical sufficiency is a theoretical idea. This idea corresponds to the practical intention that estimates of measures be as free as possible of the context from which they were obtained. To belittle sufficiency is to reject the goal of liberating measures from the local particulars of the measuring instrument and environment.

Halmos PR & Savage LJ. 1949. Application of the Radon-Nikodyna theorem to the theory of sufficient statistics. Annals of Math. Stat. 20, p.225-241.

Why Fuss about Statistical Sufficiency?, J Linacre … Rasch Measurement Transactions, 1992, 6:3 p. 230

Rasch Publications
Rasch Measurement Transactions (free, online) Rasch Measurement research papers (free, online) Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch Applying the Rasch Model 3rd. Ed., Bond & Fox Best Test Design, Wright & Stone
Rating Scale Analysis, Wright & Masters Introduction to Rasch Measurement, E. Smith & R. Smith Introduction to Many-Facet Rasch Measurement, Thomas Eckes Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, George Engelhard, Jr. Statistical Analyses for Language Testers, Rita Green
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar Journal of Applied Measurement Rasch models for measurement, David Andrich Constructing Measures, Mark Wilson Rasch Analysis in the Human Sciences, Boone, Stave, Yale
in Spanish: Análisis de Rasch para todos, Agustín Tristán Mediciones, Posicionamientos y Diagnósticos Competitivos, Juan Ramón Oreja Rodríguez

 Forum Rasch Measurement Forum to discuss any Rasch-related topic

Go to Top of Page
Go to index of all Rasch Measurement Transactions
AERA members: Join the Rasch Measurement SIG and receive the printed version of RMT
Some back issues of RMT are available as bound volumes
Subscribe to Journal of Applied Measurement

Go to Institute for Objective Measurement Home Page. The Rasch Measurement SIG (AERA) thanks the Institute for Objective Measurement for inviting the publication of Rasch Measurement Transactions on the Institute's website, www.rasch.org.

Coming Rasch-related Events
March 31, 2017, Fri. Conference: 11th UK Rasch Day, Warwick, UK, www.rasch.org.uk
April 2-3, 2017, Sun.-Mon. Conference: Validity Evidence for Measurement in Mathematics Education (V-M2Ed), San Antonio, TX, Information
April 26-30, 2017, Wed.-Sun. NCME, San Antonio, TX, www.ncme.org - April 29: Ben Wright book
April 27 - May 1, 2017, Thur.-Mon. AERA, San Antonio, TX, www.aera.net
May 26 - June 23, 2017, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 30 - July 29, 2017, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com
July 31 - Aug. 3, 2017, Mon.-Thurs. Joint IMEKO TC1-TC7-TC13 Symposium 2017: Measurement Science challenges in Natural and Social Sciences, Rio de Janeiro, Brazil, imeko-tc7-rio.org.br
Aug. 7-9, 2017, Mon-Wed. In-person workshop and research coloquium: Effect size of family and school indexes in writing competence using TERCE data (C. Pardo, A. Atorressi, Winsteps), Bariloche Argentina. Carlos Pardo, Universidad Catòlica de Colombia
Aug. 7-9, 2017, Mon-Wed. PROMS 2017: Pacific Rim Objective Measurement Symposium, Sabah, Borneo, Malaysia, proms.promsociety.org/2017/
Aug. 10, 2017, Thurs. In-person Winsteps Training Workshop (M. Linacre, Winsteps), Sydney, Australia. www.winsteps.com/sydneyws.htm
Aug. 11 - Sept. 8, 2017, Fri.-Fri. On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com
Aug. 18-21, 2017, Fri.-Mon. IACAT 2017: International Association for Computerized Adaptive Testing, Niigata, Japan, iacat.org
Sept. 15-16, 2017, Fri.-Sat. IOMC 2017: International Outcome Measurement Conference, Chicago, jampress.org/iomc2017.htm
Oct. 13 - Nov. 10, 2017, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
Jan. 5 - Feb. 2, 2018, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
Jan. 10-16, 2018, Wed.-Tues. In-person workshop: Advanced Course in Rasch Measurement Theory and the application of RUMM2030, Perth, Australia (D. Andrich), Announcement
Jan. 17-19, 2018, Wed.-Fri. Rasch Conference: Seventh International Conference on Probabilistic Models for Measurement, Matilda Bay Club, Perth, Australia, Website
May 25 - June 22, 2018, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 29 - July 27, 2018, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com
Aug. 10 - Sept. 7, 2018, Fri.-Fri. On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com
Oct. 12 - Nov. 9, 2018, Fri.-Fri. On-line workshop: Practical Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com