Skip to main content
Book cover

Handbook of Diagnostic Classification Models

Models and Model Extensions, Applications, Software Packages

  • Book
  • © 2019

Overview

  • A major collection that describes the state of the art of diagnostic classification models (DCMs)
  • Provides chapters on the majority of popular DCMs as well as cutting edge model extensions developed by leading experts in the field
  • Covers important research topics such as inferences and learning about the Q-matrix structure, tests for item-level model selection, model identifiability and identifiability conditions
  • Includes chapters on application of diagnostic models in large scale assessments, adaptive testing, and process data analysis
  • Describes specialized software packages such as R as well as the use of general purpose latent modeling software for diagnostic modeling

This is a preview of subscription content, log in via an institution to check access.

Access this book

eBook USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access

Licence this eBook for your library

Institutional subscriptions

Table of contents (31 chapters)

  1. Approaches to Cognitive Diagnosis

Keywords

About this book

This handbook provides an overview of major developments around diagnostic classification models (DCMs) with regard to modeling, estimation, model checking, scoring, and applications. It brings together not only the current state of the art, but also the theoretical background and models developed for diagnostic classification. The handbook also offers applications and special topics and practical guidelines how to plan and conduct research studies with the help of DCMs.

Commonly used models in educational measurement and psychometrics typically assume a single latent trait or at best a small number of latent variables that are aimed at describing individual differences in observed behavior. While this allows simple rankings of test takers along one or a few dimensions, it does not provide a detailed picture of strengths and weaknesses when assessing complex cognitive skills.

DCMs, on the other hand, allow the evaluation of test taker performance relative to a potentially large number of skill domains. Most diagnostic models provide a binary mastery/non-mastery classification for each of the assumed test taker attributes representing these skill domains. Attribute profiles can be used for formative decisions as well as for summative purposes, for example in a multiple cut-off procedure that requires mastery on at least a certain subset of skills.

The number of DCMs discussed in the literature and applied to a variety of assessment data has been increasing over the past decades, and their appeal to researchers and practitioners alike continues to grow. These models have been used in English language assessment, international large scale assessments, and for feedback for practice exams in preparation of college admission testing, just to name a few.

Nowadays, technology-based assessments provide increasingly rich data on a multitude of skills and allow collection of data with respect to multiple types of behaviors. Diagnostic models can be understood as an ideal match for these types of data collections to provide more in-depth information about test taker skills and behavioral tendencies.


Editors and Affiliations

  • National Board of Medical Examiners (NBME), Philadelphia, USA

    Matthias von Davier

  • Teachers College, Columbia University, New York, USA

    Young-Sun Lee

About the editors

Matthias von Davier is Distinguished Research Scientist at the National Board of Medical Examiners (NBME), in Philadelphia, Pennsylvania. Until 2016, he was a senior research director in the Research & Development Division at Educational Testing Service (ETS), and co-director of the center for Global Assessment at ETS, leading psychometric research and operations of the center. He earned his Ph.D. at the University of Kiel, Germany, in 1996, specializing in psychometrics. In the Center for Advanced Assessment at NBME, he works on psychometric methodologies for analyzing data from technology-based high-stakes assessments. He is one of the editors of the Springer journal Large Scale Assessments in Education, which is jointly published by the International Association for the Evaluation of Educational Achievement (IEA) and ETS. He is also editor-in-chief of the British Journal of Mathematical and Statistical Psychology (BJMSP), and co-editor of the Springer book series Methodology of Educational Measurement and Assessment. Dr. von Davier received the 2006 ETS Research Scientist award and the 2012 NCME Brad Hanson Award for contributions to educational measurement. His areas of expertise include topics such as item response theory, latent class analysis, diagnostic classification models, and, more broadly, classification and mixture distribution models, computational statistics, person-fit, item-fit, and model checking, hierarchical extension of models for categorical data analysis, and the analytical methodologies used in large scale educational surveys. 

Dr. Lee is an Associate Professor in the program of Measurement, Statistics & Evaluation, in the Department of Human Development at Teachers College, Columbia University. She received her Ph.D. in Quantitative Methods at the University of Wisconsin-Madison, with a minor in Statistics. Her research interests are primarily on psychometric approaches to solving practical problems in educational and psychological testing. Her areas of expertise include topics such as development and applications of diagnostic classification models, item response theory, latent class models, and analytical methodologies used in large scale assessments. In addition to her own research, Dr. Lee collaborates on various projects on the use of latent variable models for purposes of scale development/test construction and for validity studies. 

Bibliographic Information

Publish with us