ASCB logo LSE Logo

Letter to the EditorFree Access

The Answer is “No”: A Comment on Peugh and Feldon (2020)

    Published Online:https://doi.org/10.1187/cbe.24-07-0182

    To the editor:

    In Peugh and Feldon's (2020) publication, “‘How Well Does Your Structural Equation Model Fit Your Data?’: Is Marcoulides and Yuan's Equivalence Test the Answer?” the authors (we) reviewed longstanding problems with common strategies for assessment model fit using chi-square, comparative fit index (CFI), and root mean square error of approximation (RMSEA), then introduced and demonstrated a prospective alternative solution: equivalence testing as developed by Marcoulides and Yuan (2017). We were cautiously optimistic as to whether equivalence testing (EQT) was a better technique for assessing model fit versus previous strategies but cautioned that further empirical testing was needed. Based on research published since (McNeish, 2023; Peugh et al., 2023; McNeish and Wolf, 2024), our response to the question “Is Marcoulides and Yuan's equivalence test the answer?”, is “No.” While the research field would still benefit from a more reliable test of model fit, robust evidence suggests that the equivalence testing approach described by Peugh and Feldon does not meet that need.

    Decades of extensive research since the publication of Hu and Bentler's (1995, 1998, 1999) landmark research has shown fit index cut-off points (e.g., CFI ≥ 0.96, RMSEA < 0.08) have failed both to supplant chi-square as an inferential model fit test and to be reliable measures of model fit. For example, Hu and Bentler's fit index cut-off points will show acceptable fit for structural equation models based on analysis variables with unacceptable internal consistency reliability, as well as models that poorly reproduce analysis variable relationships known to exist in the sample data (Hancock and Mueller, 2011; McNeish and Wolf, 2024). However, despite these dismal empirical realities, fit index cut-off points are universally alive and well in the minds of research editors, reviewers, and authors because nothing else more quickly answers a model fit question and little (if anything) else is available.

    The goal of our previous publication (Peugh and Feldon, 2020) was to introduce and demonstrate EQT (Marcoulides and Yuan, 2017), which showed initial promise as a replacement for the poor existing mechanisms to assess model fit quality (Marcoulides and Yuan, 2020, 2023). However, subsequent Monte Carlo simulation testing showed EQT fit statistics (CFIt and RMSEAt) produce excessive Type-I (rejecting well-fitting models) and Type-II (accepting poorly fitting models) errors and failed to outperform the chi-square test of model fit in scenarios commonly occurring in applied research (e.g., N ≤ 300, missing data, and skewed data; McNeish, 2023; Peugh et al., 2023; McNeish and Wolf, 2024). Accordingly, we felt it necessary to provide the most complete information available for researchers eager to evaluate the quality of statistical analysis models.

    Researchers wondering, “Well, now what?” should be aware that a new fit assessment technique, direct-discrepancy dynamic fit indices (3DFI; McNeish 2023; McNeish and Wolf, 2024) seems to have avoided the pitfalls of both traditional fit indices and the EQT approach. Briefly, McNeish and colleagues operationalized Millsap's (2007, 2013) conjecture that Hu and Bentler's Monte Carlo simulation procedures could accurately and reliably determine the fit of a single data analysis model. Further, rather than absolute cut-off points, the authors followed Marcoulides and Yuan's (2017) example and created interpretational categories or “bins” within which 3DFI simulation results could be placed and from which a model fit determination could be made.

    3DFI is currently available either as a Shiny application (https://dynamicfit.app; Wolf and McNeish, 2021), or as an R package (“dynamic,” Wolf and McNeish, 2023), but it is not available for multiple-group or multilevel/mixed linear model analyses. Preliminary Monte Carlo testing shows 3DFI has potential as an inferential model fit test. However, how 3DFI performs in simulations across a wide range of SEM types under various sample sizes, data distributions, number of analysis variables, and missing data conditions—as we used when testing (unsuccessfully) the utility of EQT (Peugh et al., 2023)—is currently an open question. As such, we must repeat the cautionary warning we offered previously for EQT: model fit assessment methods should not be adopted until rigorous empirical Monte Carlo testing establishes both the reliability and validity of the procedure.

    REFERENCES