Preregistering Structural Equation Modeling (SEM) Studies: A Comprehensive Guide
Structural Equation Modeling (SEM) is a powerful analytical tool, capable of modeling complex latent structures and causal relationships between variables. From psychology to marketing, SEM is used in diverse fields to test theoretical models with observed data. Yet, the same flexibility that makes SEM attractive also opens the door to excessive researcher degrees of freedom. Without constraints, analysts can tweak specifications post hoc–knowingly or unknowingly–to produce more favorable results.
Preregistration addresses this issue by setting in stone the analysis plan before seeing the data. In the context of frequentist statistics, this step is crucial: data-contingent modeling decisions can invalidate p-values and confidence intervals by increasing the risk of false positives. By formally documenting decisions in advance, researchers commit to a confirmatory route, while still retaining space for transparent exploratory analysis.
This article outlines the major decision domains to address in a preregistered SEM study. Each domain includes actionable recommendations to make your research more reproducible and credible.
1. Locking the Software Environment
A reproducible analysis begins with a stable computational environment. SEM models are sensitive not only to the software used but also to subtle changes across versions of packages, operating systems, and numerical libraries.
Specify the software and exact version you will use–such as lavaan 0.6-18
in R 4.4.1
, Mplus 8.10
, or semopy 2.x
in Python. Don’t stop at the modeling package; include the operating system, any hardware dependencies, and versions of supporting math libraries (e.g., BLAS, OpenBLAS, MKL) that may affect floating-point operations.
Use tools like renv::snapshot()
in R, requirements.txt
in Python, or conda
environments to freeze dependencies. For maximal reproducibility, build a Docker container and share the Dockerfile
or image.
Also include the exact script or notebook you intend to run during analysis. This provides a literal representation of your analysis plan–helpful not only for others but also for future-you.
2. Defining the Scientific and Structural Model
Before data ever enters the picture, you need a clear theoretical model. This involves both conceptual and graphical representations of expected relationships.
Develop a complete path diagram showing hypothesized relationships between latent constructs and observed variables. Each latent variable should have a definition rooted in prior literature, and every item should be justified in terms of what it captures. This process helps clarify construct validity and prevents arbitrary inclusion of items during analysis.
Declare whether your model is directional (i.e., causal paths), and specify which variables are exogenous. For studies involving group comparisons or longitudinal designs, indicate plans for assessing measurement invariance.
To preserve confirmatory integrity, clearly state that no additional paths will be added unless in a predefined exploratory phase. If modifications are planned, make them conditional on specified thresholds or theoretical rationales.
3. Operationalizing the Statistical Model
The theoretical structure must be translated into a formal statistical model. This includes selecting the appropriate SEM framework, specifying assumptions, and handling practical modeling choices.
Indicate the model type: Confirmatory Factor Analysis (CFA), full SEM, latent growth curve models, MIMIC, multilevel SEM, or network SEM. Each requires different identification strategies and introduces different assumptions.
Define your assumptions about variable distributions. For instance, will ordinal items be treated as continuous, or will you use polychoric correlations? Will you allow for non-normality or heavy-tailed distributions?
Declare how residuals are treated–are any error covariances theory-justified? Describe the strategy for missing data, whether Full Information Maximum Likelihood (FIML), multiple imputation, or listwise deletion.
Also fix your identification strategy: marker-variable (loading fixed to 1) or unit-variance scaling (latent variance fixed to 1). Changes to these decisions post hoc can affect parameter estimates, so preregistration helps avoid retrofitting models to the data.
4. Estimation Methods and Robustness Considerations
Choosing an estimator isn’t just a technical detail–it affects parameter accuracy, standard errors, and fit indices. Preregister your primary estimation method, such as Maximum Likelihood (ML), Robust ML (MLR), Diagonally Weighted Least Squares (DWLS/WLSMV), Unweighted Least Squares (ULS), or Bayesian estimation.
If you anticipate potential violations of assumptions, specify robustness corrections ahead of time. For instance, include Satorra–Bentler scaled chi-square if using MLR. Document the maximum number of iterations, convergence thresholds, and behavior in case of non-convergence.
For studies involving bootstrapping, state how many samples will be used (e.g., 5,000 BCa resamples), which statistics will be bootstrapped, and how the results will be interpreted.
Robustness checks should be planned–not reactive. They belong in a separate sensitivity analysis tier rather than as an opportunistic fix after primary analyses fail.
5. Measurement Model Decisions and Fit Criteria
A major temptation in SEM is to “tune” the model post hoc to improve fit. Preregistration prevents this by locking in the criteria by which model fit will be judged.
Declare your primary fit indices: CFI, TLI, RMSEA, and SRMR are common. Specify thresholds for both good and acceptable fit (e.g., CFI > 0.95 for good, > 0.90 for acceptable).
State whether any correlated errors or cross-loadings are allowed based on theory. Describe how (or whether) modification indices (MI) will be consulted. If MIs are to be used, define a strict rule–for example, MI > 10 and theoretical justification must both be met.
A best practice is to employ a two-tiered strategy: analyze the confirmatory model as preregistered, and if fit is poor, then conduct a clearly labeled exploratory refinement. Keep the confirmatory and exploratory results separate in interpretation and reporting.
6. Predefined Backup Plans and Contingency Responses
Even well-specified models can fail. Convergence issues, inadmissible solutions, or severe model misfit are not uncommon in SEM. Rather than improvising fixes, define contingency plans in advance to preserve the integrity of your confirmatory claims.
Start by specifying a tiered approach to non-convergence. This might involve modifying starting values, switching optimization algorithms (e.g., from nlminb
to BFGS
in lavaan
), or simplifying the model by removing problematic latent variables or paths.
Plan for the appearance of Heywood cases, such as negative variance estimates or standardized loadings exceeding 1. You might specify that if a Heywood case appears and is minor (e.g., loading = 1.01), you will retain the solution; but if it exceeds a certain threshold, a predefined reduced model will be estimated instead.
If model fit is poor based on your preregistered criteria, define whether and how you will proceed. For example, you could specify that if RMSEA > 0.10, you will estimate a simplified version of the model that excludes certain weakly identified paths, provided the revision is consistent with theoretical expectations.
Contingencies can also include assumption violations, such as non-normality or outliers. If these are detected using predefined diagnostics (e.g., Mardia’s test, Q-Q plots), you may move to robust estimation or data transformation–again, only if such actions were laid out in the preregistration.
Use a decision table mapping common problems to specific, predefined responses. This reduces the need for subjective choices once the data are visible.
7. Multiverse and Sensitivity Analyses
To demonstrate that your results are not fragile, preregister a multiverse analysis–a systematic variation of defensible analytical decisions. This goes beyond robustness checks by explicitly modeling the uncertainty introduced by researcher degrees of freedom.
List all plausible alternatives in areas such as:
- Missing data strategy: FIML vs. multiple imputation
- Treatment of ordinal items: as continuous vs. polychoric-based CFA
- Scaling method: marker variable vs. unit variance
- Grouping strategy: multigroup vs. covariate modeling
- Outlier handling: Winsorizing, robust Mahalanobis distance exclusion
Create a plan to fit all combinations of these decisions and extract a key parameter of interest (e.g., path coefficient from A → B). Then visualize its distribution across specifications via a specification curve.
Include robustness checks such as leave-one-indicator-out analysis, where the measurement model is re-estimated repeatedly while omitting one indicator at a time to test for over-reliance on specific items.
The goal here is not to eliminate all variation, but to demonstrate that your conclusions hold across reasonable decision spaces. Automate this process before analyzing real data using scripts or workflows that can be rerun unchanged.
8. Additional Preregistration Components
A high-quality preregistration does more than lock analytic decisions–it anticipates all aspects of confirmatory research.
Specify your sample size and the method used to determine it. This may involve a Monte Carlo simulation to assess statistical power for detecting your hypothesized effects under specific assumptions about model structure and measurement quality.
State your primary outcomes and hypotheses clearly. A good rule of thumb is one primary effect per hypothesis. Secondary effects should be labeled exploratory unless they are also preregistered with equal rigor.
If testing multiple effects, plan for multiplicity correction. This could involve controlling the false discovery rate (FDR), using Bonferroni or Holm corrections, or adopting a Bayesian approach with posterior probabilities.
Define your reporting plan, including how confirmatory results will be separated from exploratory ones. Make clear which figures, tables, and model variations will be included in the final paper.
Finally, consider uploading the preregistration to a public registry, such as AsPredicted, OSF Registries, or a journal-specific format if submitting under a Registered Reports model.
9. Final Thoughts on Transparency and Rigor
Preregistration does not limit scientific creativity–it clarifies it. By defining your confirmatory analysis plan in advance, you create a clean boundary between hypothesis testing and hypothesis generation. Readers can trust that the results you present as confirmatory were not achieved through post hoc modifications.
A robust SEM preregistration spans more than just model syntax. It includes your computational setup, theoretical justifications, modeling assumptions, contingency plans, and robustness checks. It acknowledges the complexity of SEM and uses structure to prevent that complexity from becoming a liability.
Think of your preregistration as a recipe. If another researcher followed it precisely, without speaking to you, they should arrive at the same confirmatory results. When this happens, science advances–not just with findings, but with trust.