Last updated: 2019-11-21

Checks: 2 0

Knit directory: mcfa-fit/

This reproducible R Markdown analysis was created with workflowr (version 1.4.0). The Checks tab describes the reproducibility checks that were applied when the results were created. The Past versions tab lists the development history.


Great! Since the R Markdown file has been committed to the Git repository, you know the exact version of the code that produced these results.

Great! You are using Git for version control. Tracking code development and connecting the code version to the results is critical for reproducibility. The version displayed above was the version of the Git repository at the time these results were generated.

Note that you need to be careful to ensure that all relevant files for the analysis have been committed to Git prior to generating the results (you can use wflow_publish or wflow_git_commit). workflowr only checks the R Markdown file, but you know if there are other scripts or data files that it depends on. Below is the status of the Git repository when the results were generated:


Ignored files:
    Ignored:    .RData
    Ignored:    .RDataTmp
    Ignored:    .Rhistory
    Ignored:    .Rproj.user/

Note that any generated files, e.g. HTML, png, CSS, etc., are not included in this status report because it is ok for generated content to have uncommitted changes.


These are the previous versions of the R Markdown and HTML files. If you’ve configured a remote Git repository (see ?wflow_git_remote), click on the hyperlinks in the table below to view them.

File Version Author Date Message
Rmd 332a2d6 noah-padgett 2019-11-22 updated about page
html 332a2d6 noah-padgett 2019-11-22 updated about page
html b534b90 noah-padgett 2019-09-29 updated publish
Rmd 982c8f1 noah-padgett 2019-05-18 roc analyses completed
html 982c8f1 noah-padgett 2019-05-18 roc analyses completed
html 451ccf7 noah-padgett 2019-05-07 Build site.
Rmd 8b2f44d noah-padgett 2019-05-07 intial commit
Rmd 3ffee99 noah-padgett 2019-05-07 Start workflowr project.

This study is a continuation from R. Noah Padgett’s Master’s Thesis. The project is generally about the performance of commonly used fit statistics in multilevel CFA models. The work is part of a large simulation study that will be reported in multiple manuscripts throughout the next year (at least that is the hope). This study is related solely to how the fit indices varied across conditions and estimation methods.

Abstract

Within a multilevel confirmatory factor analysis framework, we investigated the ability of commonly used fit indices to discriminate between correctly specified models and misspecified models. Receiver operating characteristics (ROC) analyses were used to evaluate the performance of fit indices. Combining ROC analyses with checks of the convergence rates across Monte Carlo replications and ANOVA for investigating the variation in fit scores across replications, we found converging evidence for the utility of the investigated fit indices.
Optimal thresholds based on maximizing sensitivity and specificity for detection of the true model were identified by the highest sensitivity and specificity and found to vary across different robust estimation methods (i.e., MLR, ULSMV, and WLSMV). The estimation method and sample size influenced the performance of common fit indices to detect misspecification of the level-1 model. All fit indices investigated performed poorly for detecting misspecification of the level-2 model when the level-2 sample size was below 100. We offer recommendations of commonly reported fit indices to use (and not use), cut-off criteria to use for specific estimation methods, and cautions about the use of recommended cut-off criteria for ML-CFA.