Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ensure it is trivial to do GSEM with umxRAM #146

Open
tbates opened this issue Dec 23, 2020 · 1 comment
Open

Ensure it is trivial to do GSEM with umxRAM #146

tbates opened this issue Dec 23, 2020 · 1 comment
Projects

Comments

@tbates
Copy link
Owner

tbates commented Dec 23, 2020

Consider wrappers for the GWAS pipeline also.
https://github.com/MichelNivard/GenomicSEM/wiki

@mcneale
Copy link
Collaborator

mcneale commented Dec 23, 2020

Yes, there are some weirdnesses with Lavaan or OpenMx WLS(MV) implementations that make them disagree. Others have reported inconsistencies between Lavaan and MPlus. At least we can triangulate a bit - and with two of them being open source we can actually figure out WHY they differ if they do.

For more complicated modeling, OpenMx will be faster than Lavaan due to using C++ instead of R for computation & optimization. However, for WLS, where the weight matrix has to be inverted only once, most of the heavy lifting has already been done, with WLS model-fitting proceeding very rapidly.

@tbates tbates added this to To do in new models via automation Dec 25, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
new models
  
To do
Development

No branches or pull requests

2 participants