-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Statistical analyses #16
Comments
@agahkarakuzu until @matteomancini accepts the invitation to the repo, would you mind writing down some of your thoughts (and those that were discussed at our initial meeting)? Just everything that you remember or comes to mind would be fine. |
I think that the structure of the data (T1 values estimated across sites and scanners, with several additional details available) would be well suited from a mixed/fixed (depending on the hypothesis) effects linear model. |
Summary of what we have so far:NIST
Human
Interesting datasets combinations, but maybe not enough time to analyse yet
|
Language to useVery likey R + RShiny for visualisations |
Statistical analyses proposals
To do
|
@agahkarakuzu an idea for you to maybe replace Figure 7 with, or to supplement Figure 7. This is something that Juan didn't feel comfortable doing (nor I), but since you have more stats experience it could be easy to do with the pre-saved pandas ROI & config details databases I have in this repo. |
Beyond the comfort in implementing this, I think the main issue is whether we need this or what it adds to our analysis. Ground truth T1 is a direct determinator of the T1 measured using the gold standard IR (strongly autocorrelated), so they should not be on the opposite sides of a linear model. |
So, when considering this question, I think there are 4 main things to consider that I'm aware of currently:
Phantom version could be another (hypothesizing that there might actually be difference between phantoms; the other two studies mentioned above used a single phantom). Maybe also "submitter"/"implementor" of the protocol (since some phantoms were shared between submissions. One thing I didn't collect in the JSON but may be present in the DICOMS was pre-scan settings, which as you know, would likely be a signifiant factor if the wrong settings are used. |
This issue originated from a meeting between me, @agahkarakuzu, @matteomancini, and @stikov, which we're moving here so that anyone can get involved with the discussion.
The idea is to identify if the data collected by the challenge may be open for some potentially interesting statistical analysis, and to discuss how to best implement these (and in particular, using open-source tools). Also, we could also maybe identify some statistical analyses that would be interesting but that we don't have sufficient data for, and leave that for an open challenge for people to collect more data for.
I think we should start by describing the datasets we have at hand and some of the remaining corrections or post-processing steps that should be done, and then explore some statistical analysis ideas that are well suited for this dataset and doesn't overlap with other similar studies (such as Banes et al. 2017 that used the NIST phantom on multisites but with much stricter protocol implementation rules that our current challenge, which was to investigate the differences or robustness against cross-site implementations). We also have some human datasets to compare with, which could also be explored (human<->human and/or NIST<->human).
The text was updated successfully, but these errors were encountered: