On Wed, 06 Oct 2010, MS Al-Rawi wrote: > First, I have a question about the data > in [1]http://data.pymvpa.org/datasets/haxby2001/ > Was it preprocessed for; motion correction, slice-timing, normalized > to MNI, or smoothed? or any other preprocessing?
AFAIK nothing was done to it that subj1 was taken for the tutorial_data and there it was slightly preprocessed: http://data.pymvpa.org/datasets/tutorial_data/ > The max-responsive masks it contains were obtained per subject via GLM > contrast based localizer maps, this goes for mask4_vt.nii too. Why > there isn't one mask (per category or vt) for all subjects via group > analysis in order to use it in MVPA analysis? because no one created such I guess ;-) and each subject VT is different (if you want to be more or less precise) > Also, is it possible to > obtain such masks using MVPA via searchlight or any other alternative? sure ;) but, be warned that to avoid the independence and circularity in the data analysis, if you are to derive functional localizers masks (e.g. FFA etc) you better use a separate dataset (run), or do it along with your MVPA analysis within cross-validation folds (thus possibly having different masks across folds) -- .-. =------------------------------ /v\ ----------------------------= Keep in touch // \\ (yoh@|www.)onerussian.com Yaroslav Halchenko /( )\ ICQ#: 60653192 Linux User ^^-^^ [175555] _______________________________________________ Pkg-ExpPsy-PyMVPA mailing list [email protected] http://lists.alioth.debian.org/mailman/listinfo/pkg-exppsy-pymvpa

