Hello, 

I'd like to pick the group's brain about an issue related to cross-validation. 

I just read a paper in Current Biology (Chadwick et al., 2010, Decoding 
Individual Episodic Memory Traces in the Hippocampus) in which the following 
cross-validation procedure was used: 

"We used a standard k-fold cross- validation testing regime [10] wherein k 
equalled the number of experimental trials, with the data from each trial set 
aside in turn as the test data, and the remaining data used as the training set 
(on each fold, the feature selection step was performed using only data from 
this training set)."

In other words, it seems to me that on each cross-validation fold, training was 
performed on all trials except one, and then the classifier is tested on one 
single trial.  Does this sort of approach make sense?  It seems to me that with 
only one test for each classifier you are not adequately assessing the 
performance of any of the classifiers.   I suppose the idea is that across all 
of the folds you get a measure of how the classifiers work in general.  I'd 
like to know if this is considered a reasonable approach since I have a dataset 
with a small number of trials that might benefit from maximizing the number of 
training trials. 

Thanks!

Jonas

----
Jonas Kaplan, Ph.D.
Research Assistant Professor
Brain & Creativity Institute
University of Southern California

_______________________________________________
Pkg-ExpPsy-PyMVPA mailing list
[email protected]
http://lists.alioth.debian.org/mailman/listinfo/pkg-exppsy-pymvpa

Reply via email to