Hi all,

I recently asked a question on dealing with unbalanced datasets and here's
a follow-up question.
So let's say I have empty runs, or runs where there are zero samples for
one of the conditions. This leads to problems if that run happens to be the
test run on a leave-one-run-out cross-validation procedure.

My workaround for that was this: if I had one of such runs with empty
conditions, then I would set NFoldPartitioner(cvtype=2), together with
Balancer() so that any combination of two runs would have at least one
sample per condition. But if I had two of such runs with empty conditions,
then I would set cvtype=3, and so on. However this means I have less data
for the training set on each classification fold.

Is there any other possible solution for this? In fact, is it possible to
do leave-n-samples-out classification: So on each fold I randomly select n
samples per condition to test on, and use the remaining samples (after
balancing) for training, disregarding the chunks structure?

Thanks!
-Edmund
_______________________________________________
Pkg-ExpPsy-PyMVPA mailing list
[email protected]
http://lists.alioth.debian.org/cgi-bin/mailman/listinfo/pkg-exppsy-pymvpa

Reply via email to