At the heart of this you have a problem in incomplete conditioning. You are computing things like Prob(X > x) when you know X=x. Working with a statistician who is well versed in probability models will undoubtedly help.

Frank

Frank E Harrell Jr   Professor and Chairman        School of Medicine
                     Department of Biostatistics   Vanderbilt University

On Thu, 19 Aug 2010, Assa Yeroslaviz wrote:

Hello everybody,

yes I'm sorry. I can see it is not so easy to understand.
I'l try to explain a bit more. The experiment was used to compare two
(protein domain) data bases and find out whether or not the results founded
in one are comparable to the second DB.
the first column shows the list of the various inputs in the DB, the second
lists the various domains for each gene. the p-value column calculates the
probability that the found in column four (Expected) to be found by chance.
in column five the expected values was listed.
The calculation of the TP,TN,FP,FN was made many times, each time with a
different p-value (from p=1,...,p=10E-12) as a threshold to calculate the
various values of TP,TN, etc.

The goal of this calculation was to find the optimal p-value wit ha maximum
of TP and a minimum of FP.

To do so I thought about making the column of p-values my predictions and
the values in the column Is.Expected (TRUE,FALSE) to my labels.
This how I calculated my first ROC curve:
pValue <- read.delim(file = "p=1.txt", as.is= TRUE)
desc1 <- pValue[["p.value"]]
label1 <- pValue[["Is.Expected"]] # after changing the values of TRUE = 0,
FALSE = 1

pred <- prediction(desc1, label1)
perf <- performance(pred, "tpr", "fpr")
plot(perf, colorsize = TRUE)

my question are as follow:
1. Am I right in my way of thinkning, that the p-values here are
predictions?
I know you said I need to decided it for myself, but I'm not sure. If they
are, than I will have the same predictions for each and every calculation of
ROCR. Will it make any difference at the prediction?
2. how can i calculate the other p-values thresholds? Do I need to do each
separately, or is there a way of combining them?

I hope you can still help we with some hints or further advieces.

Thanks

Assa

On Wed, Aug 18, 2010 at 07:55, Claudia Beleites <cbelei...@units.it> wrote:

Dear Assa,

you need to call prediction with continuous predictions and a _binary_ true
class label.

You are the only one who can tell whether the p-values are actually
predictions  and what the class labels are. For the list readers p is just
the name of whatever variable, and you didn't even vaguely say what you try
to classify, nor did you offer any explanation of what the columns are.

The only information we get from your table is that p-value has small and
continuous values. From what I see the p-values could also be fitting errors
of the predictions (e.g. expressed as a probability that the similarity to
the predicted class is random).

Claudia

Assa Yeroslaviz wrote:

Dear Claudia,

thank you for your fast answer.
I add again the table of the data as an example.

Protein ID      Pfam Domain     p-value         Expected        Is
Expected     True Postive False Negative     False Positive  True Negative
NP_000011.2     APH     1.15E-05        APH     TRUE    1       0       0
      0
NP_000011.2     MutS_V  0.0173  APH     FALSE   0       0       1       0
NP_000062.1     CBS     9.40E-08        CBS     TRUE    1       0       0
      0
NP_000066.1     APH     3.83E-06        APH     TRUE    1       0       0
      0
NP_000066.1     CobU    0.009   APH     FALSE   0       0       1       0
NP_000066.1     FeoA    0.3975  APH     FALSE   0       0       1       0
NP_000066.1     Phage_integr_N  0.0219  APH     FALSE   0       0       1
      0
NP_000161.2     Beta_elim_lyase         6.25E-12        Beta_elim_lyase
      TRUE    1       0       0       0
NP_000161.2     Glyco_hydro_6   0.002   Beta_elim_lyase         FALSE   0
      0       1       0
NP_000161.2     SurE    0.0059  Beta_elim_lyase         FALSE   0       0
      1       0
NP_000161.2     SapB_2  0.0547  Beta_elim_lyase         FALSE   0       0
      1       0
NP_000161.2     Runt    0.1034  Beta_elim_lyase         FALSE   0       0
      1       0
NP_000204.3     EGF     0.004666118     EGF     TRUE    1       0       0
      0
NP_000229.1     PAS     3.13E-06        PAS     TRUE    1       0       0
      0
NP_000229.1     zf-CCCH         0.2067  PAS     FALSE   0       1       1
      0
NP_000229.1     E_raikovi_mat   0.0206  PAS     FALSE   0       0       0
      0
NP_000388.2     NAD_binding_1   8.21E-24        NAD_binding_1   TRUE    1
      0       0       0
NP_000388.2     ABM     1.40E-08        NAD_binding_1   FALSE   0       0
      1       0
NP_000483.3     MMR_HSR1        1.98E-05        MMR_HSR1        TRUE    1
      0       0       0
NP_000483.3     DEAD    2.30E-05        MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     APS_kinase      1.80E-09        MMR_HSR1        FALSE   0
      0       1       0
NP_000483.3     CbiA    0.0003  MMR_HSR1        FALSE   0       0       1
      0
NP_000483.3     CoaE    1.28E-07        MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     FMN_red         4.61E-08        MMR_HSR1        FALSE   0
      0       1       0
NP_000483.3     Fn_bind         0.3855  MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     Invas_SpaK      0.2431  MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     PEP-utilizers   0.127   MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     NIR_SIR_ferr    0.1661  MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     AAA     0.0031  MMR_HSR1        FALSE   0       0       1
      0
NP_000483.3     DUF448  0.0021  MMR_HSR1        FALSE   0       0       1
      0
NP_000483.3     CBF_beta        0.1201  MMR_HSR1        FALSE   0       0
      1       0
NP_000483.3     zf-C3HC4        0.0959  MMR_HSR1        FALSE   0       0
      1       0
NP_000560.5     ig      5.69E-39        ig      TRUE    1       0       0
      0
NP_000704.1     Epimerase       4.40E-21        Epimerase       TRUE    1
      0       0       0
NP_000704.1     Lipase_GDSL     6.63E-11        Epimerase       FALSE   0
      0       1       0

...

this is a shorted list from one of the 10 lists I have for different
p-values.

As you can see I have separate p-value experiments and probably need to
calculate for each of them a separate ROC. But I don't know how to calculate
these characteristics for the p-values.
How do I assign the predictions to each of the single p-value experiments?

I would appreciate any help

Thanks
Assa


On Tue, Aug 17, 2010 at 12:55, Claudia Beleites <cbelei...@units.it<mailto:
cbelei...@units.it>> wrote:

   Dear Assa,



       I am having a problem building a ROC curve with my data using
       the ROCR
       package.

       I have 10 lists of proteins such as attached (proteinlist.xls).
       each of the

   your file didn't make it to the list.



       lists was calculated with a different p-value.
       The goal is to find the optimal p-value for the highest number
       of true
       positives as well as lowaest number of false positives.


       As far as I understood the explanations from the vignette of
       ROCR, my data
       of TP and FP are the labels of the prediction function. But I
       don't know how
       to assign the right predictions to these labels.


   I assume the p-values are different cutoffs that you use for
   "hardening" (= making yes/no predictions) from some soft (=
   continuous class membership) output of your classifier.

   Usually, ROCR calculates the curves as function of the
   cutoff/threshold itself from the continuos predictions. If you have
   these soft predictions, let ROCR do the calculation for you.

   If you don't have them, ROCR can calculate your characteristics
   (sens, spec, precision, recall, whatever) for each of the p-values.
   While you could combine the results "by hand" into a
   ROCR-performance object and let ROCR do the plotting, it is then
   probably easier if you plot directly yourself.

   Don't be shy to look into the prediction and performance objects, I
   find them pretty obvious. Maybe start with the objects produced by
   the examples.

   Also, note ROCR works with binary validation data only. If your data
   has more than one class, you need to make two-class-problems first
   (e.g. protein xy ./. not protein xy).



       BTW, Is there a way of finding the optimum in the curve? I mean
       to find the
       exact value in the ROC curve (see sheet 2 in the excel file for
       the ROC
       curve).


   Someone asked for optimum on ROC a couple of months ago, RSiteSearch
   on the mailing list with ROC and optimal or optimum should get you
   answers.



       I would like to thank for any help in advance

   You're welcome.

   Claudia

   --     Claudia Beleites
   Dipartimento dei Materiali e delle Risorse Naturali
   Università degli Studi di Trieste
   Via Alfonso Valerio 6/a
   I-34127 Trieste

   phone: +39 0 40 5 58-37 68
   email: cbelei...@units.it <mailto:cbelei...@units.it>




--
Claudia Beleites
Dipartimento dei Materiali e delle Risorse Naturali
Università degli Studi di Trieste
Via Alfonso Valerio 6/a
I-34127 Trieste

phone: +39 0 40 5 58-37 68
email: cbelei...@units.it


       [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to