It wouldn't hurt to have p-values returned, but personally, I don't miss them
in scikit-learn. I think that's a classic "ML vs. statistics" discussion --
what I mean is the inference vs. prediction stuff. To me, scikit-learn is
primarily a machine learning library.
> On Apr 19, 2015, at 12:53 A
wrote:
> Good, I was reading your previous comments on the topic as being
> against all frequentist null hypothesis testing.
In the frequentist paradigm I prefer to use model selection instead of
classical hypothesis testing with p-values. My focus is on building useful
models which are able to
wrote:
> Note. The editors of Basic and Applied Social Psychology are also
> banning confidence intervals.
I know. I am not sure I agree on that. I don't particularly like confidence
intervals very much, but I don't hate them with a passion.
Pro: Even though confidence intervals have a bizarre
On Sat, Apr 18, 2015 at 9:45 PM, Sturla Molden wrote:
> wrote:
>
>> (I just went through some articles to see how we can produce p-values
>> after feature selection with penalized least squares or maximum
>> penalized likelihood. :)
>
> If you have used penalized least squares or penalized likeli
On Sat, Apr 18, 2015 at 9:25 PM, Sturla Molden wrote:
> wrote:
>
>>> Re. "We should therefore never compute p-values": I assume that you meant
>>> that within the narrow context of regression, and not, e.g., in the context
>>> of tests of distribution.
>>
>> Sturla means: No null hypothesis testi
wrote:
> (I just went through some articles to see how we can produce p-values
> after feature selection with penalized least squares or maximum
> penalized likelihood. :)
If you have used penalized least squares or penalized likelihood, you have
already pruned the model for parameters that only
wrote:
>> Re. "We should therefore never compute p-values": I assume that you meant
>> that within the narrow context of regression, and not, e.g., in the context
>> of tests of distribution.
>
> Sturla means: No null hypothesis testing at all
Not really, I mean "no p-values for inferential sta
On Sat, Apr 18, 2015 at 6:40 PM, Phillip Feldman
wrote:
> This is a very nice explanation. Thanks!!
>
> Re. "We should therefore never compute p-values": I assume that you meant
> that within the narrow context of regression, and not, e.g., in the context
> of tests of distribution.
Sturla means
This is a very nice explanation. Thanks!!
Re. "We should therefore never compute p-values": I assume that you meant
that within the narrow context of regression, and not, e.g., in the context
of tests of distribution.
On Sat, Apr 18, 2015 at 3:31 PM, Sturla Molden
wrote:
> Phillip Feldman
> w
Phillip Feldman
wrote:
> When using logistic regression, I'm often trying to establish whether a
> given feature has any effect.
Compare models with and without the feature: Cross-validation, BIC, AIC,
PRESS, Bayes factor, etc. By the rules of inductive reasoning (cf. lex
parsimoniae, Occam's
I was able to accomplish what I needed using
`statsmodels.discrete.discrete_model.Logit`. Thanks!
On Sat, Apr 18, 2015 at 11:35 AM, Michael Kneier
wrote:
> Hi Phillip,
>
> Have you checked out statsmodel? That might be a better fit for your needs.
>
> Sent from my iPhone
>
> > On Apr 18, 2015,
Hi Phillip,
Have you checked out statsmodel? That might be a better fit for your needs.
Sent from my iPhone
> On Apr 18, 2015, at 8:31 PM, Phillip Feldman
> wrote:
>
> When using logistic regression, I'm often trying to establish whether a given
> feature has any effect. R and Matlab give m
When using logistic regression, I'm often trying to establish whether a
given feature has any effect. R and Matlab give me p-values, but
Scikit-learn does not. I would love to be able to do all of my statistical
processing in Python. Please consider adding this feature.
Phillip M. Feldman
-
Hi Jason,
memory is a problem in our implementation of MNIST. I sent a detailed
list of the required memory to this mailing list some month ago. You can
find it here:
http://sourceforge.net/p/scikit-learn/mailman/message/33090573/
The number of features is irrelevant. Only the number of sample
14 matches
Mail list logo