In this case, SGD just means a linear model that is fit using stochastic
gradient descent instead of batch gradient methods. If you want to have
more control about the combination of model / loss function / optimization
algorithm, http://contrib.scikit-learn.org/lightning/ is better oriented
for t
Your options are to either pick a clustering algorithm that supports a
pre-computed distance matrix, or, find some kind of projection from C -> R,
embed your data in R, then cluster your embedded data and transfer the
labels back to C.
On Sat, Oct 19, 2019 at 11:44 AM ahmad qassemi
wrote:
> Dear
k-means isn't a convex problem, unless you freeze the initialization, you
are going to get very different solutions (depending on the dataset) with
different initializations.
On Thu, Jun 6, 2019 at 12:05 AM lampahome wrote:
> I tried MiniBatchKMeans with two order:
> fit -> partial_fit
> partial
Hi everyone,
I wanted to use the scikit-learn transformer API to clean up some messy
data as input to a neural network. One of the steps involves converting
categorical variables (of very high cardinality) into integers for use in
an embedding layer.
Unfortunately, I cannot quite use LabelEncode
As a caveat, a lot of clustering algorithms assume that the distance matrix
is a proper metric. If your distance is not a proper metric then the
results might be meaningless (the narrative docs do a good job of
discussing this).
On Mon, 12 Feb 2018 at 13:30 prince gosavi wrote:
> Hi,
> Thanks f
n loss function. Never mentioned about classification and
> hinge loss.
>
> On 13 September 2017 at 23:51, federico vaggi
> wrote:
>
>> You are confusing the kernel with the loss function. SVM minimize a well
>> defined hinge loss on a space that's implicitly defined
You are confusing the kernel with the loss function. SVM minimize a well
defined hinge loss on a space that's implicitly defined by a kernel mapping
(or, in feature space if you use a linear kernel).
On Wed, 13 Sep 2017 at 14:31 Thomas Evangelidis wrote:
> What about the SVM? I use an SVR at th
Hey Gurhan,
sklearn doesn't really neatly separate optimizers from the models they
optimize at the level of API (except in a few cases). In order to make the
package more friendly to newer user, each model has excellent optimizer
defaults that you can use, and only in a few cases does it make sen
If you want to use the model from C++ code, the easiest way is to probably
use Boost/Python (
http://www.boost.org/doc/libs/1_62_0/libs/python/doc/html/index.html).
Alternatively, use another gradient boosting library that has a C++ API
(like XGBoost).
Keep in mind, if you want to call Python code
That's a really good point. Do you know of any systematic studies about
the two different encodings?
Finally: wasn't there a PR for RF to accept categorical variables as inputs?
On Wed, 29 Mar 2017 at 11:57, Olivier Grisel
wrote:
> Integer coding will indeed make the DT assume an arbitrary ord
I imagine he is suggesting to have an iterator that yields results while
it's running, instead of only getting the result at the end of the run.
On Sun, 19 Mar 2017 at 11:46 Joel Nothman wrote:
> Not sure what you mean. Have you used cv_results_
>
> On 18 March 2017 at 08:46, Carlton Banks wrot
Probably not the most principled way to handle it, but: can't you treat 1
dimensional complex numbers as 2 dimensional real numbers, and then try to
cluster those with the GMM?
On Mon, 9 Jan 2017 at 20:34 Rory Smith wrote:
> Hi All,
>
> I’d like to set up a GMM using mixture.BayesianGaussianMixt
That would be most helpful. Maybe also explain the logic?
On Tue, 3 Jan 2017 at 18:19 Andy wrote:
> Should probably be called n_samples_train?
>
>
> On 01/02/2017 04:10 PM, Joel Nothman wrote:
>
> n_indexed means the number of samples in the X passed to fit. It needs to
> be able to compare eac
As long as the feature ordering has a meaningful spatial component (as is
almost always the case when you are dealing with raw pixels as features)
CNNs will almost always be better. CNNs actually have a lot fewer
parameters than MLPs (depending on architecture of course) because of
weight sharing
Hi,
if you want to have the full posterior distribution over the values of the
hyper parameters, there is a good example on how to do that with George +
emcee, another GP package for Python.
http://dan.iel.fm/george/current/user/hyper/
On Tue, 8 Nov 2016 at 16:10 Quaglino Alessio
wrote:
> Hell
This might be interesting to you:
http://blaze.pydata.org/blog/2015/10/19/dask-learn/
On Sun, 7 Aug 2016 at 10:42 Vlad Ionescu wrote:
> Hello,
>
> I am interested in scaling grid searches on an HPC LSF cluster with about
> 60 nodes, each with 20 cores. I thought i could just set n_jobs=1000 th
Another point about the dependency: the dependency is not required for run
time - it is only required to run the type checker. You could easily put
it in a try/catch block and people running scikit-learn wouldn't need it.
On Thu, 4 Aug 2016 at 13:41 Daniel Moisset wrote:
> If the dependency is
I've been using mypy on a much smaller codebase I've been developing.
The main benefits are:
1- Much nicer IDE experience when using something like pycharm. I
expect more text editors to start supporting this in the future.
2- An additional way to catch some compile time errors early on. For
a
My main issue with the upgrade is that if there was a slightly newer
version of numpy/scipy it would try to upgrade my numpy/scipy linked
against MKL/blas to a vanilla version downloaded from the cheese shop. It
was a huge pain.
On Thu, 28 Jul 2016 at 21:17 Matthew Brett wrote:
> On Thu, Jul 28
Hi everyone,
The call for contributions (talks, posters, sprints) is still open until
June 24th.
EuroSciPy 2016 takes place in Erlangen, Germany, from the 23 to the 27 of
August
and consists of two days of tutorials (beginner and advanced tracks) and two
days of conference representing many field
Err, sorry - mu1, mu2, sigma1, sigma2, where mu1, sigma1 are the
mean/standard deviation of the first distribution, and mu2, sigma2 are the
mean and standard deviation of the second distribution.
On Thu, 26 May 2016 at 09:26 federico vaggi
wrote:
> If you are talking about finding the values
- Thanks. will do that
>
> (2) - I am fitting the distribution for 2 different set of values.. I will
> find the distribution as mentioned by you in (1).. But, now having 2
> curves, how do i find the meetings point(s) ?
>
> Regards,
> Sanant
>
> On Thu, May 26, 2016 at 12:16
1) The normal distribution is parametrized by standard deviation and mean.
Simply take the mean and standard deviation of the log of your values?
2) Which curves? You only mentioned a single log normal distribution.
On Thu, 26 May 2016 at 08:42 Startup Hire wrote:
> Hi Michael,
>
> :)
>
>
> (1
23 matches
Mail list logo