Software developers are sometimes compensated based on the degree of
complexity that they deal with.
And managers are sometimes compensated based on the number of people they
manage, as well as the degree of complexity of what they manage.
And... training organizations can charge more and have a larger pool of
eager customers when the subject matter has higher complexity.
And... consultants and contractors will be in higher demand and able to
charge more, based on the degree of complexity that they have mastered.
So, more complexity results in greater opportunity for higher income!
(Oh, and, writers and book authors have more to write about and readers are
more eager to purchase those writings as well, especially if the subject
matter is constantly changing.)
Somebody please remind me I said this any time you catch me trying to argue
for Solr to be made simpler and easier to use!
-- Jack Krupansky
-----Original Message-----
From: Walter Underwood
Sent: Friday, July 05, 2013 12:11 PM
To: solr-user@lucene.apache.org
Subject: Re: Sending Documents via SolrServer as MapReduce Jobs at Solrj
Why is it better to require another large software system (Hadoop), when it
works fine without it?
That just sounds like more stuff to configure, misconfigure, and cause
problems with indexing.
wunder
On Jul 5, 2013, at 4:48 AM, Furkan KAMACI wrote:
We are using Nutch to crawl web sites and it stores documents at Hbase.
Nutch uses Solrj to send documents to be indexed. We have Hadoop at our
ecosystem as well. I think that there should be an implementation at Solrj
that sends documents (via CloudSolrServer or something like that) as
MapReduce jobs. Is there any implentation for it or is it not a good idea?