Hi,
I am trying to understand what the possible root causes for the
following exception could be.
java.io.FileNotFoundException: File does not exist:
hdfs://*/*/*/*/data/index/_2h.si
I had some long GC pauses while executing some queries which took some
of the replicas down. But how can that a
Okay. Thanks Shawn.
I am using Chef for deploying SolrCloud as a service. The chef-client runs
every 30 minutes and hence the script "install_solr_service" runs every 30
minutes. I changed that.
On Fri, May 19, 2017 at 5:20 PM, Shawn Heisey wrote:
> On 5/19/2017 5:05 PM, Cheta
Perhaps Solr is
already setup as a service on this host? To upgrade Solr use the -f option."
*exit 0*
fi
Thanks!
On Fri, May 19, 2017 at 1:59 PM, Chetas Joshi
wrote:
> Hello,
>
> I am trying to set up a solrCloud (6.5.0/6.5.1). I have installed Solr as
> a service.
>
Hello,
I am trying to set up a solrCloud (6.5.0/6.5.1). I have installed Solr as a
service.
Every time I start solr servers, they come up but one by one the
coreContainers start shutting down on their own within 1-2 minutes of their
being up.
Here are the solr logs
2017-05-19 20:45:30.926 INFO
amples of its usage in StreamingTest.java.
>
> this should
> 1> complete much, much faster. The design goal is 400K rows/second but YMMV
> 2> use vastly less memory on your Solr instances.
> 3> only require _one_ query
>
> Best,
> Erick
>
> On Wed, Apr 12, 2017 at 7:36
Erick
>
> On Wed, Apr 12, 2017 at 12:59 PM, Chetas Joshi
> wrote:
> > I am running a query that returns 10 MM docs in total and the number of
> > rows per page is 100K.
> >
> > On Wed, Apr 12, 2017 at 12:53 PM, Mikhail Khludnev
> wrote:
> >
>
I am running a query that returns 10 MM docs in total and the number of
rows per page is 100K.
On Wed, Apr 12, 2017 at 12:53 PM, Mikhail Khludnev wrote:
> And what is the rows parameter?
>
> 12 апр. 2017 г. 21:32 пользователь "Chetas Joshi"
> написал:
>
> > Tha
fussy.
>
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/ (my blog)
>
>
> > On Apr 11, 2017, at 8:22 PM, Shawn Heisey wrote:
> >
> > On 4/11/2017 2:56 PM, Chetas Joshi wrote:
> >> I am using Solr (5.5.0) on HDFS. SolrC
Hello,
I am using Solr (5.5.0) on HDFS. SolrCloud of 80 nodes. Sold collection
with number of shards = 80 and replication Factor=2
Sold JVM heap size = 20 GB
solr.hdfs.blockcache.enabled = true
solr.hdfs.blockcache.direct.memory.allocation = true
MaxDirectMemorySize = 25 GB
I am querying a solr
Yes, it is scala.
And yes, I just wanted to confirm that I had to add exception handling and
break out of the loop.
Chetas.
On Wed, Feb 22, 2017 at 4:25 PM, Shawn Heisey wrote:
> On 2/22/2017 4:59 PM, Chetas Joshi wrote:
> > 2017-02-22 15:27:06,994 ERROR o.a.s.c.solrj.impl.CloudS
Hello,
I am using Solr 5.5.1. Solr Cloud of 80 nodes deployed on HDFS.
To get back results from Solr, I use the cursor approach and the
cloudSolrClient object. While a query was running, I took the solr Cloud
down. The client got stuck in a loop with the following exception:
2017-02-22 15:27:06,
Is this visible in the logs? I mean how do I find out that a "DELETE
collection" API call was made?
Is the following indicative of the fact that the API call was made?
2017-01-20 20:42:39,822 INFO org.apache.solr.cloud.
ShardLeaderElectionContextBase: Removing leader registration node on
cancel:
Hello,
I have been running Solr (5.5.0) on HDFS.
Recently a collection just went missing with all the instanceDirs and
Datadirs getting deleted. The following logs in the solrCloud overseer.
2017-01-20 20:42:39,515 INFO org.apache.solr.core.SolrCore:
[3044_01_17_shard4_replica1] CLOSING SolrCor
Jan 12, 2017 at 8:42 AM, Shawn Heisey wrote:
> > On 1/11/2017 7:14 PM, Chetas Joshi wrote:
> >> This is what I understand about how Solr works on HDFS. Please correct
> me
> >> if I am wrong.
> >>
> >> Although solr shard replication Factor = 1, HDFS defaul
which I am not able to see.
After a couple of hours a node was assigned to host the shard but the
status of the shard is still "down" and the instance dir is missing on that
node for that particular shard_replica.
Thanks!
On Wed, Jan 11, 2017 at 5:03 PM, Shawn Heisey wrote:
> On 1/
Hello,
I have deployed a SolrCloud (solr 5.5.0) on hdfs using cloudera 5.4.7. The
cloud has 86 nodes.
This is my config for the collection
numShards=80
ReplicationFactor=1
maxShardsPerNode=1
autoAddReplica=true
I recently decommissioned a node to resolve some disk issues. The shard
that was bei
of either, it's a matter of
> which one you're more comfortable with.
>
> Erick
>
>
>
> On Tue, Jan 10, 2017 at 11:12 PM, Shawn Heisey
> wrote:
> > On 1/10/2017 5:28 PM, Chetas Joshi wrote:
> >> I have got 2 shards having hash range set to null due
Hello,
I have got 2 shards having hash range set to null due to some index
corruption.
I am trying to manually get, edit and put the file.
./zkcli.sh -zkhost ${zkhost} -cmd getfile /collections/colName/state.json
~/colName_state.json
./zkcli.sh -zkhost ${zkhost} -cmd clear /collections/colName
Want to add a couple of things
1) Shards were not deleted using the delete replica collection API
endpoint.
2) instanceDir and dataDir exist for all 20 shards.
On Tue, Jan 10, 2017 at 11:34 AM, Chetas Joshi
wrote:
> Hello,
>
> The following is my config
>
> Solr 5.5.0 on HDFS (
Hello,
The following is my config
Solr 5.5.0 on HDFS (SolrCloud of 25 nodes)
collection with shards=20, maxShards per node=1, replicationFactor=1,
autoAddReplicas=true
The ingestion process had been working fine for the last 3 months.
Yesterday, the ingestion process started throwing the follow
?
Thanks!
On Wed, Jan 4, 2017 at 4:11 PM, Shawn Heisey wrote:
> On 1/4/2017 1:43 PM, Chetas Joshi wrote:
> > while creating a new collection, it fails to spin up solr cores on some
> > nodes due to "insufficient direct memory".
> >
> > Here is the error:
>
Hello,
while creating a new collection, it fails to spin up solr cores on some
nodes due to "insufficient direct memory".
Here is the error:
- *3044_01_17_shard42_replica1:*
org.apache.solr.common.SolrException:org.apache.solr.common.SolrException:
The max direct memory is likely too low.
roduce null pointers if a segment had
> no values for a sort field. This is also fixed in the Solr 6x branch.
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Sat, Dec 17, 2016 at 2:44 PM, Chetas Joshi
> wrote:
>
> > Here is the stack trace.
> >
> &
client.solrj.io.stream.CloudSolrStream.read(CloudSolrStream.java:353)
Thanks!
On Fri, Dec 16, 2016 at 11:45 PM, Reth RM wrote:
> If you could provide the json parse exception stack trace, it might help to
> predict issue there.
>
>
> On Fri, Dec 16, 2016 at 5:52 PM, Chetas Joshi
may have been throwing exceptions because the JSON
> special characters were not escaped. This was fixed in Solr 6.0.
>
>
>
>
>
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Fri, Dec 16, 2016 at 4:34 PM, Chetas Joshi
> wrote:
>
> > Hello,
>
Hello,
I am running Solr 5.5.0.
It is a solrCloud of 50 nodes and I have the following config for all the
collections.
maxShardsperNode: 1
replicationFactor: 1
I was using Streaming API to get back results from Solr. It worked fine for
a while until the index data size reached beyond 40 GB per sh
roach?
Thanks!
On Fri, Dec 16, 2016 at 6:52 AM, Shawn Heisey wrote:
> On 12/14/2016 11:58 AM, Chetas Joshi wrote:
> > I am running Solr 5.5.0 on HDFS. It is a solrCloud of 50 nodes and I have
> > the following config.
> > maxShardsperNode: 1
> > replicationFactor: 1
&g
Hi everyone,
I am running Solr 5.5.0 on HDFS. It is a solrCloud of 50 nodes and I have
the following config.
maxShardsperNode: 1
replicationFactor: 1
I have been ingesting data into Solr for the last 3 months. With increase
in data, I am observing increase in the query time. Currently the size of
cs is assembled and
> returned to the client.
> - this sucks up bandwidth and resources
> - that's bad enough, but especially if your ZK nodes are on the same
> box as your Solr nodes they're even more like to have a timeout issue.
>
>
> Best,
> Erick
>
> On Fri, Nov 1
ssing is that you have too much going on
> somehow and you're overloading your system and
> getting a timeout. So increasing the timeout
> is definitely a possibility, or reducing the ingestion load
> as a test.
>
> Best,
> Erick
>
> On Fri, Nov 18, 2016 at 4:5
Hi,
I have a SolrCloud (on HDFS) of 50 nodes and a ZK quorum of 5 nodes. The
SolrCloud is having difficulties talking to ZK when I am ingesting data
into the collections. At that time I am also running queries (that return
millions of docs). The ingest job is crying with the the following exceptio
> same data directory. Whichever one gets there first will block any
> later cores with the
> message you see. So check your core.properties files and your HDFS magic
> to see
> how this is occurring would be my first guess.
>
> Best,
> Erick
>
> On Wed, Nov 16, 2016 at
Hi,
I have a SolrCloud (on HDFS) of 52 nodes. I have 3 collections each with 50
shards and maxShards per node for every collection is 1.
I am having problem restarting a solr shard for a collection.
When I restart, there is always a particular shard of a particular
collection that remains down.
Thanks everyone!
The discussion is really helpful.
Hi Toke, can you explain exactly what you mean by "the aggressive IO for
the memory mapping caused the kernel to start swapping parts of the JVM
heap to get better caching of storage data"?
Which JVM are you talking about? Solr shard? I have other
ust form N queries that were
> restricted to N disjoint subsets of the data and process them all in
> parallel, either with /export or /select.
>
> Best,
> Erick
>
> On Mon, Nov 14, 2016 at 3:53 PM, Chetas Joshi
> wrote:
> > Thanks Joel for the explanation.
> >
gt; > Joel Bernstein
> > http://joelsolr.blogspot.com/
> >
> > On Tue, Nov 8, 2016 at 6:17 PM, Erick Erickson
> > wrote:
> >
> >> Hmm, that should work fine. Let us know what the logs show if anything
> >> because this is w
Hi,
I have a SolrCloud (Solr 5.5.0) of 50 nodes. The JVM heap memory usage of
my solr shards is never more than 50% of the total heap. However, the hosts
on which my solr shards are deployed often run into 99% swap space issue.
This causes the solr shards go down. Why solr shards are so sensitive
in MMapDirectory space
> so will be much, much faster. As of Solr 5.5. You can override the
> decompression stuff, see:
> https://issues.apache.org/jira/browse/SOLR-8220 for fields that are
> both stored and docvalues...
>
> Best,
> Erick
>
> On Sat, Nov 5, 2016 at 6:41 PM
lica1 as long as
> the collection1_shard# parts match you should be fine. If this isn't
> done correctly, the symptom will be that when you update an existing
> document, you may have two copies returned eventually.
>
> Best,
> Erick
>
> On Mon, Nov 7, 2016 at 1:47 PM, Che
DDREPLICA to expand your collection, that'll handle the
> copying from the leader correctly.
>
> Best,
> Erick
>
> On Mon, Nov 7, 2016 at 12:49 PM, Chetas Joshi
> wrote:
> > I have a Solr Cloud deployed on top of HDFS.
> >
> > I accidentally deleted a coll
I have a Solr Cloud deployed on top of HDFS.
I accidentally deleted a collection using the collection API. So, ZooKeeper
cluster has lost all the info related to that collection. I don't have a
backup that I can restore from. However, I have indices and transaction
logs on HDFS.
If I create a new
> No, you can't get cursor-marks ahead of time.
> > They are the serialized representation of the last sort values
> > encountered (hence not known ahead of time).
> >
> > -Yonik
> >
> >
> > On Fri, Nov 4, 2016 at 8:48 PM, Chetas Joshi
> w
Hi,
I am using the cursor approach to fetch results from Solr (5.5.0). Most of
my queries return millions of results. Is there a way I can read the pages
in parallel? Is there a way I can get all the cursors well in advance?
Let's say my query returns 2M documents and I have set rows=100,000.
Can
Hello,
I have the following configuration for the Solr cloud and a Solr collection
This is Solr on HDFS and Solr version I am using is 5.5.0
No. of hosts: 52 (Solr Cloud)
shard count: 50
replicationFactor: 1
MaxShardsPerNode: 1
autoAddReplicas: true
Now, one of my shar
Just to the add to my previous question: I used dynamic shard splitting
while consuming data from the Solr collection using /export handler.
On Fri, Oct 21, 2016 at 2:27 PM, Chetas Joshi
wrote:
> Thanks Joel.
>
> I will migrate to Solr 6.0.0.
>
> However, I have one more question
ng Expressions.
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Thu, Oct 20, 2016 at 5:49 PM, Chetas Joshi
> wrote:
>
> > Hello,
> >
> > I am using /export handler to stream data using CloudSolrStream.
> >
> > I am using fl=uuid,space,tim
You just need to have indexed=true. It will use the inverted index to
delete the expired documents. You don't need stored=true as all the info
required by the DocExpirationUpdateProcessorFactory to delete a document is
there in the inverted index.
On Thu, Oct 20, 2016 at 4:26 PM, Brent wrote:
>
Hello,
I am using /export handler to stream data using CloudSolrStream.
I am using fl=uuid,space,timestamp where uuid and space are Strings and
timestamp is long. My query (q=...) is not on these fields.
While reading the results from the Solr cloud, I get the following errors
org.noggit.JS
Is this happening because I have set replicationFactor=1?
So even if I manually add replica for the shard that's down, it will just
create a dataDir but would not copy any of the data into the dataDir?
On Tue, Sep 13, 2016 at 6:07 PM, Chetas Joshi
wrote:
> Hi,
>
> I just started
Hi,
I just started experimenting with solr cloud.
I have a solr cloud of 20 nodes. I have one collection with 18 shards
running on 18 different nodes with replication factor=1.
When one of my shards goes down, I create a replica using the Solr UI. On
HDFS I see a core getting added. But the data
50 matches
Mail list logo