SplitShardCmd assumes that its main phase (when the Lucene index is being 
split) always executes on the local file system of the shard leader, and indeed 
the ShardSplitCmd.checkDiskSpace() checks the local file system’s free disk 
space - even though in reality in your case the actual data is written to the 
HDFS Directory so it (almost) doesn’t affect the local FS…

Please file a JIRA request to improve this. For now you simply have to make 
sure that you have at least 2x the index size of free disk space available on 
the shard leader.

> On 16 Sep 2019, at 18:15, Joe Obernberger <joseph.obernber...@gmail.com> 
> wrote:
> 
> Hi All - added a couple more solr nodes to an existing solr cloud cluster 
> where the index is in HDFS.  When I try to a split a shard, I get an error 
> saying there is not enough disk space.  It looks like it is looking on the 
> local file system, and not in HDFS.
> 
> "Operation splitshard casued 
> exception:":"org.apache.solr.common.SolrException:org.apache.solr.common.SolrException:
>  not enough free disk space to perform index split on node
> 
> -Joe
> 

Reply via email to