Abhi
Check your lib directives. 
https://lucene.apache.org/solr/guide/6_6/lib-directives-in-solrconfig.html#lib-directives-in-solrconfig

I suspect your jars are not in a lib dir mentioned in solrconfig.xml
Cheers -- Rick

On March 23, 2018 11:12:17 AM EDT, Abhi Basu <9000r...@gmail.com> wrote:
>MS Azure does not support Solr 4.9 on HDI, so I am posting here. I
>would
>like to write index collection data to HDFS (hosted on ADL).
>
>Note: I am able to get to ADL from hadoop fs command like, so hadoop is
>configured correctly to get to ADL:
>hadoop fs -ls adl://
>
>This is what I have done so far:
>1. Copied all required jars to sol ext lib folder:
>sudo cp -f /usr/hdp/current/hadoop-client/*.jar
>/usr/hdp/current/solr/example/lib/ext
>sudo cp -f /usr/hdp/current/hadoop-client/lib/*.jar
>/usr/hdp/current/solr/example/lib/ext
>sudo cp -f /usr/hdp/current/hadoop-hdfs-client/*.jar
>/usr/hdp/current/solr/example/lib/ext
>sudo cp -f /usr/hdp/current/hadoop-hdfs-client/lib/*.jar
>/usr/hdp/current/solr/example/lib/ext
>sudo cp -f
>/usr/hdp/current/storm-client/contrib/storm-hbase/storm-hbase*.jar
>/usr/hdp/current/solr/example/lib/ext
>sudo cp -f /usr/hdp/current/phoenix-client/lib/phoenix*.jar
>/usr/hdp/current/solr/example/lib/ext
>sudo cp -f /usr/hdp/current/hbase-client/lib/hbase*.jar
>/usr/hdp/current/solr/example/lib/ext
>
>This includes the Azure active data lake jars also.
>
>2. Edited my solr-config.xml file for my collection:
>
><dataDir>${solr.core.name}/data/</dataDir>
>
><directoryFactory name="DirectoryFactory"
>class="solr.HdfsDirectoryFactory">
><str
>name="solr.hdfs.home">adl://esodevdleus2.azuredatalakestore.net/clusters/esohadoopdeveus2/solr/</str>
>  <str name="solr.hdfs.confdir">/usr/hdp/2.6.2.25-1/hadoop/conf</str>
><str
>name="solr.hdfs.blockcache.global">${solr.hdfs.blockcache.global:true}</str>
>  <bool name="solr.hdfs.blockcache.enabled">true</bool>
>  <int name="solr.hdfs.blockcache.slab.count">1</int>
> <bool name="solr.hdfs.blockcache.direct.memory.allocation">true</bool>
>  <int name="solr.hdfs.blockcache.blocksperbank">16384</int>
>  <bool name="solr.hdfs.blockcache.read.enabled">true</bool>
>  <bool name="solr.hdfs.nrtcachingdirectory.enable">true</bool>
>  <int name="solr.hdfs.nrtcachingdirectory.maxmergesizemb">16</int>
></directoryFactory>
>
>
>When this collection is deployed to solr, I see this error message:
>
><response>
><lst name="responseHeader">
><int name="status">0</int>
><int name="QTime">2189</int></lst>
><lst name="failure">
><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error
>CREATEing SolrCore 'ems-collection_shard2_replica2':
>Unable to create core: ems-collection_shard2_replica2 Caused by: Class
>org.apache.hadoop.fs.adl.HdiAdlFileSystem not
>found</str><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error
>CREATEing SolrCore 'ems-collection_shard2_replica1': Unable to create
>core: ems-collection_shard2_replica1 Caused by: Class
>org.apache.hadoop.fs.adl.HdiAdlFileSystem not
>found</str><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error
>CREATEing SolrCore 'ems-collection_shard1_replica1': Unable to create
>core: ems-collection_shard1_replica1 Caused by: Class
>org.apache.hadoop.fs.adl.HdiAdlFileSystem not
>found</str><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error
>CREATEing SolrCore 'ems-collection_shard1_replica2': Unable to create
>core: ems-collection_shard1_replica2 Caused by: Class
>org.apache.hadoop.fs.adl.HdiAdlFileSystem not found</str>
></lst>
></response>
>
>
>Has anyone done this and can help me out?
>
>Thanks,
>
>Abhi
>
>
>-- 
>Abhi Basu

-- 
Sorry for being brief. Alternate email is rickleir at yahoo dot com 

Reply via email to