Hi Abhishek,
Try below spark submit.
spark-submit --master yarn --deploy-mode cluster --files
hdfs://abc.com:8020/tmp/abc.drl <http://abc.com:8020/tmp/abc.drl>
--class com.abc.StartMain abc-0.0.1-SNAPSHOT-jar-with-dependencies.jar
abc.drl <http://abc.com:8020/tmp/abc.drl>
On Friday 23 September 2016 07:29 PM, ABHISHEK wrote:
Thanks for your response Aditya and Steve.
Steve:
I have tried specifying both /tmp/filename in hdfs and local path but
it didn't work.
You may be write that Kie session is configured to access files from
Local path.
I have attached code here for your reference and if you find some
thing wrong, please help to correct it.
Aditya:
I have attached code here for reference. --File option will
distributed reference file to all node but Kie session is not able
to pickup it.
Thanks,
Abhishek
On Fri, Sep 23, 2016 at 2:25 PM, Steve Loughran
<ste...@hortonworks.com <mailto:ste...@hortonworks.com>> wrote:
On 23 Sep 2016, at 08:33, ABHISHEK <abhi...@gmail.com
<mailto:abhi...@gmail.com>> wrote:
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.FileNotFoundException:
hdfs:/abc.com:8020/user/abhietc/abc.drl
<http://abc.com:8020/user/abhietc/abc.drl>(No such file or directory)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.<init>(FileInputStream.java:146)
at
org.drools.core.io.impl.FileSystemResource.getInputStream(FileSystemResource.java:123)
at
org.drools.compiler.kie.builder.impl.KieFileSystemImpl.write(KieFileSystemImpl.java:58)
Looks like this .KieFileSystemImpl class only works with local
files, so when it gets an HDFS path in it tries to open it and
gets confused.
you may need to write to a local FS temp file then copy it into HDFS