Thanks Abhishek.
I will check it out.

Thank you,
Naresh

On Sat, Jul 27, 2019 at 9:21 PM Abhishek Somani <abhisheksoman...@gmail.com>
wrote:

> Hey Naresh,
>
> There is a `shaded-dependecies` project inside the root directory. You
> need to go into that and build and publish that to local first.
>
> cd shaded-dependencies
>> sbt clean publishLocal
>>
>
> After that, come back out to the root directory and build that project.
> The spark-acid-shaded-dependencies jar will now be found:
>
>> cd ..
>> sbt assembly
>
>
> This will create the jar which you can use.
>
> On another note, unless you are making changes in the code, you don't need
> to build yourself as the jar is published in
> https://spark-packages.org/package/qubole/spark-acid. So you can just use
> it as:
>
> spark-shell --packages qubole:spark-acid:0.4.0-s_2.11
>
>
> ...and it will be automatically fetched and used.
>
> Thanks,
> Abhishek
>
>
> On Sun, Jul 28, 2019 at 4:42 AM naresh Goud <nareshgoud.du...@gmail.com>
> wrote:
>
>> It looks there is some internal dependency missing.
>>
>> libraryDependencies ++= Seq(
>> "com.qubole" %% "spark-acid-shaded-dependencies" % "0.1"
>> )
>>
>> How do we get it?
>>
>>
>> Thank you,
>> Naresh
>>
>>
>>
>>
>> Thanks,
>> Naresh
>> www.linkedin.com/in/naresh-dulam
>> http://hadoopandspark.blogspot.com/
>>
>>
>>
>> On Sat, Jul 27, 2019 at 5:34 PM naresh Goud <nareshgoud.du...@gmail.com>
>> wrote:
>>
>>> Hi Abhishek,
>>>
>>>
>>> We are not able to build jar using git hub code with below error?
>>>
>>> Any others able to build jars? Is there anything else missing?
>>>
>>>
>>>
>>> Note: Unresolved dependencies path:
>>> [warn]          com.qubole:spark-acid-shaded-dependencies_2.11:0.1
>>> (C:\Data\Hadoop\spark-acid-master\build.sbt#L51-54)
>>> [warn]            +- com.qubole:spark-acid_2.11:0.4.0
>>> sbt.ResolveException: unresolved dependency:
>>> com.qubole#spark-acid-shaded-dependencies_2.11;0.1: not found
>>>         at sbt.IvyActions$.sbt$IvyActions$$resolve(IvyActions.scala:313)
>>>         at
>>> sbt.IvyActions$$anonfun$updateEither$1.apply(IvyActions.scala:191)
>>>         at
>>> sbt.IvyActions$$anonfun$updateEither$1.apply(IvyActions.scala:168)
>>>         at sbt.IvySbt$Module$$anonfun$withModule$1.apply(Ivy.scala:156)
>>>         at sbt.IvySbt$Module$$anonfun$withModule$1.apply(Ivy.scala:156)
>>>         at sbt.IvySbt$$anonfun$withIvy$1.apply(Ivy.scala:133)
>>>         at sbt.IvySbt.sbt$IvySbt$$action$1(Ivy.scala:57)
>>>         at sbt.IvySbt$$anon$4.call(Ivy.scala:65)
>>>         at xsbt.boot.Locks$GlobalLock.withChannel$1(Locks.scala:93)
>>>         at
>>> xsbt.boot.Locks$GlobalLock.xsbt$boot$Locks$GlobalLock$$withChannelRetries$1(Locks.scala:78)
>>>         at
>>> xsbt.boot.Locks$GlobalLock$$anonfun$withFileLock$1.apply(Locks.scala:97)
>>>         at xsbt.boot.Using$.withResource(Using.scala:10)
>>>         at xsbt.boot.Using$.apply(Using.scala:9)
>>>         at
>>> xsbt.boot.Locks$GlobalLock.ignoringDeadlockAvoided(Locks.scala:58)
>>>         at xsbt.boot.Locks$GlobalLock.withLock(Locks.scala:48)
>>>         at xsbt.boot.Locks$.apply0(Locks.scala:31)
>>>         at xsbt.boot.Locks$.apply(Locks.scala:28)
>>>         at sbt.IvySbt.withDefaultLogger(Ivy.scala:65)
>>>         at sbt.IvySbt.withIvy(Ivy.scala:128)
>>>         at sbt.IvySbt.withIvy(Ivy.scala:125)
>>>         at sbt.IvySbt$Module.withModule(Ivy.scala:156)
>>>         at sbt.IvyActions$.updateEither(IvyActions.scala:168)
>>>         at
>>> sbt.Classpaths$$anonfun$sbt$Classpaths$$work$1$1.apply(Defaults.scala:1541)
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> Thanks,
>>> Naresh
>>> www.linkedin.com/in/naresh-dulam
>>> http://hadoopandspark.blogspot.com/
>>>
>>>
>>>
>>> On Sat, Jul 27, 2019 at 3:25 PM Nicolas Paris <nicolas.pa...@riseup.net>
>>> wrote:
>>>
>>>> Congrats
>>>>
>>>> The read/write feature with hive3 is highly interesting
>>>>
>>>> On Fri, Jul 26, 2019 at 06:07:55PM +0530, Abhishek Somani wrote:
>>>> > Hi All,
>>>> >
>>>> > We at Qubole have open sourced a datasource that will enable users to
>>>> work on
>>>> > their Hive ACID Transactional Tables using Spark.
>>>> >
>>>> > Github: https://github.com/qubole/spark-acid
>>>> >
>>>> > Hive ACID tables allow users to work on their data transactionally,
>>>> and also
>>>> > gives them the ability to Delete, Update and Merge data efficiently
>>>> without
>>>> > having to rewrite all of their data in a table, partition or file. We
>>>> believe
>>>> > that being able to work on these tables from Spark is a much desired
>>>> value add,
>>>> > as is also apparent in
>>>> https://issues.apache.org/jira/browse/SPARK-15348 and
>>>> > https://issues.apache.org/jira/browse/SPARK-16996 with multiple
>>>> people looking
>>>> > for it. Currently the datasource supports reading from these ACID
>>>> tables only,
>>>> > and we are working on adding the ability to write into these tables
>>>> via Spark
>>>> > as well.
>>>> >
>>>> > The datasource is also available as a spark package, and instructions
>>>> on how to
>>>> > use it are available on the Github page.
>>>> >
>>>> > We welcome your feedback and suggestions.
>>>> >
>>>> > Thanks,
>>>> > Abhishek Somani
>>>>
>>>> --
>>>> nicolas
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org
>>>>
>>>> --
Thanks,
Naresh
www.linkedin.com/in/naresh-dulam
http://hadoopandspark.blogspot.com/

Reply via email to