Hi leesf,

Thanks for your reply and the instructions. I'll give it a try.

On Sun, Sep 20, 2020 at 9:04 PM leesf <[email protected]> wrote:

> Hi Rui,
>
> 1. It is because utilities.bundle.hive.scope is provided by default, thus
> not included in hudi-utilities-bundle-xxx.jar, and you would set 
> -Dutilities.bundle.hive.scope=runtime
> to re-compile from master branch to include hive dependencies, also I think
> adding hive-common jar to class path should be fine without setting -D
> utilities.bundle.hive.scope=runtime. But as you pointed out that even you
> don't sync to hive, you still get the error, I think we would document it
> at least.
> 2. There is a PR to fix the jetty version conflict, please check it out,
> https://github.com/apache/hudi/pull/1990.
>
> Rui Li <[email protected]> 于2020年9月17日周四 下午8:26写道:
>
>> Hello,
>>
>> I'm a newbie and just trying to run my very first example with Hudi -- to
>> ingest CSV DFS source to a Hudi table. I hit the following problems in the
>> process.
>>
>> 1. Class not found exception for HiveConf. I didn't enable hive sync but
>> still got this error. I guess it's because the class is imported in
>> DeltaSync. I solved this by adding hive-common to class path. (I tried
>> hive-exec at first but that caused conflicts with Parquet)
>>
>> 2. No such method error for jetty SessionHandler::setHttpOnly. It's
>> because the jetty-server version conflicts between Hudi and my Hadoop. I
>> solved this problem by setting spark.driver.userClassPathFirst=true.
>>
>> Although I've managed to make the program run successfully, I wonder
>> whether I'm doing it right and what's the recommended way to add
>> dependencies.
>>
>> The components I'm using:
>> Spark 2.4.6 w/o Hadoop
>> Hadoop 3.0.3
>> Hive 2.3.4
>> Hudi latest master code
>>
>> Thanks in advance!
>>
>> --
>> Cheers,
>> Rui Li
>>
>

-- 
Cheers,
Rui Li

Reply via email to