I am able to use blockjoin API and it does not throw compilation error

val viEventsWithListings: RDD[(Long, (DetailInputRecord, VISummary, Long))]
= lstgItem.blockJoin(viEvents,1,1).map {

}

Here viEvents is highly skewed and both are on HDFS.

What should be the optimal values of replication, i gave 1,1



On Sun, Jun 28, 2015 at 1:47 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> wrote:

> I incremented the version of spark from 1.4.0 to 1.4.0.1 and ran
>
>  ./make-distribution.sh  --tgz -Phadoop-2.4 -Pyarn  -Phive
> -Phive-thriftserver
>
> Build was successful but the script faild. Is there a way to pass the
> incremented version ?
>
>
> [INFO] BUILD SUCCESS
>
> [INFO]
> ------------------------------------------------------------------------
>
> [INFO] Total time: 09:56 min
>
> [INFO] Finished at: 2015-06-28T13:45:29-07:00
>
> [INFO] Final Memory: 84M/902M
>
> [INFO]
> ------------------------------------------------------------------------
>
> + rm -rf /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist
>
> + mkdir -p /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib
>
> + echo 'Spark 1.4.0.1 built for Hadoop 2.4.0'
>
> + echo 'Build flags: -Phadoop-2.4' -Pyarn -Phive -Phive-thriftserver
>
> + cp
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/assembly/target/scala-2.10/spark-assembly-1.4.0.1-hadoop2.4.0.jar
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/
>
> + cp
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/examples/target/scala-2.10/spark-examples-1.4.0.1-hadoop2.4.0.jar
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/
>
> + cp
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/network/yarn/target/scala-2.10/spark-1.4.0.1-yarn-shuffle.jar
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/
>
> + mkdir -p
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/examples/src/main
>
> + cp -r /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/examples/src/main
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/examples/src/
>
> + '[' 1 == 1 ']'
>
> + cp
> '/Users/dvasthimal/ebay/projects/ep/spark-1.4.0/lib_managed/jars/datanucleus*.jar'
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/
>
> cp:
> /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/lib_managed/jars/datanucleus*.jar:
> No such file or directory
>
> LM-SJL-00877532:spark-1.4.0 dvasthimal$ ./make-distribution.sh  --tgz
> -Phadoop-2.4 -Pyarn  -Phive -Phive-thriftserver
>
>
>
> On Sun, Jun 28, 2015 at 1:41 PM, Koert Kuipers <ko...@tresata.com> wrote:
>
>> you need 1) to publish to inhouse maven, so your application can depend
>> on your version, and 2) use the spark distribution you compiled to launch
>> your job (assuming you run with yarn so you can launch multiple versions of
>> spark on same cluster)
>>
>> On Sun, Jun 28, 2015 at 4:33 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com>
>> wrote:
>>
>>> How can i import this pre-built spark into my application via maven as i
>>> want to use the block join API.
>>>
>>> On Sun, Jun 28, 2015 at 1:31 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com>
>>> wrote:
>>>
>>>> I ran this w/o maven options
>>>>
>>>> ./make-distribution.sh  --tgz -Phadoop-2.4 -Pyarn  -Phive
>>>> -Phive-thriftserver
>>>>
>>>> I got this spark-1.4.0-bin-2.4.0.tgz in the same working directory.
>>>>
>>>> I hope this is built with 2.4.x hadoop as i did specify -P
>>>>
>>>> On Sun, Jun 28, 2015 at 1:10 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com>
>>>> wrote:
>>>>
>>>>>  ./make-distribution.sh  --tgz --*mvn* "-Phadoop-2.4 -Pyarn
>>>>> -Dhadoop.version=2.4.0 -Phive -Phive-thriftserver -DskipTests clean 
>>>>> package"
>>>>>
>>>>>
>>>>> or
>>>>>
>>>>>
>>>>>  ./make-distribution.sh  --tgz --*mvn* -Phadoop-2.4 -Pyarn
>>>>> -Dhadoop.version=2.4.0 -Phive -Phive-thriftserver -DskipTests clean 
>>>>> package"
>>>>> ​Both fail with
>>>>>
>>>>> + echo -e 'Specify the Maven command with the --mvn flag'
>>>>>
>>>>> Specify the Maven command with the --mvn flag
>>>>>
>>>>> + exit -1
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Deepak
>>>>
>>>>
>>>
>>>
>>> --
>>> Deepak
>>>
>>>
>>
>
>
> --
> Deepak
>
>


-- 
Deepak

Reply via email to