Thanks gents.

Removal of 'clean package -U' made the difference.

On Tue, Oct 27, 2015 at 6:39 PM, Todd Nist <tsind...@gmail.com> wrote:

> I issued the same basic command and it worked fine.
>
> RADTech-MBP:spark $ ./make-distribution.sh --name hadoop-2.6 --tgz -Pyarn
> -Phadoop-2.6 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests
>
> Which created: spark-1.6.0-SNAPSHOT-bin-hadoop-2.6.tgz in the root
> directory of the project.
>
> FWIW, the environment was an MBP with OS X 10.10.5 and Java:
>
> java version "1.8.0_51"
> Java(TM) SE Runtime Environment (build 1.8.0_51-b16)
> Java HotSpot(TM) 64-Bit Server VM (build 25.51-b03, mixed mode)
>
> -Todd
>
> On Tue, Oct 27, 2015 at 12:17 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> I used the following command:
>> make-distribution.sh --name custom-spark --tgz -Phadoop-2.4 -Phive
>> -Phive-thriftserver -Pyarn
>>
>> spark-1.6.0-SNAPSHOT-bin-custom-spark.tgz was generated (with patch from
>> SPARK-11348)
>>
>> Can you try above command ?
>>
>> Thanks
>>
>> On Tue, Oct 27, 2015 at 7:03 AM, Kayode Odeyemi <drey...@gmail.com>
>> wrote:
>>
>>> Ted, I switched to this:
>>>
>>> ./make-distribution.sh --name spark-latest --tgz -Dhadoop.version=2.6.0
>>> -Phadoop-2.6 -Phive -Phive-thriftserver -Pyarn -DskipTests clean package -U
>>>
>>> Same error. No .gz file. Here's the bottom output log:
>>>
>>> + rm -rf /home/emperor/javaprojects/spark/dist
>>> + mkdir -p /home/emperor/javaprojects/spark/dist/lib
>>> + echo 'Spark [WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin (git revision
>>> 3689beb) built for Hadoop [WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Pl
>>> + echo 'Build flags: -Dhadoop.version=2.6.0' -Phadoop-2.6 -Phive
>>> -Phive-thriftserver -Pyarn -DskipTests clean package -U
>>> + cp
>>> /home/emperor/javaprojects/spark/assembly/target/scala-2.10/spark-assembly-1.6.0-SNAPSHOT-hadoop2.6.0.jar
>>> /home/emperor/javaprojects/spark/dist/lib/
>>> + cp
>>> /home/emperor/javaprojects/spark/examples/target/scala-2.10/spark-examples-1.6.0-SNAPSHOT-hadoop2.6.0.jar
>>> /home/emperor/javaprojects/spark/dist/lib/
>>> + cp
>>> /home/emperor/javaprojects/spark/network/yarn/target/scala-2.10/spark-1.6.0-SNAPSHOT-yarn-shuffle.jar
>>> /home/emperor/javaprojects/spark/dist/lib/
>>> + mkdir -p /home/emperor/javaprojects/spark/dist/examples/src/main
>>> + cp -r /home/emperor/javaprojects/spark/examples/src/main
>>> /home/emperor/javaprojects/spark/dist/examples/src/
>>> + '[' 1 == 1 ']'
>>> + cp
>>> /home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-api-jdo-3.2.6.jar
>>> /home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-core-3.2.10.jar
>>> /home/emperor/javaprojects
>>> ed/jars/datanucleus-rdbms-3.2.9.jar
>>> /home/emperor/javaprojects/spark/dist/lib/
>>> + cp /home/emperor/javaprojects/spark/LICENSE
>>> /home/emperor/javaprojects/spark/dist
>>> + cp -r /home/emperor/javaprojects/spark/licenses
>>> /home/emperor/javaprojects/spark/dist
>>> + cp /home/emperor/javaprojects/spark/NOTICE
>>> /home/emperor/javaprojects/spark/dist
>>> + '[' -e /home/emperor/javaprojects/spark/CHANGES.txt ']'
>>> + cp -r /home/emperor/javaprojects/spark/data
>>> /home/emperor/javaprojects/spark/dist
>>> + mkdir /home/emperor/javaprojects/spark/dist/conf
>>> + cp /home/emperor/javaprojects/spark/conf/docker.properties.template
>>> /home/emperor/javaprojects/spark/conf/fairscheduler.xml.template
>>> /home/emperor/javaprojects/spark/conf/log4j.properties
>>> emperor/javaprojects/spark/conf/metrics.properties.template
>>> /home/emperor/javaprojects/spark/conf/slaves.template
>>> /home/emperor/javaprojects/spark/conf/spark-defaults.conf.template /home/em
>>> ts/spark/conf/spark-env.sh.template
>>> /home/emperor/javaprojects/spark/dist/conf
>>> + cp /home/emperor/javaprojects/spark/README.md
>>> /home/emperor/javaprojects/spark/dist
>>> + cp -r /home/emperor/javaprojects/spark/bin
>>> /home/emperor/javaprojects/spark/dist
>>> + cp -r /home/emperor/javaprojects/spark/python
>>> /home/emperor/javaprojects/spark/dist
>>> + cp -r /home/emperor/javaprojects/spark/sbin
>>> /home/emperor/javaprojects/spark/dist
>>> + cp -r /home/emperor/javaprojects/spark/ec2
>>> /home/emperor/javaprojects/spark/dist
>>> + '[' -d /home/emperor/javaprojects/spark/R/lib/SparkR ']'
>>> + '[' false == true ']'
>>> + '[' true == true ']'
>>> + TARDIR_NAME='spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest
>>> '
>>> + TARDIR='/home/emperor/javaprojects/spark/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest
>>> '
>>> + rm -rf '/home/emperor/javaprojects/spark/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest
>>> '
>>> + cp -r /home/emperor/javaprojects/spark/dist
>>> '/home/emperor/javaprojects/spark/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest
>>> '
>>> cp: cannot create directory
>>> `/home/emperor/javaprojects/spark/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest':
>>> No such file or directory
>>>
>>>
>>> On Tue, Oct 27, 2015 at 2:14 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>
>>>> Can you try the same command shown in the pull request ?
>>>>
>>>> Thanks
>>>>
>>>> On Oct 27, 2015, at 12:40 AM, Kayode Odeyemi <drey...@gmail.com> wrote:
>>>>
>>>> Thank you.
>>>>
>>>> But I'm getting same warnings and it's still preventing the archive
>>>> from being generated.
>>>>
>>>> I've ran this on both OSX Lion and Ubuntu 12. Same error. No .gz file
>>>>
>>>> On Mon, Oct 26, 2015 at 9:10 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>>
>>>>> Looks like '-Pyarn' was missing in your command.
>>>>>
>>>>> On Mon, Oct 26, 2015 at 12:06 PM, Kayode Odeyemi <drey...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> I used this command which is synonymous to what you have:
>>>>>>
>>>>>> ./make-distribution.sh --name spark-latest --tgz --mvn mvn
>>>>>> -Dhadoop.version=2.6.0 -Phadoop-2.6 -Phive -Phive-thriftserver 
>>>>>> -DskipTests
>>>>>> clean package -U
>>>>>>
>>>>>> But I still see WARNINGS like this in the output and no .gz file
>>>>>> created:
>>>>>>
>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/.part-r-00005.gz.parquet.crc:
>>>>>> No such file or directory
>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet:
>>>>>> No such file or directory
>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>>>> No such file or directory
>>>>>> cp:
>>>>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>>>> unable to copy extended attributes to
>>>>>> /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>>>> No such file or directory
>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>>>> No such file or directory
>>>>>> cp:
>>>>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>>>> unable to copy extended attributes to
>>>>>> /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>>>> No such file or directory
>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/.part-r-00007.gz.parquet.crc:
>>>>>> No such file or directory
>>>>>>
>>>>>> On Mon, Oct 26, 2015 at 8:58 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>>>>
>>>>>>> If you use the command shown in:
>>>>>>> https://github.com/apache/spark/pull/9281
>>>>>>>
>>>>>>> You should have got the following:
>>>>>>>
>>>>>>>
>>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/part-r-00008.gz.parquet
>>>>>>>
>>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/part-r-00007.gz.parquet
>>>>>>>
>>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00004.gz.parquet
>>>>>>>
>>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00002.gz.parquet
>>>>>>>
>>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet
>>>>>>>
>>>>>>> On Mon, Oct 26, 2015 at 11:47 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I see a lot of stuffs like this after the a successful maven build:
>>>>>>>>
>>>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/
>>>>>>>> part-r-00008.gz.parquet: No such file or directory
>>>>>>>>
>>>>>>>> Seems it fails when it tries to package the build as an archive.
>>>>>>>>
>>>>>>>> I'm using the latest code on github master.
>>>>>>>>
>>>>>>>> Any ideas please?
>>>>>>>>
>>>>>>>> On Mon, Oct 26, 2015 at 6:20 PM, Yana Kadiyska <
>>>>>>>> yana.kadiy...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> In 1.4 ./make_distribution produces a .tgz file in the root
>>>>>>>>> directory (same directory that make_distribution is in)
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Oct 26, 2015 at 8:46 AM, Kayode Odeyemi <drey...@gmail.com
>>>>>>>>> > wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> The ./make_distribution task completed. However, I can't seem to
>>>>>>>>>> locate the
>>>>>>>>>> .tar.gz file.
>>>>>>>>>>
>>>>>>>>>> Where does Spark save this? or should I just work with the dist
>>>>>>>>>> directory?
>>>>>>>>>>
>>>>>>>>>> On Fri, Oct 23, 2015 at 4:23 PM, Kayode Odeyemi <
>>>>>>>>>> drey...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> I saw this when I tested manually (without ./make-distribution)
>>>>>>>>>>>
>>>>>>>>>>> Detected Maven Version: 3.2.2 is not in the allowed range 3.3.3.
>>>>>>>>>>>
>>>>>>>>>>> So I simply upgraded maven to 3.3.3.
>>>>>>>>>>>
>>>>>>>>>>> Resolved. Thanks
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Oct 23, 2015 at 3:17 PM, Sean Owen <so...@cloudera.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> This doesn't show the actual error output from Maven. I have a
>>>>>>>>>>>> strong
>>>>>>>>>>>> guess that you haven't set MAVEN_OPTS to increase the memory
>>>>>>>>>>>> Maven can
>>>>>>>>>>>> use.
>>>>>>>>>>>>
>>>>>>>>>>>> On Fri, Oct 23, 2015 at 6:14 AM, Kayode Odeyemi <
>>>>>>>>>>>> drey...@gmail.com> wrote:
>>>>>>>>>>>> > Hi,
>>>>>>>>>>>> >
>>>>>>>>>>>> > I can't seem to get a successful maven build. Please see
>>>>>>>>>>>> command output
>>>>>>>>>>>> > below:
>>>>>>>>>>>> >
>>>>>>>>>>>> > bash-3.2$ ./make-distribution.sh --name spark-latest --tgz
>>>>>>>>>>>> --mvn mvn
>>>>>>>>>>>> > -Dhadoop.version=2.7.0 -Phadoop-2.7 -Phive
>>>>>>>>>>>> -Phive-thriftserver -DskipTests
>>>>>>>>>>>> > clean package
>>>>>>>>>>>> > +++ dirname ./make-distribution.sh
>>>>>>>>>>>> > ++ cd .
>>>>>>>>>>>> > ++ pwd
>>>>>>>>>>>> > + SPARK_HOME=/usr/local/spark-latest
>>>>>>>>>>>> > + DISTDIR=/usr/local/spark-latest/dist
>>>>>>>>>>>> > + SPARK_TACHYON=false
>>>>>>>>>>>> > + TACHYON_VERSION=0.7.1
>>>>>>>>>>>> > + TACHYON_TGZ=tachyon-0.7.1-bin.tar.gz
>>>>>>>>>>>> > +
>>>>>>>>>>>> > TACHYON_URL=
>>>>>>>>>>>> https://github.com/amplab/tachyon/releases/download/v0.7.1/tachyon-0.7.1-bin.tar.gz
>>>>>>>>>>>> > + MAKE_TGZ=false
>>>>>>>>>>>> > + NAME=none
>>>>>>>>>>>> > + MVN=/usr/local/spark-latest/build/mvn
>>>>>>>>>>>> > + ((  12  ))
>>>>>>>>>>>> > + case $1 in
>>>>>>>>>>>> > + NAME=spark-latest
>>>>>>>>>>>> > + shift
>>>>>>>>>>>> > + shift
>>>>>>>>>>>> > + ((  10  ))
>>>>>>>>>>>> > + case $1 in
>>>>>>>>>>>> > + MAKE_TGZ=true
>>>>>>>>>>>> > + shift
>>>>>>>>>>>> > + ((  9  ))
>>>>>>>>>>>> > + case $1 in
>>>>>>>>>>>> > + MVN=mvn
>>>>>>>>>>>> > + shift
>>>>>>>>>>>> > + shift
>>>>>>>>>>>> > + ((  7  ))
>>>>>>>>>>>> > + case $1 in
>>>>>>>>>>>> > + break
>>>>>>>>>>>> > + '[' -z
>>>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>>>>> > + '[' -z
>>>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>>>>> > ++ command -v git
>>>>>>>>>>>> > + '[' /usr/bin/git ']'
>>>>>>>>>>>> > ++ git rev-parse --short HEAD
>>>>>>>>>>>> > + GITREV=487d409
>>>>>>>>>>>> > + '[' '!' -z 487d409 ']'
>>>>>>>>>>>> > + GITREVSTRING=' (git revision 487d409)'
>>>>>>>>>>>> > + unset GITREV
>>>>>>>>>>>> > ++ command -v mvn
>>>>>>>>>>>> > + '[' '!' /usr/bin/mvn ']'
>>>>>>>>>>>> > ++ mvn help:evaluate -Dexpression=project.version
>>>>>>>>>>>> -Dhadoop.version=2.7.0
>>>>>>>>>>>> > -Phadoop-2.7 -Phive -Phive-thriftserver -DskipTests clean
>>>>>>>>>>>> package
>>>>>>>>>>>> > ++ grep -v INFO
>>>>>>>>>>>> > ++ tail -n 1
>>>>>>>>>>>> > + VERSION='[ERROR] [Help 1]
>>>>>>>>>>>> >
>>>>>>>>>>>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
>>>>>>>>>>>> '
>>>>>>>>>>>> >
>>>>>>>>>>>> > Same output error with JDK 7
>>>>>>>>>>>> >
>>>>>>>>>>>> > Appreciate your help.
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to