You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build.
Maintainer: de...@freebsd.org Last committer: de...@freebsd.org Ident: $FreeBSD: branches/2015Q4/devel/hadoop2/Makefile 396358 2015-09-08 09:16:36Z demon $ Log URL: http://beefy3.nyi.freebsd.org/data/93i386-quarterly/400677/logs/hadoop2-2.7.1.log Build URL: http://beefy3.nyi.freebsd.org/build.html?mastername=93i386-quarterly&build=400677 Log: ====>> Building devel/hadoop2 build started at Tue Nov 3 01:46:07 UTC 2015 port directory: /usr/ports/devel/hadoop2 building for: FreeBSD 93i386-quarterly-job-03 9.3-RELEASE-p29 FreeBSD 9.3-RELEASE-p29 i386 maintained by: de...@freebsd.org Makefile ident: $FreeBSD: branches/2015Q4/devel/hadoop2/Makefile 396358 2015-09-08 09:16:36Z demon $ Poudriere version: 3.1.9 Host OSVERSION: 1100079 Jail OSVERSION: 903000 ---Begin Environment--- SHELL=/bin/csh UNAME_p=i386 UNAME_m=i386 OSVERSION=903000 UNAME_v=FreeBSD 9.3-RELEASE-p29 UNAME_r=9.3-RELEASE-p29 BLOCKSIZE=K MAIL=/var/mail/root STATUS=1 SAVED_TERM= MASTERMNT=/usr/local/poudriere/data/.m/93i386-quarterly/ref PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin POUDRIERE_BUILD_TYPE=bulk PKGNAME=hadoop2-2.7.1 OLDPWD=/ PWD=/usr/local/poudriere/data/.m/93i386-quarterly/ref/.p/pool MASTERNAME=93i386-quarterly SCRIPTPREFIX=/usr/local/share/poudriere USER=root HOME=/root POUDRIERE_VERSION=3.1.9 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh LIBEXECPREFIX=/usr/local/libexec/poudriere LOCALBASE=/usr/local PACKAGE_BUILDING=yes ---End Environment--- ---Begin OPTIONS List--- ===> The following configuration options are available for hadoop2-2.7.1: EXAMPLES=on: Build and/or install examples ===> Use 'make config' to modify these settings ---End OPTIONS List--- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- HADOOP_PROTOC_PATH=/usr/local/protobuf25/bin/protoc XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" NO_PIE=yes SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local LIBDIR="/usr/lib" CC="cc" CFLAGS="-O2 -pipe -fno-strict-aliasing" CPP="cpp" CPPFLAGS="" LDFLAGS="" LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fno-strict-aliasing" MANPREFIX="/usr/local" BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 444" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- PORTVERSION="2.7.1" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVASHAREDIR="share/java" JAVAJARDIR="share/java/classes" OSREL=9.3 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local PORTDOCS="" PORTEXAMPLES="" LIB32DIR=lib DOCSDIR="share/doc/hadoop" EXAMPLESDIR="share/examples/hadoop" DATADIR="share/hadoop" WWWDIR="www/hadoop" ETCDIR="etc/hadoop" --End PLIST_SUB-- --SUB_LIST-- HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVA_HOME="/usr/local/openjdk7" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" JAVASHAREDIR="/usr/local/share/java" JAVAJARDIR="/usr/local/share/java/classes" JAVALIBDIR="/usr/local/share/java/classes" JAVA_VERSION="1.7" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/hadoop DOCSDIR=/usr/local/share/doc/hadoop EXAMPLESDIR=/usr/local/share/examples/hadoop WWWDIR=/usr/local/www/hadoop ETCDIR=/usr/local/etc/hadoop --End SUB_LIST-- ---Begin make.conf--- MACHINE=i386 MACHINE_ARCH=i386 ARCH=${MACHINE_ARCH} USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles #### /usr/local/etc/poudriere.d/make.conf #### DISABLE_MAKE_JOBS=poudriere ---End make.conf--- =======================<phase: check-sanity >============================ ===> License APACHE20 accepted by the user =========================================================================== =======================<phase: pkg-depends >============================ ===> hadoop2-2.7.1 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-1.6.1.txz [93i386-quarterly-job-03] Installing pkg-1.6.1... [93i386-quarterly-job-03] Extracting pkg-1.6.1: .......... done Message from pkg-1.6.1: If you are upgrading from the old package format, first run: # pkg2ng ===> hadoop2-2.7.1 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of hadoop2-2.7.1 =========================================================================== =======================<phase: fetch-depends >============================ =========================================================================== =======================<phase: fetch >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.7.1 for building =========================================================================== =======================<phase: checksum >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.7.1 for building => SHA256 Checksum OK for hadoop/hadoop-2.7.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.7.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.41.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =======================<phase: extract-depends>============================ =========================================================================== =======================<phase: extract >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.7.1 for building ===> Extracting for hadoop2-2.7.1 => SHA256 Checksum OK for hadoop/hadoop-2.7.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.7.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.41.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =======================<phase: patch-depends >============================ =========================================================================== =======================<phase: patch >============================ ===> Patching for hadoop2-2.7.1 ===> Applying FreeBSD patches for hadoop2-2.7.1 /usr/bin/sed -i.bak -e "s#/bin/bash#/usr/local/bin/bash#" /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DefaultContainerExecutor.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/launcher/ContainerLaunch.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Shell.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh =========================================================================== =======================<phase: build-depends >============================ ===> hadoop2-2.7.1 depends on file: /usr/local/share/java/maven3/bin/mvn - not found ===> Installing existing package /packages/All/maven3-3.0.5.txz [93i386-quarterly-job-03] Installing maven3-3.0.5... [93i386-quarterly-job-03] `-- Installing maven-wrapper-1_2... [93i386-quarterly-job-03] `-- Extracting maven-wrapper-1_2: . done [93i386-quarterly-job-03] `-- Installing openjdk-7.80.15_1,1... [93i386-quarterly-job-03] | `-- Installing libXt-1.1.4_3,1... [93i386-quarterly-job-03] | | `-- Installing xproto-7.0.27... [93i386-quarterly-job-03] | | `-- Extracting xproto-7.0.27: .......... done [93i386-quarterly-job-03] | | `-- Installing libSM-1.2.2_3,1... [93i386-quarterly-job-03] | | `-- Installing libICE-1.0.9_1,1... [93i386-quarterly-job-03] | | `-- Extracting libICE-1.0.9_1,1: .......... done [93i386-quarterly-job-03] | | `-- Extracting libSM-1.2.2_3,1: .......... done [93i386-quarterly-job-03] | | `-- Installing libX11-1.6.2_3,1... [93i386-quarterly-job-03] | | `-- Installing kbproto-1.0.6... [93i386-quarterly-job-03] | | `-- Extracting kbproto-1.0.6: .......... done [93i386-quarterly-job-03] | | `-- Installing libXdmcp-1.1.2... [93i386-quarterly-job-03] | | `-- Extracting libXdmcp-1.1.2: ......... done [93i386-quarterly-job-03] | | `-- Installing libxcb-1.11_1... [93i386-quarterly-job-03] | | | `-- Installing libxml2-2.9.2_3... [93i386-quarterly-job-03] | | | `-- Installing libiconv-1.14_8... [93i386-quarterly-job-03] | | | `-- Extracting libiconv-1.14_8: .......... done [93i386-quarterly-job-03] | | | `-- Extracting libxml2-2.9.2_3: .......... done [93i386-quarterly-job-03] | | | `-- Installing libpthread-stubs-0.3_6... [93i386-quarterly-job-03] | | | `-- Extracting libpthread-stubs-0.3_6: ..... done [93i386-quarterly-job-03] | | | `-- Installing libXau-1.0.8_3... [93i386-quarterly-job-03] | | | `-- Extracting libXau-1.0.8_3: .......... done [93i386-quarterly-job-03] | | `-- Extracting libxcb-1.11_1: .......... done [93i386-quarterly-job-03] | | `-- Extracting libX11-1.6.2_3,1: .......... done [93i386-quarterly-job-03] | `-- Extracting libXt-1.1.4_3,1: .......... done [93i386-quarterly-job-03] | `-- Installing libXtst-1.2.2_3... [93i386-quarterly-job-03] | | `-- Installing libXext-1.3.3_1,1... [93i386-quarterly-job-03] | | `-- Installing xextproto-7.3.0... [93i386-quarterly-job-03] | | `-- Extracting xextproto-7.3.0: .......... done [93i386-quarterly-job-03] | | `-- Extracting libXext-1.3.3_1,1: .......... done [93i386-quarterly-job-03] | | `-- Installing inputproto-2.3.1... [93i386-quarterly-job-03] | | `-- Extracting inputproto-2.3.1: ..... done [93i386-quarterly-job-03] | | `-- Installing libXi-1.7.4_1,1... [93i386-quarterly-job-03] | | `-- Installing libXfixes-5.0.1_3... <snip> Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-frame.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-noframe.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/index.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/help-doc.html... [INFO] Building jar: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/hadoop-auth-examples-2.7.1-javadoc.jar [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Apache Hadoop Common 2.7.1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-os) @ hadoop-common --- [INFO] [INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) @ hadoop-common --- [INFO] Executing tasks main: [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/test-dir [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/test/data [INFO] Executed tasks [INFO] [INFO] --- hadoop-maven-plugins:2.7.1:protoc (compile-protoc) @ hadoop-common --- [INFO] [INFO] --- hadoop-maven-plugins:2.7.1:version-info (version-info) @ hadoop-common --- [WARNING] [svn, info] failed: java.io.IOException: Cannot run program "svn": error=2, No such file or directory [WARNING] [git, branch] failed: java.io.IOException: Cannot run program "git": error=2, No such file or directory [INFO] SCM: NONE [INFO] Computed MD5: 4e4e024af7121717eca2e5756702dc5 [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hadoop-common --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 7 resources [INFO] Copying 1 resource [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ hadoop-common --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 849 source files to /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/classes [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[46,16] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[50,19] sun.net.dns.ResolverConfiguration is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[51,20] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[25,16] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[21,16] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[22,16] sun.misc.SignalHandler is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[44,43] sun.misc.SignalHandler is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[331,39] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[332,17] sun.misc.Cleaner is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[333,25] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[687,17] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[689,7] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[689,24] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[507,9] sun.net.dns.ResolverConfiguration is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[525,11] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[527,21] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[529,18] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[531,21] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[136,20] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[142,22] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[147,29] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CryptoStreamUtils.java:[39,37] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CryptoStreamUtils.java:[40,21] sun.misc.Cleaner is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CryptoStreamUtils.java:[41,23] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[46,19] sun.misc.SignalHandler is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[50,39] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[50,21] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[59,24] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/LdapGroupsMapping.java:[265,28] com.sun.jndi.ldap.LdapCtxFactory is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/compiler/ant/RccTask.java: Some input files use or override a deprecated API. [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/compiler/ant/RccTask.java: Recompile with -Xlint:deprecation for details. [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java: Some input files use unchecked or unsafe operations. [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java: Recompile with -Xlint:unchecked for details. [INFO] [INFO] --- native-maven-plugin:1.0-alpha-8:javah (default) @ hadoop-common --- [INFO] /bin/sh -c cd /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common && /usr/local/openjdk7/bin/javah -d /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/native/javah -classpath /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/classes:/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-annotations/target/hadoop-annotations-2.7.1.jar:/usr/local/openjdk7/jre/../lib/tools.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/guava/guava/11.0.2/guava-11.0.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-httpclient/commons-httpclient/3.1/commons-h ttpclient-3.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-codec/commons-codec/1.4/commons-codec-1.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-io/commons-io/2.4/commons-io-2.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-net/commons-net/3.1/commons-net-3.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/servlet/servlet-api/2.5/servlet-api-2.5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/mortbay/jetty/jetty/6.1.26/jetty-6.1.26.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/sun/jersey/jersey-json/1.9/jersey-json-1.9.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/wrkdirs/usr/ports/devel/hadoop2/wor k/m2/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb! -impl-2.2.3-1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/activation/activation/1.1/activation-1.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/sun/jersey/jersey-server/1.9/jersey-server-1.9.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/asm/asm/3.2/asm-3.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/log4j/log4j/1.2.17/log4j-1.2.17.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/httpcompo nents/httpclient/4.2.5/httpclient-4.2.5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/httpcomponents/httpcore/4.2.5/httpcore-4.2.5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/slf4j/slf4j-api/1.7.10/slf4j-api-1.7.10.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13. jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2! /org/code! haus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/avro/avro/1.7.4/avro-1.7.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/xerial/snappy/snappy-java/1.0.4.1/snappy-java-1.0.4.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/code/gson/gson/2.2.4/gson-2.2.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth/target/hadoop-auth-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/server/apacheds-kerberos-codec/2.0.0-M15/apacheds-kerberos-codec-2.0.0-M15.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/server/apacheds-i18n/2.0.0-M15/apacheds-i18n-2.0.0-M15.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api -asn1-api/1.0.0-M20/api-asn1-api-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-util/1.0.0-M20/api-util-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/curator/curator-framework/2.7.1/curator-framework-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-i18n/1.0.0-M20/api-i18n-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-ldap-model/1.0.0-M20/api-ldap-model-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/mina/mina-core/2.0.0-M5/mina-core-2.0.0-M5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/net/sf/ehcache/ehcache-core/2.4.4/ehcache-core-2.4.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/antlr/antlr/2.7.7/antlr-2.7.7.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-asn1-ber/1.0.0-M20/api-asn1-ber-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/jcraft/jsch/0.1.42/jsch-0.1.42.jar:/wrkdirs/usr/ports/devel/hadoop2/w ork/m2/org/apache/curator/curator-client/2.7! .1/curato! r-client-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/curator/curator-recipes/2.7.1/curator-recipes-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/code/findbugs/jsr305/3.0.0/jsr305-3.0.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/htrace/htrace-core/3.1.0-incubating/htrace-core-3.1.0-incubating.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/io/netty/netty/3.6.2.Final/netty-3.6.2.Final.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/tukaani/xz/1.0/xz-1.0.jar org.apache.hadoop.io.compress.zlib.ZlibCompressor org.apache.hadoop.io.compress.zlib.ZlibDecompressor org.apache.hadoop.io.compress.bzip2.Bzip2Compressor org.apache.hadoop.io.compress.bzip2.Bzip2Decompressor org.apache.hadoop.security.JniBasedUnixGroupsMapping org.apache.hadoop.io.na tiveio.NativeIO org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMapping org.apache.hadoop.io.compress.snappy.SnappyCompressor org.apache.hadoop.io.compress.snappy.SnappyDecompressor org.apache.hadoop.io.compress.lz4.Lz4Compressor org.apache.hadoop.io.compress.lz4.Lz4Decompressor org.apache.hadoop.crypto.OpensslCipher org.apache.hadoop.crypto.random.OpensslSecureRandom org.apache.hadoop.util.NativeCrc32 org.apache.hadoop.net.unix.DomainSocket org.apache.hadoop.net.unix.DomainSocketWatcher [INFO] [INFO] --- maven-antrun-plugin:1.7:run (make) @ hadoop-common --- [INFO] Executing tasks main: [exec] -- The C compiler identification is GNU 4.2.1 [exec] -- The CXX compiler identification is GNU 4.2.1 [exec] -- Check for working C compiler: /usr/bin/cc [exec] -- Check for working C compiler: /usr/bin/cc -- works [exec] -- Detecting C compiler ABI info [exec] -- Detecting C compiler ABI info - done [exec] -- Detecting C compile features [exec] -- Detecting C compile features - done [exec] -- Check for working CXX compiler: /usr/bin/c++ [exec] -- Check for working CXX compiler: /usr/bin/c++ -- works [exec] -- Detecting CXX compiler ABI info [exec] -- Detecting CXX compiler ABI info - done [exec] -- Detecting CXX compile features [exec] -- Detecting CXX compile features - done [exec] -- Found JNI: /usr/local/openjdk7/jre/lib/i386/libjawt.so [exec] -- Found ZLIB: /usr/lib/libz.so (found version "1.2.8") [exec] -- Looking for sync_file_range [exec] -- Looking for sync_file_range - not found [exec] -- Looking for posix_fadvise [exec] -- Looking for posix_fadvise - found [exec] -- Looking for dlopen in dl [exec] CUSTOM_OPENSSL_PREFIX = [exec] The OpenSSL library installed at /usr/lib/libcrypto.so is too old. You need a version at least new eno-- Looking for dlopen in dl - not found [exec] -- Performing Test HAS_NEW_ENOUGH_OPENSSL [exec] -- Performing Test HAS_NEW_ENOUGH_OPENSSL - Faugh to have EVP_aes_256_ctr. [exec] Cannot find a usable OpenSSL library. OPENSSL_LIBRARY=/usr/lib/libcrypto.so, OPENSSL_INCLUDE_DIR=/iled [exec] -- Configuring incomplete, errors occurred! [exec] See also "/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-usr/include, CUSTOM_OPENSSL_LIB=, CUSTOM_OPENSSL_PREFIX=, CUSTOM_OPENSSL_INCLUDE= [exec] CMake Error at CMakeLists.txt:207 (MESSAGE): [exec] project/hadoop-common/target/native/CMakeFiles/CMakeOutput.log". [exec] See also "/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-sr Terminating build because require.openssl was specified. [exec] [exec] [exec] c/hadoop-common-project/hadoop-common/target/native/CMakeFiles/CMakeError.log". [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop Main ................................ SUCCESS [1.316s] [INFO] Apache Hadoop Project POM ......................... SUCCESS [1.010s] [INFO] Apache Hadoop Annotations ......................... SUCCESS [3.033s] [INFO] Apache Hadoop Assemblies .......................... SUCCESS [0.211s] [INFO] Apache Hadoop Project Dist POM .................... SUCCESS [1.345s] [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS [2.694s] [INFO] Apache Hadoop MiniKDC ............................. SUCCESS [2.534s] [INFO] Apache Hadoop Auth ................................ SUCCESS [3.982s] [INFO] Apache Hadoop Auth Examples ....................... SUCCESS [3.228s] [INFO] Apache Hadoop Common .............................. FAILURE [17.607s] [INFO] Apache Hadoop NFS ................................. SKIPPED [INFO] Apache Hadoop KMS ................................. SKIPPED [INFO] Apache Hadoop Common Project ...................... SKIPPED [INFO] Apache Hadoop HDFS ................................ SKIPPED [INFO] Apache Hadoop HttpFS .............................. SKIPPED [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED [INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED [INFO] Apache Hadoop HDFS Project ........................ SKIPPED [INFO] hadoop-yarn ....................................... SKIPPED [INFO] hadoop-yarn-api ................................... SKIPPED [INFO] hadoop-yarn-common ................................ SKIPPED [INFO] hadoop-yarn-server ................................ SKIPPED [INFO] hadoop-yarn-server-common ......................... SKIPPED [INFO] hadoop-yarn-server-nodemanager .................... SKIPPED [INFO] hadoop-yarn-server-web-proxy ...................... SKIPPED [INFO] hadoop-yarn-server-applicationhistoryservice ...... SKIPPED [INFO] hadoop-yarn-server-resourcemanager ................ SKIPPED [INFO] hadoop-yarn-server-tests .......................... SKIPPED [INFO] hadoop-yarn-client ................................ SKIPPED [INFO] hadoop-yarn-server-sharedcachemanager ............. SKIPPED [INFO] hadoop-yarn-applications .......................... SKIPPED [INFO] hadoop-yarn-applications-distributedshell ......... SKIPPED [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SKIPPED [INFO] hadoop-yarn-site .................................. SKIPPED [INFO] hadoop-yarn-registry .............................. SKIPPED [INFO] hadoop-yarn-project ............................... SKIPPED [INFO] hadoop-mapreduce-client ........................... SKIPPED [INFO] hadoop-mapreduce-client-core ...................... SKIPPED [INFO] hadoop-mapreduce-client-common .................... SKIPPED [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED [INFO] hadoop-mapreduce-client-app ....................... SKIPPED [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED [INFO] hadoop-mapreduce-client-hs-plugins ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] hadoop-mapreduce .................................. SKIPPED [INFO] Apache Hadoop MapReduce Streaming ................. SKIPPED [INFO] Apache Hadoop Distributed Copy .................... SKIPPED [INFO] Apache Hadoop Archives ............................ SKIPPED [INFO] Apache Hadoop Rumen ............................... SKIPPED [INFO] Apache Hadoop Gridmix ............................. SKIPPED [INFO] Apache Hadoop Data Join ........................... SKIPPED [INFO] Apache Hadoop Ant Tasks ........................... SKIPPED [INFO] Apache Hadoop Extras .............................. SKIPPED [INFO] Apache Hadoop Pipes ............................... SKIPPED [INFO] Apache Hadoop OpenStack support ................... SKIPPED [INFO] Apache Hadoop Amazon Web Services support ......... SKIPPED [INFO] Apache Hadoop Azure support ....................... SKIPPED [INFO] Apache Hadoop Client .............................. SKIPPED [INFO] Apache Hadoop Mini-Cluster ........................ SKIPPED [INFO] Apache Hadoop Scheduler Load Simulator ............ SKIPPED [INFO] Apache Hadoop Tools Dist .......................... SKIPPED [INFO] Apache Hadoop Tools ............................... SKIPPED [INFO] Apache Hadoop Distribution ........................ SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 39.352s [INFO] Finished at: Tue Nov 03 01:47:15 GMT 2015 [INFO] Final Memory: 60M/226M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-antrun-plugin:1.7:run (make) on project hadoop-common: An Ant BuildException has occured: exec returned: 1 [ERROR] around Ant part ...<exec dir="/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/native" executable="cmake" failonerror="true">... @ 4:158 in /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/antrun/build-main.xml [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-common *** [do-build] Error code 1 Stop in /usr/ports/devel/hadoop2. _______________________________________________ freebsd-pkg-fallout@freebsd.org mailing list https://lists.freebsd.org/mailman/listinfo/freebsd-pkg-fallout To unsubscribe, send any mail to "freebsd-pkg-fallout-unsubscr...@freebsd.org"