没有log,只有err和out,out为空 ________________________________ zjfpla...@hotmail.com
发件人: tison<mailto:wander4...@gmail.com> 发送时间: 2020-01-24 10:03 收件人: user-zh<mailto:user-zh@flink.apache.org> 抄送: zhisheng2018<mailto:zhisheng2...@gmail.com> 主题: Re: Re: flink on yarn任务启动报错 The assigned slot container_e10_1579661300080_0005_01_000002_0 was removed. 你上面的是 taskmanager.err,需要的是 taskmanager.log Best, tison. 郑 洁锋 <zjfpla...@hotmail.com> 于2020年1月23日周四 下午10:22写道: > 之前挂过 后面启动的时候 是checkpoints的文件丢了? 你是这个意思吗? > > ________________________________ > zjfpla...@hotmail.com > > 发件人: zhisheng<mailto:zhisheng2...@gmail.com> > 发送时间: 2020-01-22 16:45 > 收件人: user-zh<mailto:user-zh@flink.apache.org> > 主题: Re: flink on yarn任务启动报错 The assigned slot > container_e10_1579661300080_0005_01_000002_0 was removed. > 应该是你作业之前挂过了 > > 郑 洁锋 <zjfpla...@hotmail.com> 于2020年1月22日周三 上午11:16写道: > > > 大家好, > > flink on yarn任务启动时,发现报错了The assigned slot > > container_e10_1579661300080_0005_01_000002_0 was removed. > > 环境:flink1.8.1,cdh5.14.2,kafka0.10,jdk1.8.0_241 > > > > flink版本为1.8.1,yarn上的日志: > > > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: > > > -------------------------------------------------------------------------------- > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Starting > > YarnJobClusterEntrypoint (Version: <unknown>, Rev:7297bac, > Date:24.06.2019 > > @ 23:04:28 CST) > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: OS current user: > > cloudera-scm > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Current > > Hadoop/Kerberos user: root > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: JVM: Java > > HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.241-b07 > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Maximum heap size: > > 406 MiBytes > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: JAVA_HOME: > > /usr/java/default > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Hadoop version: > 2.6.5 > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: JVM Options: > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: -Xms424m > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: -Xmx424m > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Program Arguments: > > (none) > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Classpath: > > > core-1.8.0_release.jar:flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:kafka10-source-1.8.0_release.jar:log4j-1.2.17.jar:mysql-all-side-1.8.0_release.jar:mysql-sink-1.8.0_release.jar:slf4j-log4j12-1.7.15.jar:sql.launcher-1.0-SNAPSHOT.jar:flink.jar:flink-conf.yaml:job.graph::/etc/hadoop/conf.cloudera.yarn:/run/cloudera-scm-agent/process/1129-yarn-NODEMANAGER:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-annotations.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-auth.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-aws.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-azure-datalake.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-common-tests.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-common.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-nfs.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-nfs-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-common-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-common-2.6.0-cdh5.14.2-tests.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-azure-datalake-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-aws-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-auth-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/hadoop-annotations-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-format.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-format-sources.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-format-javadoc.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-tools.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-thrift.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-test-hadoop2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-scrooge_2.10.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-scala_2.10.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-protobuf.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-pig.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-pig-bundle.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-jackson.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-hadoop.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-hadoop-bundle.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-generator.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-encoding.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-common.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-column.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-cascading.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/parquet-avro.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-digester-1.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-lang-2.6.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/curator-client-2.7.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-logging-1.1.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/hamcrest-core-1.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-net-3.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-configuration-1.6.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jackson-xc-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/httpclient-4.2.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-math3-3.1.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jersey-core-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/guava-11.0.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-io-2.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jsp-api-2.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jsch-0.1.42.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jetty-6.1.26.cloudera.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jersey-json-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/htrace-core4-4.0.1-incubating.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-httpclient-3.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/gson-2.2.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-el-1.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/zookeeper.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/avro.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/hue-plugins-3.9.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/slf4j-api-1.7.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/netty-3.10.5.Final.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/logredactor-1.0.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/junit-4.11.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jetty-util-6.1.26.cloudera.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jets3t-0.9.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jaxb-api-2.2.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/curator-recipes-2.7.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-compress-1.4.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/xmlenc-0.52.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/servlet-api-2.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/paranamer-2.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-collections-3.2.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-codec-1.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-cli-1.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/commons-beanutils-1.9.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/azure-data-lake-store-sdk-2.2.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/aws-java-sdk-bundle-1.11.134.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/asm-3.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/api-util-1.0.0-M20.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/activation-1.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/slf4j-log4j12.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/xz-1.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/stax-api-1.0-2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/protobuf-java-2.5.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/mockito-all-1.8.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/log4j-1.2.17.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jsr305-3.0.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jettison-1.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/jersey-server-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/java-xmlbuilder-0.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/httpcore-4.2.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop/lib/curator-framework-2.7.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/hadoop-hdfs-nfs.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/hadoop-hdfs-tests.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/hadoop-hdfs.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/hadoop-hdfs-nfs-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/hadoop-hdfs-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/hadoop-hdfs-2.6.0-cdh5.14.2-tests.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/xmlenc-0.52.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/xml-apis-1.3.04.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/servlet-api-2.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/protobuf-java-2.5.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/netty-3.10.5.Final.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/log4j-1.2.17.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/leveldbjni-all-1.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jsr305-3.0.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jsp-api-2.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jetty-util-6.1.26.cloudera.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jetty-6.1.26.cloudera.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jersey-server-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jersey-core-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jasper-runtime-5.5.23.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jackson-mapper-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/jackson-core-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/htrace-core4-4.0.1-incubating.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/guava-11.0.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-logging-1.1.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-lang-2.6.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-io-2.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-el-1.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-codec-1.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/commons-cli-1.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-hdfs/lib/asm-3.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-api.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-applications-unmanaged-am-launcher.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-client.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-common.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-registry.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-applicationhistoryservice.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-common.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-nodemanager.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-resourcemanager.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-tests.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-web-proxy.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-web-proxy-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-tests-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-resourcemanager-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-nodemanager-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-common-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-server-applicationhistoryservice-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-registry-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-common-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-client-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/hadoop-yarn-api-2.6.0-cdh5.14.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/spark-yarn-shuffle.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/xz-1.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/stax-api-1.0-2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/servlet-api-2.5.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/protobuf-java-2.5.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/log4j-1.2.17.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jsr305-3.0.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jline-2.11.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jetty-util-6.1.26.cloudera.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jetty-6.1.26.cloudera.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jettison-1.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jersey-server-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jersey-json-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jersey-guice-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jersey-core-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jersey-client-1.9.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jaxb-api-2.2.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/javax.inject-1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jackson-xc-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jackson-mapper-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jackson-jaxrs-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/jackson-core-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/guice-servlet-3.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/guice-3.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/guava-11.0.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-logging-1.1.3.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-lang-2.6.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-io-2.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-compress-1.4.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-collections-3.2.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-codec-1.4.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/commons-cli-1.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/asm-3.2.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/aopalliance-1.0.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/activation-1.1.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/zookeeper.jar:/opt/cloudera/parcels/CDH-5.14.2-1.cdh5.14.2.p0.3/lib/hadoop-yarn/lib/spark-1.6.0-cdh5.14.2-yarn-shuffle.jar > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: > > > -------------------------------------------------------------------------------- > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Registered UNIX > > signal handlers for [TERM, HUP, INT] > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: YARN daemon is > > running as: root Yarn client user obtainer: root > > 20/01/22 11:07:53 INFO configuration.GlobalConfiguration: Loading > > configuration property: time.characteristic, EventTime > > 20/01/22 11:07:53 INFO configuration.GlobalConfiguration: Loading > > configuration property: internal.cluster.execution-mode, DETACHED > > 20/01/22 11:07:53 INFO configuration.GlobalConfiguration: Loading > > configuration property: high-availability.cluster-id, > > application_1579661300080_0005 > > 20/01/22 11:07:53 INFO configuration.GlobalConfiguration: Loading > > configuration property: taskmanager.numberOfTaskSlots, 1 > > 20/01/22 11:07:53 INFO configuration.GlobalConfiguration: Loading > > configuration property: taskmanager.heap.size, 1024m > > 20/01/22 11:07:53 WARN configuration.Configuration: Config uses > deprecated > > configuration key 'web.port' instead of proper key 'rest.bind-port' > > 20/01/22 11:07:53 INFO clusterframework.BootstrapTools: Setting > > directories for temporary files to: > > /yarn/nm/usercache/root/appcache/application_1579661300080_0005 > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Starting > > YarnJobClusterEntrypoint. > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Install default > > filesystem. > > 20/01/22 11:07:53 INFO modules.HadoopModule: Hadoop user set to root > > (auth:SIMPLE) > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Initializing cluster > > services. > > 20/01/22 11:07:53 INFO akka.AkkaRpcServiceUtils: Trying to start actor > > system at uf30-3:0 > > 20/01/22 11:07:54 INFO slf4j.Slf4jLogger: Slf4jLogger started > > 20/01/22 11:07:54 INFO remote.Remoting: Starting remoting > > 20/01/22 11:07:54 INFO remote.Remoting: Remoting started; listening on > > addresses :[akka.tcp://flink@uf30-3:61028] > > 20/01/22 11:07:54 INFO akka.AkkaRpcServiceUtils: Actor system started at > > akka.tcp://flink@uf30-3:61028 > > 20/01/22 11:07:54 WARN configuration.Configuration: Config uses > deprecated > > configuration key 'web.port' instead of proper key 'rest.port' > > 20/01/22 11:07:54 INFO blob.BlobServer: Created BLOB server storage > > directory > > > /yarn/nm/usercache/root/appcache/application_1579661300080_0005/blobStore-bda7ba98-c1ee-4ad7-b04e-22b2fa1c6268 > > 20/01/22 11:07:54 INFO blob.BlobServer: Started BLOB server at > > 0.0.0.0:15790 - max concurrent requests: 50 - max backlog: 1000 > > 20/01/22 11:07:54 INFO metrics.MetricRegistryImpl: No metrics reporter > > configured, no metrics will be exposed/reported. > > 20/01/22 11:07:54 INFO entrypoint.ClusterEntrypoint: Trying to start > actor > > system at uf30-3:0 > > 20/01/22 11:07:54 INFO slf4j.Slf4jLogger: Slf4jLogger started > > 20/01/22 11:07:54 INFO remote.Remoting: Starting remoting > > 20/01/22 11:07:54 INFO remote.Remoting: Remoting started; listening on > > addresses :[akka.tcp://flink-metrics@uf30-3:26151] > > 20/01/22 11:07:54 INFO entrypoint.ClusterEntrypoint: Actor system started > > at akka.tcp://flink-metrics@uf30-3:26151 > > 20/01/22 11:07:54 INFO blob.TransientBlobCache: Created BLOB cache > storage > > directory > > > /yarn/nm/usercache/root/appcache/application_1579661300080_0005/blobStore-cc2030ec-c73c-4383-a4df-30358745cd17 > > 20/01/22 11:07:54 WARN configuration.Configuration: Config uses > deprecated > > configuration key 'web.port' instead of proper key 'rest.bind-port' > > 20/01/22 11:07:54 WARN jobmaster.MiniDispatcherRestEndpoint: Upload > > directory > > /tmp/flink-web-383e26d9-e789-4756-8f69-1b03462e27f6/flink-web-upload does > > not exist, or has been deleted externally. Previously uploaded files are > no > > longer available. > > 20/01/22 11:07:54 INFO jobmaster.MiniDispatcherRestEndpoint: Created > > directory > > /tmp/flink-web-383e26d9-e789-4756-8f69-1b03462e27f6/flink-web-upload for > > file uploads. > > 20/01/22 11:07:54 INFO jobmaster.MiniDispatcherRestEndpoint: Starting > rest > > endpoint. > > 20/01/22 11:07:54 WARN webmonitor.WebMonitorUtils: Log file environment > > variable 'log.file' is not set. > > 20/01/22 11:07:54 WARN webmonitor.WebMonitorUtils: JobManager log files > > are unavailable in the web dashboard. Log file location not found in > > environment variable 'log.file' or configuration key 'Key: > 'web.log.path' , > > default: null (fallback keys: [{key=jobmanager.web.log.path, > > isDeprecated=true}])'. > > 20/01/22 11:07:54 INFO jobmaster.MiniDispatcherRestEndpoint: Rest > endpoint > > listening at uf30-3:17001 > > 20/01/22 11:07:54 INFO jobmaster.MiniDispatcherRestEndpoint: > > http://uf30-3:17001 was granted leadership with > > leaderSessionID=00000000-0000-0000-0000-000000000000 > > 20/01/22 11:07:54 INFO jobmaster.MiniDispatcherRestEndpoint: Web frontend > > listening at http://uf30-3:17001. > > 20/01/22 11:07:54 INFO akka.AkkaRpcService: Starting RPC endpoint for > > org.apache.flink.yarn.YarnResourceManager at > > akka://flink/user/resourcemanager . > > 20/01/22 11:07:54 INFO akka.AkkaRpcService: Starting RPC endpoint for > > org.apache.flink.runtime.dispatcher.MiniDispatcher at > > akka://flink/user/dispatcher . > > 20/01/22 11:07:54 INFO dispatcher.MiniDispatcher: Dispatcher > > akka.tcp://flink@uf30-3:61028/user/dispatcher was granted leadership > with > > fencing token 00000000-0000-0000-0000-000000000000 > > 20/01/22 11:07:54 INFO dispatcher.MiniDispatcher: Recovering all > persisted > > jobs. > > 20/01/22 11:07:54 INFO akka.AkkaRpcService: Starting RPC endpoint for > > org.apache.flink.runtime.jobmaster.JobMaster at > > akka://flink/user/jobmanager_0 . > > 20/01/22 11:07:54 INFO jobmaster.JobMaster: Initializing job xctest > > (e1b2df526572dd9e93be25763519ee35). > > 20/01/22 11:07:54 INFO jobmaster.JobMaster: Using restart strategy > > FailureRateRestartStrategy(failuresInterval=360000 msdelayInterval=10000 > > msmaxFailuresPerInterval=3) for xctest > (e1b2df526572dd9e93be25763519ee35). > > 20/01/22 11:07:54 INFO executiongraph.ExecutionGraph: Job recovers via > > failover strategy: full graph restart > > 20/01/22 11:07:54 INFO client.ConfiguredRMFailoverProxyProvider: Failing > > over to rm225 > > 20/01/22 11:07:54 INFO jobmaster.JobMaster: Running initialization on > > master for job xctest (e1b2df526572dd9e93be25763519ee35). > > 20/01/22 11:07:54 INFO jobmaster.JobMaster: Successfully ran > > initialization on master in 0 ms. > > 20/01/22 11:07:54 INFO jobmaster.JobMaster: No state backend has been > > configured, using default (Memory / JobManager) MemoryStateBackend (data > in > > heap memory / checkpoints to JobManager) (checkpoints: 'null', > savepoints: > > 'null', asynchronous: TRUE, maxStateSize: 5242880) > > 20/01/22 11:07:54 INFO jobmaster.JobManagerRunner: JobManager runner for > > job xctest (e1b2df526572dd9e93be25763519ee35) was granted leadership with > > session id 00000000-0000-0000-0000-000000000000 at > akka.tcp://flink@uf30-3 > > :61028/user/jobmanager_0. > > 20/01/22 11:07:54 INFO jobmaster.JobMaster: Starting execution of job > > xctest (e1b2df526572dd9e93be25763519ee35) under job master id > > 00000000000000000000000000000000. > > 20/01/22 11:07:54 INFO executiongraph.ExecutionGraph: Job xctest > > (e1b2df526572dd9e93be25763519ee35) switched from state CREATED to > RUNNING. > > 20/01/22 11:07:54 INFO executiongraph.ExecutionGraph: Source: > > testFlink_kafkaTable -> Map -> to: Tuple2 -> Map -> from: (a_v1, a_v2, > > a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME) -> select: (a_v1, > > a_v2, a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME(PROCTIME) AS > > PROCTIME) -> to: Tuple2 -> Map -> Flat Map -> Map -> select: (a_v1 AS > r_v1, > > a_v2 AS r_v2, a_v3 AS r_v3, a_i1 AS r_i1, a_i2 AS r_i2, curr_time, > > msg_index, send_time) -> to: Tuple2 -> Sink: MyResult (1/1) > > (083db3e18b24bc9329931aa39bf3109e) switched from CREATED to SCHEDULED. > > 20/01/22 11:07:55 INFO slotpool.SlotPoolImpl: Cannot serve slot request, > > no ResourceManager connected. Adding as pending request > > [SlotRequestId{ff60413f2edc00a134b584d1a5953d77}] > > 20/01/22 11:07:55 INFO jobmaster.JobMaster: Connecting to ResourceManager > > akka.tcp://flink@uf30-3 > > :61028/user/resourcemanager(00000000000000000000000000000000) > > 20/01/22 11:07:55 INFO yarn.YarnResourceManager: Recovered 0 containers > > from previous attempts ([]). > > 20/01/22 11:07:55 INFO impl.ContainerManagementProtocolProxy: > > yarn.client.max-cached-nodemanagers-proxies : 0 > > 20/01/22 11:07:55 INFO yarn.YarnResourceManager: ResourceManager > > akka.tcp://flink@uf30-3:61028/user/resourcemanager was granted > leadership > > with fencing token 00000000000000000000000000000000 > > 20/01/22 11:07:55 INFO slotmanager.SlotManager: Starting the SlotManager. > > 20/01/22 11:07:55 INFO jobmaster.JobMaster: Resolved ResourceManager > > address, beginning registration > > 20/01/22 11:07:55 INFO jobmaster.JobMaster: Registration at > > ResourceManager attempt 1 (timeout=100ms) > > 20/01/22 11:07:55 INFO yarn.YarnResourceManager: Registering job manager > > 00000000000000000000000000000...@akka.tcp://flink@uf30-3 > :61028/user/jobmanager_0 > > for job e1b2df526572dd9e93be25763519ee35. > > 20/01/22 11:07:55 INFO yarn.YarnResourceManager: Registered job manager > > 00000000000000000000000000000...@akka.tcp://flink@uf30-3 > :61028/user/jobmanager_0 > > for job e1b2df526572dd9e93be25763519ee35. > > 20/01/22 11:07:55 INFO jobmaster.JobMaster: JobManager successfully > > registered at ResourceManager, leader id: > 00000000000000000000000000000000. > > 20/01/22 11:07:55 INFO slotpool.SlotPoolImpl: Requesting new slot > > [SlotRequestId{ff60413f2edc00a134b584d1a5953d77}] and profile > > ResourceProfile{cpuCores=-1.0, heapMemoryInMB=-1, directMemoryInMB=0, > > nativeMemoryInMB=0, networkMemoryInMB=0} from resource manager. > > 20/01/22 11:07:55 INFO yarn.YarnResourceManager: Request slot with > profile > > ResourceProfile{cpuCores=-1.0, heapMemoryInMB=-1, directMemoryInMB=0, > > nativeMemoryInMB=0, networkMemoryInMB=0} for job > > e1b2df526572dd9e93be25763519ee35 with allocation id > > 2394a48465851f57cb3592402df11112. > > 20/01/22 11:07:55 INFO yarn.YarnResourceManager: Requesting new > > TaskExecutor container with resources <memory:1024, vCores:1>. Number > > pending requests 1. > > 20/01/22 11:07:56 INFO impl.AMRMClientImpl: Received new token for : > > uf30-3:8041 > > 20/01/22 11:07:56 INFO yarn.YarnResourceManager: Received new container: > > container_e10_1579661300080_0005_01_000002 - Remaining pending container > > requests: 1 > > 20/01/22 11:07:56 INFO yarn.YarnResourceManager: Removing container > > request Capability[<memory:1024, vCores:1>]Priority[1]. Pending container > > requests 0. > > 20/01/22 11:07:57 INFO yarn.YarnResourceManager: Creating container > launch > > context for TaskManagers > > 20/01/22 11:07:57 INFO yarn.YarnResourceManager: Starting TaskManagers > > 20/01/22 11:07:57 INFO impl.ContainerManagementProtocolProxy: Opening > > proxy : uf30-3:8041 > > 20/01/22 11:07:59 INFO yarn.YarnResourceManager: Registering TaskManager > > with ResourceID container_e10_1579661300080_0005_01_000002 > > (akka.tcp://flink@uf30-3:25536/user/taskmanager_0) at ResourceManager > > 20/01/22 11:07:59 INFO executiongraph.ExecutionGraph: Source: > > testFlink_kafkaTable -> Map -> to: Tuple2 -> Map -> from: (a_v1, a_v2, > > a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME) -> select: (a_v1, > > a_v2, a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME(PROCTIME) AS > > PROCTIME) -> to: Tuple2 -> Map -> Flat Map -> Map -> select: (a_v1 AS > r_v1, > > a_v2 AS r_v2, a_v3 AS r_v3, a_i1 AS r_i1, a_i2 AS r_i2, curr_time, > > msg_index, send_time) -> to: Tuple2 -> Sink: MyResult (1/1) > > (083db3e18b24bc9329931aa39bf3109e) switched from SCHEDULED to DEPLOYING. > > 20/01/22 11:07:59 INFO executiongraph.ExecutionGraph: Deploying Source: > > testFlink_kafkaTable -> Map -> to: Tuple2 -> Map -> from: (a_v1, a_v2, > > a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME) -> select: (a_v1, > > a_v2, a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME(PROCTIME) AS > > PROCTIME) -> to: Tuple2 -> Map -> Flat Map -> Map -> select: (a_v1 AS > r_v1, > > a_v2 AS r_v2, a_v3 AS r_v3, a_i1 AS r_i1, a_i2 AS r_i2, curr_time, > > msg_index, send_time) -> to: Tuple2 -> Sink: MyResult (1/1) (attempt #0) > to > > container_e10_1579661300080_0005_01_000002 @ uf30-3 (dataPort=58080) > > 20/01/22 11:07:59 INFO executiongraph.ExecutionGraph: Source: > > testFlink_kafkaTable -> Map -> to: Tuple2 -> Map -> from: (a_v1, a_v2, > > a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME) -> select: (a_v1, > > a_v2, a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME(PROCTIME) AS > > PROCTIME) -> to: Tuple2 -> Map -> Flat Map -> Map -> select: (a_v1 AS > r_v1, > > a_v2 AS r_v2, a_v3 AS r_v3, a_i1 AS r_i1, a_i2 AS r_i2, curr_time, > > msg_index, send_time) -> to: Tuple2 -> Sink: MyResult (1/1) > > (083db3e18b24bc9329931aa39bf3109e) switched from DEPLOYING to RUNNING. > > 20/01/22 11:08:49 INFO yarn.YarnResourceManager: The heartbeat of > > TaskManager with id container_e10_1579661300080_0005_01_000002 timed out. > > 20/01/22 11:08:49 INFO yarn.YarnResourceManager: Closing TaskExecutor > > connection container_e10_1579661300080_0005_01_000002 because: The > > heartbeat of TaskManager with id > > container_e10_1579661300080_0005_01_000002 timed out. > > 20/01/22 11:08:49 INFO executiongraph.ExecutionGraph: Source: > > testFlink_kafkaTable -> Map -> to: Tuple2 -> Map -> from: (a_v1, a_v2, > > a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME) -> select: (a_v1, > > a_v2, a_v3, a_i1, curr_time, msg_index, send_time, PROCTIME(PROCTIME) AS > > PROCTIME) -> to: Tuple2 -> Map -> Flat Map -> Map -> select: (a_v1 AS > r_v1, > > a_v2 AS r_v2, a_v3 AS r_v3, a_i1 AS r_i1, a_i2 AS r_i2, curr_time, > > msg_index, send_time) -> to: Tuple2 -> Sink: MyResult (1/1) > > (083db3e18b24bc9329931aa39bf3109e) switched from RUNNING to FAILED. > > org.apache.flink.util.FlinkException: The assigned slot > > container_e10_1579661300080_0005_01_000002_0 was removed. > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.removeSlot(SlotManager.java:899) > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.removeSlots(SlotManager.java:869) > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.internalUnregisterTaskManager(SlotManager.java:1080) > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.unregisterTaskManager(SlotManager.java:391) > > at > > > org.apache.flink.runtime.resourcemanager.ResourceManager.closeTaskManagerConnection(ResourceManager.java:845) > > at > > > org.apache.flink.runtime.resourcemanager.ResourceManager$TaskManagerHeartbeatListener.notifyHeartbeatTimeout(ResourceManager.java:1187) > > at > > > org.apache.flink.runtime.heartbeat.HeartbeatManagerImpl$HeartbeatMonitor.run(HeartbeatManagerImpl.java:318) > > at > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > at > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:392) > > at > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:185) > > at > > > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74) > > at > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.onReceive(AkkaRpcActor.java:147) > > at > > > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.onReceive(FencedAkkaRpcActor.java:40) > > at > > > akka.actor.UntypedActor$$anonfun$receive$1.applyOrElse(UntypedActor.scala:165) > > at akka.actor.Actor$class.aroundReceive(Actor.scala:502) > > at akka.actor.UntypedActor.aroundReceive(UntypedActor.scala:95) > > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526) > > at akka.actor.ActorCell.invoke(ActorCell.scala:495) > > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257) > > at akka.dispatch.Mailbox.run(Mailbox.scala:224) > > at akka.dispatch.Mailbox.exec(Mailbox.scala:234) > > at > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > > at > > > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at > > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > > at > > > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > 20/01/22 11:08:49 INFO executiongraph.ExecutionGraph: Job xctest > > (e1b2df526572dd9e93be25763519ee35) switched from state RUNNING to > FAILING. > > org.apache.flink.util.FlinkException: The assigned slot > > container_e10_1579661300080_0005_01_000002_0 was removed. > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.removeSlot(SlotManager.java:899) > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.removeSlots(SlotManager.java:869) > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.internalUnregisterTaskManager(SlotManager.java:1080) > > at > > > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager.unregisterTaskManager(SlotManager.java:391) > > at > > > org.apache.flink.runtime.resourcemanager.ResourceManager.closeTaskManagerConnection(ResourceManager.java:845) > > at > > > org.apache.flink.runtime.resourcemanager.ResourceManager$TaskManagerHeartbeatListener.notifyHeartbeatTimeout(ResourceManager.java:1187) > > at > > > org.apache.flink.runtime.heartbeat.HeartbeatManagerImpl$HeartbeatMonitor.run(HeartbeatManagerImpl.java:318) > > at > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > at > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:392) > > at > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:185) > > at > > > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74) > > at > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.onReceive(AkkaRpcActor.java:147) > > at > > > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.onReceive(FencedAkkaRpcActor.java:40) > > at > > > akka.actor.UntypedActor$$anonfun$receive$1.applyOrElse(UntypedActor.scala:165) > > at akka.actor.Actor$class.aroundReceive(Actor.scala:502) > > at akka.actor.UntypedActor.aroundReceive(UntypedActor.scala:95) > > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526) > > at akka.actor.ActorCell.invoke(ActorCell.scala:495) > > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257) > > at akka.dispatch.Mailbox.run(Mailbox.scala:224) > > at akka.dispatch.Mailbox.exec(Mailbox.scala:234) > > at > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > > at > > > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at > > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > > at > > > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > 20/01/22 11:08:49 INFO executiongraph.ExecutionGraph: Try to restart or > > fail the job xctest (e1b2df526572dd9e93be25763519ee35) if no longer > > possible. > > 20/01/22 11:08:49 INFO executiongraph.ExecutionGraph: Job xctest > > (e1b2df526572dd9e93be25763519ee35) switched from state FAILING to > > RESTARTING. > > 20/01/22 11:08:49 INFO executiongraph.ExecutionGraph: Restarting the job > > xctest (e1b2df526572dd9e93be25763519ee35). > > 20/01/22 11:08:59 INFO executiongraph.ExecutionGraph: Job xctest > > (e1b2df526572dd9e93be25763519ee35) switched from state RESTARTING to > > CREATED. > > 20/01/22 11:08:59 INFO executiongraph.ExecutionGraph: Job xctest > > (e1b2df526572dd9e93be25763519ee35) switched from state CREATED to > RUNNING. > > > > jobmanager.err: > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: > > > -------------------------------------------------------------------------------- > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Starting > > YarnJobClusterEntrypoint (Version: <unknown>, Rev:7297bac, > Date:24.06.2019 > > @ 23:04:28 CST) > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: OS current user: > > cloudera-scm > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Current > > Hadoop/Kerberos user: root > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: JVM: Java > > HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.241-b07 > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Maximum heap size: > > 406 MiBytes > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: JAVA_HOME: > > /usr/java/default > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Hadoop version: > 2.6.5 > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: JVM Options: > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: -Xms424m > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: -Xmx424m > > 20/01/22 11:07:53 INFO entrypoint.ClusterEntrypoint: Program Arguments: > > (none) > > > > taskmanager.err: > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: > > > -------------------------------------------------------------------------------- > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: Starting YARN > > TaskExecutor runner (Version: <unknown>, Rev:7297bac, Date:24.06.2019 @ > > 23:04:28 CST) > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: OS current user: > > cloudera-scm > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: Current > > Hadoop/Kerberos user: root > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: JVM: Java > HotSpot(TM) > > 64-Bit Server VM - Oracle Corporation - 1.8/25.241-b07 > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: Maximum heap size: > > 345 MiBytes > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: JAVA_HOME: > > /usr/java/default > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: Hadoop version: > 2.6.5 > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: JVM Options: > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: -Xms360m > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: -Xmx360m > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: > > -XX:MaxDirectMemorySize=664m > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: Program Arguments: > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: --configDir > > 20/01/22 11:07:57 INFO yarn.YarnTaskExecutorRunner: . > > > > 网上搜了下,这报错一般都是内存的问题,请问下这个是跟yarn上的内存设置造成的吗? > > > > ________________________________ > > zjfpla...@hotmail.com > > >