没有人遇到这类问题吗?
[docker@master flink-1.13.2]$ ./bin/yarn-session.sh SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/home/docker/flink-1.13.2/lib/log4j-slf4j-impl-2.12.1.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/home/docker/hadoop-3.2.1/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] 2021-11-08 14:36:43,651 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.rpc.address, master 2021-11-08 14:36:43,657 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.rpc.port, 6123 2021-11-08 14:36:43,657 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.memory.process.size, 1600m 2021-11-08 14:36:43,657 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.memory.process.size, 1728m 2021-11-08 14:36:43,657 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.numberOfTaskSlots, 8 2021-11-08 14:36:43,657 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: parallelism.default, 1 2021-11-08 14:36:43,658 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: execution.checkpointing.interval, 30s 2021-11-08 14:36:43,658 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: execution.checkpointing.unaligned, true 2021-11-08 14:36:43,658 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: execution.checkpointing.timeout, 1200s 2021-11-08 14:36:43,658 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend, filesystem 2021-11-08 14:36:43,658 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.checkpoints.dir, oss://datalake-huifu/hudi/flink/checkpoints 2021-11-08 14:36:43,659 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.savepoints.dir, oss://datalake-huifu/hudi/flink/savepoints 2021-11-08 14:36:43,659 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.incremental, true 2021-11-08 14:36:43,659 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.execution.failover-strategy, region 2021-11-08 14:36:43,659 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: rest.port, 18081 2021-11-08 14:36:43,660 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: classloader.resolve-order, parent-first 2021-11-08 14:36:43,660 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: web.submit.enabled, false 2021-11-08 14:36:43,660 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: fs.oss.endpoint, oss-cn-shanghai.aliyuncs.com 2021-11-08 14:36:43,660 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: fs.oss.accessKeyId, LTAI5tJ4k9pk1KwZVsLd8NHd 2021-11-08 14:36:43,661 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: fs.oss.accessKeySecret, ****** 2021-11-08 14:36:44,042 WARN org.apache.hadoop.util.NativeCodeLoader [] - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2021-11-08 14:36:44,155 INFO org.apache.flink.runtime.security.modules.HadoopModule [] - Hadoop user set to docker (auth:SIMPLE) 2021-11-08 14:36:44,171 INFO org.apache.flink.runtime.security.modules.JaasModule [] - Jaas file will be created as /tmp/jaas-5414790866026859677.conf. 2021-11-08 14:36:44,213 WARN org.apache.flink.yarn.configuration.YarnLogConfigUtil [] - The configuration directory ('/home/docker/flink-1.13.2/conf') already contains a LOG4J config file.If you want to use logback, then please delete or rename the log configuration file. 2021-11-08 14:36:44,340 INFO org.apache.hadoop.yarn.client.RMProxy [] - Connecting to ResourceManager at master/192.168.16.191:8032 2021-11-08 14:36:44,711 INFO org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils [] - The derived from fraction jvm overhead memory (160.000mb (167772162 bytes)) is less than its min value 192.000mb (201326592 bytes), min value will be used instead 2021-11-08 14:36:44,734 INFO org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils [] - The derived from fraction jvm overhead memory (172.800mb (181193935 bytes)) is less than its min value 192.000mb (201326592 bytes), min value will be used instead 2021-11-08 14:36:45,025 INFO org.apache.hadoop.conf.Configuration [] - resource-types.xml not found 2021-11-08 14:36:45,026 INFO org.apache.hadoop.yarn.util.resource.ResourceUtils [] - Unable to find 'resource-types.xml'. 2021-11-08 14:36:45,091 INFO org.apache.flink.yarn.YarnClusterDescriptor [] - The configured JobManager memory is 1600 MB. YARN will allocate 2048 MB to make up an integer multiple of its minimum allocation memory (1024 MB, configured via 'yarn.scheduler.minimum-allocation-mb'). The extra 448 MB may not be used by Flink. 2021-11-08 14:36:45,092 INFO org.apache.flink.yarn.YarnClusterDescriptor [] - The configured TaskManager memory is 1728 MB. YARN will allocate 2048 MB to make up an integer multiple of its minimum allocation memory (1024 MB, configured via 'yarn.scheduler.minimum-allocation-mb'). The extra 320 MB may not be used by Flink. 2021-11-08 14:36:45,092 INFO org.apache.flink.yarn.YarnClusterDescriptor [] - Cluster specification: ClusterSpecification{masterMemoryMB=1600, taskManagerMemoryMB=1728, slotsPerTaskManager=8} 2021-11-08 14:36:45,847 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:45,979 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,004 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,032 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,191 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,248 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,287 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,325 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,368 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 2021-11-08 14:36:46,546 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false ............. 一直输出 INFO org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted = false 查看yarn控台没有flink session集群创建 At 2021-11-04 21:18:03, "casel.chen" <casel_c...@126.com> wrote: >flink-conf.yaml 内容如下 > > >################################################################################ ># Licensed to the Apache Software Foundation (ASF) under one ># or more contributor license agreements. See the NOTICE file ># distributed with this work for additional information ># regarding copyright ownership. The ASF licenses this file ># to you under the Apache License, Version 2.0 (the ># "License"); you may not use this file except in compliance ># with the License. You may obtain a copy of the License at ># ># http://www.apache.org/licenses/LICENSE-2.0 ># ># Unless required by applicable law or agreed to in writing, software ># distributed under the License is distributed on an "AS IS" BASIS, ># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ># See the License for the specific language governing permissions and ># limitations under the License. >################################################################################ > > > > >#============================================================================== ># Common >#============================================================================== > > ># The external address of the host on which the JobManager runs and can be ># reached by the TaskManagers and any clients which want to connect. This >setting ># is only used in Standalone mode and may be overwritten on the JobManager side ># by specifying the --host <hostname> parameter of the bin/jobmanager.sh >executable. ># In high availability mode, if you use the bin/start-cluster.sh script and >setup ># the conf/masters file, this will be taken care of automatically. Yarn/Mesos ># automatically configure the host name based on the hostname of the node >where the ># JobManager runs. > > >jobmanager.rpc.address: master > > ># The RPC port where the JobManager is reachable. > > >jobmanager.rpc.port: 6123 > > > > ># The total process memory size for the JobManager. ># ># Note this accounts for all memory usage within the JobManager process, >including JVM metaspace and other overhead. > > >jobmanager.memory.process.size: 1600m > > > > ># The total process memory size for the TaskManager. ># ># Note this accounts for all memory usage within the TaskManager process, >including JVM metaspace and other overhead. > > >taskmanager.memory.process.size: 1728m > > ># To exclude JVM metaspace and overhead, please, use total Flink memory size >instead of 'taskmanager.memory.process.size'. ># It is not recommended to set both 'taskmanager.memory.process.size' and >Flink memory. ># ># taskmanager.memory.flink.size: 1280m > > ># The number of task slots that each TaskManager offers. Each slot runs one >parallel pipeline. > > >taskmanager.numberOfTaskSlots: 8 > > ># The parallelism used for programs that did not specify and other parallelism. > > >parallelism.default: 1 > > ># The default file system scheme and authority. ># ># By default file paths without scheme are interpreted relative to the local ># root file system 'file:///'. Use this to override the default and interpret ># relative paths relative to a different file system, ># for example 'hdfs://mynamenode:12345' ># ># fs.default-scheme > > >#============================================================================== ># High Availability >#============================================================================== > > ># The high-availability mode. Possible options are 'NONE' or 'zookeeper'. ># ># high-availability: zookeeper > > ># The path where metadata for master recovery is persisted. While ZooKeeper >stores ># the small ground truth for checkpoint and leader election, this location >stores ># the larger objects, like persisted dataflow graphs. ># ># Must be a durable file system that is accessible from all nodes ># (like HDFS, S3, Ceph, nfs, ...) ># ># high-availability.storageDir: hdfs:///flink/ha/ > > ># The list of ZooKeeper quorum peers that coordinate the high-availability ># setup. This must be a list of the form: ># "host1:clientPort,host2:clientPort,..." (default clientPort: 2181) ># ># high-availability.zookeeper.quorum: localhost:2181 > > > > ># ACL options are based on >https://zookeeper.apache.org/doc/r3.1.2/zookeeperProgrammers.html#sc_BuiltinACLSchemes ># It can be either "creator" (ZOO_CREATE_ALL_ACL) or "open" >(ZOO_OPEN_ACL_UNSAFE) ># The default value is "open" and it can be changed to "creator" if ZK >security is enabled ># ># high-availability.zookeeper.client.acl: open > > >#============================================================================== ># Fault tolerance and checkpointing >#============================================================================== > > >execution.checkpointing.interval: 60s > > >execution.checkpointing.unaligned: true > > >execution.checkpointing.timeout: 1200s > > ># The backend that will be used to store operator state checkpoints if ># checkpointing is enabled. ># ># Supported backends are 'jobmanager', 'filesystem', 'rocksdb', or the ># <class-name-of-factory>. ># >state.backend: filesystem > > ># Directory for checkpoints filesystem, when using any of the default bundled ># state backends. ># >state.checkpoints.dir: oss://datalake-huifu/hudi/flink/checkpoints > > ># Default target directory for savepoints, optional. ># >state.savepoints.dir: oss://datalake-huifu/hudi/flink/savepoints > > ># Flag to enable/disable incremental checkpoints for backends that ># support incremental checkpoints (like the RocksDB state backend). ># >state.backend.incremental: true > > ># The failover strategy, i.e., how the job computation recovers from task >failures. ># Only restart tasks that may have been affected by the task failure, which >typically includes ># downstream tasks and potentially upstream tasks if their produced data is no >longer available for consumption. > > >jobmanager.execution.failover-strategy: region > > >#============================================================================== ># Rest & web frontend >#============================================================================== > > ># The port to which the REST client connects to. If rest.bind-port has ># not been specified, then the server will bind to this port as well. ># >rest.port: 18081 > > ># The address to which the REST client will connect to ># >#rest.address: 0.0.0.0 > > ># Port range for the REST and web server to bind to. ># >#rest.bind-port: 8080-8090 > > ># The address that the REST & web server binds to ># >#rest.bind-address: 0.0.0.0 > > ># Flag to specify whether job submission is enabled from the web-based ># runtime monitor. Uncomment to disable. > > >#web.submit.enable: false > > >#============================================================================== ># Advanced >#============================================================================== > > ># Override the directories for temporary files. If not specified, the ># system-specific Java temporary directory (java.io.tmpdir property) is taken. ># ># For framework setups on Yarn or Mesos, Flink will automatically pick up the ># containers' temp directories without any need for configuration. ># ># Add a delimited list for multiple directories, using the system directory ># delimiter (colon ':' on unix) or a comma, e.g.: ># /data1/tmp:/data2/tmp:/data3/tmp ># ># Note: Each directory entry is read from and written to by a different I/O ># thread. You can include the same directory multiple times in order to create ># multiple I/O threads against that directory. This is for example relevant for ># high-throughput RAIDs. ># ># io.tmp.dirs: /tmp > > ># The classloading resolve order. Possible values are 'child-first' (Flink's >default) ># and 'parent-first' (Java's default). ># ># Child first classloading allows users to use different dependency/library ># versions in their application than those in the classpath. Switching back ># to 'parent-first' may help with debugging dependency issues. ># >classloader.resolve-order: parent-first > > ># The amount of memory going to the network stack. These numbers usually need ># no tuning. Adjusting them may be necessary in case of an "Insufficient number ># of network buffers" error. The default min is 64MB, the default max is 1GB. ># ># taskmanager.memory.network.fraction: 0.1 ># taskmanager.memory.network.min: 64mb ># taskmanager.memory.network.max: 1gb > > >#============================================================================== ># Flink Cluster Security Configuration >#============================================================================== > > ># Kerberos authentication for various components - Hadoop, ZooKeeper, and >connectors - ># may be enabled in four steps: ># 1. configure the local krb5.conf file ># 2. provide Kerberos credentials (either a keytab or a ticket cache w/ kinit) ># 3. make the credentials available to various JAAS login contexts ># 4. configure the connector to use JAAS/SASL > > ># The below configure how Kerberos credentials are provided. A keytab will be >used instead of ># a ticket cache if the keytab path and principal are set. > > ># security.kerberos.login.use-ticket-cache: true ># security.kerberos.login.keytab: /path/to/kerberos/keytab ># security.kerberos.login.principal: flink-user > > ># The configuration below defines which JAAS login contexts > > ># security.kerberos.login.contexts: Client,KafkaClient > > >#============================================================================== ># ZK Security Configuration >#============================================================================== > > ># Below configurations are applicable if ZK ensemble is configured for security > > ># Override below configuration to provide custom ZK service name if configured ># zookeeper.sasl.service-name: zookeeper > > ># The configuration below must match one of the values set in >"security.kerberos.login.contexts" ># zookeeper.sasl.login-context-name: Client > > >#============================================================================== ># HistoryServer >#============================================================================== > > ># The HistoryServer is started and stopped via bin/historyserver.sh >(start|stop) > > ># Directory to upload completed jobs to. Add this directory to the list of ># monitored directories of the HistoryServer as well (see below). >#jobmanager.archive.fs.dir: hdfs:///completed-jobs/ > > ># The address under which the web-based HistoryServer listens. >#historyserver.web.address: 0.0.0.0 > > ># The port under which the web-based HistoryServer listens. >#historyserver.web.port: 8082 > > ># Comma separated list of directories to monitor for completed jobs. >#historyserver.archive.fs.dir: hdfs:///completed-jobs/ > > ># Interval in milliseconds for refreshing the monitored directories. >#historyserver.archive.fs.refresh-interval: 10000 > > >fs.oss.endpoint: oss-cn-shanghai.aliyuncs.com >fs.oss.accessKeyId: xxx >fs.oss.accessKeySecret: xxx > > >metrics.reporter.promgateway.class: >org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporter >metrics.reporter.promgateway.host: 192.168.24.41 >metrics.reporter.promgateway.port: 9091 >metrics.reporter.promgateway.jobName: PerfTest >metrics.reporter.promgateway.randomJobNameSuffix: true >metrics.reporter.promgateway.deleteOnShutdown: false > > > > > > > > > > > > > > > > > > > > > > > > > >在 2021-11-04 19:32:45,"Caizhi Weng" <tsreape...@gmail.com> 写道: >>Hi! >> >>没有在邮件里发现附件,可以考虑把 flink-conf.yaml 的内容贴在邮件里,或者外部剪贴板。 >> >>casel.chen <casel_c...@126.com> 于2021年11月4日周四 下午6:42写道: >> >>> flink 1.13.2 + hadoop 3.2.1 >>> yarn上已经成功跑了hive和spark作业 >>> flink上通过运行 bin/yarn-session.sh 启动yarn session集群的时候一直报如下INFO日志,查看yarn web >>> console发现并没有启flink-session集群,我的flink-conf.yaml配置如附件,hadoop集群并没有开启认证SSL之类的,改用standalone模式是可以启动3节点集群的,请问这会是什么原因造成的?要怎么修复?谢谢! >>> >>> >>> >>> 2021-11-04 16:51:39,964 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:39,986 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,004 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,020 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,041 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,059 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,078 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,097 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,114 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,134 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,155 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,175 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,193 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> 2021-11-04 16:51:40,212 INFO >>> org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient [] >>> - SASL encryption trust check: localHostTrusted = false, remoteHostTrusted >>> = false >>> >>> >>> >>> >>> >>>