Hello,

Please help me upgrade to 1.10 in AWS EMR.

On Fri, May 1, 2020 at 4:05 PM aj <ajainje...@gmail.com> wrote:

> Hi Yang,
>
> I am attaching the logs for your reference, please help me what i am doing
> wrong.
>
> Thanks,
> Anuj
>
> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <danrtsey...@gmail.com> wrote:
>
>> Hi Anuj,
>>
>> I think the exception you come across still because the hadoop version
>> is 2.4.1. I have checked the hadoop code, the code line are exactly same.
>> For 2.8.1, i also have checked the ruleParse. It could work.
>>
>> /**
>>  * A pattern for parsing a auth_to_local rule.
>>  */
>> private static final Pattern ruleParser =
>>   
>> Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>
>>
>> Could you share the jobmanager logs so that i could check the classpath
>> and hadoop version?
>>
>> Best,
>> Yang
>>
>> aj <ajainje...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>
>>> Hello Yang,
>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>> and I have put in flink/lib.   flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>
>>> then I am getting below error :
>>>
>>> SLF4J: Class path contains multiple SLF4J bindings.
>>> SLF4J: Found binding in
>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> SLF4J: Found binding in
>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>> explanation.
>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>> Exception in thread "main" java.lang.IllegalArgumentException: Invalid
>>> rule: /L
>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>       DEFAULT
>>>         at
>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>         at
>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>         at
>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>         at
>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>         at
>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>
>>>
>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from lib
>>> then i get below error:
>>>
>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     -  Classpath:
>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     -
>>> --------------------------------------------------------------------------------
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: jobmanager.heap.size, 1024m
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: taskmanager.memory.process.size, 1568m
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: parallelism.default, 1
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: jobmanager.execution.failover-strategy, region
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: classloader.resolve-order, parent-first
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: s3.secret-key, ******
>>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>>                     - Could not load CLI class
>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>> java.lang.NoClassDefFoundError:
>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>         at java.lang.Class.forName0(Native Method)
>>>         at java.lang.Class.forName(Class.java:264)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>> Caused by: java.lang.ClassNotFoundException:
>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>         at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>         ... 5 more
>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>                     - Hadoop is not in the classpath/dependencies. The
>>> extended set of supported File Systems via Hadoop is not available.
>>> 2020-04-27 16:59:37,458 INFO
>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>>> create Hadoop Security Module because Hadoop cannot be found in the
>>> Classpath.
>>> 2020-04-27 16:59:37,476 INFO
>>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>> 2020-04-27 16:59:37,480 INFO
>>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>> Classpath.
>>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     - Running 'run' command.
>>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     - Building program from JAR file
>>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>>                     - Invalid command line arguments.
>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>> program from JAR file.
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>         at
>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>> Caused by: java.io.FileNotFoundException: JAR file does not exist: -ynm
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>         ... 4 more
>>>
>>>
>>> Can you please help.
>>>
>>> Thanks,
>>> Anuj
>>>
>>>
>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <danrtsey...@gmail.com> wrote:
>>>
>>>> Hi Anuj,
>>>>
>>>> It seems that you are using hadoop version 2.4.1. I think "L" could not
>>>> be supported in
>>>> this version. Could you upgrade your hadoop version to 2.8 and have a
>>>> try? If your
>>>> YARN cluster version is 2.8+, then you could directly remove the
>>>> flink-shaded-hadoop
>>>> in your lib directory. Otherwise, you need to download the
>>>> flink-shaded-hadoop with
>>>> version 2.8 here[1].
>>>>
>>>>
>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>
>>>> Best,
>>>> Yang
>>>>
>>>> aj <ajainje...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>
>>>>> Hi Robert,
>>>>> attached the full application log file.
>>>>>
>>>>> Thanks,
>>>>> Anuj
>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>
>>>>
>>>
>>> --
>>> Thanks & Regards,
>>> Anuj Jain
>>> Mob. : +91- 8588817877
>>> Skype : anuj.jain07
>>> <http://www.oracle.com/>
>>>
>>>
>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>
>>
>
> --
> Thanks & Regards,
> Anuj Jain
> Mob. : +91- 8588817877
> Skype : anuj.jain07
> <http://www.oracle.com/>
>
>
> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>


-- 
Thanks & Regards,
Anuj Jain
Mob. : +91- 8588817877
Skype : anuj.jain07
<http://www.oracle.com/>


<http://www.cse.iitm.ac.in/%7Eanujjain/>

Reply via email to