flink-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From aj <ajainje...@gmail.com>
Subject Re: upgrade flink from 1.9.1 to 1.10.0 on EMR
Date Thu, 14 May 2020 17:11:40 GMT
Hi Yang,

I am able to resolve the issue by removing Hadoop dependency as you
mentioned.

1. Removed hadoop-common dependency and

    <groupId>org.apache.flink</groupId>
    <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
    <version>${flink.version}</version>
    <exclusions>
        <exclusion>
                <groupId>org.apache.flink</groupId>
                <artifactId>flink-hadoop-fs</artifactId>
        </exclusion>
    </exclusions>

</dependency>


2. After the above changes, I am able to submit job on yarn but facing
issues with s3 plugin that I resolved by putting *
flink-s3-fs-hadoop-1.9.0.jar *i n the plugins/s3-fs-hadoop  directory.

Thanks for your support.

Any update when will flink.10 officially supported in EMR.  Even in new EMR
version(emr 6.0)  flink has been removed.



On Sat, May 9, 2020 at 1:36 PM aj <ajainjecrc@gmail.com> wrote:

> Hello Yang,
>
> I have attached my pom file and I did not see that I am using any Hadoop
> dependency. Can you please help me.
>
> On Wed, May 6, 2020 at 1:22 PM Yang Wang <danrtsey.wy@gmail.com> wrote:
>
>> Hi aj,
>>
>> From the logs you have provided, the hadoop version is still 2.4.1.
>> Could you check the user jar(i.e. events-processor-1.0-SNAPSHOT.jar) have
>> some
>> hadoop classes? If it is, you need to exclude the hadoop dependency.
>>
>>
>> Best,
>> Yang
>>
>> aj <ajainjecrc@gmail.com> 于2020年5月6日周三 下午3:38写道:
>>
>>> Hello,
>>>
>>> Please help me upgrade to 1.10 in AWS EMR.
>>>
>>> On Fri, May 1, 2020 at 4:05 PM aj <ajainjecrc@gmail.com> wrote:
>>>
>>>> Hi Yang,
>>>>
>>>> I am attaching the logs for your reference, please help me what i am
>>>> doing wrong.
>>>>
>>>> Thanks,
>>>> Anuj
>>>>
>>>> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <danrtsey.wy@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Anuj,
>>>>>
>>>>> I think the exception you come across still because the hadoop version
>>>>> is 2.4.1. I have checked the hadoop code, the code line are exactly
>>>>> same.
>>>>> For 2.8.1, i also have checked the ruleParse. It could work.
>>>>>
>>>>> /**
>>>>>  * A pattern for parsing a auth_to_local rule.
>>>>>  */
>>>>> private static final Pattern ruleParser =
>>>>>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>>>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>>>>
>>>>>
>>>>> Could you share the jobmanager logs so that i could check the
>>>>> classpath and hadoop version?
>>>>>
>>>>> Best,
>>>>> Yang
>>>>>
>>>>> aj <ajainjecrc@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>>>>
>>>>>> Hello Yang,
>>>>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>>>>> and I have put in flink/lib.
>>>>>>  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>>>>
>>>>>> then I am getting below error :
>>>>>>
>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>> SLF4J: Found binding in
>>>>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>> SLF4J: Found binding in
>>>>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for
an
>>>>>> explanation.
>>>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>>>> Exception in thread "main" java.lang.IllegalArgumentException:
>>>>>> Invalid rule: /L
>>>>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>>>>       DEFAULT
>>>>>>         at
>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>>>         at
>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>>         at
>>>>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>>>>         at
>>>>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>>>>         at
>>>>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>>>>
>>>>>>
>>>>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from
lib
>>>>>> then i get below error:
>>>>>>
>>>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       -  Classpath:
>>>>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       -
>>>>>> --------------------------------------------------------------------------------
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: jobmanager.heap.size, 1024m
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: taskmanager.memory.process.size, 1568m
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: parallelism.default, 1
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: jobmanager.execution.failover-strategy, region
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: classloader.resolve-order, parent-first
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            -
Loading
>>>>>> configuration property: s3.secret-key, ******
>>>>>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Could not load CLI class
>>>>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>>>>> java.lang.NoClassDefFoundError:
>>>>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>>>>         at java.lang.Class.forName0(Native Method)
>>>>>>         at java.lang.Class.forName(Class.java:264)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>>>>         at
>>>>>> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>>>>         ... 5 more
>>>>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>>>>                       - Hadoop is not in the classpath/dependencies.
The
>>>>>> extended set of supported File Systems via Hadoop is not available.
>>>>>> 2020-04-27 16:59:37,458 INFO
>>>>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  -
Cannot
>>>>>> create Hadoop Security Module because Hadoop cannot be found in the
>>>>>> Classpath.
>>>>>> 2020-04-27 16:59:37,476 INFO
>>>>>>  org.apache.flink.runtime.security.modules.JaasModule          -
Jaas file
>>>>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>>>>> 2020-04-27 16:59:37,480 INFO
>>>>>>  org.apache.flink.runtime.security.SecurityUtils               -
Cannot
>>>>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>>>>> Classpath.
>>>>>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Running 'run' command.
>>>>>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Building program from JAR file
>>>>>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Invalid command line arguments.
>>>>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>>>>> program from JAR file.
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>>>>         at
>>>>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>>>>> Caused by: java.io.FileNotFoundException: JAR file does not exist:
>>>>>> -ynm
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>>>>         ... 4 more
>>>>>>
>>>>>>
>>>>>> Can you please help.
>>>>>>
>>>>>> Thanks,
>>>>>> Anuj
>>>>>>
>>>>>>
>>>>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <danrtsey.wy@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Anuj,
>>>>>>>
>>>>>>> It seems that you are using hadoop version 2.4.1. I think "L"
could
>>>>>>> not be supported in
>>>>>>> this version. Could you upgrade your hadoop version to 2.8 and
have
>>>>>>> a try? If your
>>>>>>> YARN cluster version is 2.8+, then you could directly remove
the
>>>>>>> flink-shaded-hadoop
>>>>>>> in your lib directory. Otherwise, you need to download the
>>>>>>> flink-shaded-hadoop with
>>>>>>> version 2.8 here[1].
>>>>>>>
>>>>>>>
>>>>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>>>>
>>>>>>> Best,
>>>>>>> Yang
>>>>>>>
>>>>>>> aj <ajainjecrc@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>>>>
>>>>>>>> Hi Robert,
>>>>>>>> attached the full application log file.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Anuj
>>>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Thanks & Regards,
>>>>>> Anuj Jain
>>>>>> Mob. : +91- 8588817877
>>>>>> Skype : anuj.jain07
>>>>>> <http://www.oracle.com/>
>>>>>>
>>>>>>
>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>
>>>>>
>>>>
>>>> --
>>>> Thanks & Regards,
>>>> Anuj Jain
>>>> Mob. : +91- 8588817877
>>>> Skype : anuj.jain07
>>>> <http://www.oracle.com/>
>>>>
>>>>
>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>
>>>
>>>
>>> --
>>> Thanks & Regards,
>>> Anuj Jain
>>> Mob. : +91- 8588817877
>>> Skype : anuj.jain07
>>> <http://www.oracle.com/>
>>>
>>>
>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>
>>
>
> --
> Thanks & Regards,
> Anuj Jain
> Mob. : +91- 8588817877
> Skype : anuj.jain07
> <http://www.oracle.com/>
>
>
> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>


-- 
Thanks & Regards,
Anuj Jain
Mob. : +91- 8588817877
Skype : anuj.jain07
<http://www.oracle.com/>


<http://www.cse.iitm.ac.in/%7Eanujjain/>

Mime
View raw message