[jira] [Commented] (FLINK-8812) Possible resource leak in Flip6
[ https://issues.apache.org/jira/browse/FLINK-8812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16403555#comment-16403555 ] ASF GitHub Bot commented on FLINK-8812: --- GitHub user tillrohrmann opened a pull request: https://github.com/apache/flink/pull/5713 [FLINK-8812] [flip6] Set managed memory for TaskExecutor to 80 MB in MiniCluster ## What is the purpose of the change In order to avoid problems with OOM exceptions, this commit sets the managed memory to 80 MB for TaskExecutors started by the MiniCluster. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) You can merge this pull request into a Git repository by running: $ git pull https://github.com/tillrohrmann/flink findResourceLeak Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5713.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5713 commit 47f3f515ea75db20528079fef289de5120ae95c5 Author: Till RohrmannDate: 2018-03-16T18:16:49Z [FLINK-9016] [flip6] Properly unregister jobs from JobMetricGroup This commit properly removes jobs from the JobMetricGroup once a job has reached a terminal state. commit e42749cea4a0d35850c9482c8b31274aa081dbfa Author: Till Rohrmann Date: 2018-03-17T16:53:35Z [FLINK-8812] [flip6] Set managed memory for TaskExecutor to 80 MB in MiniCluster In order to avoid problems with OOM exceptions, this commit sets the managed memory to 80 MB for TaskExecutors started by the MiniCluster. > Possible resource leak in Flip6 > --- > > Key: FLINK-8812 > URL: https://issues.apache.org/jira/browse/FLINK-8812 > Project: Flink > Issue Type: Bug > Components: Tests >Reporter: Chesnay Schepler >Assignee: Till Rohrmann >Priority: Blocker > Labels: flip-6 > Fix For: 1.5.0 > > > In this build (https://travis-ci.org/zentol/flink/builds/347373839) I set the > codebase to flip6 for half the profiles to find failing tests. > The "libraries" job (https://travis-ci.org/zentol/flink/jobs/347373851) > failed with an OutOfMemoryError. > This could mean that there is a memory-leak somewhere. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5713: [FLINK-8812] [flip6] Set managed memory for TaskEx...
GitHub user tillrohrmann opened a pull request: https://github.com/apache/flink/pull/5713 [FLINK-8812] [flip6] Set managed memory for TaskExecutor to 80 MB in MiniCluster ## What is the purpose of the change In order to avoid problems with OOM exceptions, this commit sets the managed memory to 80 MB for TaskExecutors started by the MiniCluster. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) You can merge this pull request into a Git repository by running: $ git pull https://github.com/tillrohrmann/flink findResourceLeak Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5713.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5713 commit 47f3f515ea75db20528079fef289de5120ae95c5 Author: Till RohrmannDate: 2018-03-16T18:16:49Z [FLINK-9016] [flip6] Properly unregister jobs from JobMetricGroup This commit properly removes jobs from the JobMetricGroup once a job has reached a terminal state. commit e42749cea4a0d35850c9482c8b31274aa081dbfa Author: Till Rohrmann Date: 2018-03-17T16:53:35Z [FLINK-8812] [flip6] Set managed memory for TaskExecutor to 80 MB in MiniCluster In order to avoid problems with OOM exceptions, this commit sets the managed memory to 80 MB for TaskExecutors started by the MiniCluster. ---
[jira] [Created] (FLINK-9016) Unregister jobs from JobMetricGroup after termination
Till Rohrmann created FLINK-9016: Summary: Unregister jobs from JobMetricGroup after termination Key: FLINK-9016 URL: https://issues.apache.org/jira/browse/FLINK-9016 Project: Flink Issue Type: Bug Components: Metrics Affects Versions: 1.5.0 Reporter: Till Rohrmann Assignee: Till Rohrmann Fix For: 1.5.0 In order to free resources and unregister metrics, jobs should be properly unregistered from the {{JobMetricGroup}} once they have reached terminal state. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9009) Error| You are creating too many HashedWheelTimer instances. HashedWheelTimer is a shared resource that must be reused across the application, so that only a few insta
[ https://issues.apache.org/jira/browse/FLINK-9009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16403461#comment-16403461 ] Pankaj commented on FLINK-9009: --- Stacktrace Sink: Cassandra Sink, Filter -> Flat Map -> Sink: Unnamed, Filter -> Flat Map -> Sink: Cassandra Sink, Filter -> Flat Map -> Sink: Unnamed, Filter -> Flat Map -> Sink: *Cassandra Sink)(1/10*) switched to RUNNING 17.03.2018 14:33:59.740 [OUT] [ERROR] [ ] [ ] *io.netty.util.ResourceLeakDetector LEAK: You are creating too many HashedWheelTimer instances. HashedWheelTimer is a shared resource that must be reused across the JVM,so that only a few instances are create*d. # # java.lang.OutOfMemoryError: Java heap space # -XX:OnOutOfMemoryError="/opt/tomcat/bin/shutdown.sh 5" # Executing /bin/sh -c "/opt/tomcat/bin/shutdown.sh 5"... Mar 17, 2018 2:34:28 PM org.apache.catalina.startup.Catalina stopServer SEVERE: Could not contact localhost:8005. Tomcat may not be running. Mar 17, 2018 2:34:28 PM org.apache.catalina.startup.Catalina stopServer SEVERE: Catalina.stop: java.net.ConnectException: Connection refused (Connection refused) at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) at java.net.Socket.connect(Socket.java:589) at java.net.Socket.connect(Socket.java:538) --- "cluster16-nio-worker-1" #142 prio=5 os_prio=0 tid=0x7f97fc386000 nid=0xbe waiting for monitor entry [0x7f9782e86000] java.lang.Thread.State: BLOCKED (on object monitor) at *com.datastax.driver.core.Connection$10.operationComplete(Connection.java:547) at* com.datastax.driver.core.Connection$10.operationComplete(Connection.java:534) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:680) at io.netty.util.concurrent.DefaultPromise.notifyLateListener(DefaultPromise.java:621) at io.netty.util.concurrent.DefaultPromise.addListener(DefaultPromise.java:138) at io.netty.channel.DefaultChannelPromise.addListener(DefaultChannelPromise.java:93) at io.netty.channel.DefaultChannelPromise.addListener(DefaultChannelPromise.java:28) at com.datastax.driver.core.Connection$Flusher.run(Connection.java:870) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:358) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:112) at java.lang.Thread.run(Thread.java:748) io.netty.handler.codec.MessageToMessageEncoder.write(MessageToMessageEncoder.java:89) at io.netty.channel.AbstractChannelHandlerContext.invokeWrite(AbstractChannelHandlerContext.java:643) at io.netty.channel.AbstractChannelHandlerContext.write(AbstractChannelHandlerContext.java:700) at io.netty.channel.AbstractChannelHandlerContext.write(AbstractChannelHandlerContext.java:636) at io.netty.handler.timeout.IdleStateHandler.write(IdleStateHandler.java:284) at io.netty.channel.AbstractChannelHandlerContext.invokeWrite(AbstractChannelHandlerContext.java:643) at io.netty.channel.AbstractChannelHandlerContext.write(AbstractChannelHandlerContext.java:700) at io.netty.channel.AbstractChannelHandlerContext.write(AbstractChannelHandlerContext.java:636) at io.netty.channel.AbstractChannelHandlerContext.write(AbstractChannelHandlerContext.java:622) at io.netty.channel.DefaultChannelPipeline.write(DefaultChannelPipeline.java:939) at io.netty.channel.AbstractChannel.write(AbstractChannel.java:234) at com.datastax.driver.core.Connection$Flusher.run(Connection.java:870) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:358) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:112) at java.lang.Thread.run(Thread.java:748) > Error| You are creating too many HashedWheelTimer instances. > HashedWheelTimer is a shared resource that must be reused across the > application, so that only a few instances are created. > - > > Key: FLINK-9009 > URL: https://issues.apache.org/jira/browse/FLINK-9009 > Project: Flink > Issue Type: Bug > Environment: Pass platform: Openshit >Reporter: Pankaj >Priority: Blocker > > Steps to reproduce: > 1- Flink with Kafka as a consumer -> Writing stream to Cassandra using flink >
[jira] [Commented] (FLINK-7151) FLINK SQL support create temporary function and table
[ https://issues.apache.org/jira/browse/FLINK-7151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16403335#comment-16403335 ] Shuyi Chen commented on FLINK-7151: --- [~yinhua], yes, you should be able to implement your own TableSourceFactory to do so. > FLINK SQL support create temporary function and table > - > > Key: FLINK-7151 > URL: https://issues.apache.org/jira/browse/FLINK-7151 > Project: Flink > Issue Type: New Feature > Components: Table API SQL >Reporter: yuemeng >Assignee: yuemeng >Priority: Major > > Based on create temporary function and table.we can register a udf,udaf,udtf > use sql: > {code} > CREATE TEMPORARY function 'TOPK' AS > 'com..aggregate.udaf.distinctUdaf.topk.ITopKUDAF'; > INSERT INTO db_sink SELECT id, TOPK(price, 5, 'DESC') FROM kafka_source GROUP > BY id; > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)