[ 
https://issues.apache.org/jira/browse/TINKERPOP-2030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simone Rondelli updated TINKERPOP-2030:
---------------------------------------
    Description: 
The Gremlin Driver communicates with the server via {{Connection.write()}}.

The {{write()}} method has a logic in the end to schedule a task to keep the 
connection alive, which by default is scheduled to be run after 30min, every 30 
min.


{code:java}
  // try to keep the connection alive if the channel allows such things - 
websockets will
        if (channelizer.supportsKeepAlive() && keepAliveInterval > 0) {

            final ScheduledFuture oldKeepAliveFuture = 
keepAliveFuture.getAndSet(cluster.executor().scheduleAtFixedRate(() -> {
                logger.debug("Request sent to server to keep {} alive", 
thisConnection);
                try {
                    channel.writeAndFlush(channelizer.createKeepAliveMessage());
                } catch (Exception ex) {
                    // will just log this for now - a future real request can 
be responsible for the failure that
                    // marks the host as dead. this also may not mean the host 
is actually dead. more robust handling
                    // is in play for real requests, not this simple ping
                    logger.warn(String.format("Keep-alive did not succeed on 
%s", thisConnection), ex);
                }
            }, keepAliveInterval, keepAliveInterval, TimeUnit.MILLISECONDS));

            // try to cancel the old future if it's still un-executed - no need 
to ping since a new write has come
            // through on the connection
            if (oldKeepAliveFuture != null) oldKeepAliveFuture.cancel(true);
        }
{code}

The problem with this is that on every call of the {{wirte()}} method, which 
basically means on every query executed, the {[Connection}} schedule a 
KeepAlive task which won't be run for at least 30 minutes. This lead the 
{{Cluster.executor()}} queue to fill up with tasks waiting for completion.

One possible solution to fix this would be to schedule avoid this KeepAlive 
task to be instantiated more than once per connection:

{code:java}
final class Connection {
    ...
    private final AtomicBoolean keepAliveInitialized = new AtomicBoolean(false);

    public ChannelPromise write(final RequestMessage requestMessage, final 
CompletableFuture<ResultSet> future) {
        ...
        
        // FIX HERE: with keepAliveInitialized.compareAndSet(false, true) we 
ensure the keepAlive task is run only once per connection.
        if (channelizer.supportsKeepAlive() && keepAliveInterval > 0 && 
keepAliveInitialized.compareAndSet(false, true)) {
            ...
        }
    }
}


}
{code}


  was:
The Gremlin Driver communicates with the server via {{Connection.write()}}.

The {{write()}} method has a logic in the end to schedule a task to keep the 
connection alive, which by default is scheduled to be run after 30min, every 30 
min.


{code:java}
  // try to keep the connection alive if the channel allows such things - 
websockets will
        if (channelizer.supportsKeepAlive() && keepAliveInterval > 0) {

            final ScheduledFuture oldKeepAliveFuture = 
keepAliveFuture.getAndSet(cluster.executor().scheduleAtFixedRate(() -> {
                logger.debug("Request sent to server to keep {} alive", 
thisConnection);
                try {
                    channel.writeAndFlush(channelizer.createKeepAliveMessage());
                } catch (Exception ex) {
                    // will just log this for now - a future real request can 
be responsible for the failure that
                    // marks the host as dead. this also may not mean the host 
is actually dead. more robust handling
                    // is in play for real requests, not this simple ping
                    logger.warn(String.format("Keep-alive did not succeed on 
%s", thisConnection), ex);
                }
            }, keepAliveInterval, keepAliveInterval, TimeUnit.MILLISECONDS));

            // try to cancel the old future if it's still un-executed - no need 
to ping since a new write has come
            // through on the connection
            if (oldKeepAliveFuture != null) oldKeepAliveFuture.cancel(true);
        }
{code}

The problem with this is that on every call of the {{wirte()}} method, which 
basically means on every query executed, the {[Connection}} schedule a 
KeepAlive task which won't be run for at least 30 minutes. This lead the 
{{Cluster.executor()}} queue to fill up with tasks waiting for completion.

One possible solution to fix this would be to schedule avoid this KeepAlive 
task to be instantiated more than once per connection:

{code:java}
final class Connection {
    ...
    private final AtomicBoolean keepAliveInitialized = new AtomicBoolean(false);

    public ChannelPromise write(final RequestMessage requestMessage, final 
CompletableFuture<ResultSet> future) {
    ...

        if (channelizer.supportsKeepAlive() && keepAliveInterval > 0 && 
keepAliveInitialized.compareAndSet(false, true)) {
        ...
        }
    ...
    }
   ...
}


}
{code}



> KeepAlive task executed for every Connection.write call
> -------------------------------------------------------
>
>                 Key: TINKERPOP-2030
>                 URL: https://issues.apache.org/jira/browse/TINKERPOP-2030
>             Project: TinkerPop
>          Issue Type: Bug
>          Components: driver
>    Affects Versions: 3.3.2
>            Reporter: Simone Rondelli
>            Priority: Major
>
> The Gremlin Driver communicates with the server via {{Connection.write()}}.
> The {{write()}} method has a logic in the end to schedule a task to keep the 
> connection alive, which by default is scheduled to be run after 30min, every 
> 30 min.
> {code:java}
>   // try to keep the connection alive if the channel allows such things - 
> websockets will
>         if (channelizer.supportsKeepAlive() && keepAliveInterval > 0) {
>             final ScheduledFuture oldKeepAliveFuture = 
> keepAliveFuture.getAndSet(cluster.executor().scheduleAtFixedRate(() -> {
>                 logger.debug("Request sent to server to keep {} alive", 
> thisConnection);
>                 try {
>                     
> channel.writeAndFlush(channelizer.createKeepAliveMessage());
>                 } catch (Exception ex) {
>                     // will just log this for now - a future real request can 
> be responsible for the failure that
>                     // marks the host as dead. this also may not mean the 
> host is actually dead. more robust handling
>                     // is in play for real requests, not this simple ping
>                     logger.warn(String.format("Keep-alive did not succeed on 
> %s", thisConnection), ex);
>                 }
>             }, keepAliveInterval, keepAliveInterval, TimeUnit.MILLISECONDS));
>             // try to cancel the old future if it's still un-executed - no 
> need to ping since a new write has come
>             // through on the connection
>             if (oldKeepAliveFuture != null) oldKeepAliveFuture.cancel(true);
>         }
> {code}
> The problem with this is that on every call of the {{wirte()}} method, which 
> basically means on every query executed, the {[Connection}} schedule a 
> KeepAlive task which won't be run for at least 30 minutes. This lead the 
> {{Cluster.executor()}} queue to fill up with tasks waiting for completion.
> One possible solution to fix this would be to schedule avoid this KeepAlive 
> task to be instantiated more than once per connection:
> {code:java}
> final class Connection {
>     ...
>     private final AtomicBoolean keepAliveInitialized = new 
> AtomicBoolean(false);
>     public ChannelPromise write(final RequestMessage requestMessage, final 
> CompletableFuture<ResultSet> future) {
>         ...
>         
>         // FIX HERE: with keepAliveInitialized.compareAndSet(false, true) we 
> ensure the keepAlive task is run only once per connection.
>         if (channelizer.supportsKeepAlive() && keepAliveInterval > 0 && 
> keepAliveInitialized.compareAndSet(false, true)) {
>             ...
>         }
>     }
> }
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to