[jira] [Comment Edited] (NIFI-12685) NiFi Cluster - java.net.SocketTimeoutException: timeout

2024-01-31 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-12685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17812750#comment-17812750
 ] 

Giovanni edited comment on NIFI-12685 at 1/31/24 3:27 PM:
--

After further analysis I found a recurring error on the primary node:
{code:java}
2024-01-31 15:14:21,158 ERROR [Load-Balanced Client Thread-3] 
o.a.n.c.q.c.c.a.n.NioAsyncLoadBalanceClient Unable to connect to 
tor1nifi1.skl.one:8443 for load balancing {code}
It was due to iptables that was blocking port 6342 
(nifi.cluster.load.balance.port) on every node. Adding the rule to iptables 
resolved the issue.

 

 


was (Author: JIRAUSER300198):
After further analysis I found a recurring error on the primary node:

2024-01-31 15:14:21,158 ERROR [Load-Balanced Client Thread-3] 
o.a.n.c.q.c.c.a.n.NioAsyncLoadBalanceClient Unable to connect to 
tor1nifi1.skl.one:8443 for load balancing

It was due to iptables that was blocking port 6342 
(nifi.cluster.load.balance.port) on every node. Adding the rule to iptables 
resolved the issue.



 

 

> NiFi Cluster - java.net.SocketTimeoutException: timeout 
> 
>
> Key: NIFI-12685
> URL: https://issues.apache.org/jira/browse/NIFI-12685
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.22.0
>Reporter: Giovanni
>Priority: Blocker
> Attachments: image-2024-01-30-10-19-30-651.png
>
>
> Hi,
> I have a 3 nodes cluster and every time I try to edit a workflow one of the 
> nodes goes timeout and UI return the following error:
> !image-2024-01-30-10-19-30-651.png!
>  
> At the same time one of the nodes reports the following log in the 
> nifi-app.log:
> {code:java}
> 2024-01-30 09:11:28,489 WARN [Replicate Request Thread-496] 
> o.a.n.c.c.h.r.ThreadPoolRequestReplicator 
> java.net.SocketTimeoutException: timeout
>         at okio.SocketAsyncTimeout.newTimeoutException(JvmOkio.kt:147)
>         at okio.AsyncTimeout.access$newTimeoutException(AsyncTimeout.kt:158)
>         at okio.AsyncTimeout$source$1.read(AsyncTimeout.kt:337)
>         at okio.RealBufferedSource.indexOf(RealBufferedSource.kt:427)
>         at 
> okio.RealBufferedSource.readUtf8LineStrict(RealBufferedSource.kt:320)
>         at okhttp3.internal.http1.HeadersReader.readLine(HeadersReader.kt:29)
>         at 
> okhttp3.internal.http1.Http1ExchangeCodec.readResponseHeaders(Http1ExchangeCodec.kt:180)
>         at 
> okhttp3.internal.connection.Exchange.readResponseHeaders(Exchange.kt:110)
>         at 
> okhttp3.internal.http.CallServerInterceptor.intercept(CallServerInterceptor.kt:93)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.kt:34)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.kt:95)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.kt:83)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.kt:76)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.connection.RealCall.getResponseWithInterceptorChain$okhttp(RealCall.kt:201)
>         at okhttp3.internal.connection.RealCall.execute(RealCall.kt:154)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.okhttp.OkHttpReplicationClient.replicate(OkHttpReplicationClient.java:136)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.okhttp.OkHttpReplicationClient.replicate(OkHttpReplicationClient.java:130)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.ThreadPoolRequestReplicator.replicateRequest(ThreadPoolRequestReplicator.java:645)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.ThreadPoolRequestReplicator$NodeHttpRequest.run(ThreadPoolRequestReplicator.java:869)
>         at 
> java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
>         at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>         at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>         at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>         at java.base/java.lang.Thread.run(Thread.java:829)
> Caused by: java.net.SocketException: Socket closed
>         at 
> java.base/java.net.SocketInputStream.read(SocketInputStream.java:183)
>   

[jira] [Resolved] (NIFI-12685) NiFi Cluster - java.net.SocketTimeoutException: timeout

2024-01-31 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-12685?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni resolved NIFI-12685.
-
Resolution: Not A Bug

After further analysis I found a recurring error on the primary node:

2024-01-31 15:14:21,158 ERROR [Load-Balanced Client Thread-3] 
o.a.n.c.q.c.c.a.n.NioAsyncLoadBalanceClient Unable to connect to 
tor1nifi1.skl.one:8443 for load balancing

It was due to iptables that was blocking port 6342 
(nifi.cluster.load.balance.port) on every node. Adding the rule to iptables 
resolved the issue.



 

 

> NiFi Cluster - java.net.SocketTimeoutException: timeout 
> 
>
> Key: NIFI-12685
> URL: https://issues.apache.org/jira/browse/NIFI-12685
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.22.0
>Reporter: Giovanni
>Priority: Blocker
> Attachments: image-2024-01-30-10-19-30-651.png
>
>
> Hi,
> I have a 3 nodes cluster and every time I try to edit a workflow one of the 
> nodes goes timeout and UI return the following error:
> !image-2024-01-30-10-19-30-651.png!
>  
> At the same time one of the nodes reports the following log in the 
> nifi-app.log:
> {code:java}
> 2024-01-30 09:11:28,489 WARN [Replicate Request Thread-496] 
> o.a.n.c.c.h.r.ThreadPoolRequestReplicator 
> java.net.SocketTimeoutException: timeout
>         at okio.SocketAsyncTimeout.newTimeoutException(JvmOkio.kt:147)
>         at okio.AsyncTimeout.access$newTimeoutException(AsyncTimeout.kt:158)
>         at okio.AsyncTimeout$source$1.read(AsyncTimeout.kt:337)
>         at okio.RealBufferedSource.indexOf(RealBufferedSource.kt:427)
>         at 
> okio.RealBufferedSource.readUtf8LineStrict(RealBufferedSource.kt:320)
>         at okhttp3.internal.http1.HeadersReader.readLine(HeadersReader.kt:29)
>         at 
> okhttp3.internal.http1.Http1ExchangeCodec.readResponseHeaders(Http1ExchangeCodec.kt:180)
>         at 
> okhttp3.internal.connection.Exchange.readResponseHeaders(Exchange.kt:110)
>         at 
> okhttp3.internal.http.CallServerInterceptor.intercept(CallServerInterceptor.kt:93)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.kt:34)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.kt:95)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.kt:83)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.kt:76)
>         at 
> okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>         at 
> okhttp3.internal.connection.RealCall.getResponseWithInterceptorChain$okhttp(RealCall.kt:201)
>         at okhttp3.internal.connection.RealCall.execute(RealCall.kt:154)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.okhttp.OkHttpReplicationClient.replicate(OkHttpReplicationClient.java:136)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.okhttp.OkHttpReplicationClient.replicate(OkHttpReplicationClient.java:130)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.ThreadPoolRequestReplicator.replicateRequest(ThreadPoolRequestReplicator.java:645)
>         at 
> org.apache.nifi.cluster.coordination.http.replication.ThreadPoolRequestReplicator$NodeHttpRequest.run(ThreadPoolRequestReplicator.java:869)
>         at 
> java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
>         at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>         at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>         at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>         at java.base/java.lang.Thread.run(Thread.java:829)
> Caused by: java.net.SocketException: Socket closed
>         at 
> java.base/java.net.SocketInputStream.read(SocketInputStream.java:183)
>         at 
> java.base/java.net.SocketInputStream.read(SocketInputStream.java:140)
>         at 
> java.base/sun.security.ssl.SSLSocketInputRecord.read(SSLSocketInputRecord.java:484)
>         at 
> java.base/sun.security.ssl.SSLSocketInputRecord.readHeader(SSLSocketInputRecord.java:478)
>         at 
> java.base/sun.security.ssl.SSLSocketInputRecord.bytesInCompletePacket(SSLSocketInputRecord.java:70)
>         at 
> java.base/sun.security.ssl.SSLSocketImpl.readApplicationRecord(SSLSocketImpl.java:1455)

[jira] [Created] (NIFI-12685) NiFi Cluster - java.net.SocketTimeoutException: timeout

2024-01-30 Thread Giovanni (Jira)
Giovanni created NIFI-12685:
---

 Summary: NiFi Cluster - java.net.SocketTimeoutException: timeout 
 Key: NIFI-12685
 URL: https://issues.apache.org/jira/browse/NIFI-12685
 Project: Apache NiFi
  Issue Type: Bug
Affects Versions: 1.22.0
Reporter: Giovanni
 Attachments: image-2024-01-30-10-19-30-651.png

Hi,

I have a 3 nodes cluster and every time I try to edit a workflow one of the 
nodes goes timeout and UI return the following error:

!image-2024-01-30-10-19-30-651.png!

 

At the same time one of the nodes reports the following log in the nifi-app.log:


{code:java}
2024-01-30 09:11:28,489 WARN [Replicate Request Thread-496] 
o.a.n.c.c.h.r.ThreadPoolRequestReplicator 
java.net.SocketTimeoutException: timeout
        at okio.SocketAsyncTimeout.newTimeoutException(JvmOkio.kt:147)
        at okio.AsyncTimeout.access$newTimeoutException(AsyncTimeout.kt:158)
        at okio.AsyncTimeout$source$1.read(AsyncTimeout.kt:337)
        at okio.RealBufferedSource.indexOf(RealBufferedSource.kt:427)
        at okio.RealBufferedSource.readUtf8LineStrict(RealBufferedSource.kt:320)
        at okhttp3.internal.http1.HeadersReader.readLine(HeadersReader.kt:29)
        at 
okhttp3.internal.http1.Http1ExchangeCodec.readResponseHeaders(Http1ExchangeCodec.kt:180)
        at 
okhttp3.internal.connection.Exchange.readResponseHeaders(Exchange.kt:110)
        at 
okhttp3.internal.http.CallServerInterceptor.intercept(CallServerInterceptor.kt:93)
        at 
okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
        at 
okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.kt:34)
        at 
okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
        at 
okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.kt:95)
        at 
okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
        at 
okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.kt:83)
        at 
okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
        at 
okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.kt:76)
        at 
okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
        at 
okhttp3.internal.connection.RealCall.getResponseWithInterceptorChain$okhttp(RealCall.kt:201)
        at okhttp3.internal.connection.RealCall.execute(RealCall.kt:154)
        at 
org.apache.nifi.cluster.coordination.http.replication.okhttp.OkHttpReplicationClient.replicate(OkHttpReplicationClient.java:136)
        at 
org.apache.nifi.cluster.coordination.http.replication.okhttp.OkHttpReplicationClient.replicate(OkHttpReplicationClient.java:130)
        at 
org.apache.nifi.cluster.coordination.http.replication.ThreadPoolRequestReplicator.replicateRequest(ThreadPoolRequestReplicator.java:645)
        at 
org.apache.nifi.cluster.coordination.http.replication.ThreadPoolRequestReplicator$NodeHttpRequest.run(ThreadPoolRequestReplicator.java:869)
        at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
        at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
        at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: java.net.SocketException: Socket closed
        at java.base/java.net.SocketInputStream.read(SocketInputStream.java:183)
        at java.base/java.net.SocketInputStream.read(SocketInputStream.java:140)
        at 
java.base/sun.security.ssl.SSLSocketInputRecord.read(SSLSocketInputRecord.java:484)
        at 
java.base/sun.security.ssl.SSLSocketInputRecord.readHeader(SSLSocketInputRecord.java:478)
        at 
java.base/sun.security.ssl.SSLSocketInputRecord.bytesInCompletePacket(SSLSocketInputRecord.java:70)
        at 
java.base/sun.security.ssl.SSLSocketImpl.readApplicationRecord(SSLSocketImpl.java:1455)
        at 
java.base/sun.security.ssl.SSLSocketImpl$AppInputStream.read(SSLSocketImpl.java:1066)
        at okio.InputStreamSource.read(JvmOkio.kt:94)
        at okio.AsyncTimeout$source$1.read(AsyncTimeout.kt:125)
        ... 26 common frames omitted {code}

As I thought it was a performance issue, I doubled each nodes resources but the 
error is still occurring.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-12528) Funnels bug

2023-12-21 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-12528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-12528:

Description: 
I've notice a bug in our working cluster of 4 nodes.
When two funnels are connected, with no other connections as themselves, like 
in attached image, is not possible to delete them. 
After trying to modify one of their connections, the Node where the action is 
performed crashes, and will not turn on unless the entire cluster gets 
restarted.

  was:
I've notice a bug in our working cluster of 4 nodes.
When two funnels are connected, with no other connections as themselves, like 
in attached image, is not possible to delete them. 
After trying to modify one of their connections, the Node where the action is 
performed crashes, and will not turn on unless the entire cluster restarts.


> Funnels bug
> ---
>
> Key: NIFI-12528
> URL: https://issues.apache.org/jira/browse/NIFI-12528
> Project: Apache NiFi
>  Issue Type: Bug
>  Components: Core Framework
>Affects Versions: 1.23.2
> Environment: NiFi cluster
>Reporter: Giovanni
>Priority: Critical
> Attachments: image001.png
>
>
> I've notice a bug in our working cluster of 4 nodes.
> When two funnels are connected, with no other connections as themselves, like 
> in attached image, is not possible to delete them. 
> After trying to modify one of their connections, the Node where the action is 
> performed crashes, and will not turn on unless the entire cluster gets 
> restarted.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (NIFI-12528) Funnels bug

2023-12-20 Thread Giovanni (Jira)
Giovanni created NIFI-12528:
---

 Summary: Funnels bug
 Key: NIFI-12528
 URL: https://issues.apache.org/jira/browse/NIFI-12528
 Project: Apache NiFi
  Issue Type: Bug
  Components: Core Framework
Affects Versions: 1.23.2
 Environment: NiFi cluster
Reporter: Giovanni
 Attachments: image001.png

I've notice a bug in our working cluster of 4 nodes.
When two funnels are connected, with no other connections as themselves, like 
in attached image, is not possible to delete them. 
After trying to modify one of their connections, the Node where the action is 
performed crashes, and will not turn on unless the entire cluster restarts.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-12442) Adding support for RocksDB

2023-12-10 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-12442?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17795080#comment-17795080
 ] 

Giovanni commented on NIFI-12442:
-

Dear David, 

As you suggested, I've implemented the DistributedMapCacheClient interface, 
while still keeping the old components within the project.

There's a pre-release version, feel free to try it.
When creating the services, you can add dynamic properties to add RocksOptions 
to the RocksDb instance. I suggest you to use the option: "setCreateIfMissing", 
"true", to create automatically the RocksDb instance if not existing. 

I'm up to other suggestions. Soon I'll release it to maven central.

Here's the link to the pre-release:
[Pre-release 
v1.0.1|https://github.com/kommpn/nifi-rocksdb-manager/releases/tag/v1.0.1]

> Adding support for RocksDB
> --
>
> Key: NIFI-12442
> URL: https://issues.apache.org/jira/browse/NIFI-12442
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.23.2
>Reporter: Giovanni
>Priority: Minor
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> I would like to suggest the creation of 3 new components:
> The first one is a service, which opens an existing RocksDb or, eventually, 
> using RocksOptions, create it from scratch. It will manage all the open 
> options (classic, read/write, only read, secondary).
> The second one, is a RocksDbReader, that uses the service to communicate with 
> the RocksDb in order to retrieve informations through a lookup. It can save 
> the searched content inside an attribute or inside the flowFile content. It 
> will be capable of using both APIs, such as simple "db.get" and via 
> RocksIterator.
> The last one, is a RocksDbWriter, that uses the same service as the reader, 
> but can write values inside the RocksDb, both from flowFile attribute or 
> flowFile content, using properties to determine the key to use.
>  
> Feel free to express your opinions, if you think this will be useful or 
> useless.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (NIFI-12442) Adding support for RocksDB

2023-12-10 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-12442?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17795080#comment-17795080
 ] 

Giovanni edited comment on NIFI-12442 at 12/10/23 1:59 PM:
---

Dear David, 

As you suggested, I've implemented the DistributedMapCacheClient interface, 
while still keeping the old components within the project.

There's a pre-release version, feel free to try it.
When creating the services, you can add dynamic properties to add RocksOptions 
to the RocksDb instance. I suggest you to use the option: "setCreateIfMissing", 
"true", to create automatically the RocksDb instance if not existing. 

I'm up to other suggestions. Soon I'll release it to maven central.

Here's the link to the pre-release with the NAR file in it:
[Pre-release 
v1.0.1|https://github.com/kommpn/nifi-rocksdb-manager/releases/tag/v1.0.1]


was (Author: JIRAUSER301883):
Dear David, 

As you suggested, I've implemented the DistributedMapCacheClient interface, 
while still keeping the old components within the project.

There's a pre-release version, feel free to try it.
When creating the services, you can add dynamic properties to add RocksOptions 
to the RocksDb instance. I suggest you to use the option: "setCreateIfMissing", 
"true", to create automatically the RocksDb instance if not existing. 

I'm up to other suggestions. Soon I'll release it to maven central.

Here's the link to the pre-release:
[Pre-release 
v1.0.1|https://github.com/kommpn/nifi-rocksdb-manager/releases/tag/v1.0.1]

> Adding support for RocksDB
> --
>
> Key: NIFI-12442
> URL: https://issues.apache.org/jira/browse/NIFI-12442
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.23.2
>Reporter: Giovanni
>Priority: Minor
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> I would like to suggest the creation of 3 new components:
> The first one is a service, which opens an existing RocksDb or, eventually, 
> using RocksOptions, create it from scratch. It will manage all the open 
> options (classic, read/write, only read, secondary).
> The second one, is a RocksDbReader, that uses the service to communicate with 
> the RocksDb in order to retrieve informations through a lookup. It can save 
> the searched content inside an attribute or inside the flowFile content. It 
> will be capable of using both APIs, such as simple "db.get" and via 
> RocksIterator.
> The last one, is a RocksDbWriter, that uses the same service as the reader, 
> but can write values inside the RocksDb, both from flowFile attribute or 
> flowFile content, using properties to determine the key to use.
>  
> Feel free to express your opinions, if you think this will be useful or 
> useless.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-12442) Adding support for RocksDB

2023-12-07 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-12442?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17794129#comment-17794129
 ] 

Giovanni commented on NIFI-12442:
-

Dear [~exceptionfactory], I've created a simple repository with a pre-release. 
Feel free to give me feedbacks. 
https://github.com/kommpn/nifi-rocksdb-manager

> Adding support for RocksDB
> --
>
> Key: NIFI-12442
> URL: https://issues.apache.org/jira/browse/NIFI-12442
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.23.2
>Reporter: Giovanni
>Priority: Minor
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> I would like to suggest the creation of 3 new components:
> The first one is a service, which opens an existing RocksDb or, eventually, 
> using RocksOptions, create it from scratch. It will manage all the open 
> options (classic, read/write, only read, secondary).
> The second one, is a RocksDbReader, that uses the service to communicate with 
> the RocksDb in order to retrieve informations through a lookup. It can save 
> the searched content inside an attribute or inside the flowFile content. It 
> will be capable of using both APIs, such as simple "db.get" and via 
> RocksIterator.
> The last one, is a RocksDbWriter, that uses the same service as the reader, 
> but can write values inside the RocksDb, both from flowFile attribute or 
> flowFile content, using properties to determine the key to use.
>  
> Feel free to express your opinions, if you think this will be useful or 
> useless.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-12442) Adding support for RocksDB

2023-12-03 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-12442?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17792522#comment-17792522
 ] 

Giovanni commented on NIFI-12442:
-

Thanks for your considerations.

Surely the advantage is the speed of calculation. A data flow possibly enrich 
data with half a million checks on 30 seconds with 2 clustered nodes (in my 
private experience, with my local clustered components). It's way efficient 
than other solutions. Obviously is up to the user to choose which one solution 
fits better for its purpose. At FS space cost and possibly RAM (depending of 
which RocksOptions the user choose), it could delivery a fast output.

I understand the problematic relative to the maintainability and compatibility 
could be a problem.

As you said, it's probably a better solution to create an external library, 
that receive constant updates, that a possible RocksDb chooser user could use.

Thank you, at this point I think that this Ticket could be closed.

> Adding support for RocksDB
> --
>
> Key: NIFI-12442
> URL: https://issues.apache.org/jira/browse/NIFI-12442
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.23.2
>Reporter: Giovanni
>Priority: Minor
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> I would like to suggest the creation of 3 new components:
> The first one is a service, which opens an existing RocksDb or, eventually, 
> using RocksOptions, create it from scratch. It will manage all the open 
> options (classic, read/write, only read, secondary).
> The second one, is a RocksDbReader, that uses the service to communicate with 
> the RocksDb in order to retrieve informations through a lookup. It can save 
> the searched content inside an attribute or inside the flowFile content. It 
> will be capable of using both APIs, such as simple "db.get" and via 
> RocksIterator.
> The last one, is a RocksDbWriter, that uses the same service as the reader, 
> but can write values inside the RocksDb, both from flowFile attribute or 
> flowFile content, using properties to determine the key to use.
>  
> Feel free to express your opinions, if you think this will be useful or 
> useless.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-12442) Adding support for RocksDB

2023-11-30 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-12442?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17791717#comment-17791717
 ] 

Giovanni commented on NIFI-12442:
-

Thanks David for answering this question. 
The use case that you proposed is exactly the one I was thinking of. Working on 
multiple nodes, data will be written and read in all nodes FS, maintaining 
databases consistency. 
A potential use case can be the one where a Rocksdb is created in a separate 
instance, loaded inside NiFi fs via S3, and read/updated via data flow, with 
updated data, and then re-uploaded to S3. This is a simple use case, where a 
user has total management power on his RocksDb. Then data inside them can be 
used to enrich data retrieved from the internet. 
As you said, this could probably not be much viable in the official repo. But 
can still be useful in certain cases.

> Adding support for RocksDB
> --
>
> Key: NIFI-12442
> URL: https://issues.apache.org/jira/browse/NIFI-12442
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.23.2
>Reporter: Giovanni
>Priority: Minor
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> I would like to suggest the creation of 3 new components:
> The first one is a service, which opens an existing RocksDb or, eventually, 
> using RocksOptions, create it from scratch. It will manage all the open 
> options (classic, read/write, only read, secondary).
> The second one, is a RocksDbReader, that uses the service to communicate with 
> the RocksDb in order to retrieve informations through a lookup. It can save 
> the searched content inside an attribute or inside the flowFile content. It 
> will be capable of using both APIs, such as simple "db.get" and via 
> RocksIterator.
> The last one, is a RocksDbWriter, that uses the same service as the reader, 
> but can write values inside the RocksDb, both from flowFile attribute or 
> flowFile content, using properties to determine the key to use.
>  
> Feel free to express your opinions, if you think this will be useful or 
> useless.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (NIFI-12442) Adding support for RocksDB

2023-11-30 Thread Giovanni (Jira)
Giovanni created NIFI-12442:
---

 Summary: Adding support for RocksDB
 Key: NIFI-12442
 URL: https://issues.apache.org/jira/browse/NIFI-12442
 Project: Apache NiFi
  Issue Type: Improvement
Affects Versions: 1.23.2
Reporter: Giovanni


I would like to suggest the creation of 3 new components:
The first one is a service, which opens an existing RocksDb or, eventually, 
using RocksOptions, create it from scratch. It will manage all the open options 
(classic, read/write, only read, secondary).

The second one, is a RocksDbReader, that uses the service to communicate with 
the RocksDb in order to retrieve informations through a lookup. It can save the 
searched content inside an attribute or inside the flowFile content. It will be 
capable of using both APIs, such as simple "db.get" and via RocksIterator.

The last one, is a RocksDbWriter, that uses the same service as the reader, but 
can write values inside the RocksDb, both from flowFile attribute or flowFile 
content, using properties to determine the key to use.

 

Feel free to express your opinions, if you think this will be useful or useless.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (NIFI-11764) flowfile storage size mismatch

2023-06-29 Thread Giovanni (Jira)
Giovanni created NIFI-11764:
---

 Summary: flowfile storage size mismatch
 Key: NIFI-11764
 URL: https://issues.apache.org/jira/browse/NIFI-11764
 Project: Apache NiFi
  Issue Type: Improvement
Reporter: Giovanni
 Attachments: image-2023-06-29-10-24-00-879.png, 
image-2023-06-29-10-25-06-665.png, image-2023-06-29-10-27-17-702.png, 
image-2023-06-29-10-28-45-547.png, image-2023-06-29-10-30-07-266.png, 
image-2023-06-29-10-31-33-526.png

Hi,

Nifi is reporting a wrong value for the flowfile storage used space.

If I check the dashboard it reports 528.05 MB on all nodes:

!image-2023-06-29-10-24-00-879.png|width=574,height=206!

The values are also confirmed by the nodes status history:

!image-2023-06-29-10-25-06-665.png|width=584,height=431!

However my monitoring tool reports 52KB only:

!image-2023-06-29-10-27-17-702.png|width=997,height=165!

These low values are confirmed on the hosts themselves:

!image-2023-06-29-10-28-45-547.png|width=485,height=179!

!image-2023-06-29-10-30-07-266.png|width=484,height=178!

!image-2023-06-29-10-31-33-526.png|width=485,height=176!

The flowfile repository settings are the same on all nodes:
 * nifi1

{code:java}
# FlowFile Repository
nifi.flowfile.repository.implementation=org.apache.nifi.controller.repository.WriteAheadFlowFileRepository
nifi.flowfile.repository.wal.implementation=org.apache.nifi.wali.SequentialAccessWriteAheadLog
nifi.flowfile.repository.directory=/var/nifi/flowfile_repo/data
nifi.flowfile.repository.checkpoint.interval=20 secs
nifi.flowfile.repository.always.sync=false
nifi.flowfile.repository.retain.orphaned.flowfiles=truenifi.swap.manager.implementation=org.apache.nifi.controller.FileSystemSwapManager
nifi.queue.swap.threshold=2{code}

 * nifi2

{code:java}
# FlowFile Repository
nifi.flowfile.repository.implementation=org.apache.nifi.controller.repository.WriteAheadFlowFileRepository
nifi.flowfile.repository.wal.implementation=org.apache.nifi.wali.SequentialAccessWriteAheadLog
nifi.flowfile.repository.directory=/var/nifi/flowfile_repo/data
nifi.flowfile.repository.checkpoint.interval=20 secs
nifi.flowfile.repository.always.sync=false
nifi.flowfile.repository.retain.orphaned.flowfiles=truenifi.swap.manager.implementation=org.apache.nifi.controller.FileSystemSwapManager
nifi.queue.swap.threshold=2 {code}

 * nifi3

{code:java}
# FlowFile Repository
nifi.flowfile.repository.implementation=org.apache.nifi.controller.repository.WriteAheadFlowFileRepository
nifi.flowfile.repository.wal.implementation=org.apache.nifi.wali.SequentialAccessWriteAheadLog
nifi.flowfile.repository.directory=/var/nifi/flowfile_repo/data
nifi.flowfile.repository.checkpoint.interval=20 secs
nifi.flowfile.repository.always.sync=false
nifi.flowfile.repository.retain.orphaned.flowfiles=truenifi.swap.manager.implementation=org.apache.nifi.controller.FileSystemSwapManager
nifi.queue.swap.threshold=2 {code}

I also monitored the disk usage in realtime and it reports less than 100MB on 
each node for a very few seconds. However still far from the steady values 
reported in the GUI.

 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-27 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17737509#comment-17737509
 ] 

Giovanni commented on NIFI-11530:
-

I got it.

Thanks for your help [~joewitt] 

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-27 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni resolved NIFI-11530.
-
Resolution: Not A Bug

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17737221#comment-17737221
 ] 

Giovanni edited comment on NIFI-11530 at 6/26/23 4:06 PM:
--

After I configured the volumes I also set the provenance max size to 9GB:
{code:java}
# Persistent Provenance Repository Properties
nifi.provenance.repository.directory.default=/var/nifi/provenance_repo/data
nifi.provenance.repository.max.storage.time=30 days
nifi.provenance.repository.max.storage.size=9 GB
nifi.provenance.repository.rollover.time=10 mins
nifi.provenance.repository.rollover.size=100 MB
nifi.provenance.repository.query.threads=2
nifi.provenance.repository.index.threads=2
nifi.provenance.repository.compress.on.rollover=true
nifi.provenance.repository.always.sync=false {code}
So 8.98GB could be a normal size according to the setting or should I expect 
the repo size to be lower?

 


was (Author: JIRAUSER300198):
After I configured the volumes I also set the provenance max size to 9GB:

 
{code:java}
# Persistent Provenance Repository Properties
nifi.provenance.repository.directory.default=/var/nifi/provenance_repo/data
nifi.provenance.repository.max.storage.time=30 days
nifi.provenance.repository.max.storage.size=9 GB
nifi.provenance.repository.rollover.time=10 mins
nifi.provenance.repository.rollover.size=100 MB
nifi.provenance.repository.query.threads=2
nifi.provenance.repository.index.threads=2
nifi.provenance.repository.compress.on.rollover=true
nifi.provenance.repository.always.sync=false {code}
So 8.98GB could be a normal size according to the setting or should I expect 
the repo size to be lower?

 

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17737221#comment-17737221
 ] 

Giovanni commented on NIFI-11530:
-

After I configured the volumes I also set the provenance max size to 9GB:

 
{code:java}
# Persistent Provenance Repository Properties
nifi.provenance.repository.directory.default=/var/nifi/provenance_repo/data
nifi.provenance.repository.max.storage.time=30 days
nifi.provenance.repository.max.storage.size=9 GB
nifi.provenance.repository.rollover.time=10 mins
nifi.provenance.repository.rollover.size=100 MB
nifi.provenance.repository.query.threads=2
nifi.provenance.repository.index.threads=2
nifi.provenance.repository.compress.on.rollover=true
nifi.provenance.repository.always.sync=false {code}
So 8.98GB could be a normal size according to the setting or should I expect 
the repo size to be lower?

 

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: image-2023-06-26-11-16-48-186.png

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17737060#comment-17737060
 ] 

Giovanni commented on NIFI-11530:
-

Update:

I reconfigured the repository to have a logical volume for each:

!image-2023-06-26-11-13-19-289.png!

The overall performance are improved.

However the provenance repository is still problematic:

!image-2023-06-26-11-15-00-379.png!

The other repos are ok though:

!image-2023-06-26-11-16-21-096.png!

!image-2023-06-26-11-16-35-791.png!

!image-2023-06-26-11-16-48-186.png!

 

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: image-2023-06-26-11-16-21-096.png

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: image-2023-06-26-11-16-35-791.png

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, image-2023-06-26-11-16-21-096.png, 
> image-2023-06-26-11-16-35-791.png, image-2023-06-26-11-16-48-186.png, 
> jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: image-2023-06-26-11-15-00-379.png

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, 
> image-2023-06-26-11-15-00-379.png, jvm.jpg, nifi1-app.log, nifi2-app.log, 
> nifi3-app.log, nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-26 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: image-2023-06-26-11-13-19-289.png

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, image-2023-06-26-11-13-19-289.png, jvm.jpg, 
> nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-08 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17730426#comment-17730426
 ] 

Giovanni edited comment on NIFI-11530 at 6/8/23 7:40 AM:
-

Hi Joe,

so what is the recommended configuration?

Should separate logical volumes be enough? Also 1 volume per repository?

Consider my nifi setup is a 3 nodes cluster.


was (Author: JIRAUSER300198):
Hi Joe,

so what is the recommended configuration?

Consider my nifi setup is a 3 nodes cluster.

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-08 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17730426#comment-17730426
 ] 

Giovanni commented on NIFI-11530:
-

Hi Joe,

so what is the recommended configuration?

Consider my nifi setup is a 3 nodes cluster.

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-06 Thread Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17729762#comment-17729762
 ] 

Giovanni commented on NIFI-11530:
-

Hi Joe,

thanks for your reply.

Here are my settings:

 
{code:java}
# FlowFile Repository
nifi.flowfile.repository.implementation=org.apache.nifi.controller.repository.WriteAheadFlowFileRepository
nifi.flowfile.repository.wal.implementation=org.apache.nifi.wali.SequentialAccessWriteAheadLog
nifi.flowfile.repository.directory=./flowfile_repository
nifi.flowfile.repository.checkpoint.interval=20 secs
nifi.flowfile.repository.always.sync=false
nifi.flowfile.repository.retain.orphaned.flowfiles=true
nifi.swap.manager.implementation=org.apache.nifi.controller.FileSystemSwapManager
nifi.queue.swap.threshold=2{code}
{code:java}
# Content Repository
nifi.content.repository.implementation=org.apache.nifi.controller.repository.FileSystemRepository
nifi.content.claim.max.appendable.size=50 KB
nifi.content.repository.directory.default=./content_repository
nifi.content.repository.archive.max.retention.period=7 days
nifi.content.repository.archive.max.usage.percentage=50%
nifi.content.repository.archive.enabled=true
nifi.content.repository.always.sync=false
nifi.content.viewer.url=../nifi-content-viewer/{code}
{code:java}
# Provenance Repository Properties
nifi.provenance.repository.implementation=org.apache.nifi.provenance.WriteAheadProvenanceRepository
# Persistent Provenance Repository Properties
nifi.provenance.repository.directory.default=./provenance_repository
nifi.provenance.repository.max.storage.time=30 days
nifi.provenance.repository.max.storage.size=10 GB
nifi.provenance.repository.rollover.time=10 mins
nifi.provenance.repository.rollover.size=100 MB
nifi.provenance.repository.query.threads=2
nifi.provenance.repository.index.threads=2
nifi.provenance.repository.compress.on.rollover=true
nifi.provenance.repository.always.sync=false
# Comma-separated list of fields. Fields that are not indexed will not be 
searchable. Valid fields are:
# EventType, FlowFileUUID, Filename, TransitURI, ProcessorID, 
AlternateIdentifierURI, Relationship, Details
nifi.provenance.repository.indexed.fields=EventType, FlowFileUUID, Filename, 
ProcessorID, Relationship
# FlowFile Attributes that should be indexed and made searchable.  Some 
examples to consider are filename, uuid, mime.type
nifi.provenance.repository.indexed.attributes=
# Large values for the shard size will result in more Java heap usage when 
searching the Provenance Repository
# but should provide better performance
nifi.provenance.repository.index.shard.size=500 MB
# Indicates the maximum length that a FlowFile attribute can be when retrieving 
a Provenance Event from
# the repository. If the length of any attribute exceeds this value, it will be 
truncated when the event is retrieved.
nifi.provenance.repository.max.attribute.length=65536
nifi.provenance.repository.concurrent.merge.threads=2

# Volatile Provenance Respository Properties
nifi.provenance.repository.buffer.size=10{code}
In response to your point, I confirm each node has only a single disk.

 

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-05 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Description: 
Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.

 

UPDATE:

I'm having the problem again.

Every archive size is more than 50% on each node, with 70%+ peak on coordinator 
node (see attachments).

I'm also attaching nifi-app.log this time.

  was:
Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.

UPDATE:

I'm having the problem again.

Every archive size is more than 50% on each node, with 70%+ peak on coordinator 
node (see attachments).

I'm also attaching nifi-app.log this time.


> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
>  
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-05 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Description: 
Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.

UPDATE:

I'm having the problem again.

Every archive size is more than 50% on each node, with 70%+ peak on coordinator 
node (see attachments).

I'm also attaching nifi-app.log this time.

  was:
Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.

UPDATE:

I'm having the problem again.

Every archive size is more than 50% on each node, with 70%+ peak on coordinator 
node (see attachments).

I'm also attaching nifi-app.log that is reporting the following error:


> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log this time.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-05 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: 20230605_disk_usage.jpg

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: 20230605_disk_usage.jpg, content_archive.jpg, 
> flowfile_archive.jpg, jvm.jpg, nifi1-app.log, nifi2-app.log, nifi3-app.log, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log that is reporting the following error:



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-05 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Description: 
Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.

UPDATE:

I'm having the problem again.

Every archive size is more than 50% on each node, with 70%+ peak on coordinator 
node (see attachments).

I'm also attaching nifi-app.log that is reporting the following error:

  was:
Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.


> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: content_archive.jpg, flowfile_archive.jpg, jvm.jpg, 
> nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.
> UPDATE:
> I'm having the problem again.
> Every archive size is more than 50% on each node, with 70%+ peak on 
> coordinator node (see attachments).
> I'm also attaching nifi-app.log that is reporting the following error:



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-05 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: nifi1-app.log
nifi2-app.log
nifi3-app.log

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: content_archive.jpg, flowfile_archive.jpg, jvm.jpg, 
> nifi1-app.log, nifi2-app.log, nifi3-app.log, nifi_bug.jpg, 
> provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-06-05 Thread Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-11530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni updated NIFI-11530:

Attachment: content_archive.jpg
flowfile_archive.jpg
jvm.jpg
provenance_archive.jpg

> Disk full even with nifi.content.repository.archive.max.usage.percentage set 
> to 50%
> ---
>
> Key: NIFI-11530
> URL: https://issues.apache.org/jira/browse/NIFI-11530
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.20.0
> Environment: Ubuntu 20.04.5 LTS
>Reporter: Giovanni
>Priority: Major
> Attachments: content_archive.jpg, flowfile_archive.jpg, jvm.jpg, 
> nifi_bug.jpg, provenance_archive.jpg
>
>
> Nifi primary node reports disk full causing all nodes to stop working.
> Restarting nifi service does not resolve.
> Restarting the VM does not resolve.
> The only way to fix is to clean te content_repository dir:
> rm -rf ./nifi/content_repository/*
>  
> Unfortunately I have no logs of the issue ongoing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (NIFI-11530) Disk full even with nifi.content.repository.archive.max.usage.percentage set to 50%

2023-05-08 Thread Giovanni (Jira)
Giovanni created NIFI-11530:
---

 Summary: Disk full even with 
nifi.content.repository.archive.max.usage.percentage set to 50%
 Key: NIFI-11530
 URL: https://issues.apache.org/jira/browse/NIFI-11530
 Project: Apache NiFi
  Issue Type: Bug
Affects Versions: 1.20.0
 Environment: Ubuntu 20.04.5 LTS
Reporter: Giovanni
 Attachments: nifi_bug.jpg

Nifi primary node reports disk full causing all nodes to stop working.

Restarting nifi service does not resolve.

Restarting the VM does not resolve.

The only way to fix is to clean te content_repository dir:

rm -rf ./nifi/content_repository/*

 

Unfortunately I have no logs of the issue ongoing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-10869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi De Giovanni updated NIFI-10869:
-
Priority: Minor  (was: Major)

> ExtractText processor - RegEx captured twice
> 
>
> Key: NIFI-10869
> URL: https://issues.apache.org/jira/browse/NIFI-10869
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.18.0
>Reporter: Luigi De Giovanni
>Priority: Minor
> Attachments: image (1).png, image (2).png, image.png
>
>
> Hi,
> There is an issue with the ExtractText processor, in the attempt of creating 
> FlowFile attributes from FlowFile contents.
> When creating a custom property with a regex value that does not contain 
> named groups, the value captured in the group is added as 2 different 
> attributes.
> E.G.
> ||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
> |param.alpha||param.alpha=alpha|param.alpha=alpha
> param.alpha.1=alpha|
> Reading the documentation, this might even be an expected behaviour, but if 
> so, it is preferable to have the captured value only as a single attribute, 
> without duplication.
> Please see the attachment for an example.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-10869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi De Giovanni updated NIFI-10869:
-
Issue Type: Improvement  (was: Bug)

> ExtractText processor - RegEx captured twice
> 
>
> Key: NIFI-10869
> URL: https://issues.apache.org/jira/browse/NIFI-10869
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.18.0
>Reporter: Luigi De Giovanni
>Priority: Major
> Attachments: image (1).png, image (2).png, image.png
>
>
> Hi,
> There is an issue with the ExtractText processor, in the attempt of creating 
> FlowFile attributes from FlowFile contents.
> When creating a custom property with a regex value that does not contain 
> named groups, the value captured in the group is added as 2 different 
> attributes.
> E.G.
> ||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
> |param.alpha||param.alpha=alpha|param.alpha=alpha
> param.alpha.1=alpha|
> Reading the documentation, this might even be an expected behaviour, but if 
> so, it is preferable to have the captured value only as a single attribute, 
> without duplication.
> Please see the attachment for an example.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-10869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi De Giovanni updated NIFI-10869:
-
Description: 
Hi,

There is an issue with the ExtractText processor, in the attempt of creating 
FlowFile attributes from FlowFile contents.

When creating a custom property with a regex value that does not contain named 
groups, the value captured in the group is added as 2 different attributes.

E.G.
||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
|param.alpha||param.alpha=alpha|param.alpha=alpha
param.alpha.1=alpha|

Reading the documentation, this might even be an expected behaviour, but if so, 
it is preferable to have the captured value only as a single attribute, without 
duplication.

Please see the attachment for an example.

 

  was:
Hi,

There is an issue with the ExtractText processor, in the attempt of creating 
FlowFile attributes from FlowFile contents.

When creating a custom property with a regex value that does not contain named 
groups, the value captured in the group is added as 2 different attributes.

E.G.
||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
|request.param.alpha||request.param.alpha=alpha|request.param.alpha=alpha
request.param.alpha.1=alpha|

Reading the documentation, this might even be an expected behaviour, but if so, 
it is preferable to have the captured value only as a single attribute, without 
duplication.

Please see the attachment for an example.

 


> ExtractText processor - RegEx captured twice
> 
>
> Key: NIFI-10869
> URL: https://issues.apache.org/jira/browse/NIFI-10869
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.18.0
>Reporter: Luigi De Giovanni
>Priority: Major
> Attachments: image (1).png, image (2).png, image.png
>
>
> Hi,
> There is an issue with the ExtractText processor, in the attempt of creating 
> FlowFile attributes from FlowFile contents.
> When creating a custom property with a regex value that does not contain 
> named groups, the value captured in the group is added as 2 different 
> attributes.
> E.G.
> ||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
> |param.alpha||param.alpha=alpha|param.alpha=alpha
> param.alpha.1=alpha|
> Reading the documentation, this might even be an expected behaviour, but if 
> so, it is preferable to have the captured value only as a single attribute, 
> without duplication.
> Please see the attachment for an example.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-10869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi De Giovanni updated NIFI-10869:
-
Description: 
Hi,

There is an issue with the ExtractText processor, in the attempt of creating 
FlowFile attributes from FlowFile contents.

When creating a custom property with a regex value that does not contain named 
groups, the value captured in the group is added as 2 different attributes.

E.G.
||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
|request.param.alpha||request.param.alpha=alpha|request.param.alpha=alpha
request.param.alpha.1=alpha|

Reading the documentation, this might even be an expected behaviour, but if so, 
it is preferable to have the captured value only as a single attribute, without 
duplication.

Please see the attachment for an example.

 

  was:
Hi,

There is an issue with the ExtractText processor, in the attempt of creating 
FlowFile attributes from FlowFile contents.

When creating a custom property with a regex value that does not contain named 
groups, the value captured in the group is added as 2 different attributes.

E.G.
||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
|request.param.alpha||request.param.alpha=alpha|request.param.alpha=alpha
request.param.alpha.1=alpha|


Reading the documentation, this might even be an expected behaviour, but if so, 
it is preferable to have the captured value only as a single attribute, without 
duplication.

Please see the attachment for a real example

 

 


> ExtractText processor - RegEx captured twice
> 
>
> Key: NIFI-10869
> URL: https://issues.apache.org/jira/browse/NIFI-10869
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.18.0
>Reporter: Luigi De Giovanni
>Priority: Minor
> Attachments: image (1).png, image (2).png, image.png
>
>
> Hi,
> There is an issue with the ExtractText processor, in the attempt of creating 
> FlowFile attributes from FlowFile contents.
> When creating a custom property with a regex value that does not contain 
> named groups, the value captured in the group is added as 2 different 
> attributes.
> E.G.
> ||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
> |request.param.alpha| alpha>|request.param.alpha=alpha|request.param.alpha=alpha
> request.param.alpha.1=alpha|
> Reading the documentation, this might even be an expected behaviour, but if 
> so, it is preferable to have the captured value only as a single attribute, 
> without duplication.
> Please see the attachment for an example.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-10869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi De Giovanni updated NIFI-10869:
-
Priority: Major  (was: Minor)

> ExtractText processor - RegEx captured twice
> 
>
> Key: NIFI-10869
> URL: https://issues.apache.org/jira/browse/NIFI-10869
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.18.0
>Reporter: Luigi De Giovanni
>Priority: Major
> Attachments: image (1).png, image (2).png, image.png
>
>
> Hi,
> There is an issue with the ExtractText processor, in the attempt of creating 
> FlowFile attributes from FlowFile contents.
> When creating a custom property with a regex value that does not contain 
> named groups, the value captured in the group is added as 2 different 
> attributes.
> E.G.
> ||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
> |request.param.alpha| alpha>|request.param.alpha=alpha|request.param.alpha=alpha
> request.param.alpha.1=alpha|
> Reading the documentation, this might even be an expected behaviour, but if 
> so, it is preferable to have the captured value only as a single attribute, 
> without duplication.
> Please see the attachment for an example.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)
Luigi De Giovanni created NIFI-10869:


 Summary: ExtractText processor - RegEx captured twice
 Key: NIFI-10869
 URL: https://issues.apache.org/jira/browse/NIFI-10869
 Project: Apache NiFi
  Issue Type: Bug
Affects Versions: 1.18.0
Reporter: Luigi De Giovanni
 Attachments: image (1).png, image (2).png, image.png

Hi,

There is an issue with the ExtractText processor, in the attempt of creating 
FlowFile attributes from FlowFile contents.

When creating a custom property with a regex value that does not contain named 
groups, the value captured in the group is added as 2 different attributes.

E.G.
||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
|request.param.alpha||request.param.alpha=alpha|request.param.alpha=alpha
request.param.alpha.1=alpha|


Reading the documentation, this might even be an expected behaviour, but if so, 
it is preferable to have the captured value only as a single attribute, without 
duplication.

Please see the attachment for a real example

 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (NIFI-10869) ExtractText processor - RegEx captured twice

2022-11-23 Thread Luigi De Giovanni (Jira)


 [ 
https://issues.apache.org/jira/browse/NIFI-10869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi De Giovanni updated NIFI-10869:
-
Priority: Minor  (was: Major)

> ExtractText processor - RegEx captured twice
> 
>
> Key: NIFI-10869
> URL: https://issues.apache.org/jira/browse/NIFI-10869
> Project: Apache NiFi
>  Issue Type: Bug
>Affects Versions: 1.18.0
>Reporter: Luigi De Giovanni
>Priority: Minor
> Attachments: image (1).png, image (2).png, image.png
>
>
> Hi,
> There is an issue with the ExtractText processor, in the attempt of creating 
> FlowFile attributes from FlowFile contents.
> When creating a custom property with a regex value that does not contain 
> named groups, the value captured in the group is added as 2 different 
> attributes.
> E.G.
> ||Property Name||Property Value||Expected FlowFile Attribute||Actual outcome||
> |request.param.alpha| alpha>|request.param.alpha=alpha|request.param.alpha=alpha
> request.param.alpha.1=alpha|
> Reading the documentation, this might even be an expected behaviour, but if 
> so, it is preferable to have the captured value only as a single attribute, 
> without duplication.
> Please see the attachment for a real example
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-3620) InvokeHttp should be able to send multipart forms

2022-11-23 Thread Luigi De Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-3620?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17637755#comment-17637755
 ] 

Luigi De Giovanni commented on NIFI-3620:
-

Hi,

it would be great if this could be fixed.
I have experienced the same issue recently and I managed to work around it.
I am pasting below some info I shared on the slack channel.

"I am dealing with an API which takes only POST requests and needs parameters 
passed in the body, as multipart/form-data .The InvokeHTTP processor doesn't 
work - with the API I am dealing with - if I define the request parameters as 
additional properties, like post:form:.
This is because the form-data is attached by the processor to the URL.
However, since the InvokeHTTP processor takes the contents of the input 
FlowFile as request body, I worked around that by creating the 
multipart/form-data text in a GenerateFlowFile or ReplaceText processor, with 
proper mime type and boundary.
Passing this to the InvokeHTTP without custom attributes makes that API react 
correctly."

The trick is basically in formatting the text of the body as an actual 
multipart/form mime type.
I reproduced it by checking the standard specification for multipart/form-data 
available publicly.

It would be great if the processor was able to do this formatting by itself.
Also, multipart/form-data shouldn't be attached to the URL, but remain in the 
request body only.

For any question, feel free

> InvokeHttp should be able to send multipart forms
> -
>
> Key: NIFI-3620
> URL: https://issues.apache.org/jira/browse/NIFI-3620
> Project: Apache NiFi
>  Issue Type: Improvement
>Reporter: Andre F de Miranda
>Assignee: Andre F de Miranda
>Priority: Major
>
> Currently InvokeHTTP is unable to send multipart forms, meaning it cannot be 
> used to a number of REST APIs endpoints.
> This has been identified by users 
> https://lists.apache.org/thread.html/32654510792f62ab855a785265c5ad5747cc0490da225fd2d44ac10a@%3Cdev.nifi.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] (NIFI-5901) Write JSON record in database

2022-11-18 Thread Luigi De Giovanni (Jira)


[ https://issues.apache.org/jira/browse/NIFI-5901 ]


Luigi De Giovanni deleted comment on NIFI-5901:
-

was (Author: JIRAUSER298268):
It would be great to know at least a workaround for this case, if this feature 
is not being implemented.
JOLT transformations don't seem to work, handling the JSON as string, as all 
the field names lose the double quotes.

> Write JSON record in database
> -
>
> Key: NIFI-5901
> URL: https://issues.apache.org/jira/browse/NIFI-5901
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.8.0
>Reporter: Flo Rance
>Assignee: Mike Thomsen
>Priority: Minor
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> It would be good to be able to store a whole json record in databases that 
> implement it (e.g. postgresql). This would require to define the field in the 
> shema as json/jsonb and then let PutDatabaseRecord inserts the json value in 
> the json/jsonb field.
> At the moment, it's possible to store a json/jsonb through Postgresql JDBC 
> using the Java sql type 'OTHER':
> Object data = "\{...}"; // the JSON document
> PreparedStatement.setObject(1, data, java.sql.Types.OTHER);



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (NIFI-5901) Write JSON record in database

2022-11-09 Thread Luigi De Giovanni (Jira)


[ 
https://issues.apache.org/jira/browse/NIFI-5901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17631131#comment-17631131
 ] 

Luigi De Giovanni commented on NIFI-5901:
-

It would be great to know at least a workaround for this case, if this feature 
is not being implemented.
JOLT transformations don't seem to work, handling the JSON as string, as all 
the field names lose the double quotes.

> Write JSON record in database
> -
>
> Key: NIFI-5901
> URL: https://issues.apache.org/jira/browse/NIFI-5901
> Project: Apache NiFi
>  Issue Type: Improvement
>Affects Versions: 1.8.0
>Reporter: Flo Rance
>Assignee: Mike Thomsen
>Priority: Minor
>  Time Spent: 3.5h
>  Remaining Estimate: 0h
>
> It would be good to be able to store a whole json record in databases that 
> implement it (e.g. postgresql). This would require to define the field in the 
> shema as json/jsonb and then let PutDatabaseRecord inserts the json value in 
> the json/jsonb field.
> At the moment, it's possible to store a json/jsonb through Postgresql JDBC 
> using the Java sql type 'OTHER':
> Object data = "\{...}"; // the JSON document
> PreparedStatement.setObject(1, data, java.sql.Types.OTHER);



--
This message was sent by Atlassian Jira
(v8.20.10#820010)