We already exclude CBC, because of the POODLE attack, and that leaves us
with 32 ciphers. Excluding DH as well leaves us with only four.
AES256-GCM-SHA384
AES256-SHA256
AES128-GCM-SHA256
AES128-SHA256
Why are ECDH ciphers missing? That list has no cipher featuring PFS,
I will encourage to do it before this patch gets into the codebase.
The duplicate peerinfo object issue is different from the problems
in the current generation number scheme. I don't see why this patch
needs to wait. We may need to keep volume-snapshot-clone.t disabled
until the time the
hi,
git fetch on local repo fails with the following error. I asked on
#gluster-dev, some of the people online now face the same error.
pk1@localhost - ~/workspace/gerrit-repo (cooperative-locking-3.7)
08:54:14 :( ⚡ git fetch
ssh_exchange_identification: Connection closed by remote host
Looks like gerrit is down. review.gluster.org is not accessible.
On 05/27/2015 08:58 AM, Pranith Kumar Karampuri wrote:
hi,
git fetch on local repo fails with the following error. I asked
on #gluster-dev, some of the people online now face the same error.
pk1@localhost -
On 26/05/15 20:44, Niels de Vos wrote:
On Tue, May 26, 2015 at 10:17:05AM -0400, Jiffin Thottan wrote:
- Original Message -
From: Vijay Bellur vbel...@redhat.com
To: Krishnan Parthasarathi kpart...@redhat.com, Shyam
srang...@redhat.com
Cc: Gluster Devel gluster-devel@gluster.org
During graph change we migrate all the fds opened on old subvol. In this case
we are trying to migrate fds opened on virtual .meta subtree. This has
resulted in crash as these inodes and fds have to be handled differently than
the ones in real volumes.
- Original Message -
From:
On 05/27/2015 08:55 AM, Krishnan Parthasarathi wrote:
I will encourage to do it before this patch gets into the codebase.
The duplicate peerinfo object issue is different from the problems
in the current generation number scheme. I don't see why this patch
needs to wait. We may need to
Is there a way to get coredump?
- Raghavendra Gowdappa rgowd...@redhat.com wrote:
Will take a look.
- Original Message -
From: Vijay Bellur vbel...@redhat.com
To: Gluster Devel gluster-devel@gluster.org
Sent: Monday, May 25, 2015 11:08:46 PM
Subject: [Gluster-devel] Crash
On 05/26/2015 12:02 PM, Raghavendra Gowdappa wrote:
Is there a way to get coredump?
Running the same test in a loop seems to trigger the crash (1 out of 40
runs approximately).
I will upload the core to a bz but I figure the above will be faster.
Thanks,
Vijay
- Raghavendra Gowdappa
- Original Message -
Pranith/Xavi,
./tests/basic/ec/ec.t is part of
https://public.pad.fsfe.org/p/gluster-spurious-failures
but is not part of is_bad_test() in run-tests.sh. This has repeatedly
failed an other independent spurious regression failure fix,
Hi
the logjam attack comes in two part. One is about downgrading
cipher choice because a TLS setup allows EWPORT ciphers. glusterfs
can thwart this by setting ssl.cipher-list to something hardened enough
(ECDH:DH:!TLSv1:!aNULL!eNULL seems nice);
Second part is about using pre-computed DH
I happen to see tests/bugs/nfs/bug-904065.t failing for one of the test
run [1]
I believe this failure has nothing to do with [2]
[1]
http://build.gluster.org/job/rackspace-regression-2GB-triggered/9527/consoleFull
[2] http://review.gluster.org/#/c/10820/
--
~Atin
I have $subject @ http://review.gluster.org/#/c/10917/ because we have
only one known failure for this test @
http://build.gluster.org/job/rackspace-regression-2GB-triggered/8735/consoleFull
and the tests that failed are `gluster volume stop` and `gluster volume
delete`. Not able to repro the
Hi all,
This meeting is scheduled for anyone that is interested in learning more
about, or assisting with the Bug Triage.
Meeting details:
- location: #gluster-meeting on Freenode IRC
( https://webchat.freenode.net/?channels=gluster-meeting )
- date: every Tuesday
- time: 12:00 UTC
Pranith,
I sent a patch to fix the spurious failure in entry-self-heal.t @
http://review.gluster.org/#/c/10916/ .
Requesting a review on the same.
-Krutika
___
Gluster-devel mailing list
Gluster-devel@gluster.org
All,
The following are the regression test failures that are being looked
at by Atin, Avra, Kaushal and self.
1) ./tests/bugs/glusterd/bug-974007.t to be fixed by
http://review.gluster.org/10872
2) ./tests/basic/volume-snapshot-clone. to be fixed (partially) by
Here is the status on quota test-case spurious failure:
There were 3 issues
1) Quota exceeding the limit because of parallel writes - Merged
Upstream, patch submitted to release-3.7 #10910
./tests/bugs/quota/bug-1038598.t
./tests/bugs/distribute/bug-1161156.t
2) Quoting
Status on crash in dht_fsync reported by Vijay Bellur:
I am not able to reproduce the issue. However I am consistently hitting a
failure in the last test of ./tests/performance/open-behind.t (test 18). The
test reads as:
gluster volume top $V0 open | grep -w $F0 /dev/null 21
TEST [ $? -eq 0
Hi gluster team,
Proposal:
Using Clang static analyzer tool for gluster project
About Clang:
From a very high level view, Clang has two features
1. Clang as a compiler
2. Clang as a code analyzer
The Idea hear is to use second point i.e Clang as code analyzer and still gcc
will be our
Hi,
We already are running nightly builds using both clang compile and clang
analyze.
see http://download.gluster.org/pub/gluster/glusterfs/static-analysis/
On 05/26/2015 08:00 AM, Prasanna Kalever wrote:
Hi gluster team,
Proposal:
Using Clang static analyzer tool for gluster project
These are my findings regarding the issue KP mentioned with
volume-snapshot-clone.t.
From what I gathered, the double peerinfo objects issue was observed
only via logs. Additional logging was added to
glusterd_peer_rpc_notify to log the objects address and
peerinfo-hostname . I did the same and
Hi,
What you are refering with the glusterd being restarted might be true,
but I don't remember if the two different peer addresses were before and
after restarting glusterd, or after restarting glusterd and before/after
peer status. We might need to re-do this exercise on a rackspace vm to
On 05/26/2015 04:05 PM, Atin Mukherjee wrote:
Hi all,
This meeting is scheduled for anyone that is interested in learning more
about, or assisting with the Bug Triage.
Meeting details:
- location: #gluster-meeting on Freenode IRC
(
Hi Louis,
AFAIK, we never tested nfs-ganesha+glusterfs using windows client. May
be it would be good if you can collect and provide packet trace/cores or
logs (with nfs-ganesha atleast in NIV_DEBUG level) on the server side
while you run these tests to debug futher.
Thanks,
Soumya
On
the logjam attack comes in two part. One is about downgrading
cipher choice because a TLS setup allows EWPORT ciphers. glusterfs
can thwart this by setting ssl.cipher-list to something hardened enough
(ECDH:DH:!TLSv1:!aNULL!eNULL seems nice);
Our default is already based on HIGH, which
Sorry for the wrong incomplete message sent by mistake earlier.
Hi Sussant,
Extremely sorry for the belated reply. Thanks for your input. We will try
having rebalance then create a set of small files with a random pattern
generation and check where it falls in by DM_TYPE DHT.
I have one
Hi Sussant,
Extremely sorry for the belated reply. Thanks for your input. We will try
having rebalance then create a set of small files with a random pattern
generation and check where it falls in by DM_TYPE DHT.
I have one short query:
We were also thinking to use (in case we need )
Hi Everyone,
I am trying to look in to the aspect of optimizing GlusterFS with respect
to the underlying file system (mainly EXT4). Could you please let me know
whether any such cross-layer (between GlusterFS and underlying FS)
optimizations have been already done in GlusterFS?
Regards
Toms
Hi All,
Could you please guide us to solve the following DHT and brick layout problem
we are dealing with ? Questions are marked bold.
Problem statement :
1. We have a requirement to achieve maximum write and read performance and
we have to meet some committed performance metrics.
- Original Message -
From: Vijay Bellur vbel...@redhat.com
To: Krishnan Parthasarathi kpart...@redhat.com, Shyam
srang...@redhat.com
Cc: Gluster Devel gluster-devel@gluster.org
Sent: Friday, 22 May, 2015 9:49:15 AM
Subject: Re: [Gluster-devel] Spurious regression: Checking on 3 test
On 26 May 2015 18:35, Avra Sengupta aseng...@redhat.com wrote:
Hi,
What you are refering with the glusterd being restarted might be true,
but I don't remember if the two different peer addresses were before and
after restarting glusterd, or after restarting glusterd and before/after
peer
On Tue, May 26, 2015 at 10:17:05AM -0400, Jiffin Thottan wrote:
- Original Message -
From: Vijay Bellur vbel...@redhat.com
To: Krishnan Parthasarathi kpart...@redhat.com, Shyam
srang...@redhat.com
Cc: Gluster Devel gluster-devel@gluster.org
Sent: Friday, 22 May, 2015 9:49:15 AM
Hi gluster team,
Proposal:
Using Clang static analyzer tool for gluster project
About Clang:
From a very high level view, Clang has two features
1. Clang as a compiler
2. Clang as a code analyzer
The Idea hear is to use second point i.e Clang as code analyzer and still gcc
will be our
hi,
tests/bugs/fuse/bug-924726.t failed in
http://build.gluster.org/job/rackspace-regression-2GB-triggered/9553/consoleFull
Could you take a look.
Pranith
___
Gluster-devel mailing list
Gluster-devel@gluster.org
Prasanna, have you compared the results to the ones we see via coverity?
-Ira
- Original Message -
Hi gluster team,
Proposal:
Using Clang static analyzer tool for gluster project
About Clang:
From a very high level view, Clang has two features
1. Clang as a compiler
On 26 May 2015 17:30, Prasanna Kalever pkale...@redhat.com wrote:
Hi gluster team,
Proposal:
Using Clang static analyzer tool for gluster project
About Clang:
From a very high level view, Clang has two features
1. Clang as a compiler
2. Clang as a code analyzer
The Idea hear is to
36 matches
Mail list logo