Re: [Gluster-devel] How does read-subvol-entry.t works?

2015-03-03 Thread Emmanuel Dreyfus
On Tue, Mar 03, 2015 at 07:47:15AM +0530, Ravishankar N wrote:
 If the afr xattrs on the dir is clean on all bricks, then the dir is chosen
 by afr_read_subvol_select_by_policy().

It seems there is very weird stuff going on there: it fails because 
in afr_inode_refresh_subvol_cbk (after a lookup), we have a valid 
reply from brick 0 with op_ret = 0.

But the brick 0 server process was killed. that makes no sense.


-- 
Emmanuel Dreyfus
m...@netbsd.org
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] GSoC-2015 organization application rejected

2015-03-03 Thread Kaushal M
Hi everyone.

I'm disappointed to announce that our application to be a mentoring
organization for GSoC-2015 has been rejected. Google has accepted 137
organizations [1] this year (down from 190 last year) and couldn't find
place to accommodate us.

The main reason behind the rejection was the presentation of our ideas list
[2]. Though we had a good number of ideas, Google expected them to be more
fleshed out and followed guidelines given at [3]. This some thing we must
do and doing this will help new contributors.

This was our communities first ever attempt, so we were bound to have some
missteps. We'll prepare better for the next time and try again.

In the meantime, anyone who was hoping to contribute via GSoC, but would
still like to contribute to the community anyway, take a look at our ideas
list [2]. If you'd like to work on something, get in touch with the listed
mentor or with development mailing list.

I want to thank the community members who volunteered to be mentors, and
who came up with the ideas. I want to thank Spot, for helping me get our
application submitted in a rush.

Thanks everyone. Let's try again next year.

- Kaushal

[1]: https://www.google-melange.com/gsoc/org/list/public/google/gsoc2015
[2]: http://www.gluster.org/community/documentation/index.php/Projects
[3]: http://en.flossmanuals.net/GSoCMentoring/making-your-ideas-page/
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] Spurious failure report for master branch - 2015-03-03

2015-03-03 Thread Justin Clift
Ran 20 x regression tests on our GlusterFS master branch code
as of a few hours ago, commit 95d5e60afb29aedc29909340e7564d54a6a247c2.

5 of them were successful (25%), 15 of them failed in various ways
(75%).

We need to get this down to about 5% or less (preferably 0%), as it's
killing our development iteration speed.  We're wasting huge amounts
of time working around this. :(


Spurious failures
*

  * 5 x tests/bugs/distribute/bug-1117851.t 
  (Wstat: 0 Tests: 24 Failed: 1)
Failed test:  15

This one is causing a 25% failure rate all by itself. :(

This needs fixing soon. :)


  * 3 x tests/bugs/geo-replication/bug-877293.t 
  (Wstat: 0 Tests: 15 Failed: 1)
Failed test:  11

  * 2 x tests/basic/afr/entry-self-heal.t   
  (Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128

  * 1 x tests/basic/ec/ec-12-4.t
  (Wstat: 0 Tests: 541 Failed: 2)
Failed tests:  409, 441

  * 1 x tests/basic/fops-sanity.t   
  (Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10

  * 1 x tests/basic/uss.t   
  (Wstat: 0 Tests: 160 Failed: 1)
Failed test:  26

  * 1 x tests/performance/open-behind.t 
  (Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17

  * 1 x tests/bugs/distribute/bug-884455.t  
  (Wstat: 0 Tests: 22 Failed: 1)
Failed test:  11

  * 1 x tests/bugs/fuse/bug-1126048.t   
  (Wstat: 0 Tests: 12 Failed: 1)
Failed test:  10

  * 1 x tests/bugs/quota/bug-1038598.t  
  (Wstat: 0 Tests: 28 Failed: 1)
Failed test:  28


2 x Coredumps
*

  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk5/

IP - 104.130.74.142

This coredump run also failed on:

  * tests/basic/fops-sanity.t   
  (Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10

  * tests/bugs/glusterfs-server/bug-861542.t
  (Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10

  * tests/performance/open-behind.t 
  (Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17

  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk8/

IP - 104.130.74.143

This coredump run also failed on:

  * tests/basic/afr/entry-self-heal.t   
  (Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128

  * tests/bugs/glusterfs-server/bug-861542.t
  (Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10

Both VMs are also online, in case they're useful to log into
for investigation (root / the jenkins slave pw).

If they're not, please let me know so I can blow them away. :)


1 x hung host
*

Hung on tests/bugs/posix/bug-1113960.t

root  12497  1290  0 Mar03 ?  S  0:00  \_ /bin/bash /opt/qa/regression.sh
root  12504 12497  0 Mar03 ?  S  0:00  \_ /bin/bash ./run-tests.sh
root  12519 12504  0 Mar03 ?  S  0:03  \_ /usr/bin/perl /usr/bin/prove 
-rf --timer ./tests
root  22018 12519  0 00:17 ?  S  0:00  \_ /bin/bash 
./tests/bugs/posix/bug-1113960.t
root  30002 22018  0 01:57 ?  S  0:00  \_ mv 
/mnt/glusterfs/0/longernamedir1/longernamedir2/longernamedir3/

This VM (23.253.53.111) is still online + untouched (still hung),
if someone wants to log in to investigate.  (root / the jenkins
slave pw)

Hope that's helpful. :)

Regards and best wishes,

Justin Clift

--
GlusterFS - http://www.gluster.org

An open source, distributed file system scaling to several
petabytes, and handling thousands of clients.

My personal twitter: twitter.com/realjustinclift

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Spurious failure report for master branch - 2015-03-03

2015-03-03 Thread Justin Clift
Thanks. :)

If you need a VM setup in Rackspace for you to investigate on, it's easy
to do.  Let me know if so. :)

+ Justin


On 4 Mar 2015, at 04:37, Nithya Balachandran nbala...@redhat.com wrote:
 I'll take a look at tests/bugs/distribute/bug-1117851.t
 
 Regards,
 Nithya
 
 - Original Message -
 From: Justin Clift jus...@gluster.org
 To: Gluster Devel gluster-devel@gluster.org
 Sent: Wednesday, 4 March, 2015 9:57:00 AM
 Subject: [Gluster-devel] Spurious failure report for master branch -  
 2015-03-03
 
 Ran 20 x regression tests on our GlusterFS master branch code
 as of a few hours ago, commit 95d5e60afb29aedc29909340e7564d54a6a247c2.
 
 5 of them were successful (25%), 15 of them failed in various ways
 (75%).
 
 We need to get this down to about 5% or less (preferably 0%), as it's
 killing our development iteration speed.  We're wasting huge amounts
 of time working around this. :(
 
 
 Spurious failures
 *
 
  * 5 x tests/bugs/distribute/bug-1117851.t
(Wstat: 0 Tests: 24 Failed: 1)
Failed test:  15
 
This one is causing a 25% failure rate all by itself. :(
 
This needs fixing soon. :)
 
 
  * 3 x tests/bugs/geo-replication/bug-877293.t
(Wstat: 0 Tests: 15 Failed: 1)
Failed test:  11
 
  * 2 x tests/basic/afr/entry-self-heal.t  
(Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128
 
  * 1 x tests/basic/ec/ec-12-4.t   
(Wstat: 0 Tests: 541 Failed: 2)
Failed tests:  409, 441
 
  * 1 x tests/basic/fops-sanity.t  
(Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10
 
  * 1 x tests/basic/uss.t  
(Wstat: 0 Tests: 160 Failed: 1)
Failed test:  26
 
  * 1 x tests/performance/open-behind.t
(Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17
 
  * 1 x tests/bugs/distribute/bug-884455.t 
(Wstat: 0 Tests: 22 Failed: 1)
Failed test:  11
 
  * 1 x tests/bugs/fuse/bug-1126048.t  
(Wstat: 0 Tests: 12 Failed: 1)
Failed test:  10
 
  * 1 x tests/bugs/quota/bug-1038598.t 
(Wstat: 0 Tests: 28 Failed: 1)
Failed test:  28
 
 
 2 x Coredumps
 *
 
  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk5/
 
IP - 104.130.74.142
 
This coredump run also failed on:
 
  * tests/basic/fops-sanity.t  
(Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10
 
  * tests/bugs/glusterfs-server/bug-861542.t   
(Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10
 
  * tests/performance/open-behind.t
(Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17
 
  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk8/
 
IP - 104.130.74.143
 
This coredump run also failed on:
 
  * tests/basic/afr/entry-self-heal.t  
(Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128
 
  * tests/bugs/glusterfs-server/bug-861542.t   
(Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10
 
 Both VMs are also online, in case they're useful to log into
 for investigation (root / the jenkins slave pw).
 
 If they're not, please let me know so I can blow them away. :)
 
 
 1 x hung host
 *
 
 Hung on tests/bugs/posix/bug-1113960.t
 
 root  12497  1290  0 Mar03 ?  S  0:00  \_ /bin/bash /opt/qa/regression.sh
 root  12504 12497  0 Mar03 ?  S  0:00  \_ /bin/bash ./run-tests.sh
 root  12519 12504  0 Mar03 ?  S  0:03  \_ /usr/bin/perl 
 /usr/bin/prove -rf --timer ./tests
 root  22018 12519  0 00:17 ?  S  0:00  \_ /bin/bash 
 ./tests/bugs/posix/bug-1113960.t
 root  30002 22018  0 01:57 ?  S  0:00  \_ mv 
 /mnt/glusterfs/0/longernamedir1/longernamedir2/longernamedir3/
 
 This VM (23.253.53.111) is still online + untouched (still hung),
 if someone wants to log in to investigate.  (root / the jenkins
 slave pw)
 
 Hope that's helpful. :)
 
 Regards and best wishes,
 
 Justin Clift
 
 --
 GlusterFS - http://www.gluster.org
 
 An open source, distributed file system scaling to several
 petabytes, and handling thousands of clients.
 
 My personal twitter: twitter.com/realjustinclift
 
 ___
 Gluster-devel mailing list
 Gluster-devel@gluster.org
 http://www.gluster.org/mailman/listinfo/gluster-devel

--
GlusterFS - http://www.gluster.org

An open source, distributed file system scaling to several

Re: [Gluster-devel] Spurious failure report for master branch - 2015-03-03

2015-03-03 Thread Poornima Gurusiddaiah
Few more test cases causing spurious failures:

./tests/basic/ec/ec-5-1.t 
Failed test:  69

./tests/basic/ec/ec-5-2.t
Failed test:  69

./tests/bugs/disperse/bug-1187474.t
 Failed tests:  11-12

./tests/basic/ec/nfs.t
 Failed test:  9

The above failures were seen for the patches which were ineffective,
i.e. the code that was modified was never executed as it had no callers.

Regards,
Poornima

- Original Message -
 From: Justin Clift jus...@gluster.org
 To: Gluster Devel gluster-devel@gluster.org
 Sent: Wednesday, March 4, 2015 9:57:00 AM
 Subject: [Gluster-devel] Spurious failure report for master branch -  
 2015-03-03
 
 Ran 20 x regression tests on our GlusterFS master branch code
 as of a few hours ago, commit 95d5e60afb29aedc29909340e7564d54a6a247c2.
 
 5 of them were successful (25%), 15 of them failed in various ways
 (75%).
 
 We need to get this down to about 5% or less (preferably 0%), as it's
 killing our development iteration speed.  We're wasting huge amounts
 of time working around this. :(
 
 
 Spurious failures
 *
 
   * 5 x tests/bugs/distribute/bug-1117851.t
   (Wstat: 0 Tests: 24 Failed:
   1)
 Failed test:  15
 
 This one is causing a 25% failure rate all by itself. :(
 
 This needs fixing soon. :)
 
 
   * 3 x tests/bugs/geo-replication/bug-877293.t
   (Wstat: 0 Tests: 15 Failed: 1)
 Failed test:  11
 
   * 2 x tests/basic/afr/entry-self-heal.t
   (Wstat: 0 Tests: 180
   Failed: 2)
 Failed tests:  127-128
 
   * 1 x tests/basic/ec/ec-12-4.t
   (Wstat: 0 Tests:
   541 Failed: 2)
 Failed tests:  409, 441
 
   * 1 x tests/basic/fops-sanity.t
   (Wstat: 0 Tests:
   11 Failed: 1)
 Failed test:  10
 
   * 1 x tests/basic/uss.t
   (Wstat: 0
   Tests: 160 Failed: 1)
 Failed test:  26
 
   * 1 x tests/performance/open-behind.t
   (Wstat: 0 Tests: 17
   Failed: 1)
 Failed test:  17
 
   * 1 x tests/bugs/distribute/bug-884455.t
   (Wstat: 0 Tests: 22 Failed:
   1)
 Failed test:  11
 
   * 1 x tests/bugs/fuse/bug-1126048.t
   (Wstat: 0 Tests: 12
   Failed: 1)
 Failed test:  10
 
   * 1 x tests/bugs/quota/bug-1038598.t
   (Wstat: 0 Tests: 28
   Failed: 1)
 Failed test:  28
 
 
 2 x Coredumps
 *
 
   * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk5/
 
 IP - 104.130.74.142
 
 This coredump run also failed on:
 
   * tests/basic/fops-sanity.t
   (Wstat: 0
   Tests: 11 Failed: 1)
 Failed test:  10
 
   * tests/bugs/glusterfs-server/bug-861542.t
   (Wstat: 0 Tests: 13 Failed:
   1)
 Failed test:  10
 
   * tests/performance/open-behind.t
   (Wstat: 0 Tests: 17
   Failed: 1)
 Failed test:  17
 
   * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk8/
 
 IP - 104.130.74.143
 
 This coredump run also failed on:
 
   * tests/basic/afr/entry-self-heal.t
   (Wstat: 0 Tests: 180
   Failed: 2)
 Failed tests:  127-128
 
   * tests/bugs/glusterfs-server/bug-861542.t
   (Wstat: 0 Tests: 13 Failed:
   1)
 Failed test:  10
 
 Both VMs are also online, in case they're useful to log into
 for investigation (root / the jenkins slave pw).
 
 If they're not, please let me know so I can blow them away. :)
 
 
 1 x hung host
 *
 
 Hung on tests/bugs/posix/bug-1113960.t
 
 root  12497  1290  0 Mar03 ?  S  0:00  \_ /bin/bash /opt/qa/regression.sh
 root  12504 12497  0 Mar03 ?  S  0:00  \_ /bin/bash ./run-tests.sh
 root  12519 12504  0 Mar03 ?  S  0:03  \_ /usr/bin/perl
 /usr/bin/prove -rf --timer ./tests
 root  22018 12519  0 00:17 ?  S  0:00  \_ /bin/bash
 ./tests/bugs/posix/bug-1113960.t
 root  30002 22018  0 01:57 ?  S  0:00  \_ mv
 /mnt/glusterfs/0/longernamedir1/longernamedir2/longernamedir3/
 
 This VM (23.253.53.111) is still online + untouched (still hung),
 if someone wants to log in to investigate.  (root / the jenkins
 slave pw)
 
 Hope that's helpful. :)
 
 Regards and best wishes,
 
 Justin Clift
 
 --
 GlusterFS - http://www.gluster.org
 
 An open source, distributed file system scaling to several
 petabytes, and handling thousands of clients.
 
 My personal twitter: twitter.com/realjustinclift
 
 ___
 Gluster-devel mailing list
 Gluster-devel@gluster.org
 http://www.gluster.org/mailman/listinfo/gluster-devel
 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] How does read-subvol-entry.t works?

2015-03-03 Thread Ravishankar N


On 03/04/2015 10:29 AM, Emmanuel Dreyfus wrote:

Emmanuel Dreyfus m...@netbsd.org wrote:


It seems there is very weird stuff going on there: it fails because
in afr_inode_refresh_subvol_cbk (after a lookup), we have a valid
reply from brick 0 with op_ret = 0.

But the brick 0 server process was killed. that makes no sense.

Looking at a kernel trace I can now tell that the brick0 server process
indeed gets a SIGKILL, but then glusterd spawn a new process for brick0
that answers the requests.

glusterd log confirms that: first it starts the two bricks;
[glusterd-pmap.c:227:pmap_registry_bind] 0-pmap: adding brick 
/d/backends/brick0 on port 49152
[glusterd-pmap.c:227:pmap_registry_bind] 0-pmap: adding brick 
/d/backends/brick1 on port 49153

- Killing brick0
[glusterd-handler.c:4388:__glusterd_brick_rpc_notify] 0-management: Brick 
nbslave73.cloud.gluster.org:/d/backends/brick0 has disconnected from glusterd.

- And here it restarts!

[glusterd-pmap.c:227:pmap_registry_bind] 0-pmap: adding brick 
/d/backends/brick0 on port 49152

- test terminate and kill all bricks:

[glusterd-pmap.c:271:pmap_registry_remove] 0-pmap: removing brick 
/d/backends/brick0 on port 49152
[glusterd-pmap.c:271:pmap_registry_remove] 0-pmap: removing brick 
/d/backends/brick1 on port 49153

Hence it ould be a glusterd bug? Why would it restart a brick on its own?

Not sure, CC'ing Atin who might be able to shed some light on the 
glusterd logs. If the brick gets restarted as you say, the brick log 
will also contain something like I [glusterfsd.c:1959:main] 
0-/usr/local/sbin/glusterfsd: Started running 
/usr/local/sbin/glusterfsd  and the graph information etc. Does it? And 
does volume status show the brick as online again?


-Ravi
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Spurious failure report for master branch - 2015-03-03

2015-03-03 Thread Nithya Balachandran
I'll take a look at tests/bugs/distribute/bug-1117851.t

Regards,
Nithya

- Original Message -
From: Justin Clift jus...@gluster.org
To: Gluster Devel gluster-devel@gluster.org
Sent: Wednesday, 4 March, 2015 9:57:00 AM
Subject: [Gluster-devel] Spurious failure report for master branch -
2015-03-03

Ran 20 x regression tests on our GlusterFS master branch code
as of a few hours ago, commit 95d5e60afb29aedc29909340e7564d54a6a247c2.

5 of them were successful (25%), 15 of them failed in various ways
(75%).

We need to get this down to about 5% or less (preferably 0%), as it's
killing our development iteration speed.  We're wasting huge amounts
of time working around this. :(


Spurious failures
*

  * 5 x tests/bugs/distribute/bug-1117851.t 
  (Wstat: 0 Tests: 24 Failed: 1)
Failed test:  15

This one is causing a 25% failure rate all by itself. :(

This needs fixing soon. :)


  * 3 x tests/bugs/geo-replication/bug-877293.t 
  (Wstat: 0 Tests: 15 Failed: 1)
Failed test:  11

  * 2 x tests/basic/afr/entry-self-heal.t   
  (Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128

  * 1 x tests/basic/ec/ec-12-4.t
  (Wstat: 0 Tests: 541 Failed: 2)
Failed tests:  409, 441

  * 1 x tests/basic/fops-sanity.t   
  (Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10

  * 1 x tests/basic/uss.t   
  (Wstat: 0 Tests: 160 Failed: 1)
Failed test:  26

  * 1 x tests/performance/open-behind.t 
  (Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17

  * 1 x tests/bugs/distribute/bug-884455.t  
  (Wstat: 0 Tests: 22 Failed: 1)
Failed test:  11

  * 1 x tests/bugs/fuse/bug-1126048.t   
  (Wstat: 0 Tests: 12 Failed: 1)
Failed test:  10

  * 1 x tests/bugs/quota/bug-1038598.t  
  (Wstat: 0 Tests: 28 Failed: 1)
Failed test:  28


2 x Coredumps
*

  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk5/

IP - 104.130.74.142

This coredump run also failed on:

  * tests/basic/fops-sanity.t   
  (Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10

  * tests/bugs/glusterfs-server/bug-861542.t
  (Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10

  * tests/performance/open-behind.t 
  (Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17

  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk8/

IP - 104.130.74.143

This coredump run also failed on:

  * tests/basic/afr/entry-self-heal.t   
  (Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128

  * tests/bugs/glusterfs-server/bug-861542.t
  (Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10

Both VMs are also online, in case they're useful to log into
for investigation (root / the jenkins slave pw).

If they're not, please let me know so I can blow them away. :)


1 x hung host
*

Hung on tests/bugs/posix/bug-1113960.t

root  12497  1290  0 Mar03 ?  S  0:00  \_ /bin/bash /opt/qa/regression.sh
root  12504 12497  0 Mar03 ?  S  0:00  \_ /bin/bash ./run-tests.sh
root  12519 12504  0 Mar03 ?  S  0:03  \_ /usr/bin/perl /usr/bin/prove 
-rf --timer ./tests
root  22018 12519  0 00:17 ?  S  0:00  \_ /bin/bash 
./tests/bugs/posix/bug-1113960.t
root  30002 22018  0 01:57 ?  S  0:00  \_ mv 
/mnt/glusterfs/0/longernamedir1/longernamedir2/longernamedir3/

This VM (23.253.53.111) is still online + untouched (still hung),
if someone wants to log in to investigate.  (root / the jenkins
slave pw)

Hope that's helpful. :)

Regards and best wishes,

Justin Clift

--
GlusterFS - http://www.gluster.org

An open source, distributed file system scaling to several
petabytes, and handling thousands of clients.

My personal twitter: twitter.com/realjustinclift

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Spurious failure report for master branch - 2015-03-03

2015-03-03 Thread Nithya Balachandran
Looks like I will need a VM - the test passes consistently on my local setup.

Can you please send me the details?

Regards,
Nithya

- Original Message -
From: Justin Clift jus...@gluster.org
To: Nithya Balachandran nbala...@redhat.com
Cc: Gluster Devel gluster-devel@gluster.org
Sent: Wednesday, 4 March, 2015 10:12:17 AM
Subject: Re: [Gluster-devel] Spurious failure report for master branch - 
2015-03-03

Thanks. :)

If you need a VM setup in Rackspace for you to investigate on, it's easy
to do.  Let me know if so. :)

+ Justin


On 4 Mar 2015, at 04:37, Nithya Balachandran nbala...@redhat.com wrote:
 I'll take a look at tests/bugs/distribute/bug-1117851.t
 
 Regards,
 Nithya
 
 - Original Message -
 From: Justin Clift jus...@gluster.org
 To: Gluster Devel gluster-devel@gluster.org
 Sent: Wednesday, 4 March, 2015 9:57:00 AM
 Subject: [Gluster-devel] Spurious failure report for master branch -  
 2015-03-03
 
 Ran 20 x regression tests on our GlusterFS master branch code
 as of a few hours ago, commit 95d5e60afb29aedc29909340e7564d54a6a247c2.
 
 5 of them were successful (25%), 15 of them failed in various ways
 (75%).
 
 We need to get this down to about 5% or less (preferably 0%), as it's
 killing our development iteration speed.  We're wasting huge amounts
 of time working around this. :(
 
 
 Spurious failures
 *
 
  * 5 x tests/bugs/distribute/bug-1117851.t
(Wstat: 0 Tests: 24 Failed: 1)
Failed test:  15
 
This one is causing a 25% failure rate all by itself. :(
 
This needs fixing soon. :)
 
 
  * 3 x tests/bugs/geo-replication/bug-877293.t
(Wstat: 0 Tests: 15 Failed: 1)
Failed test:  11
 
  * 2 x tests/basic/afr/entry-self-heal.t  
(Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128
 
  * 1 x tests/basic/ec/ec-12-4.t   
(Wstat: 0 Tests: 541 Failed: 2)
Failed tests:  409, 441
 
  * 1 x tests/basic/fops-sanity.t  
(Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10
 
  * 1 x tests/basic/uss.t  
(Wstat: 0 Tests: 160 Failed: 1)
Failed test:  26
 
  * 1 x tests/performance/open-behind.t
(Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17
 
  * 1 x tests/bugs/distribute/bug-884455.t 
(Wstat: 0 Tests: 22 Failed: 1)
Failed test:  11
 
  * 1 x tests/bugs/fuse/bug-1126048.t  
(Wstat: 0 Tests: 12 Failed: 1)
Failed test:  10
 
  * 1 x tests/bugs/quota/bug-1038598.t 
(Wstat: 0 Tests: 28 Failed: 1)
Failed test:  28
 
 
 2 x Coredumps
 *
 
  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk5/
 
IP - 104.130.74.142
 
This coredump run also failed on:
 
  * tests/basic/fops-sanity.t  
(Wstat: 0 Tests: 11 Failed: 1)
Failed test:  10
 
  * tests/bugs/glusterfs-server/bug-861542.t   
(Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10
 
  * tests/performance/open-behind.t
(Wstat: 0 Tests: 17 Failed: 1)
Failed test:  17
 
  * http://mirror.salasaga.org/gluster/master/2015-03-03/bulk8/
 
IP - 104.130.74.143
 
This coredump run also failed on:
 
  * tests/basic/afr/entry-self-heal.t  
(Wstat: 0 Tests: 180 Failed: 2)
Failed tests:  127-128
 
  * tests/bugs/glusterfs-server/bug-861542.t   
(Wstat: 0 Tests: 13 Failed: 1)
Failed test:  10
 
 Both VMs are also online, in case they're useful to log into
 for investigation (root / the jenkins slave pw).
 
 If they're not, please let me know so I can blow them away. :)
 
 
 1 x hung host
 *
 
 Hung on tests/bugs/posix/bug-1113960.t
 
 root  12497  1290  0 Mar03 ?  S  0:00  \_ /bin/bash /opt/qa/regression.sh
 root  12504 12497  0 Mar03 ?  S  0:00  \_ /bin/bash ./run-tests.sh
 root  12519 12504  0 Mar03 ?  S  0:03  \_ /usr/bin/perl 
 /usr/bin/prove -rf --timer ./tests
 root  22018 12519  0 00:17 ?  S  0:00  \_ /bin/bash 
 ./tests/bugs/posix/bug-1113960.t
 root  30002 22018  0 01:57 ?  S  0:00  \_ mv 
 /mnt/glusterfs/0/longernamedir1/longernamedir2/longernamedir3/
 
 This VM (23.253.53.111) is still online + untouched (still hung),
 if someone wants to log in to investigate.  (root / the jenkins
 slave pw)
 
 Hope that's helpful. :)
 
 Regards and best wishes,
 
 Justin Clift
 
 --
 GlusterFS - http://www.gluster.org
 
 An