Atin,

I am not sure about the docker containers, but both the failures you mentioned are in slave29, which as Talur explained is missing the appropriate backend filesystem. Owing to this, op-errno.t is just the tip of the iceberg, and every other test that uses lvm will fail in this particular slave will fail too.

Talur,
Thanks for looking into it. It is indeed strange this. I checked the dmesg and the /var/log/messages in this slave and I couldn't find any relevant log.

On 07/12/2016 05:29 PM, Raghavendra Talur wrote:
I checked the machine.

Here is the df -hT output
[jenkins@slave29 ~]$ cat /etc/fstab
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
/dev/xvda1 / ext3 defaults,noatime,barrier=0 1 1
tmpfs                   /dev/shm                tmpfs defaults        0 0
devpts /dev/pts devpts gid=5,mode=620 0 0
sysfs                   /sys                    sysfs defaults        0 0
proc                    /proc                   proc  defaults        0 0
#/dev/xvdc1             none                    swap  sw              0 0


We don't see a xfs device mounted at /d and / is of type ext3 which does not support fallocate. The uptime of the machine is 73 days though. I don't know how the /d xfs partition vanished.

On Tue, Jul 12, 2016 at 4:54 PM, Atin Mukherjee <amukh...@redhat.com <mailto:amukh...@redhat.com>> wrote:

    
https://build.gluster.org/job/rackspace-regression-2GB-triggered/22156/consoleFull
    - another failure

    On Tue, Jul 12, 2016 at 4:42 PM, Atin Mukherjee
    <amukh...@redhat.com <mailto:amukh...@redhat.com>> wrote:



        On Tue, Jul 12, 2016 at 4:36 PM, Avra Sengupta
        <aseng...@redhat.com <mailto:aseng...@redhat.com>> wrote:

            Hi Atin,

            Please check the testcase result in the console. It
            clearly states the reason of the failure. A quick search
            of 30815, as shown in the testcase shows that the error
            that is generated is a thinp issue, and we can see
            fallocate failing and lvm not properly being setup in the
            environment.


        While this is valid for my docker containers, I am just
        wondering why did this happen in jenkins slave?


            Regards,
            Avra

            P.S Here are the logs from the console stating so.

            *02:50:34* [09:50:34] Running tests in file 
./tests/basic/op_errnos.t
            *02:50:41* fallocate: /d/backends/patchy_snap_vhd: fallocate 
failed: Operation not supported
            *02:50:41* losetup: /d/backends/patchy_snap_vhd: warning: file 
smaller than 512 bytes, the loop device maybe be useless or invisible for 
system tools.
            *02:50:41*    Device /d/backends/patchy_snap_loop not found (or 
ignored by filtering).
            *02:50:41*    Device /d/backends/patchy_snap_loop not found (or 
ignored by filtering).
            *02:50:41*    Unable to add physical volume 
'/d/backends/patchy_snap_loop' to volume group 'patchy_snap_vg_1'.
            *02:50:41*    Volume group "patchy_snap_vg_1" not found
            *02:50:41*    Cannot process volume group patchy_snap_vg_1
            *02:50:42*    Volume group "patchy_snap_vg_1" not found
            *02:50:42*    Cannot process volume group patchy_snap_vg_1
            *02:50:42* /dev/patchy_snap_vg_1/brick_lvm: No such file or 
directory
            *02:50:42* Usage: mkfs.xfs
            *02:50:42* /* blocksize */          [-b log=n|size=num]
            *02:50:42* /* data subvol */        [-d 
agcount=n,agsize=n,file,name=xxx,size=num,
            *02:50:42*                      
(sunit=value,swidth=value|su=num,sw=num),
            *02:50:42*                      sectlog=n|sectsize=num
            *02:50:42* /* inode size */ [-i 
log=n|perblock=n|size=num,maxpct=n,attr=0|1|2,
            *02:50:42*                      projid32bit=0|1]
            *02:50:42* /* log subvol */ [-l 
agnum=n,internal,size=num,logdev=xxx,version=n
            *02:50:42*                      
sunit=value|su=num,sectlog=n|sectsize=num,
            *02:50:42*                      lazy-count=0|1]
            *02:50:42* /* label */              [-L label (maximum 12 
characters)]
            *02:50:42* /* naming */             [-n log=n|size=num,version=2|ci]
            *02:50:42* /* prototype file */     [-p fname]
            *02:50:42* /* quiet */              [-q]
            *02:50:42* /* realtime subvol */    [-r 
extsize=num,size=num,rtdev=xxx]
            *02:50:42* /* sectorsize */ [-s log=n|size=num]
            *02:50:42* /* version */            [-V]
            *02:50:42*                  devicename
            *02:50:42* <devicename> is required unless -d name=xxx is given.
            *02:50:42* <num> is xxx (bytes), xxxs (sectors), xxxb (fs blocks), 
xxxk (xxx KiB),
            *02:50:42*        xxxm (xxx MiB), xxxg (xxx GiB), xxxt (xxx TiB) or 
xxxp (xxx PiB).
            *02:50:42* <value> is xxx (512 byte blocks).
            *02:50:42* mount: special device /dev/patchy_snap_vg_1/brick_lvm 
does not exist
            *02:50:53* ./tests/basic/op_errnos.t ..
            *02:50:53* 1..21
            *02:50:53* ok 1, LINENUM:12
            *02:50:53* ok 2, LINENUM:13
            *02:50:53* ok 3, LINENUM:14
            *02:50:53* ok 4, LINENUM:16
            *02:50:53* ok 5, LINENUM:18
            *02:50:53* ok 6, LINENUM:19
            *02:50:53* ok 7, LINENUM:20



            On 07/12/2016 03:47 PM, Atin Mukherjee wrote:
            Hi Avra,

            The above fails locally as well along with few regression
            failures I observed and one of them are at [1]

            not ok 12 Got " 30807" instead of "30809", LINENUM:26
            FAILED COMMAND: 30809 get-op_errno-xml snapshot restore snap1

            not ok 17 Got " 30815" instead of "30812", LINENUM:31
            FAILED COMMAND: 30812 get-op_errno-xml snapshot create
            snap1 patchy no-timestamp

            [1]
            
https://build.gluster.org/job/rackspace-regression-2GB-triggered/22154/console

            --Atin




--
        --Atin




--
    --Atin

    _______________________________________________
    Gluster-devel mailing list
    Gluster-devel@gluster.org <mailto:Gluster-devel@gluster.org>
    http://www.gluster.org/mailman/listinfo/gluster-devel



_______________________________________________
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

Reply via email to