Please file a bug against project-infrastructure. On Tue, Jul 12, 2016 at 6:50 PM, Raghavendra Talur <rta...@redhat.com> wrote:
> Nigel/Misc, > > Could you please look into this? > slave29 does not seem to have a xfs formatted backend for tests. > > Thanks, > Raghavendra Talur > > On Tue, Jul 12, 2016 at 6:41 PM, Avra Sengupta <aseng...@redhat.com> > wrote: > >> Atin, >> >> I am not sure about the docker containers, but both the failures you >> mentioned are in slave29, which as Talur explained is missing the >> appropriate backend filesystem. Owing to this, op-errno.t is just the tip >> of the iceberg, and every other test that uses lvm will fail in this >> particular slave will fail too. >> >> Talur, >> Thanks for looking into it. It is indeed strange this. I checked the >> dmesg and the /var/log/messages in this slave and I couldn't find any >> relevant log. >> >> >> On 07/12/2016 05:29 PM, Raghavendra Talur wrote: >> >> I checked the machine. >> >> Here is the df -hT output >> [jenkins@slave29 ~]$ cat /etc/fstab >> # Accessible filesystems, by reference, are maintained under '/dev/disk' >> # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more >> info >> # >> /dev/xvda1 / ext3 >> defaults,noatime,barrier=0 1 1 >> tmpfs /dev/shm tmpfs defaults 0 >> 0 >> devpts /dev/pts devpts gid=5,mode=620 0 >> 0 >> sysfs /sys sysfs defaults 0 >> 0 >> proc /proc proc defaults 0 >> 0 >> #/dev/xvdc1 none swap sw 0 >> 0 >> >> >> We don't see a xfs device mounted at /d and / is of type ext3 which does >> not support fallocate. The uptime of the machine is 73 days though. I don't >> know how the /d xfs partition vanished. >> >> On Tue, Jul 12, 2016 at 4:54 PM, Atin Mukherjee < <amukh...@redhat.com> >> amukh...@redhat.com> wrote: >> >>> >>> https://build.gluster.org/job/rackspace-regression-2GB-triggered/22156/consoleFull >>> - another failure >>> >>> On Tue, Jul 12, 2016 at 4:42 PM, Atin Mukherjee < <amukh...@redhat.com> >>> amukh...@redhat.com> wrote: >>> >>>> >>>> >>>> On Tue, Jul 12, 2016 at 4:36 PM, Avra Sengupta < <aseng...@redhat.com> >>>> aseng...@redhat.com> wrote: >>>> >>>>> Hi Atin, >>>>> >>>>> Please check the testcase result in the console. It clearly states the >>>>> reason of the failure. A quick search of 30815, as shown in the testcase >>>>> shows that the error that is generated is a thinp issue, and we can see >>>>> fallocate failing and lvm not properly being setup in the environment. >>>>> >>>> >>>> While this is valid for my docker containers, I am just wondering why >>>> did this happen in jenkins slave? >>>> >>>> >>>>> Regards, >>>>> Avra >>>>> >>>>> P.S Here are the logs from the console stating so. >>>>> >>>>> *02:50:34* [09:50:34] Running tests in file >>>>> ./tests/basic/op_errnos.t*02:50:41* fallocate: >>>>> /d/backends/patchy_snap_vhd: fallocate failed: Operation not >>>>> supported*02:50:41* losetup: /d/backends/patchy_snap_vhd: warning: file >>>>> smaller than 512 bytes, the loop device maybe be useless or invisible for >>>>> system tools.*02:50:41* Device /d/backends/patchy_snap_loop not found >>>>> (or ignored by filtering).*02:50:41* Device >>>>> /d/backends/patchy_snap_loop not found (or ignored by >>>>> filtering).*02:50:41* Unable to add physical volume >>>>> '/d/backends/patchy_snap_loop' to volume group >>>>> 'patchy_snap_vg_1'.*02:50:41* Volume group "patchy_snap_vg_1" not >>>>> found*02:50:41* Cannot process volume group patchy_snap_vg_1*02:50:42* >>>>> Volume group "patchy_snap_vg_1" not found*02:50:42* Cannot process >>>>> volume group patchy_snap_vg_1*02:50:42* /dev/patchy_snap_vg_1/brick_lvm: >>>>> No such file or directory*02:50:42* Usage: mkfs.xfs*02:50:42* /* >>>>> blocksize */ [-b log=n|size=num]*02:50:42* /* data subvol */ >>>>> [-d agcount=n,agsize=n,file,name=xxx,size=num,*02:50:42* >>>>> (sunit=value,swidth=value|su=num,sw=num),*02:50:42* >>>>> sectlog=n|sectsize=num*02:50:42* /* inode size */ [-i >>>>> log=n|perblock=n|size=num,maxpct=n,attr=0|1|2,*02:50:42* >>>>> projid32bit=0|1]*02:50:42* /* log subvol */ [-l >>>>> agnum=n,internal,size=num,logdev=xxx,version=n*02:50:42* >>>>> sunit=value|su=num,sectlog=n|sectsize=num,*02:50:42* >>>>> lazy-count=0|1]*02:50:42* /* label */ [-L label >>>>> (maximum 12 characters)]*02:50:42* /* naming */ [-n >>>>> log=n|size=num,version=2|ci]*02:50:42* /* prototype file */ [-p >>>>> fname]*02:50:42* /* quiet */ [-q]*02:50:42* /* realtime >>>>> subvol */ [-r extsize=num,size=num,rtdev=xxx]*02:50:42* /* sectorsize >>>>> */ [-s log=n|size=num]*02:50:42* /* version */ >>>>> [-V]*02:50:42* devicename*02:50:42* <devicename> is >>>>> required unless -d name=xxx is given.*02:50:42* <num> is xxx (bytes), >>>>> xxxs (sectors), xxxb (fs blocks), xxxk (xxx KiB),*02:50:42* xxxm >>>>> (xxx MiB), xxxg (xxx GiB), xxxt (xxx TiB) or xxxp (xxx PiB).*02:50:42* >>>>> <value> is xxx (512 byte blocks).*02:50:42* mount: special device >>>>> /dev/patchy_snap_vg_1/brick_lvm does not exist*02:50:53* >>>>> ./tests/basic/op_errnos.t .. *02:50:53* 1..21*02:50:53* ok 1, >>>>> LINENUM:12*02:50:53* ok 2, LINENUM:13*02:50:53* ok 3, >>>>> LINENUM:14*02:50:53* ok 4, LINENUM:16*02:50:53* ok 5, >>>>> LINENUM:18*02:50:53* ok 6, LINENUM:19*02:50:53* ok 7, LINENUM:20 >>>>> >>>>> >>>>> >>>>> >>>>> On 07/12/2016 03:47 PM, Atin Mukherjee wrote: >>>>> >>>>> Hi Avra, >>>>> >>>>> The above fails locally as well along with few regression failures I >>>>> observed and one of them are at [1] >>>>> >>>>> not ok 12 Got " 30807" instead of "30809", LINENUM:26 >>>>> FAILED COMMAND: 30809 get-op_errno-xml snapshot restore snap1 >>>>> >>>>> not ok 17 Got " 30815" instead of "30812", LINENUM:31 >>>>> FAILED COMMAND: 30812 get-op_errno-xml snapshot create snap1 patchy >>>>> no-timestamp >>>>> >>>>> [1] >>>>> <https://build.gluster.org/job/rackspace-regression-2GB-triggered/22154/console> >>>>> https://build.gluster.org/job/rackspace-regression-2GB-triggered/22154/console >>>>> >>>>> --Atin >>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> >>>> --Atin >>>> >>> >>> >>> >>> -- >>> >>> --Atin >>> >>> _______________________________________________ >>> Gluster-devel mailing list >>> Gluster-devel@gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-devel >>> >> >> >> > -- nigelb
_______________________________________________ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel