Have attached in the Bug https://bugzilla.redhat.com/show_bug.cgi?id=1611635
On Thu, 2 Aug 2018, 22:21 Raghavendra Gowdappa, <rgowd...@redhat.com> wrote: > > > On Thu, Aug 2, 2018 at 5:48 PM, Kotresh Hiremath Ravishankar < > khire...@redhat.com> wrote: > >> I am facing different issue in softserve machines. The fuse mount itself >> is failing. >> I tried day before yesterday to debug geo-rep failures. I discussed with >> Raghu, >> but could not root cause it. >> > > Where can I find the complete client logs for this? > > So none of the tests were passing. It happened on >> both machine instances I tried. >> >> ------------------------ >> [2018-07-31 10:41:49.288117] D [fuse-bridge.c:5407:notify] 0-fuse: got >> event 6 on graph 0 >> [2018-07-31 10:41:49.289427] D [fuse-bridge.c:4990:fuse_get_mount_status] >> 0-fuse: mount status is 0 >> [2018-07-31 10:41:49.289555] D [fuse-bridge.c:4256:fuse_init] >> 0-glusterfs-fuse: Detected support for FUSE_AUTO_INVAL_DATA. Enabling >> fopen_keep_cache automatically. >> [2018-07-31 10:41:49.289591] T [fuse-bridge.c:278:send_fuse_iov] >> 0-glusterfs-fuse: writev() result 40/40 >> [2018-07-31 10:41:49.289610] I [fuse-bridge.c:4314:fuse_init] >> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.24 kernel >> 7.22 >> [2018-07-31 10:41:49.289627] I [fuse-bridge.c:4948:fuse_graph_sync] >> 0-fuse: switched to graph 0 >> [2018-07-31 10:41:49.289696] T [MSGID: 0] [syncop.c:1261:syncop_lookup] >> 0-stack-trace: stack-address: 0x7f36e4001058, winding from fuse to >> meta-autoload >> [2018-07-31 10:41:49.289743] T [MSGID: 0] >> [defaults.c:2716:default_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from meta-autoload to master >> [2018-07-31 10:41:49.289787] T [MSGID: 0] >> [io-stats.c:2788:io_stats_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from master to master-md-cache >> [2018-07-31 10:41:49.289833] T [MSGID: 0] >> [md-cache.c:513:mdc_inode_iatt_get] 0-md-cache: mdc_inode_ctx_get failed >> (00000000-0000-0000-0000-000000000001) >> [2018-07-31 10:41:49.289923] T [MSGID: 0] [md-cache.c:1200:mdc_lookup] >> 0-stack-trace: stack-address: 0x7f36e4001058, winding from master-md-cache >> to master-open-behind >> [2018-07-31 10:41:49.289946] T [MSGID: 0] >> [defaults.c:2716:default_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from master-open-behind to master-quick-read >> [2018-07-31 10:41:49.289973] T [MSGID: 0] [quick-read.c:556:qr_lookup] >> 0-stack-trace: stack-address: 0x7f36e4001058, winding from >> master-quick-read to master-io-cache >> [2018-07-31 10:41:49.290002] T [MSGID: 0] [io-cache.c:298:ioc_lookup] >> 0-stack-trace: stack-address: 0x7f36e4001058, winding from master-io-cache >> to master-readdir-ahead >> [2018-07-31 10:41:49.290034] T [MSGID: 0] >> [defaults.c:2716:default_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from master-readdir-ahead to master-read-ahead >> [2018-07-31 10:41:49.290052] T [MSGID: 0] >> [defaults.c:2716:default_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from master-read-ahead to master-write-behind >> [2018-07-31 10:41:49.290077] T [MSGID: 0] [write-behind.c:2439:wb_lookup] >> 0-stack-trace: stack-address: 0x7f36e4001058, winding from >> master-write-behind to master-dht >> [2018-07-31 10:41:49.290156] D [MSGID: 0] >> [dht-common.c:3674:dht_do_fresh_lookup] 0-master-dht: /: no subvolume in >> layout for path, checking on all the subvols to see if it is a directory >> [2018-07-31 10:41:49.290180] D [MSGID: 0] >> [dht-common.c:3688:dht_do_fresh_lookup] 0-master-dht: /: Found null hashed >> subvol. Calling lookup on all nodes. >> [2018-07-31 10:41:49.290199] T [MSGID: 0] >> [dht-common.c:3695:dht_do_fresh_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from master-dht to master-replicate-0 >> [2018-07-31 10:41:49.290245] I [MSGID: 108006] >> [afr-common.c:5582:afr_local_init] 0-master-replicate-0: no subvolumes up >> [2018-07-31 10:41:49.290291] D [MSGID: 0] >> [afr-common.c:3212:afr_discover] 0-stack-trace: stack-address: >> 0x7f36e4001058, master-replicate-0 returned -1 error: Transport endpoint is >> not conne >> cted [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290323] D [MSGID: 0] >> [dht-common.c:1391:dht_lookup_dir_cbk] 0-master-dht: lookup of / on >> master-replicate-0 returned error [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290350] T [MSGID: 0] >> [dht-common.c:3695:dht_do_fresh_lookup] 0-stack-trace: stack-address: >> 0x7f36e4001058, winding from master-dht to master-replicate-1 >> [2018-07-31 10:41:49.290381] I [MSGID: 108006] >> [afr-common.c:5582:afr_local_init] 0-master-replicate-1: no subvolumes up >> [2018-07-31 10:41:49.290403] D [MSGID: 0] >> [afr-common.c:3212:afr_discover] 0-stack-trace: stack-address: >> 0x7f36e4001058, master-replicate-1 returned -1 error: Transport endpoint is >> not connected [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290427] D [MSGID: 0] >> [dht-common.c:1391:dht_lookup_dir_cbk] 0-master-dht: lookup of / on >> master-replicate-1 returned error [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290452] D [MSGID: 0] >> [dht-common.c:1574:dht_lookup_dir_cbk] 0-stack-trace: stack-address: >> 0x7f36e4001058, master-dht returned -1 error: Transport endpoint is not >> connected [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290477] D [MSGID: 0] >> [write-behind.c:2393:wb_lookup_cbk] 0-stack-trace: stack-address: >> 0x7f36e4001058, master-write-behind returned -1 error: Transport endpoint >> is not connected [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290504] D [MSGID: 0] [io-cache.c:268:ioc_lookup_cbk] >> 0-stack-trace: stack-address: 0x7f36e4001058, master-io-cache returned -1 >> error: Transport endpoint is not connected [Transport endpoint is not >> connected] >> [2018-07-31 10:41:49.290530] D [MSGID: 0] >> [quick-read.c:515:qr_lookup_cbk] 0-stack-trace: stack-address: >> 0x7f36e4001058, master-quick-read returned -1 error: Transport endpoint is >> not connected [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290554] D [MSGID: 0] >> [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: >> 0x7f36e4001058, master-md-cache returned -1 error: Transport endpoint is >> not connected [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290581] D [MSGID: 0] >> [io-stats.c:2276:io_stats_lookup_cbk] 0-stack-trace: stack-address: >> 0x7f36e4001058, master returned -1 error: Transport endpoint is not >> connected [Transport endpoint is not connected] >> [2018-07-31 10:41:49.290626] E [fuse-bridge.c:4382:fuse_first_lookup] >> 0-fuse: first lookup on root failed (Transport endpoint is not connected) >> --------------------------------------------- >> >> On Thu, Aug 2, 2018 at 5:35 PM, Nigel Babu <nig...@redhat.com> wrote: >> >>> On Thu, Aug 2, 2018 at 5:12 PM Kotresh Hiremath Ravishankar < >>> khire...@redhat.com> wrote: >>> >>>> Don't know, something to do with perf xlators I suppose. It's not >>>> repdroduced on my local system with brick-mux enabled as well. But it's >>>> happening on Xavis' system. >>>> >>>> Xavi, >>>> Could you try with the patch [1] and let me know whether it fixes the >>>> issue. >>>> >>>> [1] https://review.gluster.org/#/c/20619/1 >>>> >>> >>> If you cannot reproduce it on your laptop, why don't you request a >>> machine from softserve[1] and try it out? >>> >>> [1]: >>> https://github.com/gluster/softserve/wiki/Running-Regressions-on-clean-Centos-7-machine >>> >>> -- >>> nigelb >>> >> >> >> >> -- >> Thanks and Regards, >> Kotresh H R >> >> _______________________________________________ >> maintainers mailing list >> maintain...@gluster.org >> https://lists.gluster.org/mailman/listinfo/maintainers >> >> >
_______________________________________________ Gluster-devel mailing list Gluster-devel@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-devel