Thanks again for your prompt response My apologies for wasting your time with trivial question but the repo provided does not contain rpms but a bunch of compressed files ( like 2f241a8387cf35372fd709be4ef6ec83b8a00cc744bb90f31d82bb27bdd80531-other.sqlite.bz2 ) and a repomd.xml
How/what exactly should I download/install ? Steven On 22 January 2018 at 09:18, Jason Dillaman <jdill...@redhat.com> wrote: > You can use these repos [1][2][3] > > [1] https://shaman.ceph.com/repos/python-rtslib/master/ > [2] https://shaman.ceph.com/repos/ceph-iscsi-config/master/ > [3] https://shaman.ceph.com/repos/ceph-iscsi-cli/master/ > > targetcli isn't used for iSCSI over RBD (gwcli from ceph-iscsi-cli > replaces it), so you hopefully shouldn't need to update it. > > On Mon, Jan 22, 2018 at 9:06 AM, Steven Vacaroaia <ste...@gmail.com> > wrote: > > Excellent news > > Many thanks for all your efforts > > > > If you do not mind, please confirm the following steps ( for centos 7, > > kernel version 3.10.0-693.11.6.el7.x86_64) > > > > - download and Install the RPMs from x86_64 repositories you provided > > - do a git clone and, if new version available "pip install . --upgrade" > for > > ceph-iscsi-cli > > ceph-iscsi-config > > rtslib-fb > > targetcli-fb > > - reboot > > > > Steven > > > > > > On 22 January 2018 at 08:53, Jason Dillaman <jdill...@redhat.com> wrote: > >> > >> The v4.13-based kernel with the necessary bug fixes and TCMU changes > >> is available here [1] and tcmu-runner v1.3.0 is available here [2]. > >> > >> [1] https://shaman.ceph.com/repos/kernel/ceph-iscsi-test/ > >> [2] https://shaman.ceph.com/repos/tcmu-runner/master/ > >> > >> On Sat, Jan 20, 2018 at 7:33 AM, Marc Roos <m.r...@f1-outsourcing.eu> > >> wrote: > >> > > >> > > >> > Sorry for me asking maybe the obvious but is this the kernel available > >> > in elrepo? Or a different one? > >> > > >> > > >> > > >> > > >> > > >> > -----Original Message----- > >> > From: Mike Christie [mailto:mchri...@redhat.com] > >> > Sent: zaterdag 20 januari 2018 1:19 > >> > To: Steven Vacaroaia; Joshua Chen > >> > Cc: ceph-users > >> > Subject: Re: [ceph-users] iSCSI over RBD > >> > > >> > On 01/19/2018 02:12 PM, Steven Vacaroaia wrote: > >> >> Hi Joshua, > >> >> > >> >> I was under the impression that kernel 3.10.0-693 will work with > >> > iscsi > >> >> > >> > > >> > That kernel works with RHCS 2.5 and below. You need the rpms from that > >> > or the matching upstream releases. Besides trying to dig out the > >> > versions and matching things up, the problem with those releases is > that > >> > they were tech previewed or only supports linux initiators. > >> > > >> > It looks like you are using the newer upstream tools or RHCS 3.0 > tools. > >> > For them you need the RHEL 7.5 beta or newer kernel or an upstream > one. > >> > For upstream all the patches got merged into the target layer > >> > maintainer's tree yesterday. A new tcmu-runner release has been made. > >> > And I just pushed a test kernel with all the patches based on 4.13 > (4.14 > >> > had a bug in the login code which is being fixed still) to github, so > >> > people do not have to wait for the next-next kernel release to come > out. > >> > > >> > Just give us a couple days for the kernel build to be done, to make > the > >> > needed ceph-iscsi-* release (current version will fail to create rbd > >> > images with the current tcmu-runner release) and get the documentation > >> > updated because some links are incorrect and some version info needs > to > >> > be updated. > >> > > >> > > >> >> Unfortunately I still cannot create a disk because qfull_time_out is > >> >> not supported > >> >> > >> >> What am I missing / do it wrong ? > >> >> > >> >> 2018-01-19 15:06:45,216 INFO [lun.py:601:add_dev_to_lio()] - > >> >> (LUN.add_dev_to_lio) Adding image 'rbd.disk2' to LIO > >> >> 2018-01-19 15:06:45,295 ERROR [lun.py:634:add_dev_to_lio()] - > Could > >> >> not set LIO device attribute cmd_time_out/qfull_time_out for device: > >> >> rbd.disk2. Kernel not supported. - error(Cannot find attribute: > >> >> qfull_time_out) > >> >> 2018-01-19 15:06:45,300 ERROR [rbd-target-api:731:_disk()] - LUN > >> >> alloc problem - Could not set LIO device attribute > >> >> cmd_time_out/qfull_time_out for device: rbd.disk2. Kernel not > >> > supported. > >> >> - error(Cannot find attribute: qfull_time_out) > >> >> > >> >> > >> >> Many thanks > >> >> > >> >> Steven > >> >> > >> >> On 4 January 2018 at 22:40, Joshua Chen <csc...@asiaa.sinica.edu.tw > >> >> <mailto:csc...@asiaa.sinica.edu.tw>> wrote: > >> >> > >> >> Hello Steven, > >> >> I am using CentOS 7.4.1708 with kernel 3.10.0-693.el7.x86_64 > >> >> and the following packages: > >> >> > >> >> ceph-iscsi-cli-2.5-9.el7.centos.noarch.rpm > >> >> ceph-iscsi-config-2.3-12.el7.centos.noarch.rpm > >> >> libtcmu-1.3.0-0.4.el7.centos.x86_64.rpm > >> >> libtcmu-devel-1.3.0-0.4.el7.centos.x86_64.rpm > >> >> python-rtslib-2.1.fb64-2.el7.centos.noarch.rpm > >> >> python-rtslib-doc-2.1.fb64-2.el7.centos.noarch.rpm > >> >> targetcli-2.1.fb47-0.1.20170815.git5bf3517.el7.centos.noarch.rpm > >> >> tcmu-runner-1.3.0-0.4.el7.centos.x86_64.rpm > >> >> tcmu-runner-debuginfo-1.3.0-0.4.el7.centos.x86_64.rpm > >> >> > >> >> > >> >> Cheers > >> >> Joshua > >> >> > >> >> > >> >> On Fri, Jan 5, 2018 at 2:14 AM, Steven Vacaroaia < > ste...@gmail.com > >> >> <mailto:ste...@gmail.com>> wrote: > >> >> > >> >> Hi Joshua, > >> >> > >> >> How did you manage to use iSCSI gateway ? > >> >> I would like to do that but still waiting for a patched > kernel > >> > > >> >> > >> >> What kernel/OS did you use and/or how did you patch it ? > >> >> > >> >> Tahnsk > >> >> Steven > >> >> > >> >> On 4 January 2018 at 04:50, Joshua Chen > >> >> <csc...@asiaa.sinica.edu.tw > >> > <mailto:csc...@asiaa.sinica.edu.tw>> > >> >> wrote: > >> >> > >> >> Dear all, > >> >> Although I managed to run gwcli and created some iqns, > >> > or > >> >> luns, > >> >> but I do need some working config example so that my > >> >> initiator could connect and get the lun. > >> >> > >> >> I am familiar with targetcli and I used to do the > >> >> following ACL style connection rather than password, > >> >> the targetcli setting tree is here: > >> >> > >> >> (or see this page > >> >> <http://www.asiaa.sinica.edu.tw/~cschen/targetcli.html>) > >> >> > >> >> #targetcli ls > >> >> o- / > >> >> > >> > ............................................................ > ............ > >> > ................................................. > >> >> [...] > >> >> o- backstores > >> >> > >> > ............................................................ > ............ > >> > ...................................... > >> >> [...] > >> >> | o- block > >> >> > >> > ............................................................ > ............ > >> > .......................... > >> >> [Storage Objects: 1] > >> >> | | o- vmware_5t > >> >> .............................. > ............................ > >> >> [/dev/rbd/rbd/vmware_5t (5.0TiB) write-thru activated] > >> >> | | o- alua > >> >> > >> > ............................................................ > ............ > >> > ........................... > >> >> [ALUA Groups: 1] > >> >> | | o- default_tg_pt_gp > >> >> > >> > ............................................................ > ........... > >> >> [ALUA state: Active/optimized] > >> >> | o- fileio > >> >> > >> > ............................................................ > ............ > >> > ......................... > >> >> [Storage Objects: 0] > >> >> | o- pscsi > >> >> > >> > ............................................................ > ............ > >> > .......................... > >> >> [Storage Objects: 0] > >> >> | o- ramdisk > >> >> > >> > ............................................................ > ............ > >> > ........................ > >> >> [Storage Objects: 0] > >> >> | o- user:rbd > >> >> > >> > ............................................................ > ............ > >> > ....................... > >> >> [Storage Objects: 0] > >> >> o- iscsi > >> >> > >> > ............................................................ > ............ > >> > .................................... > >> >> [Targets: 1] > >> >> | o- iqn.2017-12.asiaa.cephosd1:vmware5t > >> >> > >> > ............................................................ > ............ > >> > ....... > >> >> [TPGs: 1] > >> >> | o- tpg1 > >> >> > >> > ............................................................ > ............ > >> > .......................... > >> >> [gen-acls, no-auth] > >> >> | o- acls > >> >> > >> > ............................................................ > ............ > >> > ................................. > >> >> [ACLs: 12] > >> >> | | o- iqn.1994-05.com.redhat:15dbed23be9e > >> >> > >> > .................................................................. > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1994-05.com.redhat:15dbed23be9e-ovirt1 > >> >> > >> > ........................................................... > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- > >> >> iqn.1994-05.com.redhat:2af344ba6ae5-ceph-admin-test > >> >> .................................................. > [Mapped > >> >> LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1994-05.com.redhat:67669afedddf > >> >> > >> > .................................................................. > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1994-05.com.redhat:67669afedddf-ovirt3 > >> >> > >> > ........................................................... > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1994-05.com.redhat:a7c1ec3c43f7 > >> >> > >> > .................................................................. > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1994-05.com.redhat:a7c1ec3c43f7-ovirt2 > >> >> > >> > ........................................................... > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- > >> > iqn.1994-05.com.redhat:b01662ec2129-ceph-node2 > >> >> ....................................................... > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- > >> > iqn.1994-05.com.redhat:d46b42a1915b-ceph-node3 > >> >> ....................................................... > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- > >> > iqn.1994-05.com.redhat:e7692a10f661-ceph-node1 > >> >> ....................................................... > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1998-01.com.vmware:localhost-0f904dfd > >> >> > >> > ............................................................ > >> >> [Mapped LUNs: 1] > >> >> | | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | | o- iqn.1998-01.com.vmware:localhost-6af62e4c > >> >> > >> > ............................................................ > >> >> [Mapped LUNs: 1] > >> >> | | o- mapped_lun0 > >> >> > >> > ............................................................ > ............ > >> > ..... > >> >> [lun0 block/vmware_5t (rw)] > >> >> | o- luns > >> >> > >> > ............................................................ > ............ > >> > .................................. > >> >> [LUNs: 1] > >> >> | | o- lun0 > >> >> .................................................... > >> >> [block/vmware_5t (/dev/rbd/rbd/vmware_5t) > >> > (default_tg_pt_gp)] > >> >> | o- portals > >> >> > >> > ............................................................ > ............ > >> > ............................ > >> >> [Portals: 1] > >> >> | o- 172.20.0.12:3260 <http://172.20.0.12:3260> > >> >> > >> > ............................................................ > ............ > >> > ......................... > >> >> [OK] > >> >> o- loopback > >> >> > >> > ............................................................ > ............ > >> > ................................. > >> >> [Targets: 0] > >> >> o- xen_pvscsi > >> >> > >> > ............................................................ > ............ > >> > ............................... > >> >> [Targets: 0] > >> >> > >> >> > >> >> > >> >> > >> >> > >> >> > >> >> My targetcli setup procedure is like this, could someone > >> >> translate it to gwcli equivalent procedure? > >> >> sorry for asking for this due to lack of documentation > and > >> >> examples. > >> >> thanks in adavance > >> >> > >> >> Cheers > >> >> Joshua > >> >> > >> >> > >> >> > >> >> > >> >> targetcli /backstores/block create name=vmware_5t > >> >> dev=/dev/rbd/rbd/vmware_5t > >> >> targetcli /iscsi/ create > >> > iqn.2017-12.asiaa.cephosd1:vmware5t > >> >> targetcli > >> >> /iscsi/iqn.2017-12.asiaa.cephosd1:vmware5t/tpg1/portals > >> >> delete ip_address=0.0.0.0 ip_port=3260 > >> >> > >> >> targetcli > >> >> cd /iscsi/iqn.2017-12.asiaa.cephosd1:vmware5t/tpg1 > >> >> portals/ create 172.20.0.12 > >> >> acls/ > >> >> create > >> >> iqn.1994-05.com.redhat:e7692a10f661-ceph-node1 > >> >> create > >> >> iqn.1994-05.com.redhat:b01662ec2129-ceph-node2 > >> >> create > >> >> iqn.1994-05.com.redhat:d46b42a1915b-ceph-node3 > >> >> create iqn.1994-05.com.redhat: > 15dbed23be9e > >> >> create iqn.1994-05.com.redhat: > a7c1ec3c43f7 > >> >> create iqn.1994-05.com.redhat: > 67669afedddf > >> >> create > >> >> iqn.1994-05.com.redhat:15dbed23be9e-ovirt1 > >> >> create > >> >> iqn.1994-05.com.redhat:a7c1ec3c43f7-ovirt2 > >> >> create > >> >> iqn.1994-05.com.redhat:67669afedddf-ovirt3 > >> >> create > >> >> iqn.1994-05.com.redhat:2af344ba6ae5-ceph-admin-test > >> >> create > >> > iqn.1998-01.com.vmware:localhost-6af62e4c > >> >> create > >> > iqn.1998-01.com.vmware:localhost-0f904dfd > >> >> cd .. > >> >> set attribute generate_node_acls=1 > >> >> cd luns > >> >> create /backstores/block/vmware_5t > >> >> > >> >> > >> >> > >> >> > >> >> On Thu, Jan 4, 2018 at 10:55 AM, Joshua Chen > >> >> <csc...@asiaa.sinica.edu.tw > >> >> <mailto:csc...@asiaa.sinica.edu.tw>> wrote: > >> >> > >> >> I had the same problem before, mine is CentOS, and > >> > when > >> >> I created > >> >> /iscsi/create iqn_bla-bla > >> >> it goes > >> >> ocal LIO instance already has LIO configured with a > >> >> target - unable to continue > >> >> > >> >> > >> >> > >> >> then finally the solution happened to be, turn off > >> >> target service > >> >> > >> >> systemctl stop target > >> >> systemctl disable target > >> >> > >> >> > >> >> somehow they are doing the same thing, you need to > >> >> disable 'target' service (targetcli) in order to > allow > >> >> gwcli (rbd-target-api) do it's job. > >> >> > >> >> Cheers > >> >> Joshua > >> >> > >> >> On Thu, Jan 4, 2018 at 2:39 AM, Mike Christie > >> >> <mchri...@redhat.com > >> > <mailto:mchri...@redhat.com>> wrote: > >> >> > >> >> On 12/25/2017 03:13 PM, Joshua Chen wrote: > >> >> > Hello folks, > >> >> > I am trying to share my ceph rbd images > >> > through > >> >> iscsi protocol. > >> >> > > >> >> > I am trying iscsi-gateway > >> >> > > >> >> > >> > http://docs.ceph.com/docs/master/rbd/iscsi-overview/ > >> >> > >> > <http://docs.ceph.com/docs/master/rbd/iscsi-overview/> > >> >> > > >> >> > > >> >> > now > >> >> > > >> >> > systemctl start rbd-target-api > >> >> > is working and I could run gwcli > >> >> > (at a CentOS 7.4 osd node) > >> >> > > >> >> > gwcli > >> >> > /> ls > >> >> > o- / > >> >> > > >> >> > >> > ............................................................ > ............ > >> > ................................................. > >> >> > [...] > >> >> > o- clusters > >> >> > > >> >> > >> > ............................................................ > ............ > >> > ................................ > >> >> > [Clusters: 1] > >> >> > | o- ceph > >> >> > > >> >> > >> > ............................................................ > ............ > >> > .................................... > >> >> > [HEALTH_OK] > >> >> > | o- pools > >> >> > > >> >> > >> > ............................................................ > ............ > >> > .................................. > >> >> > [Pools: 1] > >> >> > | | o- rbd > >> >> > > >> >> > >> > ............................................................ > ............ > >> > ... > >> >> > [(x3), Commit: 0b/25.9T (0%), Used: 395M] > >> >> > | o- topology > >> >> > > >> >> > >> > ............................................................ > ............ > >> > ........................ > >> >> > [OSDs: 9,MONs: 3] > >> >> > o- disks > >> >> > > >> >> > >> > ............................................................ > ............ > >> > .................................. > >> >> > [0b, Disks: 0] > >> >> > o- iscsi-target > >> >> > > >> >> > >> > ............................................................ > ............ > >> > ............................. > >> >> > [Targets: 0] > >> >> > > >> >> > > >> >> > but when I created iscsi-target, I got > >> >> > > >> >> > Local LIO instance already has LIO configured > >> > with > >> >> a target - unable to > >> >> > continue > >> >> > > >> >> > > >> >> > /> /iscsi-target create > >> >> > > >> >> > >> > iqn.2003-01.org.linux-iscsi.ceph-node1.x8664:sn.571e1ab51af2 > >> >> > Local LIO instance already has LIO configured > >> > with > >> >> a target - unable to > >> >> > continue > >> >> > /> > >> >> > > >> >> > >> >> > >> >> Could you send the output of > >> >> > >> >> targetcli ls > >> >> > >> >> ? > >> >> > >> >> What distro are you using? > >> >> > >> >> You might just have a target setup from a non > >> > gwcli > >> >> source. Maybe from > >> >> the distro targetcli systemd tools. > >> >> > >> >> > >> >> > >> >> > >> >> _______________________________________________ > >> >> ceph-users mailing list > >> >> ceph-users@lists.ceph.com > >> > <mailto:ceph-users@lists.ceph.com> > >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> >> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> > >> >> > >> >> > >> >> > >> >> > >> > > >> > _______________________________________________ > >> > ceph-users mailing list > >> > ceph-users@lists.ceph.com > >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > >> > > >> > _______________________________________________ > >> > ceph-users mailing list > >> > ceph-users@lists.ceph.com > >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > >> > >> > >> -- > >> Jason > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > -- > Jason >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com