HI Yair,
  This is my conf:

[client.rgw.cephrgw]
host = cephrgw01
rgw_frontends = "civetweb port=8080s
ssl_certificate=/etc/pki/tls/cephrgw01.crt"
rgw_zone = RegionOne
keyring = /etc/ceph/ceph.client.rgw.cephrgw.keyring
log_file = /var/log/ceph/client.rgw.cephrgw.log
rgw_keystone_url = https://keystone.xxx,xx:5000
rgw_keystone_admin_user = cloud
rgw_keystone_admin_password = XXXXXXXXXXXXXXX
rgw_keystone_admin_tenant = service
rgw_keystone_accepted_roles = admin member _member_
rgw keystone admin project = service
rgw keystone admin domain = admin
rgw keystone api version = 2
rgw_s3_auth_use_keystone = true
nss_db_path = /var/ceph/nss/
rgw_keystone_verify_ssl = true

Regards, I

2017-03-13 17:28 GMT+01:00 Yair Magnezi <yair.magn...@kenshoo.com>:

> But per the doc the client stanza should include
>  client.radosgw.instance_name
>
>
> [client.rgw.ceph-rgw-02]
> host = ceph-rgw-02
> keyring = /etc/ceph/ceph.client.radosgw.keyring
> log file = /var/log/radosgw/client.radosgw.gateway.log
> rgw_frontends = "civetweb port=8080"
>
>
> "For example, if your node name is gateway-node1, add a section like this
> after the [global] section:
>
> [client.rgw.gateway-node1]
> rgw_frontends = "civetweb port=80" "
>
>
> Here, {hostname} is the short hostname (output of command hostname -s) of
> the node that is going to provide the gateway service i.e, the gateway
> host.
>
> The [client.radosgw.gateway] portion of the gateway instance identifies
> this portion of the Ceph configuration file as configuring a Ceph Storage
> Cluster client where the client type is a Ceph Object Gateway (i.e.,
> radosgw).
>
>
> Does anyone run the RGW on a different port and can share his
> configuration ?
>
> Thanks
>
>
>
>
>
>
>
> On Mon, Mar 13, 2017 at 5:28 PM, Abhishek Lekshmanan <abhis...@suse.com>
> wrote:
>
>>
>>
>> On 03/13/2017 04:06 PM, Yair Magnezi wrote:
>>
>>> Thank you Abhishek
>>>
>>> But still ...
>>>
>>> root@ceph-rgw-02:/var/log/ceph# ps -ef | grep rgw
>>> ceph      1332     1  1 14:59 ?        00:00:00 /usr/bin/radosgw
>>> --cluster=ceph --id *rgw.ceph-rgw-02* -f --setuser ceph --setgroup ceph
>>>
>>>
>>> root@ceph-rgw-02:/var/log/ceph# cat /etc/ceph/ceph.conf
>>> [global]
>>> fsid = 00c167db-aea1-41b4-903b-69b0c86b6a0f
>>> mon_initial_members = ceph-osd-01 ceph-osd-02
>>> mon_host = 10.83.1.78,10.83.1.79
>>> auth_cluster_required = cephx
>>> auth_service_required = cephx
>>> auth_client_required = cephx
>>> public_network = 10.83.1.0/24 <http://10.83.1.0/24>
>>> rbd default features = 3
>>> #debug ms = 1
>>> #debug rgw = 20
>>>
>>> [client.radosgw.*rgw.ceph-rgw-02*]
>>> host = ceph-rgw-02
>>> keyring = /etc/ceph/ceph.client.radosgw.keyring
>>> log file = /var/log/radosgw/client.radosgw.gateway.log
>>> rgw_frontends = "civetweb port=8080"
>>>
>>
>>
>> Try client.rgw.ceph-rgw-02 here (and similar in id), ie. basically what
>> you pass as the id should be what the ceph.conf section should look like.
>>
>>
>>> root@ceph-rgw-02:/var/log/ceph# netstat -an | grep 80
>>> tcp        0      0 0.0.0.0:7480 <http://0.0.0.0:7480>
>>>  0.0.0.0:*               LISTEN
>>> tcp        0      0 10.83.1.100:56884 <http://10.83.1.100:56884>
>>> 10.83.1.78:6800 <http://10.83.1.78:6800>         ESTABLISHED
>>> tcp        0      0 10.83.1.100:47842 <http://10.83.1.100:47842>
>>> 10.83.1.78:6804 <http://10.83.1.78:6804>         TIME_WAIT
>>> tcp        0      0 10.83.1.100:47846 <http://10.83.1.100:47846>
>>> 10.83.1.78:6804 <http://10.83.1.78:6804>         ESTABLISHED
>>> tcp        0      0 10.83.1.100:44791 <http://10.83.1.100:44791>
>>> 10.83.1.80:6804 <http://10.83.1.80:6804>         ESTABLISHED
>>> tcp        0      0 10.83.1.100:44782 <http://10.83.1.100:44782>
>>> 10.83.1.80:6804 <http://10.83.1.80:6804>         TIME_WAIT
>>> tcp        0      0 10.83.1.100:38082 <http://10.83.1.100:38082>
>>> 10.83.1.80:6789 <http://10.83.1.80:6789>         ESTABLISHED
>>> tcp        0      0 10.83.1.100:41999 <http://10.83.1.100:41999>
>>> 10.83.1.80:6800 <http://10.83.1.80:6800>         ESTABLISHED
>>> tcp        0      0 10.83.1.100:59681 <http://10.83.1.100:59681>
>>> 10.83.1.79:6800 <http://10.83.1.79:6800>         ESTABLISHED
>>> tcp        0      0 10.83.1.100:37590 <http://10.83.1.100:37590>
>>> 10.83.1.79:6804 <http://10.83.1.79:6804>         ESTABLISHED
>>>
>>>
>>> 2017-03-13 15:05:23.836844 7f5c2fc80900  0 starting handler: civetweb
>>> 2017-03-13 15:05:23.838497 7f5c11379700  0 -- 10.83.1.100:0/2130438046
>>> <http://10.83.1.100:0/2130438046> submit_message
>>> mon_subscribe({osdmap=48}) v2 remote, 10.83.1.78:6789/0
>>> <http://10.83.1.78:6789/0>, failed lossy con, dropping message
>>> 0x7f5bfc011850
>>> 2017-03-13 15:05:23.842769 7f5c11379700  0 monclient: hunting for new mon
>>> 2017-03-13 15:05:23.846976 7f5c2fc80900  0 starting handler: fastcgi
>>> 2017-03-13 15:05:23.849245 7f5b87a6a700  0 ERROR: no socket server point
>>> defined, cannot start fcgi frontend
>>>
>>>
>>>
>>>
>>> Any more ideas
>>>
>>> Thanks
>>>
>>>
>>>
>>>
>>> *
>>> *
>>> **
>>>
>>>
>>>
>>> On Mon, Mar 13, 2017 at 4:34 PM, Abhishek Lekshmanan <abhis...@suse.com
>>> <mailto:abhis...@suse.com>> wrote:
>>>
>>>
>>>
>>>     On 03/13/2017 03:26 PM, Yair Magnezi wrote:
>>>
>>>         Hello Wido
>>>
>>>         yes , the is my  /etc/cep/ceph.conf
>>>
>>>         and yes  radosgw.ceph-rgw-02 is the running instance .
>>>
>>>         root@ceph-rgw-02:/var/log/ceph# ps -ef | grep -i rgw
>>>         ceph     17226     1  0 14:02 ?        00:00:01 /usr/bin/radosgw
>>>         --cluster=ceph --id rgw.ceph-rgw-02 -f --setuser ceph --setgroup
>>>         ceph
>>>
>>>
>>>     The ID passed to rgw here is `rgw.ceph-rgw-02`, whereas your conf
>>>     has a section named `radosgw.ceph-rgw-02` try running this service
>>>     (systemctl start ceph-rado...@radosgw.ceph-rgw-02 maybe?)
>>>
>>>     --
>>>     Abhishek Lekshmanan
>>>     SUSE Linux GmbH, GF: Felix Imendörffer, Jane Smithard, Graham
>>>     Norton, HRB 21284 (AG Nürnberg)
>>>
>>>
>>>         Thanks
>>>
>>>
>>>
>>>         *Yair Magnezi *
>>>         *Storage & Data Protection TL   // *Kenshoo*
>>>         *Office* +972 7 32862423 <tel:%2B972%207%2032862423>   //
>>>         *Mobile* +972 50 575-2955 <tel:%2B972%2050%20575-2955>
>>>         __________________________________________
>>>         *
>>>         **
>>>
>>>
>>>
>>>         On Mon, Mar 13, 2017 at 4:06 PM, Wido den Hollander
>>>         <w...@42on.com <mailto:w...@42on.com>
>>>         <mailto:w...@42on.com <mailto:w...@42on.com>>> wrote:
>>>
>>>
>>>             > Op 13 maart 2017 om 15:03 schreef Yair Magnezi
>>>             <yair.magn...@kenshoo.com <mailto:yair.magn...@kenshoo.com>
>>>         <mailto:yair.magn...@kenshoo.com
>>>
>>>         <mailto:yair.magn...@kenshoo.com>>>:
>>>             >
>>>             >
>>>             > Hello Cephers .
>>>             >
>>>             > I'm trying to modify the   civetweb default  port to 80
>>>         but from some
>>>             > reason it insists on listening on the default 7480 port
>>>             >
>>>             > My configuration is quiet  simple ( experimental  ) and
>>>         looks like this :
>>>             >
>>>             >
>>>             > [global]
>>>             > fsid = 00c167db-aea1-41b4-903b-69b0c86b6a0f
>>>             > mon_initial_members = ceph-osd-01 ceph-osd-02
>>>             > mon_host = 10.83.1.78,10.83.1.79
>>>             > auth_cluster_required = cephx
>>>             > auth_service_required = cephx
>>>             > auth_client_required = cephx
>>>             > public_network = 10.83.1.0/24 <http://10.83.1.0/24>
>>>         <http://10.83.1.0/24>
>>>             > rbd default features = 3
>>>             > #debug ms = 1
>>>             > #debug rgw = 20
>>>             >
>>>             > [client.radosgw.ceph-rgw-02]
>>>             > host = ceph-rgw-02
>>>             > keyring = /etc/ceph/ceph.client.radosgw.keyring
>>>             > log file = /var/log/radosgw/client.radosgw.gateway.log
>>>             > rgw_frontends = "civetweb port=80"
>>>             >
>>>             >
>>>
>>>             Are you sure this is in /etc/ceph/ceph.conf?
>>>
>>>             In addition, are you also sure the RGW is running as user
>>>             'radosgw.ceph-rgw-02' ?
>>>
>>>             Wido
>>>
>>>             > after restart still the same :
>>>             >
>>>             > root@ceph-rgw-02:/var/log/ceph# netstat -an |  grep 80
>>>             > *tcp        0      0 0.0.0.0:7480 <http://0.0.0.0:7480>
>>>         <http://0.0.0.0:7480>
>>>             <http://0.0.0.0:7480>
>>>             >  0.0.0.0:*               LISTEN*
>>>             > tcp        0      0 10.83.1.100:56697
>>>         <http://10.83.1.100:56697> <http://10.83.1.100:56697>
>>>                  10.83.1.79:6804 <http://10.83.1.79:6804>
>>>         <http://10.83.1.79:6804>
>>>             > ESTABLISHED
>>>             > tcp        0      0 10.83.1.100:59482
>>>         <http://10.83.1.100:59482> <http://10.83.1.100:59482>
>>>                  10.83.1.79:6800 <http://10.83.1.79:6800>
>>>         <http://10.83.1.79:6800>
>>>             > TIME_WAIT
>>>             > tcp        0      0 10.83.1.100:33129
>>>         <http://10.83.1.100:33129> <http://10.83.1.100:33129>
>>>                  10.83.1.78:6804 <http://10.83.1.78:6804>
>>>         <http://10.83.1.78:6804>
>>>             > ESTABLISHED
>>>             > tcp        0      0 10.83.1.100:56318
>>>         <http://10.83.1.100:56318> <http://10.83.1.100:56318>
>>>                  10.83.1.80:6804 <http://10.83.1.80:6804>
>>>         <http://10.83.1.80:6804>
>>>             > TIME_WAIT
>>>             > tcp        0      0 10.83.1.100:56324
>>>         <http://10.83.1.100:56324> <http://10.83.1.100:56324>
>>>                  10.83.1.80:6804 <http://10.83.1.80:6804>
>>>         <http://10.83.1.80:6804>
>>>             > ESTABLISHED
>>>             > tcp        0      0 10.83.1.100:60990
>>>         <http://10.83.1.100:60990> <http://10.83.1.100:60990>
>>>                  10.83.1.78:6800 <http://10.83.1.78:6800>
>>>         <http://10.83.1.78:6800>
>>>             > ESTABLISHED
>>>             > tcp        0      0 10.83.1.100:60985
>>>         <http://10.83.1.100:60985> <http://10.83.1.100:60985>
>>>                  10.83.1.78:6800 <http://10.83.1.78:6800>
>>>         <http://10.83.1.78:6800>
>>>             > TIME_WAIT
>>>             > tcp        0      0 10.83.1.100:56691
>>>         <http://10.83.1.100:56691> <http://10.83.1.100:56691>
>>>                  10.83.1.79:6804 <http://10.83.1.79:6804>
>>>         <http://10.83.1.79:6804>
>>>             > TIME_WAIT
>>>             > tcp        0      0 10.83.1.100:33123
>>>         <http://10.83.1.100:33123> <http://10.83.1.100:33123>
>>>                  10.83.1.78:6804 <http://10.83.1.78:6804>
>>>         <http://10.83.1.78:6804>
>>>             > TIME_WAIT
>>>             > tcp        0      0 10.83.1.100:59494
>>>         <http://10.83.1.100:59494> <http://10.83.1.100:59494>
>>>                  10.83.1.79:6800 <http://10.83.1.79:6800>
>>>         <http://10.83.1.79:6800>
>>>             > ESTABLISHED
>>>             > tcp        0      0 10.83.1.100:55924
>>>         <http://10.83.1.100:55924> <http://10.83.1.100:55924>
>>>                  10.83.1.80:6800 <http://10.83.1.80:6800>
>>>         <http://10.83.1.80:6800>
>>>             > ESTABLISHED
>>>             > tcp        0      0 10.83.1.100:57629
>>>         <http://10.83.1.100:57629> <http://10.83.1.100:57629>
>>>                  10.83.1.80:6789 <http://10.83.1.80:6789>
>>>         <http://10.83.1.80:6789>
>>>             > ESTABLISHED
>>>             >
>>>             >
>>>             > Besides that it also looks like the service tries  to
>>>         start the
>>>             fcgi  (
>>>             > besides the civetweb ) is there a reason for that ?  (
>>>         fastcgi &
>>>             Apache are
>>>             > not  installed )  ?
>>>             >
>>>             >
>>>             > 2017-03-13 13:44:35.938897 7f05f3fd7700  1 handle_sigterm
>>> set
>>>             alarm for 120
>>>             > 2017-03-13 13:44:35.938916 7f06692c7900 -1 shutting down
>>>             > 2017-03-13 13:44:36.170559 7f06692c7900  1 final shutdown
>>>             > 2017-03-13 13:45:13.980814 7fbdb2e6c900  0 deferred set
>>>         uid:gid to
>>>             > 64045:64045 (ceph:ceph)
>>>             > 2017-03-13 13:45:13.980992 7fbdb2e6c900  0 ceph version
>>> 10.2.6
>>>             > (656b5b63ed7c43bd014bcafd81b001959d5f089f), process
>>>         radosgw, pid 16995
>>>             > *2017-03-13 13:45:14.115639 7fbdb2e6c900  0 starting
>>> handler:
>>>             civetweb*
>>>             > 2017-03-13 13:45:14.117003 7fbd8ffff700  0 --
>>>         10.83.1.100:0/3241919968 <http://10.83.1.100:0/3241919968>
>>>         <http://10.83.1.100:0/3241919968 <http://10.83.1.100:0/32419199
>>> 68>>
>>>
>>>             > submit_message mon_subscribe({osdmap=48}) v2 remote,
>>>         10.83.1.78:6789/0 <http://10.83.1.78:6789/0>
>>>         <http://10.83.1.78:6789/0>,
>>>
>>>             > failed lossy con, dropping message 0x7fbd78011cc0
>>>             > 2017-03-13 13:45:14.117644 7fbd8ffff700  0 monclient:
>>>         hunting for new mon
>>>             > *2017-03-13 13:45:14.120767 7fbdb2e6c900  0 starting
>>>         handler: fastcgi*
>>>             > *2017-03-13 13:45:14.123000 7fbd09f7b700  0 ERROR: no
>>> socket
>>>             server point
>>>             > defined, cannot start fcgi frontend*
>>>             >
>>>             >
>>>             > Any idea what do i miss here
>>>             >
>>>             >
>>>             > ceph version --> 10.2.6
>>>             > ceph-rgw-02 is the gateway .
>>>             >
>>>             > Many Thanks
>>>             >
>>>             > Yair
>>>             >
>>>             > --
>>>             > This e-mail, as well as any attached document, may contain
>>>             material which
>>>             > is confidential and privileged and may include trademark,
>>>             copyright and
>>>             > other intellectual property rights that are proprietary to
>>>         Kenshoo
>>>             Ltd,
>>>             >  its subsidiaries or affiliates ("Kenshoo"). This e-mail
>>>         and its
>>>             > attachments may be read, copied and used only by the
>>>         addressee for the
>>>             > purpose(s) for which it was disclosed herein. If you have
>>>         received
>>>             it in
>>>             > error, please destroy the message and any attachment, and
>>>         contact us
>>>             > immediately. If you are not the intended recipient, be
>>>         aware that any
>>>             > review, reliance, disclosure, copying, distribution or use
>>>         of the
>>>             contents
>>>             > of this message without Kenshoo's express permission is
>>>         strictly
>>>             prohibited.
>>>             > _______________________________________________
>>>             > ceph-users mailing list
>>>             > ceph-users@lists.ceph.com
>>>         <mailto:ceph-users@lists.ceph.com>
>>>         <mailto:ceph-users@lists.ceph.com
>>>
>>>         <mailto:ceph-users@lists.ceph.com>>
>>>             > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>         <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>>             <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>         <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>>
>>>
>>>
>>>
>>>         This e-mail, as well as any attached document, may contain
>>> material
>>>         which is confidential and privileged and may include trademark,
>>>         copyright and other intellectual property rights that are
>>>         proprietary to
>>>         Kenshoo Ltd,  its subsidiaries or affiliates ("Kenshoo"). This
>>>         e-mail
>>>         and its attachments may be read, copied and used only by the
>>>         addressee
>>>         for the purpose(s) for which it was disclosed herein. If you have
>>>         received it in error, please destroy the message and any
>>>         attachment, and
>>>         contact us immediately. If you are not the intended recipient,
>>>         be aware
>>>         that any review, reliance, disclosure, copying, distribution or
>>>         use of
>>>         the contents of this message without Kenshoo's express
>>> permission is
>>>         strictly prohibited.
>>>
>>>
>>>         _______________________________________________
>>>         ceph-users mailing list
>>>         ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>>         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>         <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>>
>>>
>>>
>>> This e-mail, as well as any attached document, may contain material
>>> which is confidential and privileged and may include trademark,
>>> copyright and other intellectual property rights that are proprietary to
>>> Kenshoo Ltd,  its subsidiaries or affiliates ("Kenshoo"). This e-mail
>>> and its attachments may be read, copied and used only by the addressee
>>> for the purpose(s) for which it was disclosed herein. If you have
>>> received it in error, please destroy the message and any attachment, and
>>> contact us immediately. If you are not the intended recipient, be aware
>>> that any review, reliance, disclosure, copying, distribution or use of
>>> the contents of this message without Kenshoo's express permission is
>>> strictly prohibited.
>>>
>>
>> --
>> Abhishek Lekshmanan
>> SUSE Linux GmbH, GF: Felix Imendörffer, Jane Smithard, Graham Norton, HRB
>> 21284 (AG Nürnberg)
>>
>
>
> This e-mail, as well as any attached document, may contain material which
> is confidential and privileged and may include trademark, copyright and
> other intellectual property rights that are proprietary to Kenshoo Ltd,
>  its subsidiaries or affiliates ("Kenshoo"). This e-mail and its
> attachments may be read, copied and used only by the addressee for the
> purpose(s) for which it was disclosed herein. If you have received it in
> error, please destroy the message and any attachment, and contact us
> immediately. If you are not the intended recipient, be aware that any
> review, reliance, disclosure, copying, distribution or use of the contents
> of this message without Kenshoo's express permission is strictly prohibited.
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
############################################################################
Iban Cabrillo Bartolome
Instituto de Fisica de Cantabria (IFCA)
Santander, Spain
Tel: +34942200969
PGP PUBLIC KEY:
http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
############################################################################
Bertrand Russell:*"El problema con el mundo es que los estúpidos están
seguros de todo y los inteligentes están **llenos de dudas*"
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to