On Fri, Apr 12, 2019 at 09:18 RUIZ LOPEZ Noel wrote: > Hello > > First of all, thanks for your quick answer. > > I have tried to upgrade from 3.0 to 3.2, but I get this error when I try to > install the 3.2 > "Error: centos-release-ceph-luminous conflicts with > centos-release-ceph-jewel-1.0-1.el7.centos.noarch" >
Oops, that was addressed by an upgrade note for the 3.1 release, as explained here: https://www.softwarefactory-project.io/releases/3.1/ (search for "Upgrade Notes") You have to do: "yum remove -y centos-release-ceph-jewel" before installing the sf-release-3.2.rpm Regards, -Tristan > I see that I can fix it adding --skip-broken but.... I am not pretty sure If > this is a good idea...I will wait for your opinión. In my first approach I > tried to deploy directly sf 3.2 but with our arch I get some errors and at > the end I give up. Anyway I can try again. > > > By other hand, I remount /srv/host-rootfs to try but, this doesn't works. > > > I attach our arch.yaml in case it could help > > ________________________________________ > De: Tristan Cacqueray [[email protected]] > Enviado: viernes, 12 de abril de 2019 2:55 > Para: Javier Pena; RUIZ LOPEZ Noel; [email protected] > Asunto: Re: [Softwarefactory-dev] Zuul NODE FAILURE > > On Thu, Apr 11, 2019 at 12:08 Javier Pena wrote: >> ----- Original Message ----- >> >>> Hello, >> >>> I have just deployed software factory and when I tried to test how zuul >>> works, I got the following zuul error: >> >>> NODE_FAILURE >> >>> Now, I can see that nodes always keep "building" state. >> >>> nodepool log : >> >>> 2019-04-11 17:55:18,203 ERROR nodepool.NodeLauncher-0000000046: Launch >>> attempt 9/9 failed for node 0000000046: >>> Traceback (most recent call last): >>> File >>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/handler.py", >>> line 40, in launch >>> self.handler.pool, hostid, port, self.label) >>> File >>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/provider.py", >>> line 149, in createContainer >>> "Manager %s failed to initialized" % self.provider.name) >>> RuntimeError: Manager oci-provider-hypervisor-oci failed to initialized >>> 2019-04-11 17:55:19,208 ERROR nodepool.NodeLauncher-0000000046: Launch >>> failed >>> for node 0000000046: >>> Traceback (most recent call last): >>> File >>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/__init__.py", >>> line 659, in run >>> self.launch() >>> File >>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/handler.py", >>> line 57, in launch >>> self.node.host_keys = key >>> UnboundLocalError: local variable 'key' referenced before assignment >>> 2019-04-11 17:55:19,208 ERROR nodepool.NodeLauncher-0000000045: Launch >>> failed >>> for node 0000000045: >>> Traceback (most recent call last): >>> File >>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/__init__.py", >>> line 659, in run >>> self.launch() >>> File >>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/handler.py", >>> line 57, in launch >>> self.node.host_keys = key >>> UnboundLocalError: local variable 'key' referenced before assignment >>> 2019-04-11 17:55:22,918 INFO nodepool.DeletedNodeWorker: Deleting failed >>> instance 0000000045-centos-oci-100-0000000045 from >>> oci-provider-hypervisor-oci >>> 2019-04-11 17:55:22,926 INFO nodepool.NodeDeleter: Deleting ZK node >>> id=0000000045, state=deleting, >>> external_id=0000000045-centos-oci-100-0000000045 >>> 2019-04-11 17:55:22,934 INFO nodepool.DeletedNodeWorker: Deleting failed >>> instance 0000000046-centos-oci-100-0000000046 from >>> oci-provider-hypervisor-oci >>> 2019-04-11 17:55:22,940 INFO nodepool.NodeDeleter: Deleting ZK node >>> id=0000000046, state=deleting, >>> external_id=0000000046-centos-oci-100-0000000046 >>> 2019-04-11 17:55:26,276 INFO nodepool.NodePool: Creating requests for 2 >>> centos-oci nodes >>> 2019-04-11 17:55:29,822 INFO >>> nodepool.PoolWorker.oci-provider-hypervisor-oci-main: Assigning node request >>> <NodeRequest {'id': '100-0000000047', 'node_types': ['centos-oci'], 'state': >>> 'requested', 'state_time': 1554998126.2781763, 'stat': >>> ZnodeStat(czxid=11466, mzxid=11466, ctime=1554998126279, >>> mtime=1554998126279, version=0, cversion=0, aversion=0, ephemeralOwner=0, >>> dataLength=217, numChildren=0, pzxid=11466), 'nodes': [], 'reuse': False, >>> 'declined_by': [], 'requestor': 'NodePool:min-ready'}> >>> 2019-04-11 17:55:29,845 WARNING nodepool.driver.oci.OpenContainerProvider: >>> Creating container when provider isn't ready >> >>> Any idea? > > Hello Noel, > > NODE_ERROR indicates a failure to start the nodes, and the exception you > found in the logs is an issue that has been fixed in newer version. > It seems like you deployed Software Factory version 3.0, since 3.1 the > drivers has been renamed runC and greatly improved. > Can you try to upgrade to version 3.2: > > https://www.softwarefactory-project.io/docs/3.2/operator/upgrade.html > > After the upgrade process, please restart the instance (that's because > we don't support upgrade from 3.0, and restart is needed to refresh the > services). > > >> >> Hi Noel, >> >> I'm not sure if it's the same situation, but last time I tried to use oci >> containers I had to remount /srv/host-rootfs as read-write before it would >> work the first time (it is mounted as read-only by default). After this and >> a reboot, it worked fine as ro. >> >> So can you try a quick "mount -o remount,rw /srv/host-rootfs" and see if it >> fixes it? >> > > Since version 3.1 (and the rename to runC), we fixed a critical issue > with bubblewrap and this remount shouldn't be needed anymore. > > Regards, > -Tristan > description: Minimal Software Factory deployment > inventory: > - hostname: managesf.sftests.com > ip: 10.6.71.81 > name: managesf > public_url: https://sftests.com > roles: > - install-server > - mysql > - zookeeper > - gateway > - cauth > - managesf > - etherpad > - lodgeit > - gitweb > - gerrit > - gerritbot > - logserver > - zuul-scheduler > - zuul-executor > - zuul-web > - nodepool-launcher > - murmur > - mirror > - kibana > - repoxplorer > - hydrant > - firehose > - grafana > - rabbitmq > - storyboard > - storyboard-webclient > - hostname: elk.sftests.com > ip: 192.168.71.82 > name: elk > public_url: http://elk.sftests.com > roles: > - elasticsearch > - logstash > - influxdb > - hostname: nodepool-builder.sftests.com > ip: 192.168.71.83 > name: nodepool-builder > public_url: http://nodepool-builder.sftests.com > roles: > - nodepool-builder > - hostname: zuul-merger.sftests.com > ip: 192.168.71.84 > name: zuul-merger > public_url: http://zuul-merger.sftests.com > roles: > - zuul-merger > - hostname: hypervisor-oci.sftests.com > ip: 192.168.71.86 > max-servers: 10 > name: hypervisor-oci > public_url: http://hypervisor-oci.sftests.com > remote: true > roles: > - hypervisor-oci
signature.asc
Description: PGP signature
_______________________________________________ Softwarefactory-dev mailing list [email protected] https://www.redhat.com/mailman/listinfo/softwarefactory-dev
