Hello Yipei, "*octavia.amphorae.backends.agent.api_server.listener [-] Failed to verify haproxy file: Command '['haproxy', '-c', '-L', 'NK20KVuD6oi5NrRP7KOVflM* *3MsQ', '-f', '/var/lib/octavia/bca2c985-471a-4477-8217-92fa71d04cb7/haproxy.cfg.new']' returned non-zero exit status 1*"
Verification of haproxy cfg file is failing. You can create a dummy file from the haproxy template files(jinja2 files) and verify on any system with haproxy installed. *haproxy -c -f "filename"* Regards, Ganpat On Wed, Jun 28, 2017 at 3:19 PM, Yipei Niu <newy...@gmail.com> wrote: > Hi, Michael, > > Thanks for your help. I have already created a load balancer successfully, > but failed creating a listener. The detailed errors of amphora-agent and > syslog in the amphora are as follows. > > In amphora-agent.log: > > [2017-06-28 08:54:12 +0000] [1209] [INFO] Starting gunicorn 19.7.0 > [2017-06-28 08:54:13 +0000] [1209] [DEBUG] Arbiter booted > [2017-06-28 08:54:13 +0000] [1209] [INFO] Listening at: http://[::]:9443 > (1209) > [2017-06-28 08:54:13 +0000] [1209] [INFO] Using worker: sync > [2017-06-28 08:54:13 +0000] [1209] [DEBUG] 1 workers > [2017-06-28 08:54:13 +0000] [1816] [INFO] Booting worker with pid: 1816 > [2017-06-28 08:54:15 +0000] [1816] [DEBUG] POST /0.5/plug/vip/10.0.1.8 > ::ffff:192.168.0.12 - - [28/Jun/2017:08:54:59 +0000] "POST /0.5/plug/vip/ > 10.0.1.8 HTTP/1.1" 202 78 "-" "Octavia HaProxy Rest Client/0.5 ( > https://wiki.openstack.org/wiki/Octavia)" > [2017-06-28 08:59:18 +0000] [1816] [DEBUG] PUT > /0.5/listeners/9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4/ > bca2c985-471a-4477-8217-92fa71d04cb7/haproxy > ::ffff:192.168.0.12 - - [28/Jun/2017:08:59:19 +0000] "PUT > /0.5/listeners/9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4/ > bca2c985-471a-4477-8217-92fa71d04cb7/haproxy HTTP/1.1" 400 414 "-" > "Octavia HaProxy Rest Client/0.5 (https://wiki.openstack.org/wiki/Octavia > )" > > In syslog: > > Jun 28 08:57:14 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: > ############################################################# > Jun 28 08:57:14 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: > -----BEGIN SSH HOST KEY FINGERPRINTS----- > Jun 28 08:57:14 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: 1024 > SHA256:qDQcKq2Je/CzlpPndccMf0aR0u/KPJEEIAl4RraAgVc > root@amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 (DSA) > Jun 28 08:57:15 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: 256 > SHA256:n+5tCCdJwASMaD/kJ6fm0kVNvXDh4aO0si2Uls4MXkI > root@amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 (ECDSA) > Jun 28 08:57:15 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: 256 > SHA256:7RWMBOW+QKzeolI6BDSpav9dVZuon58weIQJ9/peVxE > root@amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 (ED25519) > Jun 28 08:57:16 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: 2048 > SHA256:9z+EcAAUyTENKJRctKCzPslK6Yf4c7s9R8sEflDITIU > root@amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 (RSA) > Jun 28 08:57:16 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: > -----END SSH HOST KEY FINGERPRINTS----- > Jun 28 08:57:16 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 ec2: > ############################################################# > Jun 28 08:57:17 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 > cloud-init[2092]: Cloud-init v. 0.7.9 running 'modules:final' at Wed, 28 > Jun 2017 08:57:03 +0000. Up 713.82 seconds. > Jun 28 08:57:17 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 > cloud-init[2092]: Cloud-init v. 0.7.9 finished at Wed, 28 Jun 2017 08:57:16 > +0000. Datasource DataSourceConfigDrive [net,ver=2][source=/dev/sr0]. Up > 727.30 seconds > Jun 28 08:57:19 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 systemd[1]: > Started Execute cloud user/final scripts. > Jun 28 08:57:19 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 systemd[1]: > Reached target Cloud-init target. > Jun 28 08:57:19 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 systemd[1]: > Startup finished in 52.054s (kernel) + 11min 17.647s (userspace) = 12min > 9.702s. > Jun 28 08:59:19 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 > amphora-agent[1209]: 2017-06-28 08:59:19.243 1816 ERROR > octavia.amphorae.backends.agent.api_server.listener [-] Failed to verify > haproxy file: Command '['haproxy', '-c', '-L', 'NK20KVuD6oi5NrRP7KOVflM > 3MsQ', '-f', > '/var/lib/octavia/bca2c985-471a-4477-8217-92fa71d04cb7/haproxy.cfg.new']' > returned non-zero exit status 1 > Jun 28 09:00:11 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 systemd[1]: > Starting Cleanup of Temporary Directories... > Jun 28 09:00:12 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 > systemd-tmpfiles[3040]: [/usr/lib/tmpfiles.d/var.conf:14] Duplicate line > for path "/var/log", ignoring. > Jun 28 09:00:15 amphora-9ed4f0a5-6b1e-4832-97cc-fb8d1518cbd4 systemd[1]: > Started Cleanup of Temporary Directories. > > Look forward to your valuable comments. > > Best regards, > Yipei > > On Tue, Jun 27, 2017 at 2:33 PM, Yipei Niu <newy...@gmail.com> wrote: > >> Hi, Micheal, >> >> Thanks a lot for your help, but I still have one question. >> >> In Octavia, once the controller worker fails plugging VIP to the amphora, >> the amphora is deleted immediately, making it impossible to trace the >> error. How to prevent Octavia from stopping and deleting the amphora? >> >> Best regards, >> Yipei >> >> On Mon, Jun 26, 2017 at 11:21 AM, Yipei Niu <newy...@gmail.com> wrote: >> >>> Hi, all, >>> >>> I am trying to create a load balancer in octavia. The amphora can be >>> booted successfully, and can be reached via icmp. However, octavia fails to >>> plug vip to the amphora through the amphora client api and returns 500 >>> status code, causing some errors as follows. >>> >>> |__Flow >>> 'octavia-create-loadbalancer-flow': InternalServerError: Internal >>> Server Error >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> Traceback (most recent call last): >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> File >>> "/usr/local/lib/python2.7/dist-packages/taskflow/engines/action_engine/executor.py", >>> line 53, in _execute_task >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> result = task.execute(**arguments) >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> File >>> "/opt/stack/octavia/octavia/controller/worker/tasks/amphora_driver_tasks.py", >>> line 240, in execute >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> amphorae_network_config) >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> File >>> "/opt/stack/octavia/octavia/controller/worker/tasks/amphora_driver_tasks.py", >>> line 219, in execute >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> amphora, loadbalancer, amphorae_network_config) >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> File >>> "/opt/stack/octavia/octavia/amphorae/drivers/haproxy/rest_api_driver.py", >>> line 137, in post_vip_plug >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> net_info) >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> File >>> "/opt/stack/octavia/octavia/amphorae/drivers/haproxy/rest_api_driver.py", >>> line 378, in plug_vip >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> return exc.check_exception(r) >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> File "/opt/stack/octavia/octavia/amphorae/drivers/haproxy/exceptions.py", >>> line 32, in check_exception >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> raise responses[status_code]() >>> 2017-06-21 09:49:35.864 25411 ERROR >>> octavia.controller.worker.controller_worker >>> InternalServerError: Internal Server Error >>> 2017-06-21 09:49:35.864 25411 ERROR octavia.controller.worker.cont >>> roller_worker >>> >>> To fix the problem, I log in the amphora and find that there is one http >>> server process is listening on port 9443, so I think the amphora api >>> services is active. But do not know how to further investigate what error >>> happens inside the amphora api service and solve it? Look forward to your >>> valuable comments. >>> >>> Best regards, >>> Yipei >>> >> >> > > __________________________________________________________________________ > OpenStack Development Mailing List (not for usage questions) > Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > >
__________________________________________________________________________ OpenStack Development Mailing List (not for usage questions) Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev