Never mind folks, it was that I had an older version of nginx running with
a different config file. After I added the upload module and restarted it,
now I could simply go to the UI by just specifying the host name.

But...how do I stop/start this still remains unclear. please help if you
could. :-)

I see that I have so many python processes and postgresql processes:

bioinfo+ 101567 101415  0 19:41 ?        00:00:04
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python2
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/uwsgi --virtualenv
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv --ini-paste
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini --logdate
--thunder-lock --master --processes 2 --threads 2 --logto
/media/libraryfiles/bioinfoadmin/bioinfoadmin/uwsgi.log --socket
127.0.0.1:4001 --pythonpath lib --stats 127.0.0.1:9191 -b 16384
bioinfo+ 101568 101415  1 19:41 ?        00:01:38
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python
./lib/galaxy/main.py -c
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini
--server-name=handler0
--log-file=/media/libraryfiles/bioinfoadmin/bioinfoadmin/handler0.log
bioinfo+ 101569 101415  1 19:41 ?        00:01:37
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python
./lib/galaxy/main.py -c
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini
--server-name=handler1
--log-file=/media/libraryfiles/bioinfoadmin/bioinfoadmin/handler1.log
bioinfo+ 101570 101415  1 19:41 ?        00:01:38
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python
./lib/galaxy/main.py -c
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini
--server-name=handler2
--log-file=/media/libraryfiles/bioinfoadmin/bioinfoadmin/handler2.log
bioinfo+ 101571 101415  1 19:41 ?        00:01:38
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python
./lib/galaxy/main.py -c
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini
--server-name=handler3
--log-file=/media/libraryfiles/bioinfoadmin/bioinfoadmin/handler3.log
bioinfo+ 102477 101567  1 19:41 ?        00:01:39
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python2
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/uwsgi --virtualenv
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv --ini-paste
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini --logdate
--thunder-lock --master --processes 2 --threads 2 --logto
/media/libraryfiles/bioinfoadmin/bioinfoadmin/uwsgi.log --socket
127.0.0.1:4001 --pythonpath lib --stats 127.0.0.1:9191 -b 16384
bioinfo+ 102478 101567  1 19:41 ?        00:01:39
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/python2
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/bin/uwsgi --virtualenv
/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv --ini-paste
/media/libraryfiles/bioinfoadmin/bioinfoadmin/config/galaxy.ini --logdate
--thunder-lock --master --processes 2 --threads 2 --logto
/media/libraryfiles/bioinfoadmin/bioinfoadmin/uwsgi.log --socket
127.0.0.1:4001 --pythonpath lib --stats 127.0.0.1:9191 -b 16384

$ ps -ef | grep -i post
postgres 102744      1  0 19:41 ?        00:00:00
/usr/lib/postgresql/10/bin/postgres -D /var/lib/postgresql/10/main -c
config_file=/etc/postgresql/10/main/postgresql.conf
postgres 102747 102744  0 19:41 ?        00:00:00 postgres: 10/main:
checkpointer process
postgres 102748 102744  0 19:41 ?        00:00:00 postgres: 10/main: writer
process
postgres 102749 102744  0 19:41 ?        00:00:00 postgres: 10/main: wal
writer process
postgres 102750 102744  0 19:41 ?        00:00:00 postgres: 10/main:
autovacuum launcher process
postgres 102751 102744  0 19:41 ?        00:00:00 postgres: 10/main: stats
collector process
postgres 102752 102744  0 19:41 ?        00:00:00 postgres: 10/main:
bgworker: logical replication launcher
postgres 103679 101415  0 19:42 ?        00:00:00
/usr/lib/postgresql/9.5/bin/postmaster -D /var/lib/postgresql/9.5/main -c
config_file=/etc/postgresql/9.5/main/postgresql.conf
postgres 103689 103679  0 19:42 ?        00:00:00 postgres: checkpointer
process
postgres 103690 103679  0 19:42 ?        00:00:00 postgres: writer process
postgres 103691 103679  0 19:42 ?        00:00:00 postgres: wal writer
process
postgres 103692 103679  0 19:42 ?        00:00:00 postgres: autovacuum
launcher process
postgres 103693 103679  0 19:42 ?        00:00:01 postgres: stats collector
process
postgres 103694 103679  0 19:42 ?        00:00:04 postgres: bioinfoadmin
galaxy 127.0.0.1(36144) idle
postgres 103695 103679  0 19:42 ?        00:00:04 postgres: bioinfoadmin
galaxy 127.0.0.1(36146) idle
postgres 103696 103679  0 19:42 ?        00:00:04 postgres: bioinfoadmin
galaxy 127.0.0.1(36148) idle
postgres 103697 103679  0 19:42 ?        00:00:04 postgres: bioinfoadmin
galaxy 127.0.0.1(36150) idle
postgres 103698 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36152) idle
postgres 103699 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36154) idle
postgres 103700 103679  0 19:42 ?        00:00:12 postgres: bioinfoadmin
galaxy 127.0.0.1(36156) idle
postgres 103701 103679  0 19:42 ?        00:00:12 postgres: bioinfoadmin
galaxy 127.0.0.1(36158) idle
postgres 103702 103679  0 19:42 ?        00:00:12 postgres: bioinfoadmin
galaxy 127.0.0.1(36160) idle
postgres 103703 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36162) idle
postgres 103704 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36164) idle
postgres 103705 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36166) idle
postgres 103706 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36168) idle
postgres 103707 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36170) idle
postgres 103708 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36172) idle
postgres 103709 103679  0 19:42 ?        00:00:11 postgres: bioinfoadmin
galaxy 127.0.0.1(36174) idle
postgres 103710 103679  0 19:42 ?        00:00:04 postgres: bioinfoadmin
galaxy 127.0.0.1(36176) idle
postgres 103711 103679  0 19:42 ?        00:00:04 postgres: bioinfoadmin
galaxy 127.0.0.1(36178) idle

Is there a script to do the work?

Thanks,
Rui

On Fri, Oct 12, 2018 at 8:10 PM Rui Wang <ruiwang...@gmail.com> wrote:

> Hi Folks,
>
> So I managed to ran through the ansible. What I did was to remove certain
> steps. For example, I don't need slurm-drmaa and docker, then I skipped the
> update cache. After these minor changes, it finished successfully(?) with
> an error message it ignored. Then I tried to access the UI, but nothing
> worked. I pasted it below for your reference. If anyone has seen this
> before, please help. :-)
>
> Sorry for the beginner question, so once the ansible playbook ran
> through...seems the galaxy and the postgresql etc were started. How do I
> start/stop them manually? Do I have to run the playbook every time I want
> to run galaxy or only after I modify the playbook? Also, if I don't look at
> the log, I don't even know that the UI is at 127.0.0.1:4001.  Is there
> any documentation for this?
>
> Thanks,
> Rui
>
>
> error message:
>
> galaxy.web.stack INFO 2018-10-12 19:41:53,874 [p:102477,w:1,m:0]
> [MainThread] Galaxy server instance 'main.web.1' is running
> Starting server in PID 101567.
> serving on uwsgi://127.0.0.1:4001
> galaxy.jobs.handler ERROR 2018-10-12 19:42:48,487 [p:102477,w:1,m:0]
> [JobHandlerQueue.monitor_thread] Exception in monitor_step
> Traceback (most recent call last):
>   File "lib/galaxy/jobs/handler.py", line 213, in __monitor
>     self.__monitor_step()
>   File "lib/galaxy/jobs/handler.py", line 272, in __monitor_step
>     .order_by(model.Job.id).all()
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/orm/query.py",
> line 2737, in all
>     return list(self)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/orm/query.py",
> line 2889, in __iter__
>     return self._execute_and_instances(context)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/orm/query.py",
> line 2912, in _execute_and_instances
>     result = conn.execute(querycontext.statement, self._params)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 948, in execute
>     return meth(self, multiparams, params)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/sql/elements.py",
> line 269, in _execute_on_connection
>     return connection._execute_clauseelement(self, multiparams, params)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1060, in _execute_clauseelement
>     compiled_sql, distilled_params
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1200, in _execute_context
>     context)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1413, in _handle_dbapi_exception
>     exc_info
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/util/compat.py",
> line 203, in raise_from_cause
>     reraise(type(exception), exception, tb=exc_tb, cause=cause)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1193, in _execute_context
>     context)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/default.py",
> line 507, in do_execute
>     cursor.execute(statement, parameters)
> OperationalError: (psycopg2.OperationalError) server closed the connection
> unexpectedly
>         This probably means the server terminated abnormally
>         before or while processing the request.
>  [SQL: 'SELECT EXISTS (SELECT history_dataset_association.id,
> history_dataset_association.history_id,
> history_dataset_association.dataset_id,
> history_dataset_association.create_time,
> history_dataset_association.update_time, history_dataset_association.state,
> history_dataset_association.copied_from_history_dataset_association_id,
> history_dataset_association.copied_from_library_dataset_dataset_association_id,
> history_dataset_association.name, history_dataset_association.info,
> history_dataset_association.blurb, history_dataset_association.peek,
> history_dataset_association.tool_version,
> history_dataset_association.extension,
> history_dataset_association.metadata,
> history_dataset_association.parent_id,
> history_dataset_association.designation,
> history_dataset_association.deleted, history_dataset_association.visible,
> history_dataset_association.extended_metadata_id,
> history_dataset_association.version, history_dataset_association.hid,
> history_dataset_association.purged,
> history_dataset_association.hidden_beneath_collection_instance_id \nFROM
> history_dataset_association, job_to_output_dataset \nWHERE job.id =
> job_to_output_dataset.job_id AND history_dataset_association.id =
> job_to_output_dataset.dataset_id AND history_dataset_association.deleted =
> true) AS anon_1, EXISTS (SELECT history_dataset_collection_association.id
> \nFROM history_dataset_collection_association,
> job_to_output_dataset_collection \nWHERE job.id =
> job_to_output_dataset_collection.job_id AND
> history_dataset_collection_association.id =
> job_to_output_dataset_collection.dataset_collection_id AND
> history_dataset_collection_association.deleted = true) AS anon_2, job.id
> AS job_id, job.create_time AS job_create_time, job.update_time AS
> job_update_time, job.history_id AS job_history_id, job.library_folder_id AS
> job_library_folder_id, job.tool_id AS job_tool_id, job.tool_version AS
> job_tool_version, job.state AS job_state, job.info AS job_info,
> job.copied_from_job_id AS job_copied_from_job_id, job.command_line AS
> job_command_line, job.dependencies AS job_dependencies, job.param_filename
> AS job_param_filename, job.runner_name AS job_runner_name_1, job.stdout AS
> job_stdout, job.stderr AS job_stderr, job.exit_code AS job_exit_code,
> job.traceback AS job_traceback, job.session_id AS job_session_id,
> job.user_id AS job_user_id, job.job_runner_name AS job_job_runner_name,
> job.job_runner_external_id AS job_job_runner_external_id,
> job.destination_id AS job_destination_id, job.destination_params AS
> job_destination_params, job.object_store_id AS job_object_store_id,
> job.imported AS job_imported, job.params AS job_params, job.handler AS
> job_handler \nFROM job \nWHERE job.state = %(state_1)s AND job.handler =
> %(handler_1)s AND job.id NOT IN (SELECT job.id \nFROM job JOIN
> job_to_input_dataset ON job.id = job_to_input_dataset.job_id JOIN
> history_dataset_association ON history_dataset_association.id =
> job_to_input_dataset.dataset_id JOIN dataset ON dataset.id =
> history_dataset_association.dataset_id \nWHERE job.state = %(state_2)s AND
> (history_dataset_association.state = %(_state_1)s OR
> history_dataset_association.deleted = true OR dataset.state != %(state_3)s
> OR dataset.deleted = true)) AND job.id NOT IN (SELECT job.id \nFROM job
> JOIN job_to_input_library_dataset ON job.id =
> job_to_input_library_dataset.job_id JOIN
> library_dataset_dataset_association ON
> library_dataset_dataset_association.id =
> job_to_input_library_dataset.ldda_id JOIN dataset ON dataset.id =
> library_dataset_dataset_association.dataset_id \nWHERE job.state =
> %(state_4)s AND (library_dataset_dataset_association.state IS NOT NULL OR
> library_dataset_dataset_association.deleted = true OR dataset.state !=
> %(state_5)s OR dataset.deleted = true)) ORDER BY job.id'] [parameters:
> {'state_3': 'ok', 'handler_1': 'main.web.1', 'state_1': 'new', '_state_1':
> 'failed_metadata', 'state_2': 'new', 'state_5': 'ok', 'state_4': 'new'}]
> (Background on this error at: http://sqlalche.me/e/e3q8)
> galaxy.jobs.handler ERROR 2018-10-12 19:42:48,488 [p:102478,w:2,m:0]
> [JobHandlerQueue.monitor_thread] Exception in monitor_step
> Traceback (most recent call last):
>   File "lib/galaxy/jobs/handler.py", line 213, in __monitor
>     self.__monitor_step()
>   File "lib/galaxy/jobs/handler.py", line 272, in __monitor_step
>     .order_by(model.Job.id).all()
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/orm/query.py",
> line 2737, in all
>     return list(self)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/orm/query.py",
> line 2889, in __iter__
>     return self._execute_and_instances(context)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/orm/query.py",
> line 2912, in _execute_and_instances
>     result = conn.execute(querycontext.statement, self._params)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 948, in execute
>     return meth(self, multiparams, params)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/sql/elements.py",
> line 269, in _execute_on_connection
>     return connection._execute_clauseelement(self, multiparams, params)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1060, in _execute_clauseelement
>     compiled_sql, distilled_params
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1200, in _execute_context
>     context)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1413, in _handle_dbapi_exception
>     exc_info
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/util/compat.py",
> line 203, in raise_from_cause
>     reraise(type(exception), exception, tb=exc_tb, cause=cause)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
> line 1193, in _execute_context
>     context)
>   File
> "/media/libraryfiles/bioinfoadmin/bioinfoadmin/.venv/local/lib/python2.7/site-packages/sqlalchemy/engine/default.py",
> line 507, in do_execute
>     cursor.execute(statement, parameters)
> OperationalError: (psycopg2.OperationalError) server closed the connection
> unexpectedly
>         This probably means the server terminated abnormally
>         before or while processing the request.
>  [SQL: 'SELECT EXISTS (SELECT history_dataset_association.id,
> history_dataset_association.history_id,
> history_dataset_association.dataset_id,
> history_dataset_association.create_time,
> history_dataset_association.update_time, history_dataset_association.state,
> history_dataset_association.copied_from_history_dataset_association_id,
> history_dataset_association.copied_from_library_dataset_dataset_association_id,
> history_dataset_association.name, history_dataset_association.info,
> history_dataset_association.blurb, history_dataset_association.peek,
> history_dataset_association.tool_version,
> history_dataset_association.extension,
> history_dataset_association.metadata,
> history_dataset_association.parent_id,
> history_dataset_association.designation,
> history_dataset_association.deleted, history_dataset_association.visible,
> history_dataset_association.extended_metadata_id,
> history_dataset_association.version, history_dataset_association.hid,
> history_dataset_association.purged,
> history_dataset_association.hidden_beneath_collection_instance_id \nFROM
> history_dataset_association, job_to_output_dataset \nWHERE job.id =
> job_to_output_dataset.job_id AND history_dataset_association.id =
> job_to_output_dataset.dataset_id AND history_dataset_association.deleted =
> true) AS anon_1, EXISTS (SELECT history_dataset_collection_association.id
> \nFROM history_dataset_collection_association,
> job_to_output_dataset_collection \nWHERE job.id =
> job_to_output_dataset_collection.job_id AND
> history_dataset_collection_association.id =
> job_to_output_dataset_collection.dataset_collection_id AND
> history_dataset_collection_association.deleted = true) AS anon_2, job.id
> AS job_id, job.create_time AS job_create_time, job.update_time AS
> job_update_time, job.history_id AS job_history_id, job.library_folder_id AS
> job_library_folder_id, job.tool_id AS job_tool_id, job.tool_version AS
> job_tool_version, job.state AS job_state, job.info AS job_info,
> job.copied_from_job_id AS job_copied_from_job_id, job.command_line AS
> job_command_line, job.dependencies AS job_dependencies, job.param_filename
> AS job_param_filename, job.runner_name AS job_runner_name_1, job.stdout AS
> job_stdout, job.stderr AS job_stderr, job.exit_code AS job_exit_code,
> job.traceback AS job_traceback, job.session_id AS job_session_id,
> job.user_id AS job_user_id, job.job_runner_name AS job_job_runner_name,
> job.job_runner_external_id AS job_job_runner_external_id,
> job.destination_id AS job_destination_id, job.destination_params AS
> job_destination_params, job.object_store_id AS job_object_store_id,
> job.imported AS job_imported, job.params AS job_params, job.handler AS
> job_handler \nFROM job \nWHERE job.state = %(state_1)s AND job.handler =
> %(handler_1)s AND job.id NOT IN (SELECT job.id \nFROM job JOIN
> job_to_input_dataset ON job.id = job_to_input_dataset.job_id JOIN
> history_dataset_association ON history_dataset_association.id =
> job_to_input_dataset.dataset_id JOIN dataset ON dataset.id =
> history_dataset_association.dataset_id \nWHERE job.state = %(state_2)s AND
> (history_dataset_association.state = %(_state_1)s OR
> history_dataset_association.deleted = true OR dataset.state != %(state_3)s
> OR dataset.deleted = true)) AND job.id NOT IN (SELECT job.id \nFROM job
> JOIN job_to_input_library_dataset ON job.id =
> job_to_input_library_dataset.job_id JOIN
> library_dataset_dataset_association ON
> library_dataset_dataset_association.id =
> job_to_input_library_dataset.ldda_id JOIN dataset ON dataset.id =
> library_dataset_dataset_association.dataset_id \nWHERE job.state =
> %(state_4)s AND (library_dataset_dataset_association.state IS NOT NULL OR
> library_dataset_dataset_association.deleted = true OR dataset.state !=
> %(state_5)s OR dataset.deleted = true)) ORDER BY job.id'] [parameters:
> {'state_3': 'ok', 'handler_1': 'main.web.2', 'state_1': 'new', '_state_1':
> 'failed_metadata', 'state_2': 'new', 'state_5': 'ok', 'state_4': 'new'}]
> (Background on this error at: http://sqlalche.me/e/e3q8)
>
___________________________________________________________
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  https://lists.galaxyproject.org/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/

Reply via email to