On Friday, September 26, 2025 at 5:14:41 AM UTC-4 Nikolaus Rath wrote:



On Fri, 19 Sep 2025, at 13:38, '[email protected]' via s3ql wrote:



On Tuesday, September 16, 2025 at 3:21:24 PM UTC-4 Nikolaus Rath wrote:



On Tue, 16 Sep 2025, at 17:36, '[email protected]' via s3ql wrote:

I don't know what the metadata block size is I'm afraid :(  It looks like I 
made this filesystem in Sep 2021 using s3ql 3.7.3 and bash history just 
says mkfs.s3ql  gs://bucketname/...


It's probably the default  from "s3qladm upgrade", 64 kB. 

Best,
-Nikolaus



I think I got this on again - this time I ran mount.s3ql with the --debug 
option, logs are available here:

 https://www.maher.org.uk/~joseph/s3ql/


With "this", do you mean:

2025-09-19 04:18:24.153 2075958:MainThread __main__.excepthook: Uncaught 
top-level exception: + Exception Group Traceback (most recent call last): | 
File "/usr/bin/mount.s3ql", line 33, in <module> | 
sys.exit(load_entry_point('s3ql==5.4.0', 'console_scripts', 
'mount.s3ql')()) | 
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^ | File 
"/usr/lib/python3/dist-packages/s3ql/mount.py", line 170, in main | 
trio.run(main_async, options, stdout_log_handler) | 
~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | File 
"/usr/lib/python3/dist-packages/trio/_core/_run.py", line 2423, in run | 
raise runner.main_task_outcome.error | File 
"/usr/lib/python3/dist-packages/s3ql/mount.py", line 238, in main_async | 
async with trio.open_nursery() as nursery, AsyncExitStack() as cm: | 
~~~~~~~~~~~~~~~~~^^ | File 
"/usr/lib/python3/dist-packages/trio/_core/_run.py", line 1058, in 
__aexit__ | raise combined_error_from_nursery | ExceptionGroup: Exceptions 
from Trio nursery (1 sub-exception) +-+---------------- 1 ---------------- 
| Traceback (most recent call last): | File 
"/usr/lib/python3/dist-packages/s3ql/mount.py", line 336, in main_async | 
raise exc_inst.with_traceback(exc_tb) | File 
"/usr/lib/python3/dist-packages/s3ql/mount.py", line 93, in 
run_with_except_hook | run_old(*args, **kw) | ~~~~~~~^^^^^^^^^^^^^ | File 
"/usr/lib/python3.13/threading.py", line 994, in run | 
self._target(*self._args, **self._kwargs) | 
~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | File 
"/usr/lib/python3/dist-packages/s3ql/block_cache.py", line 325, in 
_upload_loop | self._do_upload(*tmp) | ~~~~~~~~~~~~~~~^^^^^^ | File 
"/usr/lib/python3/dist-packages/s3ql/block_cache.py", line 375, in 
_do_upload | obj_size = backend.write_fh('s3ql_data_%d' % obj_id, el) | 
File "/usr/lib/python3/dist-packages/s3ql/backends/comprenc.py", line 279, 
in write_fh | return self.backend.write_fh(key, fh, meta_raw, len_=len_) | 
~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | File 
"/usr/lib/python3/dist-packages/s3ql/backends/gs.py", line 425, in write_fh 
| return self._write_fh(key, fh, off, len_, metadata or {}) | 
~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | File 
"/usr/lib/python3/dist-packages/s3ql/backends/common.py", line 114, in 
wrapped | return method(*a, **kw) | File 
"/usr/lib/python3/dist-packages/s3ql/backends/gs.py", line 503, in 
_write_fh | exc = _parse_error_response(resp, self.conn.co_readall()) | 
File "/usr/lib/python3/dist-packages/s3ql/backends/gs.py", line 108, in 
_parse_error_response | json_resp = _parse_json_response(resp, body) | File 
"/usr/lib/python3/dist-packages/s3ql/backends/gs.py", line 140, in 
_parse_json_response | body_text = body.decode(charset) | ^^^^^^^^^^^ | 
AttributeError: 'coroutine' object has no attribute 'decode' 
+------------------------------------

If so, that's the same error as in your other email (which is already 
fixed), rather than https://github.com/s3ql/s3ql/issues/392.

Btw, to debug the latter it's enough to run with `--debug-modules 
s3ql.backends,s3ql.http` .



Best,
-Nikolaus


Yes - you are right, this is the one you have already fixed.  I will post 
the debug log if I get a repeat of the one for issue 392.

Joseph 

-- 
You received this message because you are subscribed to the Google Groups 
"s3ql" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion visit 
https://groups.google.com/d/msgid/s3ql/56bbc33b-fa50-42b5-848c-29710404d5e6n%40googlegroups.com.

Reply via email to