Bug#1082027: ITP: webext-folder-account -- Manage folder accounts
- Le 17 Sep 24, à 20:02, Mechtilde Stehmann mechti...@debian.org a écrit : > Package: wnpp > Severity: wishlist > Owner: Mechtilde Stehmann > X-Debbugs-Cc: debian-de...@lists.debian.org, mechti...@debian.org > > * Package name: webext-folder-account > Version : 12.0 > Upstream Contact: Hartmut Welpmann > * URL : https://github.com/Welpy-cw/Folder-Account> > * License : MIT > Programming Lang: Javascript > Description : Manage folder accounts > > Lets you associate user accounts and identities with specific folders. > Great for multiple users who share TB, or for managing mailing lists. > > - Some friends and I use it? > > - I plan to maintain it inside the webext-team
Bug#1078268: [Debian-pan-maintainers] Bug#1078268: pyfai: please make the package build reproducible
> That is true. I don't consider the bug to be the use of a temporary > directory, > but instead the embedding of build paths. I'll do some more investigation to > try to figure out whether that can be resolved within the build toolchain. thanks > It appears my fix here is incomplete or inaccurate, since a subsequent > reprotest build has failed -- but it is possible that Sphinx could change the > setting's default, provided with enough support and reasoning to do so. Good luck > > I'll investigate the cause of that failure and add more details soon.
Bug#1078268: [Debian-pan-maintainers] Bug#1078268: pyfai: please make the package build reproducible
> The meson-python build generates and uses a temporary, randomized build > directory path only when no build-dir setting is already configured. So we > may > be able to resolve this problem by choosing an appropriate static build-dir. why not fixing meson-python directly. This way it is simpler instead of patching all package whcih are using it. > The second non-deterministic output appears to be due to dynamic-evaluation of > a 'threadpoolsize' default argument value for a Python initializer[6], which > is set to the current host's CPU count. This would generally only be relevant > at runtime, except for the fact that the documentation package, built using > the Sphinx 'autodoc' extension, renders the evaluated value in the output HTML > docs. That behaviour can be disabled using the 'autodoc_preserve_defaults'[7] > configuration setting. Idem here, maybe sophinx should have decent default when it comes to reproducible build :). cheers Fred
Bug#1078268: [Debian-pan-maintainers] Bug#1078268: pyfai: please make the package build reproducible
If you are part of the Debian science just push to the repository. It is easyer for me to revert from git, than to remember to check for MR :) thanks a lot for these investigation. it is very appreciable Frederic - Le 9 Aoû 24, à 13:18, James Addison via Debian-pan-maintainers debian-pan-maintain...@alioth-lists.debian.net a écrit : > Source: pyfai > Severity: wishlist > X-Debbugs-Cc: reproducible-b...@lists.alioth.debian.org > User: reproducible-bui...@lists.alioth.debian.org > Usertags: buildpath cpu > > Dear Maintainer, > > I'm an occasional volunteer contributor to the Reproducible Builds[1] project, > and noticed that the pyFAI package failed a recent automated build > reproducibility test[2], as also reported in recent Salsa CI reprotest > results[3]. > > I'm aware of two causes of non-determinism in the build so far: > > > The first cause of non-determinism is the use of temporary directory paths by > the meson-python build process; the resulting paths are embedded in some of > the > extension modules (for example, this[4] CRC32 module) produced in the output > binary package. > > The meson-python build generates and uses a temporary, randomized build > directory path only when no build-dir setting is already configured. So we > may > be able to resolve this problem by choosing an appropriate static build-dir. > > It appears that the Debian numpy package encountered a similar problem and > applied[5] that same style of fix. > > > The second non-deterministic output appears to be due to dynamic-evaluation of > a 'threadpoolsize' default argument value for a Python initializer[6], which > is set to the current host's CPU count. This would generally only be relevant > at runtime, except for the fact that the documentation package, built using > the Sphinx 'autodoc' extension, renders the evaluated value in the output HTML > docs. That behaviour can be disabled using the 'autodoc_preserve_defaults'[7] > configuration setting. > > > I'll prepare patches and corresponding merge requests on Salsa to configure a > fixed build-dir for the meson-python build process, and to configure Sphinx > autodoc to preserve argument default signatures as they appear (unevaluated) > in the source code. > > Regards, > James > > [1] - https://reproducible-builds.org/ > > [2] - > https://tests.reproducible-builds.org/debian/rb-pkg/unstable/amd64/pyfai.html > > [3] - https://salsa.debian.org/science-team/pyfai/-/jobs/6077934 > > [4] - > https://sources.debian.org/src/pyfai/2024.05-3/src/pyFAI/ext/meson.build/#L106-L107 > > [5] - > https://salsa.debian.org/python-team/packages/numpy/-/commit/ece4cae7ceb221783be5ab56b46b1daf196dbc1f > > [6] - > https://sources.debian.org/src/pyfai/2024.05-3/src/pyFAI/multi_geometry.py/?hl=62#L59-L62 > > [7] - > https://www.sphinx-doc.org/en/master/usage/extensions/autodoc.html#confval-autodoc_preserve_defaults > > -- > Debian-pan-maintainers mailing list > debian-pan-maintain...@alioth-lists.debian.net > https://alioth-lists.debian.net/cgi-bin/mailman/listinfo/debian-pan-maintainers
Bug#982601: python3-paraview Conflicts: python3-vtk9
Hello, I am the facet-analyser maintainer/upstream (support) If fact the conflict comes from facet-analyser -> libinsighttoolkit5-dev -> lib-vtk9-dev -> python3-vtk9 facet analyser which is a PV plugins depends also on paraview-dev. facet-analyser -> paraview-dev -> conflict with python3-vtk9 the missing piece is or parview built with the system vtk9 (glop glop) or a libinsighttoolkit5-pvX.YY-dev build against the vtk of paraview... (pas glop) cheers Frederic
Bug#1078077: jupyterlab: Fail to get yarn configuration. node:internal/modules/cjs/loader:1148
the packaging of the next version is done here https://salsa.debian.org/js-team/node-jupyterlab in the branch merge-python-and-node to build it git clone -b merge-python-and-node https://salsa.debian.org/js-team/node-jupyterlab and cd node-jupyterlab gbp buildpackage --git-ignore-branch \ --git-builder="sbuild -s --build-dep-resolver=aspcud \ -j8 --no-apt-update -d unstable -c unstable-amd64-sbuild \ --no-clean-source --no-run-lintian --source-only-changes " \ --git-export=WC and voila
Bug#1078079: [Debian-pan-maintainers] Jupyterlab status ?
Running the upstream script I have this $ bash ./ci_script.sh + set -o pipefail + export YARN_ENABLE_GLOBAL_CACHE=1 + YARN_ENABLE_GLOBAL_CACHE=1 + export YARN_ENABLE_INLINE_BUILDS=1 + YARN_ENABLE_INLINE_BUILDS=1 + [[ '' != nonode ]] + python -c 'from jupyterlab.commands import build_check; build_check()' Fail to get yarn configuration. node:internal/modules/cjs/loader:1148 throw err; ^ Error: Cannot find module '/usr/share/nodejs/corepack/dist/yarn.js' at Module._resolveFilename (node:internal/modules/cjs/loader:1145:15) at Module._load (node:internal/modules/cjs/loader:986:27) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:174:12) at node:internal/main/run_main_module:28:49 { code: 'MODULE_NOT_FOUND', requireStack: [] } Node.js v20.15.1 + [[ '' == python ]] + [[ '' == js* ]] + [[ '' == docs ]] + [[ '' == integrity ]] + [[ '' == lint ]] + [[ '' == integrity2 ]] + [[ '' == integrity3 ]] + [[ '' == release_test ]] + [[ '' == examples ]] + [[ '' == usage ]] + [[ '' == usage2 ]] + [[ '' == splice_source ]] + [[ '' == interop ]] + [[ '' == nonode ]] It is possible to run the python test by starting the ci_scripts.sh like this GROUP=python bash ./ci_script.hs
Bug#1039087: removing embeded version of yajl
Here the upstream point of view about the CVE. https://github.com/epics-base/epics-base/issues/405 check with the security team, if their analyse is ok ? Fred
Bug#1039087: Info received (removing embeded version of yajl)
Here the diff between the epics version (debian patch unapplyed) and the current 2.1.0 version of yajl (debian patch unapplyed). not that simple...diff --git a/src/yajl.c b/src/yajl.c index d477893..fdad3f6 100644 --- a/src/yajl.c +++ b/src/yajl.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2007-2014, Lloyd Hilaiel + * Copyright (c) 2007-2011, Lloyd Hilaiel * * Permission to use, copy, modify, and/or distribute this software for any * purpose with or without fee is hereby granted, provided that the above @@ -14,16 +14,16 @@ * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. */ -#include "api/yajl_parse.h" -#include "yajl_lex.h" -#include "yajl_parser.h" -#include "yajl_alloc.h" - #include #include #include #include +#include "yajl_parse.h" +#include "yajl_lex.h" +#include "yajl_parser.h" +#include "yajl_alloc.h" + const char * yajl_status_to_string(yajl_status stat) { @@ -62,16 +62,19 @@ yajl_alloc(const yajl_callbacks * callbacks, } hand = (yajl_handle) YA_MALLOC(afs, sizeof(struct yajl_handle_t)); +if (hand == NULL) { +return NULL; +} /* copy in pointers to allocation routines */ memcpy((void *) &(hand->alloc), (void *) afs, sizeof(yajl_alloc_funcs)); hand->callbacks = callbacks; hand->ctx = ctx; -hand->lexer = NULL; +hand->lexer = NULL; hand->bytesConsumed = 0; hand->decodeBuf = yajl_buf_alloc(&(hand->alloc)); -hand->flags = 0; +hand->flags = yajl_allow_json5 | yajl_allow_comments; yajl_bs_init(hand->stateStack, &(hand->alloc)); yajl_bs_push(hand->stateStack, yajl_state_start); @@ -79,13 +82,16 @@ yajl_alloc(const yajl_callbacks * callbacks, } int -yajl_config(yajl_handle h, yajl_option opt, ...) +yajl_config(yajl_handle h, int option, ...) { +yajl_option opt = option; /* UB to use an enum in va_start */ int rv = 1; va_list ap; -va_start(ap, opt); +va_start(ap, option); switch(opt) { +case yajl_allow_json5: +opt |= yajl_allow_comments; /* JSON5 allows comments */ case yajl_allow_comments: case yajl_dont_validate_strings: case yajl_allow_trailing_garbage: @@ -124,7 +130,11 @@ yajl_parse(yajl_handle hand, const unsigned char * jsonText, if (hand->lexer == NULL) { hand->lexer = yajl_lex_alloc(&(hand->alloc), hand->flags & yajl_allow_comments, - !(hand->flags & yajl_dont_validate_strings)); + !(hand->flags & yajl_dont_validate_strings), + hand->flags & yajl_allow_json5); +} +if (hand->lexer == NULL) { +return yajl_status_error; } status = yajl_do_parse(hand, jsonText, jsonTextLen); @@ -144,7 +154,8 @@ yajl_complete_parse(yajl_handle hand) if (hand->lexer == NULL) { hand->lexer = yajl_lex_alloc(&(hand->alloc), hand->flags & yajl_allow_comments, - !(hand->flags & yajl_dont_validate_strings)); + !(hand->flags & yajl_dont_validate_strings), + hand->flags & yajl_allow_json5); } return yajl_do_finish(hand); diff --git a/src/yajl_alloc.c b/src/yajl_alloc.c index 96ad1d3..2388814 100644 --- a/src/yajl_alloc.c +++ b/src/yajl_alloc.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2007-2014, Lloyd Hilaiel + * Copyright (c) 2007-2011, Lloyd Hilaiel * * Permission to use, copy, modify, and/or distribute this software for any * purpose with or without fee is hereby granted, provided that the above @@ -20,25 +20,23 @@ * free */ -#include "yajl_alloc.h" #include +#include "yajl_alloc.h" + static void * yajl_internal_malloc(void *ctx, size_t sz) { -(void)ctx; return malloc(sz); } static void * yajl_internal_realloc(void *ctx, void * previous, size_t sz) { -(void)ctx; return realloc(previous, sz); } static void yajl_internal_free(void *ctx, void * ptr) { -(void)ctx; free(ptr); } diff --git a/src/yajl_alloc.h b/src/yajl_alloc.h index 203c2f9..406310a 100644 --- a/src/yajl_alloc.h +++ b/src/yajl_alloc.h @@ -1,5 +1,5 @@ /* - * Copyright (c) 2007-2014, Lloyd Hilaiel + * Copyright (c) 2007-2011, Lloyd Hilaiel * * Permission to use, copy, modify, and/or distribute this software for any * purpose with or without fee is hereby granted, provided that the above @@ -16,19 +16,31 @@ /** * \file yajl_alloc.h - * default memory allocation routines for yajl which use malloc/realloc and - * free + * \brief Memory allocation macros for yajl + * \author Lloyd Hilaiel + * + * These macros are used inside YAJL instead of directly calling + * malloc(), realloc() or free(). They call the equivalent method + * in their \ref yajl_alloc_funcs parameter \a afs. */ #ifndef __YAJ
Bug#1039087: removing embeded version of yajl
following a different path..., I added this in the rules file -export POSIX_CFLAGS+=$(CFLAGS) +export POSIX_CFLAGS+=$(CFLAGS) $(shell pkgconf --cflags yajl) export POSIX_CFLAGS+=$(CPPFLAGS) export POSIX_CPPFLAGS+=$(CPPFLAGS) -export POSIX_LDFLAGS+=$(LDFLAGS) +export POSIX_LDFLAGS+=$(LDFLAGS) $(shell pkgconf --libs yajl) but it ends up like this /usr/bin/gcc -D_GNU_SOURCE -D_DEFAULT_SOURCE -D_X86_64_ -DUNIX -Dlinux -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/<>=. -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/yajl -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/<>=. -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/yajl -Wdate-time -D_FORTIFY_SOURCE=2 -O3 -g -Wall -Werror-implicit-function-declaration -mtune=generic -m64 -I. -I../O.Common -I. -I. -I.. -I../../../../include/compiler/gcc -I../../../../include/os/Linux -I../../../../include -c ../yajl_test.c ../yajl_test.c: In function ‘main’: ../yajl_test.c:209:23: error: ‘yajl_allow_json5’ undeclared (first use in this function) 209 | yajl_config(hand, yajl_allow_json5, 0); | ^~~~ ../yajl_test.c:209:23: note: each undeclared identifier is reported only once for each function it appears in the embeded version seems to provide at least this extra symbols. what about integrating this symbol in the yajl library ?
Bug#1063631: [Debian-pan-maintainers] Bug#1063631: closing 1063631
I am working on it at the upstream level need a few more days. Cheers Fred
Bug#1069144: ADvice about htis pocl issue on armel
Hello Andreas, I have another autopkgtest failure on armel with silx and pocl The content of check_atomic32 is def check_atomic32(device): try: ctx = pyopencl.Context(devices=[device]) except: return False, f"Unable to create context on {device}" else: queue = pyopencl.CommandQueue(ctx) src = """ kernel void check_atomic32(global int* ary){ int res = atom_inc(ary); } """ try: prg = pyopencl.Program(ctx, src).build() except Exception as err: return False, f"{type(err)}: {err}" a = numpy.zeros(1, numpy.int32) d = cla.to_device(queue, a) prg.check_atomic32(queue, (1024,), (32,), d.data).wait() value = d.get()[0] return value == 1024, f"Got the proper value 1024=={value}" 1019s Testing with python3.12: 1024s pocl warning: encountered incomplete implementation in ./lib/CL/clGetDeviceInfo.c:98 1026s dlopen("/tmp/autopkgtest-lxc.1tg98r_u/downtmp/autopkgtest_tmp/.cache/pocl/kcache/GI/PODCELNLNPFCGBFNGLEIHDEIECBKNFCDBNAPP/check_atomic32/0-0-0/check_atomic32.so") failed with '/tmp/autopkgtest-lxc.1tg98r_u/downtmp/autopkgtest_tmp/.cache/pocl/kcache/GI/PODCELNLNPFCGBFNGLEIHDEIECBKNFCDBNAPP/check_atomic32/0-0-0/check_atomic32.so: undefined symbol: __atomic_fetch_add_4'. 1026s note: missing symbols in the kernel binary might be reported as 'file not found' errors. I found something similar about scipy here https://github.com/scipy/scipy/issues/19982 I am wondering if this is an issue with pocl or llvm on armel. Fred
Bug#1060318: Could be related to LLVM rather than PoCL: works with LLVM-15
bravo !!! This is team works. :)) Cheers Frederic
Bug#1065724: epics-base: FTBFS on amd64: Tests failed
Here an analyse of the FTBFS On the amd64, I have two failures dureing the test Test Summary Report --- testPVAServer.t(Wstat: 0 Tests: 0 Failed: 0) Parse errors: No plan found in TAP output Files=6, Tests=129, 1 wallclock secs ( 0.05 usr 0.01 sys + 0.09 cusr 0.06 csys = 0.21 CPU) Result: FAIL --- and Test Summary Report --- testInetAddressUtils.t (Wstat: 0 Tests: 65 Failed: 0) TODO passed: 64 testChannelAccess.t (Wstat: 0 Tests: 152 Failed: 0) TODO passed: 45 testServerContext.t (Wstat: 0 Tests: 0 Failed: 0) Parse errors: No plan found in TAP output Files=12, Tests=6381, 27 wallclock secs ( 0.38 usr 0.03 sys + 0.42 cusr 0.18 csys = 1.01 CPU) Result: FAIL --- on arm64, the first test seems to work... testPVAServer.t .. 1..1 pvAccess Server v7.1.7 Active configuration (w/ defaults) EPICS_PVAS_INTF_ADDR_LIST = 0.0.0.0:5075 EPICS_PVAS_BEACON_ADDR_LIST = EPICS_PVAS_AUTO_BEACON_ADDR_LIST = YES EPICS_PVAS_BEACON_PERIOD = 15 EPICS_PVAS_BROADCAST_PORT = 5076 EPICS_PVAS_SERVER_PORT = 5075 EPICS_PVAS_PROVIDER_NAMES = local ok 1 - ctx.get()!=0 ok I am wondering if this is not something related to the network configuration. i386 Test Summary Report --- printfTest.t (Wstat: 256 (exited 1) Tests: 97 Failed: 1) Failed test: 70 Non-zero exit status: 1 Files=22, Tests=5033, 29 wallclock secs ( 0.42 usr 0.04 sys + 1.74 cusr 0.34 csys = 2.54 CPU) Result: FAIL --- Test Summary Report --- testInetAddressUtils.t (Wstat: 0 Tests: 65 Failed: 0) TODO passed: 64 testChannelAccess.t (Wstat: 0 Tests: 152 Failed: 0) TODO passed: 45 testServerContext.t (Wstat: 0 Tests: 0 Failed: 0) Parse errors: No plan found in TAP output Files=12, Tests=6381, 26 wallclock secs ( 0.42 usr 0.02 sys + 0.46 cusr 0.20 csys = 1.10 CPU) Result: FAIL --- Test Summary Report --- testPVAServer.t(Wstat: 0 Tests: 0 Failed: 0) Parse errors: No plan found in TAP output Files=6, Tests=129, 0 wallclock secs ( 0.08 usr 0.02 sys + 0.10 cusr 0.04 csys = 0.24 CPU) Result: FAIL --- And there is a bunch of unsupported architectures. /<>/src/tools/EpicsHostArch.pl: Architecture 'mips64el-linux-gnuabi64-thread-multi' not recognized /<>/src/tools/EpicsHostArch.pl: Architecture 'powerpc64le-linux-gnu-thread-multi' not recognized /<>/src/tools/EpicsHostArch.pl: Architecture 'riscv64-linux-gnu-thread-multi' not recognized /<>/src/tools/EpicsHostArch.pl: Architecture 's390x-linux-gnu-thread-multi' not recognized
Bug#1060318: Info received (Bug#1060318: Info received (Bug#1060318: Info received (Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12))))
A workaround for now is to use this POCL_WORK_GROUP_METHOD=cbs Jerome is helping also here trying to understand the problem... https://github.com/silx-kit/silx/issues/4073
Bug#1060318: Info received (Bug#1060318: Info received (Bug#1060318: Info received (Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12))))
POCL_WORK_GROUP_METHOD=cbs python3 test.py make it works $ POCL_WORK_GROUP_METHOD=cbs python3 test.py [SubCFG] Form SubCFGs in bsort_all [SubCFG] Form SubCFGs in bsort_horizontal [SubCFG] Form SubCFGs in bsort_vertical [SubCFG] Form SubCFGs in bsort_book [SubCFG] Form SubCFGs in bsort_file [SubCFG] Form SubCFGs in medfilt2d [SubCFG] Form SubCFGs in medfilt2d
Bug#1060318: Info received (Bug#1060318: Info received (Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12)))
With latest version (PAS OK) $ dpkg -l | grep pocl ii libpocl2-common5.0-2.1 all common files for the pocl library ii libpocl2t64:amd64 5.0-2.1 amd64Portable Computing Language library ii pocl-opencl-icd:amd64 5.0-2.1 amd64pocl ICD
Bug#1060318: Info received (Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12))
Debian12 (OK) $ dpkg -l | grep pocl ii libpocl2:amd64 3.1-3+deb12u1 amd64Portable Computing Language library ii libpocl2-common 3.1-3+deb12u1 all common files for the pocl library ii pocl-opencl-icd:amd643.1-3+deb12u1 amd64pocl ICD unstable (NOT OK) $ dpkg -l | grep pocl ii libpocl2:amd64 5.0-2 amd64Portable Computing Language library ii libpocl2-common5.0-2 all common files for the pocl library ii pocl-opencl-icd:amd64 5.0-2 amd64pocl ICD
Bug#1060318: Info received (Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12))
On Debian12 it works out of the box $ POCL_DEBUG=1 python3 test.py [2024-03-11 10:05:31.837738936]POCL: in fn pocl_install_sigfpe_handler at line 229: | GENERAL | Installing SIGFPE handler... [2024-03-11 10:05:31.868890390]POCL: in fn POclCreateCommandQueue at line 98: | GENERAL | Created Command Queue 3 (0x1ee13c0) on device 0 [2024-03-11 10:05:31.868917030]POCL: in fn POclCreateContext at line 227: | GENERAL | Created Context 2 (0x1ee0e40) [2024-03-11 10:05:31.868966549]POCL: in fn POclCreateCommandQueue at line 98: | GENERAL | Created Command Queue 4 (0x1f31f10) on device 0 [2024-03-11 10:05:31.874596495]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel s8_to_float (0x1fc5540) [2024-03-11 10:05:31.874606285]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel u8_to_float (0x1fc5610) [2024-03-11 10:05:31.874617005]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel s16_to_float (0x1fc5730) [2024-03-11 10:05:31.874622275]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel u16_to_float (0x1f81e70) [2024-03-11 10:05:31.874632075]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel u32_to_float (0x1f81fb0) [2024-03-11 10:05:31.874638955]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel s32_to_float (0x1f820f0) [2024-03-11 10:05:31.874646635]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections (0x1f82230) [2024-03-11 10:05:31.874654714]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections2 (0x1f82590) [2024-03-11 10:05:31.874663744]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections3Poisson (0x1f82990) [2024-03-11 10:05:31.874669284]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections3 (0x1f82d90) [2024-03-11 10:05:31.874673814]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_all (0x201ded0) [2024-03-11 10:05:31.874681154]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_horizontal (0x201e010) [2024-03-11 10:05:31.874685604]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_vertical (0x201e150) [2024-03-11 10:05:31.874691454]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_book (0x201e290) [2024-03-11 10:05:31.874699564]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_file (0x201e3d0) [2024-03-11 10:05:31.874709654]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel medfilt2d (0x201e510) [2024-03-11 10:05:31.877001426]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel s8_to_float (0x1fdf150) [2024-03-11 10:05:31.877011365]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel u8_to_float (0x20103f0) [2024-03-11 10:05:31.877019735]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel s16_to_float (0x1f60f90) [2024-03-11 10:05:31.877025545]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel u16_to_float (0x1f61060) [2024-03-11 10:05:31.877030655]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel u32_to_float (0x1f5f1f0) [2024-03-11 10:05:31.877038395]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel s32_to_float (0x1f5f310) [2024-03-11 10:05:31.877043475]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections (0x1f60500) [2024-03-11 10:05:31.877055965]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections2 (0x200efa0) [2024-03-11 10:05:31.877061275]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections3Poisson (0x200f3a0) [2024-03-11 10:05:31.877064514]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel corrections3 (0x200f7a0) [2024-03-11 10:05:31.877071304]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_all (0x200fc20) [2024-03-11 10:05:31.877079984]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_horizontal (0x200fd60) [2024-03-11 10:05:31.877087744]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_vertical (0x1f613b0) [2024-03-11 10:05:31.877094244]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_book (0x1f614f0) [2024-03-11 10:05:31.877098614]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel bsort_file (0x1f61630) [2024-03-11 10:05:31.877102884]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel medfilt2d (0x1f61770) [2024-03-11 10:05:31.877723934]POCL: in fn POclCreateKernel at line 138: | GENERAL | Created Kernel medfilt2d (0x1f61e00) [2024-03-11 10:05:31.878064028]POCL: in fn POclSetKernelArg at line 107: | GENERAL | Kernel me
Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12)
We already had the warning message [2024-03-10 14:26:18.189651850]POCL: in fn void appendToProgramBuildLog(cl_program, unsigned int, std::string&) at line 111: | ERROR | warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:861:14: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:893:14: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:933:16: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:1266:26: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI [2024-03-10 14:26:18.195708864]POCL: in fn llvm::Module* getKernelLibrary(cl_device_id, PoclLLVMContextData*) at line 992: | LLVM | Using /lib/x86_64-linux-gnu/../../share/pocl/kernel-x86_64-pc-linux-gnu-sse41.bc as the built-in lib. [2024-03-10 14:26:20.314065808]POCL: in fn int pocl_llvm_build_program(cl_program, unsigned int, cl_uint, _cl_program* const*, const char**, int) at line 756: | LLVM | Writing program.bc to /home/picca/.cache/pocl/kcache/LK/MONFDAKCFIMDEBOPEIHEOILBLCLBMGGNLPDID/program.bc. /usr/lib/python3/dist-packages/pyopencl/cache.py:417: CompilerWarning: From-source build succeeded, but resulted in non-empty logs: Build on succeeded, but said: warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:861:14: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:893:14: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:933:16: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_msXjLw.cl:1266:26: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI prg.build(options_bytes, [devices[i] for i in to_be_built_indices])
Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12)
Here a log with POCL_DEBUG=all picca@cush:/tmp$ python3 test.py [2024-03-10 14:22:19.462191847]POCL: in fn pocl_install_sigfpe_handler at line 265: | GENERAL | Installing SIGFPE handler... [2024-03-10 14:22:19.475550217]POCL: in fn POclCreateCommandQueue at line 103: | GENERAL | Created Command Queue 3 (0x27d55b0) on device 0 [2024-03-10 14:22:19.475690904]POCL: in fn void pocl_llvm_create_context(cl_context) at line 592: | LLVM | Created context 2 (0x27d4960) [2024-03-10 14:22:19.475732695]POCL: in fn POclCreateContext at line 232: | GENERAL | Created Context 2 (0x27d4960) [2024-03-10 14:22:19.475822461]POCL: in fn POclRetainContext at line 32: | REFCOUNTS | Retain Context 2 (0x27d4960), Refcount: 2 [2024-03-10 14:22:19.475856682]POCL: in fn POclCreateCommandQueue at line 103: | GENERAL | Created Command Queue 4 (0x27d77b0) on device 0 [2024-03-10 14:22:19.492655607]POCL: in fn POclRetainContext at line 32: | REFCOUNTS | Retain Context 2 (0x27d4960), Refcount: 3 [2024-03-10 14:22:19.492795776]POCL: in fn compile_and_link_program at line 718: | LLVM | building program with options -I /usr/lib/python3/dist-packages/pyopencl/cl [2024-03-10 14:22:19.492824004]POCL: in fn compile_and_link_program at line 755: | LLVM | building program for 1 devs with options -I /usr/lib/python3/dist-packages/pyopencl/cl [2024-03-10 14:22:19.492847621]POCL: in fn compile_and_link_program at line 759: | LLVM | BUILDING for device: cpu [2024-03-10 14:22:19.497354940]POCL: in fn POclRetainContext at line 32: | REFCOUNTS | Retain Context 2 (0x27d4960), Refcount: 4 [2024-03-10 14:22:19.497919687]POCL: in fn POclRetainContext at line 32: | REFCOUNTS | Retain Context 2 (0x27d4960), Refcount: 5 [2024-03-10 14:22:19.497963205]POCL: in fn POclCreateBuffer at line 292: |MEMORY | Created Buffer 6 (0x2801b90), MEM_HOST_PTR: (nil), device_ptrs[0]: (nil), SIZE 4, FLAGS 1 [2024-03-10 14:22:19.498110078]POCL: in fn pocl_driver_alloc_mem_obj at line 428: |MEMORY | Basic device ALLOC 0x27f7380 / size 4 [2024-03-10 14:22:19.498162446]POCL: in fn POclRetainCommandQueue at line 35: | REFCOUNTS | Retain Command Queue 4 (0x27d77b0), Refcount: 2 [2024-03-10 14:22:19.498187844]POCL: in fn pocl_create_event at line 527: |EVENTS | Created event 1 (0x27e4e60) Command write_buffer [2024-03-10 14:22:19.498211789]POCL: in fn pocl_create_command_struct at line 648: |EVENTS | event pointer provided [2024-03-10 14:22:19.498232543]POCL: in fn pocl_create_command_struct at line 668: |EVENTS | Created immediate command struct: CMD 0x27e4d50 (event 1 / 0x27e4e60, type: write_buffer) [2024-03-10 14:22:19.498259772]POCL: in fn pocl_command_enqueue at line 1290: |EVENTS | In-order Q; adding event syncs [2024-03-10 14:22:19.498280767]POCL: in fn pocl_command_enqueue at line 1335: |EVENTS | Pushed Event 1 to CQ 4. [2024-03-10 14:22:19.498303076]POCL: in fn pocl_update_event_queued at line 2177: |EVENTS | Event queued: 1 [2024-03-10 14:22:19.498326609]POCL: in fn pocl_update_event_submitted at line 2197: |EVENTS | Event submitted: 1 [2024-03-10 14:22:19.498451579]POCL: in fn pocl_update_event_running_unlocked at line 2216: |EVENTS | Event running: 1 [2024-03-10 14:22:19.498484119]POCL: in fn pocl_update_event_finished at line 2368: |EVENTS | cpu: Command complete, event 1 [2024-03-10 14:22:19.498509038]POCL: in fn pocl_exec_command at line 343: |TIMING | >>>32.497 usEvent Write Buffer [2024-03-10 14:22:19.498531904]POCL: in fn POclReleaseMemObject at line 53: | REFCOUNTS | Release Memory Object 6 (0x2801b90), Refcount: 1 [2024-03-10 14:22:19.498562333]POCL: in fn POclReleaseEvent at line 39: | REFCOUNTS | Release Event 1 (0x27e4e60), Refcount: 2 [2024-03-10 14:22:19.498656679]POCL: in fn POclCreateKernel at line 133: | GENERAL | Created Kernel check_atomic32 (0x27f74c0) [2024-03-10 14:22:19.501056049]POCL: in fn POclRetainContext at line 32: | REFCOUNTS | Retain Context 2 (0x27d4960), Refcount: 6 [2024-03-10 14:22:19.501139297]POCL: in fn POclReleaseContext at line 53: | REFCOUNTS | Release Context 2 (0x27d4960), Refcount: 5 [2024-03-10 14:22:19.503196833]POCL: in fn POclSetKernelArg at line 107: | GENERAL | Kernel check_atomic32 || SetArg idx 0 || int* || Local 0 || Size 8 || Value 0x7fff7b47ae20 || Pointer 0x2801b90 || *(uint32*)Value:0 || *(uint64*)Value:0 || Hex Value: 901B8002 [2024-03-10 14:22:19.503275428]POCL: in fn pocl_kernel_calc_wg_size at line 182: | GENERAL | Preparing kernel check_atomic32 with local size 32 x 1 x 1 group sizes 32 x 1 x 1... [2024-03-10 14:22:19.503311773]POCL: in fn POclRetainCommandQueue at line 35: | REFCOUNTS | Retain Command Queue 4 (0x27d77b0), Refcount: 3 [2024-03-10 14:22:19.503350256]POCL: in fn pocl_create_event at line 527: |EVENTS | Created event 2 (0x27f8c
Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12)
It seems that here is an error here [2024-03-10 14:22:19.550588408]POCL: in fn int pocl_llvm_build_program(cl_program, unsigned int, cl_uint, _cl_program* const*, const char**, int) at line 420: | LLVM | all build options: -Dcl_khr_int64 -DPOCL_DEVICE_ADDRESS_BITS=64 -D__USE_CLANG_OPENCL_C_H -xcl -Dinline= -I. -cl-kernel-arg-info -opaque-pointers -D NIMAGE=1 -I /usr/lib/python3/dist-packages/pyopencl/cl -D__ENDIAN_LITTLE__=1 -D__IMAGE_SUPPORT__=1 -DCL_DEVICE_MAX_GLOBAL_VARIABLE_SIZE=64000 -D__OPENCL_VERSION__=300 -cl-std=CL3.0 -D__OPENCL_C_VERSION__=300 -Dcl_khr_byte_addressable_store=1 -Dcl_khr_global_int32_base_atomics=1 -Dcl_khr_global_int32_extended_atomics=1 -Dcl_khr_local_int32_base_atomics=1 -Dcl_khr_local_int32_extended_atomics=1 -Dcl_khr_3d_image_writes=1 -Dcl_khr_command_buffer=1 -Dcl_pocl_pinned_buffers=1 -Dcl_khr_subgroups=1 -Dcl_intel_unified_shared_memory=1 -Dcl_khr_subgroup_ballot=1 -Dcl_khr_subgroup_shuffle=1 -Dcl_intel_subgroups=1 -Dcl_intel_required_subgroup_size=1 -Dcl_ext_float_atomics=1 -Dcl_khr_spir=1 -Dcl_khr_fp64=1 -Dcl_khr_int64_base_atomics=1 -Dcl_khr_int64_extended_atomics=1 -D__opencl_c_3d_image_writes=1 -D__opencl_c_images=1 -D__opencl_c_atomic_order_acq_rel=1 -D__opencl_c_atomic_order_seq_cst=1 -D__opencl_c_atomic_scope_device=1 -D__opencl_c_program_scope_global_variables=1 -D__opencl_c_generic_address_space=1 -D__opencl_c_subgroups=1 -D__opencl_c_atomic_scope_all_devices=1 -D__opencl_c_read_write_images=1 -D__opencl_c_fp64=1 -D__opencl_c_ext_fp32_global_atomic_add=1 -D__opencl_c_ext_fp32_local_atomic_add=1 -D__opencl_c_ext_fp32_global_atomic_min_max=1 -D__opencl_c_ext_fp32_local_atomic_min_max=1 -D__opencl_c_ext_fp64_global_atomic_add=1 -D__opencl_c_ext_fp64_local_atomic_add=1 -D__opencl_c_ext_fp64_global_atomic_min_max=1 -D__opencl_c_ext_fp64_local_atomic_min_max=1 -D__opencl_c_int64=1 -cl-ext=-all,+cl_khr_byte_addressable_store,+cl_khr_global_int32_base_atomics,+cl_khr_global_int32_extended_atomics,+cl_khr_local_int32_base_atomics,+cl_khr_local_int32_extended_atomics,+cl_khr_3d_image_writes,+cl_khr_command_buffer,+cl_pocl_pinned_buffers,+cl_khr_subgroups,+cl_intel_unified_shared_memory,+cl_khr_subgroup_ballot,+cl_khr_subgroup_shuffle,+cl_intel_subgroups,+cl_intel_required_subgroup_size,+cl_ext_float_atomics,+cl_khr_spir,+cl_khr_fp64,+cl_khr_int64_base_atomics,+cl_khr_int64_extended_atomics,+__opencl_c_3d_image_writes,+__opencl_c_images,+__opencl_c_atomic_order_acq_rel,+__opencl_c_atomic_order_seq_cst,+__opencl_c_atomic_scope_device,+__opencl_c_program_scope_global_variables,+__opencl_c_generic_address_space,+__opencl_c_subgroups,+__opencl_c_atomic_scope_all_devices,+__opencl_c_read_write_images,+__opencl_c_fp64,+__opencl_c_ext_fp32_global_atomic_add,+__opencl_c_ext_fp32_local_atomic_add,+__opencl_c_ext_fp32_global_atomic_min_max,+__opencl_c_ext_fp32_local_atomic_min_max,+__opencl_c_ext_fp64_global_atomic_add,+__opencl_c_ext_fp64_local_atomic_add,+__opencl_c_ext_fp64_global_atomic_min_max,+__opencl_c_ext_fp64_local_atomic_min_max,+__opencl_c_int64 -fno-builtin -triple=x86_64-pc-linux-gnu -target-cpu penryn 4 warnings generated. [2024-03-10 14:22:20.986369997]POCL: in fn void appendToProgramBuildLog(cl_program, unsigned int, std::string&) at line 111: | ERROR | warning: /home/picca/.cache/pocl/kcache/tempfile_NcEztR.cl:861:14: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_NcEztR.cl:893:14: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_NcEztR.cl:933:16: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI warning: /home/picca/.cache/pocl/kcache/tempfile_NcEztR.cl:1266:26: AVX vector argument of type '__private float8' (vector of 8 'float' values) without 'avx' enabled changes the ABI [2024-03-10 14:22:20.992890946]POCL: in fn llvm::Module* getKernelLibrary(cl_device_id, PoclLLVMContextData*) at line 992: | LLVM | Using /lib/x86_64-linux-gnu/../../share/pocl/kernel-x86_64-pc-linux-gnu-sse41.bc as the built-in lib. [2024-03-10 14:22:23.151001890]POCL: in fn int pocl_llvm_build_program(cl_program, unsigned int, cl_uint, _cl_program* const*, const char**, int) at line 756: | LLVM | Writing program.bc to /home/picca/.cache/pocl/kcache/OO/KDMNEJOLAKKIBKBOIDNJJPAEHMJELJCBLMGBG/program.bc. /usr/lib/python3/dist-packages/pyopencl/cache.py:417: CompilerWarning: Non-empty compiler output encountered. Set the environment variable PYOPENCL_COMPILER_OUTPUT=1 to see more. prg.build(options_bytes, [devices[i] for i in to_be_built_indices]) let export the PYOPENCL_COMPILER_OUTPUT
Bug#1060318: Info received (silx: autopkgtest failure with Python 3.12)
Here a small script which trigger the errorfrom silx.image import medianfilter import numpy IMG = numpy.arange(1.0).reshape(100, 100) KERNEL = (1, 1) res = medianfilter.medfilt2d( image=IMG, kernel_size=KERNEL, engine="opencl", )
Bug#1060318: silx: autopkgtest failure with Python 3.12
In order to reproduce the bug, install python3-silx 2.0.0+dfsg-1 python3-pytest-xvfb pocl-opencl-icd then $ pytest --pyargs silx.image.test.test_medianfilter -v === test session starts === platform linux -- Python 3.11.8, pytest-8.0.2, pluggy-1.4.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /home/picca/debian/science-team/pyvkfft plugins: anyio-4.2.0, dials-data-2.4.0, xvfb-3.0.0 collected 2 items ::TestMedianFilterEngines::testCppMedFilt2d PASSED [ 50%] ::TestMedianFilterEngines::testOpenCLMedFilt2d Abandon the OpenCL test fails
Bug#1060318: silx: autopkgtest failure with Python 3.12
With the silx 2.0.0 version the failire is located in the OpenCL part the backtrace is this one when running the median filter # build the packag eintht echroot and enter into it once build dgit --gbp sbuild --finished-build-commands '%SBUILD_SHELL' run this command to obtain the backtrace... DEBUGINFOD_URLS="https://debuginfod.debian.net"; PYTHONPATH=. gdb --args python3.11 -m pytest --pyarg silx silx/image/test/test_medianfilter.py here the backtrace. Thread 1 "python3.11" received signal SIGABRT, Aborted. 0x77d3516c in ?? () from /lib/x86_64-linux-gnu/libc.so.6 (gdb) bt #0 0x77d3516c in ?? () from /lib/x86_64-linux-gnu/libc.so.6 #1 0x77ce7472 in raise () from /lib/x86_64-linux-gnu/libc.so.6 #2 0x77cd14b2 in abort () from /lib/x86_64-linux-gnu/libc.so.6 #3 0x77cd13d5 in ?? () from /lib/x86_64-linux-gnu/libc.so.6 #4 0x77ce03a2 in __assert_fail () from /lib/x86_64-linux-gnu/libc.so.6 #5 0x7344770d in pocl::Kernel::createParallelRegionBefore (this=, B=0x8980180) at ../llvmopencl/./lib/llvmopencl/Kernel.cc:129 #6 pocl::Kernel::getParallelRegions (this=, LI=..., ParallelRegions=0x7fff4860) at ../llvmopencl/./lib/llvmopencl/Kernel.cc:193 #7 0x7346ab82 in pocl::WorkitemLoopsImpl::processFunction (this=this@entry=0x7fff47c0, F=...) at ../llvmopencl/./lib/llvmopencl/WorkitemLoops.cc:445 #8 0x7346cb8d in pocl::WorkitemLoopsImpl::runOnFunction (this=0x7fff47c0, F=...) at ../llvmopencl/./lib/llvmopencl/WorkitemLoops.cc:183 #9 0x7346ecac in pocl::WorkitemLoops::run (this=, F=..., AM=...) at ../llvmopencl/./lib/llvmopencl/WorkitemLoops.cc:1490 #10 0x7346ede5 in llvm::detail::PassModel>::run(llvm::Function&, llvm::AnalysisManager&) (this=, IR=..., AM=...) at /usr/lib/llvm-16/include/llvm/IR/PassManagerInternal.h:89 #11 0x7fffe91d7579 in run () at llvm/include/llvm/IR/PassManager.h:517 #12 0x7fffeaeedb01 in llvm::detail::PassModel>, llvm::PreservedAnalyses, llvm::AnalysisManager>::run(llvm::Function&, llvm::AnalysisManager&) () at llvm/include/llvm/IR/PassManagerInternal.h:89 #13 0x7fffe91dade6 in run () at build-llvm/tools/clang/stage2-bins/llvm/lib/IR/PassManager.cpp:124 #14 0x7fffeaeed921 in llvm::detail::PassModel>::run(llvm::Module&, llvm::AnalysisManager&) () at llvm/include/llvm/IR/PassManagerInternal.h:89 #15 0x7348cd85 in llvm::PassManager>::run(llvm::Module&, llvm::AnalysisManager&) (AM=..., IR=..., this=0x7fff50b8) at /usr/include/c++/13/bits/unique_ptr.h:199 #16 PoCLModulePassManager::run (this=0x7fff4f98, Bitcode=...) at ./lib/CL/pocl_llvm_wg.cc:322 #17 0x73494b14 in TwoStagePoCLModulePassManager::run (Bitcode=..., this=0x7fff4e10) at ./lib/CL/pocl_llvm_wg.cc:386 #18 runKernelCompilerPasses (Device=Device@entry=0x151f050, Mod=...) at ./lib/CL/pocl_llvm_wg.cc:727 #19 0x73496302 in pocl_llvm_run_pocl_passes(llvm::Module*, _cl_command_run*, llvm::LLVMContext*, PoclLLVMContextData*, _cl_kernel*, _cl_device_id*, int) [clone .isra.0] ( Bitcode=Bitcode@entry=0x16d6ef0, RunCommand=RunCommand@entry=0x7fffbd40, PoclCtx=PoclCtx@entry=0x157e340, Kernel=Kernel@entry=0x7fffbcb0, Device=Device@entry=0x151f050, Specialize=Specialize@entry=0, LLVMContext=) at ./lib/CL/pocl_llvm_wg.cc:1101 #20 0x7348ff32 in pocl_llvm_generate_workgroup_function_nowrite (DeviceI=DeviceI@entry=0, Device=Device@entry=0x151f050, Kernel=Kernel@entry=0x7fffbcb0, Command=Command@entry=0x7fffbd40, Output=Output@entry=0x7fff6548, Specialize=Specialize@entry=0) at ./lib/CL/pocl_llvm_wg.cc:1147 #21 0x73424b2f in llvm_codegen (output=output@entry=0x30d19f0 "/sbuild-nonexistent/.cache/pocl/kcache/BC/KCELIMKPIAEADDLPJHGMOMPOPMNFLCMCBIOCK/medfilt2d/0-0-0/medfilt2d.so", device_i=device_i@entry=0, kernel=kernel@entry=0x7fffbcb0, device=0x151f050, command=command@entry=0x7fffbd40, specialize=specialize@entry=0) at ./lib/CL/devices/common.c:137 #22 0x7342778e in pocl_check_kernel_disk_cache (command=command@entry=0x7fffbd40, specialized=specialized@entry=0) at ./lib/CL/devices/common.c:983 #23 0x73427e7a in pocl_check_kernel_dlhandle_cache (command=command@entry=0x7fffbd40, retain=retain@entry=0, specialize=specialize@entry=0) at ./lib/CL/devices/common.c:1108 #24 0x7fffe477fc3d in pocl_basic_compile_kernel (cmd=0x7fffbd40, kernel=0x7fffbcb0, device=, specialize=0) at ./lib/CL/devices/basic/basic.c:682 #25 0x7342c71f in pocl_driver_build_poclbinary (program=0x15a6170, device_i=) at ./lib/CL/devices/common_driver.c:969 #26 0x733f291e in get_binary_sizes (sizes=, program=) at ./lib/CL/clGetProgramInfo.c:54 #27 POclGetProgramInfo (program=0x15a6170, param_name=, param_value_size=, param_value=0x15116f0, param_value_size_ret=0x7fffbf70) at ./lib/CL/clGetProgramInfo.c:143 #28 0x736a46ae in pyopencl::program::get_info (this=0x
Bug#1043240: transition: pandas 1.5 -> 2.1 - please upload fixes
for dials it seems that the CI works with pandas 2.1 from experimental. https://ci.debian.net/packages/d/dials/unstable/amd64/41962612/#S4 - Le 30 Jan 24, à 9:05, Rebecca N. Palmer rebecca_pal...@zoho.com a écrit : > I intend to upload pandas 2.x to unstable soon. These packages have a > patch in their bug - please upload them (I'm a DM, I can't do that), or > if you think this patch won't work or isn't a good idea, tell me why: > dials influxdb-python python-altair python-feather-format seaborn tqdm > > In particular, I'd like the seaborn fix uploaded before pandas, so I can > set Breaks for it. (The pandas documentation build-depends on seaborn.)
Bug#1058794: [Debian-pan-maintainers] [Help] Re: python-future: FTBFS: dh_auto_test: error: pybuild --test --test-pytest -i python{version} -p 3.11 returned exit code 13
ok for me - Le 4 Jan 24, à 13:19, Alexandre Detiste alexandre.deti...@gmail.com a écrit : > Le jeu. 4 janv. 2024 à 07:48, Andreas Tille a écrit : >> > @Vincent: this one package "gtextfsm" is yours >> > do you green light an upload ? >> >> If you ask me the package is team maintained and a "Team upload" >> should be fine. > > Hi, I just try to follow the rules I agreed on last month. > > https://salsa.debian.org/python-team/tools/python-modules/blob/master/policy.rst#id2 > >| Team in Uploaders is a weak statement of collaboration. Help in > maintaining the package is appreciated, >| commits to the Git repository are freely welcomed, but before > uploading, please contact the Maintainer for the green light. > > There are not so many packages where "Uploader = DPT" to begin with, > so this might not well a well-known practice... > > So I'm formally asking Ana & PaN for approval to upload "lexicon" and > "dioptas". > (lexicon is a one line change, dioptas needs to package a new release) > > @Vincent: thanks. > > Greetings > > - > > Debian Python Team > dioptas (U) > gtextfsm (U) > lexicon (U) > > Ana Custura > lexicon > > Debian PaN Maintainers > dioptas > > -- > Debian-pan-maintainers mailing list > debian-pan-maintain...@alioth-lists.debian.net > https://alioth-lists.debian.net/cgi-bin/mailman/listinfo/debian-pan-maintainers
Bug#1056792: bitshuffle: ftbfs with cython 3.0.x
It seems to me that the FTBFS was not due to cython 3.x but related to this bug https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=1054716 now that this bug is solved, can you re run the build for bitshuffle ? Frederic
Bug#1051342: Issue with the genx software
Hello, here you should find the informations. platform: Debina unstable python: ~$ python3 Python 3.11.5 (main, Aug 29 2023, 15:31:31) [GCC 13.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> $ dpkg -l | grep wx ii libwxbase3.2-1:amd64 3.2.2+dfsg-4 amd64wxBase library (runtime) - non-GUI support classes of wxWidgets toolkit ii libwxgtk-gl3.2-1:amd64 3.2.2+dfsg-4 amd64wxWidgets Cross-platform C++ GUI toolkit (GTK 3 gl library runtime) ii libwxgtk3.2-1:amd643.2.2+dfsg-4 amd64wxWidgets Cross-platform C++ GUI toolkit (GTK 3 runtime) ii python3-wxgtk4.0 4.2.1+dfsg-1 amd64Python 3 interface to the wxWidgets Cross-platform C++ GUI toolkit ii python3-wxutils0.3.0-1 all wxPython utilities and convenience functions (Python 3) > I suspect that something changed in the WX API causing this issue. I don't > think > this should prevent you from packaging, as this part of the code only gets > called upon user request to exit. (There might be no request to save or > missing > storage of window state, but no big impact on the functionality). I already uploaded the version into Debian and you can see that there is issues with numba... https://buildd.debian.org/status/package.php?p=genx one probleme at a time :)) Cheers
Bug#1051342: Issue with the genx software
Hello, I am preparing the packaging of genx 3.6.22. When I try to quit the application I have this error message CRITICAL: uncought python error Traceback (most recent call last): File "/usr/lib/python3/dist-packages/genx/gui/main_window.py", line 1418, in eh_mb_quit if event.CanVeto() and not self.model_control.saved: ^ AttributeError: 'CommandEvent' object has no attribute 'CanVeto' Traceback (most recent call last): File "/usr/lib/python3/dist-packages/genx/gui/main_window.py", line 1418, in eh_mb_quit if event.CanVeto() and not self.model_control.saved: ^ So I would like your help in order to solve this issue. thanks for considering. Frédéric
Bug#1041803: [Debian-pan-maintainers] Bug#1041803: hyperspy: FTBFS test_image fails
the old and new hyperspy is not compatible with imagio > 0.28. I kindly opened a bug report about the situation at the upstream git repository.
Bug#1042740: [Debian-pan-maintainers] Bug#1042740: dmrgpp: Please package the new 6.04 version
I filled a bug report about the new upstream. https://github.com/g1257/dmrgpp/issues/40 It is already fixed in the upstream git https://github.com/g1257/dmrgpp/commit/528501e4a5814d4cbb80e2cf16ea407f9e012ee6
Bug#1026864: dmrgpp: flaky autopkgtest on amd64: times out
and a comment about this issue https://github.com/g1257/dmrgpp/issues/38#issuecomment-1655740289
Bug#1040159: epics-base: embedded yajl is vulnerable to CVE-2017-16516 and CVE-2022-24795
Hello, I dicovered that upstream modifier yajl in order to support json5. I am wondering if their modification could not be integrated in our yajl. I fill a burg report about this idea here https://github.com/epics-base/epics-base/issues/405 Tell me what is your opinion about this. Cheers Fred
Bug#1041443: [Debian-pan-maintainers] Bug#1041443: pyfai_2023.5.0+dfsg1-3_all-buildd.changes REJECTED
> I am just the messenger here, if you disagree, please feel free to > contact ftpmasters or lintian maintainers. This was not a rant about this, I just wanted to understand what is going on :). > Your package has been built successfully on (some) buildds, but then the > binaries upload got rejected by dak, that's why they are still in > "Uploaded" state. Overall it's just like if pyfai hasn't been built or > fails to build from source. So until a new upstream source is available with other timestamp, it will not be uploadable. In you opinion, can we discuss about this on debian-devel ? Cheers Fred
Bug#1041443: [Debian-pan-maintainers] Bug#1041443: Bug#1041443: pyfai_2023.5.0+dfsg1-3_all-buildd.changes REJECTED
I just check this date is in the upstream tar file https://files.pythonhosted.org/packages/54/84/ea12e176489b35c4610625ce56aa2a1d91ab235b0caa71846317bfd1192f/pyfai-2023.5.0.tar.gz
Bug#1041443: [Debian-pan-maintainers] Bug#1041443: pyfai_2023.5.0+dfsg1-3_all-buildd.changes REJECTED
ok, it seems that I generated an orig.tag.gz with this (Thu Jan 1 00:00:00 1970). I can not remember which tool I used to generate this file. gbp import-orig --uscan or deb-new-upstream Nevertheless, why is it a serious bug ? thanks Frederic
Bug#1033600: emacs: File mode specification error: (file-missing Cannot open load file Aucun fichier ou dossier de ce type pylint)
If I remove the pylint package, no more error...
Bug#1024859: change in the extention importation with 3.11
There is a fix from the upstream around enum. https://github.com/boostorg/python/commit/a218babc8daee904a83f550fb66e5cb3f1cb3013 Fix enum_type_object type on Python 3.11 The enum_type_object type inherits from PyLong_Type which is not tracked by the GC. Instances doesn't have to be tracked by the GC: remove the Py_TPFLAGS_HAVE_GC flag. The Python C API documentation says: "To create a container type, the tp_flags field of the type object must include the Py_TPFLAGS_HAVE_GC and provide an implementation of the tp_traverse handler." https://docs.python.org/dev/c-api/gcsupport.html The new exception was introduced in Python 3.11 by: python/cpython#88429 an opinion ?
Bug#1024859: change in the extention importation with 3.11
in order to debug this, I started gdb set a breakpoint in init_module_scitbx_linalg_ext then a catch throw and I end up with this backtrace Catchpoint 2 (exception thrown), 0x770a90a1 in __cxxabiv1::__cxa_throw (obj=0xb542e0, tinfo=0x772d8200 , dest=0x772c1290 ) at ../../../../src/libstdc++-v3/libsupc++/eh_throw.cc:81 81 ../../../../src/libstdc++-v3/libsupc++/eh_throw.cc: Le dossier n'est pas vide. (gdb) bt #0 0x770a90a1 in __cxxabiv1::__cxa_throw (obj=0xb542e0, tinfo=0x772d8200 , dest=0x772c1290 ) at ../../../../src/libstdc++-v3/libsupc++/eh_throw.cc:81 #1 0x772ad089 in boost::python::throw_error_already_set () at libs/python/src/errors.cpp:61 #2 0x772b6f05 in boost::python::objects::(anonymous namespace)::new_enum_type (doc=0x0, name=0x7743ddf9 "bidiagonal_matrix_kind") at libs/python/src/object/enum.cpp:169 #3 boost::python::objects::enum_base::enum_base (this=this@entry=0x7fffcee0, name=name@entry=0x7743ddf9 "bidiagonal_matrix_kind", to_python=to_python@entry=0x7741f720 ::to_python(void const*)>, convertible=convertible@entry=0x77422e50 ::convertible_from_python(_object*)>, construct=construct@entry=0x7741fb60 ::construct(_object*, boost::python::converter::rvalue_from_python_stage1_data*)>, id=..., doc=0x0) at libs/python/src/object/enum.cpp:204 #4 0x774203cb in boost::python::enum_::enum_ (this=0x7fffcee0, name=0x7743ddf9 "bidiagonal_matrix_kind", doc=0x0) at /usr/include/boost/python/enum.hpp:45 #5 0x77428330 in scitbx::matrix::boost_python::bidiagonal_matrix_svd_decomposition_wrapper::wrap (name=name@entry=0x7743dbd0 "svd_decomposition_of_bidiagonal_matrix") at ./scitbx/linalg/boost_python/svd.cpp:19 #6 0x7741f6b0 in scitbx::matrix::boost_python::wrap_svd () at ./scitbx/linalg/boost_python/svd.cpp:66 #7 0x773f8aa3 in scitbx::matrix::boost_python::(anonymous namespace)::init_module () at ./scitbx/linalg/boost_python/linalg_ext.cpp:19 #8 0x772c13e3 in boost::function0::operator() (this=0x7fffd2b0) at ./boost/function/function_template.hpp:763 #9 boost::python::handle_exception_impl (f=...) at libs/python/src/errors.cpp:25 #10 0x772c1b69 in boost::python::handle_exception (f=) at ./boost/function/function_template.hpp:635 #11 boost::python::detail::(anonymous namespace)::init_module_in_scope (init_function=0x773f8ac0 , m=) at libs/python/src/module.cpp:24 #12 boost::python::detail::init_module (moduledef=..., init_function=0x773f8ac0 ) at libs/python/src/module.cpp:43 not crystal clear to me :)
Bug#1013158: facet-analyser: vtk[6,7] removal
Hello Anton, I have just pushed a few dependencies in the -dev package in the salsa repo I did not updated the changelog. Cheers Fred
Bug#1013158: facet-analyser: vtk[6,7] removal
Hello Anton, I try to checkout paraview in order to add the -dev dependencies but I have this message $ git clone https://salsa.debian.org/science-team/paraview Clonage dans 'paraview'... remote: Enumerating objects: 175624, done. remote: Counting objects: 100% (78929/78929), done. remote: Compressing objects: 100% (38687/38687), done. remote: Total 175624 (delta 47039), reused 65625 (delta 39190), pack-reused 96695 Réception d'objets: 100% (175624/175624), 246.21 Mio | 12.11 Mio/s, fait. Résolution des deltas: 100% (109096/109096), fait. [attr]our-c-style whitespace=tab-in-indent,-blank-at-eol format.clang-format non permis : ThirdParty/QtTesting/vtkqttesting/.gitattributes : 8 [attr]our-c-style whitespace=tab-in-indent,-blank-at-eol format.clang-format=9 non permis : ThirdParty/catalyst/vtkcatalyst/catalyst/.gitattributes : 4 [attr]our-c-style whitespace=tab-in-indent,-blank-at-eol format.clang-format=8 non permis : VTK/.gitattributes : 10 [attr]our-c-style whitespace=tab-in-indent format.clang-format=9 non permis : VTK/ThirdParty/vtkm/vtkvtkm/vtk-m/.gitattributes : 2 Mise à jour des fichiers: 100% (30828/30828), fait. [attr]our-c-style whitespace=tab-in-indent,-blank-at-eol format.clang-format=8 non permis : VTK/.gitattributes : 10 [attr]our-c-style whitespace=tab-in-indent format.clang-format=9 non permis : VTK/ThirdParty/vtkm/vtkvtkm/vtk-m/.gitattributes : 2 Downloading VTK/ThirdParty/vtkm/vtkvtkm/vtk-m/data/README.md (643 B) Error downloading object: VTK/ThirdParty/vtkm/vtkvtkm/vtk-m/data/README.md (b30a14a): Smudge error: Error downloading VTK/ThirdParty/vtkm/vtkvtkm/vtk-m/data/README.md (b30a14a308f64c6fc2969e2b959d79dacdc5affda1d1c0e24f8e176304147146): [b30a14a308f64c6fc2969e2b959d79dacdc5affda1d1c0e24f8e176304147146] Object does not exist on the server or you don't have permissions to access it: [404] Object does not exist on the server or you don't have permissions to access it Errors logged to /home/experiences/instrumentation/picca/debian/science-team/paraview/.git/lfs/logs/20221101T101535.441130442.log Use `git lfs logs last` to view the log. error: le filtre externe 'git-lfs filter-process' a échoué fatal: VTK/ThirdParty/vtkm/vtkvtkm/vtk-m/data/README.md : le filtre smudge 'lfs' a échoué warning: Le clone a réussi, mais l'extraction a échoué. Vous pouvez inspecter ce qui a été extrait avec 'git status' et réessayer avec 'git restore --source=HEAD :/'
Bug#1016598: [Debian-pan-maintainers] Bug#1016598: binoculars: vtk[6, 7] removal
Hello François, thanks a lot, I removed the NMU number and release a -2 package. (uploaded) thanks for your contribution to Debian. Fred
Bug#1008119: [Debian-pan-maintainers] Bug#1008119: Bug#1008119: src:pyfai: fails to migrate to testing for too long: autopkgtest regression
It seems that it failing now https://ci.debian.net/packages/p/pyfai/ I am on 0.21.2 but I do not know if it solve this mask issue. Cheers Fred
Bug#1003061: [Debian-pan-maintainers] Bug#1003061: bug 1003061: dmrgpp: autopkgtest failure on armhf: segmentation fault
Hello Paul, just for info, I have already reported this issue here https://github.com/g1257/dmrgpp/issues/38 cheers Fred.
Bug#1001168: [Debian-pan-maintainers] Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py
Is it not better to use the DEB__MAINT_APPEND variable in order to deal with this issue ?
Bug#1003037:
It seems that this is an issue in gcc has observed when compiling tensorflow https://zenn.dev/nbo/scraps/8f1505e365d961
Bug#1001168: Info received (Bug#1001168: Info received (Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py))
Built with gcc-11 and -fno-lto it doesn not work. (sid_mips64el-dchroot)picca@eller:~/matplotlib/build/lib.linux-mips64-3.9$ ../../../test.py Segmentation fault (sid_mips64el-dchroot)picca@eller:~/matplotlib/build/lib.linux-mips64-3.9$ PYTHONPATH=. ../../../test.py Segmentation fault
Bug#1001168: Info received (Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py)
I tested matplotlib built with numpy 0.17 0.19 0.21. each time I got the segfault. another difference was the gcc compiler. So I switched to gcc-10 (sid_mips64el-dchroot)picca@eller:~/matplotlib$ CC=gcc-10 python3 setup.py build if failed with this error lto1: fatal error: bytecode stream in file ‘build/temp.linux-mips64-3.9/matplotlib.backends._backend_agg/extern/agg24-svn/src/agg_bezier_arc.o’ generated with LTO version 9.4 instead of the expected 11.2 So I unactivated lto with this CFLAGS="-fno-lto" CC=gcc-10 python3 setup.py build at the end it seems that is does not segfault :) (sid_mips64el-dchroot)picca@eller:~/matplotlib/build/lib.linux-mips64-3.9$ ../../../test.py Segmentation fault (sid_mips64el-dchroot)picca@eller:~/matplotlib/build/lib.linux-mips64-3.9$ PYTHONPATH=. ../../../test.py (sid_mips64el-dchroot)picca@eller:~/matplotlib/build/lib.linux-mips64-3.9$ ls matplotlib mpl_toolkits pylab.py toto.png Cheers
Bug#1001168: Info received (Bug#1001168: Info received ())
If I run in the sid chroot, but with the binaryed built from bullseye, it works. (sid_mips64el-dchroot)picca@eller:~/matplotlib-3.5.0/build/lib.linux-mips64-3.9$ rm toto.png (sid_mips64el-dchroot)picca@eller:~/matplotlib-3.5.0/build/lib.linux-mips64-3.9$ python3 test.py (sid_mips64el-dchroot)picca@eller:~/matplotlib-3.5.0/build/lib.linux-mips64-3.9$ ls matplotlib mpl_toolkits pylab.py test.py toto.png
Bug#1001168: Info received ()
Here no error during the build of numpy 1.19.5 = 10892 passed, 83 skipped, 108 deselected, 19 xfailed, 2 xpassed, 2 warnings in 1658.41s (0:27:38) = but 109 for numpy 1.21... = 14045 passed, 397 skipped, 1253 deselected, 20 xfailed, 2 xpassed, 2 warnings, 109 errors in 869.47s (0:14:29) =
Bug#1001168: Info received ()
I investigated a bit more, it seems that cover is wrong. In a bullseye chroot it works $ python3 ./test.py (bullseye_mips64el-dchroot)picca@eller:~/matplotlib-3.5.0/build/lib.linux-mips64-3.9$ ls matplotlib mpl_toolkits pylab.py test.py toto.png I found that the test failed between the 3.3.4-2 and 3.3.4-2+b1 rebuild This binNMU was about python3.10 support, but in the same time numpy has changed from python3-numpy (= 1:1.19.5-1), to python3-numpy (= 1:1.21.4-2), I think that there is a non negligeable possibility for this bug to be triggered by the new numpy.
Bug#1001168: full python backtrace and print locals
the full python backtrace #8 #14 Frame 0x120debd80, for file /home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/lines.py, line 2888, in draw (self=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=None, _snap=None, _sketch=None, _path_effects=[], _sticky_edges=<_XYPair at remote 0xfff51bacc0>, _in_layout=True, _suptitle=None, _supxlabel=None, _supylabel=None, _align_label_groups={'x': , 'y': }, _g...(truncated) #20 Frame 0xfff448f230, for file /home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/artist.py, line 50, in draw_wrapper (artist=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=None, _snap=None, _sketch=None, _path_effects=[], _sticky_edges=<_XYPair at remote 0xfff51bacc0>, _in_layout=True, _suptitle=None, _supxlabel=None, _supylabel=None, _align_label_groups={'x': , 'y': , _axes=, _axes=<...>, figure=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=None, _snap=None, _sketch=None, _path_effects=[], _sticky_edges=<_XYPair at remote 0xfff51bacc0>, _in_layout=True, _suptitle=None, _supxlabel=None, _supylabel=None, _align_label_group...(truncated) 'matplotlib.ticker.Locator') #33 Frame 0xfff448f040, for file /home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/artist.py, line 50, in draw_wrapper (artist=, _axes=, _axes=<...>, figure=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=None, _snap=None, _sketch=None, _path_effects=[], _sticky_edges=<_XYPair at remote 0xfff51bacc0>, _in_layout=True, _suptitle=None, _supxlabel=None, _supylabel=None, _align_...(truncated) return draw(artist, renderer) #40 Frame 0xfff46217c0, for file /home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/axis.py, line 1419, in draw (self=, _axes=, _axes=<...>, figure=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=None, _snap=None, _sketch=None, _path_effects=[], _sticky_edges=<_XYPair at remote 0xfff51bacc0>, _in_layout=True, _suptitle=None, _supxlabel=None, _supylabel=None, _align_label_grou...(truncated) elif not visible: # something false-like but not None #47 Frame 0xfff461a230, for file /home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/artist.py, line 50, in draw_wrapper (artist=, _axes=, _axes=<...>, figure=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=None, _snap=None, _sketch=None, _path_effects=[], _sticky_edges=<_XYPair at remote 0xfff51bacc0>, _in_layout=True, _suptitle=None, _supxlabel=None, _supylabel=None, _align_...(truncated) return draw(artist, renderer) #54 Frame 0xfff447a040, for file /home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/image.py, line 388, in _draw_list_compositing_images (artists=[, _axes=, _axes=<...>, figure=, figure=<...>, _transform=None, _transformSet=False, _visible=True, _animated=False, _alpha=None, clipbox=None, _clippath=None, _clipon=True, _label='', _picker=None, _rasterized=False, _agg_filter=None, _mouseover=False, _callbacks=, callbacks={}, _cid_gen=, _func_cid_map={}, _pickled_cids=set()) at remote 0xfff7a3be20>, _remove_method=None, _url=None, _gid=N
Bug#1001168:
Here the py-bt (gdb) py-bt Traceback (most recent call first): File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/lines.py", line 2888, in draw File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/artist.py", line 50, in draw_wrapper return draw(artist, renderer) File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/axis.py", line 555, in draw 'matplotlib.ticker.Locator') File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/artist.py", line 50, in draw_wrapper return draw(artist, renderer) File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/axis.py", line 1419, in draw elif not visible: # something false-like but not None File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/artist.py", line 50, in draw_wrapper return draw(artist, renderer) File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/image.py", line 388, in _draw_list_compositing_images extra_height = (out_height - out_height_base) / out_height_base File "/home/picca/matplotlib-3.5.0/build/lib.linux-mips64-3.9/matplotlib/axes/_base.py", line 4106, in draw Python Exception 'ascii' codec can't decode byte 0xc2 in position 2280: ordinal not in range(128): Error occurred in Python: 'ascii' codec can't decode byte 0xc2 in position 2280: ordinal not in range(128)
Bug#1001168: Info received (Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py)
I can confirm that the bullseye matplotlib does not produce a segfault
Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py
This small script trigger the segfault. #!/usr/bin/env python3 import matplotlib import matplotlib.pyplot as plt plt.figure() plt.title("foo") plt.savefig("toto.png")
Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py
bugs report are already filled on matplotlib #1000774 and #1000435 I will try to see if this is identical...
Bug#1001168: hkl: FTBFS on mipsel: FAIL: trajectory.py
Here the backtrace on mips64el #0 agg::pixfmt_alpha_blend_rgba, agg::order_rgba>, agg::row_accessor >::blend_solid_hspan(int, int, unsigned int, agg::rgba8T const&, unsigned char const*) (covers=0x100 , c=..., len=, y=166, x=, this=) at extern/agg24-svn/include/agg_color_rgba.h:395 #1 agg::renderer_base, agg::order_rgba>, agg::row_accessor > >::blend_solid_hspan(int, int, int, agg::rgba8T const&, unsigned char const*) (covers=, c=..., len=, y=166, x=, this=0x12123abf8) at extern/agg24-svn/include/agg_renderer_base.h:294 #2 agg::render_scanline_aa_solid::embedded_scanline, agg::renderer_base, agg::order_rgba>, agg::row_accessor > >, agg::rgba8T >(agg::serialized_scanlines_adaptor_aa::embedded_scanline const&, agg::renderer_base, agg::order_rgba>, agg::row_accessor > >&, agg::rgba8T const&) (color=..., ren=..., sl=...) at extern/agg24-svn/include/agg_renderer_scanline.h:40 #3 agg::renderer_scanline_aa_solid, agg::order_rgba>, agg::row_accessor > > >::render::embedded_scanline>(agg::serialized_scanlines_adaptor_aa::embedded_scanline const&) (sl=..., this=0x12123ac10) at extern/agg24-svn/include/agg_renderer_scanline.h:130 #4 agg::render_scanlines, agg::serialized_scanlines_adaptor_aa::embedded_scanline, agg::renderer_scanline_aa_solid, agg::order_rgba>, agg::row_accessor > > > >(agg::serialized_scanlines_adaptor_aa&, agg::serialized_scanlines_adaptor_aa::embedded_scanline&, agg::renderer_scanline_aa_solid, agg::order_rgba>, agg::row_accessor > > >&) [clone .part.0] [clone .lto_priv.0] (ras=..., sl=..., ren=...) at extern/agg24-svn/include/agg_renderer_scanline.h:446 #5 0x00fff49a367c in agg::render_scanlines, agg::serialized_scanlines_adaptor_aa::embedded_scanline, agg::renderer_scanline_aa_solid, agg::order_rgba>, agg::row_accessor > > > >(agg::serialized_scanlines_adaptor_aa&, agg::serialized_scanlines_adaptor_aa::embedded_scanline&, agg::renderer_scanline_aa_solid, agg::order_rgba>, agg::row_accessor > > >&) (ren=..., sl=..., ras=...) at extern/agg24-svn/include/agg_renderer_scanline.h:440 #6 RendererAgg::draw_markers(GCAgg&, py::PathIterator&, agg::trans_affine&, py::PathIterator&, agg::trans_affine&, agg::rgba) (color=..., trans=..., path=..., marker_trans=..., marker_path=..., gc=..., this=0x12123aaa0) at src/_backend_agg.h:658 #7 PyRendererAgg_draw_markers(PyRendererAgg*, _object*) (self=, args=) at src/_backend_agg_wrapper.cpp:285 #8 0x0001202e36a0 in cfunction_call (func=0xfff11f08b0, args=, kwargs=) at ../Objects/methodobject.c:552 #9 0x00012003891c in _PyObject_MakeTpCall (tstate=0x1205a97c0, callable=0xfff11f08b0, args=, nargs=, keywords=0x0) at ../Objects/call.c:191 #10 0x00012002841c in _PyObject_VectorcallTstate (kwnames=0x0, nargsf=9223372036854775814, args=0x121281590, callable=0xfff11f08b0, tstate=) at ../Include/cpython/abstract.h:116 #11 _PyObject_VectorcallTstate (kwnames=0x0, nargsf=9223372036854775814, args=0x121281590, callable=0xfff11f08b0, tstate=) at ../Include/cpython/abstract.h:103 #12 PyObject_Vectorcall (kwnames=0x0, nargsf=9223372036854775814, args=0x121281590, callable=0xfff11f08b0) at ../Include/cpython/abstract.h:127 #13 call_function (kwnames=0x0, oparg=, pp_stack=, tstate=) at ../Python/ceval.c:5075 #14 _PyEval_EvalFrameDefault (tstate=, f=, throwflag=) at ../Python/ceval.c:3487 #15 0x00012001d898 in _PyEval_EvalFrame (throwflag=0, f=0x121281330, tstate=0x1205a97c0) at ../Include/internal/pycore_ceval.h:40 #16 function_code_fastcall (tstate=0x1205a97c0, co=, args=0xfff120cd78, nargs=2, globals=) at ../Objects/call.c:330 #17 0x00012002535c in _PyObject_VectorcallTstate (kwnames=0x0, nargsf=, args=0xfff120cd68, callable=0xfff520a310, tstate=0x1205a97c0) at ../Include/cpython/abstract.h:118 #18 PyObject_Vectorcall (kwnames=0x0, nargsf=, args=0xfff120cd68, callable=0xfff520a310) at ../Include/cpython/abstract.h:127 #19 call_function (kwnames=0x0, oparg=, pp_stack=, tstate=) at ../Python/ceval.c:5075 #20 _PyEval_EvalFrameDefault (tstate=, f=, throwflag=) at ../Python/ceval.c:3518 #21 0x000120107de0 in _PyEval_EvalFrame (throwflag=0, f=0xfff120cbe0, tstate=0x1205a97c0) at ../Include/internal/pycore_ceval.h:40 #22 _PyEval_EvalCode (tstate=0x1205a97c0, _co=0xfff5231500, globals=, locals=, args=, argcount=2, kwnames=0x0, kwargs=0xfff1069720, kwcount=0, kwstep=1, defs=0x0, defcount=0, kwdefs=0x0, closure=0xfff51fafa0, name=0xfff7953bb0, qualname=0xfff5264730) at ../Python/ceval.c:4327 #23 0x000120039e20 in _PyFunction_Vectorcall (func=, stack=, nargsf=, kwnames=) at ../Objects/call.c:396 #24 0x0001200262dc in _PyObject_VectorcallTstate (kwnames=0x0, nargsf=, args=0xfff1069710, callable=0xfff520a3a0, tstate=0x1205a97c0) at ../Include/cpython/abstract.h:118 #25 PyObject_Vectorcall (kwnames=0x0, nargsf=, args=0xfff1069710, callable=0xfff520a3a0) at ../Include/cpython/abstract.h:127 #26 call_function (kwnames=0x0, oparg=, p
Bug#994226:
Hello reading this I do not understand what need to be changed in the init script https://www.freedesktop.org/wiki/Software/systemd/NetworkTarget/ Whenever systemd encounters a $network dependency in LSB headers of init scripts it will translate this to a Wants= and After= dependency on network-online.target hence staying relatively close to traditional LSB behaviour. So it seems to me that systemd already translate the network target into network-online cheers Fred
Bug#994882: ITS: vitables
ack ;) sorry for the delay. You can commit directly to the repository. Cheers Fred De : Benda Xu [o...@debian.org] Envoyé : vendredi 24 septembre 2021 10:24 À : Anton Gladky Cc : 994...@bugs.debian.org; PICCA Frederic-Emmanuel Objet : Re: Bug#994882: ITS: vitables Anton Gladky writes: > Thanks for your contribution. I have approved and merged your MR. Also > I have added you to the Debian Science group on salsa. > > @PICCA Frederic-Emmanuel, would you want also to check those changes? Thanks Anton! I will wait for Picca's ack for a month before I go ahead to add myself to uploaders list. Yours, Benda
Bug#961183: metis: providing a 64-bit build
Hello, why there is no pkgconfig files provided with metis and metis64. this simplify the configuration of packages depending on these libraries. Cheers Fred
Bug#988404: Info received (Bug#988404: Info received (Bug#988404: Acknowledgement (linux: can't claim BAR 15 [mem 0x2df80000000-0x2e0b1ffffff 64bit pref]: no compatible bridge window)))
lspci -tnnkv lspci Description: lspci
Bug#988404: Acknowledgement (linux: can't claim BAR 15 [mem 0x2df80000000-0x2e0b1ffffff 64bit pref]: no compatible bridge window)
$ lspci -t -+-[:e0]-+-00.0 | +-00.2 | +-01.0 | +-02.0 | +-03.0 | +-03.2-[e1-e2]--+-00.0 | | \-00.1 | +-04.0 | +-05.0 | +-07.0 | +-07.1-[e3]--+-00.0 | |\-00.2 | +-08.0 | \-08.1-[e4]--+-00.0 |\-00.2 +-[:c0]-+-00.0 | +-00.2 | +-01.0 | +-01.1-[c1]00.0 | +-02.0 | +-03.0 | +-04.0 | +-05.0 | +-07.0 | +-07.1-[c2]--+-00.0 | |\-00.2 | +-08.0 | +-08.1-[c3]--+-00.0 | |\-00.2 | \-08.2-[c4]00.0 +-[:a0]-+-00.0 | +-00.2 | +-01.0 | +-02.0 | +-03.0 | +-03.1-[a1]-- | +-03.2-[a2]-- | +-03.3-[a3]-- | +-03.4-[a4]-- | +-04.0 | +-05.0 | +-07.0 | +-07.1-[a5]--+-00.0 | |\-00.2 | +-08.0 | \-08.1-[a6]--+-00.0 |+-00.1 |\-00.2 +-[:80]-+-00.0 | +-00.2 | +-01.0 | +-01.1-[81-87]00.0-[82-87]--+-04.0-[83]00.0 | | +-08.0-[84]00.0 | | +-0c.0-[85]--+-00.0 | | |\-00.1 | | +-10.0-[86]--+-00.0 | | |\-00.1 | | \-14.0-[87]00.0 | +-02.0 | +-03.0 | +-04.0 | +-05.0 | +-07.0 | +-07.1-[88]--+-00.0 | |\-00.2 | +-08.0 | \-08.1-[89]--+-00.0 |\-00.2 +-[:60]-+-00.0 | +-00.2 | +-01.0 | +-02.0 | +-03.0 | +-03.2-[61]00.0 | +-03.5-[62-63]--+-00.0 | | \-00.1 | +-03.6-[64-65]00.0-[65]00.0 | +-04.0 | +-05.0 | +-07.0 | +-07.1-[66]--+-00.0 | |\-00.2 | +-08.0 | \-08.1-[67]--+-00.0 |\-00.2 +-[:40]-+-00.0 | +-00.2 | +-01.0 | +-01.1-[41-47]00.0-[42-47]--+-04.0-[43]00.0 | | +-08.0-[44]00.0 | | +-0c.0-[45]00.0 | | +-10.0-[46]--+-00.0 | | |\-00.1 | | \-14.0-[47]-- | +-02.0 | +-03.0 | +-04.0 | +-05.0 | +-07.0 | +-07.1-[48]--+-00.0 | |\-00.2 | +-08.0 | +-08.1-[49]--+-00.0 | |\-00.2 | \-08.2-[4a]00.0 +-[:20]-+-00.0 | +-00.2 | +-01.0 | +-02.0 | +-03.0 | +-04.0 | +-05.0 | +-07.0 | +-07.1-[21]--+-00.0 | |\-00.2 | +-08.0 | \-08.1-[22]--+-00.0 |+-00.1 |+-00.2 |\-00.3 \-[:00]-+-00.0 +-00.2 +-01.0 +-01.1-[01]00.0 +-01.2-[02]00.0 +-01.3-[03]-- +-01.4-[04]-- +-02.0 +-03.0 +-04.0 +-05.0 +-07.0 +-07.1-[05]--+-00.0 |\-00.2 +-08.0 +-08.1-[06]--+-00.0 |+-00.2 |\-00.3 +-14.0 +-14.3 +-18.0 +-18.1 +-18.2 +-18.3 +-18.4 +-18.5 +-18.6 +-18.7 +-19.0 +-19.1 +-19.2 +-19.3 +-19.4 +-19.5 +-19.6 \-19.7
Bug#988404: linux: can't claim BAR 15 [mem 0x2df80000000-0x2e0b1ffffff 64bit pref]: no compatible bridge window
Source: linux Severity: normal Dear Maintainer, For futher information, you can have a look here https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=986312 this bug makes it impossible to use all the graphical cards in my computer. The computer contain 2 kinds of graphicalcards 3 x RTX 3090 and 6 x T4. The proble is present also with the nouveau driver only. This is why I think this is not related to the nvidia driver. We checked that this problem is not present on a CentOS6 system so it must be related to a specificity of the Debian kernel. do not hesitate to ask for further information in order to help solve this issue. thanks for considering -- System Information: Debian Release: bullseye/sid APT prefers testing APT policy: (500, 'testing') Architecture: amd64 (x86_64) Kernel: Linux 5.10.0-6-amd64 (SMP w/192 CPU threads) Kernel taint flags: TAINT_PROPRIETARY_MODULE, TAINT_OOT_MODULE, TAINT_UNSIGNED_MODULE Locale: LANG=fr_FR.UTF-8, LC_CTYPE=fr_FR.UTF-8 (charmap=UTF-8), LANGUAGE not set Shell: /bin/sh linked to /usr/bin/dash Init: systemd (via /run/systemd/system) LSM: AppArmor: enabled
Bug#986785: binoculars gui does not work
Hello Andrius. I am on it, I am also the upstream of this software :)). Thanks for your help, nevertheless I am a bit like you, I did not wrote this part of the code :)) and the matplotlib changelog is sort of useless :)) Cheers Fred
Bug#986312: NVRM: request_mem_region failed for 0M @ 0x0.
I have also this information in the syslog Mar 29 08:01:51 re-grades-01 kernel: [1.762583] pci :42:04.0: BAR 15: no space for [mem size 0x13800 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762584] pci :42:04.0: BAR 15: failed to assign [mem size 0x13800 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762586] pci :42:08.0: BAR 15: no space for [mem size 0x13800 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762587] pci :42:08.0: BAR 15: failed to assign [mem size 0x13800 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762589] pci :42:0c.0: BAR 15: no space for [mem size 0x13800 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762590] pci :42:0c.0: BAR 15: failed to assign [mem size 0x13800 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762592] pci :43:00.0: BAR 1: no space for [mem size 0x1000 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762593] pci :43:00.0: BAR 1: trying firmware assignment [mem 0x2e08000-0x2e08fff 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762594] pci :43:00.0: BAR 1: [mem 0x2e08000-0x2e08fff 64bit pref] conflicts with PCI Bus :40 [mem 0x260c020-0x2e0c01\ f window] Mar 29 08:01:51 re-grades-01 kernel: [1.762595] pci :43:00.0: BAR 1: failed to assign [mem size 0x1000 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762596] pci :43:00.0: BAR 8: no space for [mem size 0x1 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762597] pci :43:00.0: BAR 8: trying firmware assignment [mem 0x2df8000-0x2e07fff 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762598] pci :43:00.0: BAR 8: [mem 0x2df8000-0x2e07fff 64bit pref] conflicts with PCI Bus :40 [mem 0x260c020-0x2e0c01\ f window] Mar 29 08:01:51 re-grades-01 kernel: [1.762599] pci :43:00.0: BAR 8: failed to assign [mem size 0x1 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762600] pci :43:00.0: BAR 3: no space for [mem size 0x0200 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762601] pci :43:00.0: BAR 3: trying firmware assignment [mem 0x2e0b000-0x2e0b1ff 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762602] pci :43:00.0: BAR 3: [mem 0x2e0b000-0x2e0b1ff 64bit pref] conflicts with PCI Bus :40 [mem 0x260c020-0x2e0c01\ f window] Mar 29 08:01:51 re-grades-01 kernel: [1.762603] pci :43:00.0: BAR 3: failed to assign [mem size 0x0200 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762604] pci :43:00.0: BAR 10: no space for [mem size 0x2000 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762604] pci :43:00.0: BAR 10: trying firmware assignment [mem 0x2e09000-0x2e0afff 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762606] pci :43:00.0: BAR 10: [mem 0x2e09000-0x2e0afff 64bit pref] conflicts with PCI Bus :40 [mem 0x260c020-0x2e0c01fff\ ff window] Mar 29 08:01:51 re-grades-01 kernel: [1.762606] pci :43:00.0: BAR 10: failed to assign [mem size 0x2000 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762607] pci :42:04.0: PCI bridge to [bus 43] Mar 29 08:01:51 re-grades-01 kernel: [1.762611] pci :42:04.0: bridge window [mem 0xca00-0xcb3f] Mar 29 08:01:51 re-grades-01 kernel: [1.762619] pci :44:00.0: BAR 1: no space for [mem size 0x1000 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762620] pci :44:00.0: BAR 1: trying firmware assignment [mem 0x2df4000-0x2df4fff 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762621] pci :44:00.0: BAR 1: [mem 0x2df4000-0x2df4fff 64bit pref] conflicts with PCI Bus :40 [mem 0x260c020-0x2e0c01\ f window] Mar 29 08:01:51 re-grades-01 kernel: [1.762622] pci :44:00.0: BAR 1: failed to assign [mem size 0x1000 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762622] pci :44:00.0: BAR 8: no space for [mem size 0x1 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762623] pci :44:00.0: BAR 8: trying firmware assignment [mem 0x2de4000-0x2df3fff 64bit pref] Mar 29 08:01:51 re-grades-01 kernel: [1.762624] pci :44:00.0: BAR 8: [mem 0x2de4000-0x2df3fff 64bit pref] conflicts with PCI Bus :40 [mem 0x260c020-0x2e0c01\ Maybe the conflict is responsible of this issue. Do you have an idea of what should cause this conflict ?
Bug#986312: NVRM: request_mem_region failed for 0M @ 0x0.
Hello andreas > what's the NUMA layout of the machine? what cpus do you have in there? I attached the lstopo output. > Great. Half of the bus ids are in hex, half in decimal. > How many cards do you have in there? And how many pci devices are there > per card? the list is in the reportbug :) 3 x RTX 3090 and 6 x T4 > Have you tried different drivers? (tesla-450, tesla-460) only the 460 > Have you tried with only one kind of cards (Tesla or Geforce) in there? no > Have you tried with only a single card in there? no Fred re-grades-01.pdf Description: re-grades-01.pdf
Bug#976735: spyder needs source upload for Python 3.9 transition
> Strangely enough, I've already done that ;-) my bad. Cheers Fred
Bug#976735: spyder needs source upload for Python 3.9 transition
> I have a package of Spyder 4 waiting to upload, but it requires five > packages to be accepted into unstable from NEW first (pyls-server, > pyls-black, pyls-spyder, abydos, textdistance); once that happens, the > rest of the packages are almost ready to go. Maybe you can contact the ftpmaster team and request a review of these packages. In order to avoid the spyder removal. Cheers
Bug#976952: [Help] Re: lmfit-py: FTBFS on ppc64el (arch:all-only src pkg): dh_auto_test: error: pybuild --test --test-pytest -i python{version} -p 3.9 returned exit code 13
Ok, in that case, I think that a comment in the d/rules files is enough in order to keep in mind that we have this issue with ppc64el.
Bug#976952: [Help] Re: lmfit-py: FTBFS on ppc64el (arch:all-only src pkg): dh_auto_test: error: pybuild --test --test-pytest -i python{version} -p 3.9 returned exit code 13
> Well, the test is obviously broken and upstream currently can't be bothered > to fix > it on non-x86 targets. He will certainly have to do it at some point given > that ARM64 > is replacing more and more x86_64 systems, but I wouldn't bother, personally. so what is the best solution in order to have lmfit-py in bulleyes ?
Bug#976952: [Help] Re: lmfit-py: FTBFS on ppc64el (arch:all-only src pkg): dh_auto_test: error: pybuild --test --test-pytest -i python{version} -p 3.9 returned exit code 13
> Yes, good catch. The spec file for the openSUSE package has this [1]: so it does not fit with our policy: do not hide problems ;) The problem is that I do not have enougt time to investigate... on a porter box
Bug#976952: [Help] Re: lmfit-py: FTBFS on ppc64el (arch:all-only src pkg): dh_auto_test: error: pybuild --test --test-pytest -i python{version} -p 3.9 returned exit code 13
Hello looking at the Opensuze log, I can find this [ 93s] + pytest-3.8 --ignore=_build.python2 --ignore=_build.python3 --ignore=_build.pypy3 -v -k 'not speed and not (test_model_nan_policy or test_shgo_scipy_vs_lmfit_2)' [ 97s] = test session starts == does it means that the failling test on Debian is skipped during the build on OBS ? === FAILURES === _ TestUserDefiniedModel.test_model_nan_policy __ cheers Frederic
Bug#936609: cblas / gsl hint needed (Was: Bug#936609: Ported ghmm to Python3 but issues with clapack)
Hello Andreas, I just built ghmm by removing --with-gsl. It seems that the gsl implementation of blas conflict with the one provided in atlas. so --enable-gsl + --enable-atlas seems wrong... +--+ | Summary | +--+ Build Architecture: amd64 Build Type: binary Build-Space: 39004 Build-Time: 31 Distribution: UNRELEASED Host Architecture: amd64 Install-Time: 14 Job: /tmp/ghmm_0.9~rc3-3.dsc Machine Architecture: amd64 Package: ghmm Package-Time: 49 Source-Version: 0.9~rc3-3 Space: 39004 Status: successful Version: 0.9~rc3-3
Bug#976849: ITP: density-fitness -- Calculates per-residue electron density scores real-space R, real-space correlation coefficient, EDIAm, and OPIA
Hello, We maintain a Debian blends dedicated to photo and neutron facilities DebianPAN[1] via a dedicated team on salsa[2] do not hesitate to tell me in which category[2] you think density-fitness belongs cheers Frederic [2] https://salsa.debian.org/pan-team [1] https://salsa.debian.org/blends-team/pan [3] https://blends.debian.org/pan/tasks/
Bug#972017: pyfai ftbfs with python3.9 as supported python3 version
Looking at the build log of pyfai, I can find this https://launchpadlibrarian.net/501584970/buildlog_ubuntu-groovy-armhf.pyfai_0.19.0+dfsg1-3build1_BUILDING.txt.gz Selecting previously unselected package python3-silx. Preparing to unpack .../238-python3-silx_0.13.1+dfsg-1_armhf.deb ... Unpacking python3-silx (0.13.1+dfsg-1) ... Selecting previously unselected package python3-silx-dbg. Preparing to unpack .../239-python3-silx-dbg_0.13.1+dfsg-1_armhf.deb ... Unpacking python3-silx-dbg (0.13.1+dfsg-1) ... so it seems that this was rebuild with the wrong version on silx. not the one rebuilt with python3.9 support. I guess this by looking at other packages which containes Selecting previously unselected package python3-scipy-dbg:armhf. Preparing to unpack .../237-python3-scipy-dbg_1.5.2-2build1_armhf.deb ... Unpacking python3-scipy-dbg:armhf (1.5.2-2build1) ... the build1 version :)
Bug#972017: pyfai ftbfs with python3.9 as supported python3 version
Hello, the error comes from here. > ModuleNotFoundError: No module named 'silx.image.marchingsquares._mergeimpl' did you rebuilt it with a silx packages already rebuilt with python3.9 ?
Bug#946035: Request fo sponsor to uplad python-jsonrpc-server
Hello, It compile fine :), but I have a concern about the license. It is considère a good practice to use the same license for the debian directory and the source code. In your case, you chooses GPL-3+, but the code is MIT. Fred
Bug#949549: Not fixed
Hello Thomas, I am wondering if this is not a false positive. all the code is compiled with -D_FORTIFY_SOURCE=2 https://salsa.debian.org/science-team/tango/-/jobs/954394/raw Can you confirm that it is a false positive ? I am not that confident when it comes to hardening flags. for the record I checked also the Database and it is also affected by this. cheers
Bug#950094:
I tryed a new build and I end up with this error gpgv: unknown type of key resource 'trustedkeys.kbx' gpgv: keyblock resource '/tmp/dpkg-verify-sig.Wwlhs1jL/trustedkeys.kbx': General error gpgv: Signature made Mon Dec 16 20:17:19 2019 UTC gpgv:using RSA key E8FC295C86B8D7C049F97BA7A35DAFFBAD29E8DE gpgv: Can't check signature: No public key dpkg-source: warning: failed to verify signature on ./ipywidgets_6.0.0-6.dsc dpkg-source: info: extracting ipywidgets in /<> dpkg-source: info: unpacking ipywidgets_6.0.0.orig.tar.gz dpkg-source: info: unpacking ipywidgets_6.0.0-6.debian.tar.xz dpkg-source: info: using patch list from debian/patches/series dpkg-source: info: applying 0001-Unconditionally-import-setuptools-to-pick-up-depende.patch dpkg-source: info: applying 0002-Don-t-build-extension.js-in-widgetsnbextension-setup.patch dpkg-source: info: applying 0003-Use-local-MathJax.patch dpkg-source: info: applying 0004-Tweak-package.json-so-the-upstream-build-works-in-De.patch dpkg-source: info: applying 0005-Import-specific-jupyterlab-service-types-so-we-only-.patch dpkg-source: info: applying 0006-tsconfig-es2015-iterable.patch dpkg-source: error: pathname '/<>/debian/fakewebpack-unpacked/html2canvas' points outside source root (to '/usr/lib/nodejs/html2canvas') E: FAILED [dpkg-source died]
Bug#965964: ITP: geant4 -- physics simulation toolit from CERN
did you started from here ? https://salsa.debian.org/science-team/geant4 cheers Frederic
Bug#945467: Fix available (fwd)
you can look also at the CI, now that it works :) https://salsa.debian.org/science-team/veusz/pipelines/137494 Cheers Frederic
Bug#958322: Info received (Bug#958322: Info received ())
OK, with the previous NMU it works, so we can unbrand without problem the lzf library
Bug#958322: Info received ()
I am not sure at all of this in fact I red this closed bug https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=955456 I discovered that the NMU was not included in the git repository.
Bug#958322:
Hello, I tryed to unbrand the lzf library but it end up with failing unit tests I do not know the internals of liblzf. Timo Röhling can you have a look at this issue ? Cheers Frederic It seems that the embeded version of lzf and the one from your liblzf package are different. test_filter (bitshuffle.tests.test_h5filter.TestFilter) ... ERROR test_with_block_size (bitshuffle.tests.test_h5filter.TestFilter) ... ERROR test_with_compression (bitshuffle.tests.test_h5filter.TestFilter) ... ERROR test_plugins (bitshuffle.tests.test_h5plugin.TestFilterPlugins) ... ERROR test_regression (bitshuffle.tests.test_regression.TestAll) ... ok == ERROR: test_filter (bitshuffle.tests.test_h5filter.TestFilter) -- Traceback (most recent call last): File "/<>/.pybuild/cpython3_3.8_bitshuffle/build/bitshuffle/tests/test_h5filter.py", line 26, in test_filter f = h5py.File(fname) File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 387, in __init__ fid = make_fid(name, mode, userblock_size, File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 173, in make_fid fid = h5f.open(name, flags, fapl=fapl) File "h5py/_debian_h5py_serial/_objects.pyx", line 54, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/_objects.pyx", line 55, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/h5f.pyx", line 88, in h5py._debian_h5py_serial.h5f.open OSError: Unable to open file (unable to open file: name = 'tmp_test_filters.h5', errno = 2, error message = 'No such file or directory', flags = 0, o_flags = 0) == ERROR: test_with_block_size (bitshuffle.tests.test_h5filter.TestFilter) -- Traceback (most recent call last): File "/<>/.pybuild/cpython3_3.8_bitshuffle/build/bitshuffle/tests/test_h5filter.py", line 46, in test_with_block_size f = h5py.File(fname) File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 387, in __init__ fid = make_fid(name, mode, userblock_size, File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 173, in make_fid fid = h5f.open(name, flags, fapl=fapl) File "h5py/_debian_h5py_serial/_objects.pyx", line 54, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/_objects.pyx", line 55, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/h5f.pyx", line 88, in h5py._debian_h5py_serial.h5f.open OSError: Unable to open file (unable to open file: name = 'tmp_test_filters.h5', errno = 2, error message = 'No such file or directory', flags = 0, o_flags = 0) == ERROR: test_with_compression (bitshuffle.tests.test_h5filter.TestFilter) -- Traceback (most recent call last): File "/<>/.pybuild/cpython3_3.8_bitshuffle/build/bitshuffle/tests/test_h5filter.py", line 68, in test_with_compression f = h5py.File(fname) File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 387, in __init__ fid = make_fid(name, mode, userblock_size, File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 173, in make_fid fid = h5f.open(name, flags, fapl=fapl) File "h5py/_debian_h5py_serial/_objects.pyx", line 54, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/_objects.pyx", line 55, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/h5f.pyx", line 88, in h5py._debian_h5py_serial.h5f.open OSError: Unable to open file (unable to open file: name = 'tmp_test_filters.h5', errno = 2, error message = 'No such file or directory', flags = 0, o_flags = 0) == ERROR: test_plugins (bitshuffle.tests.test_h5plugin.TestFilterPlugins) -- Traceback (most recent call last): File "/<>/.pybuild/cpython3_3.8_bitshuffle/build/bitshuffle/tests/test_h5plugin.py", line 37, in test_plugins f = h5py.File(fname) File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 387, in __init__ fid = make_fid(name, mode, userblock_size, File "/usr/lib/python3/dist-packages/h5py/_debian_h5py_serial/_hl/files.py", line 173, in make_fid fid = h5f.open(name, flags, fapl=fapl) File "h5py/_debian_h5py_serial/_objects.pyx", line 54, in h5py._debian_h5py_serial._objects.with_phil.wrapper File "h5py/_debian_h5py_serial/_objects.pyx", line 55, in h5py._
Bug#956462: pkg-config for Qhull
Done :)) sorry for the delay De : Timo Röhling [t...@gaussglocke.de] Envoyé : samedi 25 avril 2020 14:06 À : PICCA Frederic-Emmanuel Cc : 956...@bugs.debian.org Objet : Re: pkg-config for Qhull Hi Frédéric! On 12.04.20 14:22, PICCA Frederic-Emmanuel wrote: > I will add both path in order to make the package backportable :) > Just pinging: Can you prepare an upload soon(ish), so I can continue with the Qhull transition? The "new" include path will work with the current 2015.2 release, too, so you won't break anything. Or you can use both paths as you suggested, which won't hurt either. Cheers Timo
Bug#956462: pymca: FTBFS with Qhull 2019.1
Do you provide a pkgconfig file with qhull ? De : debian-science-maintainers [debian-science-maintainers-bounces+picca=synchrotron-soleil...@alioth-lists.debian.net] de la part de Timo Röhling [t...@gaussglocke.de] Envoyé : samedi 11 avril 2020 19:08 À : sub...@bugs.debian.org Objet : Bug#956462: pymca: FTBFS with Qhull 2019.1 Source: pymca Version: 5.5.4+dfsg-1 Severity: important Tags: ftbfs, patch Dear maintainer, your package uses a deprecated include path for Qhull, which will no longer build with the latest release. Attached is a patch that will fix the build with the qhull package in experimental.
Bug#954352: pymca: Missing dependency "python3-scipy", probably in "python3-silx"
A work around for now is to install by hand apt install python3-scipy reassign -1 silx thanks
Bug#933484:
Hello, I just want to know what is the status of this backport ? thanks Frederic
Bug#936924: libsvm: Python2 removal in sid/bullseye - reopen 936924
Hello, if it is like for my ufo-core package, this could be due to a script file with a shebang using python instead of python3 Cheers Fred
Bug#938743: ufo-core: Python2 removal in sid/bullseye - reopen 938743
Maybe this is due to this picca@cush:~/Debian/ufo-core/ufo-core/bin$ rgrep python * ufo-mkfilter.in:#!/usr/bin/python ufo-prof:#!/usr/bin/env python I will replace python -> python3 and see what is going on
Bug#938743: ufo-core: Python2 removal in sid/bullseye - reopen 938743
Hello Sandro this is strange because, I have this in the control file Package: libufo-bin Architecture: any Depends: ${misc:Depends}, ${python3:Depends}, ${shlibs:Depends} Suggests: ufo-core-doc Description: Library for high-performance, GPU-based computing - tools The UFO data processing framework is a C library suited to build general purpose streams data processing on heterogeneous architectures such as CPUs, GPUs or clusters. It is extensively used at the Karlsruhe Institute of Technology for Ultra-fast X-ray Imaging (radiography, tomography and laminography). . A gobject-instrospection binding is also provided to write scripts or user interfaces. . This package contains binaries to run JSON descriptions of task graphs. Package: libufo-data So it seems that this python dependency comes from python3:Depends. is it normal ? Cheers Fred
Bug#943786: lmfit-py: failing tests with python3.8
Hello andreas, I will try to find some time during these holydays in order to upload the 1.0.0 version :) cheers Fred and happy new year. De : debian-science-maintainers [debian-science-maintainers-bounces+picca=synchrotron-soleil...@alioth-lists.debian.net] de la part de Andreas Tille [andr...@an3as.eu] Envoyé : dimanche 22 décembre 2019 10:48 À : PICCA Frederic-Emmanuel Cc : 943...@bugs.debian.org; MARIE Alexandre Objet : Bug#943786: lmfit-py: failing tests with python3.8 On Sun, Dec 22, 2019 at 07:54:23AM +, PICCA Frederic-Emmanuel wrote: > Hello andreas, In fact we were wayting for the pacakging of ipywidget 7.x > the jupyter-sphinx extension expected by lmfit-py require a newer version of > ipywidget. > > So maybe the best solution for now is to not produce the documentation until > this dependency is ok. Everything is fine for me. Just push a change that builds and upload (or ping me for sponsoring). I was just wondering about fixes in Git that were not uploaded. Kind regards Andreas. -- http://fam-tille.de -- debian-science-maintainers mailing list debian-science-maintain...@alioth-lists.debian.net https://alioth-lists.debian.net/cgi-bin/mailman/listinfo/debian-science-maintainers
Bug#943786: lmfit-py: failing tests with python3.8
Hello andreas, In fact we were wayting for the pacakging of ipywidget 7.x the jupyter-sphinx extension expected by lmfit-py require a newer version of ipywidget. So maybe the best solution for now is to not produce the documentation until this dependency is ok. cheers Frederic
Bug#946422: silx: autopkgtest regression: pocl error
looking in picca@sixs7:~/Debian/silx/silx/silx/opencl/test/test_addition.py def setUp(self): if ocl is None: return self.shape = 4096 self.data = numpy.random.random(self.shape).astype(numpy.float32) self.d_array_img = pyopencl.array.to_device(self.queue, self.data) self.d_array_5 = pyopencl.array.zeros_like(self.d_array_img) - 5 self.program = pyopencl.Program(self.ctx, get_opencl_code("addition")).build() I found that commenting this line # self.d_array_5 = pyopencl.array.zeros_like(self.d_array_img) - 5 remove the pocl issue. I remove everythings from the unit test. @unittest.skipUnless(ocl, "pyopencl is missing") def test_add(self): self.assetTrue(True) that means, only ther setUp and the tearDown are done. with the line uncomment (sid_amd64-dchroot)picca@barriere:~$ PYOPENCL_COMPILER_OUTPUT=1 PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v pocl error: lt_dlopen("(null)") or lt_dlsym() failed with 'can't close resident module'. note: missing symbols in the kernel binary might be reported as 'file not found' errors. Aborted with the line commented (sid_amd64-dchroot)picca@barriere:~$ PYOPENCL_COMPILER_OUTPUT=1 PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v .Maximum valid workgroup size 0 on device -- Ran 1 test in 0.013s OK Test suite succeeded If now I do not import silx.io before there is no issue with or without the commented line (sid_amd64-dchroot)picca@barriere:~$ PYOPENCL_COMPILER_OUTPUT=1 PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v .Maximum valid workgroup size 0 on device -- Ran 1 test in 0.021s OK Test suite succeeded So what is going on when executing this line ??? self.d_array_5 = pyopencl.array.zeros_like(self.d_array_img) - 5
Bug#946422: silx: autopkgtest regression: pocl error
I decided to concentrate myself on one opencl test (addition) So I deactivated all other test by commenting the test in silx/opencl/__init__.py If I do not import silxs.io, this test works (sid_amd64-dchroot)picca@barriere:~$ PYOPENCL_COMPILER_OUTPUT=1 PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v .Maximum valid workgroup size 2048 on device -- Ran 1 test in 0.030s OK If I add silxs.io, it fails (sid_amd64-dchroot)picca@barriere:~$ PYOPENCL_COMPILER_OUTPUT=1 PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v pocl error: lt_dlopen("(null)") or lt_dlsym() failed with 'can't close resident module'. note: missing symbols in the kernel binary might be reported as 'file not found' errors. Aborted so this test is ok in order to investigate this issue.
Bug#946422: silx: autopkgtest regression: pocl error
With the silx.io import I have this (sid_amd64-dchroot)picca@barriere:~$ PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py pocl error: lt_dlopen("(null)") or lt_dlsym() failed with 'can't close resident module'. note: missing symbols in the kernel binary might be reported as 'file not found' errors. without, I have this (sid_amd64-dchroot)picca@barriere:~$ PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v .Maximum valid workgroup size 2048 on device FThe gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. .The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. .The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. ../home/picca/silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build/silx/opencl/test/test_linalg.py:69: FutureWarning: Using a non-tuple sequence for multidimensional indexing is deprecated; use `arr[tuple(seq)]` instead of `arr[seq]`. In the future this will be interpreted as an array index, `arr[np.array(seq)]`, which will result either in an error or a different result. gradient[slice_all] = np.diff(img, axis=d) 4096 ...7 warnings generated. /usr/lib/python3/dist-packages/pyopencl/__init__.py:235: CompilerWarning: Non-empty compiler output encountered. Set the environment variable PYOPENCL_COMPILER_OUTPUT=1 to see more. "to see more.", CompilerWarning) ..7 warnings generated. .. == FAIL: test_medfilt (silx.opencl.test.test_medfilt.TestMedianFilter) -- Traceback (most recent call last): File "/home/picca/silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build/silx/opencl/test/test_medfilt.py", line 115, in test_medfilt self.assertEqual(r.error, 0, 'Results are correct') AssertionError: 3.4028235e+38 != 0 : Results are correct -- Ran 217 tests in 175.525s FAILED (failures=1, skipped=48) I will run it with the PYOPENCL_COMPILEr_OUTPUT=1 to check that compiler warning (sid_amd64-dchroot)picca@barriere:~$ PYOPENCL_COMPILER_OUTPUT=1 PYTHONPATH=silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build python3 test.py -v .Maximum valid workgroup size 2048 on device FThe gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. .The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. .The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. The gpyfft module was not found. The Fourier transforms will be done on CPU. For more performances, it is advised to install gpyfft. ../home/picca/silx-0.11.0+dfsg/.pybuild/cpython3_3.7_silx/build/silx/opencl/test/test_linalg.py:69: FutureWarning: Using a non-tuple sequence for multidimensional indexing is deprecated; use `arr[tuple(seq)]` instead of `arr[seq]`. In the future this will be
Bug#946422: silx: autopkgtest regression: pocl error
not better test cpp engine for medfilt2d ... ok testOpenCLMedFilt2d (silx.image.test.test_medianfilter.TestMedianFilterEngines) test cpp engine for medfilt2d ... pocl error: lt_dlopen("(null)") or lt_dlsym() failed with 'can't close resident module'. note: missing symbols in the kernel binary might be reported as 'file not found' errors. The next step will be to ding in the code and find the culprite.
Bug#946422: silx: autopkgtest regression: pocl error
It seems that this test does not PASS @unittest.skipUnless(ocl, "PyOpenCl is missing") def testOpenCLMedFilt2d(self): """test cpp engine for medfilt2d""" res = medianfilter.medfilt2d( image=TestMedianFilterEngines.IMG, kernel_size=TestMedianFilterEngines.KERNEL, engine='opencl') self.assertTrue(numpy.array_equal(res, TestMedianFilterEngines.IMG)) testOpenCLMedFilt2d (silx.image.test.test_medianfilter.TestMedianFilterEngines) test cpp engine for medfilt2d ... pocl error: lt_dlopen("(null)") or lt_dlsym() failed with 'can't close resident module'. note: missing symbols in the kernel binary might be reported as 'file not found' errors. Aborted E: pybuild pybuild:341: test: plugin custom failed with: exit code=134: env PYTHONPATH=/home/picca/silx-0.11.0+dfsg/.pybuild/cpython3_3.8_silx/build WITH_QT_TEST=False xvfb-run -a --server-args="-screen 0 1024x768x24" python3.8 run_tests.py -vv --installed dh_auto_test: pybuild --test -i python{version} -p "3.8 3.7" -s custom "--test-args=env PYTHONPATH={build_dir} WITH_QT_TEST=False xvfb-run -a --server-args=\"-screen 0 1024x768x24\" {interpreter} run_tests.py -vv --installed" returned exit code 13 make[1]: *** [debian/rules:70: override_dh_auto_test] Error 255 make[1]: Leaving directory '/home/picca/silx-0.11.0+dfsg' make: *** [debian/rules:27: build] Error 2 the code of medfilt2d is there def medfilt2d(image, kernel_size=3, engine='cpp'): """Apply a median filter on an image. This median filter is using a 'nearest' padding for values past the array edges. If you want more padding options or functionalities for the median filter (conditional filter for example) please have a look at :mod:`silx.math.medianfilter`. :param numpy.ndarray image: the 2D array for which we want to apply the median filter. :param kernel_size: the dimension of the kernel. Kernel size must be odd. If a scalar is given, then it is used as the size in both dimension. Default: (3, 3) :type kernel_size: A int or a list of 2 int (kernel_height, kernel_width) :param engine: the type of implementation to use. Valid values are: 'cpp' (default) and 'opencl' :returns: the array with the median value for each pixel. .. note:: if the opencl implementation is requested but is not present or fails, the cpp implementation is called. """ if engine not in MEDFILT_ENGINES: err = 'silx doesn\'t have an implementation for the requested engine: ' err += '%s' % engine raise ValueError(err) if len(image.shape) is not 2: raise ValueError('medfilt2d deals with arrays of dimension 2 only') if engine == 'cpp': return medianfilter_cpp.medfilt(data=image, kernel_size=kernel_size, conditional=False) elif engine == 'opencl': if medfilt_opencl is None: wrn = 'opencl median filter not available. ' wrn += 'Launching cpp implementation.' _logger.warning(wrn) # instead call the cpp implementation return medianfilter_cpp.medfilt(data=image, kernel_size=kernel_size, conditional=False) else: try: medianfilter = medfilt_opencl.MedianFilter2D(image.shape, devicetype="gpu") res = medianfilter.medfilt2d(image, kernel_size) except(RuntimeError, MemoryError, ImportError): wrn = 'Exception occured in opencl median filter. ' wrn += 'To get more information see debug log.' wrn += 'Launching cpp implementation.' _logger.warning(wrn) _logger.debug("median filter - openCL implementation issue.", exc_info=True) # instead call the cpp implementation res = medianfilter_cpp.medfilt(data=image, kernel_size=kernel_size, conditional=False) return res in our case we have engine = 'opencl' and no warning message, so medfil_opencl should not be None. it comes from here from silx.opencl import medfilt as medfilt_opencl In this code we have :param devicetype: type of device, can be "CPU", "GPU", "ACC" or "ALL" So let's do a first test by replacing gpu by cpu to see if it change something during the test.
Bug#938116:
Use salsa-ci, python-qtconsole FTBFS due to pyzmq https://salsa.debian.org/python-team/modules/python-qtconsole/-/jobs/435758