Re: [ClusterLabs] pcsd web interface not working on EL 9.3
On 19/02/2024 09:06, Strahil Nikolov via Users wrote: Hi All, Is there a specific setup I missed in order to setup the web interface ? Usually, you just login with the hacluster user on https://fqdn:2224 but when I do a curl, I get an empty response. Best Regards, Strahil Nikolov ___ Was it giving out some stuff before? I've never did curl on it. Won't guess why do that. Yes it is _empty_for me too - though I use it with reverse proxy. While _main_ URL curl returns no content this does get: -> $ curl https://pcs-ha.mine.priv/ui/ ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] clone_op_key pcmk__notify_key - Triggered fatal assertion
Hi guys. Everything seems to be working a ok yet pacemakers logs ... error: clone_op_key: Triggered fatal assertion at pcmk_graph_producer.c:207 : (n_type != NULL) && (n_task != NULL) error: pcmk__notify_key: Triggered fatal assertion at actions.c:187 : op_type != NULL error: clone_op_key: Triggered fatal assertion at pcmk_graph_producer.c:207 : (n_type != NULL) && (n_task != NULL) error: pcmk__notify_key: Triggered fatal assertion at actions.c:187 : op_type != NULL ... error: pcmk__create_history_xml: Triggered fatal assertion at pcmk_sched_actions.c:1163 : n_type != NULL error: pcmk__create_history_xml: Triggered fatal assertion at pcmk_sched_actions.c:1164 : n_task != NULL error: pcmk__notify_key: Triggered fatal assertion at actions.c:187 : op_type != NULL ... Looks critical, is it it - would you know? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
VGA - disable display functions ?
Hi guys. Would you know how to disable/tweak bits in kernel/module so display(& audio) is completely disabled? I have AMD RX 6600 which I hope I could give completely to ROCm for computing. many thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] trigger something at ?
On 31/01/2024 16:37, lejeczek via Users wrote: On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:02:12 +0100 lejeczek via Users wrote: On 29/01/2024 17:22, Ken Gaillot wrote: On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote: Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L. You could use node status alerts for that, but it's risky for alert agents to change the configuration (since that may result in more alerts and potentially some sort of infinite loop). Pacemaker has no concept of a full cluster start/stop, only node start/stop. You could approximate that by checking whether the node receiving the alert is the only active node. Another possibility would be to write a resource agent that does what you want and order everything else after it. However it's even more risky for a resource agent to modify the configuration. Finally you could write a systemd unit to do what you want and order it after pacemaker. What's wrong with leaving the constraints permanently configured? yes, that would be for a node start/stop I struggle with using constraints to move pgsql (PAF) master onto a given node - seems that co/locating paf's master results in troubles (replication brakes) at/after node shutdown/reboot (not always, but way too often) What? What's wrong with colocating PAF's masters exactly? How does it brake any replication? What's these constraints you are dealing with? Could you share your configuration? Constraints beyond/above of what is required by PAF agent itself, say... you have multiple pgSQL cluster with PAF - thus multiple (separate, for each pgSQL cluster) masters and you want to spread/balance those across HA cluster (or in other words - avoid having more that 1 pgsql master per HA node) These below, I've tried, those move the master onto chosen node but.. then the issues I mentioned. -> $ pcs constraint location PGSQL-PAF-5438-clone prefers ubusrv1=1002 or -> $ pcs constraint colocation set PGSQL-PAF-5435-clone PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master require-all=false setoptions score=-1000 Wanted to share an observation - not a measurement of anything, I did not take those - of different, latest pgSQL version which I put in place of version 14 which I've been using all this time. (also with that upgrade - from Postgres own repos - came update of PAF) So, with pgSQL ver. 16 and the same of everything else - now paf/pgSQL resources behave a lot lot better, survives just fine all those cases - with ! extra constraints of course - where previously it had replication failures. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] colocation constraint - do I get it all wrong?
On 01/01/2024 18:28, Ken Gaillot wrote: On Fri, 2023-12-22 at 17:02 +0100, lejeczek via Users wrote: hi guys. I have a colocation constraint: -> $ pcs constraint ref DHCPD Resource: DHCPD colocation-DHCPD-GATEWAY-NM-link-INFINITY and the trouble is... I thought DHCPD is to follow GATEWAY-NM-link, always! If that is true that I see very strange behavior, namely. When there is an issue with DHCPD resource, cannot be started, then GATEWAY-NM-link gets tossed around by the cluster. Is that normal & expected - is my understanding of _colocation_ completely wrong - or my cluster is indeed "broken"? many thanks, L. Pacemaker considers the preferences of colocated resources when assigning a resource to a node, to ensure that as many resources as possible can run. So if a colocated resource becomes unable to run on a node, the primary resource might move to allow the colocated resource to run. So what is the way to "fix" this - is it simply low/er score for such constraint? In my case _dhcpd_ is important but if fails sometimes as it's often tampered with, so... make _dhcpd_ flow gateway_link but just fail _dhcp_ (it it keeps failing) and leave _gateway_link_ alone if/where it's good. Or perhaps there a global config/param for whole cluster behaviour? ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] to heal gfid which is not there -- IGNORE wrong list
On 04/02/2024 12:57, lejeczek via Users wrote: hi guys. So, I've manged to make my volume go haywire, here: -> $ gluster volume heal VMAIL info Brick 10.1.1.100:/devs/00.GLUSTERs/VMAIL Status: Connected Number of entries: 0 Brick 10.1.1.101:/devs/00.GLUSTERs/VMAIL /dovecot-uidlist Status: Connected Number of entries: 1 Brick 10.1.1.99:/devs/00.GLUSTERs/VMAIL-arbiter Status: Connected Number of entries: 0 That _gfid_ nor _dovecot-uidlist_ do not exist on any of the bricks, but I've noticed that all bricks have: .glusterfs-anonymous-inode-462a1850-a31a-4a17-934d-26f3996dc9b8 which is an empty dir. That volume I fear won't heal on its own and might need some intervention - would you know how/where to start? many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home:https://www.clusterlabs.org/ ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] to heal gfid which is not there
hi guys. So, I've manged to make my volume go haywire, here: -> $ gluster volume heal VMAIL info Brick 10.1.1.100:/devs/00.GLUSTERs/VMAIL Status: Connected Number of entries: 0 Brick 10.1.1.101:/devs/00.GLUSTERs/VMAIL /dovecot-uidlist Status: Connected Number of entries: 1 Brick 10.1.1.99:/devs/00.GLUSTERs/VMAIL-arbiter Status: Connected Number of entries: 0 That _gfid_ nor _dovecot-uidlist_ do not exist on any of the bricks, but I've noticed that all bricks have: .glusterfs-anonymous-inode-462a1850-a31a-4a17-934d-26f3996dc9b8 which is an empty dir. That volume I fear won't heal on its own and might need some intervention - would you know how/where to start? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] trigger something at ?
On 01/02/2024 15:02, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 18:23:40 +0100 lejeczek via Users wrote: On 31/01/2024 17:13, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:37:21 +0100 lejeczek via Users wrote: On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:02:12 +0100 lejeczek via Users wrote: On 29/01/2024 17:22, Ken Gaillot wrote: On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote: Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L. You could use node status alerts for that, but it's risky for alert agents to change the configuration (since that may result in more alerts and potentially some sort of infinite loop). Pacemaker has no concept of a full cluster start/stop, only node start/stop. You could approximate that by checking whether the node receiving the alert is the only active node. Another possibility would be to write a resource agent that does what you want and order everything else after it. However it's even more risky for a resource agent to modify the configuration. Finally you could write a systemd unit to do what you want and order it after pacemaker. What's wrong with leaving the constraints permanently configured? yes, that would be for a node start/stop I struggle with using constraints to move pgsql (PAF) master onto a given node - seems that co/locating paf's master results in troubles (replication brakes) at/after node shutdown/reboot (not always, but way too often) What? What's wrong with colocating PAF's masters exactly? How does it brake any replication? What's these constraints you are dealing with? Could you share your configuration? Constraints beyond/above of what is required by PAF agent itself, say... you have multiple pgSQL cluster with PAF - thus multiple (separate, for each pgSQL cluster) masters and you want to spread/balance those across HA cluster (or in other words - avoid having more that 1 pgsql master per HA node) ok These below, I've tried, those move the master onto chosen node but.. then the issues I mentioned. You just mentioned it breaks the replication, but there so little information about your architecture and configuration, it's impossible to imagine how this could break the replication. Could you add details about the issues ? -> $ pcs constraint location PGSQL-PAF-5438-clone prefers ubusrv1=1002 or -> $ pcs constraint colocation set PGSQL-PAF-5435-clone PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master require-all=false setoptions score=-1000 I suppose "collocation" constraint is the way to go, not the "location" one. This should be easy to replicate, 3 x VMs, Ubuntu 22.04 in my case No, this is not easy to replicate. I have no idea how you setup your PostgreSQL replication, neither I have your full pacemaker configuration. Please provide either detailed setupS and/or ansible and/or terraform and/or vagrant, then a detailed scenario showing how it breaks. This is how you can help and motivate devs to reproduce your issue and work on it. I will not try to poke around for hours until I find an issue that might not even be the same than yours. How about you start with the basics - strange inclination to complicate things when they are not, I hear from you - that's what I did while "stumbled" upon these "issues" How about just: a) do vanilla-default pgSQL in Ubuntu (or perhaps any other OS of your choice), I use _pg_createcluster_ b) follow PAF official guide (a single PAF resource should suffice) Have a healthy pgSQL cluster, OS _reboot_ nodes - play with that, all should be ok, moving around/electing master should work a ok. Then... add, play with "additional" co/location constraints, then OS reboots,- things should begin braking. I have 3-node HA cluster & 3-node PAF resource = 1 master + 2 slaves. Only thing I deliberately set, to alleviate pgsql replication was _wal_keep_size_ - I increased that, but this is subjective. It's fine with me if you don't feel like doing this. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] trigger something at ?
On 31/01/2024 18:11, Ken Gaillot wrote: On Wed, 2024-01-31 at 16:37 +0100, lejeczek via Users wrote: On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:02:12 +0100 lejeczek via Users wrote: On 29/01/2024 17:22, Ken Gaillot wrote: On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote: Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L. You could use node status alerts for that, but it's risky for alert agents to change the configuration (since that may result in more alerts and potentially some sort of infinite loop). Pacemaker has no concept of a full cluster start/stop, only node start/stop. You could approximate that by checking whether the node receiving the alert is the only active node. Another possibility would be to write a resource agent that does what you want and order everything else after it. However it's even more risky for a resource agent to modify the configuration. Finally you could write a systemd unit to do what you want and order it after pacemaker. What's wrong with leaving the constraints permanently configured? yes, that would be for a node start/stop I struggle with using constraints to move pgsql (PAF) master onto a given node - seems that co/locating paf's master results in troubles (replication brakes) at/after node shutdown/reboot (not always, but way too often) What? What's wrong with colocating PAF's masters exactly? How does it brake any replication? What's these constraints you are dealing with? Could you share your configuration? Constraints beyond/above of what is required by PAF agent itself, say... you have multiple pgSQL cluster with PAF - thus multiple (separate, for each pgSQL cluster) masters and you want to spread/balance those across HA cluster (or in other words - avoid having more that 1 pgsql master per HA node) These below, I've tried, those move the master onto chosen node but.. then the issues I mentioned. -> $ pcs constraint location PGSQL-PAF-5438-clone prefers ubusrv1=1002 or -> $ pcs constraint colocation set PGSQL-PAF-5435-clone PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master require-all=false setoptions score=-1000 Anti-colocation sets tend to be tricky currently -- if the first resource can't be assigned to a node, none of them can. We have an idea for a better implementation: https://projects.clusterlabs.org/T383 In the meantime, a possible workaround is to use placement- strategy=balanced and define utilization for the clones only. The promoted roles will each get a slight additional utilization, and the cluster should spread them out across nodes whenever possible. I don't know if that will avoid the replication issues but it may be worth a try. using _balanced_ causes a small mayhem to PAF/pgsql: -> $ pcs property Cluster Properties: REDIS-6380_REPL_INFO: ubusrv3 REDIS-6381_REPL_INFO: ubusrv2 REDIS-6382_REPL_INFO: ubusrv2 REDIS-6385_REPL_INFO: ubusrv1 REDIS_REPL_INFO: ubusrv1 cluster-infrastructure: corosync cluster-name: ubusrv dc-version: 2.1.2-ada5c3b36e2 have-watchdog: false last-lrm-refresh: 1706711588 placement-strategy: default stonith-enabled: false -> $ pcs resource utilization PGSQL-PAF-5438 cpu="20" -> $ pcs property set placement-strategy=balanced # when resource stops: I change it back: -> $ pcs property set placement-strategy=default and pgSQL/paf works again I've not used _utilization_ nor _placement-strategy_ before, thus chance that I'm missing something is solid. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] trigger something at ?
On 31/01/2024 17:13, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:37:21 +0100 lejeczek via Users wrote: On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:02:12 +0100 lejeczek via Users wrote: On 29/01/2024 17:22, Ken Gaillot wrote: On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote: Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L. You could use node status alerts for that, but it's risky for alert agents to change the configuration (since that may result in more alerts and potentially some sort of infinite loop). Pacemaker has no concept of a full cluster start/stop, only node start/stop. You could approximate that by checking whether the node receiving the alert is the only active node. Another possibility would be to write a resource agent that does what you want and order everything else after it. However it's even more risky for a resource agent to modify the configuration. Finally you could write a systemd unit to do what you want and order it after pacemaker. What's wrong with leaving the constraints permanently configured? yes, that would be for a node start/stop I struggle with using constraints to move pgsql (PAF) master onto a given node - seems that co/locating paf's master results in troubles (replication brakes) at/after node shutdown/reboot (not always, but way too often) What? What's wrong with colocating PAF's masters exactly? How does it brake any replication? What's these constraints you are dealing with? Could you share your configuration? Constraints beyond/above of what is required by PAF agent itself, say... you have multiple pgSQL cluster with PAF - thus multiple (separate, for each pgSQL cluster) masters and you want to spread/balance those across HA cluster (or in other words - avoid having more that 1 pgsql master per HA node) ok These below, I've tried, those move the master onto chosen node but.. then the issues I mentioned. You just mentioned it breaks the replication, but there so little information about your architecture and configuration, it's impossible to imagine how this could break the replication. Could you add details about the issues ? -> $ pcs constraint location PGSQL-PAF-5438-clone prefers ubusrv1=1002 or -> $ pcs constraint colocation set PGSQL-PAF-5435-clone PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master require-all=false setoptions score=-1000 I suppose "collocation" constraint is the way to go, not the "location" one. This should be easy to replicate, 3 x VMs, Ubuntu 22.04 in my case -> $ pcs resource config PGSQL-PAF-5438-clone Clone: PGSQL-PAF-5438-clone Meta Attrs: failure-timeout=60s master-max=1 notify=true promotable=true Resource: PGSQL-PAF-5438 (class=ocf provider=heartbeat type=pgsqlms) Attributes: bindir=/usr/lib/postgresql/16/bin datadir=/var/lib/postgresql/16/paf-5438 maxlag=1000 pgdata=/etc/postgresql/16/paf-5438 pgport=5438 Operations: demote interval=0s timeout=120s (PGSQL-PAF-5438-demote-interval-0s) methods interval=0s timeout=5 (PGSQL-PAF-5438-methods-interval-0s) monitor interval=15s role=Master timeout=10s (PGSQL-PAF-5438-monitor-interval-15s) monitor interval=16s role=Slave timeout=10s (PGSQL-PAF-5438-monitor-interval-16s) notify interval=0s timeout=60s (PGSQL-PAF-5438-notify-interval-0s) promote interval=0s timeout=30s (PGSQL-PAF-5438-promote-interval-0s) reload interval=0s timeout=20 (PGSQL-PAF-5438-reload-interval-0s) start interval=0s timeout=60s (PGSQL-PAF-5438-start-interval-0s) stop interval=0s timeout=60s (PGSQL-PAF-5438-stop-interval-0s) so, regarding PAF - 1 master + 2 slaves, have a healthy pqSQL/PAF cluster to begin with, then make resource prefer a specific node (with simplest variant of constraints I tried): -> $ pcs constraint location PGSQL-PAF-5438-clone prefers ubusrv1=1002 and play with it, rebooting node(s) with OS' _reboot_ I at some point, get HA/resource unable to start pgSQL, unable to elect a master (logs saying with replication broken) and I have to "fix" pgSQL cluster outside of PAF, using _pg_basebackup_ ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] trigger something at ?
On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote: On Wed, 31 Jan 2024 16:02:12 +0100 lejeczek via Users wrote: On 29/01/2024 17:22, Ken Gaillot wrote: On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote: Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L. You could use node status alerts for that, but it's risky for alert agents to change the configuration (since that may result in more alerts and potentially some sort of infinite loop). Pacemaker has no concept of a full cluster start/stop, only node start/stop. You could approximate that by checking whether the node receiving the alert is the only active node. Another possibility would be to write a resource agent that does what you want and order everything else after it. However it's even more risky for a resource agent to modify the configuration. Finally you could write a systemd unit to do what you want and order it after pacemaker. What's wrong with leaving the constraints permanently configured? yes, that would be for a node start/stop I struggle with using constraints to move pgsql (PAF) master onto a given node - seems that co/locating paf's master results in troubles (replication brakes) at/after node shutdown/reboot (not always, but way too often) What? What's wrong with colocating PAF's masters exactly? How does it brake any replication? What's these constraints you are dealing with? Could you share your configuration? Constraints beyond/above of what is required by PAF agent itself, say... you have multiple pgSQL cluster with PAF - thus multiple (separate, for each pgSQL cluster) masters and you want to spread/balance those across HA cluster (or in other words - avoid having more that 1 pgsql master per HA node) These below, I've tried, those move the master onto chosen node but.. then the issues I mentioned. -> $ pcs constraint location PGSQL-PAF-5438-clone prefers ubusrv1=1002 or -> $ pcs constraint colocation set PGSQL-PAF-5435-clone PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master require-all=false setoptions score=-1000 ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] trigger something at ?
On 29/01/2024 17:22, Ken Gaillot wrote: On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote: Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L. You could use node status alerts for that, but it's risky for alert agents to change the configuration (since that may result in more alerts and potentially some sort of infinite loop). Pacemaker has no concept of a full cluster start/stop, only node start/stop. You could approximate that by checking whether the node receiving the alert is the only active node. Another possibility would be to write a resource agent that does what you want and order everything else after it. However it's even more risky for a resource agent to modify the configuration. Finally you could write a systemd unit to do what you want and order it after pacemaker. What's wrong with leaving the constraints permanently configured? yes, that would be for a node start/stop I struggle with using constraints to move pgsql (PAF) master onto a given node - seems that co/locating paf's master results in troubles (replication brakes) at/after node shutdown/reboot (not always, but way too often) Ideally I'm hoping that: at node stop, stopping node could check if it's PAF's master and if yes so then remove given constraints ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] trigger something at ?
Hi guys. Is it possible to trigger some... action - I'm thinking specifically at shutdown/start. If not within the cluster then - if you do that - perhaps outside. I would like to create/remove constraints, when cluster starts & stops, respectively. many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] replica vol when a peer is lost/offed & qcow2 ?
Hi guys. I wonder if you might have any tips/tweaks for volume/cluster to make it more resilient? accommodating? to qcow2 files but! when a peer is lots or missing? I have 3-peer cluster/volume: 2 + 1 arbiter & my experience is such, that when all is good then.. well, all is good, but... when one peers was lots - even if it's the arbiter - then VMs begin to suffer from & to report, their filesystems, errors. Perhaps it's not a volume's properties/config but whole cluster? Or perhaps 3-peer vol is not good for things such a qcow2s? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Lenovo ThinkBook 16 G6 touchpad - no right click ?
Hi guys. Anybody here on a Thinkbook 16 G6? I just got one myself and it seems that the right-click is not there - that should not be new laptop with mechanical-electrical damage, right? many thanks, L. -- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
NVMe sector size ?
Hi guys. I wonder if anybody played with both sector sizes 512 & 4086 and if yes so - also had some test results / thoughts to share? There are bits about it - bit confusing & a mixed-bag - over the net but I failed to find anything Fedora-specific. thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
color profiles for screens/monitors
Hi guys How do we load & tweak color profiles in Fedora (without using external sensor-devices) ? My Lenovo Thinkbook built-in screen has different (and not adjustable in BIOS) color temperature what my external DELL shows - they are quite different thus annoying. many thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ClusterLabs] colocation constraint - do I get it all wrong?
hi guys. I have a colocation constraint: -> $ pcs constraint ref DHCPD Resource: DHCPD colocation-DHCPD-GATEWAY-NM-link-INFINITY and the trouble is... I thought DHCPD is to follow GATEWAY-NM-link, always! If that is true that I see very strange behavior, namely. When there is an issue with DHCPD resource, cannot be started, then GATEWAY-NM-link gets tossed around by the cluster. Is that normal & expected - is my understanding of _colocation_ completely wrong - or my cluster is indeed "broken"? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] colocate Redis - weird
On 19/12/2023 19:13, lejeczek via Users wrote: hi guys, Is this below not the weirdest thing? -> $ pcs constraint ref PGSQL-PAF-5435 Resource: PGSQL-PAF-5435 colocation-HA-10-1-1-84-PGSQL-PAF-5435-clone-INFINITY colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory-1 colocation_set_PePePe Here Redis master should folow pgSQL master. Which such constraint: -> $ pcs resource status PGSQL-PAF-5435 * Clone Set: PGSQL-PAF-5435-clone [PGSQL-PAF-5435] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs resource status REDIS-6385-clone * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable): * Unpromoted: [ ubusrv1 ubusrv2 ubusrv3 ] If I remove that constrain: -> $ pcs constraint delete colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY -> $ pcs resource status REDIS-6385-clone * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] and ! I can manually move Redis master around, master moves to each server just fine. I again, add that constraint: -> $ pcs constraint colocation add master REDIS-6385-clone with master PGSQL-PAF-5435-clone and the same... What there might be about that one node - resource removed, created anew and cluster insists on keeping master there. I can manually move the master anywhere but if I _clear_ the resource, no constraints then cluster move it back to the same node. I wonder about: a) "transient" node attrs & b) if this cluster is somewhat broken. On a) - can we read more about those somewhere?(not the code/internals) thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] colocate Redis - weird
hi guys, Is this below not the weirdest thing? -> $ pcs constraint ref PGSQL-PAF-5435 Resource: PGSQL-PAF-5435 colocation-HA-10-1-1-84-PGSQL-PAF-5435-clone-INFINITY colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory-1 colocation_set_PePePe Here Redis master should folow pgSQL master. Which such constraint: -> $ pcs resource status PGSQL-PAF-5435 * Clone Set: PGSQL-PAF-5435-clone [PGSQL-PAF-5435] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs resource status REDIS-6385-clone * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable): * Unpromoted: [ ubusrv1 ubusrv2 ubusrv3 ] If I remove that constrain: -> $ pcs constraint delete colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY -> $ pcs resource status REDIS-6385-clone * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] and ! I can manually move Redis master around, master moves to each server just fine. I again, add that constraint: -> $ pcs constraint colocation add master REDIS-6385-clone with master PGSQL-PAF-5435-clone and the same... ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: Gnome's terminal/console fonts ?
On 16/12/2023 21:21, George N. White III wrote: On Sat, Dec 16, 2023 at 3:36 AM lejeczek via users wrote: Hi guys. My Gnome's default terminal does only select from _mono_ type of fonts & I remember it could choose any font. Alternative solutions, such a _terminator_ do allow to select any font. In Gnome's terminal 'font scaling' also seems "broken" - certainly on my Fedora - I select "Liberation Mono" and I start size at 6 and keep increasing by small fraction and the font does not seem changing until suddenly, at some given size it "jumps" the size & ... shape too. I do not suppose it's my environment - I've been upgrading my Fedora since... forever, as opposed to "clean" install - and I think it's rather "broken" terminal, but.. I wanted to consult other Fedorians to be sure. Using a fresh install of F39, in Gnome Terminal "Preferences" I selected the "Unnamed" Profile, set "Custom font" to Liberation Mono, and click on the "Custom Font" bar to get a "Choose A Terminal Font" window. The size changes in the window and the terminal also changes to match when I close the "Chose ..." window. I have TrueType Liberation fonts: % fc-list | grep LiberationMono /usr/share/fonts/liberation-mono/LiberationMono-Regular.ttf: Liberation Mono:style=Regular /usr/share/fonts/liberation-mono/LiberationMono-Bold.ttf: Liberation Mono:style=Bold /usr/share/fonts/liberation-mono/LiberationMono-BoldItalic.ttf: Liberation Mono:style=Bold Italic /usr/share/fonts/liberation-mono/LiberationMono-Italic.ttf: Liberation Mono:style=Italic Liberation Mono as the example I might not remembered exactly: some range and no change < 6.366 > changes visibly substantially range no change < 6.875 > again changes visibly quite a bit Also does not seem that "Cell spacing" _width_ works "properly" And my question about _only_ mono fonts being available in Ghome's terminal (while other terminals choose any font) ? many thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Gnome's terminal/console fonts ?
Hi guys. My Gnome's default terminal does only select from _mono_ type of fonts & I remember it could choose any font. Alternative solutions, such a _terminator_ do allow to select any font. In Gnome's terminal 'font scaling' also seems "broken" - certainly on my Fedora - I select "Liberation Mono" and I start size at 6 and keep increasing by small fraction and the font does not seem changing until suddenly, at some given size it "jumps" the size & ... shape too. I do not suppose it's my environment - I've been upgrading my Fedora since... forever, as opposed to "clean" install - and I think it's rather "broken" terminal, but.. I wanted to consult other Fedorians to be sure. many thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ClusterLabs] resource-agents and VMs
Hi guys. my resources-agents depend like so: resource-agents-deps.target ○ ├─00\\x2dVMsy.mount ● └─virt-guest-shutdown.target when I reboot a node VMs seems to migrated off it live a ok, but.. when node comes back on after a reboot, VMs fail to migrate back to it, live. I see on such node -> $ journalctl -lf -o cat -u virtqemud.service Starting Virtualization qemu daemon... Started Virtualization qemu daemon. libvirt version: 9.5.0, package: 6.el9 (buil...@centos.org, 2023-08-25-08:53:56, ) hostname: dzien.mine.priv Path '/00-VMsy/enc.podnode3.qcow2' is not accessible: No such file or directory and I wonder if it's indeed the fact the the _path_ is absent at the moment cluster just after node start, tries to migrate VM resource... Is it possible to somehow - seemingly my _resource-agents-deps.target_ does not do - assure that cluster, perhaps on per-resource basis, will wait/check a path first? BTW, that paths is available and is made sure is available to the system, it's a glusterfs mount. many thanks, L ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] how to colocate promoted resources ?
On 08/12/2023 13:25, Jehan-Guillaume de Rorthais wrote: Hi, On Wed, 6 Dec 2023 10:36:39 +0100 lejeczek via Users wrote: How do your colocate your promoted resources with balancing underlying resources as priority? What do you mean? With a simple scenario, say 3 nodes and 3 pgSQL clusters what would be best possible way - I'm thinking most gentle at the same time, if that makes sense. I'm not sure it answers your question (as I don't understand it), but here is a doc explaining how to create and move two IP supposed to each start on secondaries, avoiding the primary node if possible, as long as secondaries nodes exists https://clusterlabs.github.io/PAF/CentOS-7-admin-cookbook.html#adding-ips-on-standbys-nodes ++ Apologies, perhaps I was quite vague. I was thinking - having a 3-node HA cluster and 3-node single-master->slaves pgSQL, now.. say, I want pgSQL masters to spread across HA cluster so I theory - having each HA node identical hardware-wise - masters' resources would nicely balance out across HA cluster. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] ethernet link up/down - ?
On 04/12/2023 20:58, Reid Wahl wrote: On Thu, Nov 30, 2023 at 10:30 AM lejeczek via Users wrote: On 07/02/2022 20:09, lejeczek via Users wrote: Hi guys How do you guys go about doing link up/down as a resource? many thanks, L. With simple tests I confirmed that indeed Linux - on my hardware at leat - can easily power down an eth link - if a @devel reads this: Is there an agent in the suite which a non-programmer could easily (for most safely) adopt for such purpose? I understand such agent has to be cloneable & promotable. The iface-bridge resource appears to do something similar for bridges. I don't see anything currently for links in general. Where can I find that agent? Any comment on that idea about adding, introducing such "link" agent into agents in the future? Should I go _github_ and suggest it there perhaps? Naturally done by "devel" it would be ideal, as opposed to, by us user/admins. thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] how to colocate promoted resources ?
Hi guys. How do your colocate your promoted resources with balancing underlying resources as priority? With a simple scenario, say 3 nodes and 3 pgSQL clusters what would be best possible way - I'm thinking most gentle at the same time, if that makes sense. many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] make promoted follow promoted resource ?
On 26/11/2023 12:20, Reid Wahl wrote: On Sun, Nov 26, 2023 at 1:32 AM lejeczek via Users wrote: Hi guys. With these: -> $ pcs resource status REDIS-6381-clone * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable): * Promoted: [ ubusrv2 ] * Unpromoted: [ ubusrv1 ubusrv3 ] -> $ pcs resource status PGSQL-PAF-5433-clone * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs constraint ref REDIS-6381-clone Resource: REDIS-6381-clone colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY basically promoted Redis should follow promoted pgSQL but it's not happening, usually it does. I presume pcs/cluster does something internally which results in disobeying/ignoring that _colocation_ constraint for these resources. I presume scoring might play a role: REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) (with-rsc-role:Master) but usually, that scoring works, only "now" it does not. Any comments I appreciate much. thanks, L. I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but cannot see explanation for this. ... notice: Calculated transition 110, saving inputs in /var/lib/pacemaker/pengine/pe-input-3729.bz2 notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete notice: State transition S_TRANSITION_ENGINE -> S_IDLE notice: State transition S_IDLE -> S_POLICY_ENGINE notice: Actions: Start REDIS-6381:0 ( ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv3 ) notice: Actions: Start REDIS-6381:2 ( ubusrv1 ) notice: Calculated transition 111, saving inputs in /var/lib/pacemaker/pengine/pe-input-3730.bz2 notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2 notice: Requesting local execution of start operation for REDIS-6381 on ubusrv2 (to redis) root on none pam_unix(su:session): session opened for user redis(uid=127) by (uid=0) pam_sss(su:session): Request to sssd failed. Connection refused pam_unix(su:session): session closed for user redis pam_sss(su:session): Request to sssd failed. Connection refused notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000 notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create master-REDIS-6381=1000: Transient attribute change INFO: demote: Setting master to 'no-such-master' notice: Result of start operation for REDIS-6381 on ubusrv2: ok notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped notice: Actions: PromoteREDIS-6381:0 ( Unpromoted -> Promoted ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv1 ) notice: Actions: Start REDIS-6381:2 ( ubusrv3 ) notice: Calculated transition 112, saving inputs in /var/lib/pacemaker/pengine/pe-input-3731.bz2 notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating start operation REDIS-6381_start_0 on ubusrv1 notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3 notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2 notice: Requesting local execution of promote operation for REDIS-6381 on ubusrv2 notice: Result of promote operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_promote_0 on ubusrv3 notice: Result of notify operation for RED
[ClusterLabs] IPaddr2 Started (disabled) ?
hi guys. A cluster thinks the resource is up: ... * HA-10-1-1-80 (ocf:heartbeat:IPaddr2): Started ubusrv3 (disabled) .. while it is not the case. What might it mean? Config is simple: -> $ pcs resource config HA-10-1-1-80 Resource: HA-10-1-1-80 (class=ocf provider=heartbeat type=IPaddr2) Attributes: cidr_netmask=24 ip=10.1.1.80 Meta Attrs: failure-timeout=20s target-role=Stopped Operations: monitor interval=10s timeout=20s (HA-10-1-1-80-monitor-interval-10s) start interval=0s timeout=20s (HA-10-1-1-80-start-interval-0s) stop interval=0s timeout=20s (HA-10-1-1-80-stop-interval-0s) I expected, well.. cluster to behave. This results in a trouble - because of that exactly I think - for when resource is re-enabled then that IP/resource is not instantiated. Is its monitor not working? Is there a way to "harden" cluster perhaps via resource config? cluster is Ubuntu VM boxes. many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
a VM with secure boot does not survive a reset
Hi guys. A VM created with: -> $ virt-install --boot uefi --name=client-efi-secureboot --os-variant centos-stream9 --memory $((4*1024)) --graphics=vnc,listen=0.0.0.0 --accelerate --cpuset 2 --cpu host-model --pxe --disk=/00-VMsy/client-efi-secureboot.qcow2 --network network=10_3_1,,model=virtio when *virsh reset* while running will power off, instead of *restart*. Whereas: -> $ virt-install --boot uefi,firmware.feature0.name=secure-boot,firmware.feature0.enabled=no --name=client-efi-no-secureboot --os-variant centos-stream9 --memory $((4*1024)) --graphics=vnc,listen=0.0.0.0 --accelerate --cpuset 2 --cpu host-model --pxe --disk=/00-VMsy/client-efi-no-secureboot.qcow2 --network network=10_3_1,,model=virtio does behave as expected when *reset*. libvirt-libs-9.5.0-6.el9.x86_64 BTW. Where should bug reports go to theses days - I'm trying Redhat's Jira but cannot file a BZ in _RHEL_ category. I get no errors nor warnings but Jira won't let me file anything there, but I can in Centos category. thanks, L. ___ Users mailing list -- users@lists.libvirt.org To unsubscribe send an email to users-le...@lists.libvirt.org
pxe efi boot - bdsdxe boot0002 access denied
Hi guys I have VMs which fail to boot with error as per the subject. I suspect this is due to: hvm <--- HERE < HERE ? type='pflash'>/usr/share/edk2/ovmf/OVMF_CODE.secboot.fd template='/usr/share/edk2/ovmf/OVMF_VARS.secboot.fd'>/var/lib/libvirt/qemu/nvram/centos-efi_VARS.fd Would you have any thoughts on how to fix, to work around it? many thanks, L.___ Users mailing list -- users@lists.libvirt.org To unsubscribe send an email to users-le...@lists.libvirt.org
Re: [ClusterLabs] ethernet link up/down - ?
On 07/02/2022 20:09, lejeczek via Users wrote: Hi guys How do you guys go about doing link up/down as a resource? many thanks, L. With simple tests I confirmed that indeed Linux - on my hardware at leat - can easily power down an eth link - if a @devel reads this: Is there an agent in the suite which a non-programmer could easily (for most safely) adopt for such purpose? I understand such agent has to be cloneable & promotable. btw. I think many would consider that a really neat & useful addition to resource-agents - if authors/devel made it into the core package. many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] ethernet link up/down - ?
On 16/02/2022 10:37, Klaus Wenninger wrote: On Tue, Feb 15, 2022 at 5:25 PM lejeczek via Users wrote: On 07/02/2022 19:21, Antony Stone wrote: > On Monday 07 February 2022 at 20:09:02, lejeczek via Users wrote: > >> Hi guys >> >> How do you guys go about doing link up/down as a resource? > I apply or remove addresses on the interface, using "IPaddr2" and "IPv6addr", > which I know is not the same thing. > > Why do you separately want to control link up/down? I can't think what I > would use this for. Just out of curiosity and as I haven't seen an answer in the thread yet - maybe I overlooked something ... Is this to control some link-triggered redundancy setup with switches? Revisiting my own question/thread. Yes. Very close to what Klaus wondered - it's a device over which I have no control and from that device perspective it's simply - link is up then I'll "serve" it. I've been thinking lowest possible layer shall be the safest way - thus asked about controlling eth link that way: down/up by means of electric power, ideally. As opposed to ha-cluster calling some middle men such as network managers. I read some eth nics/drivers can power down a port. Is there an agent & a way to do that? > > > Antony. > Kind of similar - tcp/ip and those layers configs are delivered by DHCP. I'd think it would have to be a clone resource with one master without any constraints where cluster freely decides where to put master(link up) on - which is when link gets dhcp-served. But I wonder if that would mean writing up a new resource - I don't think there is anything like that included in ready-made pcs/ocf packages. many thanks, L ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/ ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] make promoted follow promoted resource ?
On 26/11/2023 17:44, Andrei Borzenkov wrote: On 26.11.2023 12:32, lejeczek via Users wrote: Hi guys. With these: -> $ pcs resource status REDIS-6381-clone * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable): * Promoted: [ ubusrv2 ] * Unpromoted: [ ubusrv1 ubusrv3 ] -> $ pcs resource status PGSQL-PAF-5433-clone * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs constraint ref REDIS-6381-clone Resource: REDIS-6381-clone colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY basically promoted Redis should follow promoted pgSQL but it's not happening, usually it does. I presume pcs/cluster does something internally which results in disobeying/ignoring that _colocation_ constraint for these resources. I presume scoring might play a role: REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) (with-rsc-role:Master) but usually, that scoring works, only "now" it does not. Any comments I appreciate much. thanks, L. I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but cannot see explanation for this. ... notice: Calculated transition 110, saving inputs in /var/lib/pacemaker/pengine/pe-input-3729.bz2 notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete notice: State transition S_TRANSITION_ENGINE -> S_IDLE notice: State transition S_IDLE -> S_POLICY_ENGINE notice: Actions: Start REDIS-6381:0 ( ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv3 ) notice: Actions: Start REDIS-6381:2 ( ubusrv1 ) notice: Calculated transition 111, saving inputs in /var/lib/pacemaker/pengine/pe-input-3730.bz2 notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2 notice: Requesting local execution of start operation for REDIS-6381 on ubusrv2 (to redis) root on none pam_unix(su:session): session opened for user redis(uid=127) by (uid=0) pam_sss(su:session): Request to sssd failed. Connection refused pam_unix(su:session): session closed for user redis pam_sss(su:session): Request to sssd failed. Connection refused notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000 This is the only line that sets master score, so apparently ubusrv2 is the only node where your clone *can* be promoted. Whether pacemaker is expected to fail this operation because it violates constraint I do not know. notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create master-REDIS-6381=1000: Transient attribute change INFO: demote: Setting master to 'no-such-master' notice: Result of start operation for REDIS-6381 on ubusrv2: ok notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped notice: Actions: Promote REDIS-6381:0 ( Unpromoted -> Promoted ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv1 ) notice: Actions: Start REDIS-6381:2 ( ubusrv3 ) notice: Calculated transition 112, saving inputs in /var/lib/pacemaker/pengine/pe-input-3731.bz2 notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating start operation REDIS-6381_start_0 on ubusrv1 notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3 notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2 notice: Requesting local execution of promote operation for REDIS-6381 on ubusrv2 notice: Result of promote operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 o
Re: [ClusterLabs] make promoted follow promoted resource ?
On 26/11/2023 12:20, Reid Wahl wrote: On Sun, Nov 26, 2023 at 1:32 AM lejeczek via Users wrote: Hi guys. With these: -> $ pcs resource status REDIS-6381-clone * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable): * Promoted: [ ubusrv2 ] * Unpromoted: [ ubusrv1 ubusrv3 ] -> $ pcs resource status PGSQL-PAF-5433-clone * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs constraint ref REDIS-6381-clone Resource: REDIS-6381-clone colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY basically promoted Redis should follow promoted pgSQL but it's not happening, usually it does. I presume pcs/cluster does something internally which results in disobeying/ignoring that _colocation_ constraint for these resources. I presume scoring might play a role: REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) (with-rsc-role:Master) but usually, that scoring works, only "now" it does not. Any comments I appreciate much. thanks, L. I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but cannot see explanation for this. ... notice: Calculated transition 110, saving inputs in /var/lib/pacemaker/pengine/pe-input-3729.bz2 notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete notice: State transition S_TRANSITION_ENGINE -> S_IDLE notice: State transition S_IDLE -> S_POLICY_ENGINE notice: Actions: Start REDIS-6381:0 ( ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv3 ) notice: Actions: Start REDIS-6381:2 ( ubusrv1 ) notice: Calculated transition 111, saving inputs in /var/lib/pacemaker/pengine/pe-input-3730.bz2 notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2 notice: Requesting local execution of start operation for REDIS-6381 on ubusrv2 (to redis) root on none pam_unix(su:session): session opened for user redis(uid=127) by (uid=0) pam_sss(su:session): Request to sssd failed. Connection refused pam_unix(su:session): session closed for user redis pam_sss(su:session): Request to sssd failed. Connection refused notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000 notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create master-REDIS-6381=1000: Transient attribute change INFO: demote: Setting master to 'no-such-master' notice: Result of start operation for REDIS-6381 on ubusrv2: ok notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped notice: Actions: PromoteREDIS-6381:0 ( Unpromoted -> Promoted ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv1 ) notice: Actions: Start REDIS-6381:2 ( ubusrv3 ) notice: Calculated transition 112, saving inputs in /var/lib/pacemaker/pengine/pe-input-3731.bz2 notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating start operation REDIS-6381_start_0 on ubusrv1 notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3 notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2 notice: Requesting local execution of promote operation for REDIS-6381 on ubusrv2 notice: Result of promote operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_promote_0 on ubusrv3 notice: Result of notify operation for RED
Re: [ClusterLabs] make promoted follow promoted resource ?
On 26/11/2023 10:32, lejeczek via Users wrote: Hi guys. With these: -> $ pcs resource status REDIS-6381-clone * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable): * Promoted: [ ubusrv2 ] * Unpromoted: [ ubusrv1 ubusrv3 ] -> $ pcs resource status PGSQL-PAF-5433-clone * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs constraint ref REDIS-6381-clone Resource: REDIS-6381-clone colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY basically promoted Redis should follow promoted pgSQL but it's not happening, usually it does. I presume pcs/cluster does something internally which results in disobeying/ignoring that _colocation_ constraint for these resources. I presume scoring might play a role: REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) (with-rsc-role:Master) but usually, that scoring works, only "now" it does not. Any comments I appreciate much. thanks, L. I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but cannot see explanation for this. ... notice: Calculated transition 110, saving inputs in /var/lib/pacemaker/pengine/pe-input-3729.bz2 notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete notice: State transition S_TRANSITION_ENGINE -> S_IDLE notice: State transition S_IDLE -> S_POLICY_ENGINE notice: Actions: Start REDIS-6381:0 ( ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv3 ) notice: Actions: Start REDIS-6381:2 ( ubusrv1 ) notice: Calculated transition 111, saving inputs in /var/lib/pacemaker/pengine/pe-input-3730.bz2 notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2 notice: Requesting local execution of start operation for REDIS-6381 on ubusrv2 (to redis) root on none pam_unix(su:session): session opened for user redis(uid=127) by (uid=0) pam_sss(su:session): Request to sssd failed. Connection refused pam_unix(su:session): session closed for user redis pam_sss(su:session): Request to sssd failed. Connection refused notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000 notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create master-REDIS-6381=1000: Transient attribute change INFO: demote: Setting master to 'no-such-master' notice: Result of start operation for REDIS-6381 on ubusrv2: ok notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped notice: Actions: Promote REDIS-6381:0 ( Unpromoted -> Promoted ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv1 ) notice: Actions: Start REDIS-6381:2 ( ubusrv3 ) notice: Calculated transition 112, saving inputs in /var/lib/pacemaker/pengine/pe-input-3731.bz2 notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating start operation REDIS-6381_start_0 on ubusrv1 notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3 notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2 notice: Requesting local execution of promote operation for REDIS-6381 on ubusrv2 notice: Result of promote operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Setting master-REDIS-6381[ubusrv3]: (unset) -> 1 notice: Transition 112
[ClusterLabs] make promoted follow promoted resource ?
Hi guys. With these: -> $ pcs resource status REDIS-6381-clone * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable): * Promoted: [ ubusrv2 ] * Unpromoted: [ ubusrv1 ubusrv3 ] -> $ pcs resource status PGSQL-PAF-5433-clone * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable): * Promoted: [ ubusrv1 ] * Unpromoted: [ ubusrv2 ubusrv3 ] -> $ pcs constraint ref REDIS-6381-clone Resource: REDIS-6381-clone colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY basically promoted Redis should follow promoted pgSQL but it's not happening, usually it does. I presume pcs/cluster does something internally which results in disobeying/ignoring that _colocation_ constraint for these resources. I presume scoring might play a role: REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) (with-rsc-role:Master) but usually, that scoring works, only "now" it does not. Any comments I appreciate much. thanks, L. I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but cannot see explanation for this. ... notice: Calculated transition 110, saving inputs in /var/lib/pacemaker/pengine/pe-input-3729.bz2 notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete notice: State transition S_TRANSITION_ENGINE -> S_IDLE notice: State transition S_IDLE -> S_POLICY_ENGINE notice: Actions: Start REDIS-6381:0 ( ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv3 ) notice: Actions: Start REDIS-6381:2 ( ubusrv1 ) notice: Calculated transition 111, saving inputs in /var/lib/pacemaker/pengine/pe-input-3730.bz2 notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2 notice: Requesting local execution of start operation for REDIS-6381 on ubusrv2 (to redis) root on none pam_unix(su:session): session opened for user redis(uid=127) by (uid=0) pam_sss(su:session): Request to sssd failed. Connection refused pam_unix(su:session): session closed for user redis pam_sss(su:session): Request to sssd failed. Connection refused notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000 notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create master-REDIS-6381=1000: Transient attribute change INFO: demote: Setting master to 'no-such-master' notice: Result of start operation for REDIS-6381 on ubusrv2: ok notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped notice: Actions: Promote REDIS-6381:0 ( Unpromoted -> Promoted ubusrv2 ) notice: Actions: Start REDIS-6381:1 ( ubusrv1 ) notice: Actions: Start REDIS-6381:2 ( ubusrv3 ) notice: Calculated transition 112, saving inputs in /var/lib/pacemaker/pengine/pe-input-3731.bz2 notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating start operation REDIS-6381_start_0 on ubusrv1 notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3 notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2 notice: Requesting local execution of promote operation for REDIS-6381 on ubusrv2 notice: Result of promote operation for REDIS-6381 on ubusrv2: ok notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally on ubusrv2 notice: Requesting local execution of notify operation for REDIS-6381 on ubusrv2 notice: Initiating notify operation REDIS-6381_post_notify_promote_0 on ubusrv1 notice: Initiating notify operation REDIS-6381:2_post_notify_promote_0 on ubusrv3 notice: Result of notify operation for REDIS-6381 on ubusrv2: ok notice: Setting master-REDIS-6381[ubusrv3]: (unset) -> 1 notice: Transition 112 aborted by status-3-master-REDIS-6381 doing create master-REDIS-6381=1: Transient
Re: [ClusterLabs] [EXT] moving VM live fails?
On 24/11/2023 08:33, Windl, Ulrich wrote: Hi! So you have different CPUs in the cluster? We once had a similar situation with Xen using live migration: Migration failed, and the cluster "wasn't that smart" handling the situation. The solution was (with the help of support) to add some CPU flags masking in the VM configuration so that the newer CPU features were not used. Before migration worked from the older to the newer CPU and back if the VM had been started on the older CPU initially, but when it had been started on the newer CPU it could not migrate to the older one. I vaguely remember the details but the hardware was a ProLiant 380 G7 vs. 380 G8 or so. My recommendation is: If you want to size a cluster, but all the nodes you need, because if you want to add another node maybe in two years or so, the older model may not be available any more. IT lifetime is short. Regards, Ulrich -Original Message- From: Users On Behalf Of lejeczek via Users Sent: Friday, November 17, 2023 12:55 PM To: users@clusterlabs.org Cc: lejeczek Subject: [EXT] [ClusterLabs] moving VM live fails? Hi guys. I have a resource which when asked to 'move' then it fails with: virtqemud[3405456]: operation failed: guest CPU doesn't match specification: missing features: xsave but VM domain does not require (nor disable) the feature: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/ This must go down - same environment - all the way, to the kernel. I knew that but forgotten - one box did change default kernel & after reboot booted with Centos's default as opposed to other boxes running with ver. 6.x. So, my bad, not really an issue. Thanks,. L ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] node orderly shutdown
Hi guys. Having a node with a couple of _promoted_ resources - when such node is os-shutdown in an orderly manner it seems that cluster takes a while. By a "while" I mean longer than I'd expect a relatively simple 3-node cluster to move/promote a few _promoted_ resources: redis, postgresql, IP onto another. Is there somewhere one can look, tweak or measure & troubleshot, in order to "fix" this, if possible at all? From watching such a "promoted" node I see that as systemd stops all services going into power-down target - it's _pacemaker_ which as last systemd takes bit longer before complete shutdown. Or perhaps you have another & more than one approach / technique to node-with-promoted-resources shutdown, a better one? many thanks,___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
does Firefox honor local self-signed CAs ?
Hi guys I put my own CAs certs into: /etc/pki/ca-trust/source/anchors then _update-ca-trust_ My other web-browsers seem happy but Firefox always shows/warns: connection not secure as oppose to other browsers saying: is secure Something has changed in default behavior - I remember Firefox did not do it - even if it was possible to "tweak" Firefox back to "normal" would you know? thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: zbar VS qrencode
On 19/11/2023 14:17, lejeczek via users wrote: Hi guys I have a peculiar case, I think, while using national chars. I wonder which one is _playing_ up here, qrencode or zbarimg When I with: -> qrencode -l L -v 1 .. encode: ADR;LABEL="Kukólka":;;;Kukólka;Mazowieckie;PL then: -> $ zbarimg ... ... ADR;LABEL="Kuk籀lka":;;;Kuk籀lka;Mazowieckie;PL but when I encode: ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL then: -> $ zbarimg ... ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL you see - if all the "processors" between our email clients, mine your yours, did not mangle the content - it seems when there is only *single* polish char - ó - then _zbaring_ shows that weird output - a Korean or Japanese? Any other case, add just one morenational char in the line and the output, even that same - ó - mangled when single, now shows as it should. So I checked all nine polish national chars and it's: ó and ż when occur as single in a line - _zbarimg_ shows them incorrectly. (or _qrencode_ fails?) Any _qrencode_ & _zbarimg_ users here who use their national chars? I'll be glad for all comments. thanks, L. Ah, to answer my question - partly at least - seems that _zbarimg_ could use some fixing up. Trying _qrscanner_ and here I get everything shown correctly. Would anybody know if _zbarimg_ is looked after still, if it's worth filing a bug? -- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
zbar VS qrencode
Hi guys I have a peculiar case, I think, while using national chars. I wonder which one is _playing_ up here, qrencode or zbarimg When I with: -> qrencode -l L -v 1 .. encode: ADR;LABEL="Kukólka":;;;Kukólka;Mazowieckie;PL then: -> $ zbarimg ... ... ADR;LABEL="Kuk籀lka":;;;Kuk籀lka;Mazowieckie;PL but when I encode: ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL then: -> $ zbarimg ... ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL you see - if all the "processors" between our email clients, mine your yours, did not mangle the content - it seems when there is only *single* polish char - ó - then _zbaring_ shows that weird output - a Korean or Japanese? Any other case, add just one morenational char in the line and the output, even that same - ó - mangled when single, now shows as it should. So I checked all nine polish national chars and it's: ó and ż when occur as single in a line - _zbarimg_ shows them incorrectly. (or _qrencode_ fails?) Any _qrencode_ & _zbarimg_ users here who use their national chars? I'll be glad for all comments. thanks, L.-- ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ClusterLabs] non-existent attribute ?
Hi guys. My 3-node cluster had one node absent for a long time and now when it's back I cannot get _mariadb_ to start on that node. ... * MARIADB (ocf:heartbeat:galera): ORPHANED Stopped ... * MARIADB-last-committed : 147 * MARIADB-safe-to-bootstrap : 0 I wanted to start with a _cleanup_ of node's attribute - as mariadb node outside of ha-pcs starts & run a ok, but I get: -> $ pcs node attribute dzien MARIADB-last-committed='' Error: attribute: 'MARIADB-last-committed' doesn't exist for node: 'dzien' Is possible to clean that up safely? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] [EXT] Re: PAF / pgSQL fails after OS/system shutdown - FIX
On 13/11/2023 13:08, Jehan-Guillaume de Rorthais via Users wrote: On Mon, 13 Nov 2023 11:39:45 + "Windl, Ulrich" wrote: But shouldn't the RA check for that (and act appropriately)? Interesting. I'm open to discuss this. Below my thoughts so far. Why the RA should check that? There's so many way to setup the system and PostgreSQL, where should the RA stop checking for all possible way to break it? The RA checks various (maybe too many) things related to the instance itself already. I know various other PostgreSQL setups that would trigger errors in the cluster if the dba doesn't check everything is correct. I'm really reluctant to add add a fair amount of code in the RA to correctly parse and check the complex PostgreSQL's setup. This would add complexity and bugs. Or maybe I could add a specific OCF_CHECK_LEVEL sysadmins can trigger by hand before starting the cluster. But I wonder if it worth the pain, how many people will know about this and actually run it? The problem here is that few users actually realize how the postgresql-common wrapper works and what it actually does behind your back. I really appreciate this wrapper, I do. But when you setup a Pacemaker cluster, you either have to bend to it when setting up PAF (as documented), or avoid it completely. PAF is all about drawing a clear line between the sysadmin job and the dba one. Dba must build a cluster of instances ready to start/replicate with standard binaries (not wrappers) before sysadmin can set up the resource in your cluster. Thoughts? I would be the same/similar mind - which is - adding more code to account for more/all _config_ cases may not be the healthiest approach, however!! Just like some code-writers out there in around the globe here too, some do seem to _not_ appreciate or seem to completely ignore, parts which are meant do alleviate the TCO burden of any software - documentation best bits which for Unix/Linux... are MAN PAGES. This is rhetorical but I'll ask - what is a project worth when its own man pages miss critical _gotchas_ and similar gimmicks. If you read here are are a prospective/future coder-writer for Linux please make such note - your 'man pages" ought to be at least!! as good as your code. many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] moving VM live fails?
Hi guys. I have a resource which when asked to 'move' then it fails with: virtqemud[3405456]: operation failed: guest CPU doesn't match specification: missing features: xsave but VM domain does not require (nor disable) the feature: what even more interesting, _virsh_ migrate does the migration live a ok. I'm on Centos 9 with packages up to day. All thoughts much appreciated. many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown - FIX
On 10/11/2023 18:16, Jehan-Guillaume de Rorthais wrote: On Fri, 10 Nov 2023 17:17:41 +0100 lejeczek via Users wrote: ... Of course you can use "pg_stat_tmp", just make sure the temp folder exists: cat < /etc/tmpfiles.d/postgresql-part.conf # Directory for PostgreSQL temp stat files d /var/run/postgresql/14-paf.pg_stat_tmp 0700 postgres postgres - - EOF To take this file in consideration immediately without rebooting the server, run the following command: systemd-tmpfiles --create /etc/tmpfiles.d/postgresql-part.conf Then there must be something else at play here with Ubuntus, for none of the nodes has any extra/additional configs for those paths & I'm sure that those were not created manually. Indeed. This parameter is usually set by pg_createcluster command and the folder created by both pg_createcluster and pg_ctlcluster commands when needed. This is explained in PAF tutorial there: https://clusterlabs.github.io/PAF/Quick_Start-Debian-10-pcs.html#postgresql-and-cluster-stack-installation These commands comes from the postgresql-common wrapper, used in all Debian related distros, allowing to install, create and use multiple PostgreSQL versions on the same server. Perhpaphs pgSQL created these on it's own outside of HA-cluster. No, the Debian packaging did. Just create the config file I pointed you in my previous answer, systemd-tmpfiles will take care of it and you'll be fine. yes, that was a weird trip. I could not take the whole cluster down & as I kept fiddling with it trying to "fix" it - within & out of _pcs_ those paths were created by wrappers from disto packages - I know see that - but at first it did not look that way. Still that directive for the stats - I wonder if that got introduced, injected somewhere in between because the PG cluster - which I had for a while - did not experience this "issue" from the start and not until "recently" thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown - FIX
On 10/11/2023 13:13, Jehan-Guillaume de Rorthais wrote: On Fri, 10 Nov 2023 12:27:24 +0100 lejeczek via Users wrote: ... to share my "fix" for it - perhaps it was introduced by OS/packages (Ubuntu 22) updates - ? - as oppose to resource agent itself. As the logs point out - pg_stat_tmp - is missing and from what I see it's only the master, within a cluster, doing those stats. That appeared, I use the word for I did not put it into configs, on all nodes. fix = to not use _pg_stat_tmp_ directive/option at all. Of course you can use "pg_stat_tmp", just make sure the temp folder exists: cat < /etc/tmpfiles.d/postgresql-part.conf # Directory for PostgreSQL temp stat files d /var/run/postgresql/14-paf.pg_stat_tmp 0700 postgres postgres - - EOF To take this file in consideration immediately without rebooting the server, run the following command: systemd-tmpfiles --create /etc/tmpfiles.d/postgresql-part.conf Then there must be something else at play here with Ubuntus, for none of the nodes has any extra/additional configs for those paths & I'm sure that those were not created manually. Perhpaphs pgSQL created these on it's own outside of HA-cluster. Also, one node has the path missing but the other two have & between those two only - how it seems to me - the master PG actually put any data there and if that is the case - does the existence of the path alone guarantee anything, is the question. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown - FIX
On 07/11/2023 17:57, lejeczek via Users wrote: hi guys Having 3-node pgSQL cluster with PAF - when all three systems are shutdown at virtually the same time then PAF fails to start when HA cluster is operational again. from status: ... Migration Summary: * Node: ubusrv2 (2): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' * Node: ubusrv3 (3): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' * Node: ubusrv1 (1): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' Failed Resource Actions: * PGSQL-PAF-5433_stop_0 on ubusrv2 'error' (1): call=90, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=84ms * PGSQL-PAF-5433_stop_0 on ubusrv3 'error' (1): call=82, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=82ms * PGSQL-PAF-5433_stop_0 on ubusrv1 'error' (1): call=86, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=108ms and all three pgSQLs show virtually identical logs: ... 2023-11-07 16:54:45.532 UTC [24936] LOG: starting PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit 2023-11-07 16:54:45.532 UTC [24936] LOG: listening on IPv4 address "0.0.0.0", port 5433 2023-11-07 16:54:45.532 UTC [24936] LOG: listening on IPv6 address "::", port 5433 2023-11-07 16:54:45.535 UTC [24936] LOG: listening on Unix socket "/var/run/postgresql/.s.PGSQL.5433" 2023-11-07 16:54:45.547 UTC [24938] LOG: database system was interrupted while in recovery at log time 2023-11-07 15:30:56 UTC 2023-11-07 16:54:45.547 UTC [24938] HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target. 2023-11-07 16:54:45.819 UTC [24938] LOG: entering standby mode 2023-11-07 16:54:45.824 UTC [24938] FATAL: could not open directory "/var/run/postgresql/14-paf.pg_stat_tmp": No such file or directory 2023-11-07 16:54:45.825 UTC [24936] LOG: startup process (PID 24938) exited with exit code 1 2023-11-07 16:54:45.825 UTC [24936] LOG: aborting startup due to startup process failure 2023-11-07 16:54:45.826 UTC [24936] LOG: database system is shut down Is this "test" case's result, as I showed above, expected? It reproduces every time. If not - what might it be I'm missing? many thanks, L. ___ to share my "fix" for it - perhaps it was introduced by OS/packages (Ubuntu 22) updates - ? - as oppose to resource agent itself. As the logs point out - pg_stat_tmp - is missing and from what I see it's only the master, within a cluster, doing those stats. That appeared, I use the word for I did not put it into configs, on all nodes. fix = to not use _pg_stat_tmp_ directive/option at all.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown
On 07/11/2023 17:57, lejeczek via Users wrote: hi guys Having 3-node pgSQL cluster with PAF - when all three systems are shutdown at virtually the same time then PAF fails to start when HA cluster is operational again. from status: ... Migration Summary: * Node: ubusrv2 (2): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' * Node: ubusrv3 (3): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' * Node: ubusrv1 (1): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' Failed Resource Actions: * PGSQL-PAF-5433_stop_0 on ubusrv2 'error' (1): call=90, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=84ms * PGSQL-PAF-5433_stop_0 on ubusrv3 'error' (1): call=82, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=82ms * PGSQL-PAF-5433_stop_0 on ubusrv1 'error' (1): call=86, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=108ms and all three pgSQLs show virtually identical logs: ... 2023-11-07 16:54:45.532 UTC [24936] LOG: starting PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit 2023-11-07 16:54:45.532 UTC [24936] LOG: listening on IPv4 address "0.0.0.0", port 5433 2023-11-07 16:54:45.532 UTC [24936] LOG: listening on IPv6 address "::", port 5433 2023-11-07 16:54:45.535 UTC [24936] LOG: listening on Unix socket "/var/run/postgresql/.s.PGSQL.5433" 2023-11-07 16:54:45.547 UTC [24938] LOG: database system was interrupted while in recovery at log time 2023-11-07 15:30:56 UTC 2023-11-07 16:54:45.547 UTC [24938] HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target. 2023-11-07 16:54:45.819 UTC [24938] LOG: entering standby mode 2023-11-07 16:54:45.824 UTC [24938] FATAL: could not open directory "/var/run/postgresql/14-paf.pg_stat_tmp": No such file or directory 2023-11-07 16:54:45.825 UTC [24936] LOG: startup process (PID 24938) exited with exit code 1 2023-11-07 16:54:45.825 UTC [24936] LOG: aborting startup due to startup process failure 2023-11-07 16:54:45.826 UTC [24936] LOG: database system is shut down Is this "test" case's result, as I showed above, expected? It reproduces every time. If not - what might it be I'm missing? many thanks, L. Actually, the resource fails to start on a node a single node - as opposed to entire cluster shutdown as I noted originally - which was powered down in an orderly fashion and powered back on. That the the time of power-cycle the node was PAF resource master, it fails: ... 2023-11-09 20:35:04.439 UTC [17727] LOG: starting PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit 2023-11-09 20:35:04.439 UTC [17727] LOG: listening on IPv4 address "0.0.0.0", port 5433 2023-11-09 20:35:04.439 UTC [17727] LOG: listening on IPv6 address "::", port 5433 2023-11-09 20:35:04.442 UTC [17727] LOG: listening on Unix socket "/var/run/postgresql/.s.PGSQL.5433" 2023-11-09 20:35:04.452 UTC [17731] LOG: database system was interrupted while in recovery at log time 2023-11-09 20:25:21 UTC 2023-11-09 20:35:04.452 UTC [17731] HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target. 2023-11-09 20:35:04.809 UTC [17731] LOG: entering standby mode 2023-11-09 20:35:04.813 UTC [17731] FATAL: could not open directory "/var/run/postgresql/14-paf.pg_stat_tmp": No such file or directory 2023-11-09 20:35:04.814 UTC [17727] LOG: startup process (PID 17731) exited with exit code 1 2023-11-09 20:35:04.814 UTC [17727] LOG: aborting startup due to startup process failure 2023-11-09 20:35:04.815 UTC [17727] LOG: database system is shut down The master at the time node was shut down did get moved over to standby/slave node, properly, I'm on Ubuntu with: ii corosync 3.1.6-1ubuntu1 amd64 cluster engine daemon and utilities ii pacemaker 2.1.2-1ubuntu3.1 amd64 cluster resource manager ii pacemaker-cli-utils 2.1.2-1ubuntu3.1 amd64 cluster resource manager command line utilities ii pacemaker-common 2.1.2-1ubuntu3.1 all cluster resource manager common files ii pacemaker-resource-agents 2.1.2-1ubuntu3.1 all cluster res
Re: freaking TFTP
On 07/11/2023 15:42, Chris Adams wrote: Once upon a time, lejeczek said: 2nd meanwhile - I'm trying _dnsmasq_ which I'm new to thus errors are possible but... it seems that the same issue remains. with _dnsmasq_. tftp client - as with in.tftp as the server - "times out" but _dnsmasq_ server thinks and says that: I use dnsmasq's TFTP server on a bridge, with this config: bind-interfaces interface=br0 port=0 enable-tftp tftp-root=/srv/tftpboot "port=0" disables DNS, and I then don't configure DHCP, so it's just a TFTP server. I also should have added that it's Centos I'm doing all this on, with kernel from _elrepo_ Anybody here do/use Centos that way, with net bridges? ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: freaking TFTP
On 07/11/2023 15:42, Chris Adams wrote: Once upon a time, lejeczek said: 2nd meanwhile - I'm trying _dnsmasq_ which I'm new to thus errors are possible but... it seems that the same issue remains. with _dnsmasq_. tftp client - as with in.tftp as the server - "times out" but _dnsmasq_ server thinks and says that: I use dnsmasq's TFTP server on a bridge, with this config: bind-interfaces interface=br0 port=0 enable-tftp tftp-root=/srv/tftpboot "port=0" disables DNS, and I then don't configure DHCP, so it's just a TFTP server. I started with that - got logs as shown in my last email. One thing I should perhaps mention, is that libvirt uses my NM bridges, but it's simple: -> $ virsh net-dumpxml 10_1_1 10_1_1 864fb78e-0fb0-4c32-bb47-0e5ac68d9491 and I guess, but that does not "add" to on-bare-metal NM bridge. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ClusterLabs] PAF / pgSQL fails after OS/system shutdown
hi guys Having 3-node pgSQL cluster with PAF - when all three systems are shutdown at virtually the same time then PAF fails to start when HA cluster is operational again. from status: ... Migration Summary: * Node: ubusrv2 (2): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' * Node: ubusrv3 (3): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' * Node: ubusrv1 (1): * PGSQL-PAF-5433: migration-threshold=100 fail-count=100 last-failure='Tue Nov 7 17:52:38 2023' Failed Resource Actions: * PGSQL-PAF-5433_stop_0 on ubusrv2 'error' (1): call=90, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=84ms * PGSQL-PAF-5433_stop_0 on ubusrv3 'error' (1): call=82, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=82ms * PGSQL-PAF-5433_stop_0 on ubusrv1 'error' (1): call=86, status='complete', exitreason='Unexpected state for instance "PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov 7 17:52:38 2023', queued=0ms, exec=108ms and all three pgSQLs show virtually identical logs: ... 2023-11-07 16:54:45.532 UTC [24936] LOG: starting PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit 2023-11-07 16:54:45.532 UTC [24936] LOG: listening on IPv4 address "0.0.0.0", port 5433 2023-11-07 16:54:45.532 UTC [24936] LOG: listening on IPv6 address "::", port 5433 2023-11-07 16:54:45.535 UTC [24936] LOG: listening on Unix socket "/var/run/postgresql/.s.PGSQL.5433" 2023-11-07 16:54:45.547 UTC [24938] LOG: database system was interrupted while in recovery at log time 2023-11-07 15:30:56 UTC 2023-11-07 16:54:45.547 UTC [24938] HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target. 2023-11-07 16:54:45.819 UTC [24938] LOG: entering standby mode 2023-11-07 16:54:45.824 UTC [24938] FATAL: could not open directory "/var/run/postgresql/14-paf.pg_stat_tmp": No such file or directory 2023-11-07 16:54:45.825 UTC [24936] LOG: startup process (PID 24938) exited with exit code 1 2023-11-07 16:54:45.825 UTC [24936] LOG: aborting startup due to startup process failure 2023-11-07 16:54:45.826 UTC [24936] LOG: database system is shut down Is this "test" case's result, as I showed above, expected? It reproduces every time. If not - what might it be I'm missing? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: freaking TFTP
On 07/11/2023 13:43, Mauricio Tavares wrote: On Tue, Nov 7, 2023 at 7:15 AM lejeczek via users wrote: Hi guys. cannot get a client to get anything from 'in.tftp' server? Connection seems to get established, _tcpdump_ shows: ... 13:13:08.331277 IP 10.3.1.100.35176 > whale.mine.priv.tftp: TFTP, length 19, RRQ "zabbix.zip" octet 13:13:13.331324 IP 10.3.1.100.35176 > whale.mine.priv.tftp: TFTP, length 19, RRQ "zabbix.zip" octet ... journal log for tftp.service, so on the servers, shows: ... tftpd: read(ack): No route to host ... any & all thoughts on how troubleshoot & fix this are much appreciated. many thanks, L. ___ Can you connect from client host to tftp host using other services? ___ Yes, everything works as... well, as expected/normal. Meanwhile... I've read through my old notes and found: .. When you use in.tftpd on interface like bridges, this probably won't work, it at least did not for Dell PowerConnect when you wanted to upload the image, it worked on regualar eth And it today, a couple of years later and different hardware, seems that still with bridge ifaces, for both server & client, the issue remains. 2nd meanwhile - I'm trying _dnsmasq_ which I'm new to thus errors are possible but... it seems that the same issue remains. with _dnsmasq_. tftp client - as with in.tftp as the server - "times out" but _dnsmasq_ server thinks and says that: ... sent /var/lib/tftpboot/zabbix.zip to 10.3.1.100 ... This might go as far down as kernel's tweaking or perhaps bridge's iface - if possible at all - any kernel/net experts are most welcome to comment. many thanks, L. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
freaking TFTP
Hi guys. cannot get a client to get anything from 'in.tftp' server? Connection seems to get established, _tcpdump_ shows: ... 13:13:08.331277 IP 10.3.1.100.35176 > whale.mine.priv.tftp: TFTP, length 19, RRQ "zabbix.zip" octet 13:13:13.331324 IP 10.3.1.100.35176 > whale.mine.priv.tftp: TFTP, length 19, RRQ "zabbix.zip" octet ... journal log for tftp.service, so on the servers, shows: ... tftpd: read(ack): No route to host ... any & all thoughts on how troubleshoot & fix this are much appreciated. many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
MongoDB & lots of SELinux denials
Hi guys. I’ve just installed vanilla-default MongoDB (following their official docs) and right away SELinux shows denials: ... SELinux is preventing /usr/bin/mongod from search access on the directory /var/lib/containers/storage/overlay-containers/fe6f1040c1373930efe68c777805ecd4c921631e3a87b8806af1fe0cdf266d8b/userdata/shm. For complete SELinux messages run: sealert -l 5263cddc-183a-472b-9098-bf4599c8453c ... What is Mongo’s business in checking such path (and there is more paths)? - I refuse to believe (yet) that my Mongo is trojaned. Many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
LUKS - lost token?
Hi guys. I know this is most likely not best suited question for this list, but I'm hoping some experts might be able to help. I have a LUKS device which had keyslot with pass-phrase removed and token for TPM keyslot removed too - I think this is the case, for none of my passphrase works and device is as below: I hope there is a way to save & bring it back to live - device is open right now and I've access to filesystem, obviously goal would be to avoid re-format/crypt. Is it possible to restore/recreate that lost token and/or add new Keyslot somehow? I have no header backup for this device. -> $ cryptsetup luksDump /dev/nvme0n1p3 LUKS header information Version: 2 Epoch: 83 Metadata area: 16384 [bytes] Keyslots area: 1678 [bytes] UUID: 3a879268-84fd-4b48-a5d4-960cccb0caa9 Label: (no label) Subsystem: (no subsystem) Flags: (no flags) Data segments: 0: crypt offset: 16777216 [bytes] length: (whole device) cipher: aes-xts-plain64 sector: 512 [bytes] Keyslots: 1: luks2 Key: 512 bits Priority: normal Cipher: aes-xts-plain64 Cipher key: 512 bits PBKDF: pbkdf2 Hash: sha512 Iterations: 1000 Salt: a4 5b 6b cc a8 f1 6b e8 b7 3b e2 3d ca 8d 43 fb 10 52 62 b9 99 45 70 16 bd e1 0f 7a 6c 7f 3d 11 AF stripes: 4000 AF hash: sha512 Area offset:290816 [bytes] Area length:258048 [bytes] Digest ID: 0 Tokens: Digests: 0: pbkdf2 Hash: sha256 Iterations: 183317 Salt: ef 56 aa 59 c2 64 66 c7 49 57 31 4b a7 7d 00 3c fe 00 89 2e b9 e9 da bc 69 1d 19 59 96 a9 27 aa Digest: 79 aa 0c 8a 29 64 9c 83 bb 5a f8 5c b5 c6 b0 9c 5e 54 80 49 bd 21 f6 b4 5b 49 65 39 bd 6f 5f 20 ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Ungoogled Chromium - spell check?
Hi guys. Anybody here uses _Ungoogled Chromium_ ? If so - does 'spell check' work for you? many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: how to use amd-pstate - ?
On 29/10/2022 15:19, Matti Pulkkinen wrote: On Sat, 2022-10-29 at 12:01 +0100, lejeczek via users wrote: Hi guys. Does anybody use 'amd-pstate' and if yes could share a howto on how to have Fedora use that driver? many thanks, L. If you have a Zen2 or newer AMD processor, you can use a grubby command in a terminal to set the needed kernel argument: sudo grubby --update-kernel DEFAULT --args "initcall_blacklist=acpi_cpufreq_init" Then, after a reboot, you can use this command to see if it worked: cat /sys/devices/system/cpu/cpu0/cpufreq/scaling_driver If that says something other than amd-pstate then it did not work, and you can try adding another kernel argument and then rebooting: sudo grubby --update-kernel DEFAULT --args "amd_pstate.shared_mem=1" but is there a module? I cannot find it default/repo kernels. I see _amd_pstate_ut_ which says is a "test module" Is there still no official/approved p-state in f38, even with 6.5.x? ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Thunderbird & non-english characters ?
Hi guys. I'm trying to add an account in Thunderbird, address contains non-english chars and Thunderbird stops, does not proceed further, colors the field in red & sticks ! mark there. That surely must be not an issue of Thunderbird/OS but my configuration somewhere, right? Email accounts/address with non-english chars are a norm nowadays? thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] PAF / PGSQLMS and wal_level ?
On 08/09/2023 17:29, Jehan-Guillaume de Rorthais wrote: On Fri, 8 Sep 2023 16:52:53 +0200 lejeczek via Users wrote: Hi guys. Before I start fiddling and brake things I wonder if somebody knows if: pgSQL can work with: |wal_level = archive for PAF ? Or more general question with pertains to ||wal_level - can _barman_ be used with pgSQL "under" PAF? PAF needs "wal_level = replica" (or "hot_standby" on very old versions) so it can have hot standbys where it can connects and query there status. Wal level "replica" includes the archive level, so you can set up archiving. Of course you can use barman or any other tools to manage your PITR Backups, even when Pacemaker/PAF is looking at your instances. This is even the very first step you should focus on during your journey to HA. Regards, and with _barman_ specifically - is one method preferred, recommended over another: streaming VS rsync - for/with PAF? many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] PAF / PGSQLMS and wal_level ?
Hi guys. Before I start fiddling and brake things I wonder if somebody knows if: pgSQL can work with: |wal_level = archive for PAF ? Or more general question with pertains to ||wal_level - can _barman_ be used with pgSQL "under" PAF? many thanks, L. |___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] PAF / PGSQLMS on Ubuntu
On 07/09/2023 16:20, lejeczek via Users wrote: On 07/09/2023 16:09, Andrei Borzenkov wrote: On Thu, Sep 7, 2023 at 5:01 PM lejeczek via Users wrote: Hi guys. I'm trying to set ocf_heartbeat_pgsqlms agent but I get: ... Failed Resource Actions: * PGSQL-PAF-5433 stop on ubusrv3 returned 'invalid parameter' because 'Parameter "recovery_target_timeline" MUST be set to 'latest'. It is currently set to ''' at Thu Sep 7 13:58:06 2023 after 54ms I'm new to Ubuntu and I see that Ubuntu has a bit different approach to paths (in comparison to how Centos do it). I see separation between config & data, eg. 14 paf 5433 down postgres /var/lib/postgresql/14/paf /var/log/postgresql/postgresql-14-paf.log I create the resource like here: -> $ pcs resource create PGSQL-PAF-5433 ocf:heartbeat:pgsqlms pgport=5433 bindir=/usr/bin pgdata=/etc/postgresql/14/paf datadir=/var/lib/postgresql/14/paf meta failure-timeout=30s master-max=1 op start timeout=60s op stop timeout=60s op promote timeout=30s op demote timeout=120s op monitor interval=15s timeout=10s role="Promoted" op monitor interval=16s timeout=10s role="Unpromoted" op notify timeout=60s promotable notify=true failure-timeout=30s master-max=1 --disable Ubuntu 22.04.3 LTS What am I missing can you tell? Exactly what the message tells you. You need to set recovery_target=latest. and having it in 'postgresql.conf' make it all work for you? I've had it and got those errors - perhaps that has to be set some place else. In case anybody was in this situation - I was missing one important bit: _bindir_ Ubuntu's pgSQL binaries have different path - what resource/agent returns as errors is utterly confusing. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] PAF / PGSQLMS on Ubuntu
On 07/09/2023 16:09, Andrei Borzenkov wrote: On Thu, Sep 7, 2023 at 5:01 PM lejeczek via Users wrote: Hi guys. I'm trying to set ocf_heartbeat_pgsqlms agent but I get: ... Failed Resource Actions: * PGSQL-PAF-5433 stop on ubusrv3 returned 'invalid parameter' because 'Parameter "recovery_target_timeline" MUST be set to 'latest'. It is currently set to ''' at Thu Sep 7 13:58:06 2023 after 54ms I'm new to Ubuntu and I see that Ubuntu has a bit different approach to paths (in comparison to how Centos do it). I see separation between config & data, eg. 14 paf 5433 down postgres /var/lib/postgresql/14/paf /var/log/postgresql/postgresql-14-paf.log I create the resource like here: -> $ pcs resource create PGSQL-PAF-5433 ocf:heartbeat:pgsqlms pgport=5433 bindir=/usr/bin pgdata=/etc/postgresql/14/paf datadir=/var/lib/postgresql/14/paf meta failure-timeout=30s master-max=1 op start timeout=60s op stop timeout=60s op promote timeout=30s op demote timeout=120s op monitor interval=15s timeout=10s role="Promoted" op monitor interval=16s timeout=10s role="Unpromoted" op notify timeout=60s promotable notify=true failure-timeout=30s master-max=1 --disable Ubuntu 22.04.3 LTS What am I missing can you tell? Exactly what the message tells you. You need to set recovery_target=latest. and having it in 'postgresql.conf' make it all work for you? I've had it and got those errors - perhaps that has to be set some place else. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] PAF / PGSQLMS on Ubuntu
Hi guys. I'm trying to set ocf_heartbeat_pgsqlms agent but I get: ... Failed Resource Actions: * PGSQL-PAF-5433 stop on ubusrv3 returned 'invalid parameter' because 'Parameter "recovery_target_timeline" MUST be set to 'latest'. It is currently set to ''' at Thu Sep 7 13:58:06 2023 after 54ms I'm new to Ubuntu and I see that Ubuntu has a bit different approach to paths (in comparison to how Centos do it). I see separation between config & data, eg. 14 paf 5433 down postgres /var/lib/postgresql/14/paf /var/log/postgresql/postgresql-14-paf.log I create the resource like here: -> $ pcs resource create PGSQL-PAF-5433 ocf:heartbeat:pgsqlms pgport=5433 bindir=/usr/bin pgdata=/etc/postgresql/14/paf datadir=/var/lib/postgresql/14/paf meta failure-timeout=30s master-max=1 op start timeout=60s op stop timeout=60s op promote timeout=30s op demote timeout=120s op monitor interval=15s timeout=10s role="Promoted" op monitor interval=16s timeout=10s role="Unpromoted" op notify timeout=60s promotable notify=true failure-timeout=30s master-max=1 --disable Ubuntu 22.04.3 LTS What am I missing can you tell? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] last man - issue
Hi guys. That below should work, right? -> $ pcs quorum update last_man_standing=1 --skip-offline Checking corosync is not running on nodes... Warning: Unable to connect to dzien (Failed to connect to dzien port 2224: No route to host) Warning: dzien: Unable to check if corosync is not running Error: swir: corosync is running Error: whale: corosync is running Error: Errors have occurred, therefore pcs is unable to continue if it should, why does it not or what else am I missing? many thanks, L ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: SATA drive rescan?
On 17/07/2023 19:57, ToddAndMargo via users wrote: On 7/17/23 01:06, lejeczek via users wrote: On 16/07/2023 08:27, ToddAndMargo via users wrote: Hi All, I have a removable drive sleeve that I mount SATA drives mounted in that attaches to an internal SATA port. I use it backup my system to removable drives. The mechanism works identical to USB flash drives. With one exception. A drive MUST be in the carriage when I boot up. Otherwise, if I insert a drive AFTER the system is booted up, the drive is invisible to the system. Now I have seen Clonzilla re-educating the system as to what drives are where. Is there a way to goose Fedora to do the same? Is there a way to rescan without having to reboot? Many thanks, -T Bit better, clearer description of the hardware/caddie/connection you should have started with. I thought "attaches to an internal SATA port" covered that. https://www.supermicro.com/en/products/motherboard/X11SAE-M It is attached to "I-SATA1" So, it is purely all SATA (& not! e-sata) and no bridge of any kind in-between If so, if purely SATA, I'd suggest you check - if have not done so yet- your mobo's BIOS for something like SATA devices "hot-plugging" Hot plug is enabled and is the default. ___ And to clarify - CloneZilla does the trick when boot up independently of Fedora, as a live-iso, yes? If so then it rather is - as you presumed I guess - OS/software end. Then I'd presume that your BIOS has: - AHCI and not RAID - ports Spin Up Device is "Enabled" - also there is/should be "SATA Device Type" I'd play - imagine you did too - with: 'signal-first-then-power' VS 'power-first-then-signal' Other than that Fedora, Centos. which I mostly use which is older, normally do not need much - I always try not to forget to "enable" that in BIOS -> different vendor might have different nomenclature for it: hot- plug OR swap I've only glanced at your 'poking' re/scan, I think it's rather the host/controller you want to poke. If you have not tried already: echo "- - -" > /sys/class/scsi_host/host0/scan (& port/connector might want to give/chose different)___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] newly created clone waits for off-lined node
On 13/07/2023 17:33, Ken Gaillot wrote: On Wed, 2023-07-12 at 21:08 +0200, lejeczek via Users wrote: Hi guys. I have a fresh new 'galera' clone and that one would not start & cluster says: ... INFO: Waiting on node to report database status before Master instances can start. ... Is that only for newly created resources - which I guess it must be - and if so then why? Naturally, next question would be - how to make such resource start in that very circumstance? many thank, L. That is part of the agent rather than Pacemaker. Looking at the agent code, it's based on a node attribute the agent sets, so it is only empty for newly created resources that haven't yet run on a node. I'm not sure if there's a way around it. (Anyone else have experience with that?) any expert/devel would agree - this qualifies as a bug? To add to my last message - mariadb starts a okey outside of the cluster but also if I manually change 'safe_to_bootstrap' to '1' on any of the - in my case - two nodes, then cluster will also start the clone ok. But then... disable(seems that Galera cluster gets shut down ok) & enable the clone and clone fails to start with those errors as earlier - waiting for off-lined node - and leaving the clone as 'Unpromoted' ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: SATA drive rescan?
On 16/07/2023 08:27, ToddAndMargo via users wrote: Hi All, I have a removable drive sleeve that I mount SATA drives mounted in that attaches to an internal SATA port. I use it backup my system to removable drives. The mechanism works identical to USB flash drives. With one exception. A drive MUST be in the carriage when I boot up. Otherwise, if I insert a drive AFTER the system is booted up, the drive is invisible to the system. Now I have seen Clonzilla re-educating the system as to what drives are where. Is there a way to goose Fedora to do the same? Is there a way to rescan without having to reboot? Many thanks, -T Bit better, clearer description of the hardware/caddie/connection you should have started with. So, it is purely all SATA (& not! e-sata) and no bridge of any kind in-between? If so, if purely SATA, I'd suggest you check - if have not done so yet- your mobo's BIOS for something like SATA devices "hot-plugging" ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] newly created clone waits for off-lined node
On 13/07/2023 17:33, Ken Gaillot wrote: On Wed, 2023-07-12 at 21:08 +0200, lejeczek via Users wrote: Hi guys. I have a fresh new 'galera' clone and that one would not start & cluster says: ... INFO: Waiting on node to report database status before Master instances can start. ... Is that only for newly created resources - which I guess it must be - and if so then why? Naturally, next question would be - how to make such resource start in that very circumstance? many thank, L. That is part of the agent rather than Pacemaker. Looking at the agent code, it's based on a node attribute the agent sets, so it is only empty for newly created resources that haven't yet run on a node. I'm not sure if there's a way around it. (Anyone else have experience with that?) But this is bit weird -- even with 'promoted-max' & 'clone-max' to exclude (& on top of it ban the resource on the) off-lined node, clone would not start -- no? Perhaps if not a an obvious bug, I'd think this could/should be improved, the code. If anybody wanted to reproduce - have a "regular" galera clone okey, off-line a node, remove re/create the clone. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] newly created clone waits for off-lined node
Hi guys. I have a fresh new 'galera' clone and that one would not start & cluster says: ... INFO: Waiting on node to report database status before Master instances can start. ... Is that only for newly created resources - which I guess it must be - and if so then why? Naturally, next question would be - how to make such resource start in that very circumstance? many thank, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] location constraint does not move promoted resource ?
On 03/07/2023 18:55, Andrei Borzenkov wrote: On 03.07.2023 19:39, Ken Gaillot wrote: On Mon, 2023-07-03 at 19:22 +0300, Andrei Borzenkov wrote: On 03.07.2023 18:07, Ken Gaillot wrote: On Mon, 2023-07-03 at 12:20 +0200, lejeczek via Users wrote: On 03/07/2023 11:16, Andrei Borzenkov wrote: On 03.07.2023 12:05, lejeczek via Users wrote: Hi guys. I have pgsql with I constrain like so: -> $ pcs constraint location PGSQL-clone rule role=Promoted score=-1000 gateway-link ne 1 and I have a few more location constraints with that ethmonitor & those work, but this one does not seem to. When contraint is created cluster is silent, no errors nor warning, but relocation does not take place. I can move promoted resource manually just fine, to that node where 'location' should move it. Instance to promote is selected according to promotion scores which are normally set by resource agent. Documentation implies that standard location constraints are also taken in account, but there is no explanation how promotion scores interoperate with location scores. It is possible that promotion score in this case takes precedence. It seems to have kicked in with score=-1 but.. that was me just guessing. Indeed it would be great to know how those are calculated, in a way which would' be admin friendly or just obvious. thanks, L. It's a longstanding goal to have some sort of tool for explaining how scores interact in a given situation. However it's a challenging problem and there's never enough time ... Basically, all scores are added together for each node, and the node with the highest score runs the resource, subject to any placement strategy configured. These mainly include stickiness, location constraints, colocation constraints, and node health. Nodes may be And you omitted the promotion scores which was the main question. Oh right -- first, the above is used to determine the nodes on which clone instances will be placed. After that, an appropriate number of nodes are selected for the promoted role, based on promotion scores and location and colocation constraints for the promoted role. I am sorry but it does not really explain anything. Let's try concrete examples a) master clone instance has location score -1000 for a node and promotion score 1000. Is this node eligible for promoting clone instance (assuming no other scores are present)? b) promotion score is equal on two nodes A and B, but node A has better location score than node B. Is it guaranteed that clone will be promoted on A? a real-life example: ...Colocation Constraints: Started resource 'HA-10-1-1-253' with Promoted resource 'PGSQL-clone' (id: colocation-HA-10-1-1-253-PGSQL-clone-INFINITY) score=INFINITY .. Order Constraints: promote resource 'PGSQL-clone' then start resource 'HA-10-1-1-253' (id: order- PGSQL-clone-HA-10-1-1-253-Mandatory) symmetrical=0 kind=Mandatory demote resource 'PGSQL-clone' then stop resource 'HA-10-1-1-253' (id: order- PGSQL-clone-HA-10-1-1-253-Mandatory-1) symmetrical=0 kind=Mandatory I had to bump this one up to: ... resource 'PGSQL-clone' (id: location-PGSQL-clone) Rules: Rule: role=Promoted score=-1 (id: location-PGSQL-clone-rule) Expression: gateway-link ne 1 (id: location-PGSQL-clone-rule-expr) '-1000' did not seem to be good enough, '-1' was just a "lucky" guess. as earlier: I was able to 'move' the promoted and I think 'prefers' also worked. I don't know it 'pgsql' would work with any other constraints, if it was safe to try so. many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [Freeipa-users] FreeIPA on Fedora and dnf system-upgrade
On 03/07/2023 15:05, Ian Pilcher wrote: On 7/3/23 07:08, lejeczek via users wrote: I'm not sure what you mean by 'in-place' do not work? You cannot update an EL 7 FreeIPA server to EL 8, nor can you update from EL 8 to EL 9. The only supported path is to add a new replica, running the new version, and then retire the old server. (If you want the server to keep the same hostname, you then need to repeat the process.) Ah, sure, that will probably never change for Centos. I think more adequate is, as they call it, 'rolling/on' release or point-release-update, rather then in-place updates, which was a bit confusing. I'd suggest containers as a way out if it, perhaps VMs if can afford resource-wise. Sure containers are also subject to generally understood "up-grade/date" but might be easier to manage in this & migrate to/over respects - I'm inclined to claim. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [Freeipa-users] FreeIPA on Fedora and dnf system-upgrade
On 29/06/2023 22:32, Ian Pilcher via FreeIPA-users wrote: I am currently running FreeIPA on CentOS 7, and I am considering moving it to Fedora. On RHEL and derivatives, in-place upgrades are not supported. It is necessary to provision a new server, running the new OS version, add it as a FreeIPA replica, and then decommission the old system. How does this work on Fedora? Will I be able to use dnf system-upgrade, or will I find myself having to use the process described above? I'm not sure what you mean by 'in-place' do not work? My c7 days are quite a way behind but I don't remember it did no work. c8s & c9s I've been doing up-dates/grades in place for a long time and sure I might have seen some glitches - but by intent packages are ready to upgrade already deployed IPA - but that goes for any software. Then, usual best-practices - however you worked out those - such as backups, apply, how you control & order IPA packages upgrades, etc. IPA has own 'ipa-server-upgrade' which is part of packages/version upgrade process(es) Fedora's IPA stack works okey as it does on Centos - which Centos is Fedora pretty much, only back-dated/ported at it's core. Centos 9 which I've been riding for long time is great, been updating IPA since c9s release - in fact I've recently, though for different reasons, migrated IPA into containers which are Fedoras - no problems. regards, L. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] location constraint does not move promoted resource ?
On 03/07/2023 11:16, Andrei Borzenkov wrote: On 03.07.2023 12:05, lejeczek via Users wrote: Hi guys. I have pgsql with I constrain like so: -> $ pcs constraint location PGSQL-clone rule role=Promoted score=-1000 gateway-link ne 1 and I have a few more location constraints with that ethmonitor & those work, but this one does not seem to. When contraint is created cluster is silent, no errors nor warning, but relocation does not take place. I can move promoted resource manually just fine, to that node where 'location' should move it. Instance to promote is selected according to promotion scores which are normally set by resource agent. Documentation implies that standard location constraints are also taken in account, but there is no explanation how promotion scores interoperate with location scores. It is possible that promotion score in this case takes precedence. It seems to have kicked in with score=-1 but.. that was me just guessing. Indeed it would be great to know how those are calculated, in a way which would' be admin friendly or just obvious. thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] location constraint does not move promoted resource ?
Hi guys. I have pgsql with I constrain like so: -> $ pcs constraint location PGSQL-clone rule role=Promoted score=-1000 gateway-link ne 1 and I have a few more location constraints with that ethmonitor & those work, but this one does not seem to. When contraint is created cluster is silent, no errors nor warning, but relocation does not take place. I can move promoted resource manually just fine, to that node where 'location' should move it. All thoughts share are much appreciated. many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: SOLUTION: Re: journal / systemd colours ?
On 28/06/2023 20:26, stan via users wrote: On Wed, 28 Jun 2023 11:10:05 -0700 stan wrote: On Wed, 28 Jun 2023 10:17:45 -0700 stan via users wrote: On Wed, 28 Jun 2023 17:08:38 +0100 Patrick O'Callaghan wrote: See also 'man journalctl' and search for PAGER. This is also relevant. Unfortunately, it seems that journalctl ignores the input of SYSTEMD_LESS. It is hard to tell, because there is no example of the kind of input it wants or whether it only accepts options without arguments, as the default options all are. I would probably have to look at the journalctl code to determine if they are hard coded into the program. There is no configuration file to set them in any case. So, the code for journalctl shows it reading the options from the variables, and invoking less with those options, but it doesn't seem to affect the behavior of running journalctl, or at least the colors don't change. Maybe I'll look more closely at some point, but no joy for now. Operator error. Exporting the wrong name because of a cut and paste. Once I fixed that, definitely works to change colors in journalctl output, will have to tune it to get what I want. I put this in my .bashrc so everything is set on login. SYSTEMD_PAGER=less export SYSTEMD_PAGER SYSTEMD_LESS="[list of less options]" export SYSTEMD_LESS ___ Not exactly, I'd not think of it as _the_ solution - (I much prefer to up pages via actual mechanical scrolling) - as I use, always I've had, SYSTEMD_PAGER=cat so... man page for 'journalctl' has a shor section: When outputting to a tty, lines are colored according to priority: lines of level ERROR and higher are colored red; lines of level NOTICE and higher are highlighted; lines of level DEBUG are colored lighter grey; other lines are displayed normally. would be nice to be able to customize those & if 'systemd' delegates declaration of that 'highlighting' colour then these below do not do it: a) terminal-colors.d - perhaps systemd/journal ignores it altogether b) gnome-terminal has config for 'Highlight colour' It would be great/the best to have that functionality internal to systemd/journal - thus, if authors/devel might read here - please think of this conversation's subject as possible future addition/enhancement to the software. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
journal / systemd colours ?
Hi guys. I wonder if somebody might know how to customize, to change systemd journal's colours? I'm specifically thinking of changing 'WARN' warnings which get - in my opinion unfortunate, for I use other than white background colours - painted in yellow. many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ClusterLabs] silence resource ? - PGSQL
Hi guys. Having 'pgsql' set up in what I'd say is a vanilla-default confg, pacemaker's journal log is flooded with: ... pam_unix(runuser:session): session closed for user postgres pam_unix(runuser:session): session opened for user postgres(uid=26) by (uid=0) pam_unix(runuser:session): session closed for user postgres pam_unix(runuser:session): session opened for user postgres(uid=26) by (uid=0) pam_unix(runuser:session): session closed for user postgres pam_unix(runuser:session): session opened for user postgres(uid=26) by (uid=0) pam_unix(runuser:session): session closed for user postgres ... Would you have a working fix or even a suggestion on how to silence those? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Thunderbird adaptive junk control - does it work?
Hi guys. I have - as you can probably see - a Yahoo account and there, on the server I have a all my filters/rules set and those work, say.. another mailing messages list a filter/rule puts meails where I told it to, to a folder, I can confirm that when I go to Yahoo's web mail, but... Thunderbird keeps putting those message into "Bulk"(spam) folder even though I told Thunderbird not to - by marking message as 'not spam', a number of times already. I also have "Trust junk mail headers" de-activated, so it should be entirely on Thunderbird's internal business logic to spam counting. I don't think all this works in Thunderbird or if it does then it does it extremely poorly. What is your experience I wonder & I wonder if there is something, some setting/tweaking I missed which can help? (I want to avoid putting addresses into my adrress book only for the reason of helping spam filtering) many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: Gnome Nautilus and no thumbnails from connected Android storage
On 21/10/2021 11:39, lejeczek via users wrote: Hi guys. I asked this a while ago but nobody got interested yet I'm really curious - when you connect your Android Phone via USB, do you get thumbnails? I cannot get my F34's Nautilus to show thumbnails off my Huawei Honor 10 with Android 10. thanks, L. We have thumbnails back in f38 - at least on my Oppo have them - so manyyy thanks to those responsible for fixing this. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?
On 09/06/2023 09:04, Reid Wahl wrote: On Thu, Jun 8, 2023 at 10:55 PM lejeczek via Users wrote: On 09/06/2023 01:38, Reid Wahl wrote: On Thu, Jun 8, 2023 at 2:24 PM lejeczek via Users wrote: Ouch. Let's see the full output of the move command, with the whole CIB that failed to validate. For a while there I thought perhaps it was just that one pglsq resource, but it seems that any - though only a few are set up - (only clone promoted?)resource fails to move. Perhaps primarily to do with 'pcs' -> $ pcs resource move REDIS-clone --promoted podnode3 Error: cannot move resource 'REDIS-clone' 1 This is the problem: `validate-with="pacemaker-3.6"`. That old schema doesn't support role="Promoted" in a location constraint. Support begins with version 3.7 of the schema: https://github.com/ClusterLabs/pacemaker/commit/e7f1424df49ac41b2d38b72af5ff9ad5121432d2. You'll need at least Pacemaker 2.1.0. I have: corosynclib-3.1.7-1.el9.x86_64 corosync-3.1.7-1.el9.x86_64 pacemaker-schemas-2.1.6-2.el9.noarch pacemaker-libs-2.1.6-2.el9.x86_64 pacemaker-cluster-libs-2.1.6-2.el9.x86_64 pacemaker-cli-2.1.6-2.el9.x86_64 pacemaker-2.1.6-2.el9.x86_64 pcs-0.11.5-2.el9.x86_64 pacemaker-remote-2.1.6-2.el9.x86_64 and the reset is Centos 9 up-to-what-is-in-repos If all your cluster nodes are at those versions, try `pcs cluster cib-upgrade` 'cib-upgrade' might be a fix for this "issue" - but should it not happen at rpm update time? Differently but still 'move' errored out: -> $ pcs resource move REDIS-clone --promoted podnode2 Location constraint to move resource 'REDIS-clone' has been created Waiting for the cluster to apply configuration changes... Location constraint created to move resource 'REDIS-clone' has been removed Waiting for the cluster to apply configuration changes... Error: resource 'REDIS-clone' is promoted on node 'podnode3'; unpromoted on nodes 'podnode1', 'podnode2' Error: Errors have occurred, therefore pcs is unable to continue Then a node-promted-resource going into 'standby' or being rebooted - for first/one time - seemed to "fix" 'move'. -> $ pcs resource move REDIS-clone --promoted podnode2 Location constraint to move resource 'REDIS-clone' has been created Waiting for the cluster to apply configuration changes... Location constraint created to move resource 'REDIS-clone' has been removed Waiting for the cluster to apply configuration changes... resource 'REDIS-clone' is promoted on node 'podnode2'; unpromoted on nodes 'podnode1', 'podnode3' quite puzzling, sharing in case others experience this/similar. thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?
On 09/06/2023 01:38, Reid Wahl wrote: On Thu, Jun 8, 2023 at 2:24 PM lejeczek via Users wrote: Ouch. Let's see the full output of the move command, with the whole CIB that failed to validate. For a while there I thought perhaps it was just that one pglsq resource, but it seems that any - though only a few are set up - (only clone promoted?)resource fails to move. Perhaps primarily to do with 'pcs' -> $ pcs resource move REDIS-clone --promoted podnode3 Error: cannot move resource 'REDIS-clone' 1 This is the problem: `validate-with="pacemaker-3.6"`. That old schema doesn't support role="Promoted" in a location constraint. Support begins with version 3.7 of the schema: https://github.com/ClusterLabs/pacemaker/commit/e7f1424df49ac41b2d38b72af5ff9ad5121432d2. You'll need at least Pacemaker 2.1.0. I have: corosynclib-3.1.7-1.el9.x86_64 corosync-3.1.7-1.el9.x86_64 pacemaker-schemas-2.1.6-2.el9.noarch pacemaker-libs-2.1.6-2.el9.x86_64 pacemaker-cluster-libs-2.1.6-2.el9.x86_64 pacemaker-cli-2.1.6-2.el9.x86_64 pacemaker-2.1.6-2.el9.x86_64 pcs-0.11.5-2.el9.x86_64 pacemaker-remote-2.1.6-2.el9.x86_64 and the reset is Centos 9 up-to-what-is-in-repos 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216
Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?
Ouch. Let's see the full output of the move command, with the whole CIB that failed to validate. For a while there I thought perhaps it was just that one pglsq resource, but it seems that any - though only a few are set up - (only clone promoted?)resource fails to move. Perhaps primarily to do with 'pcs' -> $ pcs resource move REDIS-clone --promoted podnode3 Error: cannot move resource 'REDIS-clone' 1 validate-with="pacemaker-3.6" epoch="8212" num_updates="0" admin_epoch="0" cib-last-written="Thu Jun 8 21:59:53 2023" update-origin="podnode1" update-client="crm_attribute" have-quorum="1" update-user="root" dc-uuid="1"> 2 3 4 5 id="cib-bootstrap-options-have-watchdog" name="have-watchdog" value="false"/> 6 name="dc-version" value="2.1.6-2.el9-6fdc9deea29"/> 7 id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/> 8 id="cib-bootstrap-options-cluster-name" name="cluster-name" value="podnodes"/> 9 id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"/> 10 id="cib-bootstrap-options-last-lrm-refresh" name="last-lrm-refresh" value="1686047745"/> 11 id="cib-bootstrap-options-maintenance-mode" name="maintenance-mode" value="false"/> 12 13 14 name="REDIS_REPL_INFO" value="podnode1"/> 15 16 17 18 19 20 name="PGSQL-data-status" value="DISCONNECT"/> 21 22 23 24 25 name="PGSQL-data-status" value="DISCONNECT"/> 26 27 28 29 30 name="PGSQL-data-status" value="LATEST"/> 31 32 33 34 35 provider="heartbeat" type="IPaddr2"> 36 id="HA-10-1-1-226-meta_attributes"> 37 id="HA-10-1-1-226-meta_attributes-failure-timeout" name="failure-timeout" value="30s"/> 38 39 id="HA-10-1-1-226-instance_attributes"> 40 id="HA-10-1-1-226-instance_attributes-cidr_netmask" name="cidr_netmask" value="24"/> 41 id="HA-10-1-1-226-instance_attributes-ip" name="ip" value="10.1.1.226"/> 42 43 44 interval="10s" name="monitor" timeout="20s"/> 45 interval="0s" name="start" timeout="20s"/> 46 interval="0s" name="stop" timeout="20s"/> 47 48 49 provider="heartbeat" type="IPaddr2"> 50 id="HA-10-3-1-226-meta_attributes"> 51 id="HA-10-3-1-226-meta_attributes-failure-timeout" name="failure-timeout" value="30s"/> 52 53 id="HA-10-3-1-226-instance_attributes"> 54 id="HA-10-3-1-226-instance_attributes-cidr_netmask" name="cidr_netmask" value="24"/> 55 id="HA-10-3-1-226-instance_attributes-ip" name="ip" value="10.3.1.226"/> 56 57 58 interval="10s" name="monitor" timeout="20s"/> 59 interval="0s" name="start" timeout="20s"/> 60 interval="0s" name="stop" timeout="20s"/> 61 62 63 64 provider="heartbeat"> 65 id="REDIS-instance_attributes"> 66 name="bin" value="/usr/bin/redis-server"/> 67 id="REDIS-instance_attributes-client_bin" name="client_bin" value="/usr/bin/redis-cli"/> 68 id="REDIS-instance_attributes-config" name="config" value="/etc/redis/redis.conf"/> 69 id="REDIS-instance_attributes-rundir" name="rundir" value="/run/redis"/> 70 id="REDIS-instance_attributes-user" name="user" value="redis"/> 71 id="REDIS-instance_attributes-wait_last_known_master" name="wait_last_known_master" value="true"/> 72 73 74 timeout="120s" id="REDIS-demote-interval-0s"/> 75 interval="45s" id="REDIS-monitor-interval-45s"> 76 id="REDIS-monitor-interval-45s-instance_attributes"> 77 id="REDIS-monitor-interval-45s-instance_attributes-OCF_CHECK_LEVEL" name="OCF_CHECK_LEVEL" value="0"/> 78 79 80 timeout="60s" interval="20s" id="REDIS-monitor-interval-20s"> 81 id="REDIS-monitor-interval-20s-instance_attributes"> 82 id="REDIS-monitor-interval-20s-instance_attributes-OCF_CHECK_LEVEL" name="OCF_CHECK_LEVEL" value="0"/> 83 84 85 timeout="60s" interval="60s" id="REDIS-monitor-interval-60s"> 86 id="REDIS-monitor-interval-60s-instance_attributes"> 87 id="REDIS-monitor-interval-60s-instance_attributes-OCF_CHECK_LEVEL" name="OCF_CHECK_LEVEL" value="0"/> 88 89 90 timeout="90s" id="REDIS-notify-interval-0s"/> 91 timeout="120s" id="REDIS-promote-interval-0s"/>
Re: gimmicks for most & a question networking curious/experts - ? MACSEC
de On 06/06/2023 14:50, Chris Adams wrote: Once upon a time, lejeczek said: To start with the most basic of what I ponder over - basic in my mind though I get, that it does not have to be that or simple/obvious - is MACSEC with 'bond' as parent or in other words: macsec "on" a 'bond' network interface. Should such a "thing" work, does it work? My understanding is that you do MACsec on physical interfaces, so you would configure it on each member of a LAG, not the virtual LAG interface (e.g. bond0) itself. So I started from the bottom, went to do 'macsec' physical eth interfaces but when under 'bond' in "broadcast" I got NM failing to set such macsec ifaces up: ... [1686137957.1794] device (macsec-ten-bott): supplicant interface state: disconnected -> completed [1686137957.1829] device (macsec-ten-bott): carrier: link connected [1686137957.1829] device (macsec-ten-bott): Activation: Stage 2 of 5 (Device Configure) successful. [1686137957.2081] device (macsec-ten-bott): state change: config -> ip-config (reason 'none', sys-iface-state: 'managed') [1686137957.2149] device (macsec-ten-bott): Activation: connection 'macsec-10.1.1-slave-bott' could not be enslaved [1686137957.2150] device (macsec-ten-bott): state change: ip-config -> failed (reason 'unknown', sys-iface-state: 'managed') [1686137957.2152] device (macsec-ten-bott): released from master device bond-1011 [1686137957.2153] device (macsec-ten-bott): Activation: failed for connection 'macsec-10.1.1-slave-bott' [1686137957.2156] device (macsec-ten-bott): state change: failed -> disconnected (reason 'none', sys-iface-state: 'managed') [1686137957.2402] device (macsec-ten-bott): state change: disconnected -> unmanaged (reason 'user-requested', sys-iface-state: 'managed') And to share what I find - obvious to those with expertise - that 'macsec' "on" a bond iface (broadcast mode is my case!) works, or seems to work I shall add as it's less than 1hr of my fiddling. But I'd rather see physical MACs "macseced" and then these being bonded/bridged - as going to the lowest layer/device in a stack puts my mind at ease the best. I should add - I do all that in Centos 9. thanks, L. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
gimmicks for most & a question networking curious/experts - ? MACSEC
Hi guys. Looking at macsec I've only started - so go easy on me with my possibly trivial questions - and write here in hope, that some of you have expertise to tell... To start with the most basic of what I ponder over - basic in my mind though I get, that it does not have to be that or simple/obvious - is MACSEC with 'bond' as parent or in other words: macsec "on" a 'bond' network interface. Should such a "thing" work, does it work? All & any thoughts shared are much appreciated. many thanks, L. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?
On 05/06/2023 16:23, Ken Gaillot wrote: On Sat, 2023-06-03 at 15:09 +0200, lejeczek via Users wrote: Hi guys. I've something which I'm new to entirely - cluster which is seemingly okey errors, fails to move a resource. What pcs version are you using? I believe there was a move regression in a recent push. I'd won't contaminate here just yet with long json cluster spits when fails but a snippet: -> $ pcs resource move PGSQL-clone --promoted podnode1 Error: cannot move resource 'PGSQL-clone' 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 ... crm_resource: Error performing operation: Invalid configuration This one line: (might be more) puzzles me, as there is no such node/member in the cluster and so I try: That's not a problem. Pacemaker allows "custom" values in both cluster options and resource/action meta-attributes. I don't know whether redis is actually using that or not. -> $ pcs property unset redis_REPL_INFO --force Warning: Cannot remove property 'redis_REPL_INFO', it is not present in property set 'cib-bootstrap-options' That's because the custom options are in their own cluster_property_set. I believe pcs can only manage the options in the cluster_property_set with id="cib-bootstrap-options", so you'd have to use "pcs cluster edit" or crm_attribute to remove the custom ones. Any & all suggestions on how to fix this are much appreciated. many thanks, L. I've downgraded back to: pacemaker-2.1.6-1.el9.x86_64 pcs-0.11.4-7.el9.x86_64 but, it's either not enough - if bugs are in those that is - or issues are somewhere else. for 'move' still fails the same: ... crm_resource: Error performing operation: Invalid configuration ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?
On 05/06/2023 16:23, Ken Gaillot wrote: On Sat, 2023-06-03 at 15:09 +0200, lejeczek via Users wrote: Hi guys. I've something which I'm new to entirely - cluster which is seemingly okey errors, fails to move a resource. What pcs version are you using? I believe there was a move regression in a recent push. I'd won't contaminate here just yet with long json cluster spits when fails but a snippet: -> $ pcs resource move PGSQL-clone --promoted podnode1 Error: cannot move resource 'PGSQL-clone' 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 ... crm_resource: Error performing operation: Invalid configuration This one line: (might be more) puzzles me, as there is no such node/member in the cluster and so I try: That's not a problem. Pacemaker allows "custom" values in both cluster options and resource/action meta-attributes. I don't know whether redis is actually using that or not. -> $ pcs property unset redis_REPL_INFO --force Warning: Cannot remove property 'redis_REPL_INFO', it is not present in property set 'cib-bootstrap-options' That's because the custom options are in their own cluster_property_set. I believe pcs can only manage the options in the cluster_property_set with id="cib-bootstrap-options", so you'd have to use "pcs cluster edit" or crm_attribute to remove the custom ones. Any & all suggestions on how to fix this are much appreciated. many thanks, L. Well. that would make sense - those must have gone in & through to the repos' rpms - as I suspected thos broke after recent 'dnf' updates. I'll fiddle with 'downgrades' later on - but if true, then rather critical, no? ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] cluster okey but errors when tried to move resource - ?
Hi guys. I've something which I'm new to entirely - cluster which is seemingly okey errors, fails to move a resource. I'd won't contaminate here just yet with long json cluster spits when fails but a snippet: -> $ pcs resource move PGSQL-clone --promoted podnode1 Error: cannot move resource 'PGSQL-clone' 1 validate-with="pacemaker-3.6" epoch="8109" num_updates="0" admin_epoch="0" cib-last-written="Sat Jun 3 13:49:34 2023" update-origin="podnode2" update-client="cibadmin" have-quorum="1" update-user="root" dc-uuid="2"> 2 3 4 5 id="cib-bootstrap-options-have-watchdog" name="have-watchdog" value="false"/> 6 name="dc-version" value="2.1.6-1.el9-802a72226be"/> 7 id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/> 8 id="cib-bootstrap-options-cluster-name" name="cluster-name" value="podnodes"/> 9 id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"/> 10 id="cib-bootstrap-options-last-lrm-refresh" name="last-lrm-refresh" value="1683293193"/> 11 id="cib-bootstrap-options-maintenance-mode" name="maintenance-mode" value="false"/> 12 13 14 name="redis_REPL_INFO" value="c8kubernode1"/> 15 name="REDIS_REPL_INFO" value="podnode3"/> 16 17 ... crm_resource: Error performing operation: Invalid configuration This one line: (might be more) name="redis_REPL_INFO" value="c8kubernode1"/> puzzles me, as there is no such node/member in the cluster and so I try: -> $ pcs property unset redis_REPL_INFO --force Warning: Cannot remove property 'redis_REPL_INFO', it is not present in property set 'cib-bootstrap-options' Any & all suggestions on how to fix this are much appreciated. many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: PS1 not in .bash_profile but somewhere else ?
On 02/06/2023 13:56, lejeczek via users wrote: Hi guys I have one specific user whose PS1 if not set in .bash_profile, is taken from somewhere else. At the same time globally I set PS1 in /etc/environment - bu that, this one user does not honor. I've tried a few terminal/console apps - no difference. Where else can PS1 be set and/or how to find out here PS1 for a user/account was set? many thanks, L. A bit more fiddling.. I don't know what role in it gnome-terminal plays but, it seems that PS1 gets "global" at Gnome login time and user inherits that, if set in '/etc/environment; , if PS1 set specifically in user "profile", then takes expected effect. If set only in '/etc/environment' , PS1 is read by gnome/terminal but is malformed, as not digested, raw - suffices to 'ssh localhost' from the same gnome terminal and PS1 is set, displayed properly In terminal, "Run command as login shell" is ON for all my fiddling and makes no difference. It smells to me like a bug, I'd not presume it's intentional. thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
PS1 not in .bash_profile but somewhere else ?
Hi guys I have one specific user whose PS1 if not set in .bash_profile, is taken from somewhere else. At the same time globally I set PS1 in /etc/environment - bu that, this one user does not honor. I've tried a few terminal/console apps - no difference. Where else can PS1 be set and/or how to find out here PS1 for a user/account was set? many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ovirt-users] troubleshooting SHE - Run engine-setup with answerfile - ?
Hi guys. I'm revisiting my failed attempt to make friends with oVirt - novice so go easy on me. Deployment fails sooner that what can troubleshoot - at least I manged to find in docs so far - and I wanted to ask hot to troubleshot such failure as this: ... INFO ] TASK [ovirt.ovirt.engine_setup : Run engine-setup with answerfile] [ ERROR ] fatal: [localhost -> 192.168.1.209]: FAILED! => {"changed": true, "cmd": ["engine-setup", "--accept-defaults", "--config-append=/root/ovirt-engine-answers"], "delta": "0:08:05.997322", "end": "2023-06-01 13:07:15.932517", "msg": "non-zero return code", "rc": 1, "start": "2023-06-01 12:59:09.935195", "stderr": "", "stderr_lines": [], "stdout": "[ INFO ] Stage: Initializing\n[ INFO ] Stage: Environment setup\n Configuration files:\n ... Start with setting up Keycloak for Ovirt Engine", "[ ERROR ] Failed to execute stage 'Closing up': Command '/usr/share/ovirt-engine-keycloak/bin/kk_cli.sh' failed to execute", "[ INFO ] Stage: Clean up", " Log file is located at", " /var/log/ovirt-engine/setup/ovirt-engine-setup-20230601125911-448xvj.log", ... [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "There was a failure deploying the engine on the local engine VM. The system may not be provisioned according to the playbook results: please check the logs for the issue, fix accordingly or re-deploy from scratch.\n"} [ ERROR ] Failed to execute stage 'Closing up': Failed executing ansible-playbook [ INFO ] Stage: Clean up ... Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20230601123237-rhwx8e.log Logfile to which 'stdout' points, shows nothing or I fail to find anything in it and/or understand it: ... 2023-06-01 13:07:17,716+0100 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:113 fatal: [localhost -> 192.168.1.209]: FAILED! => {"changed": true, "cmd": ["engine-setup", "--accept-defaults", "--config-append=/root/ovirt-engine-answers"], "delta": "0:08:05.997322", "end": "2023-06-01 13:07:15.932517", "msg": "non-zero return code", "rc": 1, "start": "2023-06-01 12:59:09.935195", "stderr": "", "stderr_lines": [], "stdout": ... or Ovirt Engine\n[ ERROR ] Failed to execute stage 'Closing up': Command '/usr/share/ovirt-engine-keycloak/bin/kk_cli.sh' failed to execute\n[ INFO ] Stage: Clean up\n Log file is located at\n /var/log/ovirt-engine/setup/ovirt-engine-setup-20230601125911-448xvj.log\n[ INFO ] Generating answer file '/var/lib/ovirt-engine/setup/answers/20230601130714-setup.conf'\n[ INFO ] Stage: Pre-termination\n[ INFO ] Stage: Termination\n[ ERROR ] Execution of setup failed", "stdout_lines": ["[ INFO ] Stage: Initializing", "[ INFO ] Stage: Environment setup", " Configuration files:", " /etc/ovirt-engine-setup.conf.d/10-packaging-jboss.conf,", " /etc/ovirt-engine-setup.conf.d/10-packaging.conf," ... ... 2023-06-01 13:08:52,946+0100 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 {'msg': "The tas k includes an option with an undefined variable. The error was: 'local_vm_ip' is undefined. 'local_vm_ip' is undefined\n\nThe error appears to be in '/usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_engine_setup/tasks/sync_on_engine_ma chine.yml': line 2, column 3, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appear s to be:\n\n---\n- name: Set the name for add_host\n ^ here\n", '_ansible_no_log': False} 2023-06-01 13:08:53,047+0100 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 ignored: [localh ost]: FAILED! => {"msg": "The task includes an option with an undefined variable. The error was: 'local_vm_ip' is undefined. 'local _vm_ip' is undefined\n\nThe error appears to be in '/usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_eng ine_setup/tasks/sync_on_engine_machine.yml': line 2, column 3, but may\nbe elsewhere in the file depending on the exact syntax prob lem.\n\nThe offending line appears to be:\n\n---\n- name: Set the name for add_host\n ^ here\n"} ... 2023-06-01 13:09:16,818+0100 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 {'changed': True, 'stdout': '', 'stderr': "error: Failed to destroy pool 9ab14cca-78e0-4a8e-8a1b-00e98c300208\nerror: Requested operation is not valid: storage pool '9ab14cca-78e0-4a8e-8a1b-00e98c300208' is not active", 'rc': 1, 'cmd': ['virsh', '-c', 'qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf', 'pool-destroy', '9ab14cca-78e0-4a8e-8a1b-00e98c300208'], 'start': '2023-06-01 13:09:16.614978', 'end': '2023-06-01 13:09:16.672260', 'delta': '0:00:00.057282', 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'virsh -c
switching windows behavior changed with F38 Gnome ?
Hi guys. All this years I could switch between... say Firefox and Chromium, this way: a) had a Chromium window focused-on b) had a number of Firefox windows (as well as Chromum windows) in the background c) I switched with Alt+Tab to Firefox window (whichever window was there first, does not matter) d) could switch back - all the previous Fedora/Gnome versions - to that last Chromium window, quickly & easily with Alt+Esc but now.. with an upgrade to F38, Atl+Esc goes through _all_ Firefox windows instead. @fedorians - do you also see/get this change? @devel, if read here - was this change planned, intentional? If so then why, what for? many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Re: Firefox - detection of exiting tabs with URLs
On 10/05/2023 22:17, Samuel Sieb wrote: On 5/10/23 12:47, Tim via users wrote: On Wed, 2023-05-10 at 13:04 +0200, lejeczek via users wrote: With f38 I think my Firefox is unable to detect or is ignoring tabs which are already opened in other windows with certain URL, when I open a new tab and want to go to a website/URL. I think Firefox would then say, would offer something like "switch to tab". In the past I'd noticed a behaviour that if I opened a blank tab and typed in something like facebook, let auto-complete do its thing and picked something from the drop-down list that appeared below the address bar from its history, I'd often see the browser whiz over to an existing tab which already had that site loaded. It was that or a very similar kind of situation. Look carefully at the item you're clicking. Watch for the tag that says "Switch to Tab". ___ Yes, that is exactly what we are saying, I was as the originator of this thread. I can only add - to what I said which was "this functionality is completely gone" - that now after a longer look I see "Swtich to Tab" for some site/urls, sometimes! but not the others. This would need to be thoroughly tested (probably something only devel could do) to say, if there is pattern to it at all or all this is purely random. Eg. I've had a window with a few tabs - my Firefox is set to restore last sessions - _none_ of those tabs (sites/urls were fully loaded after restoration I made sure of) was detected as "Switch to Tab" I closed that window, opened a new one (made sure that no other auto-started/restored window/tab had those sites in question/example opened) and each site, one by one loaded/opened, then tested in already existing window and only ! then that window picked up, detected and offered "Switch to Tab" to those sites. So, purely from consumer perspective I can say this "feature" is bit flaky. many thanks, L. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ovirt-users] Re: engine setup fails: error creating bridge interface virbr0: File exists - ?
On 11/05/2023 18:18, Volenbovskyi, Konstantin wrote: Hi, Is it actually first attempt to install HE? Or maybe you ran into some (different) problem and this is like second (third, fourth…) attempt. While installation script takes care of some cleanup, I would say that you should consider doing ovirt-hosted-engine-cleanup ,checking that virbr0 is absent in ‘ip a’ output and trying once again BR, Konstantin *Von: *lejeczek via Users *Antworten an: *lejeczek *Datum: *Donnerstag, 11. Mai 2023 um 13:25 *An: *"users@ovirt.org" *Betreff: *[ovirt-users] engine setup fails: error creating bridge interface virbr0: File exists - ? Hi guys. I'm trying to setup the engine on the latest stable ovirt node(in a VM), so a clean, vanilla-default system. -> $ hosted-engine --deploy --4 ... [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate default libvirt network] [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "cmd": ["virsh", "net-start", "default"], "delta": "0:00:00.042134", "end": "2023-05-11 11:08:59.248405", "msg": "non-zero return code", "rc": 1, "start": "2023-05-11 11:08:59.206271", "stderr": "error: Failed to start network default\nerror: error creating bridge interface virbr0: File exists", "stderr_lines": ["error: Failed to start network default", "error: error creating bridge interface virbr0: File exists"], "stdout": "", "stdout_lines": []} [ ERROR ] Failed to execute stage 'Closing up': Failed getting local_vm_dir ... Any & all suggestions on how to fix/troubleshoot this are much appreciated. many thanks, L. No, it is a clean install of oVirt node in a KVM guest/VM. All I do in the node is I configure two ifaces and both are "regular" ethernets, then I start HE setup. Bare-metal host is Centos 9 Stream with everything up-to-dayte off the distro repos. I'd think this should be easily reproducible. ps. please keep replies to the bottom - this is not a conversion between two of us nor any two persons, Mailing to list - think of it a book, which always reads top-to-bottom. many thanks, L.___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/A42X23EGFVJLAJHBNBXIVF7YY5UPYXW6/
[ovirt-users] engine setup fails: error creating bridge interface virbr0: File exists - ?
Hi guys. I'm trying to setup the engine on the latest stable ovirt node(in a VM), so a clean, vanilla-default system. -> $ hosted-engine --deploy --4 ... [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate default libvirt network] [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "cmd": ["virsh", "net-start", "default"], "delta": "0:00:00.042134", "end": "2023-05-11 11:08:59.248405", "msg": "non-zero return code", "rc": 1, "start": "2023-05-11 11:08:59.206271", "stderr": "error: Failed to start network default\nerror: error creating bridge interface virbr0: File exists", "stderr_lines": ["error: Failed to start network default", "error: error creating bridge interface virbr0: File exists"], "stdout": "", "stdout_lines": []} [ ERROR ] Failed to execute stage 'Closing up': Failed getting local_vm_dir ... Any & all suggestions on how to fix/troubleshoot this are much appreciated. many thanks, L.___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/PYHBYP5R6RWW72PIENVPAUG6GB67FGOI/
Re: Firefox - detection of exiting tabs with URLs
On 10/05/2023 17:41, stan via users wrote: On Wed, 10 May 2023 13:04:53 +0200 lejeczek via users wrote: Hi guys. With f38 I think my Firefox is unable to detect or is ignoring tabs which are already opened in other windows with certain URL, when I open a new tab and want to go to a website/URL. I think Firefox would then say, would offer something like "switch to tab". Do you see it it too? I did not change any settings, configuration in Firefox so I wonder it's just me. I am using nightly, the firefox beta, and it works this way in F37. I don't remember it ever working differently. Maybe an addon you were using to do this is no longer available? Here are a few add ons that will prevent this, probably more at the mozilla addons site. I haven't used these. https://addons.mozilla.org/en-US/firefox/addon/dont-open-duplicate-tabs/?utm_source=addons.mozilla.org_medium=referral_content=search https://addons.mozilla.org/en-US/firefox/addon/smart-prevent-duplicate-tabs/?utm_source=addons.mozilla.org_medium=referral_content=search https://addons.mozilla.org/en-US/firefox/addon/duplicate-tabs-closer-updated/?utm_source=addons.mozilla.org_medium=referral_content=search And this more extensive add on seems to be able to do so also. https://addons.mozilla.org/en-US/firefox/addon/foxytab/?utm_source=addons.mozilla.org_medium=referral_content=search ___ No add-ons in my case, only a few but those I've had for long long time. add-ons: duckduckgo privacy, cookie manager, gab. privacy pass from cloudflare and obviously gnome shell integation. Firefox version firefox-112.0.2-1.fc38.x86_64 I believe that "functionality" was built-in and vanilla-defaul behavior. I've got the feeling something to do with f38 upgrade I did recently. ___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
Firefox - detection of exiting tabs with URLs
Hi guys. With f38 I think my Firefox is unable to detect or is ignoring tabs which are already opened in other windows with certain URL, when I open a new tab and want to go to a website/URL. I think Firefox would then say, would offer something like "switch to tab". Do you see it it too? I did not change any settings, configuration in Firefox so I wonder it's just me. many thanks, L.___ users mailing list -- users@lists.fedoraproject.org To unsubscribe send an email to users-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue
[ovirt-users] barely started - cannot import name 'Callable' from 'collections'
Hi guys. I've barely started, trying to deploy my first oVirt and I get: ... Please indicate the gateway IP address [10.3.1.254]: [ INFO ] Checking available network interfaces: [ ERROR ] b'[WARNING]: Skipping plugin (/usr/share/ovirt-hosted-engine-\n' [ ERROR ] b'setup/he_ansible/callback_plugins/2_ovirt_logger.py), cannot load: cannot\n' [ ERROR ] b"import name 'Callable' from 'collections'\n" [ ERROR ] b'(/usr/lib64/python3.11/collections/__init__.py)\n' [ ERROR ] b"ERROR! Unexpected Exception, this is probably a bug: cannot import name 'Callable' from 'collections' (/usr/lib64/python3.11/collections/__init__.py)\n" [ ERROR ] Failed to execute stage 'Environment customization': Failed executing ansible-playbook [ INFO ] Stage: Clean up [ INFO ] Cleaning temporary resources [ ERROR ] b'[WARNING]: Skipping plugin (/usr/share/ovirt-hosted-engine-\n' [ ERROR ] b'setup/he_ansible/callback_plugins/2_ovirt_logger.py), cannot load: cannot\n' [ ERROR ] b"import name 'Callable' from 'collections'\n" [ ERROR ] b'(/usr/lib64/python3.11/collections/__init__.py)\n' [ ERROR ] b"ERROR! Unexpected Exception, this is probably a bug: cannot import name 'Callable' from 'collections' (/usr/lib64/python3.11/collections/__init__.py)\n" [ ERROR ] Failed to execute stage 'Clean up': Failed executing ansible-playbook [ INFO ] Generating answer file '/var/lib/ovirt-hosted-engine-setup/answers/answers-20230509193552.conf' [ INFO ] Stage: Pre-termination [ INFO ] Stage: Termination [ ERROR ] Hosted Engine deployment failed Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20230509193544-s72umf.log centos 9 stream with ovirt-engine-setup-base-4.5.3.1-1.el9.noarch Any & every suggestion on what is braking here and how to troubleshoot/fix it are much appreciated. thanks, L.___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/KYRUZEF7MWYETFY3CVAPOR4UDESUZLGY/
[ClusterLabs] ocf_heartbeat_pgsql - lock file
Hi guys. I have a resource seemingly running a ok but when a node gets rebooted then cluster finds it not able, not good to start the resource. Failed Resource Actions: * PGSQL start on podnode3 returned 'error' (My data may be inconsistent. You have to remove /var/lib/pgsql/tmp/PGSQL.lock file to force start.) at Sun May 7 11:48:43 2023 after 121ms and indeed, with manual intervention, after removal of that file, cluster seems to be happy to rejoin the node into pgsql cluster. Is that intentional, by design and if yes/no then why it happens? many thanks, L.___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] OCF_HEARTBEAT_PGSQL - any good with current Postgres- ?
On 05/05/2023 10:41, Jehan-Guillaume de Rorthais wrote: On Fri, 5 May 2023 10:08:17 +0200 lejeczek via Users wrote: On 25/04/2023 14:16, Jehan-Guillaume de Rorthais wrote: Hi, On Mon, 24 Apr 2023 12:32:45 +0200 lejeczek via Users wrote: I've been looking up and fiddling with this RA but unsuccessfully so far, that I wonder - is it good for current versions of pgSQLs? As far as I know, the pgsql agent is still supported, last commit on it happen in Jan 11th 2023. I don't know about its compatibility with latest PostgreSQL versions. I've been testing it many years ago, I just remember it was quite hard to setup, understand and manage from the maintenance point of view. Also, this agent is fine in a shared storage setup where it only start/stop/monitor the instance, without paying attention to its role (promoted or not). It's not only that it's hard - which is purely due to piss-poor man page in my opinion - but it really sounds "expired". I really don't know. My feeling is that the manpage might be expired, which really doesn't help with this agent, but not the RA itself. Eg. man page speaks of 'recovery.conf' which - as I understand it - newer/current versions of pgSQL do not! even use... which makes one wonder. This has been fixed in late 2019, but with no documentation associated :/ See: https://github.com/ClusterLabs/resource-agents/commit/a43075be72683e1d4ddab700ec16d667164d359c Regards, Right.. like the rest of us admins/users going to the code is the very first thing we do :) RA/resource seems to work but I sincerely urge the programmer(s)/author(s) responsible, to update man pages so they reflect state of affairs as of today - there is nothing more discouraging - to the rest of us sysadmins/endusers - than man pages which look like stingy conservatives programmer's note. What setup of this RA does, is misleading in parts - still creates config part in a file called 'recovery.conf' on that note - does this RA not require way to much open pgSQL setup, meaning not very secure? Would anybody know? I cannot see - again ! regular man pages and not the code - how replication could be secured down by not using user 'postgres' itself and then perhaps adding authentication with passwords, at least. many thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] 99-VirtualDomain-libvirt.conf under control - ?
On 05/05/2023 10:08, Andrei Borzenkov wrote: On Fri, May 5, 2023 at 11:03 AM lejeczek via Users wrote: On 29/04/2023 21:02, Reid Wahl wrote: On Sat, Apr 29, 2023 at 3:34 AM lejeczek via Users wrote: Hi guys. I presume these are a consequence of having resource of VirtuaDomain type set up(& enabled) - but where, how cab users control presence & content of those? Yep: https://github.com/ClusterLabs/resource-agents/blob/v4.12.0/heartbeat/VirtualDomain#L674-L680 You can't really control the content, since it's set by the resource agent. (You could change it after creation but that defeats the purpose.) However, you can view it at /run/systemd/system/resource-agents-deps.target.d/libvirt.conf. You can see the systemd_drop_in definition here: https://github.com/ClusterLabs/resource-agents/blob/v4.12.0/heartbeat/ocf-shellfuncs.in#L654-L673 I wonder how much of an impact those bits have on the cluster(?) Take '99-VirtualDomain-libvirt.conf' - that one poses questions, with c9s 'libvirtd.service' is not really used or should not be, new modular approach is devised there. So, with 'resources-agents' having: After=libvirtd.service and users not being able to manage those bit - is that not asking for trouble? it does no harm (missing units are simply ignored) but it certainly does not do anything useful either. OTOH modular approach is also optional, so you could still use monolithic libvirtd on cluster nodes. So it is more a documentation issue. Not sure what you mean by 'missing unit' - unit is there only is not used, is disabled. What does 'resource-agents-deps' do with that? I don't suppose upstream, redhat & others made that effort, those changes with the suggestions to us consumers - do go back to "old" stuff. I'd suggest, if devel/contributors read here - and I'd imagine other users would reckon as well - to enhance RAs, certainly VirtualDomain, with a parameter/attribute with which users could, at least to certain extent, control those "outside" of cluster, dependencies. thanks, L. ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/