Re: [ClusterLabs] pcsd web interface not working on EL 9.3

2024-02-19 Thread lejeczek via Users




On 19/02/2024 09:06, Strahil Nikolov via Users wrote:

Hi All,

Is there a specific setup I missed in order to setup the 
web interface ?


Usually, you just login with the hacluster user on 
https://fqdn:2224 but when I do a curl, I get an empty 
response.


Best Regards,
Strahil Nikolov

___
Was it giving out some stuff before? I've never did curl on 
it. Won't guess why do that.
Yes it is _empty_for me too - though I use it with reverse 
proxy.

While _main_ URL curl returns no content this does get:
-> $ curl https://pcs-ha.mine.priv/ui/

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] clone_op_key pcmk__notify_key - Triggered fatal assertion

2024-02-17 Thread lejeczek via Users

Hi guys.

Everything seems to be working a ok yet pacemakers logs
...
 error: clone_op_key: Triggered fatal assertion at 
pcmk_graph_producer.c:207 : (n_type != NULL) && (n_task != NULL)
 error: pcmk__notify_key: Triggered fatal assertion at 
actions.c:187 : op_type != NULL
 error: clone_op_key: Triggered fatal assertion at 
pcmk_graph_producer.c:207 : (n_type != NULL) && (n_task != NULL)
 error: pcmk__notify_key: Triggered fatal assertion at 
actions.c:187 : op_type != NULL

...
 error: pcmk__create_history_xml: Triggered fatal assertion 
at pcmk_sched_actions.c:1163 : n_type != NULL
 error: pcmk__create_history_xml: Triggered fatal assertion 
at pcmk_sched_actions.c:1164 : n_task != NULL
 error: pcmk__notify_key: Triggered fatal assertion at 
actions.c:187 : op_type != NULL

...

Looks critical, is it it - would you know?
many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


VGA - disable display functions ?

2024-02-15 Thread lejeczek via users

Hi guys.

Would you know how to disable/tweak bits in kernel/module so 
display(& audio) is completely disabled?
I have AMD RX 6600 which I hope I could give completely to 
ROCm for computing.


many thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] trigger something at ?

2024-02-08 Thread lejeczek via Users



On 31/01/2024 16:37, lejeczek via Users wrote:



On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:02:12 +0100
lejeczek via Users  wrote:



On 29/01/2024 17:22, Ken Gaillot wrote:
On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users 
wrote:

Hi guys.

Is it possible to trigger some... action - I'm 
thinking specifically

at shutdown/start.
If not within the cluster then - if you do that - 
perhaps outside.
I would like to create/remove constraints, when 
cluster starts &

stops, respectively.

many thanks, L.

You could use node status alerts for that, but it's 
risky for alert
agents to change the configuration (since that may 
result in more

alerts and potentially some sort of infinite loop).

Pacemaker has no concept of a full cluster start/stop, 
only node
start/stop. You could approximate that by checking 
whether the node

receiving the alert is the only active node.

Another possibility would be to write a resource agent 
that does what
you want and order everything else after it. However 
it's even more

risky for a resource agent to modify the configuration.

Finally you could write a systemd unit to do what you 
want and order it

after pacemaker.

What's wrong with leaving the constraints permanently 
configured?

yes, that would be for a node start/stop
I struggle with using constraints to move pgsql (PAF) 
master

onto a given node - seems that co/locating paf's master
results in troubles (replication brakes) at/after node
shutdown/reboot (not always, but way too often)
What? What's wrong with colocating PAF's masters exactly? 
How does it brake any

replication? What's these constraints you are dealing with?

Could you share your configuration?
Constraints beyond/above of what is required by PAF agent 
itself, say...
you have multiple pgSQL cluster with PAF - thus multiple 
(separate, for each pgSQL cluster) masters and you want to 
spread/balance those across HA cluster
(or in other words - avoid having more that 1 pgsql master 
per HA node)
These below, I've tried, those move the master onto chosen 
node but.. then the issues I mentioned.


-> $ pcs constraint location PGSQL-PAF-5438-clone prefers 
ubusrv1=1002

or
-> $ pcs constraint colocation set PGSQL-PAF-5435-clone 
PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master 
require-all=false setoptions score=-1000


Wanted to share an observation - not a measurement of 
anything, I did not take those - of different, latest pgSQL 
version which I put in place of version 14 which I've been 
using all this time.
(also with that upgrade -  from Postgres own repos - came 
update of PAF)
So, with pgSQL ver. 16  and the same of everything else - 
now paf/pgSQL resources behave a lot lot better, survives 
just fine all those cases - with ! extra constraints of 
course - where previously it had replication failures.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] colocation constraint - do I get it all wrong?

2024-02-05 Thread lejeczek via Users




On 01/01/2024 18:28, Ken Gaillot wrote:

On Fri, 2023-12-22 at 17:02 +0100, lejeczek via Users wrote:

hi guys.

I have a colocation constraint:

-> $ pcs constraint ref DHCPD
Resource: DHCPD
   colocation-DHCPD-GATEWAY-NM-link-INFINITY

and the trouble is... I thought DHCPD is to follow GATEWAY-NM-link,
always!
If that is true that I see very strange behavior, namely.
When there is an issue with DHCPD resource, cannot be started, then
GATEWAY-NM-link gets tossed around by the cluster.

Is that normal & expected - is my understanding of _colocation_
completely wrong - or my cluster is indeed "broken"?
many thanks, L.


Pacemaker considers the preferences of colocated resources when
assigning a resource to a node, to ensure that as many resources as
possible can run. So if a colocated resource becomes unable to run on a
node, the primary resource might move to allow the colocated resource
to run.
So what is the way to "fix" this - is it simply low/er score 
for such constraint?
In my case _dhcpd_ is important but if fails sometimes as 
it's often tampered with, so... make _dhcpd_ flow 
gateway_link but just fail _dhcp_ (it it keeps failing) and 
leave _gateway_link_ alone if/where it's good.
Or perhaps there a global config/param for whole cluster 
behaviour?


___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] to heal gfid which is not there -- IGNORE wrong list

2024-02-04 Thread lejeczek via Users



On 04/02/2024 12:57, lejeczek via Users wrote:

hi guys.

So, I've manged to make my volume go haywire, here:

-> $ gluster volume heal VMAIL info
Brick 10.1.1.100:/devs/00.GLUSTERs/VMAIL
Status: Connected
Number of entries: 0

Brick 10.1.1.101:/devs/00.GLUSTERs/VMAIL
/dovecot-uidlist
Status: Connected
Number of entries: 1

Brick 10.1.1.99:/devs/00.GLUSTERs/VMAIL-arbiter
Status: Connected
Number of entries: 0

That _gfid_ nor _dovecot-uidlist_ do not exist on any of 
the bricks, but I've noticed that all bricks have:

.glusterfs-anonymous-inode-462a1850-a31a-4a17-934d-26f3996dc9b8
which is an empty dir.

That volume I fear won't heal on its own and might need 
some intervention - would you know how/where to start?


many thanks, L.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home:https://www.clusterlabs.org/
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] to heal gfid which is not there

2024-02-04 Thread lejeczek via Users

hi guys.

So, I've manged to make my volume go haywire, here:

-> $ gluster volume heal VMAIL info
Brick 10.1.1.100:/devs/00.GLUSTERs/VMAIL
Status: Connected
Number of entries: 0

Brick 10.1.1.101:/devs/00.GLUSTERs/VMAIL
/dovecot-uidlist
Status: Connected
Number of entries: 1

Brick 10.1.1.99:/devs/00.GLUSTERs/VMAIL-arbiter
Status: Connected
Number of entries: 0

That _gfid_ nor _dovecot-uidlist_ do not exist on any of the 
bricks, but I've noticed that all bricks have:

.glusterfs-anonymous-inode-462a1850-a31a-4a17-934d-26f3996dc9b8
which is an empty dir.

That volume I fear won't heal on its own and might need some 
intervention - would you know how/where to start?


many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] trigger something at ?

2024-02-01 Thread lejeczek via Users




On 01/02/2024 15:02, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 18:23:40 +0100
lejeczek via Users  wrote:


On 31/01/2024 17:13, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:37:21 +0100
lejeczek via Users  wrote:
  

On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:02:12 +0100
lejeczek via Users  wrote:
  

On 29/01/2024 17:22, Ken Gaillot wrote:

On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote:

Hi guys.

Is it possible to trigger some... action - I'm thinking specifically
at shutdown/start.
If not within the cluster then - if you do that - perhaps outside.
I would like to create/remove constraints, when cluster starts &
stops, respectively.

many thanks, L.
  

You could use node status alerts for that, but it's risky for alert
agents to change the configuration (since that may result in more
alerts and potentially some sort of infinite loop).

Pacemaker has no concept of a full cluster start/stop, only node
start/stop. You could approximate that by checking whether the node
receiving the alert is the only active node.

Another possibility would be to write a resource agent that does what
you want and order everything else after it. However it's even more
risky for a resource agent to modify the configuration.

Finally you could write a systemd unit to do what you want and order it
after pacemaker.

What's wrong with leaving the constraints permanently configured?

yes, that would be for a node start/stop
I struggle with using constraints to move pgsql (PAF) master
onto a given node - seems that co/locating paf's master
results in troubles (replication brakes) at/after node
shutdown/reboot (not always, but way too often)

What? What's wrong with colocating PAF's masters exactly? How does it
brake any replication? What's these constraints you are dealing with?

Could you share your configuration?

Constraints beyond/above of what is required by PAF agent
itself, say...
you have multiple pgSQL cluster with PAF - thus multiple
(separate, for each pgSQL cluster) masters and you want to
spread/balance those across HA cluster
(or in other words - avoid having more that 1 pgsql master
per HA node)

ok
  

These below, I've tried, those move the master onto chosen
node but.. then the issues I mentioned.

You just mentioned it breaks the replication, but there so little
information about your architecture and configuration, it's impossible to
imagine how this could break the replication.

Could you add details about the issues ?
  

-> $ pcs constraint location PGSQL-PAF-5438-clone prefers
ubusrv1=1002
or
-> $ pcs constraint colocation set PGSQL-PAF-5435-clone
PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master
require-all=false setoptions score=-1000

I suppose "collocation" constraint is the way to go, not the "location"
one.

This should be easy to replicate, 3 x VMs, Ubuntu 22.04 in
my case

No, this is not easy to replicate. I have no idea how you setup your PostgreSQL
replication, neither I have your full pacemaker configuration.

Please provide either detailed setupS and/or ansible and/or terraform and/or
vagrant, then a detailed scenario showing how it breaks. This is how you can
help and motivate devs to reproduce your issue and work on it.

I will not try to poke around for hours until I find an issue that might not
even be the same than yours.
How about you start with the basics - strange inclination to 
complicate things when they are not, I hear from you - 
that's what I did while "stumbled" upon these "issues"

How about just:
a) do vanilla-default pgSQL in Ubuntu (or perhaps any other 
OS of your choice), I use _pg_createcluster_
b) follow PAF official guide (a single PAF resource should 
suffice)
Have a healthy pgSQL cluster, OS _reboot_ nodes - play with 
that, all should be ok, moving around/electing master should 
work a ok.
Then... add, play with "additional" co/location constraints, 
then OS reboots,- things should begin braking.
I have 3-node HA cluster & 3-node PAF resource = 1 master + 
2 slaves.
Only thing I deliberately set, to alleviate pgsql 
replication was _wal_keep_size_ - I increased that, but this 
is subjective.


It's fine with me if you don't feel like doing this.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] trigger something at ?

2024-02-01 Thread lejeczek via Users



On 31/01/2024 18:11, Ken Gaillot wrote:

On Wed, 2024-01-31 at 16:37 +0100, lejeczek via Users wrote:

On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:02:12 +0100
lejeczek via Users  wrote:


On 29/01/2024 17:22, Ken Gaillot wrote:

On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote:

Hi guys.

Is it possible to trigger some... action - I'm thinking
specifically
at shutdown/start.
If not within the cluster then - if you do that - perhaps
outside.
I would like to create/remove constraints, when cluster
starts &
stops, respectively.

many thanks, L.


You could use node status alerts for that, but it's risky for
alert
agents to change the configuration (since that may result in
more
alerts and potentially some sort of infinite loop).

Pacemaker has no concept of a full cluster start/stop, only
node
start/stop. You could approximate that by checking whether the
node
receiving the alert is the only active node.

Another possibility would be to write a resource agent that
does what
you want and order everything else after it. However it's even
more
risky for a resource agent to modify the configuration.

Finally you could write a systemd unit to do what you want and
order it
after pacemaker.

What's wrong with leaving the constraints permanently
configured?

yes, that would be for a node start/stop
I struggle with using constraints to move pgsql (PAF) master
onto a given node - seems that co/locating paf's master
results in troubles (replication brakes) at/after node
shutdown/reboot (not always, but way too often)

What? What's wrong with colocating PAF's masters exactly? How does
it brake any
replication? What's these constraints you are dealing with?

Could you share your configuration?

Constraints beyond/above of what is required by PAF agent
itself, say...
you have multiple pgSQL cluster with PAF - thus multiple
(separate, for each pgSQL cluster) masters and you want to
spread/balance those across HA cluster
(or in other words - avoid having more that 1 pgsql master
per HA node)
These below, I've tried, those move the master onto chosen
node but.. then the issues I mentioned.

-> $ pcs constraint location PGSQL-PAF-5438-clone prefers
ubusrv1=1002
or
-> $ pcs constraint colocation set PGSQL-PAF-5435-clone
PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master
require-all=false setoptions score=-1000


Anti-colocation sets tend to be tricky currently -- if the first
resource can't be assigned to a node, none of them can. We have an idea
for a better implementation:

  https://projects.clusterlabs.org/T383

In the meantime, a possible workaround is to use placement-
strategy=balanced and define utilization for the clones only. The
promoted roles will each get a slight additional utilization, and the
cluster should spread them out across nodes whenever possible. I don't
know if that will avoid the replication issues but it may be worth a
try.

using _balanced_ causes a small mayhem to PAF/pgsql:

-> $ pcs property
Cluster Properties:
 REDIS-6380_REPL_INFO: ubusrv3
 REDIS-6381_REPL_INFO: ubusrv2
 REDIS-6382_REPL_INFO: ubusrv2
 REDIS-6385_REPL_INFO: ubusrv1
 REDIS_REPL_INFO: ubusrv1
 cluster-infrastructure: corosync
 cluster-name: ubusrv
 dc-version: 2.1.2-ada5c3b36e2
 have-watchdog: false
 last-lrm-refresh: 1706711588
 placement-strategy: default
 stonith-enabled: false

-> $ pcs resource utilization PGSQL-PAF-5438 cpu="20"

-> $ pcs property set placement-strategy=balanced # when 
resource stops:

I change it back:
-> $ pcs property set placement-strategy=default
and pgSQL/paf works again

I've not used _utilization_ nor _placement-strategy_ before, 
thus chance that I'm missing something is solid.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] trigger something at ?

2024-01-31 Thread lejeczek via Users



On 31/01/2024 17:13, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:37:21 +0100
lejeczek via Users  wrote:



On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:02:12 +0100
lejeczek via Users  wrote:


On 29/01/2024 17:22, Ken Gaillot wrote:

On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote:

Hi guys.

Is it possible to trigger some... action - I'm thinking specifically
at shutdown/start.
If not within the cluster then - if you do that - perhaps outside.
I would like to create/remove constraints, when cluster starts &
stops, respectively.

many thanks, L.


You could use node status alerts for that, but it's risky for alert
agents to change the configuration (since that may result in more
alerts and potentially some sort of infinite loop).

Pacemaker has no concept of a full cluster start/stop, only node
start/stop. You could approximate that by checking whether the node
receiving the alert is the only active node.

Another possibility would be to write a resource agent that does what
you want and order everything else after it. However it's even more
risky for a resource agent to modify the configuration.

Finally you could write a systemd unit to do what you want and order it
after pacemaker.

What's wrong with leaving the constraints permanently configured?

yes, that would be for a node start/stop
I struggle with using constraints to move pgsql (PAF) master
onto a given node - seems that co/locating paf's master
results in troubles (replication brakes) at/after node
shutdown/reboot (not always, but way too often)

What? What's wrong with colocating PAF's masters exactly? How does it brake
any replication? What's these constraints you are dealing with?

Could you share your configuration?

Constraints beyond/above of what is required by PAF agent
itself, say...
you have multiple pgSQL cluster with PAF - thus multiple
(separate, for each pgSQL cluster) masters and you want to
spread/balance those across HA cluster
(or in other words - avoid having more that 1 pgsql master
per HA node)

ok


These below, I've tried, those move the master onto chosen
node but.. then the issues I mentioned.

You just mentioned it breaks the replication, but there so little information
about your architecture and configuration, it's impossible to imagine how this
could break the replication.

Could you add details about the issues ?


-> $ pcs constraint location PGSQL-PAF-5438-clone prefers
ubusrv1=1002
or
-> $ pcs constraint colocation set PGSQL-PAF-5435-clone
PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master
require-all=false setoptions score=-1000

I suppose "collocation" constraint is the way to go, not the "location" one.
This should be easy to replicate, 3 x VMs, Ubuntu 22.04 in 
my case


-> $ pcs resource config PGSQL-PAF-5438-clone
 Clone: PGSQL-PAF-5438-clone
  Meta Attrs: failure-timeout=60s master-max=1 notify=true 
promotable=true
  Resource: PGSQL-PAF-5438 (class=ocf provider=heartbeat 
type=pgsqlms)
   Attributes: bindir=/usr/lib/postgresql/16/bin 
datadir=/var/lib/postgresql/16/paf-5438 maxlag=1000 
pgdata=/etc/postgresql/16/paf-5438 pgport=5438
   Operations: demote interval=0s timeout=120s 
(PGSQL-PAF-5438-demote-interval-0s)
   methods interval=0s timeout=5 
(PGSQL-PAF-5438-methods-interval-0s)
   monitor interval=15s role=Master timeout=10s 
(PGSQL-PAF-5438-monitor-interval-15s)
   monitor interval=16s role=Slave timeout=10s 
(PGSQL-PAF-5438-monitor-interval-16s)
   notify interval=0s timeout=60s 
(PGSQL-PAF-5438-notify-interval-0s)
   promote interval=0s timeout=30s 
(PGSQL-PAF-5438-promote-interval-0s)
   reload interval=0s timeout=20 
(PGSQL-PAF-5438-reload-interval-0s)
   start interval=0s timeout=60s 
(PGSQL-PAF-5438-start-interval-0s)
   stop interval=0s timeout=60s 
(PGSQL-PAF-5438-stop-interval-0s)


so, regarding PAF - 1 master + 2 slaves, have a healthy 
pqSQL/PAF cluster to begin with, then
make resource prefer a specific node (with simplest variant 
of constraints I tried):
-> $ pcs constraint location PGSQL-PAF-5438-clone prefers 
ubusrv1=1002


and play with it, rebooting node(s) with OS' _reboot_
I at some point, get HA/resource unable to start pgSQL, 
unable to elect a master (logs saying with replication 
broken) and I have to "fix" pgSQL cluster outside of PAF, 
using _pg_basebackup_



___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] trigger something at ?

2024-01-31 Thread lejeczek via Users




On 31/01/2024 16:06, Jehan-Guillaume de Rorthais wrote:

On Wed, 31 Jan 2024 16:02:12 +0100
lejeczek via Users  wrote:



On 29/01/2024 17:22, Ken Gaillot wrote:

On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote:

Hi guys.

Is it possible to trigger some... action - I'm thinking specifically
at shutdown/start.
If not within the cluster then - if you do that - perhaps outside.
I would like to create/remove constraints, when cluster starts &
stops, respectively.

many thanks, L.


You could use node status alerts for that, but it's risky for alert
agents to change the configuration (since that may result in more
alerts and potentially some sort of infinite loop).

Pacemaker has no concept of a full cluster start/stop, only node
start/stop. You could approximate that by checking whether the node
receiving the alert is the only active node.

Another possibility would be to write a resource agent that does what
you want and order everything else after it. However it's even more
risky for a resource agent to modify the configuration.

Finally you could write a systemd unit to do what you want and order it
after pacemaker.

What's wrong with leaving the constraints permanently configured?

yes, that would be for a node start/stop
I struggle with using constraints to move pgsql (PAF) master
onto a given node - seems that co/locating paf's master
results in troubles (replication brakes) at/after node
shutdown/reboot (not always, but way too often)

What? What's wrong with colocating PAF's masters exactly? How does it brake any
replication? What's these constraints you are dealing with?

Could you share your configuration?
Constraints beyond/above of what is required by PAF agent 
itself, say...
you have multiple pgSQL cluster with PAF - thus multiple 
(separate, for each pgSQL cluster) masters and you want to 
spread/balance those across HA cluster
(or in other words - avoid having more that 1 pgsql master 
per HA node)
These below, I've tried, those move the master onto chosen 
node but.. then the issues I mentioned.


-> $ pcs constraint location PGSQL-PAF-5438-clone prefers 
ubusrv1=1002

or
-> $ pcs constraint colocation set PGSQL-PAF-5435-clone 
PGSQL-PAF-5434-clone PGSQL-PAF-5433-clone role=Master 
require-all=false setoptions score=-1000




___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] trigger something at ?

2024-01-31 Thread lejeczek via Users




On 29/01/2024 17:22, Ken Gaillot wrote:

On Fri, 2024-01-26 at 13:55 +0100, lejeczek via Users wrote:

Hi guys.

Is it possible to trigger some... action - I'm thinking specifically
at shutdown/start.
If not within the cluster then - if you do that - perhaps outside.
I would like to create/remove constraints, when cluster starts &
stops, respectively.

many thanks, L.


You could use node status alerts for that, but it's risky for alert
agents to change the configuration (since that may result in more
alerts and potentially some sort of infinite loop).

Pacemaker has no concept of a full cluster start/stop, only node
start/stop. You could approximate that by checking whether the node
receiving the alert is the only active node.

Another possibility would be to write a resource agent that does what
you want and order everything else after it. However it's even more
risky for a resource agent to modify the configuration.

Finally you could write a systemd unit to do what you want and order it
after pacemaker.

What's wrong with leaving the constraints permanently configured?

yes, that would be for a node start/stop
I struggle with using constraints to move pgsql (PAF) master 
onto a given node - seems that co/locating paf's master 
results in troubles (replication brakes) at/after node 
shutdown/reboot (not always, but way too often)


Ideally I'm hoping that:
at node stop, stopping node could check if it's PAF's master 
and if yes so then remove given constraints

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] trigger something at ?

2024-01-26 Thread lejeczek via Users

Hi guys.

Is it possible to trigger some... action - I'm thinking 
specifically at shutdown/start.
If not within the cluster then - if you do that - perhaps 
outside.
I would like to create/remove constraints, when cluster 
starts & stops, respectively.


many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] replica vol when a peer is lost/offed & qcow2 ?

2024-01-17 Thread lejeczek via Users

Hi guys.

I wonder if you might have any tips/tweaks for 
volume/cluster to make it more resilient? accommodating? to 
qcow2 files but! when a peer is lots or missing?
I have 3-peer cluster/volume: 2 + 1 arbiter & my experience 
is such, that when all is good then.. well, all is good, but...
when one peers was lots - even if it's the arbiter - then 
VMs begin to suffer from & to report, their filesystems, errors.
Perhaps it's not a volume's properties/config but whole 
cluster? Or perhaps 3-peer vol is not good for things such a 
qcow2s?


many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Lenovo ThinkBook 16 G6 touchpad - no right click ?

2024-01-14 Thread lejeczek via users

Hi guys.

Anybody here on a Thinkbook 16 G6?
I just got one myself and it seems that the right-click is 
not there - that should not be new laptop with 
mechanical-electrical damage, right?


many thanks, L.
--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


NVMe sector size ?

2024-01-10 Thread lejeczek via users

Hi guys.

I wonder if anybody played with both sector sizes 512 & 4086 
and if yes so - also had some test results / thoughts to share?


There are bits about it - bit confusing & a mixed-bag -  
over the net but I failed to find anything Fedora-specific.


thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


color profiles for screens/monitors

2024-01-05 Thread lejeczek via users

Hi guys

How do we load & tweak color profiles in Fedora (without 
using external sensor-devices) ?
My Lenovo Thinkbook built-in screen has different (and not 
adjustable in BIOS) color temperature  what my external DELL 
shows - they are quite different thus annoying.


many thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ClusterLabs] colocation constraint - do I get it all wrong?

2023-12-22 Thread lejeczek via Users

hi guys.

I have a colocation constraint:

-> $ pcs constraint ref DHCPD
Resource: DHCPD
  colocation-DHCPD-GATEWAY-NM-link-INFINITY

and the trouble is... I thought DHCPD is to follow 
GATEWAY-NM-link, always!

If that is true that I see very strange behavior, namely.
When there is an issue with DHCPD resource, cannot be 
started, then GATEWAY-NM-link gets tossed around by the cluster.


Is that normal & expected - is my understanding of 
_colocation_ completely wrong - or my cluster is indeed 
"broken"?

many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] colocate Redis - weird

2023-12-20 Thread lejeczek via Users



On 19/12/2023 19:13, lejeczek via Users wrote:

hi guys,

Is this below not the weirdest thing?

-> $ pcs constraint ref PGSQL-PAF-5435
Resource: PGSQL-PAF-5435
  colocation-HA-10-1-1-84-PGSQL-PAF-5435-clone-INFINITY
  colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY
  order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory
  order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory-1
  colocation_set_PePePe

Here Redis master should folow pgSQL master.
Which such constraint:

-> $ pcs resource status PGSQL-PAF-5435
  * Clone Set: PGSQL-PAF-5435-clone [PGSQL-PAF-5435] 
(promotable):

    * Promoted: [ ubusrv1 ]
    * Unpromoted: [ ubusrv2 ubusrv3 ]
-> $ pcs resource status REDIS-6385-clone
  * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable):
    * Unpromoted: [ ubusrv1 ubusrv2 ubusrv3 ]

If I remove that constrain:
-> $ pcs constraint delete 
colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY

-> $ pcs resource status REDIS-6385-clone
  * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable):
    * Promoted: [ ubusrv1 ]
    * Unpromoted: [ ubusrv2 ubusrv3 ]

and ! I can manually move Redis master around, master 
moves to each server just fine.

I again, add that constraint:

-> $ pcs constraint colocation add master REDIS-6385-clone 
with master PGSQL-PAF-5435-clone


and the same...


What there might be about that one node - resource removed, 
created anew and cluster insists on keeping master there.
I can manually move the master anywhere but if I _clear_ the 
resource, no constraints then cluster move it back to the 
same node.


I wonder about:  a) "transient" node attrs & b) if this 
cluster is somewhat broken.
On a) - can we read more about those somewhere?(not the 
code/internals)

thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] colocate Redis - weird

2023-12-19 Thread lejeczek via Users

hi guys,

Is this below not the weirdest thing?

-> $ pcs constraint ref PGSQL-PAF-5435
Resource: PGSQL-PAF-5435
  colocation-HA-10-1-1-84-PGSQL-PAF-5435-clone-INFINITY
  colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY
  order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory
  order-PGSQL-PAF-5435-clone-HA-10-1-1-84-Mandatory-1
  colocation_set_PePePe

Here Redis master should folow pgSQL master.
Which such constraint:

-> $ pcs resource status PGSQL-PAF-5435
  * Clone Set: PGSQL-PAF-5435-clone [PGSQL-PAF-5435] 
(promotable):

    * Promoted: [ ubusrv1 ]
    * Unpromoted: [ ubusrv2 ubusrv3 ]
-> $ pcs resource status REDIS-6385-clone
  * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable):
    * Unpromoted: [ ubusrv1 ubusrv2 ubusrv3 ]

If I remove that constrain:
-> $ pcs constraint delete 
colocation-REDIS-6385-clone-PGSQL-PAF-5435-clone-INFINITY

-> $ pcs resource status REDIS-6385-clone
  * Clone Set: REDIS-6385-clone [REDIS-6385] (promotable):
    * Promoted: [ ubusrv1 ]
    * Unpromoted: [ ubusrv2 ubusrv3 ]

and ! I can manually move Redis master around, master moves 
to each server just fine.

I again, add that constraint:

-> $ pcs constraint colocation add master REDIS-6385-clone 
with master PGSQL-PAF-5435-clone


and the same...

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: Gnome's terminal/console fonts ?

2023-12-19 Thread lejeczek via users



On 16/12/2023 21:21, George N. White III wrote:
On Sat, Dec 16, 2023 at 3:36 AM lejeczek via users 
 wrote:


Hi guys.

My Gnome's default terminal does only select from
_mono_ type of fonts & I remember it could choose any
font.
Alternative solutions, such a _terminator_ do allow to
select any font.

In Gnome's terminal 'font scaling' also seems "broken"
- certainly on my Fedora - I select "Liberation Mono"
and I start size at 6 and keep increasing by small
fraction and the font does not seem changing until
suddenly, at some given size it "jumps" the size & ...
shape too.

I do not suppose it's my environment - I've been
upgrading my Fedora since... forever, as opposed to
"clean" install - and I think it's rather "broken"
terminal, but..
I wanted to consult other Fedorians to be sure.


Using a fresh install of F39, in Gnome Terminal 
"Preferences" I selected the "Unnamed" Profile, set 
"Custom font" to Liberation Mono, and click on the  
"Custom Font" bar to get a "Choose A Terminal Font" 
window. The size changes in the window and the terminal 
also changes to match when I close the "Chose ..." 
window.  I have TrueType Liberation fonts:


% fc-list | grep LiberationMono
/usr/share/fonts/liberation-mono/LiberationMono-Regular.ttf: 
Liberation Mono:style=Regular
/usr/share/fonts/liberation-mono/LiberationMono-Bold.ttf: 
Liberation Mono:style=Bold
/usr/share/fonts/liberation-mono/LiberationMono-BoldItalic.ttf: 
Liberation Mono:style=Bold Italic
/usr/share/fonts/liberation-mono/LiberationMono-Italic.ttf: 
Liberation Mono:style=Italic


Liberation Mono as the example
I might not remembered exactly:
some range and no change < 6.366 > changes visibly substantially
range no change < 6.875 > again changes visibly quite a bit
Also does not seem that "Cell spacing" _width_ works "properly"

And my question about _only_ mono fonts being available in 
Ghome's terminal (while other terminals choose any font) ?


many thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Gnome's terminal/console fonts ?

2023-12-15 Thread lejeczek via users

Hi guys.

My Gnome's default terminal does only select from _mono_ 
type of fonts & I remember it could choose any font.
Alternative solutions, such a _terminator_ do allow to 
select any font.


In Gnome's terminal 'font scaling' also seems "broken" - 
certainly on my Fedora - I select "Liberation Mono" and I 
start size at 6 and keep increasing by small fraction and 
the font does not seem changing until suddenly, at some 
given size it "jumps" the size & ... shape too.


I do not suppose it's my environment - I've been upgrading 
my Fedora since... forever, as opposed to "clean" install - 
and I think it's rather "broken" terminal, but..

I wanted to consult other Fedorians to be sure.

many thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ClusterLabs] resource-agents and VMs

2023-12-15 Thread lejeczek via Users

Hi guys.

my resources-agents depend like so:

resource-agents-deps.target
○ ├─00\\x2dVMsy.mount
● └─virt-guest-shutdown.target

when I reboot a node VMs seems to migrated off it live a ok, 
but..
when node comes back on after a reboot, VMs fail to migrate 
back to it, live.

I see on such node

-> $ journalctl -lf -o cat -u virtqemud.service
Starting Virtualization qemu daemon...
Started Virtualization qemu daemon.
libvirt version: 9.5.0, package: 6.el9 (buil...@centos.org, 
2023-08-25-08:53:56, )

hostname: dzien.mine.priv
Path '/00-VMsy/enc.podnode3.qcow2' is not accessible: No 
such file or directory


and I wonder if it's indeed the fact the the _path_ is 
absent at the moment cluster just after node start, tries to 
migrate VM resource...
Is it possible to somehow - seemingly my 
_resource-agents-deps.target_ does not do - assure that 
cluster, perhaps on per-resource basis, will wait/check a 
path first?
BTW, that paths is available and is made sure is available 
to the system, it's a glusterfs mount.


many thanks, L

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] how to colocate promoted resources ?

2023-12-08 Thread lejeczek via Users




On 08/12/2023 13:25, Jehan-Guillaume de Rorthais wrote:

Hi,

On Wed, 6 Dec 2023 10:36:39 +0100
lejeczek via Users  wrote:


How do your colocate your promoted resources with balancing
underlying resources as priority?

What do you mean?


With a simple scenario, say
3 nodes and 3 pgSQL clusters
what would be best possible way - I'm thinking most gentle
at the same time, if that makes sense.

I'm not sure it answers your question (as I don't understand it), but here is a
doc explaining how to create and move two IP supposed to each start on
secondaries, avoiding the primary node if possible, as long as secondaries nodes
exists

https://clusterlabs.github.io/PAF/CentOS-7-admin-cookbook.html#adding-ips-on-standbys-nodes

++

Apologies, perhaps I was quite vague.
I was thinking - having a 3-node HA cluster and 3-node 
single-master->slaves pgSQL, now..
say, I want pgSQL masters to spread across HA cluster so I 
theory - having each HA node identical hardware-wise - 
masters' resources would nicely balance out across HA cluster.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] ethernet link up/down - ?

2023-12-07 Thread lejeczek via Users



On 04/12/2023 20:58, Reid Wahl wrote:

On Thu, Nov 30, 2023 at 10:30 AM lejeczek via Users
 wrote:



On 07/02/2022 20:09, lejeczek via Users wrote:

Hi guys

How do you guys go about doing link up/down as a resource?

many thanks, L.


With simple tests I confirmed that indeed Linux - on my
hardware at leat - can easily power down an eth link - if
a @devel reads this:
Is there an agent in the suite which a non-programmer could
easily (for most safely) adopt for such purpose?
I understand such agent has to be cloneable & promotable.

The iface-bridge resource appears to do something similar for bridges.
I don't see anything currently for links in general.



Where can I find that agent?

Any comment on that idea about adding, introducing such 
"link" agent into agents in the future?

Should I go _github_ and suggest it there perhaps?
Naturally done by "devel" it would be ideal, as opposed to, 
by us user/admins.

thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] how to colocate promoted resources ?

2023-12-06 Thread lejeczek via Users

Hi guys.

How do your colocate your promoted resources with balancing 
underlying resources as priority?

With a simple scenario, say
3 nodes and 3 pgSQL clusters
what would be best possible way - I'm thinking most gentle 
at the same time, if that makes sense.


many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] make promoted follow promoted resource ?

2023-12-06 Thread lejeczek via Users



On 26/11/2023 12:20, Reid Wahl wrote:

On Sun, Nov 26, 2023 at 1:32 AM lejeczek via Users
 wrote:

Hi guys.

With these:

-> $ pcs resource status REDIS-6381-clone
   * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable):
 * Promoted: [ ubusrv2 ]
 * Unpromoted: [ ubusrv1 ubusrv3 ]

-> $ pcs resource status PGSQL-PAF-5433-clone
   * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable):
 * Promoted: [ ubusrv1 ]
 * Unpromoted: [ ubusrv2 ubusrv3 ]

-> $ pcs constraint ref REDIS-6381-clone
Resource: REDIS-6381-clone
   colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY

basically promoted Redis should follow promoted pgSQL but it's not happening, 
usually it does.
I presume pcs/cluster does something internally which results in 
disobeying/ignoring that _colocation_ constraint for these resources.
I presume scoring might play a role:
   REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) 
(with-rsc-role:Master)
but usually, that scoring works, only "now" it does not.
Any comments I appreciate much.
thanks, L.

I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but 
cannot see explanation for this.
...
  notice: Calculated transition 110, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3729.bz2
  notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, 
Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete
  notice: State transition S_TRANSITION_ENGINE -> S_IDLE
  notice: State transition S_IDLE -> S_POLICY_ENGINE
  notice: Actions: Start  REDIS-6381:0 (
ubusrv2 )
  notice: Actions: Start  REDIS-6381:1 (
ubusrv3 )
  notice: Actions: Start  REDIS-6381:2 (
ubusrv1 )
  notice: Calculated transition 111, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3730.bz2
  notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2
  notice: Requesting local execution of start operation for REDIS-6381 on 
ubusrv2
(to redis) root on none
pam_unix(su:session): session opened for user redis(uid=127) by (uid=0)
pam_sss(su:session): Request to sssd failed. Connection refused
pam_unix(su:session): session closed for user redis
pam_sss(su:session): Request to sssd failed. Connection refused
  notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000
  notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create 
master-REDIS-6381=1000: Transient attribute change
INFO: demote: Setting master to 'no-such-master'
  notice: Result of start operation for REDIS-6381 on ubusrv2: ok
  notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, 
Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped
  notice: Actions: PromoteREDIS-6381:0 ( Unpromoted -> Promoted 
ubusrv2 )
  notice: Actions: Start  REDIS-6381:1 (
ubusrv1 )
  notice: Actions: Start  REDIS-6381:2 (
ubusrv3 )
  notice: Calculated transition 112, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3731.bz2
  notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on 
ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Result of notify operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating start operation REDIS-6381_start_0 on ubusrv1
  notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3
  notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on 
ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1
  notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on 
ubusrv3
  notice: Result of notify operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally 
on ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1
  notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on 
ubusrv3
  notice: Result of notify operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2
  notice: Requesting local execution of promote operation for REDIS-6381 on 
ubusrv2
  notice: Result of promote operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally 
on ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Initiating notify operation REDIS-6381_post_notify_promote_0 on 
ubusrv1
  notice: Initiating notify operation REDIS-6381:2_post_notify_promote_0 on 
ubusrv3
  notice: Result of notify operation for RED

[ClusterLabs] IPaddr2 Started (disabled) ?

2023-12-04 Thread lejeczek via Users

hi guys.

A cluster thinks the resource is up:
...
  * HA-10-1-1-80    (ocf:heartbeat:IPaddr2):     Started 
ubusrv3 (disabled)

..
while it is not the case. What might it mean?
Config is simple:
-> $ pcs resource config HA-10-1-1-80
 Resource: HA-10-1-1-80 (class=ocf provider=heartbeat 
type=IPaddr2)

  Attributes: cidr_netmask=24 ip=10.1.1.80
  Meta Attrs: failure-timeout=20s target-role=Stopped
  Operations: monitor interval=10s timeout=20s 
(HA-10-1-1-80-monitor-interval-10s)
  start interval=0s timeout=20s 
(HA-10-1-1-80-start-interval-0s)
  stop interval=0s timeout=20s 
(HA-10-1-1-80-stop-interval-0s)


I expected, well.. cluster to behave.
This results in a trouble - because of that exactly I think 
- for when resource is re-enabled then that IP/resource is 
not instantiated.

Is its monitor not working?
Is there a way to "harden" cluster perhaps via resource config?
cluster is Ubuntu VM boxes.

many thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


a VM with secure boot does not survive a reset

2023-12-03 Thread lejeczek via Users

Hi guys.

A VM created with:

-> $ virt-install --boot uefi --name=client-efi-secureboot 
--os-variant centos-stream9 --memory $((4*1024)) 
--graphics=vnc,listen=0.0.0.0 --accelerate --cpuset 2 --cpu 
host-model --pxe --disk=/00-VMsy/client-efi-secureboot.qcow2 
 --network network=10_3_1,,model=virtio


when *virsh reset* while running will power off, instead of 
*restart*. Whereas:


-> $ virt-install --boot 
uefi,firmware.feature0.name=secure-boot,firmware.feature0.enabled=no 
--name=client-efi-no-secureboot --os-variant centos-stream9 
--memory $((4*1024)) --graphics=vnc,listen=0.0.0.0 
--accelerate --cpuset 2 --cpu host-model --pxe 
--disk=/00-VMsy/client-efi-no-secureboot.qcow2  --network 
network=10_3_1,,model=virtio


does behave as expected when *reset*.

libvirt-libs-9.5.0-6.el9.x86_64

BTW. Where should bug reports go to theses days - I'm trying 
Redhat's Jira but cannot file a BZ in _RHEL_ category. I get 
no errors nor warnings but Jira won't let me file anything 
there, but I can in Centos category.


thanks, L.
___
Users mailing list -- users@lists.libvirt.org
To unsubscribe send an email to users-le...@lists.libvirt.org


pxe efi boot - bdsdxe boot0002 access denied

2023-12-01 Thread lejeczek via Users

Hi guys

I have VMs which fail to boot with error as per the subject.
I suspect this is due to:

  
    hvm
    
   <--- HERE
   < HERE ?
    
    type='pflash'>/usr/share/edk2/ovmf/OVMF_CODE.secboot.fd
    template='/usr/share/edk2/ovmf/OVMF_VARS.secboot.fd'>/var/lib/libvirt/qemu/nvram/centos-efi_VARS.fd

    
    
  

Would you have any thoughts on how to fix, to work around it?
many thanks, L.___
Users mailing list -- users@lists.libvirt.org
To unsubscribe send an email to users-le...@lists.libvirt.org


Re: [ClusterLabs] ethernet link up/down - ?

2023-11-30 Thread lejeczek via Users




On 07/02/2022 20:09, lejeczek via Users wrote:

Hi guys

How do you guys go about doing link up/down as a resource?

many thanks, L.



With simple tests I confirmed that indeed Linux - on my 
hardware at leat - can easily power down an eth link - if 
a @devel reads this:
Is there an agent in the suite which a non-programmer could 
easily (for most safely) adopt for such purpose?

I understand such agent has to be cloneable & promotable.

btw. I think many would consider that a really neat & useful 
addition to resource-agents - if authors/devel made it into 
the core package.


many thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] ethernet link up/down - ?

2023-11-28 Thread lejeczek via Users



On 16/02/2022 10:37, Klaus Wenninger wrote:



On Tue, Feb 15, 2022 at 5:25 PM lejeczek via Users 
 wrote:




On 07/02/2022 19:21, Antony Stone wrote:
> On Monday 07 February 2022 at 20:09:02, lejeczek via
    Users wrote:
>
>> Hi guys
>>
>> How do you guys go about doing link up/down as a
resource?
> I apply or remove addresses on the interface, using
"IPaddr2" and "IPv6addr",
> which I know is not the same thing.
>
> Why do you separately want to control link up/down? 
I can't think what I
> would use this for.

Just out of curiosity and as I haven't seen an answer in 
the thread yet - maybe

I overlooked something ...
Is this to control some link-triggered redundancy setup 
with switches?



Revisiting my own question/thread.
Yes. Very close to what Klaus wondered - it's a device over 
which I have no control and from that device perspective 
it's simply - link is up then I'll "serve" it.
I've been thinking lowest possible layer shall be the safest 
way - thus asked about controlling eth link that way: 
down/up by means of electric power, ideally.
As opposed to ha-cluster calling some middle men such as 
network managers.

I read some eth nics/drivers can power down a port.
Is there an agent & a way to do that?



>
>
> Antony.
>
Kind of similar - tcp/ip and those layers configs are
delivered by DHCP.
I'd think it would have to be a clone resource with one
master without any constraints where cluster freely
decides
where to put master(link up) on - which is when link gets
dhcp-served.
But I wonder if that would mean writing up a new
resource -
I don't think there is anything like that included in
ready-made pcs/ocf packages.

many thanks, L
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] make promoted follow promoted resource ?

2023-11-26 Thread lejeczek via Users



On 26/11/2023 17:44, Andrei Borzenkov wrote:

On 26.11.2023 12:32, lejeczek via Users wrote:

Hi guys.

With these:

-> $ pcs resource status REDIS-6381-clone
    * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable):
      * Promoted: [ ubusrv2 ]
      * Unpromoted: [ ubusrv1 ubusrv3 ]

-> $ pcs resource status PGSQL-PAF-5433-clone
    * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433]
(promotable):
      * Promoted: [ ubusrv1 ]
      * Unpromoted: [ ubusrv2 ubusrv3 ]

-> $ pcs constraint ref REDIS-6381-clone
Resource: REDIS-6381-clone
    
colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY


basically promoted Redis should follow promoted pgSQL but
it's not happening, usually it does.
I presume pcs/cluster does something internally which
results in disobeying/ignoring that _colocation_ constraint
for these resources.
I presume scoring might play a role:
    REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001)
(rsc-role:Master) (with-rsc-role:Master)
but usually, that scoring works, only "now" it does not.
Any comments I appreciate much.
thanks, L.

I looked at pamaker log - snippet below after
REDIS-6381-clone re-enabled - but cannot see explanation for
this.
...
   notice: Calculated transition 110, saving inputs in
/var/lib/pacemaker/pengine/pe-input-3729.bz2
   notice: Transition 110 (Complete=0, Pending=0, Fired=0,
Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): 
Complete

   notice: State transition S_TRANSITION_ENGINE -> S_IDLE
   notice: State transition S_IDLE -> S_POLICY_ENGINE
   notice: Actions: Start  REDIS-6381:0
(    ubusrv2 )
   notice: Actions: Start  REDIS-6381:1
(    ubusrv3 )
   notice: Actions: Start  REDIS-6381:2
(    ubusrv1 )
   notice: Calculated transition 111, saving inputs in
/var/lib/pacemaker/pengine/pe-input-3730.bz2
   notice: Initiating start operation REDIS-6381_start_0
locally on ubusrv2
   notice: Requesting local execution of start operation for
REDIS-6381 on ubusrv2
(to redis) root on none
pam_unix(su:session): session opened for user redis(uid=127)
by (uid=0)
pam_sss(su:session): Request to sssd failed. Connection 
refused

pam_unix(su:session): session closed for user redis
pam_sss(su:session): Request to sssd failed. Connection 
refused
   notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 
1000


This is the only line that sets master score, so 
apparently ubusrv2 is the only node where your clone *can* 
be promoted. Whether pacemaker is expected to fail this 
operation because it violates constraint I do not know.



   notice: Transition 111 aborted by
status-2-master-REDIS-6381 doing create
master-REDIS-6381=1000: Transient attribute change
INFO: demote: Setting master to 'no-such-master'
   notice: Result of start operation for REDIS-6381 on
ubusrv2: ok
   notice: Transition 111 (Complete=4, Pending=0, Fired=0,
Skipped=1, Incomplete=14,
Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): 
Stopped

   notice: Actions: Promote    REDIS-6381:0 (
Unpromoted -> Promoted ubusrv2 )
   notice: Actions: Start  REDIS-6381:1
(    ubusrv1 )
   notice: Actions: Start  REDIS-6381:2
(    ubusrv3 )
   notice: Calculated transition 112, saving inputs in
/var/lib/pacemaker/pengine/pe-input-3731.bz2
   notice: Initiating notify operation
REDIS-6381_pre_notify_start_0 locally on ubusrv2
   notice: Requesting local execution of notify operation 
for

REDIS-6381 on ubusrv2
   notice: Result of notify operation for REDIS-6381 on
ubusrv2: ok
   notice: Initiating start operation REDIS-6381_start_0 on
ubusrv1
   notice: Initiating start operation 
REDIS-6381:2_start_0 on

ubusrv3
   notice: Initiating notify operation
REDIS-6381_post_notify_start_0 locally on ubusrv2
   notice: Requesting local execution of notify operation 
for

REDIS-6381 on ubusrv2
   notice: Initiating notify operation
REDIS-6381_post_notify_start_0 on ubusrv1
   notice: Initiating notify operation
REDIS-6381:2_post_notify_start_0 on ubusrv3
   notice: Result of notify operation for REDIS-6381 on
ubusrv2: ok
   notice: Initiating notify operation
REDIS-6381_pre_notify_promote_0 locally on ubusrv2
   notice: Requesting local execution of notify operation 
for

REDIS-6381 on ubusrv2
   notice: Initiating notify operation
REDIS-6381_pre_notify_promote_0 on ubusrv1
   notice: Initiating notify operation
REDIS-6381:2_pre_notify_promote_0 on ubusrv3
   notice: Result of notify operation for REDIS-6381 on
ubusrv2: ok
   notice: Initiating promote operation REDIS-6381_promote_0
locally on ubusrv2
   notice: Requesting local execution of promote operation
for REDIS-6381 on ubusrv2
   notice: Result of promote operation for REDIS-6381 on
ubusrv2: ok
   notice: Initiating notify operation
REDIS-6381_post_notify_promote_0 locally on ubusrv2
   notice: Requesting local execution of notify operation 
for

REDIS-6381 o

Re: [ClusterLabs] make promoted follow promoted resource ?

2023-11-26 Thread lejeczek via Users



On 26/11/2023 12:20, Reid Wahl wrote:

On Sun, Nov 26, 2023 at 1:32 AM lejeczek via Users
 wrote:

Hi guys.

With these:

-> $ pcs resource status REDIS-6381-clone
   * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable):
 * Promoted: [ ubusrv2 ]
 * Unpromoted: [ ubusrv1 ubusrv3 ]

-> $ pcs resource status PGSQL-PAF-5433-clone
   * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] (promotable):
 * Promoted: [ ubusrv1 ]
 * Unpromoted: [ ubusrv2 ubusrv3 ]

-> $ pcs constraint ref REDIS-6381-clone
Resource: REDIS-6381-clone
   colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY

basically promoted Redis should follow promoted pgSQL but it's not happening, 
usually it does.
I presume pcs/cluster does something internally which results in 
disobeying/ignoring that _colocation_ constraint for these resources.
I presume scoring might play a role:
   REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) (rsc-role:Master) 
(with-rsc-role:Master)
but usually, that scoring works, only "now" it does not.
Any comments I appreciate much.
thanks, L.

I looked at pamaker log - snippet below after REDIS-6381-clone re-enabled - but 
cannot see explanation for this.
...
  notice: Calculated transition 110, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3729.bz2
  notice: Transition 110 (Complete=0, Pending=0, Fired=0, Skipped=0, 
Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete
  notice: State transition S_TRANSITION_ENGINE -> S_IDLE
  notice: State transition S_IDLE -> S_POLICY_ENGINE
  notice: Actions: Start  REDIS-6381:0 (
ubusrv2 )
  notice: Actions: Start  REDIS-6381:1 (
ubusrv3 )
  notice: Actions: Start  REDIS-6381:2 (
ubusrv1 )
  notice: Calculated transition 111, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3730.bz2
  notice: Initiating start operation REDIS-6381_start_0 locally on ubusrv2
  notice: Requesting local execution of start operation for REDIS-6381 on 
ubusrv2
(to redis) root on none
pam_unix(su:session): session opened for user redis(uid=127) by (uid=0)
pam_sss(su:session): Request to sssd failed. Connection refused
pam_unix(su:session): session closed for user redis
pam_sss(su:session): Request to sssd failed. Connection refused
  notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000
  notice: Transition 111 aborted by status-2-master-REDIS-6381 doing create 
master-REDIS-6381=1000: Transient attribute change
INFO: demote: Setting master to 'no-such-master'
  notice: Result of start operation for REDIS-6381 on ubusrv2: ok
  notice: Transition 111 (Complete=4, Pending=0, Fired=0, Skipped=1, 
Incomplete=14, Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped
  notice: Actions: PromoteREDIS-6381:0 ( Unpromoted -> Promoted 
ubusrv2 )
  notice: Actions: Start  REDIS-6381:1 (
ubusrv1 )
  notice: Actions: Start  REDIS-6381:2 (
ubusrv3 )
  notice: Calculated transition 112, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3731.bz2
  notice: Initiating notify operation REDIS-6381_pre_notify_start_0 locally on 
ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Result of notify operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating start operation REDIS-6381_start_0 on ubusrv1
  notice: Initiating start operation REDIS-6381:2_start_0 on ubusrv3
  notice: Initiating notify operation REDIS-6381_post_notify_start_0 locally on 
ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Initiating notify operation REDIS-6381_post_notify_start_0 on ubusrv1
  notice: Initiating notify operation REDIS-6381:2_post_notify_start_0 on 
ubusrv3
  notice: Result of notify operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 locally 
on ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Initiating notify operation REDIS-6381_pre_notify_promote_0 on ubusrv1
  notice: Initiating notify operation REDIS-6381:2_pre_notify_promote_0 on 
ubusrv3
  notice: Result of notify operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating promote operation REDIS-6381_promote_0 locally on ubusrv2
  notice: Requesting local execution of promote operation for REDIS-6381 on 
ubusrv2
  notice: Result of promote operation for REDIS-6381 on ubusrv2: ok
  notice: Initiating notify operation REDIS-6381_post_notify_promote_0 locally 
on ubusrv2
  notice: Requesting local execution of notify operation for REDIS-6381 on 
ubusrv2
  notice: Initiating notify operation REDIS-6381_post_notify_promote_0 on 
ubusrv1
  notice: Initiating notify operation REDIS-6381:2_post_notify_promote_0 on 
ubusrv3
  notice: Result of notify operation for RED

Re: [ClusterLabs] make promoted follow promoted resource ?

2023-11-26 Thread lejeczek via Users



On 26/11/2023 10:32, lejeczek via Users wrote:

Hi guys.

With these:

-> $ pcs resource status REDIS-6381-clone
  * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable):
    * Promoted: [ ubusrv2 ]
    * Unpromoted: [ ubusrv1 ubusrv3 ]

-> $ pcs resource status PGSQL-PAF-5433-clone
  * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] 
(promotable):

    * Promoted: [ ubusrv1 ]
    * Unpromoted: [ ubusrv2 ubusrv3 ]

-> $ pcs constraint ref REDIS-6381-clone
Resource: REDIS-6381-clone
  colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY

basically promoted Redis should follow promoted pgSQL but 
it's not happening, usually it does.
I presume pcs/cluster does something internally which 
results in disobeying/ignoring that _colocation_ 
constraint for these resources.

I presume scoring might play a role:
  REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) 
(rsc-role:Master) (with-rsc-role:Master)

but usually, that scoring works, only "now" it does not.
Any comments I appreciate much.
thanks, L.

I looked at pamaker log - snippet below after 
REDIS-6381-clone re-enabled - but cannot see explanation 
for this.

...
 notice: Calculated transition 110, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3729.bz2
 notice: Transition 110 (Complete=0, Pending=0, Fired=0, 
Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete

 notice: State transition S_TRANSITION_ENGINE -> S_IDLE
 notice: State transition S_IDLE -> S_POLICY_ENGINE
 notice: Actions: Start  REDIS-6381:0 
(    ubusrv2 )
 notice: Actions: Start  REDIS-6381:1 
(    ubusrv3 )
 notice: Actions: Start  REDIS-6381:2 
(    ubusrv1 )
 notice: Calculated transition 111, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3730.bz2
 notice: Initiating start operation REDIS-6381_start_0 
locally on ubusrv2
 notice: Requesting local execution of start operation for 
REDIS-6381 on ubusrv2

(to redis) root on none
pam_unix(su:session): session opened for user 
redis(uid=127) by (uid=0)
pam_sss(su:session): Request to sssd failed. Connection 
refused

pam_unix(su:session): session closed for user redis
pam_sss(su:session): Request to sssd failed. Connection 
refused

 notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000
 notice: Transition 111 aborted by 
status-2-master-REDIS-6381 doing create 
master-REDIS-6381=1000: Transient attribute change

INFO: demote: Setting master to 'no-such-master'
 notice: Result of start operation for REDIS-6381 on 
ubusrv2: ok
 notice: Transition 111 (Complete=4, Pending=0, Fired=0, 
Skipped=1, Incomplete=14, 
Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped
 notice: Actions: Promote    REDIS-6381:0 ( 
Unpromoted -> Promoted ubusrv2 )
 notice: Actions: Start  REDIS-6381:1 
(    ubusrv1 )
 notice: Actions: Start  REDIS-6381:2 
(    ubusrv3 )
 notice: Calculated transition 112, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3731.bz2
 notice: Initiating notify operation 
REDIS-6381_pre_notify_start_0 locally on ubusrv2
 notice: Requesting local execution of notify operation 
for REDIS-6381 on ubusrv2
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating start operation REDIS-6381_start_0 on 
ubusrv1
 notice: Initiating start operation REDIS-6381:2_start_0 
on ubusrv3
 notice: Initiating notify operation 
REDIS-6381_post_notify_start_0 locally on ubusrv2
 notice: Requesting local execution of notify operation 
for REDIS-6381 on ubusrv2
 notice: Initiating notify operation 
REDIS-6381_post_notify_start_0 on ubusrv1
 notice: Initiating notify operation 
REDIS-6381:2_post_notify_start_0 on ubusrv3
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating notify operation 
REDIS-6381_pre_notify_promote_0 locally on ubusrv2
 notice: Requesting local execution of notify operation 
for REDIS-6381 on ubusrv2
 notice: Initiating notify operation 
REDIS-6381_pre_notify_promote_0 on ubusrv1
 notice: Initiating notify operation 
REDIS-6381:2_pre_notify_promote_0 on ubusrv3
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating promote operation REDIS-6381_promote_0 
locally on ubusrv2
 notice: Requesting local execution of promote operation 
for REDIS-6381 on ubusrv2
 notice: Result of promote operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating notify operation 
REDIS-6381_post_notify_promote_0 locally on ubusrv2
 notice: Requesting local execution of notify operation 
for REDIS-6381 on ubusrv2
 notice: Initiating notify operation 
REDIS-6381_post_notify_promote_0 on ubusrv1
 notice: Initiating notify operation 
REDIS-6381:2_post_notify_promote_0 on ubusrv3
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok

 notice: Setting master-REDIS-6381[ubusrv3]: (unset) -> 1
 notice: Transition 112

[ClusterLabs] make promoted follow promoted resource ?

2023-11-26 Thread lejeczek via Users

Hi guys.

With these:

-> $ pcs resource status REDIS-6381-clone
  * Clone Set: REDIS-6381-clone [REDIS-6381] (promotable):
    * Promoted: [ ubusrv2 ]
    * Unpromoted: [ ubusrv1 ubusrv3 ]

-> $ pcs resource status PGSQL-PAF-5433-clone
  * Clone Set: PGSQL-PAF-5433-clone [PGSQL-PAF-5433] 
(promotable):

    * Promoted: [ ubusrv1 ]
    * Unpromoted: [ ubusrv2 ubusrv3 ]

-> $ pcs constraint ref REDIS-6381-clone
Resource: REDIS-6381-clone
  colocation-REDIS-6381-clone-PGSQL-PAF-5433-clone-INFINITY

basically promoted Redis should follow promoted pgSQL but 
it's not happening, usually it does.
I presume pcs/cluster does something internally which 
results in disobeying/ignoring that _colocation_ constraint 
for these resources.

I presume scoring might play a role:
  REDIS-6385-clone with PGSQL-PAF-5435-clone (score:1001) 
(rsc-role:Master) (with-rsc-role:Master)

but usually, that scoring works, only "now" it does not.
Any comments I appreciate much.
thanks, L.

I looked at pamaker log - snippet below after 
REDIS-6381-clone re-enabled - but cannot see explanation for 
this.

...
 notice: Calculated transition 110, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3729.bz2
 notice: Transition 110 (Complete=0, Pending=0, Fired=0, 
Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-input-3729.bz2): Complete

 notice: State transition S_TRANSITION_ENGINE -> S_IDLE
 notice: State transition S_IDLE -> S_POLICY_ENGINE
 notice: Actions: Start  REDIS-6381:0 
(    ubusrv2 )
 notice: Actions: Start  REDIS-6381:1 
(    ubusrv3 )
 notice: Actions: Start  REDIS-6381:2 
(    ubusrv1 )
 notice: Calculated transition 111, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3730.bz2
 notice: Initiating start operation REDIS-6381_start_0 
locally on ubusrv2
 notice: Requesting local execution of start operation for 
REDIS-6381 on ubusrv2

(to redis) root on none
pam_unix(su:session): session opened for user redis(uid=127) 
by (uid=0)

pam_sss(su:session): Request to sssd failed. Connection refused
pam_unix(su:session): session closed for user redis
pam_sss(su:session): Request to sssd failed. Connection refused
 notice: Setting master-REDIS-6381[ubusrv2]: (unset) -> 1000
 notice: Transition 111 aborted by 
status-2-master-REDIS-6381 doing create 
master-REDIS-6381=1000: Transient attribute change

INFO: demote: Setting master to 'no-such-master'
 notice: Result of start operation for REDIS-6381 on 
ubusrv2: ok
 notice: Transition 111 (Complete=4, Pending=0, Fired=0, 
Skipped=1, Incomplete=14, 
Source=/var/lib/pacemaker/pengine/pe-input-3730.bz2): Stopped
 notice: Actions: Promote    REDIS-6381:0 ( 
Unpromoted -> Promoted ubusrv2 )
 notice: Actions: Start  REDIS-6381:1 
(    ubusrv1 )
 notice: Actions: Start  REDIS-6381:2 
(    ubusrv3 )
 notice: Calculated transition 112, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-3731.bz2
 notice: Initiating notify operation 
REDIS-6381_pre_notify_start_0 locally on ubusrv2
 notice: Requesting local execution of notify operation for 
REDIS-6381 on ubusrv2
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating start operation REDIS-6381_start_0 on 
ubusrv1
 notice: Initiating start operation REDIS-6381:2_start_0 on 
ubusrv3
 notice: Initiating notify operation 
REDIS-6381_post_notify_start_0 locally on ubusrv2
 notice: Requesting local execution of notify operation for 
REDIS-6381 on ubusrv2
 notice: Initiating notify operation 
REDIS-6381_post_notify_start_0 on ubusrv1
 notice: Initiating notify operation 
REDIS-6381:2_post_notify_start_0 on ubusrv3
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating notify operation 
REDIS-6381_pre_notify_promote_0 locally on ubusrv2
 notice: Requesting local execution of notify operation for 
REDIS-6381 on ubusrv2
 notice: Initiating notify operation 
REDIS-6381_pre_notify_promote_0 on ubusrv1
 notice: Initiating notify operation 
REDIS-6381:2_pre_notify_promote_0 on ubusrv3
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating promote operation REDIS-6381_promote_0 
locally on ubusrv2
 notice: Requesting local execution of promote operation 
for REDIS-6381 on ubusrv2
 notice: Result of promote operation for REDIS-6381 on 
ubusrv2: ok
 notice: Initiating notify operation 
REDIS-6381_post_notify_promote_0 locally on ubusrv2
 notice: Requesting local execution of notify operation for 
REDIS-6381 on ubusrv2
 notice: Initiating notify operation 
REDIS-6381_post_notify_promote_0 on ubusrv1
 notice: Initiating notify operation 
REDIS-6381:2_post_notify_promote_0 on ubusrv3
 notice: Result of notify operation for REDIS-6381 on 
ubusrv2: ok

 notice: Setting master-REDIS-6381[ubusrv3]: (unset) -> 1
 notice: Transition 112 aborted by 
status-3-master-REDIS-6381 doing create master-REDIS-6381=1: 
Transient 

Re: [ClusterLabs] [EXT] moving VM live fails?

2023-11-25 Thread lejeczek via Users




On 24/11/2023 08:33, Windl, Ulrich wrote:

Hi!

So you have different CPUs in the cluster? We once had a similar situation with Xen using 
live migration: Migration failed, and the cluster "wasn't that smart" handling 
the situation. The solution was (with the help of support) to add some CPU flags masking 
in the VM configuration so that the newer CPU features were not used.
Before migration worked from the older to the newer CPU and back if the VM had 
been started on the older CPU initially, but when it had been started on the 
newer CPU it could not migrate to the older one. I vaguely remember the details 
but the hardware was a ProLiant 380 G7 vs. 380 G8 or so.
My recommendation is: If you want to size a cluster, but all the nodes you 
need, because if you want to add another node maybe in two years or so, the 
older model may not be available any more. IT lifetime is short.

Regards,
Ulrich

-Original Message-
From: Users  On Behalf Of lejeczek via Users
Sent: Friday, November 17, 2023 12:55 PM
To: users@clusterlabs.org
Cc: lejeczek 
Subject: [EXT] [ClusterLabs] moving VM live fails?

Hi guys.

I have a resource which when asked to 'move' then it fails with:

  virtqemud[3405456]: operation failed: guest CPU doesn't match specification: 
missing features: xsave

but VM domain does not require (nor disable) the feature:

   
 https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/
This must go down - same environment - all the way, to the 
kernel. I knew that but forgotten - one box did change 
default kernel & after reboot booted with Centos's default 
as opposed to other boxes running with ver. 6.x.

So, my bad, not really an issue.
Thanks,. L
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] node orderly shutdown

2023-11-21 Thread lejeczek via Users

Hi guys.

Having a node with a couple of _promoted_ resources - when 
such node is os-shutdown in an orderly manner it seems that 
cluster takes a while.
By a "while" I mean longer than I'd expect a relatively 
simple 3-node cluster to move/promote a few _promoted_ 
resources:

redis, postgresql, IP
onto another.

Is there somewhere one can look, tweak or measure & 
troubleshot, in order to "fix" this, if possible at all?


From watching such a "promoted" node I see that as systemd 
stops all services going into power-down target - it's 
_pacemaker_ which as last systemd takes bit longer before 
complete shutdown.
Or perhaps you have another & more than one approach / 
technique to node-with-promoted-resources shutdown, a better 
one?


many thanks,___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


does Firefox honor local self-signed CAs ?

2023-11-19 Thread lejeczek via users

Hi guys

I put my own CAs certs into: /etc/pki/ca-trust/source/anchors
then _update-ca-trust_
My other web-browsers seem happy but Firefox always 
shows/warns: connection not secure

as oppose to other browsers saying: is secure

Something has changed in default behavior - I remember 
Firefox did not do it - even if it was possible to "tweak" 
Firefox back to "normal"


would you know?
thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: zbar VS qrencode

2023-11-19 Thread lejeczek via users



On 19/11/2023 14:17, lejeczek via users wrote:

Hi guys

I have a peculiar case, I think, while using national chars.
I wonder which one is _playing_ up here, qrencode or zbarimg

When I with:
-> qrencode -l L -v 1 ..
encode:

ADR;LABEL="Kukólka":;;;Kukólka;Mazowieckie;PL
then:

-> $ zbarimg ...
...
ADR;LABEL="Kuk籀lka":;;;Kuk籀lka;Mazowieckie;PL

but when I encode:
ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL
then:

-> $ zbarimg 
...
ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL

you see - if all the "processors" between our email 
clients, mine your yours, did not mangle the content - it 
seems when there is only *single* polish char - ó - then 
_zbaring_ shows that weird output - a Korean or Japanese?
Any other case, add just one morenational char in the line 
and the output, even that same - ó - mangled when single, 
now shows as it should.
So I checked all nine polish national chars and it's: ó 
and ż when occur as single in a line - _zbarimg_ shows 
them incorrectly. (or _qrencode_ fails?)


Any _qrencode_ & _zbarimg_ users here who use their 
national chars?

I'll be glad for all comments.

thanks, L.


Ah, to answer my question - partly at least - seems that 
_zbarimg_ could use some fixing up.

Trying _qrscanner_ and here I get everything shown correctly.

Would anybody know if _zbarimg_ is looked after still, if 
it's worth filing a bug?
--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


zbar VS qrencode

2023-11-19 Thread lejeczek via users

Hi guys

I have a peculiar case, I think, while using national chars.
I wonder which one is _playing_ up here, qrencode or zbarimg

When I with:
-> qrencode -l L -v 1 ..
encode:

ADR;LABEL="Kukólka":;;;Kukólka;Mazowieckie;PL
then:

-> $ zbarimg ...
...
ADR;LABEL="Kuk籀lka":;;;Kuk籀lka;Mazowieckie;PL

but when I encode:
ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL
then:

-> $ zbarimg 
...
ADR;LABEL="Kukółka":;;;Kukółka;Mazowieckie;PL

you see - if all the "processors" between our email clients, 
mine your yours, did not mangle the content - it seems when 
there is only *single* polish char - ó - then _zbaring_ 
shows that weird output - a Korean or Japanese?
Any other case, add just one morenational char in the line 
and the output, even that same - ó - mangled when single, 
now shows as it should.
So I checked all nine polish national chars and it's: ó and 
ż when occur as single in a line - _zbarimg_ shows them 
incorrectly. (or _qrencode_ fails?)


Any _qrencode_ & _zbarimg_ users here who use their national 
chars?

I'll be glad for all comments.

thanks, L.--
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ClusterLabs] non-existent attribute ?

2023-11-19 Thread lejeczek via Users

Hi guys.

My 3-node cluster had one node absent for a long time and 
now when it's back I cannot get _mariadb_ to start on that node.

...
    * MARIADB    (ocf:heartbeat:galera):     ORPHANED Stopped
...
    * MARIADB-last-committed  : 147
    * MARIADB-safe-to-bootstrap   : 0

I wanted to start with a _cleanup_ of node's attribute - as 
mariadb node outside of ha-pcs starts & run a ok, but I get:


-> $ pcs node attribute dzien MARIADB-last-committed=''
Error: attribute: 'MARIADB-last-committed' doesn't exist for 
node: 'dzien'


Is possible to clean that up safely?

many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] [EXT] Re: PAF / pgSQL fails after OS/system shutdown - FIX

2023-11-18 Thread lejeczek via Users




On 13/11/2023 13:08, Jehan-Guillaume de Rorthais via Users 
wrote:

On Mon, 13 Nov 2023 11:39:45 +
"Windl, Ulrich"  wrote:


But shouldn't the RA check for that (and act appropriately)?

Interesting. I'm open to discuss this. Below my thoughts so far.

Why the RA should check that? There's so many way to setup the system and
PostgreSQL, where should the RA stop checking for all possible way to break it?

The RA checks various (maybe too many) things related to the instance itself
already.

I know various other PostgreSQL setups that would trigger errors in the cluster
if the dba doesn't check everything is correct. I'm really reluctant to
add add a fair amount of code in the RA to correctly parse and check the
complex PostgreSQL's setup. This would add complexity and bugs. Or maybe I
could add a specific OCF_CHECK_LEVEL sysadmins can trigger by hand before
starting the cluster. But I wonder if it worth the pain, how many people will
know about this and actually run it?

The problem here is that few users actually realize how the postgresql-common
wrapper works and what it actually does behind your back. I really appreciate
this wrapper, I do. But when you setup a Pacemaker cluster, you either have to
bend to it when setting up PAF (as documented), or avoid it completely.

PAF is all about drawing a clear line between the sysadmin job and the
dba one. Dba must build a cluster of instances ready to start/replicate with
standard binaries (not wrappers) before sysadmin can set up the resource in your
cluster.

Thoughts?


I would be the same/similar mind - which is - adding more 
code to account for more/all _config_ cases may not be the 
healthiest approach, however!!


Just like some code-writers out there in around the globe 
here too, some do seem to _not_ appreciate or seem to 
completely ignore, parts which are meant do alleviate the 
TCO burden of any software - documentation best bits which 
for Unix/Linux... are MAN PAGES.
This is rhetorical but I'll ask - what is a project worth 
when its own man pages miss critical _gotchas_ and similar 
gimmicks.


If you read here are are a prospective/future coder-writer 
for Linux please make such note - your 'man pages" ought to 
be at least!! as good as your code.

many thanks, L.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] moving VM live fails?

2023-11-17 Thread lejeczek via Users

Hi guys.

I have a resource which when asked to 'move' then it fails with:

 virtqemud[3405456]: operation failed: guest CPU doesn't 
match specification: missing features: xsave


but VM domain does not require (nor disable) the feature:

  
    what even more interesting, _virsh_ migrate does the 
migration live a ok.


I'm on Centos 9 with packages up to day.

All thoughts much appreciated.
many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown - FIX

2023-11-10 Thread lejeczek via Users




On 10/11/2023 18:16, Jehan-Guillaume de Rorthais wrote:

On Fri, 10 Nov 2023 17:17:41 +0100
lejeczek via Users  wrote:

...

Of course you can use "pg_stat_tmp", just make sure the temp folder exists:

cat < /etc/tmpfiles.d/postgresql-part.conf
# Directory for PostgreSQL temp stat files
d /var/run/postgresql/14-paf.pg_stat_tmp 0700 postgres postgres - -
EOF

To take this file in consideration immediately without rebooting the server,
run the following command:

systemd-tmpfiles --create /etc/tmpfiles.d/postgresql-part.conf

Then there must be something else at play here with Ubuntus,
for none of the nodes has any extra/additional configs for
those paths & I'm sure that those were not created manually.

Indeed.

This parameter is usually set by pg_createcluster command and the folder
created by both pg_createcluster and pg_ctlcluster commands when needed.

This is explained in PAF tutorial there:

https://clusterlabs.github.io/PAF/Quick_Start-Debian-10-pcs.html#postgresql-and-cluster-stack-installation

These commands comes from the postgresql-common wrapper, used in all Debian
related distros, allowing to install, create and use multiple PostgreSQL
versions on the same server.


Perhpaphs pgSQL created these on it's own outside of HA-cluster.

No, the Debian packaging did.

Just create the config file I pointed you in my previous answer,
systemd-tmpfiles will take care of it and you'll be fine.


yes, that was a weird trip.
I could not take the whole cluster down & as I kept fiddling 
with it trying to "fix" it - within & out of _pcs_ those 
paths were created by wrappers from disto packages - I know 
see that - but at first it did not look that way.
Still that directive for the stats - I wonder if that got 
introduced, injected somewhere in between because the PG 
cluster - which I had for a while - did not experience this 
"issue" from the start and not until "recently"

thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown - FIX

2023-11-10 Thread lejeczek via Users



On 10/11/2023 13:13, Jehan-Guillaume de Rorthais wrote:

On Fri, 10 Nov 2023 12:27:24 +0100
lejeczek via Users  wrote:
...
  

to share my "fix" for it - perhaps it was introduced by
OS/packages (Ubuntu 22) updates - ? - as oppose to resource
agent itself.

As the logs point out - pg_stat_tmp - is missing and from
what I see it's only the master, within a cluster, doing
those stats.
That appeared, I use the word for I did not put it into
configs, on all nodes.
fix = to not use _pg_stat_tmp_ directive/option at all.

Of course you can use "pg_stat_tmp", just make sure the temp folder exists:

   cat < /etc/tmpfiles.d/postgresql-part.conf
   # Directory for PostgreSQL temp stat files
   d /var/run/postgresql/14-paf.pg_stat_tmp 0700 postgres postgres - -
   EOF

To take this file in consideration immediately without rebooting the server,
run the following command:

   systemd-tmpfiles --create /etc/tmpfiles.d/postgresql-part.conf
Then there must be something else at play here with Ubuntus, 
for none of the nodes has any extra/additional configs for 
those paths & I'm sure that those were not created manually.
Perhpaphs pgSQL created these on it's own outside of 
HA-cluster. Also, one node has the path missing but the 
other two  have & between those two only - how it seems to 
me -  the master PG actually put any data there and if that 
is the case - does the existence of the path alone guarantee 
anything, is the question.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown - FIX

2023-11-10 Thread lejeczek via Users



On 07/11/2023 17:57, lejeczek via Users wrote:

hi guys

Having 3-node pgSQL cluster with PAF - when all three 
systems are shutdown at virtually the same time then PAF 
fails to start when HA cluster is operational again.


from status:
...
Migration Summary:
  * Node: ubusrv2 (2):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'

  * Node: ubusrv3 (3):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'

  * Node: ubusrv1 (1):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'


Failed Resource Actions:
  * PGSQL-PAF-5433_stop_0 on ubusrv2 'error' (1): call=90, 
status='complete', exitreason='Unexpected state for 
instance "PGSQL-PAF-5433" (returned 1)', 
last-rc-change='Tue Nov  7 17:52:38 2023', queued=0ms, 
exec=84ms
  * PGSQL-PAF-5433_stop_0 on ubusrv3 'error' (1): call=82, 
status='complete', exitreason='Unexpected state for 
instance "PGSQL-PAF-5433" (returned 1)', 
last-rc-change='Tue Nov  7 17:52:38 2023', queued=0ms, 
exec=82ms
  * PGSQL-PAF-5433_stop_0 on ubusrv1 'error' (1): call=86, 
status='complete', exitreason='Unexpected state for 
instance "PGSQL-PAF-5433" (returned 1)', 
last-rc-change='Tue Nov  7 17:52:38 2023', queued=0ms, 
exec=108ms


and all three pgSQLs show virtually identical logs:
...
2023-11-07 16:54:45.532 UTC [24936] LOG:  starting 
PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on 
x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 
11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit
2023-11-07 16:54:45.532 UTC [24936] LOG:  listening on 
IPv4 address "0.0.0.0", port 5433
2023-11-07 16:54:45.532 UTC [24936] LOG:  listening on 
IPv6 address "::", port 5433
2023-11-07 16:54:45.535 UTC [24936] LOG:  listening on 
Unix socket "/var/run/postgresql/.s.PGSQL.5433"
2023-11-07 16:54:45.547 UTC [24938] LOG:  database system 
was interrupted while in recovery at log time 2023-11-07 
15:30:56 UTC
2023-11-07 16:54:45.547 UTC [24938] HINT:  If this has 
occurred more than once some data might be corrupted and 
you might need to choose an earlier recovery target.
2023-11-07 16:54:45.819 UTC [24938] LOG:  entering standby 
mode
2023-11-07 16:54:45.824 UTC [24938] FATAL:  could not open 
directory "/var/run/postgresql/14-paf.pg_stat_tmp": No 
such file or directory
2023-11-07 16:54:45.825 UTC [24936] LOG:  startup process 
(PID 24938) exited with exit code 1
2023-11-07 16:54:45.825 UTC [24936] LOG:  aborting startup 
due to startup process failure
2023-11-07 16:54:45.826 UTC [24936] LOG:  database system 
is shut down


Is this "test" case's result, as I showed above, expected? 
It reproduces every time.

If not - what might it be I'm missing?

many thanks, L.

___

to share my "fix" for it - perhaps it was introduced by 
OS/packages (Ubuntu 22) updates - ? - as oppose to resource 
agent itself.


As the logs point out - pg_stat_tmp - is missing and from 
what I see it's only the master, within a cluster, doing 
those stats.
That appeared, I use the word for I did not put it into 
configs, on all nodes.

fix = to not use _pg_stat_tmp_ directive/option at all.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] PAF / pgSQL fails after OS/system shutdown

2023-11-09 Thread lejeczek via Users



On 07/11/2023 17:57, lejeczek via Users wrote:

hi guys

Having 3-node pgSQL cluster with PAF - when all three 
systems are shutdown at virtually the same time then PAF 
fails to start when HA cluster is operational again.


from status:
...
Migration Summary:
  * Node: ubusrv2 (2):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'

  * Node: ubusrv3 (3):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'

  * Node: ubusrv1 (1):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'


Failed Resource Actions:
  * PGSQL-PAF-5433_stop_0 on ubusrv2 'error' (1): call=90, 
status='complete', exitreason='Unexpected state for 
instance "PGSQL-PAF-5433" (returned 1)', 
last-rc-change='Tue Nov  7 17:52:38 2023', queued=0ms, 
exec=84ms
  * PGSQL-PAF-5433_stop_0 on ubusrv3 'error' (1): call=82, 
status='complete', exitreason='Unexpected state for 
instance "PGSQL-PAF-5433" (returned 1)', 
last-rc-change='Tue Nov  7 17:52:38 2023', queued=0ms, 
exec=82ms
  * PGSQL-PAF-5433_stop_0 on ubusrv1 'error' (1): call=86, 
status='complete', exitreason='Unexpected state for 
instance "PGSQL-PAF-5433" (returned 1)', 
last-rc-change='Tue Nov  7 17:52:38 2023', queued=0ms, 
exec=108ms


and all three pgSQLs show virtually identical logs:
...
2023-11-07 16:54:45.532 UTC [24936] LOG:  starting 
PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on 
x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 
11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit
2023-11-07 16:54:45.532 UTC [24936] LOG:  listening on 
IPv4 address "0.0.0.0", port 5433
2023-11-07 16:54:45.532 UTC [24936] LOG:  listening on 
IPv6 address "::", port 5433
2023-11-07 16:54:45.535 UTC [24936] LOG:  listening on 
Unix socket "/var/run/postgresql/.s.PGSQL.5433"
2023-11-07 16:54:45.547 UTC [24938] LOG:  database system 
was interrupted while in recovery at log time 2023-11-07 
15:30:56 UTC
2023-11-07 16:54:45.547 UTC [24938] HINT:  If this has 
occurred more than once some data might be corrupted and 
you might need to choose an earlier recovery target.
2023-11-07 16:54:45.819 UTC [24938] LOG:  entering standby 
mode
2023-11-07 16:54:45.824 UTC [24938] FATAL:  could not open 
directory "/var/run/postgresql/14-paf.pg_stat_tmp": No 
such file or directory
2023-11-07 16:54:45.825 UTC [24936] LOG:  startup process 
(PID 24938) exited with exit code 1
2023-11-07 16:54:45.825 UTC [24936] LOG:  aborting startup 
due to startup process failure
2023-11-07 16:54:45.826 UTC [24936] LOG:  database system 
is shut down


Is this "test" case's result, as I showed above, expected? 
It reproduces every time.

If not - what might it be I'm missing?

many thanks, L.

Actually, the  resource fails to start on a node a single 
node - as opposed to entire cluster shutdown as I noted 
originally - which was powered down in an orderly fashion 
and powered back on.
That the the time of power-cycle the node was PAF resource 
master, it fails:

...
2023-11-09 20:35:04.439 UTC [17727] LOG:  starting 
PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on 
x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 
11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit
2023-11-09 20:35:04.439 UTC [17727] LOG:  listening on IPv4 
address "0.0.0.0", port 5433
2023-11-09 20:35:04.439 UTC [17727] LOG:  listening on IPv6 
address "::", port 5433
2023-11-09 20:35:04.442 UTC [17727] LOG:  listening on Unix 
socket "/var/run/postgresql/.s.PGSQL.5433"
2023-11-09 20:35:04.452 UTC [17731] LOG:  database system 
was interrupted while in recovery at log time 2023-11-09 
20:25:21 UTC
2023-11-09 20:35:04.452 UTC [17731] HINT:  If this has 
occurred more than once some data might be corrupted and you 
might need to choose an earlier recovery target.

2023-11-09 20:35:04.809 UTC [17731] LOG:  entering standby mode
2023-11-09 20:35:04.813 UTC [17731] FATAL:  could not open 
directory "/var/run/postgresql/14-paf.pg_stat_tmp": No such 
file or directory
2023-11-09 20:35:04.814 UTC [17727] LOG:  startup process 
(PID 17731) exited with exit code 1
2023-11-09 20:35:04.814 UTC [17727] LOG:  aborting startup 
due to startup process failure
2023-11-09 20:35:04.815 UTC [17727] LOG:  database system is 
shut down



The master at the time node was shut down did get moved over 
to standby/slave node, properly,


I'm on Ubuntu with:

ii  corosync  3.1.6-1ubuntu1   amd64    
cluster engine daemon and utilities
ii  pacemaker 2.1.2-1ubuntu3.1 amd64    
cluster resource manager
ii  pacemaker-cli-utils   2.1.2-1ubuntu3.1 amd64    
cluster resource manager command line utilities
ii  pacemaker-common  2.1.2-1ubuntu3.1 all  
cluster resource manager common files
ii  pacemaker-resource-agents 2.1.2-1ubuntu3.1 all  
cluster res

Re: freaking TFTP

2023-11-08 Thread lejeczek via users



On 07/11/2023 15:42, Chris Adams wrote:

Once upon a time, lejeczek  said:

2nd meanwhile - I'm trying _dnsmasq_ which I'm new to thus errors
are possible but... it seems that the same issue remains.

with _dnsmasq_. tftp client - as with in.tftp as the server - "times
out" but _dnsmasq_ server thinks and says that:

I use dnsmasq's TFTP server on a bridge, with this config:

bind-interfaces
interface=br0
port=0
enable-tftp
tftp-root=/srv/tftpboot

"port=0" disables DNS, and I then don't configure DHCP, so it's just a
TFTP server.

I also should have added that it's Centos I'm doing all this 
on, with kernel from _elrepo_

Anybody here do/use Centos that way, with net bridges?
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: freaking TFTP

2023-11-07 Thread lejeczek via users



On 07/11/2023 15:42, Chris Adams wrote:

Once upon a time, lejeczek  said:

2nd meanwhile - I'm trying _dnsmasq_ which I'm new to thus errors
are possible but... it seems that the same issue remains.

with _dnsmasq_. tftp client - as with in.tftp as the server - "times
out" but _dnsmasq_ server thinks and says that:

I use dnsmasq's TFTP server on a bridge, with this config:

bind-interfaces
interface=br0
port=0
enable-tftp
tftp-root=/srv/tftpboot

"port=0" disables DNS, and I then don't configure DHCP, so it's just a
TFTP server.


I started with that - got logs as shown in my last email.
One thing I should perhaps mention, is that libvirt uses my 
NM bridges, but it's simple:

-> $ virsh net-dumpxml 10_1_1

  10_1_1
  864fb78e-0fb0-4c32-bb47-0e5ac68d9491
  
  


and I guess, but that does not "add" to on-bare-metal NM 
bridge.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ClusterLabs] PAF / pgSQL fails after OS/system shutdown

2023-11-07 Thread lejeczek via Users

hi guys

Having 3-node pgSQL cluster with PAF - when all three 
systems are shutdown at virtually the same time then PAF 
fails to start when HA cluster is operational again.


from status:
...
Migration Summary:
  * Node: ubusrv2 (2):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'

  * Node: ubusrv3 (3):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'

  * Node: ubusrv1 (1):
    * PGSQL-PAF-5433: migration-threshold=100 
fail-count=100 last-failure='Tue Nov  7 17:52:38 2023'


Failed Resource Actions:
  * PGSQL-PAF-5433_stop_0 on ubusrv2 'error' (1): call=90, 
status='complete', exitreason='Unexpected state for instance 
"PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov  7 
17:52:38 2023', queued=0ms, exec=84ms
  * PGSQL-PAF-5433_stop_0 on ubusrv3 'error' (1): call=82, 
status='complete', exitreason='Unexpected state for instance 
"PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov  7 
17:52:38 2023', queued=0ms, exec=82ms
  * PGSQL-PAF-5433_stop_0 on ubusrv1 'error' (1): call=86, 
status='complete', exitreason='Unexpected state for instance 
"PGSQL-PAF-5433" (returned 1)', last-rc-change='Tue Nov  7 
17:52:38 2023', queued=0ms, exec=108ms


and all three pgSQLs show virtually identical logs:
...
2023-11-07 16:54:45.532 UTC [24936] LOG:  starting 
PostgreSQL 14.9 (Ubuntu 14.9-0ubuntu0.22.04.1) on 
x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 
11.4.0-1ubuntu1~22.04) 11.4.0, 64-bit
2023-11-07 16:54:45.532 UTC [24936] LOG:  listening on IPv4 
address "0.0.0.0", port 5433
2023-11-07 16:54:45.532 UTC [24936] LOG:  listening on IPv6 
address "::", port 5433
2023-11-07 16:54:45.535 UTC [24936] LOG:  listening on Unix 
socket "/var/run/postgresql/.s.PGSQL.5433"
2023-11-07 16:54:45.547 UTC [24938] LOG:  database system 
was interrupted while in recovery at log time 2023-11-07 
15:30:56 UTC
2023-11-07 16:54:45.547 UTC [24938] HINT:  If this has 
occurred more than once some data might be corrupted and you 
might need to choose an earlier recovery target.

2023-11-07 16:54:45.819 UTC [24938] LOG:  entering standby mode
2023-11-07 16:54:45.824 UTC [24938] FATAL:  could not open 
directory "/var/run/postgresql/14-paf.pg_stat_tmp": No such 
file or directory
2023-11-07 16:54:45.825 UTC [24936] LOG:  startup process 
(PID 24938) exited with exit code 1
2023-11-07 16:54:45.825 UTC [24936] LOG:  aborting startup 
due to startup process failure
2023-11-07 16:54:45.826 UTC [24936] LOG:  database system is 
shut down


Is this "test" case's result, as I showed above, expected? 
It reproduces every time.

If not - what might it be I'm missing?

many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: freaking TFTP

2023-11-07 Thread lejeczek via users



On 07/11/2023 13:43, Mauricio Tavares wrote:

On Tue, Nov 7, 2023 at 7:15 AM lejeczek via users
  wrote:

Hi guys.

cannot get a client to get anything from 'in.tftp' server?

Connection seems to get established, _tcpdump_ shows:
...
13:13:08.331277 IP 10.3.1.100.35176 > whale.mine.priv.tftp: TFTP, length 19, RRQ 
"zabbix.zip" octet
13:13:13.331324 IP 10.3.1.100.35176 > whale.mine.priv.tftp: TFTP, length 19, RRQ 
"zabbix.zip" octet
...
journal log for tftp.service, so on the servers, shows:
...
tftpd: read(ack): No route to host
...

any & all thoughts on how troubleshoot & fix this are much appreciated.
many thanks, L.
___

   Can you connect from client host to tftp host using other services?
___

Yes, everything works as... well, as expected/normal.
Meanwhile...

I've read through my old notes and found:

.. When you use in.tftpd on interface like bridges, this 
probably won't work, it at least did not for Dell 
PowerConnect when you wanted to upload the image, it worked 
on regualar eth


And it today, a couple of years later and different 
hardware, seems that still with bridge ifaces, for both 
server & client, the issue remains.


2nd meanwhile - I'm trying _dnsmasq_ which I'm new to thus 
errors are possible but... it seems that the same issue remains.


with _dnsmasq_. tftp client - as with in.tftp as the server 
- "times out" but _dnsmasq_ server thinks and says that:


...

sent /var/lib/tftpboot/zabbix.zip to 10.3.1.100
...

This might go as far down as kernel's tweaking or perhaps 
bridge's iface - if possible at all - any kernel/net experts 
are most welcome to comment.


many thanks, L.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


freaking TFTP

2023-11-07 Thread lejeczek via users

Hi guys.

cannot get a client to get anything from 'in.tftp' server?

Connection seems to get established, _tcpdump_ shows:
...
13:13:08.331277 IP 10.3.1.100.35176 > whale.mine.priv.tftp: 
TFTP, length 19, RRQ "zabbix.zip" octet
13:13:13.331324 IP 10.3.1.100.35176 > whale.mine.priv.tftp: 
TFTP, length 19, RRQ "zabbix.zip" octet

...
journal log for tftp.service, so on the servers, shows:
...
tftpd: read(ack): No route to host
...

any & all thoughts on how troubleshoot & fix this are much 
appreciated.

many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


MongoDB & lots of SELinux denials

2023-11-03 Thread lejeczek via users

Hi guys.

 I’ve just installed vanilla-default MongoDB (following 
their official docs) and right away SELinux shows denials:


... SELinux is preventing /usr/bin/mongod from search access 
on the directory 
/var/lib/containers/storage/overlay-containers/fe6f1040c1373930efe68c777805ecd4c921631e3a87b8806af1fe0cdf266d8b/userdata/shm. 
For complete SELinux messages run: sealert -l 
5263cddc-183a-472b-9098-bf4599c8453c ...


What is Mongo’s business in checking such path (and there is 
more paths)? - I refuse to believe (yet) that my Mongo is 
trojaned.


Many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


LUKS - lost token?

2023-10-28 Thread lejeczek via users

Hi guys.

I know this is most likely not best suited question for this 
list, but I'm hoping some experts might be able to help.


I have a LUKS device which had keyslot with pass-phrase 
removed and token for TPM keyslot removed too - I think this 
is the case, for none of my passphrase works and device is 
as below:


I hope there is a way to save & bring it back to live - 
device is open right now and I've access to filesystem, 
obviously goal would be to avoid re-format/crypt.


Is it possible to restore/recreate that lost token and/or 
add new Keyslot somehow? I have no header backup for this 
device.


-> $ cryptsetup luksDump /dev/nvme0n1p3
LUKS header information
Version:   2
Epoch:     83
Metadata area:     16384 [bytes]
Keyslots area:     1678 [bytes]
UUID:  3a879268-84fd-4b48-a5d4-960cccb0caa9
Label:     (no label)
Subsystem:     (no subsystem)
Flags:   (no flags)

Data segments:
  0: crypt
    offset: 16777216 [bytes]
    length: (whole device)
    cipher: aes-xts-plain64
    sector: 512 [bytes]

Keyslots:
  1: luks2
    Key:    512 bits
    Priority:   normal
    Cipher: aes-xts-plain64
    Cipher key: 512 bits
    PBKDF:  pbkdf2
    Hash:   sha512
    Iterations: 1000
    Salt:   a4 5b 6b cc a8 f1 6b e8 b7 3b e2 3d ca 8d 
43 fb
    10 52 62 b9 99 45 70 16 bd e1 0f 7a 6c 7f 
3d 11

    AF stripes: 4000
    AF hash:    sha512
    Area offset:290816 [bytes]
    Area length:258048 [bytes]
    Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
    Hash:   sha256
    Iterations: 183317
    Salt:   ef 56 aa 59 c2 64 66 c7 49 57 31 4b a7 7d 
00 3c
    fe 00 89 2e b9 e9 da bc 69 1d 19 59 96 a9 
27 aa
    Digest: 79 aa 0c 8a 29 64 9c 83 bb 5a f8 5c b5 c6 
b0 9c
    5e 54 80 49 bd 21 f6 b4 5b 49 65 39 bd 6f 
5f 20
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Ungoogled Chromium - spell check?

2023-09-30 Thread lejeczek via users

Hi guys.

Anybody here uses _Ungoogled Chromium_ ?
If so - does 'spell check' work for you?

many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: how to use amd-pstate - ?

2023-09-22 Thread lejeczek via users



On 29/10/2022 15:19, Matti Pulkkinen wrote:

On Sat, 2022-10-29 at 12:01 +0100, lejeczek via users wrote:

Hi guys.

Does anybody use 'amd-pstate' and if yes could share a howto
on how to have Fedora use that driver?

many thanks, L.


If you have a Zen2 or newer AMD processor, you can use a grubby command
in a terminal to set the needed kernel argument:

sudo grubby --update-kernel DEFAULT --args 
"initcall_blacklist=acpi_cpufreq_init"

Then, after a reboot, you can use this command to see if it worked:

cat /sys/devices/system/cpu/cpu0/cpufreq/scaling_driver

If that says something other than amd-pstate then it did not work, and
you can try adding another kernel argument and then rebooting:

sudo grubby --update-kernel DEFAULT --args "amd_pstate.shared_mem=1"



but is there a module? I cannot find it default/repo kernels.
I see _amd_pstate_ut_ which says is a "test module"
Is there still no official/approved p-state in f38, even 
with 6.5.x?

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Thunderbird & non-english characters ?

2023-09-16 Thread lejeczek via users

Hi guys.
I'm trying to add an account in Thunderbird, address 
contains non-english chars and Thunderbird  stops, does not 
proceed further, colors the field in red & sticks ! mark there.
That surely must be not an issue of Thunderbird/OS but my 
configuration somewhere, right? Email accounts/address with 
non-english chars are a norm nowadays?


thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] PAF / PGSQLMS and wal_level ?

2023-09-13 Thread lejeczek via Users




On 08/09/2023 17:29, Jehan-Guillaume de Rorthais wrote:

On Fri, 8 Sep 2023 16:52:53 +0200
lejeczek via Users  wrote:


Hi guys.

Before I start fiddling and brake things I wonder if
somebody knows if:
pgSQL can work with: |wal_level = archive for PAF ?
Or more general question with pertains to ||wal_level - can
_barman_ be used with pgSQL "under" PAF?

PAF needs "wal_level = replica" (or "hot_standby" on very old versions) so it
can have hot standbys where it can connects and query there status.

Wal level "replica" includes the archive level, so you can set up archiving.

Of course you can use barman or any other tools to manage your PITR Backups,
even when Pacemaker/PAF is looking at your instances. This is even the very
first step you should focus on during your journey to HA.

Regards,
and with _barman_ specifically - is one method preferred, 
recommended over another: streaming VS rsync - for/with PAF?

many thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] PAF / PGSQLMS and wal_level ?

2023-09-08 Thread lejeczek via Users

Hi guys.

Before I start fiddling and brake things I wonder if 
somebody knows if:

pgSQL can work with: |wal_level = archive for PAF ?
Or more general question with pertains to ||wal_level - can 
_barman_ be used with pgSQL "under" PAF?


many thanks, L.
|___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] PAF / PGSQLMS on Ubuntu

2023-09-08 Thread lejeczek via Users



On 07/09/2023 16:20, lejeczek via Users wrote:



On 07/09/2023 16:09, Andrei Borzenkov wrote:
On Thu, Sep 7, 2023 at 5:01 PM lejeczek via Users 
 wrote:

Hi guys.

I'm trying to set ocf_heartbeat_pgsqlms agent but I get:
...
Failed Resource Actions:
   * PGSQL-PAF-5433 stop on ubusrv3 returned 'invalid 
parameter' because 'Parameter "recovery_target_timeline" 
MUST be set to 'latest'. It is currently set to ''' at 
Thu Sep  7 13:58:06 2023 after 54ms


I'm new to Ubuntu and I see that Ubuntu has a bit 
different approach to paths (in comparison to how Centos 
do it).

I see separation between config & data, eg.

14  paf 5433 down   postgres 
/var/lib/postgresql/14/paf 
/var/log/postgresql/postgresql-14-paf.log


I create the resource like here:

-> $ pcs resource create PGSQL-PAF-5433 
ocf:heartbeat:pgsqlms pgport=5433 bindir=/usr/bin 
pgdata=/etc/postgresql/14/paf 
datadir=/var/lib/postgresql/14/paf meta 
failure-timeout=30s master-max=1 op start timeout=60s op 
stop timeout=60s op promote timeout=30s op demote 
timeout=120s op monitor interval=15s timeout=10s 
role="Promoted" op monitor interval=16s timeout=10s 
role="Unpromoted" op notify timeout=60s promotable 
notify=true failure-timeout=30s master-max=1 --disable


Ubuntu 22.04.3 LTS
What am I missing can you tell?
Exactly what the message tells you. You need to set 
recovery_target=latest.

and having it in 'postgresql.conf' make it all work for you?
I've had it and got those errors - perhaps that has to be 
set some place else.


In case anybody was in this situation - I was missing one 
important bit: _bindir_
Ubuntu's pgSQL binaries have different path - what 
resource/agent returns as errors is utterly confusing.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] PAF / PGSQLMS on Ubuntu

2023-09-07 Thread lejeczek via Users



On 07/09/2023 16:09, Andrei Borzenkov wrote:

On Thu, Sep 7, 2023 at 5:01 PM lejeczek via Users  wrote:

Hi guys.

I'm trying to set ocf_heartbeat_pgsqlms agent but I get:
...
Failed Resource Actions:
   * PGSQL-PAF-5433 stop on ubusrv3 returned 'invalid parameter' because 'Parameter 
"recovery_target_timeline" MUST be set to 'latest'. It is currently set to ''' 
at Thu Sep  7 13:58:06 2023 after 54ms

I'm new to Ubuntu and I see that Ubuntu has a bit different approach to paths 
(in comparison to how Centos do it).
I see separation between config & data, eg.

14  paf 5433 down   postgres /var/lib/postgresql/14/paf 
/var/log/postgresql/postgresql-14-paf.log

I create the resource like here:

-> $ pcs resource create PGSQL-PAF-5433 ocf:heartbeat:pgsqlms pgport=5433 bindir=/usr/bin 
pgdata=/etc/postgresql/14/paf datadir=/var/lib/postgresql/14/paf meta failure-timeout=30s master-max=1 
op start timeout=60s op stop timeout=60s op promote timeout=30s op demote timeout=120s op monitor 
interval=15s timeout=10s role="Promoted" op monitor interval=16s timeout=10s 
role="Unpromoted" op notify timeout=60s promotable notify=true failure-timeout=30s 
master-max=1 --disable

Ubuntu 22.04.3 LTS
What am I missing can you tell?

Exactly what the message tells you. You need to set recovery_target=latest.

and having it in 'postgresql.conf' make it all work for you?
I've had it and got those errors - perhaps that has to be 
set some place else.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] PAF / PGSQLMS on Ubuntu

2023-09-07 Thread lejeczek via Users

Hi guys.

I'm trying to set ocf_heartbeat_pgsqlms agent but I get:
...
Failed Resource Actions:
  * PGSQL-PAF-5433 stop on ubusrv3 returned 'invalid 
parameter' because 'Parameter "recovery_target_timeline" 
MUST be set to 'latest'. It is currently set to ''' at Thu 
Sep  7 13:58:06 2023 after 54ms


I'm new to Ubuntu and I see that Ubuntu has a bit different 
approach to paths (in comparison to how Centos do it).

I see separation between config & data, eg.

14  paf 5433 down   postgres /var/lib/postgresql/14/paf 
/var/log/postgresql/postgresql-14-paf.log


I create the resource like here:

-> $ pcs resource create PGSQL-PAF-5433 
ocf:heartbeat:pgsqlms pgport=5433 bindir=/usr/bin 
pgdata=/etc/postgresql/14/paf 
datadir=/var/lib/postgresql/14/paf meta failure-timeout=30s 
master-max=1 op start timeout=60s op stop timeout=60s op 
promote timeout=30s op demote timeout=120s op monitor 
interval=15s timeout=10s role="Promoted" op monitor 
interval=16s timeout=10s role="Unpromoted" op notify 
timeout=60s promotable notify=true failure-timeout=30s 
master-max=1 --disable


Ubuntu 22.04.3 LTS
What am I missing can you tell?
many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] last man - issue

2023-07-22 Thread lejeczek via Users

Hi guys.

That below should work, right?

-> $ pcs quorum update last_man_standing=1 --skip-offline
Checking corosync is not running on nodes...
Warning: Unable to connect to dzien (Failed to connect to 
dzien port 2224: No route to host)

Warning: dzien: Unable to check if corosync is not running
Error: swir: corosync is running
Error: whale: corosync is running
Error: Errors have occurred, therefore pcs is unable to continue

if it should, why does it not or what else am I missing?

many thanks, L
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: SATA drive rescan?

2023-07-18 Thread lejeczek via users



On 17/07/2023 19:57, ToddAndMargo via users wrote:

On 7/17/23 01:06, lejeczek via users wrote:



On 16/07/2023 08:27, ToddAndMargo via users wrote:

Hi All,

I have a removable drive sleeve that I mount SATA
drives mounted in that attaches to an internal SATA
port.  I use it backup my system to removable
drives.  The mechanism works identical to USB
flash drives.

With one exception.  A drive MUST be in the
carriage when I boot up.  Otherwise, if I insert
a drive AFTER the system is booted up, the drive
is invisible to the system.

Now I have seen Clonzilla re-educating the system
as to what drives are where.  Is there a way to
goose Fedora to do the same?  Is there
a way to rescan without having to reboot?

Many thanks,
-T

Bit better, clearer description of the 
hardware/caddie/connection you should have started with.


I thought
   "attaches to an internal SATA port"
covered that.

https://www.supermicro.com/en/products/motherboard/X11SAE-M

It is attached to "I-SATA1"

So, it is purely all SATA (& not! e-sata) and no bridge 
of any kind in-between
If so, if purely SATA, I'd suggest you check - if have 
not done so yet-    your mobo's BIOS for something like 
SATA devices "hot-plugging"


Hot plug is enabled and is the default.
___

And to clarify - CloneZilla does the trick when boot up 
independently of Fedora, as a live-iso, yes?
If so then it rather is - as you presumed I guess - 
OS/software end.

Then I'd presume that your BIOS has:
- AHCI and not RAID
- ports Spin Up Device is "Enabled"
- also there is/should be "SATA Device Type"

I'd play - imagine you did too - with: 
'signal-first-then-power' VS 'power-first-then-signal'


Other than that Fedora, Centos. which I mostly use which is 
older, normally do not need much - I always 
try not to forget to "enable" that in BIOS -> different 
vendor might have different nomenclature for it: hot- plug 
OR swap


I've only glanced at your 'poking' re/scan, I think it's 
rather the host/controller you want to poke.

If you have not tried already:

echo "- - -" > /sys/class/scsi_host/host0/scan

(& port/connector might want to give/chose different)___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] newly created clone waits for off-lined node

2023-07-17 Thread lejeczek via Users




On 13/07/2023 17:33, Ken Gaillot wrote:

On Wed, 2023-07-12 at 21:08 +0200, lejeczek via Users wrote:

Hi guys.

I have a fresh new 'galera' clone and that one would not start &
cluster says:
...
INFO: Waiting on node  to report database status before Master
instances can start.
...

Is that only for newly created resources - which I guess it must be -
and if so then why?
Naturally, next question would be - how to make such resource start
in that very circumstance?

many thank, L.

That is part of the agent rather than Pacemaker. Looking at the agent
code, it's based on a node attribute the agent sets, so it is only
empty for newly created resources that haven't yet run on a node. I'm
not sure if there's a way around it. (Anyone else have experience with
that?)

any expert/devel would agree - this qualifies as a bug?
To add to my last message - mariadb starts a okey outside of 
the cluster but also if I manually change 
'safe_to_bootstrap' to '1' on any of the - in my case - two 
nodes, then cluster will also start the clone ok.
But then... disable(seems that Galera cluster gets shut down 
ok) & enable the clone and clone fails to start with those 
errors as earlier - waiting for off-lined node - and leaving 
the clone as 'Unpromoted'

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: SATA drive rescan?

2023-07-17 Thread lejeczek via users



On 16/07/2023 08:27, ToddAndMargo via users wrote:

Hi All,

I have a removable drive sleeve that I mount SATA
drives mounted in that attaches to an internal SATA
port.  I use it backup my system to removable
drives.  The mechanism works identical to USB
flash drives.

With one exception.  A drive MUST be in the
carriage when I boot up.  Otherwise, if I insert
a drive AFTER the system is booted up, the drive
is invisible to the system.

Now I have seen Clonzilla re-educating the system
as to what drives are where.  Is there a way to
goose Fedora to do the same?  Is there
a way to rescan without having to reboot?

Many thanks,
-T

Bit better, clearer description of the 
hardware/caddie/connection you should have started with.
So, it is purely all SATA (& not! e-sata) and no bridge of 
any kind in-between?
If so, if purely SATA, I'd suggest you check - if have not 
done so yet-   your mobo's BIOS for something like SATA 
devices "hot-plugging"

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] newly created clone waits for off-lined node

2023-07-13 Thread lejeczek via Users




On 13/07/2023 17:33, Ken Gaillot wrote:

On Wed, 2023-07-12 at 21:08 +0200, lejeczek via Users wrote:

Hi guys.

I have a fresh new 'galera' clone and that one would not start &
cluster says:
...
INFO: Waiting on node  to report database status before Master
instances can start.
...

Is that only for newly created resources - which I guess it must be -
and if so then why?
Naturally, next question would be - how to make such resource start
in that very circumstance?

many thank, L.

That is part of the agent rather than Pacemaker. Looking at the agent
code, it's based on a node attribute the agent sets, so it is only
empty for newly created resources that haven't yet run on a node. I'm
not sure if there's a way around it. (Anyone else have experience with
that?)
But this is bit weird -- even with 'promoted-max' & 
'clone-max' to exclude (& on top of it ban the resource on 
the) off-lined node, clone would not start -- no?
Perhaps if not a an obvious bug, I'd think this could/should 
be improved, the code.
If anybody wanted to reproduce - have a "regular" galera 
clone okey, off-line a node, remove re/create the clone.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] newly created clone waits for off-lined node

2023-07-12 Thread lejeczek via Users

Hi guys.

I have a fresh new 'galera' clone and that one would not 
start & cluster says:

...
INFO: Waiting on node  to report database status 
before Master instances can start.

...

Is that only for newly created resources - which I guess it 
must be - and if so then why?
Naturally, next question would be - how to make such 
resource start in that very circumstance?


many thank, L.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] location constraint does not move promoted resource ?

2023-07-03 Thread lejeczek via Users



On 03/07/2023 18:55, Andrei Borzenkov wrote:

On 03.07.2023 19:39, Ken Gaillot wrote:

On Mon, 2023-07-03 at 19:22 +0300, Andrei Borzenkov wrote:

On 03.07.2023 18:07, Ken Gaillot wrote:
On Mon, 2023-07-03 at 12:20 +0200, lejeczek via Users 
wrote:

On 03/07/2023 11:16, Andrei Borzenkov wrote:

On 03.07.2023 12:05, lejeczek via Users wrote:

Hi guys.

I have pgsql with I constrain like so:

-> $ pcs constraint location PGSQL-clone rule 
role=Promoted

score=-1000 gateway-link ne 1

and I have a few more location constraints with that
ethmonitor & those work, but this one does not seem to.
When contraint is created cluster is silent, no 
errors nor

warning, but relocation does not take place.
I can move promoted resource manually just fine, to 
that

node where 'location' should move it.



Instance to promote is selected according to promotion
scores which are normally set by resource agent.
Documentation implies that standard location constraints
are also taken in account, but there is no 
explanation how
promotion scores interoperate with location scores. 
It is
possible that promotion score in this case takes 
precedence.

It seems to have kicked in with score=-1 but..
that was me just guessing.
Indeed it would be great to know how those are 
calculated,

in a way which would' be admin friendly or just obvious.

thanks, L.


It's a longstanding goal to have some sort of tool for 
explaining

how
scores interact in a given situation. However it's a 
challenging

problem and there's never enough time ...

Basically, all scores are added together for each node, 
and the

node
with the highest score runs the resource, subject to 
any placement
strategy configured. These mainly include stickiness, 
location
constraints, colocation constraints, and node health. 
Nodes may be


And you omitted the promotion scores which was the main 
question.


Oh right -- first, the above is used to determine the 
nodes on which
clone instances will be placed. After that, an 
appropriate number of
nodes are selected for the promoted role, based on 
promotion scores and

location and colocation constraints for the promoted role.



I am sorry but it does not really explain anything. Let's 
try concrete examples


a) master clone instance has location score -1000 for a 
node and promotion score 1000. Is this node eligible for 
promoting clone instance (assuming no other scores are 
present)?


b) promotion score is equal on two nodes A and B, but node 
A has better location score than node B. Is it guaranteed 
that clone will be promoted on A?




a real-life example:
...Colocation Constraints:
  Started resource 'HA-10-1-1-253' with Promoted resource 
'PGSQL-clone' (id:

  colocation-HA-10-1-1-253-PGSQL-clone-INFINITY)
    score=INFINITY
..
Order Constraints:
  promote resource 'PGSQL-clone' then start resource 
'HA-10-1-1-253' (id: order-

  PGSQL-clone-HA-10-1-1-253-Mandatory)
    symmetrical=0 kind=Mandatory
  demote resource 'PGSQL-clone' then stop resource 
'HA-10-1-1-253' (id: order-

  PGSQL-clone-HA-10-1-1-253-Mandatory-1)
    symmetrical=0 kind=Mandatory

I had to bump this one up to:
...
  resource 'PGSQL-clone' (id: location-PGSQL-clone)
    Rules:
  Rule: role=Promoted score=-1 (id: 
location-PGSQL-clone-rule)
    Expression: gateway-link ne 1 (id: 
location-PGSQL-clone-rule-expr)



'-1000' did not seem to be good enough, '-1' was just a 
"lucky" guess.


as earlier: I was able to 'move' the promoted and I think 
'prefers' also worked.
I don't know it 'pgsql' would work with any other 
constraints, if it was safe to try so.


many thanks, L.


___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [Freeipa-users] FreeIPA on Fedora and dnf system-upgrade

2023-07-03 Thread lejeczek via users



On 03/07/2023 15:05, Ian Pilcher wrote:

On 7/3/23 07:08, lejeczek via users wrote:

I'm not sure what you mean by 'in-place' do not work?


You cannot update an EL 7 FreeIPA server to EL 8, nor can 
you update
from EL 8 to EL 9.  The only supported path is to add a 
new replica,
running the new version, and then retire the old server.  
(If you want
the server to keep the same hostname, you then need to 
repeat the

process.)


Ah, sure, that will probably never change for Centos.
I think more adequate is, as they call it, 'rolling/on' 
release or point-release-update, rather then in-place 
updates, which was a bit confusing.
I'd suggest containers as a way out if it, perhaps VMs if 
can afford resource-wise.
Sure containers are also subject to generally understood 
"up-grade/date" but might be easier to manage in this & 
migrate to/over respects - I'm inclined to claim.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [Freeipa-users] FreeIPA on Fedora and dnf system-upgrade

2023-07-03 Thread lejeczek via users



On 29/06/2023 22:32, Ian Pilcher via FreeIPA-users wrote:
I am currently running FreeIPA on CentOS 7, and I am 
considering moving

it to Fedora.

On RHEL and derivatives, in-place upgrades are not 
supported.  It is
necessary to provision a new server, running the new OS 
version, add it

as a FreeIPA replica, and then decommission the old system.

How does this work on Fedora?  Will I be able to use dnf 
system-upgrade,
or will I find myself having to use the process described 
above?



I'm not sure what you mean by 'in-place' do not work?
My c7 days are quite a way behind but I don't remember it 
did no work.
c8s & c9s I've been doing up-dates/grades in place for a 
long time and sure I might have seen some glitches - but by 
intent packages are ready to upgrade already deployed IPA - 
but that goes for any software.
Then, usual best-practices - however you worked out those - 
such as backups, apply, how you control & order IPA packages 
upgrades, etc.
IPA has own 'ipa-server-upgrade' which is part of 
packages/version upgrade process(es)
Fedora's IPA stack works okey as it does on Centos - which 
Centos is Fedora pretty much, only back-dated/ported at it's 
core.
Centos 9 which I've been riding for long time is great, been 
updating IPA since c9s release - in fact I've recently, 
though for different reasons, migrated IPA into containers 
which are Fedoras - no problems.


regards, L.
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] location constraint does not move promoted resource ?

2023-07-03 Thread lejeczek via Users




On 03/07/2023 11:16, Andrei Borzenkov wrote:

On 03.07.2023 12:05, lejeczek via Users wrote:

Hi guys.

I have pgsql with I constrain like so:

-> $ pcs constraint location PGSQL-clone rule role=Promoted
score=-1000 gateway-link ne 1

and I have a few more location constraints with that
ethmonitor & those work, but this one does not seem to.
When contraint is created cluster is silent, no errors nor
warning, but relocation does not take place.
I can move promoted resource manually just fine, to that
node where 'location' should move it.



Instance to promote is selected according to promotion 
scores which are normally set by resource agent. 
Documentation implies that standard location constraints 
are also taken in account, but there is no explanation how 
promotion scores interoperate with location scores. It is 
possible that promotion score in this case takes precedence.

It seems to have kicked in with score=-1 but..
that was me just guessing.
Indeed it would be great to know how those are calculated, 
in a way which would' be admin friendly or just obvious.


thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] location constraint does not move promoted resource ?

2023-07-03 Thread lejeczek via Users

Hi guys.

I have pgsql with I constrain like so:

-> $ pcs constraint location PGSQL-clone rule role=Promoted 
score=-1000 gateway-link ne 1


and I have a few more location constraints with that 
ethmonitor & those work, but this one does not seem to.
When contraint is created cluster is silent, no errors nor 
warning, but relocation does not take place.
I can move promoted resource manually just fine, to that 
node where 'location' should move it.


All thoughts share are much appreciated.
many thanks, L.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: SOLUTION: Re: journal / systemd colours ?

2023-06-29 Thread lejeczek via users



On 28/06/2023 20:26, stan via users wrote:

On Wed, 28 Jun 2023 11:10:05 -0700
stan  wrote:


On Wed, 28 Jun 2023 10:17:45 -0700
stan via users  wrote:


On Wed, 28 Jun 2023 17:08:38 +0100
Patrick O'Callaghan  wrote:

See also 'man journalctl' and search for PAGER.

This is also relevant.  Unfortunately, it seems that journalctl
ignores the input of SYSTEMD_LESS.  It is hard to tell, because
there is no example of the kind of input it wants or whether it
only accepts options without arguments, as the default options all
are.  I would probably have to look at the journalctl code to
determine if they are hard coded into the program.  There is no
configuration file to set them in any case.

So, the code for journalctl shows it reading the options from the
variables, and invoking less with those options, but it doesn't seem
to affect the behavior of running journalctl, or at least the colors
don't change.  Maybe I'll look more closely at some point, but no joy
for now.

Operator error.  Exporting the wrong name because of a cut and paste.
Once I fixed that, definitely works to change colors in journalctl
output, will have to tune it to get what I want.
I put this in my .bashrc so everything is set on login.
SYSTEMD_PAGER=less
export SYSTEMD_PAGER
SYSTEMD_LESS="[list of less options]"
export SYSTEMD_LESS
___
Not exactly, I'd not think of it as _the_ solution - (I much 
prefer to up pages via actual mechanical scrolling) - 
as I use, always I've had, SYSTEMD_PAGER=cat so...


man page for 'journalctl' has a shor section:
When outputting to a tty, lines are colored according to 
priority: lines of level ERROR and higher are colored red;
   lines of level NOTICE and higher are highlighted; 
lines of level DEBUG are colored lighter grey; other lines are

   displayed normally.

would be nice to be able to customize those & if 'systemd' 
delegates declaration of that 'highlighting' colour then 
these below do not do it:
a) terminal-colors.d - perhaps systemd/journal ignores it 
altogether

b) gnome-terminal has config for 'Highlight colour'

It would be great/the best to have that functionality 
internal to systemd/journal - thus, if authors/devel might 
read here - please think of this conversation's subject as 
possible future addition/enhancement to the software.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


journal / systemd colours ?

2023-06-28 Thread lejeczek via users

Hi guys.

I wonder if somebody might know how to customize, to change 
systemd journal's colours?
I'm specifically thinking of changing 'WARN' warnings which 
get - in my opinion unfortunate, for I use other than white 
background colours - painted in yellow.


many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ClusterLabs] silence resource ? - PGSQL

2023-06-28 Thread lejeczek via Users

Hi guys.

Having 'pgsql' set up in what I'd say is a vanilla-default 
confg, pacemaker's journal log is flooded with:

...
pam_unix(runuser:session): session closed for user postgres
pam_unix(runuser:session): session opened for user 
postgres(uid=26) by (uid=0)

pam_unix(runuser:session): session closed for user postgres
pam_unix(runuser:session): session opened for user 
postgres(uid=26) by (uid=0)

pam_unix(runuser:session): session closed for user postgres
pam_unix(runuser:session): session opened for user 
postgres(uid=26) by (uid=0)

pam_unix(runuser:session): session closed for user postgres
...

Would you have a working fix or even a suggestion on how to 
silence those?


many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Thunderbird adaptive junk control - does it work?

2023-06-22 Thread lejeczek via users

Hi guys.

I have - as you can probably see - a Yahoo account and 
there, on the server I have a all my filters/rules set and 
those work, say..
another mailing messages list a filter/rule puts meails 
where I told it to, to a folder, I can confirm that when I 
go to Yahoo's web mail, but...
Thunderbird keeps putting those message into "Bulk"(spam) 
folder even though I told Thunderbird not to - by marking 
message as 'not spam', a number of times already.
I also have "Trust junk mail headers" de-activated, so it 
should be entirely on Thunderbird's internal business logic 
to spam counting.
I don't think all this works in Thunderbird or if it does 
then it does it extremely poorly.
What is your experience I wonder & I wonder if there is 
something, some setting/tweaking I missed which can help?
(I want to avoid putting addresses into my adrress book only 
for the reason of helping spam filtering)


many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: Gnome Nautilus and no thumbnails from connected Android storage

2023-06-11 Thread lejeczek via users



On 21/10/2021 11:39, lejeczek via users wrote:

Hi guys.

I asked this a while ago but nobody got interested yet I'm 
really curious - when you connect your Android Phone via 
USB, do you get thumbnails?
I cannot get my F34's Nautilus to show thumbnails off my 
Huawei Honor 10 with Android 10.


thanks, L.

We have thumbnails back in f38 - at least on my Oppo have 
them - so manyyy  thanks to those responsible for fixing 
this.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?

2023-06-09 Thread lejeczek via Users



On 09/06/2023 09:04, Reid Wahl wrote:

On Thu, Jun 8, 2023 at 10:55 PM lejeczek via Users
 wrote:



On 09/06/2023 01:38, Reid Wahl wrote:

On Thu, Jun 8, 2023 at 2:24 PM lejeczek via Users  wrote:



Ouch.

Let's see the full output of the move command, with the whole CIB that
failed to validate.


For a while there I thought perhaps it was just that one
pglsq resource, but it seems that any - though only a few
are set up - (only clone promoted?)resource fails to move.
Perhaps primarily to do with 'pcs'

-> $ pcs resource move REDIS-clone --promoted podnode3
Error: cannot move resource 'REDIS-clone'
  1 

This is the problem: `validate-with="pacemaker-3.6"`. That old schema
doesn't support role="Promoted" in a location constraint. Support
begins with version 3.7 of the schema:
https://github.com/ClusterLabs/pacemaker/commit/e7f1424df49ac41b2d38b72af5ff9ad5121432d2.

You'll need at least Pacemaker 2.1.0.

I have:
corosynclib-3.1.7-1.el9.x86_64
corosync-3.1.7-1.el9.x86_64
pacemaker-schemas-2.1.6-2.el9.noarch
pacemaker-libs-2.1.6-2.el9.x86_64
pacemaker-cluster-libs-2.1.6-2.el9.x86_64
pacemaker-cli-2.1.6-2.el9.x86_64
pacemaker-2.1.6-2.el9.x86_64
pcs-0.11.5-2.el9.x86_64
pacemaker-remote-2.1.6-2.el9.x86_64
and the reset is Centos 9 up-to-what-is-in-repos

If all your cluster nodes are at those versions, try `pcs cluster cib-upgrade`


'cib-upgrade' might be a fix for this "issue" - but should 
it not happen at rpm update time?

Differently but still 'move' errored out:
-> $ pcs resource move REDIS-clone --promoted podnode2
Location constraint to move resource 'REDIS-clone' has been 
created

Waiting for the cluster to apply configuration changes...
Location constraint created to move resource 'REDIS-clone' 
has been removed

Waiting for the cluster to apply configuration changes...
Error: resource 'REDIS-clone' is promoted on node 
'podnode3'; unpromoted on nodes 'podnode1', 'podnode2'

Error: Errors have occurred, therefore pcs is unable to continue

Then a node-promted-resource going into 'standby' or being 
rebooted - for first/one time - seemed to "fix" 'move'.

-> $ pcs resource move REDIS-clone --promoted podnode2
Location constraint to move resource 'REDIS-clone' has been 
created

Waiting for the cluster to apply configuration changes...
Location constraint created to move resource 'REDIS-clone' 
has been removed

Waiting for the cluster to apply configuration changes...
resource 'REDIS-clone' is promoted on node 'podnode2'; 
unpromoted on nodes 'podnode1', 'podnode3'


quite puzzling, sharing in case others experience this/similar.
thanks, L.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?

2023-06-08 Thread lejeczek via Users



On 09/06/2023 01:38, Reid Wahl wrote:

On Thu, Jun 8, 2023 at 2:24 PM lejeczek via Users  wrote:




Ouch.

Let's see the full output of the move command, with the whole CIB that
failed to validate.


For a while there I thought perhaps it was just that one
pglsq resource, but it seems that any - though only a few
are set up - (only clone promoted?)resource fails to move.
Perhaps primarily to do with 'pcs'

-> $ pcs resource move REDIS-clone --promoted podnode3
Error: cannot move resource 'REDIS-clone'
 1 

This is the problem: `validate-with="pacemaker-3.6"`. That old schema
doesn't support role="Promoted" in a location constraint. Support
begins with version 3.7 of the schema:
https://github.com/ClusterLabs/pacemaker/commit/e7f1424df49ac41b2d38b72af5ff9ad5121432d2.

You'll need at least Pacemaker 2.1.0.

I have:
corosynclib-3.1.7-1.el9.x86_64
corosync-3.1.7-1.el9.x86_64
pacemaker-schemas-2.1.6-2.el9.noarch
pacemaker-libs-2.1.6-2.el9.x86_64
pacemaker-cluster-libs-2.1.6-2.el9.x86_64
pacemaker-cli-2.1.6-2.el9.x86_64
pacemaker-2.1.6-2.el9.x86_64
pcs-0.11.5-2.el9.x86_64
pacemaker-remote-2.1.6-2.el9.x86_64
and the reset is Centos 9 up-to-what-is-in-repos



 2   
 3 
 4   
 5 
 6 
 7 
 8 
 9 
10 
11 
12   
13   
14 
15   
16 
17 
18   
19 
20   
21 
22   
23   
24 
25   
26 
27   
28   
29 
30   
31 
32   
33 
34 
35   
36 
37   
38 
39 
40   
41   
42 
43 
44   
45   
46   
47 
48   
49   
50 
51   
52 
53 
54   
55   
56 
57 
58   
59   
60   
61 
62   
63   
64 
65   
66 
67 
68 
69 
70 
71 
72   
73   
74 
75 
76   
77 
78   
79 
80 
81   
82 
83   
84 
85 
86   
87 
88   
89 
90 
91 
92 
93 
94   
95 
96 
97   
98   
99   
   100   
   101 
   102   
   103   
   104 
   105   
   106 
   107 
   108 
   109 
   110 
   111 
   112 
   113 
   114   
   115   
   116 
   117   
   118   
   119 
   120 
   121   
   122 
   123   
   124 
   125 
   126   
   127 
   128   
   129 
   130 
   131   
   132 
   133   
   134 
   135 
   136 
   137 
   138   
   139 
   140 
   141   
   142   
   143 
   144   
   145   
   146 
   147   
   148 
   149 
   150 
   151 
   152 
   153 
   154 
   155 
   156 
   157 
   158 
   159   
   160   
   161 
   162 
   163 
   164 
   165 
   166 
   167 
   168 
   169   
   170 
   171 
   172   
   173   
   174   
   175   
   176 
   177   
   178 
   179 
   180   
   181   
   182   
   183   
   184   
   185 
   186 
   187   
   188 
   189   
   190 
   191   
   192   
   193 
   194   
   195 
   196   
   197   
   198   
   199 
   200   
   201   
   202 
   203   
   204 
   205   
   206   
   207 
   208   
   209   
   210 
   211 
   212 
   213   
   214   
   215 
   216   
 

Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?

2023-06-08 Thread lejeczek via Users




Ouch.

Let's see the full output of the move command, with the whole CIB that
failed to validate.

For a while there I thought perhaps it was just that one 
pglsq resource, but it seems that any - though only a few 
are set up - (only clone promoted?)resource fails to move.

Perhaps primarily to do with 'pcs'

-> $ pcs resource move REDIS-clone --promoted podnode3
Error: cannot move resource 'REDIS-clone'
   1 validate-with="pacemaker-3.6" epoch="8212" num_updates="0" 
admin_epoch="0" cib-last-written="Thu Jun  8 21:59:53 2023" 
update-origin="podnode1" update-client="crm_attribute" 
have-quorum="1" update-user="root" dc-uuid="1">

   2   
   3 
   4   
   5 id="cib-bootstrap-options-have-watchdog" 
name="have-watchdog" value="false"/>
   6 name="dc-version" value="2.1.6-2.el9-6fdc9deea29"/>
   7 id="cib-bootstrap-options-cluster-infrastructure" 
name="cluster-infrastructure" value="corosync"/>
   8 id="cib-bootstrap-options-cluster-name" name="cluster-name" 
value="podnodes"/>
   9 id="cib-bootstrap-options-stonith-enabled" 
name="stonith-enabled" value="false"/>
  10 id="cib-bootstrap-options-last-lrm-refresh" 
name="last-lrm-refresh" value="1686047745"/>
  11 id="cib-bootstrap-options-maintenance-mode" 
name="maintenance-mode" value="false"/>

  12   
  13   
  14 name="REDIS_REPL_INFO" value="podnode1"/>

  15   
  16 
  17 
  18   
  19 
  20   name="PGSQL-data-status" value="DISCONNECT"/>

  21 
  22   
  23   
  24 
  25   name="PGSQL-data-status" value="DISCONNECT"/>

  26 
  27   
  28   
  29 
  30   name="PGSQL-data-status" value="LATEST"/>

  31 
  32   
  33 
  34 
  35   provider="heartbeat" type="IPaddr2">
  36 id="HA-10-1-1-226-meta_attributes">
  37   id="HA-10-1-1-226-meta_attributes-failure-timeout" 
name="failure-timeout" value="30s"/>

  38 
  39 id="HA-10-1-1-226-instance_attributes">
  40   id="HA-10-1-1-226-instance_attributes-cidr_netmask" 
name="cidr_netmask" value="24"/>
  41   id="HA-10-1-1-226-instance_attributes-ip" name="ip" 
value="10.1.1.226"/>

  42 
  43 
  44   interval="10s" name="monitor" timeout="20s"/>
  45   interval="0s" name="start" timeout="20s"/>
  46   interval="0s" name="stop" timeout="20s"/>

  47 
  48   
  49   provider="heartbeat" type="IPaddr2">
  50 id="HA-10-3-1-226-meta_attributes">
  51   id="HA-10-3-1-226-meta_attributes-failure-timeout" 
name="failure-timeout" value="30s"/>

  52 
  53 id="HA-10-3-1-226-instance_attributes">
  54   id="HA-10-3-1-226-instance_attributes-cidr_netmask" 
name="cidr_netmask" value="24"/>
  55   id="HA-10-3-1-226-instance_attributes-ip" name="ip" 
value="10.3.1.226"/>

  56 
  57 
  58   interval="10s" name="monitor" timeout="20s"/>
  59   interval="0s" name="start" timeout="20s"/>
  60   interval="0s" name="stop" timeout="20s"/>

  61 
  62   
  63   
  64 provider="heartbeat">
  65   id="REDIS-instance_attributes">
  66 name="bin" value="/usr/bin/redis-server"/>
  67 id="REDIS-instance_attributes-client_bin" name="client_bin" 
value="/usr/bin/redis-cli"/>
  68 id="REDIS-instance_attributes-config" name="config" 
value="/etc/redis/redis.conf"/>
  69 id="REDIS-instance_attributes-rundir" name="rundir" 
value="/run/redis"/>
  70 id="REDIS-instance_attributes-user" name="user" value="redis"/>
  71 id="REDIS-instance_attributes-wait_last_known_master" 
name="wait_last_known_master" value="true"/>

  72   
  73   
  74 timeout="120s" id="REDIS-demote-interval-0s"/>
  75 interval="45s" id="REDIS-monitor-interval-45s">
  76   id="REDIS-monitor-interval-45s-instance_attributes">
  77 id="REDIS-monitor-interval-45s-instance_attributes-OCF_CHECK_LEVEL" 
name="OCF_CHECK_LEVEL" value="0"/>

  78   
  79 
  80 timeout="60s" interval="20s" id="REDIS-monitor-interval-20s">
  81   id="REDIS-monitor-interval-20s-instance_attributes">
  82 id="REDIS-monitor-interval-20s-instance_attributes-OCF_CHECK_LEVEL" 
name="OCF_CHECK_LEVEL" value="0"/>

  83   
  84 
  85 timeout="60s" interval="60s" id="REDIS-monitor-interval-60s">
  86   id="REDIS-monitor-interval-60s-instance_attributes">
  87 id="REDIS-monitor-interval-60s-instance_attributes-OCF_CHECK_LEVEL" 
name="OCF_CHECK_LEVEL" value="0"/>

  88   
  89 
  90 timeout="90s" id="REDIS-notify-interval-0s"/>
  91 timeout="120s" id="REDIS-promote-interval-0s"/>
  

Re: gimmicks for most & a question networking curious/experts - ? MACSEC

2023-06-07 Thread lejeczek via users

de

On 06/06/2023 14:50, Chris Adams wrote:

Once upon a time, lejeczek  said:

To start with the most basic of what I ponder over - basic in my
mind though I get, that it does not have to be that or
simple/obvious - is MACSEC with 'bond' as parent or in other words:
macsec "on" a 'bond' network interface.
Should such a "thing" work, does it work?

My understanding is that you do MACsec on physical interfaces, so you
would configure it on each member of a LAG, not the virtual LAG
interface (e.g. bond0) itself.

So I started from the bottom, went to do 'macsec' physical 
eth interfaces but when under 'bond' in "broadcast" I got NM 
failing to set such macsec ifaces up:

...
  [1686137957.1794] device (macsec-ten-bott): 
supplicant interface state: disconnected -> completed
  [1686137957.1829] device (macsec-ten-bott): carrier: 
link connected
  [1686137957.1829] device (macsec-ten-bott): 
Activation: Stage 2 of 5 (Device Configure) successful.
  [1686137957.2081] device (macsec-ten-bott): state 
change: config -> ip-config (reason 'none', sys-iface-state: 
'managed')
  [1686137957.2149] device (macsec-ten-bott): 
Activation: connection 'macsec-10.1.1-slave-bott' could not 
be enslaved
  [1686137957.2150] device (macsec-ten-bott): state 
change: ip-config -> failed (reason 'unknown', 
sys-iface-state: 'managed')
  [1686137957.2152] device (macsec-ten-bott): released 
from master device bond-1011
  [1686137957.2153] device (macsec-ten-bott): 
Activation: failed for connection 'macsec-10.1.1-slave-bott'
  [1686137957.2156] device (macsec-ten-bott): state 
change: failed -> disconnected (reason 'none', 
sys-iface-state: 'managed')
  [1686137957.2402] device (macsec-ten-bott): state 
change: disconnected -> unmanaged (reason 'user-requested', 
sys-iface-state: 'managed')


And to share what I find - obvious to those with expertise - 
that 'macsec' "on" a bond iface (broadcast mode is my case!) 
works, or seems to work I shall add as it's less than 1hr of 
my fiddling.


But I'd rather see physical MACs "macseced" and then these 
being bonded/bridged - as going to the lowest layer/device 
in a stack puts my mind at ease the best.

I should add - I do all that in Centos 9.

thanks, L.
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


gimmicks for most & a question networking curious/experts - ? MACSEC

2023-06-06 Thread lejeczek via users

Hi guys.

Looking at macsec I've only started - so go easy on me with 
my possibly trivial questions -  and write here in hope, 
that some of you have expertise to tell...


To start with the most basic of what I ponder over - basic 
in my mind though I get, that it does not have to be that or 
simple/obvious - is MACSEC with 'bond' as parent or in other 
words: macsec "on" a 'bond' network interface.

Should such a "thing" work, does it work?

All & any thoughts shared are much appreciated.
many thanks, L.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?

2023-06-06 Thread lejeczek via Users




On 05/06/2023 16:23, Ken Gaillot wrote:

On Sat, 2023-06-03 at 15:09 +0200, lejeczek via Users wrote:

Hi guys.

I've something which I'm new to entirely - cluster which is seemingly
okey errors, fails to move a resource.

What pcs version are you using? I believe there was a move regression
in a recent push.


I'd won't contaminate here just yet with long json cluster spits when
fails but a snippet:

-> $ pcs resource move PGSQL-clone --promoted podnode1
Error: cannot move resource 'PGSQL-clone'
1 
2   
3 
4   
5 
6 
7 
8 
9 
   10 
   11 
   12   
   13   
   14 
   15 
   16   
   17 
...
crm_resource: Error performing operation: Invalid configuration

This one line: (might be more)

puzzles me, as there is no such node/member in the cluster and so I
try:

That's not a problem. Pacemaker allows "custom" values in both cluster
options and resource/action meta-attributes. I don't know whether redis
is actually using that or not.


-> $ pcs property unset redis_REPL_INFO --force
Warning: Cannot remove property 'redis_REPL_INFO', it is not present
in property set 'cib-bootstrap-options'

That's because the custom options are in their own
cluster_property_set. I believe pcs can only manage the options in the
cluster_property_set with id="cib-bootstrap-options", so you'd have to
use "pcs cluster edit" or crm_attribute to remove the custom ones.


Any & all suggestions on how to fix this are much appreciated.
many thanks, L.

I've downgraded back to:
pacemaker-2.1.6-1.el9.x86_64
pcs-0.11.4-7.el9.x86_64
but, it's either not enough - if bugs are in those that is - 
or issues are somewhere else. for 'move' still fails the same:

...
crm_resource: Error performing operation: Invalid configuration


___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] cluster okey but errors when tried to move resource - ?

2023-06-05 Thread lejeczek via Users




On 05/06/2023 16:23, Ken Gaillot wrote:

On Sat, 2023-06-03 at 15:09 +0200, lejeczek via Users wrote:

Hi guys.

I've something which I'm new to entirely - cluster which is seemingly
okey errors, fails to move a resource.

What pcs version are you using? I believe there was a move regression
in a recent push.


I'd won't contaminate here just yet with long json cluster spits when
fails but a snippet:

-> $ pcs resource move PGSQL-clone --promoted podnode1
Error: cannot move resource 'PGSQL-clone'
1 
2   
3 
4   
5 
6 
7 
8 
9 
   10 
   11 
   12   
   13   
   14 
   15 
   16   
   17 
...
crm_resource: Error performing operation: Invalid configuration

This one line: (might be more)

puzzles me, as there is no such node/member in the cluster and so I
try:

That's not a problem. Pacemaker allows "custom" values in both cluster
options and resource/action meta-attributes. I don't know whether redis
is actually using that or not.


-> $ pcs property unset redis_REPL_INFO --force
Warning: Cannot remove property 'redis_REPL_INFO', it is not present
in property set 'cib-bootstrap-options'

That's because the custom options are in their own
cluster_property_set. I believe pcs can only manage the options in the
cluster_property_set with id="cib-bootstrap-options", so you'd have to
use "pcs cluster edit" or crm_attribute to remove the custom ones.


Any & all suggestions on how to fix this are much appreciated.
many thanks, L.
Well. that would make sense - those must have gone in & 
through to the repos' rpms - as I suspected thos broke after 
recent 'dnf' updates.
I'll fiddle with 'downgrades' later on - but if true, then 
rather critical, no?

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] cluster okey but errors when tried to move resource - ?

2023-06-03 Thread lejeczek via Users

Hi guys.

I've something which I'm new to entirely - cluster which is 
seemingly okey errors, fails to move a resource.
I'd won't contaminate here just yet with long json cluster 
spits when fails but a snippet:


-> $ pcs resource move PGSQL-clone --promoted podnode1
Error: cannot move resource 'PGSQL-clone'
   1 validate-with="pacemaker-3.6" epoch="8109" num_updates="0" 
admin_epoch="0" cib-last-written="Sat Jun  3 13:49:34 2023" 
update-origin="podnode2" update-client="cibadmin" 
have-quorum="1" update-user="root" dc-uuid="2">

   2   
   3 
   4   
   5 id="cib-bootstrap-options-have-watchdog" 
name="have-watchdog" value="false"/>
   6 name="dc-version" value="2.1.6-1.el9-802a72226be"/>
   7 id="cib-bootstrap-options-cluster-infrastructure" 
name="cluster-infrastructure" value="corosync"/>
   8 id="cib-bootstrap-options-cluster-name" name="cluster-name" 
value="podnodes"/>
   9 id="cib-bootstrap-options-stonith-enabled" 
name="stonith-enabled" value="false"/>
  10 id="cib-bootstrap-options-last-lrm-refresh" 
name="last-lrm-refresh" value="1683293193"/>
  11 id="cib-bootstrap-options-maintenance-mode" 
name="maintenance-mode" value="false"/>

  12   
  13   
  14 name="redis_REPL_INFO" value="c8kubernode1"/>
  15 name="REDIS_REPL_INFO" value="podnode3"/>

  16   
  17 
...
crm_resource: Error performing operation: Invalid configuration

This one line: (might be more)
name="redis_REPL_INFO" value="c8kubernode1"/>
puzzles me, as there is no such node/member in the cluster 
and so I try:


-> $ pcs property unset redis_REPL_INFO --force
Warning: Cannot remove property 'redis_REPL_INFO', it is not 
present in property set 'cib-bootstrap-options'


Any & all suggestions on how to fix this are much appreciated.
many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: PS1 not in .bash_profile but somewhere else ?

2023-06-02 Thread lejeczek via users



On 02/06/2023 13:56, lejeczek via users wrote:

Hi guys

I have one specific user whose PS1 if not set in 
.bash_profile, is taken from somewhere else.
At the same time globally I set PS1 in /etc/environment - 
bu that, this one user does not honor.

I've tried a few terminal/console apps - no difference.

Where else can PS1 be set and/or how to find out here PS1 
for a user/account was set?


many thanks, L.



A bit more fiddling..
I don't know what role in it gnome-terminal plays but, it 
seems that PS1 gets "global" at Gnome login time and user 
inherits that, if set in  '/etc/environment; , if PS1 set 
specifically in user "profile", then takes expected effect.
If set only in '/etc/environment' , PS1 is read by 
gnome/terminal but is malformed, as not digested, raw - 
suffices to 'ssh localhost' from the same gnome terminal and 
PS1 is set, displayed properly
In terminal, "Run command as login shell" is ON for all my 
fiddling and makes no difference.


It smells to me like a bug, I'd not presume it's intentional.

thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


PS1 not in .bash_profile but somewhere else ?

2023-06-02 Thread lejeczek via users

Hi guys

I have one specific user whose PS1 if not set in 
.bash_profile, is taken from somewhere else.
At the same time globally I set PS1 in /etc/environment - bu 
that, this one user does not honor.

I've tried a few terminal/console apps - no difference.

Where else can PS1 be set and/or how to find out here PS1 
for a user/account was set?


many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ovirt-users] troubleshooting SHE - Run engine-setup with answerfile - ?

2023-06-01 Thread lejeczek via Users

Hi guys.

I'm revisiting my failed attempt to make friends with oVirt 
- novice so go easy on me.
Deployment fails sooner that what can troubleshoot - at 
least I manged to find in docs so far - and I wanted to ask 
hot to troubleshot such failure as this:

...
 INFO  ] TASK [ovirt.ovirt.engine_setup : Run engine-setup 
with answerfile]
[ ERROR ] fatal: [localhost -> 192.168.1.209]: FAILED! => 
{"changed": true, "cmd": ["engine-setup", 
"--accept-defaults", 
"--config-append=/root/ovirt-engine-answers"], "delta": 
"0:08:05.997322", "end": "2023-06-01 13:07:15.932517", 
"msg": "non-zero return code", "rc": 1, "start": "2023-06-01 
12:59:09.935195", "stderr": "", "stderr_lines": [], 
"stdout": "[ INFO  ] Stage: Initializing\n[ INFO  ] Stage: 
Environment setup\n  Configuration files:\n

...
Start with setting up Keycloak for Ovirt Engine", "[ ERROR ] 
Failed to execute stage 'Closing up': Command 
'/usr/share/ovirt-engine-keycloak/bin/kk_cli.sh' failed to 
execute", "[ INFO  ] Stage: Clean up", "  Log file 
is located at", " 
/var/log/ovirt-engine/setup/ovirt-engine-setup-20230601125911-448xvj.log", 


...
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, 
"msg": "There was a failure deploying the engine on the 
local engine VM. The system may not be provisioned according 
to the playbook results: please check the logs for the 
issue, fix accordingly or re-deploy from scratch.\n"}
[ ERROR ] Failed to execute stage 'Closing up': Failed 
executing ansible-playbook

[ INFO  ] Stage: Clean up
...
  Log file is located at 
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20230601123237-rhwx8e.log



Logfile to which 'stdout' points, shows nothing or I fail to 
find anything in it and/or understand it:

...
2023-06-01 13:07:17,716+0100 ERROR 
otopi.ovirt_hosted_engine_setup.ansible_utils 
ansible_utils._process_output:113 fatal: [localhost -> 
192.168.1.209]: FAILED! => {"changed": true, "cmd": 
["engine-setup", "--accept-defaults", 
"--config-append=/root/ovirt-engine-answers"], "delta": 
"0:08:05.997322", "end": "2023-06-01 13:07:15.932517", 
"msg": "non-zero return code", "rc": 1, "start": "2023-06-01 
12:59:09.935195", "stderr": "", "stderr_lines": [], "stdout":

...
or Ovirt Engine\n[ ERROR ] Failed to execute stage 'Closing 
up': Command 
'/usr/share/ovirt-engine-keycloak/bin/kk_cli.sh' failed to 
execute\n[ INFO  ] Stage: Clean up\n  Log file is 
located at\n 
/var/log/ovirt-engine/setup/ovirt-engine-setup-20230601125911-448xvj.log\n[ 
INFO  ] Generating answer file 
'/var/lib/ovirt-engine/setup/answers/20230601130714-setup.conf'\n[ 
INFO  ] Stage: Pre-termination\n[ INFO  ] Stage: 
Termination\n[ ERROR ] Execution of setup failed", 
"stdout_lines": ["[ INFO  ] Stage: Initializing", "[ INFO  ] 
Stage: Environment setup", "  Configuration files:", 
" /etc/ovirt-engine-setup.conf.d/10-packaging-jboss.conf,", 
"  /etc/ovirt-engine-setup.conf.d/10-packaging.conf,"

...
...
2023-06-01 13:08:52,946+0100 DEBUG 
otopi.ovirt_hosted_engine_setup.ansible_utils 
ansible_utils._process_output:109 {'msg': "The tas
k includes an option with an undefined variable. The error 
was: 'local_vm_ip' is undefined. 'local_vm_ip' is 
undefined\n\nThe error
 appears to be in 
'/usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_engine_setup/tasks/sync_on_engine_ma
chine.yml': line 2, column 3, but may\nbe elsewhere in the 
file depending on the exact syntax problem.\n\nThe offending 
line appear
s to be:\n\n---\n- name: Set the name for add_host\n  ^ 
here\n", '_ansible_no_log': False}
2023-06-01 13:08:53,047+0100 DEBUG 
otopi.ovirt_hosted_engine_setup.ansible_utils 
ansible_utils._process_output:109 ignored: [localh
ost]: FAILED! => {"msg": "The task includes an option with 
an undefined variable. The error was: 'local_vm_ip' is 
undefined. 'local
_vm_ip' is undefined\n\nThe error appears to be in 
'/usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_eng
ine_setup/tasks/sync_on_engine_machine.yml': line 2, column 
3, but may\nbe elsewhere in the file depending on the exact 
syntax prob
lem.\n\nThe offending line appears to be:\n\n---\n- name: 
Set the name for add_host\n  ^ here\n"}

...
2023-06-01 13:09:16,818+0100 DEBUG 
otopi.ovirt_hosted_engine_setup.ansible_utils 
ansible_utils._process_output:109 {'changed': True, 
'stdout': '', 'stderr': "error: Failed to destroy pool 
9ab14cca-78e0-4a8e-8a1b-00e98c300208\nerror: Requested 
operation is not valid: storage pool 
'9ab14cca-78e0-4a8e-8a1b-00e98c300208' is not active", 'rc': 
1, 'cmd': ['virsh', '-c', 
'qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf', 
'pool-destroy', '9ab14cca-78e0-4a8e-8a1b-00e98c300208'], 
'start': '2023-06-01 13:09:16.614978', 'end': '2023-06-01 
13:09:16.672260', 'delta': '0:00:00.057282', 'msg': 
'non-zero return code', 'invocation': {'module_args': 
{'_raw_params': 'virsh -c 

switching windows behavior changed with F38 Gnome ?

2023-05-16 Thread lejeczek via users

Hi guys.

All this years I could switch between... say Firefox and 
Chromium, this way:


a) had a Chromium window focused-on
b) had a number of Firefox windows (as well as Chromum 
windows) in the background
c) I switched with Alt+Tab to Firefox window (whichever 
window was there first, does not matter)
d) could switch back - all the previous Fedora/Gnome 
versions - to that last Chromium window, quickly & easily 
with Alt+Esc
but now.. with an upgrade to F38, Atl+Esc goes through _all_ 
Firefox windows instead.


@fedorians - do you also see/get this change?
@devel, if read here - was this change planned, intentional? 
If so then why, what for?


many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Re: Firefox - detection of exiting tabs with URLs

2023-05-12 Thread lejeczek via users



On 10/05/2023 22:17, Samuel Sieb wrote:

On 5/10/23 12:47, Tim via users wrote:

On Wed, 2023-05-10 at 13:04 +0200, lejeczek via users wrote:
With f38 I think my Firefox is unable to detect or is 
ignoring tabs
which are already opened in other windows with certain 
URL, when I

open a new tab and want to go to a website/URL.
  I think Firefox would then say, would offer something 
like "switch

to tab".


In the past I'd noticed a behaviour that if I opened a 
blank tab and
typed in something like facebook, let auto-complete do 
its thing and
picked something from the drop-down list that appeared 
below the
address bar from its history, I'd often see the browser 
whiz over to an
existing tab which already had that site loaded.  It was 
that or a very

similar kind of situation.


Look carefully at the item you're clicking.  Watch for the 
tag that says "Switch to Tab".

___

Yes, that is exactly what we are saying, I was as the 
originator of this thread.
I can only add - to what I said which was "this 
functionality is completely gone" - that now after a longer 
look I see "Swtich to Tab" for some site/urls, sometimes! 
but not the others.
This would need to be thoroughly tested (probably something 
only devel could do) to say, if there is pattern to it at 
all or all this is purely random.
Eg. I've had a window with a few tabs - my Firefox is set to 
restore last sessions - _none_ of those tabs (sites/urls 
were fully loaded after restoration I made sure of) was 
detected as "Switch to Tab"
I closed that window, opened a new one (made sure that no 
other auto-started/restored window/tab had those sites in 
question/example opened) and each site, one by one 
loaded/opened, then tested in already existing window and 
only ! then that window picked up, detected and offered 
"Switch to Tab" to those sites.
So, purely from consumer perspective I can say this 
"feature" is bit flaky.


many thanks, L.
___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ovirt-users] Re: engine setup fails: error creating bridge interface virbr0: File exists - ?

2023-05-11 Thread lejeczek via Users



On 11/05/2023 18:18, Volenbovskyi, Konstantin wrote:


Hi,

Is it actually first attempt to install HE? Or maybe you 
ran into some (different) problem and this is like second 
(third, fourth…) attempt.


While installation script takes care of some cleanup, I 
would say that you should consider doing


ovirt-hosted-engine-cleanup

,checking that virbr0 is absent in ‘ip a’ output

and trying once again

BR,

Konstantin

*Von: *lejeczek via Users 
*Antworten an: *lejeczek 
*Datum: *Donnerstag, 11. Mai 2023 um 13:25
*An: *"users@ovirt.org" 
*Betreff: *[ovirt-users] engine setup fails: error 
creating bridge interface virbr0: File exists - ?


Hi guys.

I'm trying to setup the engine on the latest stable ovirt 
node(in a VM), so a clean, vanilla-default system.


-> $  hosted-engine --deploy --4
...
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Activate 
default libvirt network]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": 
false, "cmd": ["virsh", "net-start", "default"], "delta": 
"0:00:00.042134", "end": "2023-05-11 11:08:59.248405", 
"msg": "non-zero return code", "rc": 1, "start": 
"2023-05-11 11:08:59.206271", "stderr": "error: Failed to 
start network default\nerror: error creating bridge 
interface virbr0: File exists", "stderr_lines": ["error: 
Failed to start network default", "error: error creating 
bridge interface virbr0: File exists"], "stdout": "", 
"stdout_lines": []}
[ ERROR ] Failed to execute stage 'Closing up': Failed 
getting local_vm_dir

...

Any & all suggestions on how to fix/troubleshoot this are 
much appreciated.

many thanks, L.


No, it is a clean install of oVirt node in a KVM guest/VM.
All I do in the node is I configure two ifaces and both are 
"regular" ethernets, then I start HE setup.
Bare-metal host is Centos 9 Stream with everything 
up-to-dayte off the distro repos.

I'd think this should be easily reproducible.

ps. please keep replies to the bottom - this is not a 
conversion between two of us nor any two persons, Mailing to 
list - think of it a book, which always reads top-to-bottom.


many thanks, L.___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/A42X23EGFVJLAJHBNBXIVF7YY5UPYXW6/


[ovirt-users] engine setup fails: error creating bridge interface virbr0: File exists - ?

2023-05-11 Thread lejeczek via Users

Hi guys.

I'm trying to setup the engine on the latest stable ovirt 
node(in a VM), so a clean, vanilla-default system.


-> $  hosted-engine --deploy --4
...
[ INFO  ] TASK [ovirt.ovirt.hosted_engine_setup : Activate 
default libvirt network]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, 
"cmd": ["virsh", "net-start", "default"], "delta": 
"0:00:00.042134", "end": "2023-05-11 11:08:59.248405", 
"msg": "non-zero return code", "rc": 1, "start": "2023-05-11 
11:08:59.206271", "stderr": "error: Failed to start network 
default\nerror: error creating bridge interface virbr0: File 
exists", "stderr_lines": ["error: Failed to start network 
default", "error: error creating bridge interface virbr0: 
File exists"], "stdout": "", "stdout_lines": []}
[ ERROR ] Failed to execute stage 'Closing up': Failed 
getting local_vm_dir

...

Any & all suggestions on how to fix/troubleshoot this are 
much appreciated.

many thanks, L.___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PYHBYP5R6RWW72PIENVPAUG6GB67FGOI/


Re: Firefox - detection of exiting tabs with URLs

2023-05-10 Thread lejeczek via users



On 10/05/2023 17:41, stan via users wrote:

On Wed, 10 May 2023 13:04:53 +0200
lejeczek via users  wrote:


Hi guys.

With f38 I think my Firefox is unable to detect or is
ignoring tabs which are already opened in other windows with
certain URL, when I open a new tab and want to go to a
website/URL.
I think Firefox would then say, would offer something like
"switch to tab".

Do you see it it too?
I did not change any settings, configuration in Firefox so I
wonder it's just me.

I am using nightly, the firefox beta, and it works this way in F37.  I
don't remember it ever working differently.  Maybe an addon you were
using to do this is no longer available?

Here are a few add ons that will prevent this, probably more at the
mozilla addons site.  I haven't used these.

https://addons.mozilla.org/en-US/firefox/addon/dont-open-duplicate-tabs/?utm_source=addons.mozilla.org_medium=referral_content=search

https://addons.mozilla.org/en-US/firefox/addon/smart-prevent-duplicate-tabs/?utm_source=addons.mozilla.org_medium=referral_content=search

https://addons.mozilla.org/en-US/firefox/addon/duplicate-tabs-closer-updated/?utm_source=addons.mozilla.org_medium=referral_content=search

And this more extensive add on seems to be able to do so also.

https://addons.mozilla.org/en-US/firefox/addon/foxytab/?utm_source=addons.mozilla.org_medium=referral_content=search
___

No add-ons in my case, only a few but those I've had for 
long long time.
add-ons: duckduckgo privacy, cookie manager, gab. privacy 
pass from cloudflare and obviously gnome shell integation.

Firefox version firefox-112.0.2-1.fc38.x86_64
I believe that "functionality" was built-in and 
vanilla-defaul behavior.
I've got the feeling something to do with f38 upgrade I did 
recently.

___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


Firefox - detection of exiting tabs with URLs

2023-05-10 Thread lejeczek via users

Hi guys.

With f38 I think my Firefox is unable to detect or is 
ignoring tabs which are already opened in other windows with 
certain URL, when I open a new tab and want to go to a 
website/URL.
I think Firefox would then say, would offer something like 
"switch to tab".


Do you see it it too?
I did not change any settings, configuration in Firefox so I 
wonder it's just me.


many thanks, L.___
users mailing list -- users@lists.fedoraproject.org
To unsubscribe send an email to users-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/users@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue


[ovirt-users] barely started - cannot import name 'Callable' from 'collections'

2023-05-09 Thread lejeczek via Users

Hi guys.

I've barely started, trying to deploy my first oVirt and I get:
...
  Please indicate the gateway IP address [10.3.1.254]:
[ INFO  ] Checking available network interfaces:
[ ERROR ] b'[WARNING]: Skipping plugin 
(/usr/share/ovirt-hosted-engine-\n'
[ ERROR ] 
b'setup/he_ansible/callback_plugins/2_ovirt_logger.py), 
cannot load: cannot\n'

[ ERROR ] b"import name 'Callable' from 'collections'\n"
[ ERROR ] b'(/usr/lib64/python3.11/collections/__init__.py)\n'
[ ERROR ] b"ERROR! Unexpected Exception, this is probably a 
bug: cannot import name 'Callable' from 'collections' 
(/usr/lib64/python3.11/collections/__init__.py)\n"
[ ERROR ] Failed to execute stage 'Environment 
customization': Failed executing ansible-playbook

[ INFO  ] Stage: Clean up
[ INFO  ] Cleaning temporary resources
[ ERROR ] b'[WARNING]: Skipping plugin 
(/usr/share/ovirt-hosted-engine-\n'
[ ERROR ] 
b'setup/he_ansible/callback_plugins/2_ovirt_logger.py), 
cannot load: cannot\n'

[ ERROR ] b"import name 'Callable' from 'collections'\n"
[ ERROR ] b'(/usr/lib64/python3.11/collections/__init__.py)\n'
[ ERROR ] b"ERROR! Unexpected Exception, this is probably a 
bug: cannot import name 'Callable' from 'collections' 
(/usr/lib64/python3.11/collections/__init__.py)\n"
[ ERROR ] Failed to execute stage 'Clean up': Failed 
executing ansible-playbook
[ INFO  ] Generating answer file 
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20230509193552.conf'

[ INFO  ] Stage: Pre-termination
[ INFO  ] Stage: Termination
[ ERROR ] Hosted Engine deployment failed
  Log file is located at 
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20230509193544-s72umf.log


centos 9 stream with 
ovirt-engine-setup-base-4.5.3.1-1.el9.noarch


Any & every suggestion on what is braking here and how to 
troubleshoot/fix it are much appreciated.

thanks, L.___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KYRUZEF7MWYETFY3CVAPOR4UDESUZLGY/


[ClusterLabs] ocf_heartbeat_pgsql - lock file

2023-05-07 Thread lejeczek via Users

Hi guys.

I have a resource seemingly running a ok but when a node 
gets rebooted then cluster finds it not able, not good to 
start the resource.


Failed Resource Actions:
  * PGSQL start on podnode3 returned 'error' (My data may 
be inconsistent. You have to remove 
/var/lib/pgsql/tmp/PGSQL.lock file to force start.) at Sun 
May  7 11:48:43 2023 after 121ms


and indeed, with manual intervention, after removal of that 
file, cluster seems to be happy to rejoin the node into 
pgsql cluster.
Is that intentional, by design and if yes/no then why it 
happens?


many thanks, L.___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] OCF_HEARTBEAT_PGSQL - any good with current Postgres- ?

2023-05-05 Thread lejeczek via Users




On 05/05/2023 10:41, Jehan-Guillaume de Rorthais wrote:

On Fri, 5 May 2023 10:08:17 +0200
lejeczek via Users  wrote:


On 25/04/2023 14:16, Jehan-Guillaume de Rorthais wrote:

Hi,

On Mon, 24 Apr 2023 12:32:45 +0200
lejeczek via Users  wrote:
  

I've been looking up and fiddling with this RA but
unsuccessfully so far, that I wonder - is it good for
current versions of pgSQLs?

As far as I know, the pgsql agent is still supported, last commit on it
happen in Jan 11th 2023. I don't know about its compatibility with latest
PostgreSQL versions.

I've been testing it many years ago, I just remember it was quite hard to
setup, understand and manage from the maintenance point of view.

Also, this agent is fine in a shared storage setup where it only
start/stop/monitor the instance, without paying attention to its role
(promoted or not).
  

It's not only that it's hard - which is purely due to
piss-poor man page in my opinion - but it really sounds
"expired".

I really don't know. My feeling is that the manpage might be expired, which
really doesn't help with this agent, but not the RA itself.


Eg. man page speaks of 'recovery.conf' which - as I
understand it - newer/current versions of pgSQL do not! even
use... which makes one wonder.

This has been fixed in late 2019, but with no documentation associated :/
See:
https://github.com/ClusterLabs/resource-agents/commit/a43075be72683e1d4ddab700ec16d667164d359c

Regards,


Right.. like the rest of us admins/users going to the code 
is the very first thing we do :)
RA/resource seems to work but I sincerely urge the 
programmer(s)/author(s) responsible, to update man pages so 
they reflect state of affairs as of today - there is nothing 
more discouraging - to the rest of us sysadmins/endusers - 
than man pages which look like stingy conservatives 
programmer's note.
What setup of this RA does, is misleading in parts - still 
creates config part in a file called 'recovery.conf'


on that note - does this RA not require way to much open 
pgSQL setup, meaning not very secure? Would anybody know?
I cannot see - again ! regular man pages and not the code - 
how replication could be secured down by not using user 
'postgres' itself and then perhaps adding authentication 
with passwords, at least.


many thanks, L.

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] 99-VirtualDomain-libvirt.conf under control - ?

2023-05-05 Thread lejeczek via Users



On 05/05/2023 10:08, Andrei Borzenkov wrote:

On Fri, May 5, 2023 at 11:03 AM lejeczek via Users
 wrote:



On 29/04/2023 21:02, Reid Wahl wrote:

On Sat, Apr 29, 2023 at 3:34 AM lejeczek via Users
 wrote:

Hi guys.

I presume these are a consequence of having resource of VirtuaDomain type set up(& 
enabled) - but where, how cab users control presence & content of those?

Yep: 
https://github.com/ClusterLabs/resource-agents/blob/v4.12.0/heartbeat/VirtualDomain#L674-L680

You can't really control the content, since it's set by the resource
agent. (You could change it after creation but that defeats the
purpose.) However, you can view it at
/run/systemd/system/resource-agents-deps.target.d/libvirt.conf.

You can see the systemd_drop_in definition here:
https://github.com/ClusterLabs/resource-agents/blob/v4.12.0/heartbeat/ocf-shellfuncs.in#L654-L673


I wonder how much of an impact those bits have on the
cluster(?)
Take '99-VirtualDomain-libvirt.conf' - that one poses
questions, with c9s 'libvirtd.service' is not really used or
should not be, new modular approach is devised there.
So, with 'resources-agents' having:
After=libvirtd.service
and users not being able to manage those bit - is that not
asking for trouble?


it does no harm (missing units are simply ignored) but it certainly
does not do anything useful either. OTOH modular approach is also
optional, so you could still use monolithic libvirtd on cluster nodes.
So it is more a documentation issue.
Not sure what you mean by 'missing unit' - unit is there 
only is not used, is disabled. What does 
'resource-agents-deps' do with that?
I don't suppose upstream, redhat & others made that effort, 
those changes with the suggestions to us consumers - do go 
back to "old" stuff.
I'd suggest, if devel/contributors read here - and I'd 
imagine other users would reckon as well - to enhance RAs, 
certainly VirtualDomain, with a parameter/attribute with 
which users could, at least to certain extent, control those 
"outside" of cluster, dependencies.


thanks, L.
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


  1   2   3   4   >