Hi All,
On the latest Frankfurt instance of OOM in our lab, the curator pod 
consistently fails with same error.(logs attached)
Any help/work-around pertaining to this will be very helpful.

Also, the other pod which consistently fails is aaf-sshsm-testca (error below). 
Is there any additional plugin/setup required for this and what flow this 
covers? (logs attached)
Appreciate any help in this regard.

                HW plugin NOT available to use !
The token has been initialized and is reassigned to slot 508812854
The slot ID used is 508812854
TPM hardware unavailable. Using SoftHSM implementation

Regards,
Kuldeep

From: onap-discuss@lists.onap.org <onap-discuss@lists.onap.org> On Behalf Of 
Kamel Idir via lists.onap.org
Sent: Friday, July 10, 2020 10:09 PM
To: Won <wjst...@gmail.com>; onap-discuss@lists.onap.org
Subject: Re: [onap-discuss] [Frankfurt][SDNC] Curator hosts and https issue

[CAUTION: This Email is from outside the Organization. Unless you trust the 
sender, Don’t click links or open attachments as it may be a Phishing email, 
which can steal your Information and compromise your Computer.]
I'm also having the same issue. did you manage to fix it ?
root@frankfurt-rancher:~# kubectl logs -n onap 
onap-sdnrdb-curator-1594342800-trx44
2020-07-10 01:01:18,160 INFO      Preparing Action ID: 1, "delete_indices"
2020-07-10 01:01:18,161 INFO      Creating client object and testing connection
2020-07-10 01:01:18,163 INFO      Instantiating client object
2020-07-10 01:01:18,164 INFO      Testing client connectivity
2020-07-10 01:01:18,216 ERROR     HTTP N/A error: 
HTTPConnectionPool(host='onap-sdnrdb.onap.svc.cluster.local', port=9200): Max 
retries exceeded with url: / (Caused by 
NewConnectionError('<urllib3.connection.HTTPConnection object at 
0x7f7035b38668>: Failed to establish a new connection: [Errno -2] Name or 
service not known',))
2020-07-10 01:01:18,216 CRITICAL  Curator cannot proceed. Exiting.


::DISCLAIMER::
________________________________
The contents of this e-mail and any attachment(s) are confidential and intended 
for the named recipient(s) only. E-mail transmission is not guaranteed to be 
secure or error-free as information could be intercepted, corrupted, lost, 
destroyed, arrive late or incomplete, or may contain viruses in transmission. 
The e mail and its contents (with or without referred errors) shall therefore 
not attach any liability on the originator or HCL or its affiliates. Views or 
opinions, if any, presented in this email are solely those of the author and 
may not necessarily reflect the views or opinions of HCL or its affiliates. Any 
form of reproduction, dissemination, copying, disclosure, modification, 
distribution and / or publication of this message without the prior written 
consent of authorized representative of HCL is strictly prohibited. If you have 
received this email in error please delete it and notify the sender 
immediately. Before opening any email and/or attachments, please check them for 
viruses and other defects.
________________________________

-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#21801): https://lists.onap.org/g/onap-discuss/message/21801
Mute This Topic: https://lists.onap.org/mt/75870846/21656
Group Owner: onap-discuss+ow...@lists.onap.org
Unsubscribe: https://lists.onap.org/g/onap-discuss/unsub  
[arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

ubuntu@onap-dub-controller-1:~/oom/kubernetes$ 
kubuntu@onap-dub-controller-1:~/oom/kubernetes$ kubectl -n onap logs 
jul27-sdnrdb-curator-1595984400-588qb
2020-07-29 01:00:07,516 INFO      Preparing Action ID: 1, "delete_indices"
2020-07-29 01:00:07,516 INFO      Creating client object and testing connection
2020-07-29 01:00:07,520 INFO      Instantiating client object
2020-07-29 01:00:07,521 INFO      Testing client connectivity
2020-07-29 01:00:07,590 ERROR     HTTP N/A error: 
HTTPConnectionPool(host='jul27-sdnrdb.onap.svc.cluster.local', port=9200): Max 
retries exceeded with url: / (Caused by 
NewConnectionError('<urllib3.connection.HTTPConnection object at 
0x7f0cf82d9e10>: Failed to establish a new connection: [Errno -2] Name or 
service not known',))
2020-07-29 01:00:07,591 CRITICAL  Curator cannot proceed. Exiting.
ubuntu@onap-dub-controller-1:~/oom/kubernetes$
ubuntu@onap-dub-controller-1:~/oom/kubernetes$ kubectl -n onap describe po 
jul27-sdnrdb-curator-1595984400-588qb
Name:           jul27-sdnrdb-curator-1595984400-588qb
Namespace:      onap
Priority:       0
Node:           onap-dub-worker-7/192.168.0.16
Start Time:     Wed, 29 Jul 2020 01:00:01 +0000
Labels:         app.kubernetes.io/instance=jul27
                app.kubernetes.io/managed-by=Tiller
                app.kubernetes.io/name=sdnrdb
                controller-uid=45bebfde-1808-4303-860a-6011796f445a
                helm.sh/chart=curator-6.0.0
                job-name=jul27-sdnrdb-curator-1595984400
Annotations:    cni.projectcalico.org/podIP: 192.166.5.72/32
                cni.projectcalico.org/podIPs: 192.166.5.72/32
Status:         Failed
IP:             192.166.5.72
Controlled By:  Job/jul27-sdnrdb-curator-1595984400
Containers:
  jul27-sdnrdb-curator:
    Container ID:  
docker://f1d8a27a6f71eea1926636755c89e7b95b1a3f99ac368550b01d762210cbf680
    Image:         docker.io/bitnami/elasticsearch-curator:5.8.1-debian-9-r74
    Image ID:      
docker-pullable://bitnami/elasticsearch-curator@sha256:ae29c8fadc9eb654b455579c064349e7ce1d3d629879d61d8e24cacff39938be
    Port:          <none>
    Host Port:     <none>
    Command:
      curator
    Args:
      --config
      /etc/es-curator/config.yml
      /etc/es-curator/action_file.yml
    State:          Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Wed, 29 Jul 2020 01:00:06 +0000
      Finished:     Wed, 29 Jul 2020 01:00:07 +0000
    Ready:          False
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /etc/es-curator from config-volume (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-np6vr 
(ro)
Conditions:
  Type              Status
  Initialized       True
  Ready             False
  ContainersReady   False
  PodScheduled      True
Volumes:
  config-volume:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      jul27-sdnrdb-curator
    Optional:  false
  default-token-np6vr:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  default-token-np6vr
    Optional:    false
QoS Class:       BestEffort
Node-Selectors:  <none>
Tolerations:     node.kubernetes.io/not-ready:NoExecute for 300s
                 node.kubernetes.io/unreachable:NoExecute for 300s
Events:          <none>
ubuntu@onap-dub-controller-1:~/oom/kubernetes$ 
ubuntu@onap-dub-controller-1:~/oom/kubernetes$ kubectl -n onap log 
jul27-aaf-sshsm-testca-5gv9d
prepareHWPlugin() called
Couldn't open the directory
HW plugin NOT available to use !
prepareHWPlugin() called
Couldn't open the directory
HW plugin NOT available to use !
The token has been initialized and is reassigned to slot 1248874089
The slot ID used is 1248874089
TPM hardware unavailable. Using SoftHSM implementation
cat: passphrase: No such file or directory
gpg: directory `/root/.gnupg' created
gpg: new configuration file `/root/.gnupg/gpg.conf' created
gpg: WARNING: options in `/root/.gnupg/gpg.conf' are not yet active during this 
run
gpg: keyring `/root/.gnupg/secring.gpg' created
gpg: keyring `/root/.gnupg/pubring.gpg' created
gpg: can't open `privkey.pem.gpg'
ubuntu@onap-dub-controller-1:~/oom/kubernetes$
Name:           jul27-aaf-sshsm-testca-5gv9d
Namespace:      onap
Priority:       0
Node:           onap-dub-worker-2/192.168.0.15
Start Time:     Mon, 27 Jul 2020 14:41:08 +0000
Labels:         app.kubernetes.io/instance=jul27
                app.kubernetes.io/managed-by=Tiller
                app.kubernetes.io/name=aaf-sshsm-testca
                controller-uid=14f0006c-cfaf-4dbd-a3d2-ee976bc74cf3
                helm.sh/chart=aaf-sshsm-testca-6.0.0
                job-name=jul27-aaf-sshsm-testca
Annotations:    cni.projectcalico.org/podIP: 192.166.1.175/32
                cni.projectcalico.org/podIPs: 192.166.1.175/32
Status:         Failed
IP:             192.166.1.175
Controlled By:  Job/jul27-aaf-sshsm-testca
Init Containers:
  aaf-sshsm-testca-distcenter-ready:
    Container ID:  
docker://45e564eb6378d534c54b9429ce8d7f12f883c2a156b9753268952cbb70fefaae
    Image:         oomk8s/readiness-check:2.0.2
    Image ID:      
docker-pullable://oomk8s/readiness-check@sha256:875814cc853d7d9b3227165740584c347884bc8c0d831b4148cb5f7952e4cbff
    Port:          <none>
    Host Port:     <none>
    Command:
      /root/job_complete.py
    Args:
      -j
      jul27-aaf-sshsm-distcenter
    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Mon, 27 Jul 2020 14:41:12 +0000
      Finished:     Mon, 27 Jul 2020 14:41:16 +0000
    Ready:          True
    Restart Count:  0
    Limits:
      cpu:     100m
      memory:  100Mi
    Requests:
      cpu:     3m
      memory:  20Mi
    Environment:
      NAMESPACE:  onap (v1:metadata.namespace)
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-np6vr 
(ro)
Containers:
  aaf-sshsm-testca:
    Container ID:  
docker://afb4af95f21ba4191dd9c719c2ce0f50ea53d2e596d6f615fab0d4fbc79faaf1
    Image:         nexus3.onap.org:10001/onap/aaf/testcaservice:4.0.0
    Image ID:      
docker-pullable://nexus3.onap.org:10001/onap/aaf/testcaservice@sha256:fad10cad9feb45a70e78fc6eaca0597cce93b7e9ed14c585aca3242ba7659c29
    Port:          <none>
    Host Port:     <none>
    Command:
      ./import.sh
    State:          Terminated
      Reason:       Error
      Exit Code:    2
      Started:      Mon, 27 Jul 2020 14:41:19 +0000
      Finished:     Mon, 27 Jul 2020 14:41:19 +0000
    Ready:          False
    Restart Count:  0
    Environment:
      DATA_FOLDER:     /testca/data
      SECRETS_FOLDER:  /testca/secrets
    Mounts:
      /etc/localtime from localtime (ro)
      /testca/data from jul27-aaf-sshsm-testca-data (rw)
      /testca/secrets from jul27-aaf-sshsm-testca-secrets (ro)
      /var/run/dbus from jul27-aaf-sshsm-testca-dbus (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from default-token-np6vr 
(ro)
Conditions:
  Type              Status
  Initialized       True
  Ready             False
  ContainersReady   False
  PodScheduled      True
Volumes:
  localtime:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/localtime
    HostPathType:
  jul27-aaf-sshsm-testca-data:
    Type:       PersistentVolumeClaim (a reference to a PersistentVolumeClaim 
in the same namespace)
    ClaimName:  jul27-aaf-sshsm
    ReadOnly:   false
  jul27-aaf-sshsm-testca-dbus:
    Type:       PersistentVolumeClaim (a reference to a PersistentVolumeClaim 
in the same namespace)
    ClaimName:  jul27-aaf-sshsm-dbus
    ReadOnly:   false
  jul27-aaf-sshsm-testca-secrets:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  jul27-aaf-sshsm
    Optional:    false
  default-token-np6vr:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  default-token-np6vr
    Optional:    false
QoS Class:       BestEffort
Node-Selectors:  <none>
Tolerations:     node.kubernetes.io/not-ready:NoExecute for 300s
                 node.kubernetes.io/unreachable:NoExecute for 300s
Events:          <none>

Reply via email to