Hi All,
On the latest Frankfurt instance of OOM in our lab, the curator pod
consistently fails with same error.(logs attached)
Any help/work-around pertaining to this will be very helpful.
Also, the other pod which consistently fails is aaf-sshsm-testca (error below).
Is there any additional plugin/setup required for this and what flow this
covers? (logs attached)
Appreciate any help in this regard.
HW plugin NOT available to use !
The token has been initialized and is reassigned to slot 508812854
The slot ID used is 508812854
TPM hardware unavailable. Using SoftHSM implementation
Regards,
Kuldeep
From: [email protected] <[email protected]> On Behalf Of
Kamel Idir via lists.onap.org
Sent: Friday, July 10, 2020 10:09 PM
To: Won <[email protected]>; [email protected]
Subject: Re: [onap-discuss] [Frankfurt][SDNC] Curator hosts and https issue
[CAUTION: This Email is from outside the Organization. Unless you trust the
sender, Don’t click links or open attachments as it may be a Phishing email,
which can steal your Information and compromise your Computer.]
I'm also having the same issue. did you manage to fix it ?
root@frankfurt-rancher:~# kubectl logs -n onap
onap-sdnrdb-curator-1594342800-trx44
2020-07-10 01:01:18,160 INFO Preparing Action ID: 1, "delete_indices"
2020-07-10 01:01:18,161 INFO Creating client object and testing connection
2020-07-10 01:01:18,163 INFO Instantiating client object
2020-07-10 01:01:18,164 INFO Testing client connectivity
2020-07-10 01:01:18,216 ERROR HTTP N/A error:
HTTPConnectionPool(host='onap-sdnrdb.onap.svc.cluster.local', port=9200): Max
retries exceeded with url: / (Caused by
NewConnectionError('<urllib3.connection.HTTPConnection object at
0x7f7035b38668>: Failed to establish a new connection: [Errno -2] Name or
service not known',))
2020-07-10 01:01:18,216 CRITICAL Curator cannot proceed. Exiting.
::DISCLAIMER::
________________________________
The contents of this e-mail and any attachment(s) are confidential and intended
for the named recipient(s) only. E-mail transmission is not guaranteed to be
secure or error-free as information could be intercepted, corrupted, lost,
destroyed, arrive late or incomplete, or may contain viruses in transmission.
The e mail and its contents (with or without referred errors) shall therefore
not attach any liability on the originator or HCL or its affiliates. Views or
opinions, if any, presented in this email are solely those of the author and
may not necessarily reflect the views or opinions of HCL or its affiliates. Any
form of reproduction, dissemination, copying, disclosure, modification,
distribution and / or publication of this message without the prior written
consent of authorized representative of HCL is strictly prohibited. If you have
received this email in error please delete it and notify the sender
immediately. Before opening any email and/or attachments, please check them for
viruses and other defects.
________________________________
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#21801): https://lists.onap.org/g/onap-discuss/message/21801
Mute This Topic: https://lists.onap.org/mt/75870846/21656
Group Owner: [email protected]
Unsubscribe: https://lists.onap.org/g/onap-discuss/unsub
[[email protected]]
-=-=-=-=-=-=-=-=-=-=-=-
ubuntu@onap-dub-controller-1:~/oom/kubernetes$
kubuntu@onap-dub-controller-1:~/oom/kubernetes$ kubectl -n onap logs
jul27-sdnrdb-curator-1595984400-588qb
2020-07-29 01:00:07,516 INFO Preparing Action ID: 1, "delete_indices"
2020-07-29 01:00:07,516 INFO Creating client object and testing connection
2020-07-29 01:00:07,520 INFO Instantiating client object
2020-07-29 01:00:07,521 INFO Testing client connectivity
2020-07-29 01:00:07,590 ERROR HTTP N/A error:
HTTPConnectionPool(host='jul27-sdnrdb.onap.svc.cluster.local', port=9200): Max
retries exceeded with url: / (Caused by
NewConnectionError('<urllib3.connection.HTTPConnection object at
0x7f0cf82d9e10>: Failed to establish a new connection: [Errno -2] Name or
service not known',))
2020-07-29 01:00:07,591 CRITICAL Curator cannot proceed. Exiting.
ubuntu@onap-dub-controller-1:~/oom/kubernetes$
ubuntu@onap-dub-controller-1:~/oom/kubernetes$ kubectl -n onap describe po
jul27-sdnrdb-curator-1595984400-588qb
Name: jul27-sdnrdb-curator-1595984400-588qb
Namespace: onap
Priority: 0
Node: onap-dub-worker-7/192.168.0.16
Start Time: Wed, 29 Jul 2020 01:00:01 +0000
Labels: app.kubernetes.io/instance=jul27
app.kubernetes.io/managed-by=Tiller
app.kubernetes.io/name=sdnrdb
controller-uid=45bebfde-1808-4303-860a-6011796f445a
helm.sh/chart=curator-6.0.0
job-name=jul27-sdnrdb-curator-1595984400
Annotations: cni.projectcalico.org/podIP: 192.166.5.72/32
cni.projectcalico.org/podIPs: 192.166.5.72/32
Status: Failed
IP: 192.166.5.72
Controlled By: Job/jul27-sdnrdb-curator-1595984400
Containers:
jul27-sdnrdb-curator:
Container ID:
docker://f1d8a27a6f71eea1926636755c89e7b95b1a3f99ac368550b01d762210cbf680
Image: docker.io/bitnami/elasticsearch-curator:5.8.1-debian-9-r74
Image ID:
docker-pullable://bitnami/elasticsearch-curator@sha256:ae29c8fadc9eb654b455579c064349e7ce1d3d629879d61d8e24cacff39938be
Port: <none>
Host Port: <none>
Command:
curator
Args:
--config
/etc/es-curator/config.yml
/etc/es-curator/action_file.yml
State: Terminated
Reason: Error
Exit Code: 1
Started: Wed, 29 Jul 2020 01:00:06 +0000
Finished: Wed, 29 Jul 2020 01:00:07 +0000
Ready: False
Restart Count: 0
Environment: <none>
Mounts:
/etc/es-curator from config-volume (rw)
/var/run/secrets/kubernetes.io/serviceaccount from default-token-np6vr
(ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
config-volume:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: jul27-sdnrdb-curator
Optional: false
default-token-np6vr:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-np6vr
Optional: false
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events: <none>
ubuntu@onap-dub-controller-1:~/oom/kubernetes$ ubuntu@onap-dub-controller-1:~/oom/kubernetes$ kubectl -n onap log
jul27-aaf-sshsm-testca-5gv9d
prepareHWPlugin() called
Couldn't open the directory
HW plugin NOT available to use !
prepareHWPlugin() called
Couldn't open the directory
HW plugin NOT available to use !
The token has been initialized and is reassigned to slot 1248874089
The slot ID used is 1248874089
TPM hardware unavailable. Using SoftHSM implementation
cat: passphrase: No such file or directory
gpg: directory `/root/.gnupg' created
gpg: new configuration file `/root/.gnupg/gpg.conf' created
gpg: WARNING: options in `/root/.gnupg/gpg.conf' are not yet active during this
run
gpg: keyring `/root/.gnupg/secring.gpg' created
gpg: keyring `/root/.gnupg/pubring.gpg' created
gpg: can't open `privkey.pem.gpg'
ubuntu@onap-dub-controller-1:~/oom/kubernetes$
Name: jul27-aaf-sshsm-testca-5gv9d
Namespace: onap
Priority: 0
Node: onap-dub-worker-2/192.168.0.15
Start Time: Mon, 27 Jul 2020 14:41:08 +0000
Labels: app.kubernetes.io/instance=jul27
app.kubernetes.io/managed-by=Tiller
app.kubernetes.io/name=aaf-sshsm-testca
controller-uid=14f0006c-cfaf-4dbd-a3d2-ee976bc74cf3
helm.sh/chart=aaf-sshsm-testca-6.0.0
job-name=jul27-aaf-sshsm-testca
Annotations: cni.projectcalico.org/podIP: 192.166.1.175/32
cni.projectcalico.org/podIPs: 192.166.1.175/32
Status: Failed
IP: 192.166.1.175
Controlled By: Job/jul27-aaf-sshsm-testca
Init Containers:
aaf-sshsm-testca-distcenter-ready:
Container ID:
docker://45e564eb6378d534c54b9429ce8d7f12f883c2a156b9753268952cbb70fefaae
Image: oomk8s/readiness-check:2.0.2
Image ID:
docker-pullable://oomk8s/readiness-check@sha256:875814cc853d7d9b3227165740584c347884bc8c0d831b4148cb5f7952e4cbff
Port: <none>
Host Port: <none>
Command:
/root/job_complete.py
Args:
-j
jul27-aaf-sshsm-distcenter
State: Terminated
Reason: Completed
Exit Code: 0
Started: Mon, 27 Jul 2020 14:41:12 +0000
Finished: Mon, 27 Jul 2020 14:41:16 +0000
Ready: True
Restart Count: 0
Limits:
cpu: 100m
memory: 100Mi
Requests:
cpu: 3m
memory: 20Mi
Environment:
NAMESPACE: onap (v1:metadata.namespace)
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-np6vr
(ro)
Containers:
aaf-sshsm-testca:
Container ID:
docker://afb4af95f21ba4191dd9c719c2ce0f50ea53d2e596d6f615fab0d4fbc79faaf1
Image: nexus3.onap.org:10001/onap/aaf/testcaservice:4.0.0
Image ID:
docker-pullable://nexus3.onap.org:10001/onap/aaf/testcaservice@sha256:fad10cad9feb45a70e78fc6eaca0597cce93b7e9ed14c585aca3242ba7659c29
Port: <none>
Host Port: <none>
Command:
./import.sh
State: Terminated
Reason: Error
Exit Code: 2
Started: Mon, 27 Jul 2020 14:41:19 +0000
Finished: Mon, 27 Jul 2020 14:41:19 +0000
Ready: False
Restart Count: 0
Environment:
DATA_FOLDER: /testca/data
SECRETS_FOLDER: /testca/secrets
Mounts:
/etc/localtime from localtime (ro)
/testca/data from jul27-aaf-sshsm-testca-data (rw)
/testca/secrets from jul27-aaf-sshsm-testca-secrets (ro)
/var/run/dbus from jul27-aaf-sshsm-testca-dbus (rw)
/var/run/secrets/kubernetes.io/serviceaccount from default-token-np6vr
(ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
localtime:
Type: HostPath (bare host directory volume)
Path: /etc/localtime
HostPathType:
jul27-aaf-sshsm-testca-data:
Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim
in the same namespace)
ClaimName: jul27-aaf-sshsm
ReadOnly: false
jul27-aaf-sshsm-testca-dbus:
Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim
in the same namespace)
ClaimName: jul27-aaf-sshsm-dbus
ReadOnly: false
jul27-aaf-sshsm-testca-secrets:
Type: Secret (a volume populated by a Secret)
SecretName: jul27-aaf-sshsm
Optional: false
default-token-np6vr:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-np6vr
Optional: false
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events: <none>