Hi, Seshu, I will check on those.
Thanks a lot. Xin Miao Solution Engineering Fujitsu Network Communication (W)972-479-2263 (M)469-268-5226 2811 Telecom Drive Richardson, TX 75081, USA From: Seshu m [mailto:[email protected]] Sent: Wednesday, September 1, 2021 8:34 PM To: Miao, Xin <[email protected]> Cc: Deepika S (iDEAS-ER&D) <[email protected]>; Rajewski Łukasz - Hurt <[email protected]>; onap-discuss <[email protected]>; [email protected] Subject: RE: SO pods recovery failure Hi Xin To me it looks like an issue with the Maria dB connectivity. You can check this from the describe of the catalogdb and request dB pods. Since sdc controller is dependent on this that too has failed and similar seems to be the reasoning for the other pods. ________________________________ Thanks and Regards, M Seshu Kumar Lead Architect, P&S, Cloud Network OSDT, Huawei Technologies India Pvt. Ltd. Survey No. 37, Next to EPIP Area, Kundalahalli, Whitefield Bengaluru-560066, Karnataka. Tel: + 91-80-49160700 , Mob: 9845355488 ___________________________________________________________________________________________________ This e-mail and its attachments contain confidential information from HUAWEI, which is intended only for the person or entity whose address is listed above. Any use of the information contained herein in any way (including, but not limited to, total or partial disclosure, reproduction, or dissemination) by persons other than the intended recipient(s) is prohibited. If you receive this e-mail in error, please notify the sender by phone or email immediately and delete it! ________________________________ ________________________________ ________________________________ ------------- From:Xin.Miao <[email protected]<mailto:[email protected]>> To:Seshu m <[email protected]<mailto:[email protected]>> Cc:Deepika S (iDEAS-ER&D) <[email protected]<mailto:[email protected]>>;Rajewski Łukasz - Hurt <[email protected]<mailto:[email protected]>>;onap-discuss <[email protected],[email protected]<mailto:[email protected],[email protected]>> Date:2021-08-31 23:48:13 Subject:SO pods recovery failure Hi, Seshu/Team, How are you? I ran into the following SO pods recovery failure issues. Any idea on how to bring it back w/o redeploying entire SO component? Thanks a lot! ubuntu@onap-rancher-vm:~$ kubectl get po -n onap | grep -i dev-so dev-so-64f474fb4b-ns94t 2/2 Running 0 22d dev-so-admin-cockpit-669f97bb56-n47qs 1/1 Running 0 22d dev-so-bpmn-infra-78d98f7cbd-c297v 2/2 Running 0 19d dev-so-catalog-db-adapter-c895cdcd9-ldx5t 0/1 Init:2/3 477 3d11h dev-so-cnf-adapter-c48d4d844-llllj 1/1 Running 0 3d11h dev-so-etsi-nfvo-ns-lcm-57ccd6888d-wz7q4 1/1 Running 0 3d11h dev-so-etsi-sol003-adapter-5459d685c9-jkqd2 1/1 Running 0 3d11h dev-so-etsi-sol005-adapter-64ff7c46d4-ghvn7 0/1 Init:2/3 477 3d11h dev-so-nssmf-adapter-577cbcbf96-g4z57 0/1 Init:2/3 478 3d11h dev-so-oof-adapter-74794c4f64-vt4jw 2/2 Running 0 3d11h dev-so-openstack-adapter-5769c9fc65-mtmt6 2/2 Running 0 22d dev-so-request-db-adapter-589f8f4c4f-xbhrn 0/1 Init:2/3 477 3d11h dev-so-sdc-controller-d99c85467-tmqw6 0/2 Init:2/3 477 3d11h dev-so-sdnc-adapter-674b56df8b-rgbtc 2/2 Running 0 22d What happened: 3 days ago, one of my local rancher worker node got reset for some reason. It causes all the pods running on that node got restarted. The above highlighted SO pods never came back. They still failed even after manually restarted them. Several other pods, such as, dev-ejbca, dev-holmes and some dcae pods are at the similar status. I do not want to redeploy my entire local ONAP instance just yet since I do want to lose the data and configuration information for now. Any idea to save the SO? Thanks a lot. Xin Miao Solution Engineering Fujitsu Network Communication -=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#23528): https://lists.onap.org/g/onap-discuss/message/23528 Mute This Topic: https://lists.onap.org/mt/85283270/21656 Group Owner: [email protected] Unsubscribe: https://lists.onap.org/g/onap-discuss/unsub [[email protected]] -=-=-=-=-=-=-=-=-=-=-=-
