Hi, Marko,

All mariadb data is kept under 
/dockerdata-nfs/dev-mariadb-galera/mariadb-galera/*{0..2}
I'd reather scale down mariadb to 0 pods, backup those data dirs then delete 
them for pod1 and pod2  *{1..2} keeping pod0 data dir intact. After that modify 
grastate.dat as it proposed in the log:
"edit the grastate.dat file manually and set safe_to_bootstrap to 1" and after 
that scale mariadb up to 1 pod to see if it started successfully. If it is just 
scale up to 3 or what number you need.

Regards,
Dmitry
________________________________
From: [email protected] <[email protected]> on behalf of 
Ferrero Marco Angelo via Lists.Onap.Org 
<[email protected]>
Sent: Monday, October 21, 2019 11:35
To: [email protected] <[email protected]>
Subject: [onap-discuss] #mariadb-galera #sdc #so

Hi all, today after 73 days con correct working (since August,9 to October,21) 
all the three pods of mariadb-galera are in CrashLoopbackOff state.
I tried to delete pods dev-mariadb-galera-mariadb-galera-1 and 
dev-mariadb-galera-mariadb-galera-2 hoping in pods re-creation by kubernetes.
It doesn't happened.
Pod dev-mariadb-galera-mariadb-galera-0 fails because of the readiness check. 
Mysql in not able to correctly start before the readiness check.
Here is the log of the pods dev-mariadb-galera-mariadb-galera-0

root@cp-1:~/oom/kubernetes# kubectl logs -n onap -f 
dev-mariadb-galera-mariadb-galera-0 >  faulty_mariadb_galera
root@cp-1:~/oom/kubernetes# cat faulty_mariadb_galera
+ CONTAINER_SCRIPTS_DIR=/usr/share/container-scripts/mysql
+ EXTRA_DEFAULTS_FILE=/etc/my.cnf.d/galera.cnf
+ '[' -z onap ']'
+ echo 'Galera: Finding peers'
Galera: Finding peers
++ hostname -f
++ cut -d. -f2
+ K8S_SVC_NAME=mariadb-galera
+ echo 'Using service name: mariadb-galera'
+ cp /usr/share/container-scripts/mysql/galera.cnf /etc/my.cnf.d/galera.cnf
Using service name: mariadb-galera
+ /usr/bin/peer-finder 
-on-start=/usr/share/container-scripts/mysql/configure-galera.sh 
-service=mariadb-galera
2019/10/21 09:01:46 Peer list updated
was []
now [dev-mariadb-galera-mariadb-galera-0.mariadb-galera.onap.svc.cluster.local]
2019/10/21 09:01:46 execing: 
/usr/share/container-scripts/mysql/configure-galera.sh with stdin: 
dev-mariadb-galera-mariadb-galera-0.mariadb-galera.onap.svc.cluster.local
2019/10/21 09:01:46
2019/10/21 09:01:47 Peer finder exiting
+ '[' '!' -d /var/lib/mysql/mysql ']'
+ exec mysqld
2019-10-21  9:01:47 140401625143552 [Note] mysqld (mysqld 10.1.24-MariaDB) 
starting as process 1 ...
2019-10-21  9:01:47 140401625143552 [Note] WSREP: Read nil XID from storage 
engines, skipping position init
2019-10-21  9:01:47 140401625143552 [Note] WSREP: wsrep_load(): loading 
provider library '/usr/lib64/galera/libgalera_smm.so'
2019-10-21  9:01:47 140401625143552 [Note] WSREP: wsrep_load(): Galera 
25.3.20(r3703) by Codership Oy <info@...> loaded successfully.
2019-10-21  9:01:47 140401625143552 [Note] WSREP: CRC-32C: using hardware 
acceleration.
2019-10-21  9:01:47 140401625143552 [Note] WSREP: Found saved state: 
00000000-0000-0000-0000-000000000000:-1, safe_to_bootsrap: 0
2019-10-21  9:01:47 140401625143552 [Note] WSREP: Passing config to GCS: 
base_dir = /var/lib/mysql/; base_host = 
dev-mariadb-galera-mariadb-galera-0.mariadb-galera.onap.svc.cluster.local; 
base_port = 4567; cert.log_conflicts = no; debug = no; evs.auto_evict = 0; 
evs.delay_margin = PT1S; evs.delayed_keep_period = PT30S; 
evs.inactive_check_period = PT0.5S; evs.inactive_timeout = PT15S; 
evs.join_retrans_period = PT1S; evs.max_install_timeouts = 3; evs.send_window = 
4; evs.stats_report_period = PT1M; evs.suspect_timeout = PT5S; 
evs.user_send_window = 2; evs.view_forget_timeout = PT24H; gcache.dir = 
/var/lib/mysql/; gcache.keep_pages_size = 0; gcache.mem_size = 0; gcache.name = 
/var/lib/mysql//galera.cache; gcache.page_size = 128M; gcache.recover = no; 
gcache.size = 128M; gcomm.thread_prio = ; gcs.fc_debug = 0; gcs.fc_factor = 
1.0; gcs.fc_limit = 16; gcs.fc_master_slave = no; gcs.max_packet_size = 64500; 
gcs.max_throttle = 0.25; gcs.recv_q_hard_limit = 9223372036854775807; 
gcs.recv_q_soft_limit = 0.25; gcs.sync_donor = no; gmcast.segment = 0; 
gmcast.versi
2019-10-21  9:01:47 140401625143552 [Note] WSREP: GCache history reset: 
old(00000000-0000-0000-0000-000000000000:0) -> 
new(00000000-0000-0000-0000-000000000000:-1)
2019-10-21  9:01:47 140401625143552 [Note] WSREP: Assign initial position for 
certification: -1, protocol version: -1
2019-10-21  9:01:47 140401625143552 [Note] WSREP: wsrep_sst_grab()
2019-10-21  9:01:47 140401625143552 [Note] WSREP: Start replication
2019-10-21  9:01:47 140401625143552 [Note] WSREP: Setting initial position to 
00000000-0000-0000-0000-000000000000:-1
2019-10-21  9:01:47 140401625143552 [ERROR] WSREP: It may not be safe to 
bootstrap the cluster from this node. It was not the last one to leave the 
cluster and may not contain all the updates. To force cluster bootstrap with 
this node, edit the grastate.dat file manually and set safe_to_bootstrap to 1 .
2019-10-21  9:01:47 140401625143552 [ERROR] WSREP: wsrep::connect(gcomm://) 
failed: 7
2019-10-21  9:01:47 140401625143552 [ERROR] Aborting


Someone has an idea of what whas happened ?
Somemone knows how to re-deploy mariadb-galera KEEPING all the data in the 
underlying databasse ?
Thank you for the help

Marco F.


-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#19228): https://lists.onap.org/g/onap-discuss/message/19228
Mute This Topic: https://lists.onap.org/mt/36311989/21656
Mute #sdc: https://lists.onap.org/mk?hashtag=sdc&subid=2740164
Mute #so: https://lists.onap.org/mk?hashtag=so&subid=2740164
Mute #mariadb-galera: 
https://lists.onap.org/mk?hashtag=mariadb-galera&subid=2740164
Group Owner: [email protected]
Unsubscribe: https://lists.onap.org/g/onap-discuss/unsub  
[[email protected]]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to