Hello team ,
I have some repetitive problems (daily) and don’t know why it is happening , I have tried find in the log but no informations , totally blocked
My configuration is the following ( CEE version ) :
I have one master and 2 slaves in differents network zones , the master are piloting the slaves on configuration push
I use the master to monitor the slaves ( I have installed the standard agent on the slaves ) , the slaves are considered a a normal host , seen in the master
Problems :
In the master it the backup plugins says it is CRIT but the backup has succeded
plugins output : CRIT - Backup completed, it was running for 6 m from 2020-07-05 03:30:02 till 2020-07-05 03:36:58, Size: 23.84 GB, Next run: 2020-07-06 03:30:00 CRIT
additional info (agent plugins in diagnostic ) :
mkbackup
[[[site:prd01:Daily_Backups]]]
{
“bytes_per_second”: 262779899.90978375,
“finished”: 1593913018.262849,
“next_schedule”: 1593999000.0,
“output”: “2020-07-05 03:30:02 — Starting backup (Check_MK-xxxxxx-prd01-Daily_Backups to xxxx_Backups_target) —\n2020-07-05 03:35:42 Verifying backup consistency\n2020-07-05 03:36:57 Cleaning up previously completed backup\n2020-07-05 03:36:58 — Backup completed (Duration: 0:06:55, Size: 23.84 GB, IO: 250.61 MB/s) —\n”,
“pid”: 102838,
“size”: 25602887680,
“started”: 1593912602.826015,
“state”: “finished”,
“success”: true
}
I’ve verified on the backup section on the slave , it has succeeded
I don’t know why is has happening because it non sense , if the backup has failed then it is logic but it has succeeded
Someone has any idea on how to handle / fix this ?
Kr ,
Andry