Nagios backups are failing

This support forum board is for support questions relating to Nagios XI, our flagship commercial network monitoring solution.
Locked
TSCAdmin
Posts: 155
Joined: Wed Apr 14, 2010 3:00 pm
Location: India

Nagios backups are failing

Post by TSCAdmin »

Greetings,

We are using Nagios XI 2009R1.3 on 64bit CentOS Linux 5.4.

We are monitoring 2100+ hosts and 10800 services. Our Nagios configuration/data has grown bigger in last few days which is causing all our backup jobs to fail with the following error message:

Code: Select all

zip error: Entry too big to split, read, or write (file exceeds Zip's 4GB uncompressed size limit)
The usage of nagios and nagiosxi directories is as follows:

Code: Select all

9.5G    nagios
39M     nagiosxi
Inside nagios directory share is consuming most of the space:

Code: Select all

1.4M    bin
19M     etc
8.1M    libexec
3.8M    sbin
9.0G    share
435M    var
and inside share directory perfdata is biggest:

Code: Select all

4.0K    config.inc.php
256K    contexthelp
3.2M    docs
6.8M    images
16K     includes
4.0K    index.php
124K    locale
4.0K    main.php
4.0K    media
11M     nagvis
9.0G    perfdata
1.1M    pnp
4.0K    robots.txt
8.0K    side.php
4.0K    ssi
100K    stylesheets
Is there an easy way to fix this problem or a renewed backup script? Also is there a way to keep perfdata compact?

The zip version we are using is 2.31 which comes default with Centos 5.4.

Please let me know if you require any other information from my side.

On a relative note - When we remove a host from Nagios XI configuration Nagios does not seem to remove it's directory from perfdata. Is there a specific reason for that?

Thanks
mguthrie
Posts: 4380
Joined: Mon Jun 14, 2010 10:21 am

Re: Nagios backups are failing

Post by mguthrie »

The issue with a 4GB limit was actually fixed in our latest 2011R3.3 release, which uses tar.gz files instead of .zip files, since zip has a 4gb limit. You could either upgrade to the latest version if XI, or grab the backup script from the latest tarball.

We don't currently have a cleaner to remove old performance data. Your best bet would be to write some form of a script to clean up any old performance data that hasn't been updated in X amount of days or months.
Locked