Maintenance may not be working

This support forum board is for support questions relating to Nagios Log Server, our solution for managing and monitoring critical log data.
User avatar
mike4vr
Posts: 89
Joined: Wed Feb 04, 2015 2:23 pm

Maintenance may not be working

Post by mike4vr »

We are seeing roughly a 10GB increase every day in our storage even though we have the following Maintenance settings:

Optimize indexes older than: 2 days
Close indexes older than: 30 days
Delete indexes older than: 60 days
Delete backups older than: 180 days

The backups are being done to a mounted NFS share and not on local disk.

Is there anything you can recommend checking?
User avatar
hsmith
Agent Smith
Posts: 3539
Joined: Thu Jul 30, 2015 11:09 am
Location: 127.0.0.1
Contact:

Re: Maintenance may not be working

Post by hsmith »

How long has this behavior been happening for?
Former Nagios Employee.
me.
User avatar
mike4vr
Posts: 89
Joined: Wed Feb 04, 2015 2:23 pm

Re: Maintenance may not be working

Post by mike4vr »

hsmith wrote:How long has this behavior been happening for?
Months. I changed the "Delete indexes older than" from 90 to 60 days a few weeks ago. When the scheduler ran, it shaved off quite a bit. However, the size keeps growing. It is now larger than it was when I retained 90 days of indexes.
jolson
Attack Rabbit
Posts: 2560
Joined: Thu Feb 12, 2015 12:40 pm

Re: Maintenance may not be working

Post by jolson »

We are seeing roughly a 10GB increase every day in our storage even though we have the following Maintenance settings:
Can you identify what exactly is growing? Do you know if it's your backups that are growing wildly, or is it the data that you're receiving in Nagios Log Server? Any additional information or screenshots you could provide us would be very useful in diagnosing the problem.
Twits Blog
Show me a man who lives alone and has a perpetually clean kitchen, and 8 times out of 9 I'll show you a man with detestable spiritual qualities.
User avatar
mike4vr
Posts: 89
Joined: Wed Feb 04, 2015 2:23 pm

Re: Maintenance may not be working

Post by mike4vr »

jolson wrote:
We are seeing roughly a 10GB increase every day in our storage even though we have the following Maintenance settings:
Can you identify what exactly is growing? Do you know if it's your backups that are growing wildly, or is it the data that you're receiving in Nagios Log Server? Any additional information or screenshots you could provide us would be very useful in diagnosing the problem.
The backup repository is not on local disk. It is on a mounted NFS share.

There are days where we generate more logs in record numbers, but it isn't consistent. The only consistency we have is the disk usage going up by 1% every 24 hours on 1TB disk. What sort of information and/or screenshots would you like for me to provide?
User avatar
hsmith
Agent Smith
Posts: 3539
Joined: Thu Jul 30, 2015 11:09 am
Location: 127.0.0.1
Contact:

Re: Maintenance may not be working

Post by hsmith »

This command should be helpful: find / -type f -print0 | xargs -0 du | sort -n | tail -10 | cut -f2 | xargs -I{} du -sh {}
Former Nagios Employee.
me.
User avatar
mike4vr
Posts: 89
Joined: Wed Feb 04, 2015 2:23 pm

Re: Maintenance may not be working

Post by mike4vr »

hsmith wrote:This command should be helpful: find / -type f -print0 | xargs -0 du | sort -n | tail -10 | cut -f2 | xargs -I{} du -sh {}

Code: Select all

1.2G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.09/1/index/_osj_Lucene41_0.tim
1.3G	/mnt/repo/indices/logstash-2016.03.17/4/__3r
1.3G	/mnt/repo/indices/logstash-2016.03.10/4/__46
1.3G	/mnt/repo/indices/logstash-2016.03.09/0/__3w
1.3G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.09/2/index/_pke_Lucene41_0.tim
1.3G	/mnt/repo/indices/logstash-2016.03.08/0/__4h
1.4G	/mnt/repo/indices/logstash-2016.03.17/0/__4f
1.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.09/4/index/_pfy_Lucene41_0.tim
1.4G	/mnt/repo/indices/logstash-2016.03.10/3/__3z
1.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.10/3/index/_p5q_Lucene41_0.tim
/mnt/repo is the mounted nfs share.

Here's where my disk usage is:

Code: Select all

# du -sh --exclude=/mnt/repo --exclude=/proc /*
8.9M	/bin
60M	/boot
168K	/dev
30M	/etc
48K	/home
261M	/lib
26M	/lib64
16K	/lost+found
4.0K	/media
0	/misc
4.0K	/mnt
0	/net
91M	/opt
556K	/rhome
2.4M	/root
13M	/sbin
4.0K	/selinux
4.0K	/srv
356M	/store
0	/sys
701M	/tmp
587G	/usr
511M	/var
Here's where the usage is:

Code: Select all

# du -sh /usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/*
612K	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/kibana-int
8.6G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.22
4.3G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.23
3.7G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.24
7.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.25
9.2G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.26
9.3G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.27
8.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.28
8.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.29
4.1G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.30
3.5G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.01.31
8.1G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.01
8.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.02
9.0G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.03
8.8G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.04
8.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.05
4.6G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.06
3.5G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.07
7.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.08
9.2G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.09
9.1G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.10
9.3G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.11
8.7G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.12
4.1G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.13
3.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.14
7.3G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.15
8.7G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.16
9.1G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.17
9.1G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.18
8.6G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.19
4.2G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.20
3.6G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.21
8.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.22
9.2G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.23
9.3G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.24
9.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.25
11G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.26
5.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.27
4.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.28
9.4G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.02.29
11G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.01
11G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.02
11G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.03
11G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.04
6.5G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.05
5.9G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.06
13G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.07
20G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.08
20G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.09
20G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.10
18G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.11
12G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.12
10G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.13
18G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.14
18G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.15
16G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.16
20G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.17
19G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.18
12G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.19
11G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.20
16G	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.03.21
912K	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/logstash-2016.11.25
420K	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/nagioslogserver
227M	/usr/local/nagioslogserver/elasticsearch/data/98ec1d45-73e0-4296-ba7f-d7d71953b7e8/nodes/0/indices/nagioslogserver_log
Hope that helps a bit.
User avatar
hsmith
Agent Smith
Posts: 3539
Joined: Thu Jul 30, 2015 11:09 am
Location: 127.0.0.1
Contact:

Re: Maintenance may not be working

Post by hsmith »

I was looking for something growing out of control. Can you try to go to Administration --> System --> Command Subsystem and hit the Reset All Jobs button? I just want to make sure this is working correctly.
Former Nagios Employee.
me.
User avatar
mike4vr
Posts: 89
Joined: Wed Feb 04, 2015 2:23 pm

Re: Maintenance may not be working

Post by mike4vr »

hsmith wrote:I was looking for something growing out of control. Can you try to go to Administration --> System --> Command Subsystem and hit the Reset All Jobs button? I just want to make sure this is working correctly.
Sure. Just did that. Says "Subsystem jobs have been reset." Should we wait a day now?
User avatar
hsmith
Agent Smith
Posts: 3539
Joined: Thu Jul 30, 2015 11:09 am
Location: 127.0.0.1
Contact:

Re: Maintenance may not be working

Post by hsmith »

I would wait a day to see if you see that 1% growth.
Former Nagios Employee.
me.
Locked