The logs don’t tell me a lot, I’m afraid. The current Graylog server.log, which begins 2019-10-21T06:40:46.667+02:00, contains only lines reading
WARN [LicenseChecker] License violation - Detected irregular traffic records
On 2019-10-18, when I reported the issue, it still had numerous lines like the following two:
2019-10-18T13:37:41.534+02:00 WARN [Messages] Failed to index message: index=<palo_alto_8> id= error=<{“type”:“mapper_parsing_exception”,“reason”:“failed to parse field [pa_dev_group_hierarchy_level_1] of type [integer] in document with id ‘b14fe960-f19b-11e9-9f74-005056842309’”,“caused_by”:{“type”:“number_format_exception”,“reason”:"For input string: “0x0"”}}>
2019-10-18T13:37:41.534+02:00 ERROR [Messages] Failed to index [1] messages. Please check the index error log in your web interface for the reason. Error: One or more of the items in the Bulk request failed, check BulkResult.getItems() for more information.
I’d asked about these parsing errors here before. Since I wasn’t able to correct the mapping problem, I decided to disable the extractors with the field called pa_dev_group_hierarchy_level_1. I had done that well before 2019-10-18, but since errors relating to this field were still popping up, I decided to be a bit more drastic and rebooted the system. That’s when I got the short burst of messages that made it into Graylog. And the pa_dev_group_hierarchy_level_1 mapping errors have since disappeared.
The other thing I can see in the Graylog server.log is that it rotated indices starting on 2019-10-21T02:00:06.368+02:00, but that looks fine.
As for the Elasticsearch logs, I don’t really undestand them… In /var/log/elasticsearch, I’ve got a number of files of the form gc.log.[.current], with gc.log.0.current being the one with today’s timestamp. Its last few lines read as follows:
2019-10-23T10:12:26.384+0200: 420429,081: Total time for which application threads were stopped: 0,0140923 seconds, Stopping threads took: 0,0000427 seconds
2019-10-23T10:17:06.375+0200: 420709,072: [GC (Allocation Failure) 2019-10-23T10:17:06.375+0200: 420709,072: [ParNew
Desired survivor size 8716288 bytes, new threshold 6 (max 6)
- age 1: 335880 bytes, 335880 total
- age 2: 408 bytes, 336288 total
- age 3: 32 bytes, 336320 total
: 136817K->432K(153344K), 0,0120600 secs] 526626K->390241K(1031552K), 0,0121570 secs] [Times: user=0,01 sys=0,00, real=0,01 secs]
I have no idea what that means. The other current log files in that directory are:
- graylog_deprecation.log
- graylog_index_indexing_slowlog.log (zero bytes)
- graylog_index_search_slowlog.log (zero bytes)
- graylog.log
The entire contents of graylog.log are:
[2019-10-21T02:00:07,215][INFO ][o.e.c.m.MetaDataIndexTemplateService] [Tt8LRUG] adding template [firewall-1-template] for index patterns [firewall-1_]
[2019-10-21T02:00:07,907][INFO ][o.e.c.m.MetaDataCreateIndexService] [Tt8LRUG] [firewall-1_25] creating index, cause [api], templates [firewall-1-template], shards [4]/[0], mappings [message]
[2019-10-21T02:00:08,975][INFO ][o.e.c.r.a.AllocationService] [Tt8LRUG] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[firewall-1_25][2], [firewall-1_25][3], [firewall-1_25][1]] …]).
[2019-10-21T02:00:10,199][INFO ][o.e.c.m.MetaDataIndexTemplateService] [Tt8LRUG] adding template [mswinsecaudit-template] for index patterns [mswinsecaudit_]
[2019-10-21T02:00:10,206][INFO ][o.e.c.m.MetaDataCreateIndexService] [Tt8LRUG] [mswinsecaudit_13] creating index, cause [api], templates [mswinsecaudit-template], shards [4]/[0], mappings [message]
[2019-10-21T02:00:10,348][INFO ][o.e.c.r.a.AllocationService] [Tt8LRUG] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[mswinsecaudit_13][2], [mswinsecaudit_13][1], [mswinsecaudit_13][0]] …]).
[2019-10-21T02:00:10,434][INFO ][o.e.c.m.MetaDataIndexTemplateService] [Tt8LRUG] adding template [palo_alto-template] for index patterns [palo_alto_*]
[2019-10-21T02:00:10,442][INFO ][o.e.c.m.MetaDataCreateIndexService] [Tt8LRUG] [palo_alto_9] creating index, cause [api], templates [paloalto-custom-mapping, palo_alto-template], shards [4]/[0], mappings [message]
[2019-10-21T02:00:10,609][INFO ][o.e.c.r.a.AllocationService] [Tt8LRUG] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[palo_alto_9][1], [palo_alto_9][0]] …]).
I don’t see anything alarming there, either. Any further hints?
Cheers,
Tobias