You are browsing a read-only backup copy of Wikitech. The primary site can be found at

Incident documentation/20190723-logstash: Difference between revisions

From Wikitech-static
Jump to navigation Jump to search
m (lang="text")
(5 intermediate revisions by 2 users not shown)
Line 1: Line 1:
'''document status''': {{irdoc-review}}
#REDIRECT [[Incident documentation/2019-07-23 logstash]]
Logstash became overloaded during a network outage that caused elevated error rate.  Problematic log messages were also observed, which were able to crash logstash, requiring manual filtering of the problem logs.
Logstash was unable to process incoming logs on all inputs until the the problem log type was identified and filtered.  This resulted in delayed logs, and missing logs of the affected type (MediaWiki's SlowTimer)
* 19:10 - First page (logstash failure, secondary fallout from error ingestion) - A flood of errors caused by network disruption during eqiad rack a6/a7 pdu maintenance overwhelmed logstash.
** troubleshooting of related network issue proceeds with higher priority.  logstash outage presumed to be secondary effect of network issues.
* 20:10 network issue begins to improve, however logstash does not recover on its own
* 20:23 logstash unable to process backlog and is crashing due to UTF-8 parsing error  Suspected message example:
**<syntaxhighlight lang="text">
[FATAL][logstash.runner          ] An unexpected error occurred! {:error=>#<ArgumentError: invalid byte sequence in UTF-8>
* 21:00 UTF-8 issue traced to kafka rsyslog-shipper input.  Logstash collectors restarted with kafka rsyslog-shipper input disabled.  Kafka-logging consumer lag begins to recover on non-rsyslog topics
* 21:45 UTF-8 issue traced to SlowTimer messages and a temporary fix to drop [message] =~ /^SlowTimer/ deployed
* 22:00 temporary fix is active across eqiad/codfw logstash collectors, kafka rsyslog-shipper input re-enabled.  Kafka consumer lag is recovering now on all topics
* 00:16 (next day) logstash has caught up with kafka backlog
* Get to the bottom of invalid utf8 sequences being produced to logstash. Errors were traced to SlowTimer events logging binary data, in the form of slow parser cache queries. For example <nowiki></nowiki> (TODO: Create task)
* Investigate logstash rate limiting options for logstash to avoid ingesting millions of the same error message in failures like this (TODO: Create task)
* Look into separating logstash pipelines (TODO: discuss)
* Investigate filtering/parsing options to avoid UTF-8 error observed
** <nowiki></nowiki>
** <nowiki></nowiki>

Revision as of 19:14, 20 October 2021