You are browsing a read-only backup copy of Wikitech. The live site can be found at

Incident documentation/20200319-parsercache: Difference between revisions

From Wikitech-static
Jump to navigation Jump to search
Line 1: Line 1:
#REDIRECT [[Incident documentation/2020-03-19 parsercache]]
'''document status''': {{irdoc-review}} <!--
The status field should be one of:
* {{tl|irdoc-draft}} - Initial status. When you're happy with the state of your draft, change it to {{tl|irdoc-review}}.
* {{tl|irdoc-review}} - The incident review working group will contact you then to finalise the report. See also the steps on [[Incident documentation]].
* {{tl|irdoc-final}}
== Summary ==
Parsercache databases got overloaded due to a malfunctioning host which resulted on spikes of connections on the other 2 active hosts and increased latency on our mwapps servers.
=== Impact ===
* Query latency was increased
[[File:Application-servers-red-dashboard (1).png|thumb]]
* mw app servers got their workers saturated:
* Higher than usual response time
[[File:High response time.png|thumb]]
=== Detection ===
Icinga paged for pc1008 host that was having performance degradation
18:43:14 <+icinga-wm> PROBLEM - MariaDB Slave SQL: pc2 #page on pc1008 is CRITICAL: CRITICAL slave_sql_state could not connect
== Timeline ==
'''All times in UTC.'''
* 18:00 ''' Degradation begins'''
* 18:00 pc1008 starts having performance issues and its disk latency starts increasing, connections start to pile up on pc1008
[[File:Mysql (14).png|thumb]]
[[File:Mysql (13).png|thumb]]
* 18:00 Other hosts (pc1007 and pc1009) also start suffering more idle connections as the result of pc1008 failing to handle connections as fast as usual
[[File:Mysql (15).png|thumb]]
* 18:00 Average response time increases
[[File:High response time.png|thumb]]
* 18:43  ''<+icinga-wm> PROBLEM - MariaDB Slave SQL: pc2 #page on pc1008 is CRITICAL: CRITICAL slave_sql_state could not connect''
* 18:43-19:44 A number of SREs and 2 DBAs respond and troubleshooting starts
* 19:11 DBAs Replace pc1008 with pc1010 (which is a spare for a different pc group, and has 1/3 of the key), but worth trying as there were no more ideas and pc1008 was checked for HW errors, misconfigurations and such and all looked fine anyways.
* 19:12 Response time, idle connections on other hosts, latency...they all start to get better
* 19:24 Values almost around the same before the incident (considering that 1/3 of the pc keys were gone)
* 19:24 '''Degradation stops'''
== Conclusions ==
The hardware performance degradation was hard to detect via the usual checks: broken BBU, degraded RAID, disks with errors that hasn't removed from the RAID, memory issues....
As nothing appeared to be broken, DBAs didn't consider pc1008 as the core of the issue.
The fact that all the parsercache showed similar connections spike pattern made us think that the problem was on the other side of the spectrum (MW).
We later learned thanks to Brad, that parsercache has a "double write" behaviour we didn't know of and if one of those fails, the others keep hanging until the request is processed or shutdown.
=== What went well? ===
* When we planned the parsercache refresh a year ago, we decided to buy a host to have it as a spare, precisely for these kind of situations.
=== What went poorly? ===
* DBAs were not aware of this parsercache behaviour so they didn't consider pc1008 affecting other host as a possibility (later explained by Brad on|T247788#5975667):
Each write to ParserCache sets two keys into the backend, which will probably get sharded to two different servers. Once SqlBagOStuff opens a connection to one of the servers, it keeps
that connection open until request shutdown. So if we assume that pc1008 is somehow failing in a way that has connections hang open for a while, we'd also see a smaller increase in
idle open connections on pc1007 and pc1009 for the cases where ParserCache's first write goes to pc1007/pc1009 and the second one goes to pc1008. That seems consistent with what the
three graphs show.
* Trying to get ahold of CPT via IRC wasn't possible.
* The hardware degradation pc1008 had, was hard to detect and was only detected a day after, with lots of testing (
=== Where did we get lucky? ===
* Just to try things, we decided to replace pc1008 with pc1010 but without much expectations and it worked
=== How many people were involved in the remediation? ===
* 2 DBAs
* 3 SREs
* 2 WMDE Devs
== Links to relevant documentation ==
This explanation by Brad resumes what was happening from MW side and
== Actionables ==
* [RFC] improve parsercache replication, sharding and HA:
* Investigate pc1008 for possible hardware issues / performance under high load:
** Once pc1008 is back full - repool it to make sure it is fully fixed after re-creating the raid
** Purge pc1010 old rows once it is out of rotation
* Parsercache sudden increase of connections:
{{#ifeq:{{SUBPAGENAME}}|Report Template||
[[Category:Incident documentation]]

Revision as of 19:27, 31 March 2021