You are browsing a read-only backup copy of Wikitech. The live site can be found at wikitech.wikimedia.org

Incidents/2023-02-28 GitLab data loss

From Wikitech-static
Jump to navigation Jump to search

document status: draft

Summary

Incident metadata (see Incident Scorecard)
Incident ID 2023-02-28 GitLab data loss Start 2022-02-28 00:04:00
Task T329931 End 2022-02-28 02:20:00
People paged 0 Responder count
Coordinators Affected metrics/SLOs
Impact Data loss on GitLab production host for one and a half hours, reduced availability for 20 minutes

GitLab was switched to the other data center as a planned maintenance in T329931. During the switchover some configuration had to be changed depending on the instance state (production or replica). The daily restore job was not disabled on the new production host in codfw, resulting in a backup being restored on the production host. So all actions between the backup (Feb 28th 00:04:00) and the restore (Feb 28th 02:00:00) are lost.

Timeline

All times in UTC.

  • 27 Feb 10:00 gitlab1004 is switched over to gitlab2002 as part of the planned maintenance
  • 27 Feb 12:00 maintenance is over
  • 28 Feb 00:04 backup is triggered on gitlab2002 - Incident begins
  • 28 Feb 02:00 restore is triggered on gitlab2002
  • 28 Feb 02:03 monitoring "probe down" for production host: T330717
  • 28 Feb 02:20: restore is finished - Incident ends

Detection

The issue was detected by automatic monitoring and task creation (probeDown) in https://phabricator.wikimedia.org/T330717

    alertname: ProbeDown
    instance: gitlab2002:443
    job: probes/custom
    prometheus: ops
    severity: task
    site: codfw
    source: prometheus
    team: serviceops-collab

Conclusions

What went well?

  • automatic alerting caught the restore on the production host (probe down)
  • production instance was working normally after the restore (beside data loss)
  • rsync server was disabled on the production host, preventing restore of replica-backups (which would cause more data loss)

What went poorly?

  • Manual configuration changes for the restore timer (profile::gitlab::enable_restore) were not adjusted on the new host
  • It was assumed backup is handled automatically, depending on the instance status similar to backups
  • Timers were not double checked after the switchover

Where did we get lucky?

  • Only one and a half hours of data loss, could be even worse if backups from the replica would have been restored
  • Low usage during UTC night, no user reported issues
    • One connection over SSH during that time at: Feb 28 01:48:25
    • two merges/other actions in two other projects

Links to relevant documentation

Actionables

Scorecard

Incident Engagement ScoreCard
Question Answer

(yes/no)

Notes
People Were the people responding to this incident sufficiently different than the previous five incidents? yes Yes in general, no for GitLab specifically, but that's to be expected in this case.
Were the people who responded prepared enough to respond effectively yes
Were fewer than five people paged? yes
Were pages routed to the correct sub-team(s)? no
Were pages routed to online (business hours) engineers?  Answer “no” if engineers were paged after business hours. no
Process Was the "Incident status" section atop the Google Doc kept up-to-date during the incident? no No Google doc, the response was tracked in a task only.
Was a public wikimediastatus.net entry created? no
Is there a phabricator task for the incident? yes
Are the documented action items assigned? yes
Is this incident sufficiently different from earlier incidents so as not to be a repeat occurrence? yes
Tooling To the best of your knowledge was the open task queue free of any tasks that would have prevented this incident? Answer “no” if there are

open tasks that would prevent this incident or make mitigation easier if implemented.

yes
Were the people responding able to communicate effectively during the incident with the existing tooling? yes
Did existing monitoring notify the initial responders? yes The monitoring was not meant for this specific failure, but worked nonetheless
Were the engineering tools that were to be used during the incident, available and in service? yes
Were the steps taken to mitigate guided by an existing runbook? no
Total score (count of all “yes” answers above) 10