You are browsing a read-only backup copy of Wikitech. The live site can be found at

Wikimedia Cloud Services team/EnhancementProposals/Operational Automation: Difference between revisions

From Wikitech-static
Jump to navigation Jump to search
imported>David Caro
imported>Andrew Bogott
Line 50: Line 50:

<syntaxhighlight lang=shell-session>
<syntaxhighlight lang=shell-session>
$ mkvirtualenv spicerack
$ mkvirtualenv -p python3 spicerack
$ pip install wikimedia-spicerack
$ pip install wikimedia-spicerack

Revision as of 18:17, 24 May 2021

We currently use Puppet to automate most of our tasks, but it has it's limitations. We still need a tool to automate, collect and review all our operational procedures. Some examples of such procedures are:

  • Adding a new member to a toolforge instance etcd cluster.
  • Bootstrapping a new toolforge instance.
  • Searching for the host where a backup is kept.
  • Provisioning a new cloudvirt node.
  • Re-image all/a set of the cloudvirt nodes.
  • Manage non-automated upgrades.
  • Take down a cloudvirt node for maintenance.

Problem statement

All these tasks still require manual operations, following a runbook whenever available, that easily get outdated, are prone to human error and require considerable attention to execute.


After reviewing several automation tools (spicerack, saltstack, puppet, ...), and doing a quick POC (see gerrit:647735 and gerrit:658637) for the two more relevant (summary of the experience here, I've decided to propose spicerack as the de-facto tool for WMCS operational tasks automation.


The main advantage of choosing spicerack is collaboration with the rest of the SRE teams. This comes with both the duty and privilege of becoming co-maintainers for spicerack and related projects, allowing us to have a say in the direction of the project and the use cases that will be supported. With the duty of driving, reviewing and maintaining the projects for all the users (including other SRE teams).


The Spicerack ecosystem is split in several projects:


Cumin is the lowermost layer, built on top of ClusterShell takes care of translating host expressions to hosts, running commands in them (using whatever strategy is selected) and returning the results.

This library should be pretty stable and require little to no changes.


Wmflib is a bundle of generic commonly used functions related to the wikimedia foundation, has some helpful decorators (ex. retry) and similar tools.

This library should be used for generic functions that are not bound to the spicerack library and can be reused in other non-spicerack wikimedia related projects.


Spicerack is the core library, contains more wikimedia specific libraries and a cli (cookbook) to interact with different services (ex. toolforge.etcd) and is meant to be used to store the core logic for any interaction with the services.

Here we will have to add, specially at the beginning, some libraries to interact with our services, here will also be where more of the re-usage of code and collaboration will happen. We should keep always in mind things here that can be used by other group around the foundation. Code in this library will be considerably tested, and no merges should happen without review.


The Spicerack/Cookbooks repo contains the main recipes to execute, the only logic should be orchestration, and any service management related code should be eventually moved to the above Spicerack library.

This repository of cookbooks will be shared with the rest of the SRE group, but our specific cookbooks will go under the cookbooks/wmcs directory. Any helper library should go under cookbooks/wmcs/ and we should periodically consider moving as much of the code from there to Spicerack.

Execution of the cookbooks

As of now, this cookbooks can be run locally, I'm actively considering how to provide a host/vm/... with a spicerack + cumin setup for easy usage and running long cookbooks, but as of right now, we can start locally.

If your cookbooks are only accessing bare metal machines, you can already run them on the cumin hosts for the wiki operations (ex. cumin1001.eqiad.wmnet), but those hosts have no access to the VMs as of writing this.

Local setup

To run locally the cookbooks, you will need to create a virtualenv and install the dependencies, for example:

$ mkvirtualenv -p python3 spicerack
$ pip install wikimedia-spicerack

Then clone the cookbooks repo (see

$ git clone "https://$"

Then create the relevant config files, one for cumin, wherever you prefer, I recommend ~/.config/spicerack/cumin_config.yaml, with the contents:

transport: clustershell
log_file: cumin.log  # feel free to change this to another path of your choosing
default_backend: direct
environment: {}
        # needed for vms that repeat a name
        - |
            -o StrictHostKeyChecking=no
            -o "UserKnownHostsFile=/dev/null"
            -o "LogLevel=ERROR"

And another for spicerack itself (the cookbook cli), I'll use ~/.config/spicerack/cookbook_config.yaml

# adapt to wherever you cloned the repo
cookbooks_base_dir: ~/Work/wikimedia/operations/cookbooks
logs_base_dir:  /tmp/spicerack_logs
    # pending a pip release: cumin_config: ~/.config/spicerack/cumin-config.yaml
    # for now you'll have to use the full path, change YOURUSER with your actual user
    cumin_config: /home/YOURUSER/.config/spicerack/cumin-config.yaml

With those config files, now you are able to run the client, from the root of the operations/cookbooks repository, you can list all the cookbooks:

$ cookbook -c ~/.config/spicerack/cookbook_config.yaml --list

NOTE: This is a very premature proposal, this workflow will be improved, feel free to start any questions/discussions in the talks page or ping me directly (David Caro (talk) 17:21, 5 February 2021 (UTC)).