From Fedora Project Wiki

(tweak telia position)
(redirect page to new infra-docs)
(7 intermediate revisions by one other user not shown)
Line 2: Line 2:
Every once in a while, we need to apply mass upgrades to our servers for various security and other upgrades.
== Contact Information ==
This SOP has moved to the fedora Infrastructure SOP git repo. Please see the current document at:
Owner: Fedora Infrastructure Team
Contact: #fedora-admin, sysadmin-main,, #fedora-noc
For changes, questions or comments, please contact anyone in the Fedora Infrastructure team.  
Location: All over the world.
Servers: all
Purpose: Apply kernel/other upgrades to all of our servers
== Preparation ==
# Determine which host group you are going to be doing updates/reboots on.
## Group "A" are servers that end users will see or note being down and anything that depends on them.
## Group "B" are servers that contributors will see or note being down and anything that depends on them.
## Group "C" are servers that infrastructure will notice are down, or are redundent enough to reboot some with others taking the load.
# Appoint an 'Update Leader' for the updates.
# Follow the [[Outage Infrastructure SOP]] and send advance notification to the appropriate lists.  Try to schedule the update at a time when many admins are around to help/watch for problems and when impact for the group affected is less. Do NOT do multiple groups on the same day if possible.
# Plan an order for rebooting the machines considering two factors:
#* Location of systems on the kvm or xen hosts. [You will normally reboot all systems on a host together]
#* Impact of systems going down on other services, operations and users.  Thus since the database servers and nfs servers are the backbone of many other systems, they and systems that are on the same xen boxes would be rebooted before other boxes.
# To aid in organizing a mass upgrade/reboot with many people helping, it may help to create a checklist of machines in a gobby document.
# Schedule downtime in nagios.
# Make doubly sure that various app owners are aware of the reboots
== Staging ==
Any updates that can be tested in staging or a pre-production environment should be tested there first.  Including new kernels, updates to core database applications / libraries.  Web applications, libraries, etc.
== Special Considerations ==
While this may not be a complete list, here are some special things that must be taken into account before rebooting certain systems:
Before the following machines are rebooted, all koji builders should be disabled and all running jobs allowed to complete:
* db04
* nfs01
* kojipkgs01
Builders can be removed from koji, updated and re-added. Use:
koji disable-host NAME
koji enable-host NAME
(note: you must be a koji admin).
The following machines require post-boot actions (mostly entering passphrases).  Make sure admins that have the passphrases are on hand for the reboot:
* backup-2 (LUKS passphrase on boot)
* sign-vault01 (NSS passphrase for sigul service)
* sign-bridge01 (NSS passphrase for sigul bridge service)
There is currently an autoqa01.c host on cnode01. Check with QA folks before rebooting this guest/host.
== Update Leader ==
Each update should have a Leader appointed. This person will be in charge of doing any read-write operations, and delegating to others to do tasks. If you aren't specficially asked by the Leader to reboot or change something, please don't. The Leader will assign out machine groups to reboot, or ask specific people to look at machines that didn't come back up from reboot or aren't working right after reboot. It's important to avoid multiple people operating on a single machine in a read-write manner and interfering with changes.
== Group A reboots ==
Group A machines are end user critical ones. Outages here should be planned at least a week in advance and announced to the announce list.
List of machines currently in A group (note: this is going to be automated):
(due to being on the same virt host as above)
== Group B reboots ==
This Group contains machines that contributors use. Announcements of outages here should be at least a week in advance and sent to the devel-announce list.
(due to being on the same virt host as one of above)
== Group C reboots ==
Group C are machines that infrastructure uses, or can be rebooted in such a way as to continue to provide services to others via multiple machines.
Outages here should be announced on the infrastructure list.
Group C hosts that have proxy servers on them:
<pre> (check with QA before rebooting this host/guest)
(disable each builder in turn, update and reenable).
== Doing the upgrade ==
If possible, system upgrades should be done in advance of the reboot (with relevant testing of new packages on staging).  To do the upgrades, make sure that the Infrastructure RHEL repo is updated as necessary to pull in the new packages ([[Infrastructure Yum Repo SOP]])
On puppet1, as root run:
func-yum [--host=hostname] update
--host can be specified multiple times and takes wildcards.
pinging people as necessary if you are unsure about any packages.
Additionally you can see which machines still need rebooted with:
sudo func-command --timeout=10 --oneline /usr/local/bin/ | grep yes
== Doing the reboot ==
In the order determined above, reboots will usually be grouped by the virtualization hosts that the servers are on.
You can see the guests per virt host on puppet01 in /var/log/virthost-lists.out
For each host you will want to:
* connect and verify no one is logged in and using it. If they are contact them to log off, etc
* grep default /etc/grub.conf # make sure that the kernel you upgraded to will be the one rebooted.
* shutdown -h now
This is also a good time to double check that each guest you are starting up is set to be restarted on reboot of the virt host.
== Aftermath ==
# Make sure that everything's running fine
# Reenable nagios notification as needed
# Make sure to perform any manual post-boot setup (such as entering passphrases for encrypted volumes)
# Close outage ticket.
[[Category:Infrastructure SOPs]]
[[Category:Infrastructure SOPs]]

Revision as of 18:28, 19 December 2011

Infrastructure InfrastructureTeamN1.png

This SOP has moved to the fedora Infrastructure SOP git repo. Please see the current document at:

For changes, questions or comments, please contact anyone in the Fedora Infrastructure team.