This module contains a set of tasks and custom facts to allow the automation of and reporting on operating system patching. Currently, patching works on Linux (Redhat, Suse and Debian derivatives) and Windows (Server 2008 through to 2019 have been tested).
Under the hood, it uses the OS level tools or APIs to carry out the actual patching. That does mean that you need to be sure that your nodes can search for their updates using the native tool - e.g. You still need to manage the configuration of YUM, APT, Zypper or Windows Update.
Note - Windows systems require at least PowerShell version 3.0. If you are intending to update an unpatched Windows system before Server 2012, you will need to update PowerShell first.
The wiki contains some useful background information on the module and how it works.
Puppet Enterprise tasks and Bolt have opened up methods to integrate operating system level patching into the puppet workflow. Providing automation of patch execution through tasks and the robust reporting of the state through custom facts and PuppetDB.
If you're looking for a simple way to report on your OS patch levels, this module will show all outstanding updates, including which are related to security updates. Do you want to enable self-service patching? This module will use Puppet's RBAC and orchestration and task execution facilities to give you that power.
It also uses security metadata (where available) to determine if there are security updates. On Redhat, this is provided from Redhat as additional metadata in YUM. On Debian, checks are done for which repo the updates are coming from. On Windows, this information is provided by default. There is a parameter to the os_patching::patch_server task to only apply security updates.
Blackout windows enable the support for time-based change freezes where no patching can happen. There can be multiple windows defined and each will automatically expire after reaching the defined end date.
The module provides an additional fact (os_patching
) and has a task to allow the patching of a server. When the os_patching
manifest is added to a node it installs a script and cron job (Linux) or a scheduled task (Windows) to check for available updates and generate cache data used by the os_patching
fact.
Install the module using the Puppetfile, include it on your nodes and then use the provided tasks to carry out patching.
Include the module:
include os_patching
More advanced usage:
class { 'os_patching':
patch_window => 'Week3',
blackout_windows => { 'End of year change freeze':
{
'start': '2018-12-15T00:00:00+1000',
'end': '2019-01-15T23:59:59+1000',
}
},
}
In that example, the node is assigned to a "patch window", will be forced to reboot regardless of the setting specified in the task and has a blackout window defined for the period of 2018-12-15 - 2019-01-15, during which time no patching through the task can be carried out.
Run a basic patching task from the command line:
os_patching::patch_server - Carry out OS patching on the server, optionally including a reboot and/or only applying security-related updates
USAGE:
$ puppet task run os_patching::patch_server [dpkg_params=<value>] [reboot=<value>] [security_only=<value>] [timeout=<value>] [yum_params=<value>] <[--nodes, -n <node-names>] | [--query, -q <'query'>]>
PARAMETERS:
- dpkg_params : Optional[String]
Any additional parameters to include in the dpkg command
- reboot : Optional[Variant[Boolean, Enum['always', 'never', 'patched', 'smart']]]
Should the server reboot after patching has been applied? (Defaults to "never")
- security_only : Optional[Boolean]
Limit patches to those tagged as security-related? (Defaults to false)
- timeout : Optional[Integer]
How many seconds should we wait until timing out the patch run? (Defaults to 3600 seconds)
- yum_params : Optional[String]
Any additional parameters to include in the yum upgrade command (such as including/excluding repos)
Example:
$ puppet task run os_patching::patch_server --params='{"reboot": "patched", "security_only": false}' --query="inventory[certname] { facts.os_patching.patch_window = 'Week3' and facts.os_patching.blocked = false and facts.os_patching.package_update_count > 0}"
This will run a patching task against all nodes which have facts matching:
os_patching.patch_window
of 'Week3'os_patching.blocked
equalsfalse
os_patching.package_update_count
greater than 0
The task will apply all patches (security_only=false
) and will reboot the node after patching (reboot=true
).
Most of the reporting is driven by the custom fact os_patching_data
, for example:
# facter -p os_patching
{
package_update_count => 0,
package_updates => [],
security_package_updates => [],
security_package_update_count => 0,
blocked => false,
blocked_reasons => [],
blackouts => {},
patch_window = 'Week3',
pinned_packages => [],
last_run => {
date => "2018-08-07T21:55:20+10:00",
message => "Patching complete",
return_code => "Success",
post_reboot => "false",
security_only => "false",
job_id => "60"
}
reboots => {
reboot_required => false,
apps_needing_restart => { },
app_restart_required => false
}
}
This shows there are no updates that can be applied to this server and the server doesn't need a reboot or any application restarts. When there are updates to add, you will see similar to this:
# facter -p os_patching
{
package_update_count => 6,
package_updates => [
"kernel.x86_64",
"kernel-tools.x86_64",
"kernel-tools-libs.x86_64",
"postfix.x86_64",
"procps-ng.x86_64",
"python-perf.x86_64"
]
security_package_updates => [],
security_package_update_count => 0,
blocked => false,
blocked_reasons => [],
blackouts => {
Test change freeze 2 => {
start => "2018-08-01T09:17:10+1000",
end => "2018-08-01T11:15:50+1000"
}
},
pinned_packages => [],
patch_window = 'Week3',
last_run => {
date => "2018-08-07T21:55:20+10:00",
message => "Patching complete",
return_code => "Success",
post_reboot => "false",
security_only => "false",
job_id => "60"
}
reboots => {
reboot_required => true,
apps_needing_restart => {
630 => "/usr/sbin/NetworkManager --no-daemon ",
1451 => "/usr/bin/python2 -s /usr/bin/fail2ban-server -s /var/run/fail2ban/fail2ban.sock -p /var/run/fail2ban/fail2ban.pid -x -b ",
1232 => "/usr/bin/python -Es /usr/sbin/tuned -l -P "
},
app_restart_required => true
}
}
Where it shows 6 packages with available updates, along with an array of the package names. None of the packages are tagged as security-related (requires Debian, a subscription to RHEL or a Windows system). There are no blockers to patching and the blackout window defined is not in effect.
The reboot_required flag is set to true, which means there have been changes to packages that require a reboot (libc, kernel, etc) but a reboot hasn't happened. The apps_needing_restart shows the PID and command line of applications that are using files that have been upgraded but the process hasn't been restarted.
The pinned packages entry lists any packages which have been specifically excluded from being patched, from version lock on Red Hat or by pinning in Debian.
The last run shows a summary of the information from the last os_patching::patch_server
task.
The fact os_patching.patch_window
can be used to assign nodes to an arbitrary group. The fact can be used as part of the query fed into the task to determine which nodes to patch:
$ puppet task run os_patching::patch_server --query="inventory[certname] {facts.os_patching.patch_window = 'Week3'}"
You can use the parameter os_patching::pre_patching_command
to supply a file name to be run before running the patch job. The file must be executable and should exit with a return code of 0
if the command was successful.
The entry must be a single command, with no arguments or parameters.
The logic for how to handle reboots is a little complex as it has to handle a wide range of scenarios and desired outcomes.
There are two options which can be set that control how the reboot decision is made:
The reboot parameter is set in the os_patching::patch_server
task. It takes the following options:
- "always"
- No matter what, always reboot the node during the task run, even if no patches are required
- "never" (or the legacy value
false
)- No matter what, never reboot the node during the task run, even if patches have been applied
- "patched" (or the legacy value
true
)- Reboot the node if patches have been applied
- "smart"
- Use the OS supplied tools (e.g.
needs_restarting
on RHEL, or a pending reboot check on Windows) to determine if a reboot is required, if it is, then reboot the machine, otherwise do not.
- Use the OS supplied tools (e.g.
The default value is "never".
These parameters set the default action for all nodes during the run of the task. It is possible to override the behavior on a node by using...
The reboot override fact is part of the os_patching
fact set. It is set through the os_patching manifest and has a default of "default".
If it is set to "default" it will take whatever reboot actions are listed in the os_patching::patch_server
task. The other options it takes are the same as those for the reboot parameter (always, never, patched, smart).
During the task run, any value other than "default" will override the value for the reboot
parameter. For example, if the reboot
parameter is set to "never" but the reboot_override
fact is set to "always", the node will always reboot. If the reboot
parameter is set to "never" but the reboot_override
fact is set to "default", the node will use the reboot
parameter and not reboot.
By having a reboot mode set by the task parameter, it is possible to set the behavior for all nodes in a patching run (I do 100's at once). Having the override functionality provided by the fact, you can allow individual nodes included in the patching run excluded from the reboot behavior. Maybe there are a couple of nodes you know you need to patch but you can't reboot them immediately, you can set their reboot_override fact to "never" and handle the reboot manually at another time.
If there is nothing to be done, the task will report:
{
"pinned_packages" : [ ],
"security" : false,
"return" : "Success",
"start_time" : "2018-08-08T07:52:28+10:00",
"debug" : "",
"end_time" : "2018-08-08T07:52:46+10:00",
"reboot" : "never",
"packages_updated" : "",
"job_id" : "",
"message" : "No patches to apply"
}
If patching was executed, the task will report similar to below:
{
"pinned_packages" : [ ],
"security" : false,
"return" : "Success",
"start_time" : "2018-08-07T21:55:20+10:00",
"debug" : "TRIMMED DUE TO LENGTH FOR THIS EXAMPLE, WOULD NORMALLY CONTAIN FULL COMMAND OUTPUT",
"end_time" : "2018-08-07T21:57:11+10:00",
"reboot" : "never",
"packages_updated" : [ "NetworkManager-1:1.10.2-14.el7_5.x86_64", "NetworkManager-libnm-1:1.10.2-14.el7_5.x86_64", "NetworkManager-team-1:1.10.2-14.el7_5.x86_64", "NetworkManager-tui-1:1.10.2-14.el7_5.x86_64", "binutils-2.27-27.base.el7.x86_64", "centos-release-7-5.1804.el7.centos.2.x86_64", "git-1.8.3.1-13.el7.x86_64", "gnupg2-2.0.22-4.el7.x86_64", "kernel-tools-3.10.0-862.3.3.el7.x86_64", "kernel-tools-libs-3.10.0-862.3.3.el7.x86_64", "perl-Git-1.8.3.1-13.el7.noarch", "python-2.7.5-68.el7.x86_64", "python-libs-2.7.5-68.el7.x86_64", "python-perf-3.10.0-862.3.3.el7.centos.plus.x86_64", "selinux-policy-3.13.1-192.el7_5.3.noarch", "selinux-policy-targeted-3.13.1-192.el7_5.3.noarch", "sudo-1.8.19p2-13.el7.x86_64", "yum-plugin-fastestmirror-1.1.31-45.el7.noarch", "yum-utils-1.1.31-45.el7.noarch" ],
"job_id" : "60",
"message" : "Patching complete"
}
If patching was blocked, the task will report similar to below:
Error: Task exited: 100
Patching blocked
A summary of the patch run is also written to /var/cache/os_patching/run_history
, the last line of which is used by the os_patching.last_run
fact.
2018-08-07T14:47:24+10:00|No patches to apply|Success|false|false|
2018-08-07T14:56:56+10:00|Patching complete|Success|false|false|121
2018-08-07T15:04:42+10:00|yum timeout after 2 seconds : Loaded plugins: versionlock|1|||
2018-08-07T15:05:51+10:00|yum timeout after 3 seconds : Loaded plugins: versionlock|1|||
2018-08-07T15:10:16+10:00|Patching complete|Success|false|false|127
2018-08-07T21:31:47+10:00|Patching blocked |100|||
2018-08-08T07:53:59+10:00|Patching blocked |100|||
Each system with the os_patching class applied will have several files and a directory managed by the manifest.
The script used to scan for updates and generate the fact data is stored in the following location based on the OS type:
- Linux -
/usr/local/bin/os_patching_fact_generation.sh
- Windows -
c:\ProgramData\os_patching\os_patching_fact_generation.ps1
The os_patching directory contains the various control files needed for this module and its tasks to work correctly. The locations are as follows:
- Linux -
/var/cache/os_patching
- Windows -
c:\ProgramData\os_patching
The following files are stored in this directory:
blackout_windows
: contains name, start and end time for all blackout windowspackage_updates
: a list of all package updates available, populated byos_patching_fact_generation.sh
(Linux) oros_patching_fact_generation.ps1
(Windows), triggered through cron (Linux) or task scheduler (Windows)security_package_updates
: a list of all security_package updates available, populated byos_patching_fact_generation.sh
(Linux) oros_patching_fact_generation.ps1
(Windows), triggered through cron (Linux) or task scheduler (Windows)run_history
: a summary of each run of theos_patching::patch_server
task, populated by the taskreboot_override
: if present, overrides thereboot=
parameter to the taskpatch_window
: if present, sets the value for the factos_patching.patch_window
reboot_required
: if the OS can determine that the server needs to be rebooted due to package changes, this file contains the result. Populates the fact reboot.reboot_required.apps_to_restart
: (Linux only) a list of processes (PID and command line) that haven't been restarted since the packages they use were patched. Sets the fact reboot.apps_needing_restart and .reboot.app_restart_required.
Except for the run_history file and Windows os_patching scripts, all files in the os_patching directory will be regenerated after a puppet run and a run of the os_patching_fact_generation.sh
or os_patching_fact_generation.ps1
script, which runs every hour by default. If run_history is removed, the same information can be obtained from PDB, apt/yum, syslog or the Windows event log.
As Windows includes no native command line tools to manage update installation, PowerShell scripts have been written utilizing the Windows Update agent APIs that handle the update search, download, and installation process:
os_patching_fact_generation.ps1
which scans for updates and generates fact data (as above)os_patching_windows.ps1
is utilised by thepatch_server
task and underlying ruby script to handle the update installation process
Windows Server 2008 (x86 and x64) through to 2019 have been tested. The code should also function on the equivalent client versions of Windows (e.g. Vista and newer), however, this has not been thoroughly tested.
This module does not handle the configuration of the update source or any of the other Windows Update settings - it simply triggers a search (fact generation) or search, download and install (patch_server task). It is recommended to use the puppetlabs wsus_client module to configure the following options:
- WSUS server if you are using one (although this is not strictly required)
- Set the mode to automatically download updates and notify for install (
AutoNotify
)
For example:
class { 'wsus_client':
server_url => 'http://my-wsus-server.internal:8530', # WSUS Server
enable_status_server => true # Send status to WSUS too
auto_update_option => 'AutoNotify', # automatically download updates and notify for install
}
-
RedHat 5 based systems have support but lack a lot of the yum functionality added in 6, so things like the upgraded package list and job ID will be missing.
-
PowerShell version 3.0 or newer is required on Windows Systems.
-
If updates or packages are installed outside of this script (e.g. by a user or another automated process), the results will not be captured in the facts.
-
On Windows systems, the timeout parameter of the
patch_server
task is implemented as a maintenance window end time (e.g. start time + timeout). This is used by doing a calculation before installing each update. If there is insufficient available, the update run will stop. However, at this stage, each update is estimated to take 5 minutes to install. This will be improved in a future release to perform an estimation based on update size or type (e.g. an SCCM-like 5 minutes for a hotfix, 30 minutes for a cumulative update).
Fork, develop, submit a pull request