Skip to content

Commit

Permalink
Merge #24377
Browse files Browse the repository at this point in the history
24377: rfcs: Finalize the cluster upgrade process automatically. r=windchan7 a=windchan7

This RFC proposes to automate the finalization process to upgrade a cluster.

Issue: #23912

Release note: None

Co-authored-by: Victor Chen <victor@cockroachlabs.com>
  • Loading branch information
craig[bot] and windchan7 committed Apr 12, 2018
2 parents bfa9931 + 2dd9c6b commit b3928c5
Showing 1 changed file with 144 additions and 0 deletions.
144 changes: 144 additions & 0 deletions docs/RFCS/20180411_finalize_cluster_upgrade_automatically.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,144 @@
- Feature Name: Finalize cluster upgrades automatically
- Status: in-progress
- Start Date: 2018 4/11
- Authors: Victor Chen
- RFC PR: [#24377](https://github.com/cockroachdb/cockroach/pull/24377)
- Cockroach Issue:
[#23912](https://github.com/cockroachdb/cockroach/issues/23912),
[#22686](https://github.com/cockroachdb/cockroach/issues/22686)

## Summary

We will add an auto-finalization feature to help operators finalize their
cluster upgrade process without manually running the set cluster version sql
command.

Meanwhile, we will also add a feature that allows operators to opt out of
auto-finalization to preserve the ability to downgrade.


## Motivation
In the past, in order to finalize the cluster upgrade process, operators must
manually run

```sql
SET CLUSTER SETTING version = crdb_internal.node_executable_version();
```

However, this step is easy to forget and without it, we cannot unlock
backwards-incompatible features. Therefore, it is necessary to add a version
check mechanism that automatically runs the finalization command once knowing
all the nodes in the cluster are running the new version.

Finally, because this finalization step is irreversible, we want to allow
operators to make their own decisions on whether to use this auto-finalization
feature or not. For example, operators who want to opt out of auto-upgrade from
`version 2.0` to its next version should run:

```sql
SET CLUSTER SETTING cluster.preserve_downgrade_option = '2.0';
```

Please note that the opt-out feature is on a per version basis. That is to say,
if operators want to keep use manual upgrade for `version 2.1`, they have to
re-run the above command with `2.0` replaced with `2.1`. Otherwise, the cluster
version will still be upgraded automatically.

If the operators change their mind after running the manual upgrade command,
they can run the following command to change it back to auto upgrade:

```sql
RESET CLUSTER SETTING cluster.preserve_downgrade_option;
```

## Detailed Design

### Add a new setting `cluster.preserve_downgrade_option`

The field should be added to `CLUSTER SETTING` in the format of:
`cluster.preserve_downgrade_option = 2.0`

By default, the value should be `""`, an empty string, to indicate we are always
using auto upgrade.

We should always validate the value being stored in this field. It should not
exceed the maximum version we have; nor should it fall below user's current
cluster version.

### On node startup, have a daemon that runs the following logic

- Check setting `cluster.preserve_downgrade_option`.
- **IF**:
- Did not preserve downgrade options for current cluster version.
<br>**AND**
- Look at `crdb_internal.gossip_nodes` and verify that all nodes are running
the new version.
<br>**AND**
- Look at `NodeStatusKeys` and verify that all non-decommissioned nodes are
in `gossip_nodes` (no missing nodes).
- **THEN**:
- Do upgrade:
```
SET CLUSTER SETTING version = crdb_internal.node_executable_version();
```

- Run `RESET CLUSTER SETTING cluster.preserve_downgrade_option;`.
- **ELSE**
- Exit and abort upgrade.


## Testing
Use the existing ``pkg/cmd/roachtest/version.go`` as a basis.

Test Steps:

1. Run `SET CLUSTER SETTING cluster.preserve_downgrade_option = '2.0';`

2. Perform a rolling upgrade.

3. Sleep and check `cluster version` is still the old version.

4. Run `RESET CLUSTER SETTING cluster.preserve_downgrade_option;`

5. Sleep and check `cluster version` is bumped to the new version.


## UI
- If the user opt out of the auto upgrade, we should have a banner that alert
users when all the nodes are running a higher version than the cluster version
to instruct them to manually run the upgrade command.

- If all live nodes are running the newest version but some nodes are down and
not decommissioned, we should have a banner to alert operators to either revive
or decommission the down nodes.

- Have an API that returns if the auto-upgrade is disabled or not.
- If `cluster.preserve_downgrade_option` is equal to `cluster.version`,
auto-upgrade is disabled. Otherwise it's enabled.
- If auto-upgrade is enabled, have an API that returns if the auto upgrade has
started or not.
- If all nodes have the same `node_executable_version`, auto-upgrade has started.
- If auto-upgrade is enabled, have an API that returns if the auto upgrade has
finished or not.
- If all node's `node_executable_version` is the same as `cluster.version`,
auto-upgrade has finished.


## Drawbacks
- Race condition
[#24670](https://github.com/cockroachdb/cockroach/issues/24670) (future work):
- Say an operator want to upgrade a cluster of 3 nodes from version 1.1 to
2.0.
- As soon as the operator just replaced the last node running v1.1 with a
node running v2.0, the auto-finalization will start running the set cluster
version command.
- However, if right before the command get executed, the operator erroneously
joined another node running 1.1 version. Then the command get executed and
the cluster version is set to be 2.0, with new features enabled.
- But the node running 1.1 version can not perform some of the 2.0 version
operations and it will cause trouble.
- Unreachable Node:
- We don't guarantee the upgrade will always be successful if there're
unreachable nodes.
- If a node is unreachable, we won't ever upgrade because it could be unsafe
if the unreachable node is running in a lower version.

0 comments on commit b3928c5

Please sign in to comment.