Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Operator changes for local disk clusters #6942

Merged
merged 12 commits into from
Nov 15, 2022

Conversation

joejulian
Copy link
Contributor

Cover letter

To support local disk, we need to use the automatic node id and automatic bootstrap, among other things. This is the start of that change.

Each additional body of work will be in its own commit, to make review easier.

Fixes #ISSUE-NUMBER, Fixes #ISSUE-NUMBER, ...

Backport Required

  • not a bug fix
  • issue does not exist in previous branches
  • papercut/not impactful enough to backport
  • v22.2.x
  • v22.1.x
  • v21.11.x

UX changes

Describe in plain language how this PR affects an end-user. What topic flags, configuration flags, command line flags, deprecation policies etc are added/changed.

Release notes

@joejulian joejulian force-pushed the decommission branch 6 times, most recently from 4ce5b9e to d3eaa14 Compare November 12, 2022 07:30
@RafalKorepta RafalKorepta merged commit e51d5b7 into redpanda-data:dev Nov 15, 2022
@RafalKorepta
Copy link
Contributor

/backport v22.3.x

@BenPope
Copy link
Member

BenPope commented Nov 16, 2022

@joejulian Please fill out the backport section of the cover letter

RafalKorepta added a commit to redpanda-data/redpanda-operator that referenced this pull request Jun 19, 2024
Previous usage of finalizer handlers was unreliable in the case of
flipping Kubernetes Nodes ready status. Local SSD disks that could be
attached to Redpanda Pod prevents rescheduling as the Persistent Volume
affinity bounds Pod to only one Node. In case of Kubernetes Node coming
back to live Cluster controller could already delete Redpanda data (PVC
deletion and Redpanda decommissioning). If particular Redpanda Node would
host single replica partition, then it would be a data lost.

If the majority of Redpanda process would run in unstable Kubernetes
Nodes, then Redpanda operator could break whole cluster by losing Raft
quorum.

Reference

#112
redpanda-data/redpanda#6942
RafalKorepta added a commit to redpanda-data/redpanda-operator that referenced this pull request Jun 21, 2024
Previous usage of finalizer handlers was unreliable in the case of
flipping Kubernetes Nodes ready status. Local SSD disks that could be
attached to Redpanda Pod prevents rescheduling as the Persistent Volume
affinity bounds Pod to only one Node. In case of Kubernetes Node coming
back to live Cluster controller could already delete Redpanda data (PVC
deletion and Redpanda decommissioning). If particular Redpanda Node would
host single replica partition, then it would be a data lost.

If the majority of Redpanda process would run in unstable Kubernetes
Nodes, then Redpanda operator could break whole cluster by losing Raft
quorum.

Reference

#112
redpanda-data/redpanda#6942
RafalKorepta added a commit to redpanda-data/redpanda-operator that referenced this pull request Jun 21, 2024
Previous usage of finalizer handlers was unreliable in the case of
flipping Kubernetes Nodes ready status. Local SSD disks that could be
attached to Redpanda Pod prevents rescheduling as the Persistent Volume
affinity bounds Pod to only one Node. In case of Kubernetes Node coming
back to live Cluster controller could already delete Redpanda data (PVC
deletion and Redpanda decommissioning). If particular Redpanda Node would
host single replica partition, then it would be a data lost.

If the majority of Redpanda process would run in unstable Kubernetes
Nodes, then Redpanda operator could break whole cluster by losing Raft
quorum.

Reference

#112
redpanda-data/redpanda#6942
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

6 participants