Skip to content

Latest commit

 

History

History
65 lines (47 loc) · 5.15 KB

CONTRIBUTING.md

File metadata and controls

65 lines (47 loc) · 5.15 KB

How to Contribute

This project is Apache 2.0 licensed and accepts contributions via GitHub pull requests.

Contributing Guidelines

  • Unit tests are required to accompany all logic being introduced into the pkg/ directory.
  • Functional tests are required to accompany any logic that modifies pkg/apis/ API structs.
  • Follow the tide review flow for merging commits. This flow forces a review and all required CI jobs to pass before a PR is merged.
  • Read and understand the design principles outlined in this document that pertain to the code you are contributing to.

Certificate of Origin

By contributing to this project you agree to the Developer Certificate of Origin (DCO). This document was created by the Linux Kernel community and is a simple statement that you, as a contributor, have the legal right to make the contribution. See the DCO file for details.

Reconcile Loop Design Principles

  • Never block reconcile execution. ever. This means no sleeps and no retry loops
  • All logic must be idempotent. This means discovering what has already occurred by inspecting informer cache, and only mutating to converge on desired state.
  • Never perform an apiserver List() request within the reconcile loop. This action is very apiserver intensive. Instead use informers for retrieving items from cache.
  • Avoid mutating a cluster object the loop didn't create. A exception to this would to be adding annotations to an existing object. Mutating a object's Spec that our controller didn't create should be avoided.
  • Avoid performing an apiserver Get() request within the reconcile loop if at all possible. Again use informers. There are rare exceptions.
  • Use finalizers in order to perform any cleanup tasks required before a CR is completely removed from etcd.
  • Use owner references on any objects created and owned by a CR. Owner references ensure objects are automatically garbage collected after a corresponding CR is deleted.
  • Return early to handle errors which re-enqueues the work queue key to be tried again.
  • Do not create logic to sync a key using a recurring time interval. Instead only wake up processing the key when something has changed.
  • Reconcile execution should be broken into two parts, synchronization and status reporting. Synchronization converges the CR on the desired state. Status reporting records the current state of the CR based on the synchronization execution results and the observed state of the cluster. These are two independent functions. Keep them separated. Do not modify status during the synchronization step.

API Design

  • All workflows must be declarative. This means the cluster-admin must be able to declare the exact state they want up front, and have our operator handle converging on that state.
  • No imperative actions. This means no requirement for a cluster-admin to account for the ordering or timing at which manifests are posted. If our API can't express the desired state up front, then the API is wrong.
  • Use Status.Conditions and Events to represent and record transient states that a CR exists in.
  • Avoid introducing Status.Phase enums if possible. The pitfall of a Phase enum is that it often gets treated as a FSM, when in reality that isn't always the case. Source

Unit Testing and Coverage

Unit testing allows us to validate our reconcile loops under specific conditions to ensure no unexpected changes in behavior are introduced.

  • Use mock clients to simulate reconcile loops.
  • Execute reconcile loop to validate execution under specific conditions.
  • Validate reconcile execution by introspecting API calls that were made during the execution. Example

Unit tests can be executed with make unittests. This will also create unit test coverage reports, their location will be reported. Every PR will also run unit tests with coverage, and the status will be reported on the PR. Make sure the coverage doesn't decrease, by writing unit tests for new code. PRs for increasing the coverage for existing code are very welcome, too. The current coverage on master and release branches can also be seen at coveralls.io.

Functional Testing

Functional testing allows us to exercise and validate our API on a live cluster.

  • All functional tests should be written in golang+gomega and exist in the functtests/ directory
  • Functional test scenarios should represent conditions cluster-admins will encounter.
  • Exercise and validate behavior by posting/mutating/deleting manifests and observing the desired state is reached.
  • As a general rule of thumb for test condition timeouts, take the time you typically observe an action taking and double it for the timeout value.

CI

On every Pull Request the Openshift CI will run some source code validation checks and the performance operator unit and functional tests. Please check the CI logs in case of a failed test. If you need help, don't hesitate to ping one of the maintainers on the PR.