Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

🌱 fix readme for in-memory provider #8905

Conversation

fabriziopandini
Copy link
Member

What this PR does / why we need it:
Fix readme for the in-memory provider

@k8s-ci-robot k8s-ci-robot added the cncf-cla: yes Indicates the PR's author has signed the CNCF CLA. label Jun 22, 2023
@k8s-ci-robot k8s-ci-robot added the size/XS Denotes a PR that changes 0-9 lines, ignoring generated files. label Jun 22, 2023
@sbueringer
Copy link
Member

/lgtm

@k8s-ci-robot k8s-ci-robot added the lgtm "Looks good to me", indicates that a PR is ready to be merged. label Jun 22, 2023
@k8s-ci-robot
Copy link
Contributor

LGTM label has been added.

Git tree hash: c862ac76e3f8e9f4a2390731d9471c1701331473

Copy link
Contributor

@killianmuldoon killianmuldoon left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

/lgtm
/approve

FIXME
CAPIM is a implementation of an infrastructure provider for the Cluster API project using in memory, fake objects.

**NOTE:** The In memory provider is **not** designed for production use and is intended for development environments only.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The In memory provider is not designed for production use

I'd like to see someone try 😆

@k8s-ci-robot
Copy link
Contributor

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: killianmuldoon

The full list of commands accepted by this bot can be found here.

The pull request process is described here

Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@k8s-ci-robot k8s-ci-robot added the approved Indicates a PR has been approved by an approver from all required OWNERS files. label Jun 22, 2023
@k8s-ci-robot k8s-ci-robot merged commit 266685d into kubernetes-sigs:main Jun 22, 2023
11 checks passed
@k8s-ci-robot k8s-ci-robot added this to the v1.5 milestone Jun 22, 2023
@fabriziopandini fabriziopandini deleted the fix-readme-inmemory-provider branch June 23, 2023 15:04
Copy link
Contributor

@killianmuldoon killianmuldoon left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

/area provider/infrastructure-inmemory

@k8s-ci-robot
Copy link
Contributor

@killianmuldoon: The label(s) area/provider/infrastructure-inmemory cannot be applied, because the repository doesn't have them.

In response to this:

/area provider/infrastructure-inmemory

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@killianmuldoon
Copy link
Contributor

/area provider/infrastructure-in-memory

@k8s-ci-robot k8s-ci-robot added the area/provider/infrastructure-in-memory Issues or PRs related to the in-memory infrastructure provider label Jun 27, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
approved Indicates a PR has been approved by an approver from all required OWNERS files. area/provider/infrastructure-in-memory Issues or PRs related to the in-memory infrastructure provider cncf-cla: yes Indicates the PR's author has signed the CNCF CLA. lgtm "Looks good to me", indicates that a PR is ready to be merged. size/XS Denotes a PR that changes 0-9 lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants