- Setting up the development environment
- Repository structure
- Running the tests locally
- Making a contribution
- Advanced Debugging
- Install Golang environment (check instructions on the official page https://go.dev/doc/install depending on you OS).
- Fork this repo and clone it.
- Run
make dev-setup
in the main directory of the cloned repository. - You can clean up the dev setup by running
make dev-cleanup
.
The notable technical files/directories inside the repository:
Makefile
- contains instructions to set up the developer's environment, run tests, etc.pkg/provider
- definition of the providerpkg/resources
,pkg/datasources
- definitions and tests (consult section Running the tests locally below) for resources and datasourcespkg/acceptance
- helpers for acceptance and integration testspkg/sdk
- definitions of the SDK objects (SDK is our client to Snowflake, using gosnowflake driver underneath)pkg/sdk/testint
- integration tests for the SDK (consult section Running the tests locally below)
Currently, we have three main types of tests:
- SDK unit tests (in directory
pkg/sdk
, files ending with_test.go
) - SDK integration tests (in directory
pkg/sdk/testint
, files ending with_integration_test.go
) - resource/datasource acceptance tests (in directories
pkg/resources
andpkg/datasources
, files ending with_acceptance_test.go
)
Both integration and acceptance tests require the connection to Snowflake (some of the tests require multiple accounts).
The preferred way of running particular tests locally is to create a config file ~/.snowflake/config
, with the following content.
[default]
account = "<your account in form of organisation-account_name>"
user = "<your user>"
password = "<your password>"
role = "<your role>"
host="<host of your account, e.g. organisation-account_name.snowflakecomputing.com>"
To be able to run all the tests you additionally need two more profiles [secondary_test_account]
and [incorrect_test_profile]
:
[secondary_test_account]
account = "<your secondary account in form of organisation-account_name2>"
user = "<your user on the secondary account>"
password = "<your password on the secondary account>"
role = "<your role on the secondary account>"
host="<host of your account, e.g. organisation-account_name2.snowflakecomputing.com>"
[incorrect_test_profile]
account = "<your account in form of organisation-account_name>"
user = "<non-existing user>"
password = "<bad password>"
role = "<any role, e.g. ACCOUNTADMIN>"
host="<host of your account, e.g. organisation-account_name.snowflakecomputing.com>"
We are aware that not everyone has access two multiple accounts, so the majority of tests can be run using just one account. The tests setup however, requires both profiles (default
and secondary_test_account
) to be present. You can use the same details for secondary_test_account
as in the default
one, if you don't plan to run tests requiring multiple accounts. The warning will be logged when setting up tests with just a single account.
ACCOUNTADMIN
). Otherwise, the managed objects may not be created. If you want to use lower role, you have to make sure it has all the necessary privileges added.
To run the tests we have three different commands:
make test
run unit and integration testsmake test-acceptance
run acceptance testsmake test-integration
run integration tests
You can run the particular tests form inside your chosen IDE but remember that you have to set TF_ACC=1
environment variable to run any acceptance tests (the above commands set it for you). It is also worth adding the TF_LOG=DEBUG
environment variable too, because the output of the execution is much more verbose.
It's important to establish the scope of the change before the actual implementation. We want to avoid the situations in which the PR is rejected because it contradicts some other change we are introducing.
Remember to consult our roadmap, maybe we are already working on the issue!
It's best to approach us through the GitHub issues: either by commenting the already existing one or by creating a new one.
We believe that code following the same conventions is easier to maintain and extend. When working on the given part of the provider try to follow the local solutions and not introduce too much new ideas.
To create new objects in our SDK we use quickly created generator that outputs the majority of the files needed. These files should be later edited and filled with the missing parts. We plan to improve the generator later on, but it should be enough for now. Please read more in the generator readme.
Every introduced change should be tested. Depending on the type of the change it may require (any or mix of):
- adding/modifying existing unit tests (e.g. changing the behavior of validation in the SDK)
- adding/modifying existing integration tests (e.g. adding missing SDK invocations)
- adding/modifying existing acceptance tests (e.g. fixing the parameter on the resource level)
It's best to discuss with us what checks we expect prior to making the change.
If the change requires manual actions when bumping the provider version, they should be added to the migration guide.
The documentation for the provider is generated automatically. We follow the few formatting conventions that are automatically checked with every PR. They can fail and delay the resolution of your PR. To make it much less possible, run make pre-push
before pushing your changes to GH. It will reformat your code (or suggest reformatting), generate all the missing docs, clean the dependencies, etc.
We use Conventional Commits for commit message formatting and PR titles. Please try to adhere to the standard.
Refer to the regular expression for PR title validation.
Implemented changes should be described thoroughly (we will prepare PR template for the known use cases soon):
- reference the issue that is addressed with the given change
- summary of changes
- summary of added tests
- (optional) what parts will be covered in the subsequent PRs
We check for the new PRs in our repository every day Monday-Friday. We usually need 1-2 days to leave the review comments. However, there are times when you can expect even more than a week response time. In such cases, please be patient, and ping us after a week if we do not post a reason for the delay ourselves. It's possible that we just missed it.
During our review we try to point out the unhandled special cases, missing tests, and deviations from the established conventions. Remember, review comment is like an invitation to dance: you don't have to agree but please provide the substantive reasons.
If you want to build and test the provider locally (manually, not through acceptance tests), build the binary first using make build-local
or install to the proper local directory by invoking make install-tf
(to uninstall run make uninstall-tf
).
Next, edit your ~/.terraformrc
file to include the following:
provider_installation {
dev_overrides {
"registry.terraform.io/Snowflake-Labs/snowflake" = "<path_to_binary>"
}
direct {}
}
To debug the provider with a debugger:
-
Launch the provider with the
-debug
command line argument in your debugger session. Once the provider starts, it will print instructions on setting theTF_REATTACH_PROVIDERS
environment variable.Provider started. To attach Terraform CLI, set the TF_REATTACH_PROVIDERS environment variable with the following: Command Prompt: set "TF_REATTACH_PROVIDERS={"registry.terraform.io/Snowflake-Labs/snowflake":{"Protocol":"grpc","ProtocolVersion":5,"Pid":35140,"Test":true,"Addr": {"Network":"tcp","String":"127.0.0.1:54706"}}}" PowerShell: $env:TF_REATTACH_PROVIDERS='{"registry.terraform.io/Snowflake-Labs/snowflake":{"Protocol":"grpc","ProtocolVersion":5,"Pid":35140,"Test":true,"Addr":{"Network":"tcp","String":"127.0.0.1:54706"}}}'
-
Open a terminal where you will execute Terraform and set the
TF_REATTACH_PROVIDERS
environment variable using the command from the first step. -
Run Terraform as usual from this terminal. Any breakpoints you set will halt execution, and you can troubleshoot the provider from your debugger.
Note: The TF_REATTACH_PROVIDERS
environment variable needs to be set every time you restart your debugger session as some values like the Pid
or the TCP port will change with every execution.
For further instructions, please check the official Terraform Plugin Development guide.