Thank you for your interest in contributing to Elastic Cloud on Kubernetes! The goal of this document is to provide a high-level overview on how you can get involved.
If you find an issue, check first our list of issues. If your problem has not been reported yet, open a new issue, add a detailed description on how to reproduce the problem and complete it with any additional information that might help solving the issue.
Check requirements and steps in this guide.
- Run
make lint
to make sure there are no lint warnings. - Make sure you only have two groups in your imports:
- a group for packages from the standard library
- a group for third parties
As most of the contributors are using macOS and Linux, make sure that scripts run on these two environments.
Your contributions should pass the existing tests. You must provide new tests to demonstrate bugs and fixes.
There are 3 test suites:
-
Unit tests - use standard
go test
and github.com/stretchr/testify/assert assertions. Keep them small, fast and reliable.A good practice is to have some table-driven tests, you can use gotests to quickly generate them from your code.
-
Integration tests - some tests are flagged as integration as they can take more than a few milliseconds to complete. It's usually recommended to separate them from the rest of the unit tests that run fast. Usually they include disk I/O operations, network I/O operations on a test port, or encryption computations. We also rely on the kubebuilder testing framework, that spins up etcd and the apiserver locally, and enqueues requests to a reconciliation function.
-
End-to-end tests - (e2e) allow us to test interactions between the operator and a real Kubernetes cluster. They use the standard
go test
tooling. See thetest/e2e
directory. We recommend to rely primarily on unit and integration tests, as e2e tests are slow and hard to debug because they simulate real user scenarios. To run a specific e2e test, you can use something similar tomake TESTS_MATCH=TestMetricbeatStackMonitoringRecipe clean docker-build docker-push e2e-docker-build e2e-docker-push e2e-run
. This will run the e2e test with your latest commit and is very close to how it will run in CI.A faster option is to run the operator and tests locally, with
make run
in one shell andmake e2e-local TESTS_MATCH= TestMetricbeatStackMonitoringRecipe
in another, though this does not exercise all of the same configuration (permissions etc.) that will be used in CI, so is not as thorough.
After submitting a PR, a run of unit tests, integration tests and a single E2E test (SamplesTest
) on a single provider (GKE) can be triggered by commenting the PR with jenkins test this please
.
For changes larger in scope and touching CI pipeline, e2e test runner and/or providers a full E2E test suite can be triggered by commenting the PR with run full pr build
. This will trigger the same set of pipelines that are triggered during nightly testing. Note that green checkmark next to this build result (devops-ci/full
) in GitHub only means that scheduling runs was successful. Inspect all results by going to Details
> Open Blue Ocean
.
The operator relies on controller-runtime logging instead of golang built-in log library. It uses a type of logging called structured logging, log messages must not contain variables, but they can be associated with some key/value pairs.
For example, do not write:
log.Printf("starting reconciliation for pod %s/%s", podNamespace, podName)
But instead write:
logger.Info("starting reconciliation", "pod", req.NamespacedNamed)
We only use two levels: debug
and info
. To produce a log at the debug
level use V(1)
before the Info
call:
logger.V(1).Info("starting reconciliation", "pod", req.NamespacedNamed)
We require license headers on all files that are part of the source code.
Make sure you signed the Contributor License Agreement. You only need to sign the CLA once. By signing this agreement, you give us the right to distribute your code without restriction.
Here are some good practices for a good pull request:
- Push your changes to a topic branch in your fork of the repository.
- Break your pull request into smaller PRs if it's too large.
- Run and pass unit and integration tests with
make unit
andmake integration
. - Write a short and self-explanatory title.
- Write a clear description to make the code reviewer understand what the PR is about.
New PRs should target the master
branch, then be backported as necessary. The original PR to master should contain labels of the versions it will be backported to. The actual backport PR should be labeled backport
. You can use https://github.com/sqren/backport to generate backport PRs easily. An example .backportrc.json
may be:
{
"upstream": "elastic/cloud-on-k8s",
"targetBranchChoices": [
{ "name": "1.2", "checked": true },
"1.1",
"1.0"
],
"targetPRLabels": ["backport"]
}
Whether it’s a new or existing feature, make sure you document it. Before you start, pull the latest files from these repos:
- elastic/cloud-on-k8s: Contains the docs source files for Elastic Cloud on Kubernetes.
- elastic/docs: Has the tools to publish locally your changes before committing them.
To update existing content, find the right file in the cloud-on-k8s/docs/ repo and make your change.
To create new content in a new file, add the file to cloud-on-k8s/docs/, and include it in the index.asciidoc.
NOTE: For searchability purposes, the file name should match the first top-level section ID of the document. For example:
- File name:
apm-server.asciidoc
- Section ID:
[id="{p}-apm-server"]
Test the doc build locally:
-
Move to the directory where the elastic/docs repository has been pulled
-
Run the following command:
./build_docs --asciidoctor --doc $GOPATH/src/github.com/elastic/cloud-on-k8s/docs/index.asciidoc --chunk 1 --open
Push a PR for review and add the label >docs
.
We keep track of architectural decisions through the architectural decision records. All records must apply the Markdown Architectural Decision Records format. We recommend to read these documents to understand the technical choices that we make.
Thank you for taking the time to contribute.