Contributing to Camel K
In order to build the project, you need to comply with the following requirements:
The Camel K Java runtime (camel-k-runtime) requires:
Java 11: needed for compilation
Maven: needed for building
Checks rely on
golangci-lint being installed, to install it look at the Local Installation instructions.
You can run checks via
make lint, or you can install a GIT pre-commit hook and have the checks run via pre-commit; then make sure to install the pre-commit hooks after installing pre-commit by running:
$ pre-commit install
You can create a fork of this project from GitHub, then clone your fork with the
git command line tool.
This is a high-level overview of the project structure:
Contains the Docker and Maven build configuration.
Contains the entry points (the main functions) for the camel-k binary (manager) and the kamel client tool.
Contains Kubernetes resource files, specifically for use with the operator-sdk, that are used by the kamel client during installation. The
Contains Kubernetes resource files, used by the kamel client during installation. The
Contains the documentation website based on Antora.
Include integration tests to ensure that the software interacts correctly with Kubernetes and OpenShift.
Various examples of Camel K usage.
This is where the code resides. The code is divided in multiple subpackages.
Contains scripts used during make operations for building the project.
To build the whole project you now need to run:
This executes a full build of the Go code. If you need to build the components separately you can execute:
make build-operator: to build the operator binary only.
make build-kamel: to build the
kamelclient tool only.
After a successful build, if you’re connected to a Docker daemon, you can build the operator Docker image by running:
The above command produces a
camel-k image with the name
docker.io/apache/camel-k. Sometimes you might need to produce camel-k images that need to be pushed to the custom repository e.g.
docker.io/myrepo/camel-k, to do that you can pass a parameter
imgDestination to the make as shown below:
make imgDestination='docker.io/myrepo' images
Unit tests are executed automatically as part of the build. They use the standard go testing framework.
Integration tests (aimed at ensuring that the code integrates correctly with Kubernetes and OpenShift), need special care.
The convention used in this repo is to name tests
xxx_test.go. Integration tests are all in the /e2e dir.
Since both names end with
_test.go, both would be executed by go during the build, so you need to put a special build tag to mark integration tests. An integration test should start with the following line:
// +build integration
Look into the /e2e directory for examples of integration tests.
Before running an integration test, you need to be connected to a Kubernetes/OpenShift namespace. After you log in into your cluster, you can run the following command to execute all integration tests:
If you want to install everything you have in your source code and see it running on Kubernetes, you need to run the following command:
You need to have Docker installed and running (or connected to a Docker daemon)
You need to set up Docker daemon to trust CRC’s insecure Docker registry which is exposed by default through the route
default-route-openshift-image-registry.apps-crc.testing. One way of doing that is to instruct the Docker daemon to trust the certificate:
oc extract secret/router-ca --keys=tls.crt -n openshift-ingress-operator: to extract the certificate
sudo cp tls.crt /etc/docker/certs.d/default-route-openshift-image-registry.apps-crc.testing/ca.crt: to copy the certificate for Docker daemon to trust
docker login -u kubeadmin -p $(oc whoami -t) default-route-openshift-image-registry.apps-crc.testing: to test that the certificate is trusted
make install-crc: to build the project and install it in the current namespace on CRC
You can specify a different namespace with
make install-crc project=myawesomeproject
To uninstall Camel K, run
kamel uninstall --all --olm=false
The commands assume you have an already running CRC instance and logged in correctly.
make install-minishift(or just
make install): to build the project and install it in the current namespace on Minishift
You can specify a different namespace with
make install-minishift project=myawesomeproject
This command assumes you have an already running Minishift instance.
make install-minikube: to build the project and install it in the current namespace on Minikube
This command assumes you have an already running Minikube instance.
If you need to develop and test your Camel K operator locally, you can follow the local development procedure.
Sometimes it’s useful to debug the code from the IDE when troubleshooting.
It should be straightforward: just execute the /cmd/kamel/main.go file from the IDE (e.g. Goland) in debug mode.
It is a bit more complex (but not so much).
You are going to run the operator code outside OpenShift in your IDE so, first of all, you need to stop the operator running inside:
// use kubectl in plain Kubernetes oc scale deployment/camel-k-operator --replicas 0
You can scale it back to 1 when you’re done, and you have updated the operator image.
You can set up the IDE (e.g. Goland) to execute the /cmd/manager/main.go file in debug mode with
operator as the argument.
When configuring the IDE task, make sure to add all required environment variables in the IDE task configuration screen:
KUBERNETES_CONFIGenvironment variable to point to your Kubernetes configuration file (usually
WATCH_NAMESPACEenvironment variable to a Kubernetes namespace you have access to.
OPERATOR_NAMEenvironment variable to
After you set up the IDE task, with Java 11+ to be used by default, you can run and debug the operator process.
|The operator can be fully debugged in Minishift, because it uses OpenShift S2I binary builds under the hood. The build phase cannot be (currently) debugged in Minikube because the Kaniko builder requires that the operator and the publisher pod share a common persistent volume.|
There are two formats for the publishing of the metadata:
The legacy packaging format used for deploying the operator to an OLM registry. While deprecated in Openshift 4.5+, it is still supported and used on that and other cluster types. A single CSV is generated, comprising of the operator CRDs, and additional files. All versions of the operator metadata are located into a single directory.
To generate the metadata for camel-k in the legacy format, use the following command:
The CSV and accompanying files are located in
deploy/olm-catalog. Note. these files are under version control.
The latest packaging format used for deploying the operator to an OLM registry. This generates a CSV and related metadata files in a directory named
bundle. The directory contains a Dockerfile that allows for building the bundle into a single image. It is this image that is submitted to the OLM registry.
To generate the bundle for camel-k, use the following command:
The bundle directory is created at the root of the camel-k project filesystem.