KubeVela & OAM: The Resurrection of Simplified App Management?
Imagine that you are building an Internal Developer Platform.
What would be a good user experience if, for example, one would like to deploy and manage a backend application without spending five years trying to understand all the details about Kubernetes?
If you are following along the instructions in this post, skip the commands in this section. This is only a preview of what’s coming. We’ll set up everything later.
How about saying vela up
and passing only a few lines of YAML containing only the information that actually matters without dealing with low-level details?
vela up --file app.yaml
That’s it. That’s all there is to do.
Let’s see what happened behind the scenes.
kubectl --namespace dev get all,ingresses
The output is as follows.
NAME READY STATUS RESTARTS AGE
pod/silly-demo-864f6c8f8c-hxkwv 1/1 Running 0 3m6s
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/silly-demo ClusterIP 10.96.77.159 <none> 8080/TCP 3m6s
NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/silly-demo 1/1 1 1 3m6s
NAME DESIRED CURRENT READY AGE
replicaset.apps/silly-demo-864f6c8f8c 1 1 1 3m6s
NAME REFERENCE TARGETS MINPODS MAXPODS REPLICAS AGE
horizontalpodautoscaler.autoscaling/silly-demo Deployment/silly-demo cpu: <unknown>/80%, memory: <unknown>/80% 1 5 1 3m6s
NAME CLASS HOSTS ADDRESS PORTS AGE
ingress.networking.k8s.io/silly-demo nginx dev.silly-demo.127.0.0.1.nip.io localhost 80 3m6s
We got a deployment
, which created a replicaset
which created a pod
that runs the actuall application. We also got a service
in change of internal networking, a horizontalpodautoscaler
that scales replicas of the app, and an ingress
that is in charge of redirecting external traffic into the app.
That application was deployed into the dev environment. Now, let’s say that we would like to promote it to production. All we’d have to do is to resume
the workflow
.
vela workflow resume silly-demo
Let’s see what happened by listing all the resources, this time in the prod
Namespace.
kubectl --namespace prod get all,ingresses
The output is as follows.
NAME READY STATUS RESTARTS AGE
pod/silly-demo-7bc9d88df7-4gzkd 0/1 CreateContainerConfigError 0 3m31s
pod/silly-demo-7bc9d88df7-vhxgm 0/1 CreateContainerConfigError 0 3m46s
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/silly-demo ClusterIP 10.96.99.182 <none> 8080/TCP 3m46s
NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/silly-demo 0/2 2 0 3m46s
NAME DESIRED CURRENT READY AGE
replicaset.apps/silly-demo-7bc9d88df7 2 2 0 3m46s
NAME REFERENCE TARGETS MINPODS MAXPODS REPLICAS AGE
horizontalpodautoscaler.autoscaling/silly-demo Deployment/silly-demo cpu: <unknown>/80%, memory: <unknown>/80% 2 5 2 3m46s
NAME CLASS HOSTS ADDRESS PORTS AGE
ingress.networking.k8s.io/silly-demo nginx silly-demo.127.0.0.1.nip.io localhost 80 3m46s
There are almost always some differences between dev and prod so we got a similar, yet not the same set of resources. For example, this time there are two Pods (pod
) since we need high availability. the ingress
host is different as well. More importantly, the app in production needs a database which we did not yet create. As a result, the Pods are failing since they are trying to mount a secret with database authentication and that secret does not yet exist.
Let’s fix that by deploying a database server and everything it might need by executing vela up
, passing it the manifest that is only a few lines of YAML, and making sure that we’re doing it in the prod
environment.
vela up --file db-google.yaml --env prod
That’s it. That’s all it took to deploy a backend application to both dev and prod as well as to get a database running in a hyperscaler of choice.
Here’s a question though. How did I do all that and why did I do it the way I did?
Open Application Model (OAM) and KubeVela (Revisited)
Let me start with a question. What do zombies, Jesus, and Open Application Model (OAM) have in common?
They were all resurrected from dead.
OAM, short for Open Application Model is an attempt to standardize application definitions. It’s a spec that was born in Microsoft only to be abandoned shortly afterward. Microsoft made it, thought that it is a dead-end, and then it killed it. However, OAM did not stay dead for long. Then came Alibaba and resurected it from the dead, mainly through the KubeVela project. OAM is a spec and KubeVela is the implementation of that spec.
Back in the day, I thought that KubeVela was one of the most interesting and useful projects, especially for those building Internal Developer Platforms (IDPs). Since then I switched to other projects that perform similar functions and eventually settled down on Crossplane.
Nevertheless, KubeVela kept evolving and added a lot of very interesting features and capabilities. So, I went back to it in an attempt to figure out whether I made a wrong call. I wanted to see whether it ovecame the issues that compelled me to abandon it, whether I should come back to it, and whether its new ideas might be worth adding to other similar projects.
So, today’s video is a walkthrough through KubeVela. We’ll see what it does, what’s old and what’s new, whether you should adopt it or, if you are already using it, whether you should continue investing into it.
By the end of this video you will be equipped with knowledge that will help you make a decision whether to use KubeVela as one of the building blocks of your internal developer platform.
Setup
git clone https://github.com/vfarcic/kubevela-demo-2
cd kubevela-demo-2
Make sure that Docker is up-and-running. We’ll use it to create a KinD cluster.
Watch Nix for Everyone: Unleash Devbox for Simplified Development if you are not familiar with Devbox. Alternatively, you can skip Devbox and install all the tools listed in
devbox.json
yourself.
devbox shell
Watch The Future of Shells with Nushell! Shell + Data + Programming Language if you are not familiar with Nushell. Alternatively, you can inspect the
dot.nu
script and transform the instructions in it to Bash or ZShell if you prefer not to use that Nushell script.
chmod +x dot.nu
./dot.nu setup
source .env
Define KubeVela Components and Traits
To start working with KubeVela, we need to initialize an environment, so let’s to just that with dev
,…
vela env init dev --namespace dev
…and prod
.
vela env init prod --namespace prod
Those environments are, essentially, Namespaces in Kubernetes.
Let’s confirm that by listing all those in the cluster.
kubectl get namespaces
The output is as follows.
NAME STATUS AGE
crossplane-system Active 9m41s
default Active 10m
dev Active 82s
ingress-nginx Active 10m
kube-node-lease Active 10m
kube-public Active 10m
kube-system Active 10m
kubevela Active 3m44s
local-path-storage Active 10m
prod Active 67s
vela-system Active 5m3s
We can see the dev
and prod
are there. There is nothing really special about them, apart from a label vela put into them. We could have created them without the CLI.
The core of KubeVela are components that define all the base resources it will compose for us.
Here’s an example.
cat component-app-backend.cue
The output is as follows.
"app-backend": {
attributes: {
workload: definition: {
apiVersion: "apps/v1"
kind: "Deployment"
}
status: healthPolicy: "isHealth: (context.output.status.readyReplicas > 0) && (context.output.status.readyReplicas == context.output.status.replicas)"
}
type: "component"
}
template: {
parameter: {
image: string
tag: string
port: *80 | int
host: *"devopstoolkit.live" | string
ingressClassName: string
db: secret: string
db: secretNamespace: string
}
output: {
apiVersion: "apps/v1"
kind: "Deployment"
metadata: labels: "app.kubernetes.io/name": context.name
spec: {
selector: matchLabels: "app.kubernetes.io/name": context.name
template: {
metadata: labels: "app.kubernetes.io/name": context.name
spec: containers: [{
image: parameter.image + ":" + parameter.tag
livenessProbe: httpGet: {
path: "/"
port: parameter.port
}
name: "backend"
ports: [{ containerPort: 80 }]
readinessProbe: httpGet: {
path: "/"
port: parameter.port
}
resources: {
limits: {
cpu: "250m"
memory: "256Mi"
}
requests: {
cpu: "125m"
memory: "128Mi"
}
}
if parameter.db.secret != _|_ {
env: [{
name: "DB_ENDPOINT"
valueFrom: secretKeyRef: {
key: "endpoint"
name: parameter.db.secret
}
}, {
name: "DB_PASSWORD"
valueFrom: secretKeyRef: {
key: "password"
name: parameter.db.secret
}
}, {
name: "DB_PORT"
valueFrom: secretKeyRef: {
key: "port"
name: parameter.db.secret
optional: true
}
}, {
name: "DB_USERNAME"
valueFrom: secretKeyRef: {
key: "username"
name: parameter.db.secret
}
}, {
name: "DB_NAME"
value: context.name
}]
}
}]
}
}
}
outputs: {
service: {
apiVersion: "v1"
kind: "Service"
metadata: {
name: context.name
labels: "app.kubernetes.io/name": context.name
}
spec: {
selector: "app.kubernetes.io/name": context.name
type: "ClusterIP"
ports: [{
port: parameter.port
targetPort: parameter.port
protocol: "TCP"
name: "http"
}]
}
}
ingress: {
apiVersion: "networking.k8s.io/v1"
kind: "Ingress"
metadata: {
name: context.name
labels: "app.kubernetes.io/name": context.name
annotations: "ingress.kubernetes.io/ssl-redirect": "false"
}
spec: {
if parameter.ingressClassName != _|_ {
ingressClassName: parameter.ingressClassName
}
rules: [{
host: parameter.host
http: paths: [{
path: "/"
pathType: "ImplementationSpecific"
backend: service: {
name: context.name
port: number: parameter.port
}
}]
}]
}
}
}
}
The first thing you’ll notice is, for some, a strange syntax. Everything we define with KubeVela is written in CUE, which is a language built on top of Go aiming to make it easier to write complex configurations. It’s a data-driven language that is both a superset of JSON and Go.
Over there we have app-backend
which acts as a name of a component
. The main resource it will manage is a Deployment
.
Further down we have a template
that defines all the parameters (parameter
) and the resources (output
) it will manage for us.
Think of it as Helm template that will be running in Kubernetes instead you laptop and that will create new API endpoints in Kubernetes API (sort of).
Over there, we have image
, tag
, port
, host
, ingressClassName
, and db
parameters, with db
having two sub-parameters: secret
and secretNamespace
. Those will be parameters others will be defining when they use that component. Think of those as Helm values file, except that it will be available in Kubernetes API (kind of).
The output
section contains the list of resources it will assemble for us. There is a Deployment
, a Service
, and an Ingress
.
If you ignore the fact that it is written in CUE, those would be the same definitions you would normally write as YAML files. The major difference is that some of the values are templates. Some of those are using built-in values like context.name
, while others are using parameters defined earlier. That would be, for example, parameter.image
and parameter.tag
which we’re using to construct the image
value for the Deployment.
All in all, if someone requests the app-backend
component, KubeVela will create a Deployment
, a Service
, and an Ingress
with the parameters we defined.
So far so good. Easy!
Now, let’s apply that component and, by doing that, make it available to everyone else with everyone else being users of the platform we’re building.
vela def apply component-app-backend.cue
That’s it. From now on, our users can create and manage backend applications without having to deal with Deployments, Services, and Ingresses. We’ll make it more complicated later but, for now, that’s it.
Okay. Now that we created Backend-App-as-a-Service, let’s see how can users know what to do with it.
One option is to go to the KubeVela UI, which we’ll explore later. Another option is to use the CLI to show
the information about the component.
vela show app-backend
The output is as follows.
# Specification
+------------------+-------------+-----------+----------+--------------------+
| NAME | DESCRIPTION | TYPE | REQUIRED | DEFAULT |
+------------------+-------------+-----------+----------+--------------------+
| image | | string | true | |
| tag | | string | true | |
| port | | int | false | 80 |
| host | | string | false | devopstoolkit.live |
| ingressClassName | | string | true | |
| db | | [db](#db) | true | |
+------------------+-------------+-----------+----------+--------------------+
## db
+-----------------+-------------+--------+----------+---------+
| NAME | DESCRIPTION | TYPE | REQUIRED | DEFAULT |
+-----------------+-------------+--------+----------+---------+
| secret | | string | true | |
| secretNamespace | | string | true | |
+-----------------+-------------+--------+----------+---------+
We can see the Specification
with the same fields we defined in the component. Since db
has some sub-fields, they are listed separately (## db
).
We can also list all the componentdefinitions
in the cluster.
kubectl --namespace vela-system get componentdefinitions
The output is as follows.
NAME WORKLOAD-KIND DESCRIPTION
app-backend Deployment
cron-task Describes cron jobs that run code or a script to completion.
daemon DaemonSet Describes daemonset services in Kubernetes.
k8s-objects K8s-objects allow users to specify raw K8s objects in properties
raw Raw allow users to specify raw K8s object in properties. This definition is DEPRECATED, please use 'k8s-objects' instead.
ref-objects Ref-objects allow users to specify ref objects to use. Notice that this component type have special handle logic.
task Job Describes jobs that run code or a script to completion.
webservice Deployment Describes long-running, scalable, containerized services that have a stable network endpoint to receive external network traffic from customers.
worker Deployment Describes long-running, scalable, containerized services that running at backend. They do NOT have network endpoint to receive external network traffic.
We can see that there are some pre-defined components available, as well as the app-backend
we just created.
Finally, let’s list Kubernetes crds
related to oam
.
kubectl get crds | grep oam
The output is as follows.
applicationrevisions.core.oam.dev 2025-02-14T01:17:06Z
applications.core.oam.dev 2025-02-14T01:17:06Z
componentdefinitions.core.oam.dev 2025-02-14T01:17:06Z
definitionrevisions.core.oam.dev 2025-02-14T01:17:06Z
policies.core.oam.dev 2025-02-14T01:17:06Z
policydefinitions.core.oam.dev 2025-02-14T01:17:06Z
resourcetrackers.core.oam.dev 2025-02-14T01:17:06Z
traitdefinitions.core.oam.dev 2025-02-14T01:17:06Z
workflows.core.oam.dev 2025-02-14T01:17:06Z
workflowstepdefinitions.core.oam.dev 2025-02-14T01:17:06Z
workloaddefinitions.core.oam.dev 2025-02-14T01:17:06Z
As a side note, KubeVela is an implementation of the Open Application Model (OAM) specification. That’s why we filtered the CRDs with oam.
Here’s the first, in my opinion, issue we are facing with KubeVela. It does not create new CRDs. Whatever we do is always implemented as applications.core.oam.dev
. That makes discovery harder than it should be. I would have expected a new CRD called, let’s say, AppBackend or something like that. If we would get that, we could “discover” the schema for that specific resource, we could tell Backstage, or whichever UI you’re using as a platform portal, to fetch that schema and generate fields for that resource. We could do many other things, but we can’t since it is always Application CRD no matter what we do. That’s a pity.
Nevertheless, from now on, users would be able to manage their backend applications using the component we created, but there’s more. We might want to add additional traits to the component.
Let’s say that we would like to enable users to decide whether their applications should scale horizontally. Since that would be an optional feature, we cannot add it to our component directly. If we would, scaling would be enabled all the time, making it mandatory.
KubeVela has the concept of Traits which are a way to add optional features to a component. Here’s an example.
cat trait-scaler.cue
The output is as follows.
scaler: {
attributes: {
podDisruptive: false
}
type: "trait"
}
template: {
parameter: {
min: *1 | int
max: *10 | int
}
outputs: {
hpa: {
apiVersion: "autoscaling/v2"
kind: "HorizontalPodAutoscaler"
metadata: {
name: context.name
labels: "app.kubernetes.io/name": context.name
}
spec: {
scaleTargetRef: {
apiVersion: "apps/v1"
kind: "Deployment"
name: context.name
}
minReplicas: parameter.min
maxReplicas: parameter.max
metrics: [{
type: "Resource"
resource: {
name: "cpu"
target: {
type: "Utilization"
averageUtilization: 80
}
}
}, {
type: "Resource"
resource: {
name: "memory"
target: {
type: "Utilization"
averageUtilization: 80
}
}
}]
}
}
}
}
This time we are defining a scaler
. Unlike the previous definition, this one is a trait
. Traits are a way to optionally attach additional resources to a component.
In this case, it defines parameters min
and max
, both with default values. Those who attach that trait will be able to specify the minimum and the maximum number of replicas which will scale automatically between those two numbers.
Further on, the we have a HorizontalPodAutoscaler
resource which will be created when someone attaches that trait to a component.
Let’s apply it.
vela def apply trait-scaler.cue
That’s about it, for now. We have a backend application that will compose a Deployment, a Service, and an Ingress, and, optionally, have a HorizontalPodAutoscaler.
Let’s see how it all works.
Use KubeVela Components and Traits
Now we’ll see how developers can use KubeVela to create and manage their applications.
To begin with, we’ll set dev
to be the current environment.
vela env set dev
From now on, we, as developers, will be using dev
environment and can manage backend applications either from the Web UI or using the CLI.
If we open the KubeVela UI, we’ll see the home page where nothing is happening just yet. From there on, we can select to create New Application
that opens a new modal split into two sections. First, there are common atributes like the Name
and the Description
that apply to all Kubevela Applications. The important part, in this scenario, is to select the app-backend
component we applied earlier.
If we proceed to the next step, we can fill in the fields that are specific to the component we selected. Those are exactly the same parameters as those we defined in the component earlier.
You can probably guess what happens next. If you can’t… Well… Tough luck since we’ll switch to the terminal. Looking at graphical user interfaces for too long hurts my eyes. A second reason for moving away from the UI is that it does not allow us to specify everything we can define ourselves, so off we go into the darkness of my terminal.
Here’s an example of a YAML.
cat app.yaml
The output is as follows (truncated for brevity).
apiVersion: core.oam.dev/v1beta1
kind: Application
metadata:
name: silly-demo
spec:
components:
- name: silly-demo
type: app-backend
properties:
image: ghcr.io/vfarcic/silly-demo
tag: 1.4.343
port: 8080
host: silly-demo.127.0.0.1.nip.io
ingressClassName: nginx
traits:
- type: scaler
properties:
min: 2
max: 5
policies:
...
workflow:
...
There’s a lot to digest here, so let’s break it down.
To begin with, we are defining an Application
resource. As I already mentioned, we cannot create our own CRDs with KubeVela, and that’s a pity. Instead, it’s always an Application resource.
The spec
is split into multiple sections. There are components
, policies
, and workflow
.
The last time I used KubeVela, probably a few years ago, there were only components
, so this is very exciting for me.
Using Components is straightforward. We specify the name
, the type
, properties
, and traits
.
The name is probably self-explanatory while the type is whatever we defined it to be when we created the component.
Properties are essentially the parameters we defined in the component which will be used to compose the resources.
Traits are optional. In this case, we are specifying that we would like to add the scaler
to the mix.
None of the things we explored so far are new. What comes next are capabilities that did not exist the last time I used KubeVela. Now comes the existing part.
KubeVela Policies and Workflows
Let’s continue exploring the definition of the application.
cat app.yaml
The output is as follows (truncated for brevity).
apiVersion: core.oam.dev/v1beta1
kind: Application
metadata:
name: silly-demo
spec:
...
policies:
- name: target-dev
type: topology
properties:
namespace: dev
- name: host-dev
type: override
properties:
components:
- type: app-backend
properties:
host: dev.silly-demo.127.0.0.1.nip.io
- name: target-prod
type: topology
properties:
namespace: prod
- name: scaler-dev
type: override
properties:
components:
- type: app-backend
traits:
- type: scaler
properties:
min: 1
- name: db-prod
type: override
properties:
components:
- type: app-backend
properties:
db:
secret: silly-demo-db
workflow:
steps:
- name: deploy-to-dev
type: deploy
properties:
policies:
- target-dev
- host-dev
- scaler-dev
- name: promotion
type: suspend
- name: deploy-to-prod
type: deploy
properties:
policies:
- target-prod
- db-prod
KubeVela policies are not what you might expect. They are not policies that define what can and what cannot be done like, for example, Kyverno policies. Instead, they are a way to modify the application before it is deployed.
The first policy is of type topology
. It is used to specify the destination of the application which, in this case, is namespace
dev.
The second policy is of type override
which can be used to modify properties of a component or a trait. In this case, we’re saying that the host
parameter should be dev.silly-demo.127.0.0.1.nip.io
.
Then we have yet another topology
that will deploy the application to prod
namespace.
The fourth one will reduce the min
number of replicas of the scaler
trait to 1.
Finally, the last policy will override
the db.secret
property.
In this particular case, those policies allow us to have variations between the resources that will be composed in different environments given that dev and prod are similar, yet not identical.
Now, the important thing to note is that none of those policies do anything by themselves. Policies are applied through the workflow
, which we’ll explore next.
Let’s see the workflow
part of the definition of that application.
cat app.yaml
The output is as follows (truncated for brevity).
apiVersion: core.oam.dev/v1beta1
kind: Application
metadata:
name: silly-demo
spec:
...
workflow:
steps:
- name: deploy-to-dev
type: deploy
properties:
policies:
- target-dev
- host-dev
- scaler-dev
- name: promotion
type: suspend
- name: deploy-to-prod
type: deploy
properties:
policies:
- target-prod
- db-prod
That workflow
has three steps. The first one will deploy the application to dev
environment but, before it does that, it will modify the application by applying target-dev
, host-dev
, and scaler-dev
policies.
Once it’s done with the deployment to dev, it will suspend
the workflow and wait for someone to resume it.
Once the workflow is resumed, it will deploy-to-prod
environment but, just as with dev, it will modify the application by applying target-prod
and db-prod
policies.
As a result of that workflow, dev and prod will be different and a manual approval will be required to promote the application from dev to prod.
On the first look, the idea of having policies and workflows feels like a very useful addition to KubeVela. Yet, I think it is silly. I think that the implementation is just wrong. Something that was supposed to be easy for people to consume all of a sudden turned itself into something complicated.
We’ll talk about that later. For now, imagine that I did not say anything. Fill yourself with happy thoughts, and let’s move on.
KubeVela in Action
Let’s apply that Application definition and see what we’ll get.
vela up --file app.yaml
To be clear, we could have created that Application from the Web UI. Similarly, we can observe it in the UI as well, but we won’t. We’ll stick with the terminal.
We can get all the applications
in the dev
Namespace.
kubectl --namespace dev get applications
The output is as follows.
NAME COMPONENT TYPE PHASE HEALTHY STATUS AGE
silly-demo silly-demo app-backend workflowSuspending true 2m41s
The only thing I want you to draw your attention to is the PHASE
column currently set to workflowSuspending
. If you remember the definition we explored ealier, it was to be expected for it to go into some kind of a hibernatition after composing the resources in the dev Namespace. That’s what we defined in the workflow.
The interesting part is what that application composed.
kubectl --namespace dev get all,ingresses
The output is as follows (truncated for brevity).
NAME READY STATUS RESTARTS AGE
pod/silly-demo-864f6c8f8c-hxkwv 1/1 Running 0 3m6s
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/silly-demo ClusterIP 10.96.77.159 <none> 8080/TCP 3m6s
NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/silly-demo 1/1 1 1 3m6s
NAME DESIRED CURRENT READY AGE
replicaset.apps/silly-demo-864f6c8f8c 1 1 1 3m6s
NAME REFERENCE TARGETS MINPODS MAXPODS REPLICAS AGE
horizontalpodauto... Deployment... cpu: ..., memory: ... 1 5 1 3m6s
NAME CLASS HOSTS ADDRESS PORTS AGE
ingress.networking.k8s.io/silly-demo nginx dev.silly-demo... localhost 80 3m6s
There’s a deployment
which created the replicaset
which created the pod
. We also got a service
, and an ingress
. Those are all pieces of the Component we’re using. On top of that, since we chose to add the scaler trait we got a hirozontalpodautoscaler
as well.
We might also want to pay attention to the Ingress HOST
. It is set to dev.silly-demo
because the workflow step that deployed the app to the dev environment applied one of the policies that changed the host.
Now, to be 100% clear, so far, we deployed some resources to the dev Namespace and we suspended the execution of the workflow. Nothing was deployed to production so far. Capiche?
If we ever get confused what’s going on and what happened so far, we can always ask for the current status.
vela status silly-demo
The output is as follows.
About:
Name: silly-demo
Namespace: dev
Created at: 2025-02-14 02:48:25 +0100 CET
Status: workflowSuspending
Workflow:
mode: StepByStep-DAG
finished: false
Suspend: true
Terminated: false
Steps
- id: 78w3ucqvxk
name: deploy-to-dev
type: deploy
phase: succeeded
- id: 6yg9ukgvav
name: promotion
type: suspend
phase: suspending
message: Suspended by field suspend
Services:
- Name: silly-demo
Cluster: local Namespace: dev
Type: app-backend
Healthy
Traits:
✅ scaler
The workflow deployed to dev (deploy-to-dev
) and, after that, it suspended the promotion
. As for Services
, we’re running app-backend
only in the dev
Namespace
(for now).
If we would like to see which resources were composed, we can output the status
in the tree
format.
vela status silly-demo --tree
The output is as follows.
CLUSTER NAMESPACE RESOURCE STATUS
local ─── dev ─┬─ Service/silly-demo updated
├─ Deployment/silly-demo updated
├─ HorizontalPodAutoscaler/silly-demo updated
└─ Ingress/silly-demo updated
That KubeVela Application composed a Service
, a Deployment
, a HorizontalPodAutoscaler
, and an Ingress
. We knew all that already. Nevertheless, it is nice to know that we can list all the resources managed by a KubeVela Application.
Now that dev is up and running, we would normally run some tests, do some manual reviews, or anything else we might need to be doing. It is unclear how would KubeVela trigger those processes. We might want to use HTTPDo workflow step to trigger some other processes that would do whatever we need to be done. We might also execute Apply workflow step to apply some other resources that would run tests or whatever we need to be done.
It is certainly doable to connect KubeVela workflows with other workflows, but that does not seem to be thought through. I feel that the assumption is that KubeVela workflows themselves are sufficient, which is certainly not the case.
I’ll leave that as an exercise for you and assume that we validated the release in the dev environment and that we are ready to promote it to prod.
vela workflow resume silly-demo
Let’s check the status again.
vela status silly-demo --tree
The output is as follows.
CLUSTER NAMESPACE RESOURCE STATUS
local ─┬─ dev ─┬─ Service/silly-demo updated
│ ├─ Deployment/silly-demo updated
│ ├─ HorizontalPodAutoscaler/silly-demo updated
│ └─ Ingress/silly-demo updated
└─ prod ─┬─ Service/silly-demo updated
├─ Deployment/silly-demo updated
├─ HorizontalPodAutoscaler/silly-demo updated
└─ Ingress/silly-demo updated
We can see that, this time, we got the release not only in the dev
Namespace but also in prod
.
We can confirm that further by listing all the resources in the prod
Namespace.
kubectl --namespace prod get all,ingresses
The output is as follows (truncated for brevity).
NAME READY STATUS RESTARTS AGE
pod/silly-demo-7bc9d88df7-4gzkd 0/1 CreateContainerConfigError 0 3m31s
pod/silly-demo-7bc9d88df7-vhxgm 0/1 CreateContainerConfigError 0 3m46s
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/silly-demo ClusterIP 10.96.99.182 <none> 8080/TCP 3m46s
NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/silly-demo 0/2 2 0 3m46s
NAME DESIRED CURRENT READY AGE
replicaset.apps/silly-demo-7bc9d88df7 2 2 0 3m46s
NAME REFERENCE TARGETS MINPODS MAXPODS REPLICAS AGE
horizontalpodauto... Deployment... cpu: ..., memory: ... 2 5 2 3m46s
NAME CLASS HOSTS ADDRESS PORTS AGE
ingress.networking.k8s.io/silly-demo nginx dev.silly-demo... localhost 80 3m6s
There we go. We promoted the application from dev to prod and, while doing that, we made slight changes to the resources through policies.
That being said, the Pods are not running. Kubernetes complains that their status is CreateContainerConfigError
. Those Pods are trying to mount a Secret that provides credentials to the database and, since we did not yet deploy the database, the Secret does not exist.
Let’s fix that with a new Component definition.
The outputs you’ll see in the rest of this post are based on databases in Google Cloud. If you followed alone with the instructions, you might see the different outputs if you chose AWS or Azure.
Here’s the definition of the database Component.
cat component-db-$HYPERSCALER.cue
The output is as follows.
import "encoding/base64"
"db-google": {
attributes: {
workload: definition: {
apiVersion: "sql.gcp.upbound.io/v1beta1"
kind: "DatabaseInstance"
}
}
type: "component"
}
template: {
parameter: {
region: *"us-east1" | string
size: *"small" | string
version: string
}
output: {
apiVersion: "sql.gcp.upbound.io/v1beta1"
kind: "DatabaseInstance"
metadata: {
name: context.name + "-" + context.namespace
labels: "app.kubernetes.io/name": context.name
}
spec: {
forProvider: {
region: parameter.region
databaseVersion: "POSTGRES_" + parameter.version
rootPasswordSecretRef: {
name: context.name + "-password"
namespace: context.namespace
key: "password"
}
settings: [{
if parameter.size == "small" {
tier: "db-custom-1-3840"
}
if parameter.size == "medium" {
tier: "db-custom-16-61440"
}
if parameter.size == "large" {
tier: "db-custom-64-245760"
}
availabilityType: "REGIONAL"
backupConfiguration: [{
enabled: true
binaryLogEnabled: false
}]
ipConfiguration: [{
ipv4Enabled: true
authorizedNetworks: [{
name: "all"
value: "0.0.0.0/0"
}]
}]
}]
deletionProtection: false
}
}
}
outputs: {
#Metadata: {
name: context.name + "-" + context.namespace
labels: "app.kubernetes.io/name": context.name
}
user: {
apiVersion: "sql.gcp.upbound.io/v1beta1"
kind: "User"
metadata: #Metadata
spec: {
deletionPolicy: "Orphan"
forProvider: {
passwordSecretRef: {
name: context.name + "-password"
namespace: context.namespace
key: "password"
}
instanceRef: name: context.name + "-" + context.namespace
}
}
}
secret: {
apiVersion: "kubernetes.crossplane.io/v1alpha2"
kind: "Object"
metadata: #Metadata
spec: {
references: [{
patchesFrom: {
apiVersion: "sql.gcp.upbound.io/v1beta1"
kind: "User"
name: context.name + "-" + context.namespace
namespace: "crossplane-system"
fieldPath: "metadata.name"
}
toFieldPath: "stringData.username"
}, {
patchesFrom: {
apiVersion: "v1"
kind: "Secret"
name: context.name + "-password"
namespace: context.namespace
fieldPath: "data.password"
}
toFieldPath: "data.password"
}, {
patchesFrom: {
apiVersion: "sql.gcp.upbound.io/v1beta1"
kind: "DatabaseInstance"
name: context.name + "-" + context.namespace
namespace: "crossplane-system"
fieldPath: "status.atProvider.publicIpAddress"
}
toFieldPath: "stringData.endpoint"
}]
forProvider: manifest: {
apiVersion: "v1"
kind: "Secret"
metadata: {
name: context.name
namespace: context.namespace
}
data: port: "NTQzMg=="
}
}
}
}
}
That Component follows the same pattern as the previous one. The major difference is that, this time, we are not composing core Kubernetes resources but, instead, are combining custom resources that will, ultimately, create a DatabaseInstance
in Google Cloud as the main resource, and a User
and a Secret
as the other resources. We’re using Crossplane Managed Resources for that, for two reasons. First of all, we are trying to manage a database in a hyperscaler like Google Cloud. We could have accomplished the same with other operators like Google Config Connector or AWS Controllers for Kubernetes (ACK), or Azure something something, or anything else. That brings me to the second reason. I wanted to prove that competing tools can be combined. We are using KubeVela in the similar capacity as Crossplane Compositions, but, since KubeVela does not have something simliar to Crossplane Managed Resources, we are combining the two.
Now, to be clear, KubeVela does support the usage of Terraform, but that is just silly. I have nothing against Terraform in general. I think it’s great and that it changed the industry in many ways. However, Terraform does not work well inside Kubernetes. I won’t go into details why I think so since that would result in a complete derailement from the main subject.
The resources we are composing are customizable through parameters region
, size
, and version
. Those should, hopefully, be self-explanatory.
Let’s apply that Component.
vela def apply component-db-$HYPERSCALER.cue
From now on, whenever a developer, or anyone else, needs a database, they can create a KubeVela Application that uses that Component.
We’ll see that in action in a moment. But, before we do that, since the Component we just saw assumes that there is a Secret with the root password to the database, we’ll create that Secret first.
kubectl --namespace prod apply --filename db-$HYPERSCALER-password.yaml
To be honest, that is not something anyone should do. Instead, we should have stored that password in a Secrets Store and pulled it with something like External Secret Operator (ESO). However, that would also derail us from the main subject and, at the same time, I was too lazy to add it to this demo, so we did what should never be done and created the Secret directly.
Now a developer, or anyone else, can define the Application that uses the Component. Here’s an example.
cat db-$HYPERSCALER.yaml
The output is as follows.
apiVersion: core.oam.dev/v1beta1
kind: Application
metadata:
name: silly-demo-db
spec:
components:
- name: silly-demo-db
type: db-google
properties:
region: us-east1
size: small
version: "13"
That’s easy, isn’t it? We define an Application
that uses the db-google
Component with the region
set to us-east1
, the size
set to small
, and the version
set to 13
. That manifest contains all the information needed to create the resources and is the right level of abstraction for a person that does not want to know everything required to run a database in a hyperscaler.
Now that a user of the platform defined the Application, all they have to do to make it real is to apply it,…
vela up --file db-$HYPERSCALER.yaml --env prod
…and check the status.
vela status silly-demo-db --env prod
The output is as follows.
About:
Name: silly-demo-db
Namespace: prod
Created at: 2025-02-14 03:02:07 +0100 CET
Status: running
Workflow:
mode: DAG-DAG
finished: true
Suspend: false
Terminated: false
Steps
- id: 2mfchdel4w
name: silly-demo-db
type: apply-component
phase: succeeded
Services:
- Name: silly-demo-db
Cluster: local Namespace: prod
Type: db-google
Healthy
No trait applied
That status is not actually useful since it does not show the status of all the composed resources nor the overall status of the Application. That can be fixed with healthPolicy, but would require a lot of work, especially on more complex setups. So, it’s possible, but not necessarily something many will do due to the increased complexity.
Next, we should wait for a while until all the resources are operational in whichever hyperscaler we chose.
After a while we can list them with the following command.
kubectl get managed
The output is as follows (truncated for brevity).
NAME KIND PROVIDERCONFIG SYNCED READY AGE
object.kubernetes.crossplane.io/... Secret default False 5m34s
NAME SYNCED READY EXTERNAL-NAME AGE
databaseinstance.sql.gcp.upbound.io/... True True silly-demo-db-prod 5m34s
NAME SYNCED READY EXTERNAL-NAME AGE
user.sql.gcp.upbound.io/silly-demo-db-prod True True silly-demo-db-prod 5m34s
We can see that all the resources are READY
, so we can proceed.
If, in your case, some of the resources are not
READY
, and you are following along, you should wait for a while longer before proceeding.
Let’s get back to the first resource we applied, the one that composes backend application, and check whether the Pods are now running.
kubectl --namespace prod get all
The output is as follows (truncated for brevity).
NAME READY STATUS RESTARTS AGE
pod/silly-demo-7bc9d88df7-4gzkd 1/1 Running 0 14m
pod/silly-demo-7bc9d88df7-vhxgm 1/1 Running 0 14m
...
We can see that the Pods are now Running
. They were failing before because they were trying to mount a secret with database credentials which did not exist. Now that we applied the database as well which, in turn, created that secret, the Pods are running. Hurray!
That’s it. We got two KubeVela Components that compose resources in different ways. One is managing databases in hyperscalers while the other is managing backend applications in Kuberentes clusters with those apps being connected to those databases. Platform users can easily define what they want without the clutter of the underlying low-level resources.
Does everything we saw so far make sense? Let’s talk about it.
KubeVela Pros and Cons
It’s very hard for me to put my finger on KubeVela. It’s as if the direction of every important feature is the correct one, yet we somehow end up in a wrong place. I think that’s just as valid for Application spec as for policies, workflows, and integrations. I’ll get back to that in a moment. For now, let’s talk about pros and cons, starting with those I don’t like.
Cons
- No Discoverable Schemas
- Policies and Workflows
- Poor Integrations
- Only CUE
I don’t understand why KubeVela insists on having a single CRD Application. As it is now, there is no reliable schema. An Application can have any number of components which, in turn, can have any number of traits, and none of that is enforced at the Kubernetes API level but, rather, through the KubeVela CLI and engine. As a result, other tools cannot easily deduce what is what, how to use it, what the schema is, and so on and so forth. KubeVela would improve greatly if it would adopt a model where we would be able to create new CRDs like backend-app, db, or anything else. Whether that would mean removal of the Application model, the OAM, in favor or working directly with Components or something else is beyond me. What matters is that it does not make sense to insist on pre-defined Application CRD when it’s so easy to create new ones. As it is now, KubeVela does not provide discoverable schemas, and that’s a pity.
Policies and workflows are conceptually great. Other tools in that area would benefit greatly from having something similar to what KubeVela has. I do believe that Crossplane and kro, to name a few, should add capabilities that serve the same objectives. That being said, both fail at the fundamental levels. The primary value of tools like KubeVela is the ability to create our own abstractions and simplify lives of those who will use such abstractions. KubeVela does that pretty well with Application, Components, and Traits, yet, when it comes to policies and workflows, it chose to completely change the direction. All of a sudden, low level details are not abstracted any more but moved to the forefront. Now, that’s not something necessarily bad. Some users like dealing with low-level details, while some don’t. The problem is that KubeVela end-users don’t. If they do, they picked a wrong tool. It’s all about one team, call it platform engineers, building Components and Traits while everyone else uses them. Then, all of a sudden, it all falls apart with Policies and Workflows. All of a sudden, roles are inverted so those building services are, mostly, doing nothing, while those consuming them are dealing with low-level details. KubeVela should have done with Policies and Workflows the same as what they did with Components and Traits. Move them to the server-side and expose parts that matter to end-users.
There is very poor integration between KubeVela and other tools from the ecosystem, which is probably why KubeVela is building its own, often as poor replacements for what we already have. Here’s an example. With a single CRD Application serving potentially infinite number of permutations it is close to impossible to define server-side policies like those we typically do with Kyverno. There isn’t much we can do on the Admission Controller level if everything is the same resource and there is no schema. Kyverno will not work, unless in very simple scenarios, but that’s only one out of many. You’ll have trouble using KubeVela with Argo CD, you’ll face challenges trying to observe Applications through metrics from, let’s say, Prometheus, and so on and so forth. Simply put, KubeVela ignores quite a few Kubernetes “rules” making it very hard to combine it with the rest of the ecosystem that follows those same rules. That could be the biggest downside given that the ability for everything to work with anything else is one of the main advantages of Kubernetes. KubeVela tries to solve that not by trying to make itself closer to what is expected from Kubernetes-native apps, but by building it’s own ecosystem. That feels like a waste of time.
KubeVela supports only CUE. Now, to be clear, I think that CUE is great. It is one of my favorite languages when it comes to defining the state of something. Still, the fact is that CUE is nowhere near as widely adopted as it should be. If CUE is what you like and everyone else in your organization agrees with that, great. But, if that’s not the case, KubeVela is a non-starter. There was a period when quite a few projects were formed around the idea that CUE is going to dominate the world. Most of them failed and, as a result, either stopped existings or, more often than not, recorgnized the need to support other languages. CUE is great, but is also a niche language and, as long as that keeps being the case, it is not a good idea to make it the only choice.
There are other issues which we won’t dive into today. Those that I mentioned are, in my opinion, fundamental and not something that can be fixed with a bit of polish. Others are less important.
Let’s see what’s good about it.
Pros
- CLI
- Web UI
- Policies and Workflows
To begin with, vela CLI is fantastic. It is truly great. It, in a way, shows how necessity often produces something great. Given that KubeVela itself does not work well with the rest of the ecosystem, we cannot rely on other tools, those we typically use, to help us with it. Typically, I would expect to be able to do everything with kubectl with any additional tool making it more convenient. With KubeVela CLI, it is not about it being more convenient but, rather, often being the only option. That’s the necessity I mentioned earlier. As a result of that necessity, the team was forced to pay extra attention to the CLI. It’s really great. Big thumbs up.
Web UI is, more or less, great. It is certainly better than what we get in competing projects like Crossplane and kro. That could be it’s biggest selling point, given that graphical user interfaces are the most common ways end-users interact with services.
Finally, there are policies and workflows. That might sound confusing since I put those into cons. I was very negative towards KubeVela implementation of policies and workflows, but I do give them a credit for having them. If there is anything other tools in that space should learn from KubeVela are two things. We need something like policies and workflows integrated into Compositions in case of Crossplane and Resource Groups in case of kro. Second, we should not implement them in the same way as in KubeVela. That being said, for now, KubeVela gets thumbs-up for having policies and workflows baked in.
All in all, I do think that Open Application Model (OAM) is broken on a very fundamental level. KubeVela as a project is great and, in many ways is leading in certain area. Yet, the project itself is being dragged down by the issues in the OAM model. I feel that KubeVela is trying to break away from it, but is unsure how to do it.
All that being said, KubeVela is a great way to create services for developer platforms and I encourage you to try it out if you haven’t already.
Destroy
vela delete silly-demo-db --yes --wait --env prod
vela delete silly-demo --yes --wait
./dot.nu destroy $HYPERSCALER
exit