# Reference implementation This example creates a local version of the CNOE reference implementation. ## Prerequisites Ensure you have the following tools installed on your computer. **Required** - [idpbuilder](https://github.com/cnoe-io/idpbuilder/releases/latest): version `0.3.0` or later - [kubectl](https://kubernetes.io/docs/tasks/tools/#kubectl): version `1.27` or later - Your computer should have at least 6 GB RAM allocated to Docker. If you are on Docker Desktop, see [this guide](https://docs.docker.com/desktop/settings/mac/). **Optional** - AWS credentials: Access Key and secret Key. If you want to create AWS resources in one of examples below. ## Installation **_NOTE:_** - If you'd like to run this in your web browser through Codespaces, please follow [the instructions here](./codespaces.md) to install instead. - _This example assumes that you run the reference implementation with the default port configguration of 8443 for the idpBuilder. If you happen to configure a different host or port for the idpBuilder, the manifests in the reference example need to be updated and be configured with the new host and port. you can use the [replace.sh](replace.sh) to change the port as desired prior to applying the manifest as instructed in the command above._ ```bash idpbuilder create --use-path-routing \ --package https://github.com/cnoe-io/stacks//ref-implementation ``` This will take ~6 minutes for everything to come up. To track the progress, you can go to the [ArgoCD UI](https://cnoe.localtest.me:8443/argocd/applications). ### What was installed? 1. **Argo Workflows** to enable workflow orchestrations. 1. **Backstage** as the UI for software catalog and templating. Source is available [here](https://github.com/cnoe-io/backstage-app). 1. **External Secrets** to generate secrets and coordinate secrets between applications. 1. **Keycloak** as the identity provider for applications. 1. **Spark Operator** to demonstrate an example Spark workload through Backstage. If you don't want to install a package above, you can remove the ArgoCD Application file corresponding to the package you want to remove. For example, if you want to remove Spark Operator, you can delete [this file](./spark-operator.yaml). The only package that cannot be removed this way is Keycloak because other packages rely on it. #### Accessing UIs - Argo CD: https://cnoe.localtest.me:8443/argocd - Argo Workflows: https://cnoe.localtest.me:8443/argo-workflows - Backstage: https://cnoe.localtest.me:8443/ - Gitea: https://cnoe.localtest.me:8443/gitea - Keycloak: https://cnoe.localtest.me:8443/keycloak/admin/master/console/ # Using it For this example, we will walk through a few demonstrations. Once applications are ready, go to the [backstage URL](https://cnoe.localtest.me:8443). Click on the Sign-In button, you will be asked to log into the Keycloak instance. There are two users set up in this configuration, and their password can be retrieved with the following command: ```bash idpbuilder get secrets ``` Use the username **`user1`** and the password value given by `USER_PASSWORD` field to login to the backstage instance. `user1` is an admin user who has access to everything in the cluster, while `user2` is a regular user with limited access. Both users use the same password retrieved above. If you want to create a new user or change existing users: 1. Go to the [Keycloak UI](https://cnoe.localtest.me:8443/keycloak/admin/master/console/). Login with the username `cnoe-admin`. Password is the `KEYCLOAK_ADMIN_PASSWORD` field from the command above. 2. Select `cnoe` from the realms drop down menu. 3. Select users tab. ## Basic Deployment Let's start by deploying a simple application to the cluster through Backstage. Click on the `Create...` button on the left, then select the `Create a Basic Deployment` template. ![img.png](images/backstage-templates.png) In the next screen, type `demo` for the name field, then click Review, then Create. Once steps run, click the Open In Catalog button to go to the entity page. ![img.png](images/basic-template-flow.png) In the demo entity page, you will notice a ArgoCD overview card associated with this entity. You can click on the ArgoCD Application name to see more details. ![img.png](images/demo-entity.png) ### What just happened? 1. Backstage created [a git repository](https://cnoe.localtest.me:8443/gitea/giteaAdmin/demo), then pushed templated contents to it. 2. Backstage created [an ArgoCD Application](https://cnoe.localtest.me:8443/argocd/applications/argocd/demo?) and pointed it to the git repository. 3. Backstage registered the application as [a component](https://cnoe.localtest.me:8443/gitea/giteaAdmin/demo/src/branch/main/catalog-info.yaml) in Backstage. 4. ArgoCD deployed the manifests stored in the repo to the cluster. 5. Backstage retrieved application health from ArgoCD API, then displayed it. ![image.png](images/basic-deployment.png) ## Argo Workflows and Spark Operator In this example, we will deploy a simple Apache Spark job through Argo Workflows. Click on the `Create...` button on the left, then select the `Basic Argo Workflow with a Spark Job` template. ![img.png](images/backstage-templates-spark.png) Type `demo2` for the name field, then click create. You will notice that the Backstage templating steps are very similar to the basic example above. Click on the Open In Catalog button to go to the entity page. ![img.png](images/demo2-entity.png) Deployment processes are the same as the first example. Instead of deploying a pod, we deployed a workflow to create a Spark job. In the entity page, there is a card for Argo Workflows, and it should say running or succeeded. You can click the name in the card to go to the Argo Workflows UI to view more details about this workflow run. When prompted to log in, click the login button under single sign on. Argo Workflows is configured to use SSO with Keycloak allowing you to login with the same credentials as Backstage login. Note that Argo Workflows are not usually deployed this way. This is just an example to show you how you can integrate workflows, backstage, and spark. Back in the entity page, you can view more details about Spark jobs by navigating to the Spark tab. ## Application with cloud resources. To deploy cloud resources, you can follow any of the instructions below: - [Cloud resource deployments via Crossplane](../crossplane-integrations/) - [Cloud resource deployments via Terraform](../terraform-integrations/) ## Notes - In these examples, we have used the pattern of creating a new repository for every app, then having ArgoCD deploy it. This is done for convenience and demonstration purposes only. There are alternative actions that you can use. For example, you can create a PR to an existing repository, create a repository but not deploy them yet, etc. - If Backstage's pipelining and templating mechanisms is too simple, you can use more advanced workflow engines like Tekton or Argo Workflows. You can invoke them in Backstage templates, then track progress similar to how it was described above.