Continuing from part IV of our series, this post will incorporate the DAPR runtime services into the application.

The source related to this post is contained in the azure-and-dapr-identity-example repo.

Incorporating DAPR

Up to this point, nothing we’ve done included DAPR. We’ve just used Kubernetes definitions to assemble the application components.

DAPR provides a number of building blocks that solve common development challenges for distributed applications. We will be using Service Invocation which gives us service discovery. Instead of coding routes between services explicitly, DAPR will automatically route requests coming into the Ingress to the user API and other APIs we might add in the future.

It also gives us service-to-service security. Previously we were doing TLS termination and traffic within our cluster was unencrypted. DAPR protects communciations between the sidecars using mutual TLS. This gives us another layer of security. Communications are encrypted and caller/callee both identify themselves to the other. Note that this is mutually verifying the identities of the applications, not the identity of the user accessing the service – which is still handled as before using identity flows.

DAPR provides these runtime services through a sidecar architecture. To get the DAPR runtime to create the sidecars, we just annotate our deployments with additional DAPR-specific configuration. This will include updating the Users API definitions and the Ingress definitions.

Installing DAPR Into the Cluster

First we’ll need to install DAPR into our Kubernetes cluster. Run these scripts:


These scripts encapsulate the initialization steps described in the DAPR documentation. They also create REDIS servers for state management and pub/sub events.

To ensure that DAPR is up and running, run the DAPR dashboard using the command:


Users API

For the Users API we add this configuration to the deployment definition.

annotations: "true" "users-api" "80"

Also, we no longer need the service and ingress definitions for the API. We can remove these from the cluster (from our previous work) by using the commands:

kubectl delete svc users-api-service -n azure-dapr-identity-example
kubectl delete ing users-api-ingress -n azure-dapr-identity-example


For the Ingress, we no longer need the mergeable ingress definitions. We will annotate the Ingress depolyment to allow the DAPR runtime to determine the location of the services (though we will still route requests for static web assets to the NGINX server).

We can remove the front end and master ingress definitions from the cluster (from our previous work) by using the commands:

kubectl delete ing frontend-ingress -n azure-dapr-identity-example
kubectl delete ing example-application-ingress-master -n azure-dapr-identity-example

Since the NGINX Ingress controller was installed using a Helm chart, we update the annotations that are supplied to the chart (in packages/nginx-ingress/ingress-config-template.yaml):

  annotations: "true" "nginx-ingress" "80"

We also update the ingress definition itself:

kind: Ingress
  name: example-application-ingress
  ingressClassName: nginx
    - hosts:
        - testing.local
      secretName: example-app-ingress-nginx-ingress-default-server-tls
    - host: testing.local
          - path: /
            pathType: Prefix
                name: frontend-service
                  number: 80
          - path: /v1.0/invoke
            pathType: Prefix
                name: nginx-ingress-dapr
                  number: 80

This definition defines two paths. The / path just routes requests for the static web assets to the NGINX server. The second path redirects all service invocations to the DAPR sidecar, which then routes the request to the correct endpoint within the cluster. In DAPR, service invocations are prefixed with /v1.0/invoke.

This ingress definition delegates all incoming service invocations to the DAPR runtime. DAPR does the hard part of figuring out where those services actually live.

Change in API Endpoints

Finally, this changes the end points of the users API. Previously our endpoints were:

  • /api/config/auth
  • /api/ping
  • /api/current-user/me

DAPR now augments the endpoints to be more “discoverable”. We added a DAPR application id to the users API in the annotation: "users-api"

DAPR will use this application id in the endpoint URL:

  • /v1.0/invoke/users-api/method/api/config/auth
  • /v1.0/invoke/users-api/method/api/ping
  • /v1.0/invoke/users-api/method/api/current-user/me

Note that this did not require changes to the Users Web API application itself, only to its deployment definition.

Of course, the changes to the endpoints do require a change to our front end application as well.

Running the Updated Application

Make sure the previous definitions are deleted (as described above), then run the following scripts to update the cluster:

packages/users-api/deploy/initialize-users-api.ps1 -Upgrade

Then navigate to https://testing.local:31001/.

While developing, we can still run the web application locally outside of the cluster but redirect API requests to the cluster by updating the proxy configuration:

const { createProxyMiddleware } = require("http-proxy-middleware");

module.exports = function (app) {
      target: "https://testing.local:31001",
      changeOrigin: true,
      secure: false,

Also, you can run the DAPR dashboard to see the components we’ve added: