Getting started with Akita is as simple as dropping our Agent into your service or stack. Once we’re in, we gather all the data you need and surface it in the Akita app so you can easily see what your system is doing.
Below are instructions for installing an Akita Agent using Kubernetes host networking. This will allow Akita to monitor any of the pods running on a node in your staging or production environment.
To configure a daemonset with the Akita Agent that attaches to the host network you will:
- Meet the prerequisites
- Create an Akita Project
- Generate an API key for the Akita Agent
- Add your Akita credentials as a Kubernetes secret
- Create a Daemonset definition to run the Akita Agent using host networking
- Apply the Daemonset configuration and generate an API map
- Verify that the Akita Agent is working
In order to use this method, you must have:
- Unencrypted data to the API endpoint.
- Host networking enabled in Kubernetes
Give your project a name that's easy to remember – you'll need it later, when you start the Akita Agent on the CLI.
On the same Settings page, locate and click the “API Keys” tab. Click the “Set up the Akita client” button. Copy your API key secret into your favorite password manager or somewhere else you can easily access it. Also note your API key, as you will need it later.
You will need to add a bas64-encoded version of the Akita API key ID and key secret you created in a previous step to a YAML file.
You can encode your key and key secret by inserting them in the following commands:
$ echo -n $AKITA_API_KEY_ID | base64 $ echo -n $AKITA_API_KEY_SECRET | base64 -w 0
Then create or edit a YAML file to include the below using the encoded key ID and key secret values:
apiVersion: v1 kind: Secret metadata: name: akita-secrets type: Opaque data: api-key-id: <base64 encoded key id> api-key-secret: <base64 encoded key secret>
kubectl apply to create or update the secret.
Create a new Daemonset controller by entering the following in a new
daemonset.yaml file, or adding it to an existing file. This will ensure that each node is running an instance of the pod.
--- apiVersion: apps/v1 kind: DaemonSet metadata: name: akita-capture spec: selector: matchLabels: name: akita-capture template: metadata: labels: name: akita-capture spec: containers: - image: akitasoftware/cli:latest imagePullPolicy: Always name: akita args: - apidump - --service - <your service name here> - --rate-limit - "200" env: - name: AKITA_API_KEY_ID valueFrom: secretKeyRef: name: akita-secrets key: api-key-id - name: AKITA_API_KEY_SECRET valueFrom: secretKeyRef: name: akita-secrets key: api-key-secret dnsPolicy: ClusterFirst hostNetwork: true restartPolicy: Always
Now, apply the configuration you created with
kubectl apply -f <yaml file>. You should see pods named
akita-capture-xxxxx scheduled on all of the eligible nodes.
You can use
kubectl logs to verify that the Akita Agent is collecting data.
In the Akita web console, check out the incoming data on the Model page. You should see a map of your API being generated as the Akita Agent gathers data.
Then check out the Metrics and Errors page to get real-time information on the health of your app or service.
Updated 27 days ago