At the beginning of November, we announced the availability of a Helm chart to install and run a LogStream worker group on Kubernetes. That was the first of a few things you’re going to see from Cribl with regard to Kubernetes. The next one is the availability of a Helm chart to install and configure a Cribl LogStream master instance on a Kubernetes cluster, and I’m happy to announce that that is available now in our helm-charts
repository.
The logstream-master
helm chart is a simple way to install and do basic configuration on a Cribl LogStream master instance. It allows you to, with a single command line, spin up a master node, install a license key, configure one or more worker group configurations, and set the instances’ admin password and master authentication key. If you wanna just skip to the Helm chart, take a look at the Helm Chart repo.
Unlike the logsteam-workergroup helm chart, this chart has a requirement for maintaining state. In the 2.3 release of LogStream, there are four directories where state is maintained:
We looked at two different mechanisms for maintaining the state information – 1) use an external git repository to “rehydrate” the instance upon creation, or 2) use persistent storage that can be mounted on the master pod as part of its lifecycle. For the helm chart, we chose the persistent storage approach for simplicity. Also, this way the helm chart does not require a LogStream Enterprise license (remote git access is an Enterprise feature).
Due to this requirement for persistent storage, you’ll need to set up a (preferably default) storage class. The implementation details will be different for each Kubernetes cluster, but in our case, we’re using AWS’s EKS service, the EBS CSI driver, and a default storage class called ebs-sc
. Make sure to understand the limitations and challenges for the storage driver you use. (For example, with EKS and EBS as the storage class, it’s very important that your cluster is “AZ aware”, or you’ll run into problems with nodes not being able to mount the storage volumes.)
While LogStream is typically configured in the UI, the logstream-master
helm chart provides a few options to “preconfigure” the instance. These can all be configured either by creating a values
file and giving helm install
the -f <filename>
option, or via the --set
argument. The overridable values for “preconfiguration” are:
Value | Use |
---|---|
config.adminPassword
|
The password you want set for the LogStream admin user. |
config.token
|
The auth key to set up for worker access. Setting this value will configure the instance as a “master” node. |
config.license
|
The license for your LogStream instance. If you do not set this, it will default to the “free” license. |
config.groups
|
The group names to configure for the master instance. This will create a mapping for each group, which looks for the tag <groupname> , and will create the basic structure of each group’s configuration. |
The combination of both the logstream-workergroup and logstream-master helm charts gives us a very simple and quick way to deploy a distributed LogStream environment. Let’s say that we want to create a complete distributed environment with one worker group (named all-logs
), and we want to use a generated UUID as our authentication key,
We also like using namespaces to achieve a modicum of isolation, so let’s start out by creating a logstream
namespace, via the kubectl
command:
kubectl create namespace logstream
And, we’ll use the uuid
command to generate an auth key:
helm-charts% uuid
e448c60a-2d24-11eb-a601-c78b7f337b04
Now, we’ll use the helm
command to install the master instance:
helm install ls-master cribl/logstream-master --set config.token="e448c60a-2d24-11eb-a601-c78b7f337b04" --set config.groups=\{all-logs\} --set config.license="<license key>"
Now, if you look at the output from kubectl get all -n logstream
, you’ll see something like this:
helm-charts% kubectl get all -n logstream
NAME |
READY |
STATUS |
RESTARTS |
AGE |
||
pod/ls-master-57744fcf45-qmrkn |
1/1 |
Running |
0 |
113s |
NAME |
TYPE |
CLUSTER-IP |
EXTERNAL-IP |
PORT(S) |
AGE |
||
service/ls-master |
LoadBalancer |
172.20.137.83 |
a15<...>.us-west-2.elb.amazonaws.com |
9000:32233/TCP |
113s |
||
service/ls-master-internal |
LoadBalancer |
172.20.95.84 |
a9c<...>.us-west-2.elb.amazonaws.com |
9000:30564/TCP |
4200:32002/TCP |
113s |
NAME |
READY |
UP-TO-DATE |
AVAILABLE |
AGE |
||
deployment.apps/ls-master |
1/1 |
1 |
1 |
114s |
NAME |
DESIRED |
CURRENT |
READY |
AGE |
replicaset.apps/ls-master-57744fcf45 |
1 |
1 |
1 |
114s |
You may notice that there are two services defined:
ls-master
ls-master-internal
Those services are named after the helm chart “release” that was specified in the helm install command. Since we used ls-master
, the primary service is the ls-master
service. This service is intended for end-user access to the LogStream instance, and exposes the 9000 TCP port. The other service is ls-master-internal
, which exposes the 9000 AND the 4200 TCP ports, and is intended as the endpoint for worker groups to connect to.
Now, to create our worker group, we simple run the helm install
command for the logstream-workergroup
chart, with a couple arguments:
Value | Use |
---|---|
--set config.host="ls-master-internal" |
Tells the logstream-workergroup instance to connect to the ls-master-internal service. |
--set config.tag="all-logs" |
Sets the tag and workergroup name to all-logs (which was the group we specified creating in the logstream-master install above). |
--set config.token="e448c60a-2d24-11eb-a601-c78b7f337b04" |
Sets the authentication token to use the uuid we generated before the master install. |
So, if we run:
helm install ls-wg-all-logs cribl/logstream-workergroup --set config.token="e448c60a-2d24-11eb-a601-c78b7f337b04" --set config.tag="all-logs" --set config.host=”\"ls-master-internal" -n logstream
…this sets up the worker group and connects it to the master. Looking at the output of kubectl get all -n logstream
again, you should see something like this:
helm-charts kubectl get all -n logstream
NAME | READY |
STATUS |
RESTARTS |
AGE |
pod/ls-master-57744fcf45-qmrkn | 1/1 |
Running |
0 |
5m43s |
pod/ls-wg-all-logs-logstream-workergroup-668cb65d75-9tndm | 1/1 |
Running |
0 |
21s |
NAME | TYPE |
CLUSTER-IP |
EXTERNAL-IP |
PORT(S) |
AGE |
service/ls-master | LoadBalancer |
172.20.137.83 |
a15<...>.us-west-2.elb.amazonaws.com |
9000:32233/TCP |
5m44s |
service/ls-master-internal | LoadBalancer |
172.20.95.84 |
a9c<...>.us-west-2.elb.amazonaws.com |
9000:30564/TCP,4200:32002/TCP |
5m44s |
service/ls-wg-all-logs-logstream-workergroup | LoadBalancer |
172.20.111.33 |
ac4<...>.us-west-2.elb.amazonaws.com |
10001:30082/TCP, 9997:30381/TCP, 10080:31380/TCP, 10081:32596/TCP, 5140:32025/TCP, 8125:32546/TCP, 9200:30030/TCP |
22s |
NAME | READY |
UP-TO-DATE |
AVAILABLE |
AGE |
deployment.apps/ls-master | 1/1 |
1 |
1 |
5m44s |
deployment.apps/ls-wg-all-logs-logstream-workergroup | 1/1 |
1 |
1 |
22s |
NAME | DESIRED |
CURRENT |
READY |
AGE |
replicaset.apps/ls-master-57744fcf45 | 1 |
1 |
1 |
5m44s |
replicaset.apps/ls-wg-all-logs-logstream-workergroup-668cb65d75 | 1 |
1 |
1 |
22s |
NAME | REFERENCE |
TARGETS |
MINPODS |
MAXPODS |
REPLICAS |
AGE |
horizontalpodautoscaler.autoscaling/ls-wg-all-logs-logstream-workergroup-all-logs | Deployment/ls-wg-all-logs-logstream-workergroup-all-logs |
/50% |
2 |
10 |
0 |
22s |
You’ll notice that we have a new service ls-wg-all-logs-logstream-workergroup
– this is the load balancer that you’ll need to point all your sources to. After just a few commands, you now have a working distributed deployment of Cribl LogStream!
As I mentioned when I announced the logstream-workergroup chart, this is not “fire and forget” – we’re learning as we go, and would love feedback at each step (via the #kubernetes channel in our Slack community). Please try the charts, and let us know how we can make them better and easier to use.
Experience a full version of Cribl Stream and Cribl Edge in the cloud with pre-made sources and destinations.