Under Construction
This documentation is a work in progress. Visit regularly to see the improvements we're making.
Create a Pipeline for Log Data
As they run, services generate log data to provide visibility into what’s happening. Log data can include anything from benign informational messages to full-blown errors. You want to know about errors, but too many informational messages can become distracting noise.
You’ll be able to use MDAI to filter out unnecessary log lines. First, lets set up log generators and a pipeline to collect the logs and display telemetry data. Then we’ll add a collector and collect logs.
Generate Log Data
We’ve built in some synthetics that will emulate log generation. Go to our Configs repo > synthetics directory and copy the log generators into your working directory.
Deploy Log Generators to Your Cluster
Let’s create 3 different log generators.
The first emits logs for random services named service####
, where ####
is a random number.
kubectl apply -f ./loggen_services.yaml
The second is a noisy log generator for a particular service (service1234
by default).
kubectl apply -f ./loggen_service_noisy.yaml
The third is an excessively noisy version of the second log generator, but for a different service (service4321
by default).
kubectl apply -f ./loggen_service_xtra_noisy.yaml
Verify That Logs Are Being Generated
If the log generators are running, get the pods.
kubectl get pods -n mdai -l app.kubernetes.io/part-of=mdai-log-generator
You should see a list of pods similar to the following.
NAME READY STATUS RESTARTS AGE
mdai-logger-d8bb6f897-f6jt6 1/1 Running 0 101s
mdai-logger-d8bb6f897-tzh6f 1/1 Running 0 101s
mdai-logger-d8bb6f897-w7d4n 1/1 Running 0 101s
mdai-logger-noisy-77fcbf8b9f-g4c9c 1/1 Running 0 84s
mdai-logger-noisy-77fcbf8b9f-tsg2s 1/1 Running 0 84s
mdai-logger-noisy-77fcbf8b9f-xrp8s 1/1 Running 0 84s
mdai-logger-xnoisy-686cb6465f-cdpvm 1/1 Running 0 71s
mdai-logger-xnoisy-686cb6465f-tdjkd 1/1 Running 0 71s
mdai-logger-xnoisy-686cb6465f-xdzvp 1/1 Running 0 71s
Verify that logs are being generated by using the kubectl logs
command with any one of the pods.
kubectl logs -n mdai {pod name}
For example, the logs from the pod named mdai-logger-xnoisy-686cb6465f-xdzvp
should contain a variety of log levels.
2025-02-06T04:43:50+00:00 - service4321 - INFO - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:50+00:00 - service4321 - INFO - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:50+00:00 - service4321 - INFO - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:50+00:00 - service4321 - ERROR - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:50+00:00 - service4321 - ERROR - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:50+00:00 - service4321 - WARNING - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - ERROR - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - WARNING - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - WARNING - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - WARNING - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - WARNING - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - ERROR - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - ERROR - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - INFO - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - INFO - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
2025-02-06T04:43:51+00:00 - service4321 - INFO - The algorithm successfully executed, triggering neural pathways and producing a burst of optimized data streams.
Tip
If you installed K9s, you can use it for observing the logs. To launch the K9s application, enter k9s
in the terminal window where you launched the cluster (for the correct context). Use the arrow keys to select one of the log generators, then press l (lowercase letter l). Press ESC to exit the log window.
Set Up a Collector
From the MDAI Example Config repo, copy the
otel_config.yaml
into your working directory.Deploy the Otel config to your cluster.
kubectl apply -f otel_config.yaml
Verify that the collector is running in Kubernetes.
kubectl -n mdai get pods --selector app.kubernetes.io/name=gateway-collector
Output should similar to the following.
NAME READY STATUS RESTARTS AGE gateway-collector-74f69ccf5b-896gk 1/1 Running 0 2m28s
Collect Logs
We’ll use Fluentd to capture the synthetic log streams you created in step 1 in stdout/stderr, and forward them to the collector.
From the MDAI Example Config repo, copy the
loggen_fluent_config.yaml
into your working directory.helm upgrade --install --repo https://fluent.github.io/helm-charts fluent fluentd -f loggen_fluent_config.yaml
Confirm that Fluentd is running in the default namespace.
kubectl get pods
This also gives you the name of the Fluentd pod.
Look at the Fluentd logs, which should indicate that various pod log files are being accessed.
kubectl logs fluent-fluentd-<your_pod_id_here>
You should see log lines similar to the following.
2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/mdai-prometheus-node-exporter-957tk_mdai_node-exporter-5b9ba8e8394e00b48b54534c075097cdc1bff99d2e634684d450645062a3a390.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/mdai-prometheus-node-exporter-957tk_mdai_node-exporter-f6b7eacef004e60992050358d14a1e0780fa56c5cab236b66dcc4aa8debf7ecd.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/noisy-mclogface-7ccd56984f-6wmrb_mdai_noisy-mclogface-c60bf1edeacfca17682f7bdb87213485ec4ec18571b334c5953ba74b25dda802.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/noisy-mclogface-7ccd56984f-6wmrb_mdai_noisy-mclogface-c85adcb072728b92991a967a4e7503c88e4be5ce0e9d324e10cb56d67065a6fd.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/noisy-mclogface-7ccd56984f-7x4qd_mdai_noisy-mclogface-3d159b3e51d63dae5686122bfa03e21fd3f5faf99ae5563b588bbaaeab50d2d1.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/noisy-mclogface-7ccd56984f-7x4qd_mdai_noisy-mclogface-b343a4e7d4e995249a934119898368dc11a827a695fb8cb20309dbe67a50a7b9.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/noisy-mclogface-7ccd56984f-fwnxk_mdai_noisy-mclogface-609ac69e83dd09e70948f3ad19439132246c2b4ee8072f7a21b845224b0aee00.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/noisy-mclogface-7ccd56984f-fwnxk_mdai_noisy-mclogface-ff2407b2005ed2f2f00157840743748800c78078343846acdf0b8539518632d6.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/opentelemetry-operator-57bd64b65d-mnf5d_mdai_manager-e1d482d5de3054b584edf8aee1501856d5fe0b61f3c80626c539c5435c48f7f3.log 2025-02-10 01:41:18 +0000 [info]: #0 following tail of /var/log/containers/prometheus-kube-prometheus-stack-prometheus-0_mdai_config-reloader-3f50b71d40173e561f5c2110c9ec00241e2057768a4ec6f5267e88589efc6812.log
Step 4: Verify your OTel collector is receiving data
You just finished connecting your fluentD instance to your Otel collector. You should see a healthy stream of data flowing through the collector.
kubectl logs gateway-collector-<your_pod_id_here> --tail 10
You should see log lines similar to the following.
2025-03-26T21:32:16.442Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 185}
2025-03-26T21:32:17.446Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 184}
2025-03-26T21:32:20.444Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 173}
2025-03-26T21:32:22.440Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 12}
2025-03-26T21:32:23.442Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 61}
2025-03-26T21:32:24.403Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/observer", "resource logs": 114, "log records": 694}
2025-03-26T21:32:24.442Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 31}
2025-03-26T21:32:25.444Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 238}
2025-03-26T21:32:26.446Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/storage", "resource logs": 1, "log records": 101}
2025-03-26T21:32:27.509Z info Logs {"kind": "exporter", "data_type": "logs", "name": "debug/observer", "resource logs": 95, "log records": 3228}
Success
You’re now sending logs to mdai. See our usage guide for more information about taking control of your telemetry streams.