Troubleshoot Guide#
1. Seeing “No Data” in Grafana#
Issue
Grafana panels show “No Data” even though the container/stack is running.
Reason
The system date/time is incorrect on the device. If the system time is wrong, data timestamps fall outside Grafana’s query window.
Solution
Check the date/time using the command below:
date
Set the correct date/time manually:
sudo date -s 'YYYY-MM-DD HH:MM:SS'
Set date/time from the internet:
sudo date -s "$(wget --method=HEAD -qSO- --max-redirect=0 google.com 2>&1 | sed -n 's/^ *Date: *//p')"
2. Influx – Data Being Deleted Beyond Retention Policy (RP)#
Issue
Data appears to be deleted beyond the configured retention policy (RP).
InfluxDB 1.x deletes old data based on the retention policy duration and shard group duration.
Reason
Data is grouped into shards.
Shards are deleted only when all data inside them is older than the RP.
For RPs ≤ 2 days, shard group duration = 1 hour.
InfluxDB always expires data at RP + shard duration.
Example:
For a 1-hour RP: - Data written at 00:00 goes into the shard covering 00:00–01:00. - The shard closes at 01:00. - InfluxDB deletes the shard only when everything inside it is past the RP → at 02:00.
So the effective expiration time is 1 hour RP + 1 hour shard duration = 2 hours.
Retention Policy |
Shard Duration |
Actual Expiry |
|---|---|---|
1 hour |
1 hour |
2 hours |
2 days |
1 hour |
2 days + 1 hr |
30 days |
24 hours |
30 days + 24 hr |
Solution
Understand that this is normal and expected behavior in InfluxDB 1.x.
A 1-hour RP will always result in ~2 hours before deletion.
No configuration can force deletion exactly at the RP limit.
3. Time Series Analytics Microservice (Docker/Helm) – Takes Time to Start or Shows Python Packages Installing#
Issue
The Time Series Analytics Microservice takes time to start or displays messages about Python packages being installed.
Reason
UDF packages require several dependent packages to be installed during
runtime, as specified under udfs/requirements.txt. Once these
dependencies are installed, the Time Series Analytics microservice
initializes and starts inferencing.
Solution
No action required — wait for the time-series-analytics microservice to complete downloading the dependent packages and initialize Kapacitor to start inference.
4. Helm Deployment – Grafana Doesn’t Load or Shows 502 Bad Gateway#
Issue
Grafana UI fails to load after a Helm deployment, showing a 502 Bad Gateway error or taking more than 2 minutes to load.
Reason
Helm deployment takes time to initialize Grafana containers. During this period, the ingress/service may route requests before the pods are fully ready.
Solution
No action required — wait for the deployment to complete and for all pods to become ready.