Troubleshoot Guide#
1. Seeing “No Data” in Grafana#
Issue#
Grafana panels show “No Data” even though the container/stack is running.
Reason#
The system date/time is incorrect on the device. If the system time is wrong, data timestamps fall outside Grafana’s query window.
Solution#
Check the date/time using the command below:
date
Set the correct date/time manually:
sudo date -s 'YYYY-MM-DD HH:MM:SS'
Set date/time from the internet:
sudo date -s "$(wget --method=HEAD -qSO- --max-redirect=0 google.com 2>&1 | sed -n 's/^ *Date: *//p')"
2. Influx – Data Being Deleted Beyond Retention Policy (RP)#
Issue#
Data appears to be deleted beyond the configured retention policy (RP).
InfluxDB 1.x deletes old data based on the retention policy duration and shard group duration.
Reason#
Data is grouped into shards.
Shards are deleted only when all data inside them is older than the RP.
For RPs ≤ 2 days, shard group duration = 1 hour.
InfluxDB always expires data at RP + shard duration.
Example:
For a 1-hour RP: - Data written at 00:00 goes into the shard covering 00:00–01:00. - The shard closes at 01:00. - InfluxDB deletes the shard only when everything inside it is past the RP → at 02:00.
So the effective expiration time is 1 hour RP + 1 hour shard duration = 2 hours.
Retention Policy |
Shard Duration |
Actual Expiry |
|---|---|---|
1 hour |
1 hour |
2 hours |
2 days |
1 hour |
2 days + 1 hr |
30 days |
24 hours |
30 days + 24 hr |
Solution#
Understand that this is normal and expected behavior in InfluxDB 1.x.
A 1-hour RP will always result in ~2 hours before deletion.
No configuration can force deletion exactly at the RP limit.
3. Time Series Analytics Microservice (Docker/Helm) – Takes Time to Start or Shows Python Packages Installing#
Issue#
The Time Series Analytics Microservice takes time to start or displays messages about Python packages being installed.
Reason#
UDF packages require several dependent packages to be installed during
runtime, as specified under udfs/requirements.txt. Once these
dependencies are installed, the Time Series Analytics microservice
initializes and starts inferencing.
Solution#
No action required — wait for the time-series-analytics microservice to complete downloading the dependent packages and initialize Kapacitor to start inference.
4. Helm Deployment – Grafana Doesn’t Load or Shows 502 Bad Gateway#
Issue#
Grafana UI fails to load after a Helm deployment, showing a 502 Bad Gateway error or taking more than 2 minutes to load.
Reason#
Helm deployment takes time to initialize Grafana containers. During this period, the ingress/service may route requests before the pods are fully ready.
Solution#
No action required — wait for the deployment to complete and for all pods to become ready.