Skip to content

Monitoring System Health

This guide covers monitoring the F1 2025 Data Collector's health, verifying data flow, and troubleshooting common issues.

Status Bar

The status bar at the top of the Collector page provides at-a-glance system health via LED indicators.

Master Control

Located on the left side of the status bar.

  • SYSTEMS LIVE (green pulsing LED) — All collectors running
  • SYSTEMS OFFLINE (muted) — Collectors stopped
  • DEPLOYING... — Configuration change in progress

External IP

A blue LED badge displays the collector's external IP address. Use this IP in the F1 game's telemetry settings.

Redis Connection Status

  • 🟢 Green — Redis database connected and operational
  • 🔴 Red — Redis connection failed or unavailable

What Redis Does:

  • Stores driver names
  • Caches configuration settings
  • Tracks race completion status

If Redis is Down:

  • Collectors will continue to work
  • Driver names may not persist
  • Race completion tracking may not function
  • Configuration changes may not save

Troubleshooting Red Redis Status:

# Check if Redis container is running
docker ps | grep redis

# Restart the container
docker restart f1-2025

# Check container logs
docker logs f1-2025

Collectors Status

Shows how many collectors are currently running out of the total configured.

Display Format: X/Y Collectors

  • X = Number of active collectors
  • Y = Total number of configured collectors

LED Colour:

  • 🟢 Green — All collectors running (X = Y)
  • 🟠 Amber — Some collectors running (X < Y)
  • Off — No collectors running (X = 0)

Rig Card Indicators

Each rig card displays status pills and live metrics for detailed per-rig monitoring.

Endpoint Status

When Observability Cloud or HEC destinations are enabled, each rig card shows endpoint health pills:

  • O11y ✓ / HEC ✓ (green) — Data successfully sent to endpoint
  • O11y ✗ / HEC ✗ (red) — Endpoint unreachable or rejecting data
  • O11y — / HEC — (muted) — No data sent yet (pending first transmission)

Endpoint Failure Banner

If an endpoint fails, a red error banner appears at the top of the page with details about the failure. This helps identify configuration issues (wrong token, unreachable URL, etc.) immediately.

Network Queue Status

Each rig card displays a queue health pill showing the state of outbound network requests:

State Colour Meaning
QUEUE OK Green All requests flowing — no drops, inflight below 50% capacity
Q:12/50 Amber Inflight requests ≥ 50% of max — backpressure building
3 DROP Red Requests are being dropped — Splunk endpoint may be unreachable or slow

Troubleshooting Queue Issues:

  • Amber (backpressure) — Temporary network slowness. Usually resolves on its own. Check endpoint health.
  • Red (drops) — Verify Splunk endpoint is reachable. Check HEC URL and tokens. May indicate network or firewall issues.

UDP Port Status

Each rig card displays the UDP port number with colour coding:

  • 🟢 Green — Collector running and ready to receive telemetry
  • Muted — Collector not running

Race Complete Flag

A flag icon with RACE COMPLETE appears as a green pill when the race session has ended (after receiving FinalClassificationData). This unlocks the driver name for editing.

Real-time Data Verification

When collectors are running and receiving data from F1 2025, live metrics appear on each rig card.

Speed

  • Display: Current vehicle speed in mph
  • Updates: Continuously during gameplay
  • Expected Behaviour:
    • 0 mph when stationary (pits, menus)
    • Fluctuating values during racing
    • Max speeds vary by track (Monaco ~180 mph, Monza ~220 mph)

Troubleshooting:

  • No speed shown = no telemetry data being received
  • Speed stuck at 0 = driver may be in menu/paused
  • Speed frozen = possible network issue or game frozen

Current Lap

  • Display: Live lap counter
  • Updates: Increments as driver completes laps
  • Expected Behaviour:
    • Lap 0 during warm-up/out lap
    • Lap 1 after crossing start/finish line
    • Increments each lap thereafter

Troubleshooting:

  • Lap not incrementing = driver may not be in race session
  • Lap counter missing = no telemetry data

Track

  • Display: Current track name
  • Examples: "Silverstone", "Monaco", "Spa-Francorchamps", "Suzuka"
  • Updates: When track loads in game

Troubleshooting:

  • No track name = telemetry not being received
  • Wrong track name = data from different session/rig

System Health Checklist

Use this checklist to verify system health before and during events.

Pre-Event Health Check

  • [ ] Redis status shows green in status bar
  • [ ] Configuration deployed with correct credentials
  • [ ] Driver names set for all active rigs
  • [ ] Master Control toggled to SYSTEMS LIVE
  • [ ] Collectors status shows X/X (all running) in green

During Event Health Check

  • [ ] All active rig UDP ports are green
  • [ ] Real-time speed values updating
  • [ ] Lap counters incrementing as races progress
  • [ ] Track names displaying correctly
  • [ ] O11y/HEC endpoint pills showing ✓
  • [ ] Queue status showing QUEUE OK
  • [ ] Memory usage within normal range (50-200 MB per collector)

Post-Event Health Check

  • [ ] RACE COMPLETE status appeared for all completed races
  • [ ] All data visible in Splunk dashboards
  • [ ] No endpoint failure banners displayed
  • [ ] Collectors stopped cleanly when Master Control toggled off

Monitoring Best Practices

Keep Interface Visible

  • Display the Collector page on a monitor during events
  • Allows quick identification of issues
  • Real-time verification that data is flowing

Watch the Status Pills

  • Endpoint pills (O11y/HEC) are the most critical indicators
  • Green ✓ = data flowing to Splunk
  • Red ✗ = immediate attention needed
  • Queue drops = network or endpoint issue

Monitor Memory Usage

Normal Memory Usage:

  • 50-150 MB per collector (typical)
  • 150-200 MB per collector (high load)
  • Total system memory should be <1 GB for 4 collectors

High Memory Usage:

  • >300 MB per collector = potential issue
  • May indicate data buildup or network problems
  • Consider restarting collectors

How to Check:

  • Memory usage displayed as an amber pill on each rig card
  • Can also check container stats:
docker stats f1-2025

Regular Status Checks

During long events, check status every 15-30 minutes:

  • All endpoint pills green?
  • Queue status OK?
  • Real-time data still updating?
  • Memory usage normal?
  • Any error banners?

Troubleshooting Guide

UDP Port Not Receiving Data

Problem: Collector running but no telemetry data appears.

Possible Causes & Solutions:

  1. F1 2025 not running on rig
  2. Start the game
  3. Verify telemetry is enabled in game settings

  4. Wrong IP address in game telemetry settings

  5. Check External IP in the status bar (blue badge)
  6. Update game telemetry settings with correct IP

  7. Network connectivity issue

  8. Ping controller IP from rig PC
  9. Check firewall rules
  10. Verify UDP port not blocked

  11. Wrong UDP port configured

  12. Verify game is sending to correct port (20777, 20778, etc.)
  13. Each rig needs unique port number

  14. Game in menu/paused

  15. Driver must be in session (practice, qualifying, race)
  16. Telemetry doesn't stream from menus

Verification Steps:

# From the rig PC, test network connectivity
ping <controller-ip>

# Check if UDP port is accessible (requires netcat)
nc -u <controller-ip> 20777

Endpoint Status Shows ✗

Problem: O11y or HEC pill is red.

Possible Causes & Solutions:

  1. Invalid token
  2. Open Config panel and verify credentials
  3. Re-enter the token and deploy configuration

  4. Incorrect URL

  5. Verify HEC URL format: https://<instance>:8088
  6. Must use HTTPS (not HTTP)

  7. Firewall blocking connection

  8. Test connectivity to endpoint
  9. Check network/firewall rules

  10. Endpoint not enabled

  11. Open Config panel and verify the toggle is on

Test HEC connectivity:

curl -k https://<splunk-hec-url>:8088/services/collector \
  -H "Authorization: Splunk <hec-token>" \
  -d '{"event": "test"}'

Queue Showing Drops

Problem: Queue pill shows red with drop count.

Possible Causes:

  • Splunk endpoint temporarily unavailable
  • Network latency or packet loss
  • HEC endpoint overloaded

Solutions:

  1. Check endpoint status pills — if also red, fix the endpoint issue first
  2. Verify network connectivity to Splunk
  3. Restart collectors (toggle Master Control off then on)

Collector Shows Stopped

Problem: Rig card not showing as running.

Possible Causes & Solutions:

  1. Master Control is off
  2. Toggle Master Control to SYSTEMS LIVE

  3. Collector crashed

  4. Check logs for errors
  5. Restart collectors (toggle Master Control off then on)

  6. Configuration issue

  7. Verify configuration is deployed
  8. Check that the correct number of rigs is configured

Redis Status is Red

Problem: Redis LED shows red in status bar.

Impact:

  • Driver names may not save
  • Configuration changes may not persist
  • Race completion tracking disabled

Solutions:

  1. Restart the container:
docker restart f1-2025
  1. Check container logs:
docker logs f1-2025

High Memory Usage

Problem: Container memory usage >300 MB or climbing continuously.

Possible Causes:

  • Network issues causing data backup
  • Too many simultaneous connections
  • Memory leak (rare)

Solutions:

  1. Monitor over time — Is it stable or growing?

  2. Restart collectors:

  3. Toggle Master Control off
  4. Wait 10 seconds
  5. Toggle Master Control on

  6. Restart entire container:

docker restart f1-2025

Log Files

The collector generates detailed logs for troubleshooting.

Accessing Logs

Via the UI:

The Collector page has a built-in log viewer accessible via the API.

Via Docker:

docker logs f1-2025

Via API:

# JSON format (default)
curl http://localhost:8501/api/logs

# Plain text format
curl "http://localhost:8501/api/logs?format=text"

# Last 20 lines
curl "http://localhost:8501/api/logs?format=text&lines=20"

Help Panel

Click the ? button in the top-right of the navigation bar to open the Help panel. It provides a quick reference for rig card indicators, queue states, and configuration options.

Next Steps