TeleGCT collects, normalises, and scores hardware telemetry across your entire heterogeneous compute fleet — regardless of vendor, cloud, or infrastructure layer.
// live telemetry — simulated cluster snapshot
// hardware coverage
TeleGCT normalises telemetry from every major compute vendor into a single unified schema — no matter where your workloads run.
Full telemetry from training and inference GPUs. Compute utilisation, power draw, VRAM usage, temperature, and wastage scoring per device.
Inference-grade CPU telemetry for ARM and x86. Core utilisation, memory bandwidth, power envelope, and tokens-per-watt for LLM serving workloads.
Purpose-built inference accelerator telemetry. Chip utilisation, power draw, and throughput metrics from cloud and on-premise deployments.
// what telegct measures
From raw PDU outlet watts to per-token inference cost — TeleGCT connects the dots your existing tools miss.
01
Per-device wastage scoring identifies idle hardware drawing power without doing work. Correlates GPU compute utilisation against actual wall power draw for accurate waste attribution.
02
Benchmark your models across GPU, CPU, and TPU hardware. Get tokens-per-second and tokens-per-watt data to make migration decisions with confidence before you move production workloads.
03
SNMP telemetry from PDU outlets cross-correlated with GPU metrics. See stranded rack capacity and contracted power waste — the cost your cloud bill doesn't show you.
04
Connects to any Prometheus instance, DCGM exporter, ROCm exporter, or vLLM metrics endpoint. No agent required on your nodes. No K8s dependency. Plain EC2 and VMs fully supported.
// telegct scoring formulas
// data sources
TeleGCT pulls from your existing observability stack. No new agents, no new dashboards to maintain. One URL and a read-only token is all we need.
REST / HTTPS
Any Prometheus-compatible endpoint. DCGM exporter, ROCm exporter, vLLM metrics — all supported. Basic, Bearer, OAuth2, API key auth.
gRPC
Connect directly to NVIDIA DCGM on port 5555 for richer GPU telemetry without requiring a Prometheus layer.
SNMP v2c / v3
APC, Raritan, Vertiv, ServerTech, Eaton. Outlet-level watts and amps correlated with GPU metrics for wall power attribution.
REST API
AWS CloudWatch for Inferentia and Trainium. Google Cloud Monitoring for TPU v4/v5. Azure Monitor for GPU VMs.
REST / Generic
Define your own response schema mapping. Any JSON REST API can be a TeleGCT source with zero code changes.
Agent (optional)
Lightweight binary for bare EC2 instances and VMs. Auto-detects NVIDIA and AMD hardware. Exposes /metrics over HTTPS.
We're onboarding our first design partners now. Point us at your Prometheus instance and we'll show you your wastage score in under 10 minutes.