Developers staring at GPU monitoring dashboards this week are voicing a familiar frustration with tools like nvtop. The screen reports a GPU utilization of 90 percent, yet the actual throughput of the model remains stubbornly low. This discrepancy creates a visibility gap where engineers know the hardware is working hard, but they have no way of knowing why the performance is not meeting expectations. The central tension lies in the fact that a single utilization percentage is a blunt instrument that fails to reflect the intricate flow of resources within a modern GPU.
The Mechanics of Compute and Memory SOL %
Utilyze enters this space as an open-source monitoring tool built upon the NVIDIA Nsight Perf SDK to provide a more granular view of hardware efficiency. Instead of a generic utilization figure, Utilyze tracks two primary metrics: Compute SOL % and Memory SOL %. In these calculations, SOL stands for Speed-of-Light, representing the theoretical maximum performance of the hardware. The numerator consists of real-time measurements from the compute engines, such as Tensor Cores or the FP32, FP64, and INT32 pipelines, as well as the memory subsystem, including HBM, L1, and L2 caches.
When a developer observes that Compute SOL % is the dominant metric, the workload is identified as compute-bound, meaning the arithmetic units are the primary bottleneck. Conversely, if Memory SOL % is higher, the system is memory-bound, signaling that the bottleneck is the speed at which data moves through the memory bandwidth. This distinction is critical for high-end hardware like the H100, which boasts a theoretical maximum of 2,000 TFLOPS and a memory bandwidth of 3.4 TB/s. In practice, these numbers are physical limits that AI workloads rarely hit due to architectural inefficiencies.
To bridge the gap between theoretical limits and reality, Utilyze introduces the concept of Attainable SOL %. This metric accounts for unavoidable losses, such as kernel launching overhead and thread synchronization delays. By establishing a realistic ceiling based on the specific combination of model architecture and hardware, Utilyze allows developers to see the gap between current SOL % and Attainable SOL %. This gap represents the actual performance budget that can be recovered through software optimization.
Overcoming the Nsight Compute Overhead
Historically, identifying these bottlenecks required a grueling process using Nsight Compute. Because Nsight Compute uses a replay method, it must execute each kernel multiple times to gather precise data, which often slows down execution by 10 to 100 times. This makes it virtually impossible to use in a production environment where real-time traffic is flowing. While Nsight Systems offers a lower-overhead alternative by recording execution timelines, it fails to provide the specific throughput numbers necessary for deep performance tuning.
Utilyze solves this by abandoning the replay method in favor of a rolling sampling approach. Instead of re-running kernels, the tool cycles through performance counters within specific time windows, collecting and aggregating samples. This method reduces overhead to a negligible level, enabling continuous measurement in live production environments. It effectively shifts the workflow for performance engineers from offline debugging with `ncu` or AMD's Omniperf to real-time monitoring via a dashboard.
This shift fundamentally changes how teams decide between scaling infrastructure and optimizing code. Consider an inference environment running a model with 120B parameters. If the Compute SOL % is 30 percent and the Attainable SOL % is 35 percent, the developer knows that the code is nearly as optimized as it can be for that hardware, and the only way to increase performance is to purchase more GPUs. However, if the Attainable SOL % is 65 percent, it becomes immediately clear that buying more hardware is a waste of budget and that the solution lies in code optimization.
This transition from guesswork to precise measurement transforms GPU monitoring from a status check into a strategic decision-making tool.




