OpenXLA Benchmark

PJRT: The Pluggable Just-in-Time Runtime

09 Apr 2026

To deliver the “run anywhere” half of OpenXLA’s promise, the ecosystem ships PJRT—a hardware- and framework-independent interface for ML compilers and runtimes. PJRT simplifies new hardware integration by exposing a stable C API that abstracts device management, memory allocation, and executable execution.

Core abstractions

Why this matters for portability

The plugin architecture lets hardware vendors develop support independently from the main OpenXLA or framework repositories. Two concrete examples:

From the benchmarking perspective, the PJRT layer is exactly what the paper’s portability tax measurement targets: for each hardware platform we compare the OpenXLA path (JAX → StableHLO → XLA → hardware) against the native path (vLLM/CUDA on NVIDIA, PyTorch/ROCm on AMD, direct HLO on TPU) to quantify the overhead this abstraction introduces.

Details in the paper, Section 4.