Accelerated Linear Algebra
XLA is an open-source compiler for machine learning developed by the OpenXLA project. XLA is designed to improve the performance of machine learning models by optimizing the computation graphs at a lower level, making it particularly useful for large-scale computations and high-performance machine learning models. Key features of XLA include:
- Compilation of Computation Graphs: Compiles computation graphs into efficient machine code.
- Optimization Techniques: Applies operation fusion, memory optimization, and other techniques.
- Hardware Support: Optimizes models for various hardware, including CPUs, GPUs, and NPUs.
- Improved Model Execution Time: Aims to reduce machine learning models' execution time for both training and inference.
- Seamless Integration: Can be used with existing machine learning code with minimal changes.
OpenXLA Project
OpenXLA Project is an open-source machine learning compiler and infrastructure initiative intended to provide a common set of tools for compiling and deploying machine learning models across different frameworks and hardware platforms. It provides a modular compilation stack that can be used by major deep learning frameworks like JAX, PyTorch, and TensorFlow. The project focuses on supplying shared components for optimization, portability, and execution across CPUs, GPUs, and specialized accelerators. Its design emphasizes interoperability between frameworks and a standardized set of representations for model computation.Components
The OpenXLA ecosystem includes several core components:- XLA – A deep learning compiler that optimizes computational graphs for multiple hardware targets.
- PJRT – A runtime interface that allows different back-ends to connect to XLA through a consistent API.
- StableHLO – A high-level operator set intended to serve as a stable, portable representation for ML models across compilers and frameworks.
- Shardy – An MLIR-based system for describing and transforming models that run in distributed or multi-device environments.
- Additional profiling, testing, and integration tools maintained under the OpenXLA organization.
Users and adopters
Supported target devices
- x86-64
- ARM64
- NVIDIA GPU
- AMD GPU
- Intel GPU
- Apple GPU
- Google TPU
- AWS Trainium, Inferentia
- Cerebras
- Graphcore IPU
Governance