cuEquivariance
Type: Technology Tags: CUDA, NVIDIA, GPU, Equivariant Neural Networks, Geometric Neural Networks, PyTorch, JAX, Scientific AI Related: PyTorch, JAX, cuDNN, cuTENSOR, NVIDIA-Warp, cuBLAS, NVIDIA-BioNeMo, NIM-for-OpenFold3, NIM-for-Boltz2, NIM-for-OpenFold2, NIM-for-DiffDock, NIM-for-ALCHEMI-Batched-Geometry-Relaxation, NIM-for-ALCHEMI-Batched-Molecular-Dynamics, NVIDIA-CUDA Sources: https://docs.nvidia.com/cuda/cuequivariance/, https://docs.nvidia.com/cuda/cuequivariance/changelog.html, https://github.com/NVIDIA/cuEquivariance Last Updated: 2026-04-29
Summary
cuEquivariance is NVIDIA’s Python library for building high-performance geometric/equivariant neural networks with segmented polynomials, segmented tensor products, triangular operations, and optimized CUDA kernels. Current docs expose core non-ML components plus PyTorch and JAX frontends.
Detail
Purpose
Equivariant neural networks respect symmetries such as rotations and translations, which is valuable for physical, molecular, structural biology, and materials-science models. cuEquivariance gives model authors a way to describe these operations with group representations and segmented tensor-product/polynomial descriptors, then execute them through optimized CUDA-backed PyTorch or JAX paths.
Key Features
- Core package for non-ML components such as representations, irreps, layouts, descriptors, segmented tensor products, and segmented polynomials.
- PyTorch package for modules such as segmented polynomial execution, channel-wise and fully connected tensor products, linear layers, spherical harmonics, rotation/inversion utilities, triangle attention, and triangle multiplicative update.
- JAX package for representation arrays, segmented/equivariant polynomial execution, spherical harmonics, indexed linear operations, triangle attention, and triangle multiplicative update.
- Descriptor hierarchy based on
EquivariantPolynomial,SegmentedPolynomial, andSegmentedTensorProduct. - CUDA kernel packages for PyTorch and JAX with CUDA 12 and CUDA 13 variants.
- Open-source frontend under Apache 2.0, with NVIDIA-distributed optimized CUDA operations.
Use Cases
- Drug discovery: molecular docking and force field computation (DiffDock)
- Materials science: interatomic potential models (MACE, Allegro, NequIP)
- Protein structure prediction (Boltz, Neo-1, OpenFold)
- Quantum chemistry calculations
- Molecular dynamics with machine-learned force fields
- Catalyst design and discovery
Hardware Requirements
- Linux x86_64 or aarch64 for CUDA operations packages.
- Python 3.10-3.14.
- PyTorch 2.4.0+ for torch packages.
- JAX 0.8.1+ for JAX packages.
- CUDA 12 or CUDA 13 package variants for optimized GPU kernels.
Language Bindings
- Python via JAX frontend
- Python via PyTorch frontend
- Core Python package for non-ML representation and descriptor components
Connections
- PyTorch - cuEquivariance provides PyTorch modules and CUDA operation packages.
- JAX - cuEquivariance provides JAX execution functions and array wrappers.
- cuDNN — cuDNN provides general DNN primitives; cuEquivariance provides equivariance-specific ops
- cuTENSOR — cuTENSOR handles general tensor contractions; cuEquivariance handles irreps-specific contractions
- NVIDIA-Warp — Warp enables differentiable physics simulation; cuEquivariance enables equivariant ML on molecular data
- cuBLAS — cuBLAS handles standard GEMM; cuEquivariance handles structured tensor products not expressible as standard GEMM
- NVIDIA-BioNeMo — BioNeMo structure models use geometry-aware/equivariant neural network acceleration.
- NIM-for-OpenFold3 — current OpenFold3 NIM docs call out cuEquivariance kernels in the optimized backend.
- NIM-for-Boltz2 — Boltz-style biomolecular structure prediction is adjacent to equivariant model acceleration.
- NIM-for-OpenFold2 — OpenFold-style structure prediction is part of the BioNeMo structure-model family.
- NIM-for-DiffDock — molecular docking NIM based on equivariant/geometric pose prediction.
- NIM-for-ALCHEMI-Batched-Geometry-Relaxation and NIM-for-ALCHEMI-Batched-Molecular-Dynamics — ALCHEMI NIMs use MLIP model families such as MACE, which are adjacent to cuEquivariance acceleration.
- NVIDIA-CUDA - CUDA platform underneath the optimized kernel packages.
Source Excerpts
- NVIDIA describes cuEquivariance as a Python library for high-performance geometric neural networks using segmented polynomials, triangular operations, and optimized CUDA kernels.
- Current docs organize cuEquivariance into core, JAX, and PyTorch packages.