Skip to content

[RFC]: Updates for float16/bfloat16 and for dtypes that are lacking full support in libraries #998

@rgommers

Description

@rgommers

This issue is meant to provide context for changes to dtype support in the next version of the standard.

Overview of data types implemented in various array libraries:

Summary of dtype support across array libraries

Legend: ✓ = full support, ○ = partial support, ✗ = no support

dtype NumPy PyTorch JAX CuPy MLX ndonnx DPCTL cuTile
bool
int8
int16
int32
int64 ○ [6]
uint8
uint16 ○ [1]
uint32 ○ [1]
uint64 ○ [1] ○ [6]
float32
float64 ○ [6] ○ [2] ○ [7]
complex64
complex128 ○ [6] ○ [7]
float16 ○ [8]
bfloat16 ○ [3] ○ [4]
complex32 ○ [5]
bcomplex32

Notes:

  1. PyTorch uint16/uint32/uint64 are "shell dtypes" with limited op and backend support; primarily for use with torch.compile.
  2. MLX float64 only works with CPU operations; using float64 on the GPU raises an exception.
  3. NumPy bfloat16 support is available via the ml_dtypes package, not natively.
  4. CuPy has minimal bfloat16 support via ml_dtypes.bfloat16; some gaps remain especially in cupyx.
  5. PyTorch has torch.complex32 defined, but operator coverage is limited.
  6. JAX 64-bit dtypes require opt-in via jax.config.update('jax_enable_x64', True) or the JAX_ENABLE_X64 env var; disabled by default, and 64-bit values are silently truncated to 32-bit without it.
  7. DPCTL float64/complex128 are device-dependent; only available when the target device's has_aspect_fp64 property is True.
  8. DPCTL float16 is device-dependent; only available when the target device's has_aspect_fp16 property is True.

Conclusions

  • Only a few dtypes have no caveats at all: bool, int8/int16/int32, uint8, float32
  • Support for float64 is the biggest issue: very important for scientific computing and other fields that require high accuracy, not available at all or CPU-only on several deep learning-focused libraries
  • Complex dtype support is generally spotty
  • float16 and bfloat16 don't have universal support yet, but are consistently named

Next steps

Open a PR for discussion which brings documentation on data more in line with reality, reserves the float16/bfloat16 names, and says something about dtype support that is storage-only or partial.

Metadata

Metadata

Assignees

No one assigned

    Labels

    API changeChanges to existing functions or objects in the API.RFCRequest for comments. Feature requests and proposed changes.

    Type

    No type

    Projects

    No projects

    Milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions