xla icon indicating copy to clipboard operation
xla copied to clipboard

A machine learning compiler for GPUs, CPUs, and ML accelerators

Results 653 xla issues
Sort by recently updated
recently updated
newest added

Integrate Triton up to [243522bd](https://github.com/openai/triton/commits/243522bd337fb2e3f4245d58452b0247629f0b58)

[XLA] [NFC] Add more fine-grained tracing annotations to compilation stages Reverts 9e4f0ab3d004d1dd2948830d644e518ebab4fccd

PR #15417: Add while loop config options and optional pass pipeline immediately before unroll. Imported from GitHub PR https://github.com/openxla/xla/pull/15417 This PR adds the availability to configure while loop unroll thresholds....

Remove duplicate location of ml_dtypes - `third_party/py/ml_dtypes` as it is not used. The actively used ml_dtypes is located in `third_party/tsl/third_party/py/ml_dtypes`

See for details: https://github.com/google/jax/issues/18787

GPU

Integrate LLVM at llvm/llvm-project@0c25f85e5b88 Updates LLVM usage to match [0c25f85e5b88](https://github.com/llvm/llvm-project/commit/0c25f85e5b88)

Fix `GetExecutablePath(..)` for the case when a command flag is passed to the python executable. Currently `GetExecutablePath(..)` returns an empty string if the command `python -c ''` is executed. The...

Add `ClearEntryComputationLayout` in hlo_module_util.

Add Platform::FindExisting to find pre-existing StreamExecutors for the given Platform without creating new ones.

[RFC] [XLA][HostOffloader] Updates UsesBeforeValueDefinition to consider host offloaded computation usage Updates: the use at an async host offloaded call occurs before values that are defined in the async wrapped computation