transformers icon indicating copy to clipboard operation
transformers copied to clipboard

Improve `transformers-cli env` reporting

Open ji-huazhong opened this issue 1 year ago • 1 comments

What does this PR do?

As we're getting more issues related to specific NPUs, like:

  1. https://github.com/huggingface/accelerate/issues/2368
  2. https://github.com/hiyouga/LLaMA-Factory/issues?q=is%3Aissue+npu+is%3Aopen

This PR modifies transformers-cli env to report the NPU the user is using.

Before submitting

  • [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • [x] Did you read the contributor guideline, Pull Request section?
  • [ ] Was this discussed/approved via a Github issue or the forum? Please add a link to it if that's the case.
  • [ ] Did you make sure to update the documentation with your changes? Here are the documentation guidelines, and here are tips on formatting docstrings.
  • [ ] Did you write any new necessary tests?

Who can review?

Anyone in the community is free to review the PR once the tests have passed. Feel free to tag members/contributors who may be interested in your PR.

cc @muellerzr

ji-huazhong avatar May 24 '24 08:05 ji-huazhong

With this patch, for GPU:

(hf) lynn@LAPTOP:~/github/transformers$ transformers-cli env

Copy-and-paste the text below in your GitHub issue and FILL OUT the two last points.

- `transformers` version: 4.42.0.dev0
- Platform: Linux-5.10.16.3-microsoft-standard-WSL2-x86_64-with-glibc2.35
- Python version: 3.10.13
- Huggingface_hub version: 0.23.1
- Safetensors version: 0.4.2
- Accelerate version: 0.30.1
- Accelerate config:    not found
- PyTorch version (GPU?): 2.3.0+cu121 (True)
- Tensorflow version (GPU?): not installed (NA)
- Flax version (CPU?/GPU?/TPU?): not installed (NA)
- Jax version: not installed
- JaxLib version: not installed
- Using GPU in script?: <fill in>
- Using distributed or parallel set-up in script?: <fill in>
- GPU type: NVIDIA GeForce RTX 4060 Laptop GPU

for NPU:

(lynn) [root@localhost transformers-env]# transformers-cli env

Copy-and-paste the text below in your GitHub issue and FILL OUT the two last points.

- `transformers` version: 4.42.0.dev0
- Platform: Linux-5.10.0-60.125.0.152.oe2203.aarch64-aarch64-with-glibc2.26
- Python version: 3.8.18
- Huggingface_hub version: 0.23.0
- Safetensors version: 0.4.2
- Accelerate version: 0.30.0
- Accelerate config:    not found
- PyTorch version (GPU?): 2.1.0 (False)
- Tensorflow version (GPU?): not installed (NA)
- Flax version (CPU?/GPU?/TPU?): not installed (NA)
- Jax version: not installed
- JaxLib version: not installed
- Using GPU in script?: <fill in>
- Using distributed or parallel set-up in script?: <fill in>
- NPU type: Ascend910B1
- CANN version: 8.0.RC1

ji-huazhong avatar May 24 '24 08:05 ji-huazhong

cc @amyeroberts

ji-huazhong avatar May 25 '24 08:05 ji-huazhong

Hi @amyeroberts , it's ready for re-review. :)

ji-huazhong avatar May 29 '24 00:05 ji-huazhong

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

I think this PR is causing issues in the CI because pt_cuda_available defaults to "NA", which is a truthy value, and so the if pt_cuda_available: block is executed even if torch is not present, which causes our TF and Flax tests to fail.

Rocketknight1 avatar May 29 '24 11:05 Rocketknight1

Opened a fix at #31113!

Rocketknight1 avatar May 29 '24 12:05 Rocketknight1