Commit Graph

21 Commits

Author SHA1 Message Date
Yu, Guangye b1dc2b71a3
Add auto-detection for Intel GPUs (#14386)
## Summary

This PR intends to enable `--torch-backend=auto` to detect Intel GPUs
automatically:
- On Linux, detection is performed using the `lspci` command via
`Display controller` id.
- On Windows, ~~detection is done via a `powershell` query to
`Win32_VideoController`~~. Skip support for now—revisit once a better
solution is available.

Currently, Intel GPUs (XPU) do not rely on specific driver or toolkit
versions to distribute different PyTorch wheels.

## Test Plan

<!-- How was it tested? -->
On Linux:

![image](https://github.com/user-attachments/assets/f7f238e3-a797-42ea-b8fa-9b028dfd4db5)
~~On Windows:

![image](https://github.com/user-attachments/assets/a10d774e-1cb9-431b-bb85-e3e8225df98f)~~

---------

Co-authored-by: Charlie Marsh <charlie.r.marsh@gmail.com>
2025-07-09 13:31:08 +00:00
Charlie Marsh a82c210cab
Add auto-detection for AMD GPUs (#14176)
## Summary

Allows `--torch-backend=auto` to detect AMD GPUs. The approach is fairly
well-documented inline, but I opted for `rocm_agent_enumerator` over
(e.g.) `rocminfo` since it seems to be the recommended approach for
scripting:
https://rocm.docs.amd.com/projects/rocminfo/en/latest/how-to/use-rocm-agent-enumerator.html.

Closes https://github.com/astral-sh/uv/issues/14086.

## Test Plan

```
root@rocm-jupyter-gpu-mi300x1-192gb-devcloud-atl1:~# ./uv-linux-libc-11fb582c5c046bae09766ceddd276dcc5bb41218/uv pip install torch --torch-backend=auto
Resolved 11 packages in 251ms
Prepared 2 packages in 6ms
Installed 11 packages in 257ms
 + filelock==3.18.0
 + fsspec==2025.5.1
 + jinja2==3.1.6
 + markupsafe==3.0.2
 + mpmath==1.3.0
 + networkx==3.5
 + pytorch-triton-rocm==3.3.1
 + setuptools==80.9.0
 + sympy==1.14.0
 + torch==2.7.1+rocm6.3
 + typing-extensions==4.14.0
```

---------

Co-authored-by: Zanie Blue <contact@zanie.dev>
2025-06-21 15:21:06 +00:00
Lucas Vittor 563e9495ba
Replace cuda124 with cuda128 (#14168)
## Summary

<!-- What's the purpose of the change? What does it do, and why? -->

Replace wrong `cuda124` version to the correct `cuda128` version in
torch docs

## Test Plan

<!-- How was it tested? -->
2025-06-20 16:05:17 -04:00
Charlie Marsh 6c096246d8
Remove preview label from `--torch-backend` (#14119)
This is now used in enough places that I'm comfortable committing to
maintaining it under our versioning policy.

Closes #14091.
2025-06-17 16:49:00 -04:00
FishAlchemist d653fbb133
doc: Sync PyTorch integration index for CUDA and ROCm versions from PyTorch website. (#14100)
## Summary
Just to sync the documentation with PyTorch's officially recommended
installation method.

![image](https://github.com/user-attachments/assets/7be0aea6-b51b-4083-acae-fbe8aa79c363)
**Change:**
* CUDA 12.1 to CUDA 12.6
* CUDA 12.4 to CUDA 12.8
* ROCm 6.2 to ROCm 6.3
## Test Plan
Run doc server in local.
Result:
<details><summary>CUDA 12.6</summary>
<p>


![image](https://github.com/user-attachments/assets/962a1058-3922-4709-a57f-200939d0a397)

![image](https://github.com/user-attachments/assets/0dae693f-2dc1-4d3e-9186-d26aa84c7d73)

</p>
</details> 
<details><summary>CUDA 12.8</summary>
<p>


![image](https://github.com/user-attachments/assets/dc12d09b-f8f2-41d3-b185-cb1e4e8b062b)

![image](https://github.com/user-attachments/assets/1b5490e3-6265-4bad-8af2-eab0a207adab)

</p>
</details> 
<details><summary>ROCm6</summary>
<p>


![image](https://github.com/user-attachments/assets/3b17a24f-3210-4a99-a81b-f8f2f5578b73)

![image](https://github.com/user-attachments/assets/c7baae55-14e5-45d0-94a0-164ab31bbffc)

</p>
</details>
2025-06-17 09:55:03 -04:00
Karim Abou Zeid 5dae9ac5f2
Update referenced CUDA version in `pytorch.md` (#13899)
## Summary

torch 2.7 by default targets CUDA 12.6
2025-06-08 20:21:07 -04:00
Charlie Marsh 990c59ddb6
Add ROCm example to the PyTorch guide (#13200)
## Summary

Closes
https://github.com/astral-sh/uv/issues/13197#issuecomment-2839707240.
2025-04-29 15:13:13 -04:00
Charlie Marsh 241b013600
Upgrade PyTorch guide to CUDA 12.8 and PyTorch 2.7 (#13199) 2025-04-29 15:08:13 -04:00
Maxime ec997b6fd2
Update PyTorch/Python version tip (#12402)
PyTorch 2.6.0 supports Python 3.13, and publishes wheels for it. Update
the tip to reflect this.


## Summary

Clarify docs.

## Test Plan

Look for "cp313" at the following URLs:
- [x] https://download.pytorch.org/whl/cu124/torch/
- [x] https://download.pytorch.org/whl/torch/
2025-03-23 17:23:08 -04:00
Charlie Marsh bf12cdbd37
Add experimental `--torch-backend` to the PyTorch guide (#12317) 2025-03-19 12:52:01 -04:00
Charlie Marsh a1690203cf
Avoid fallback to PyPI in mixed CPU/CUDA example (#11115)
## Summary

This is roughly equivalent, but gets the non-`+cpu` macOS build from the
PyTorch index rather than PyPI. It seems a bit simpler? Though up for
debate.
2025-03-02 22:35:32 -05:00
Charlie Marsh 7531bb8669
Upgrade PyTorch version in documentation (#11114)
## Summary

PyTorch 2.6.0 is now GA!
2025-01-30 14:05:13 -05:00
Charlie Marsh 658ca89fb1
Use `sys_platform` in lieu of `platform_system` in PyTorch docs (#11113)
## Summary

This is what we use in the lockfile. I believe it's preferable, and it's
also more concise.
2025-01-30 19:01:52 +00:00
Charlie Marsh 9f8cd92bd3
Use positive (rather than negative) markers in PyTorch examples (#11112)
## Summary

Maybe slightly controversial because it's more verbose, but we really
want to limit these indexes to Linux and Windows, rather than ignoring
them on Darwin. E.g., we'd also want to ignore them on other platforms.

Further down, I use markers that look like this in the more complete
examples, so this feels more consistent.
2025-01-30 14:00:35 -05:00
Charlie Marsh 70507b7898
Add PyTorch XPU instructions to the PyTorch guide (#11109)
## Summary

Closes https://github.com/astral-sh/uv/issues/11079.
2025-01-30 13:48:52 -05:00
Charlie Marsh 129a75e2d0
Add meta descriptions for remaining guides (#10450)
## Summary

Closes https://github.com/astral-sh/uv/issues/10418.
2025-01-09 17:39:00 -05:00
Charlie Marsh 2c68dfd4a9
Backtrack to non-local versions when wheels are missing platform support (#10046)
## Summary

This is yet another variation on
https://github.com/astral-sh/uv/pull/9928, with a few minor changes:

1. It only applies to local versions (e.g., `2.5.1+cpu`).
2. It only _considers_ the non-local version as an alternative (e.g.,
`2.5.1`).
3. It only _considers_ the non-local alternative if it _does_ support
the unsupported platform.
4. Instead of failing, it falls back to using the local version.

So, this is far less strict, and is effectively designed to solve
PyTorch but nothing else. It's also not user-configurable, except by way
of using `environments` to exclude platforms.
2024-12-20 19:11:27 +00:00
Charlie Marsh c188b11b0a
Fix typo in PyTorch docs (#9249)
## Summary

Raised at https://x.com/stefanofusai/status/1858981714942718163/photo/1.
2024-11-19 15:41:30 -06:00
Charlie Marsh b09d5fcdd9
Add minimum version to PyTorch guide (#9247)
## Summary

Closes https://github.com/astral-sh/uv/issues/9224.
2024-11-19 15:21:43 -06:00
Charlie Marsh dea2a040f0
Add documentation for using uv with PyTorch (#9210)
## Summary

Now that we have all the pieces in place, this PR adds some dedicated
documentation to enable a variety of PyTorch setups.

This PR is downstream of #6523 and builds on the content in there; #6523
will merge first, and this PR will follow.
2024-11-19 01:09:52 +00:00
baggiponte a88a3e5eba
docs: Add PyTorch installation guide (#6523)
Hello there! First real docs PR for uv.

1. I expect this will be rewritten a gazillion times to have a
consistent tone with the rest of the docs, despite me trying to stick to
it as best as I could. Feel free to edit!
2. I went super on the verbose mode, while also providing a callout with
a TLDR on top. Scrap anything you feel it's redundant!
3. I placed the guide under `integrations` since Charlie added the
FastAPI integration there.

## Summary

<!-- What's the purpose of the change? What does it do, and why? -->
Addresses #5945

## Test Plan

<!-- How was it tested? -->
I just looked at the docs on the dev server of mkdocs if it looked nice.

**I could not test the commands that I wrote work** outside of macOS. If
someone among contributors has a Windows/Linux laptop, it should be
enough, even for the GPU-supported versions: I expect the installation
will just break once torch checks for CUDA (perhaps even at runtime).

---------

Co-authored-by: Santiago Castro <bryant1410@gmail.com>
Co-authored-by: Charlie Marsh <charlie.r.marsh@gmail.com>
2024-11-18 20:02:55 -05:00