Skip to content

Add Intel GPU prototype feature to PT2.4 release blog #1690

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jul 25, 2024
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 7 additions & 0 deletions _posts/2024-07-24-pytorch2-4.md
Original file line number Diff line number Diff line change
Expand Up @@ -105,6 +105,13 @@ Pipeline Parallelism is one of the primitive parallelism techniques for deep lea

For more information on this please refer to our [documentation](https://pytorch.org/docs/main/distributed.pipelining.html) and [tutorial](https://pytorch.org/tutorials/intermediate/pipelining_tutorial.html).

### [PROTOTYPE] Intel GPU is available through source build

Intel GPU in PyTorch on Linux systems offers fundamental functionalities on Intel® Data Center GPU Max Series: eager mode and torch.compile.

For eager mode, the commonly used Aten operators are implemented by using SYCL programming language. The most performance-critical graphs and operators are highly optimized by using oneAPI Deep Neural Network (oneDNN). For torch.compile mode, Intel GPU backend is integrated to Inductor on top of Triton.

For more information for Intel GPU source build please refer to our [blog post](https://www.intel.com/content/www/us/en/developer/articles/technical/pytorch-2-4-supports-gpus-accelerate-ai-workloads.html) and [documentation](https://pytorch.org/docs/main/notes/get_start_xpu.html).

## Performance Improvements

Expand Down