-
Notifications
You must be signed in to change notification settings - Fork 247
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
update known issues, profiler and torch.compile doc contents. --------- Co-authored-by: Ye Ting <ting.ye@intel.com>
- Loading branch information
1 parent
e8317aa
commit 8bd78d6
Showing
13 changed files
with
173 additions
and
192 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,93 +1,6 @@ | ||
Legacy Profiler Tool (Prototype) | ||
Legacy Profiler Tool (Deprecated) | ||
================================ | ||
|
||
## Introduction | ||
|
||
The legacy profiler tool is an extension of PyTorch\* legacy profiler for profiling operators' overhead on XPU devices. With this tool, users can get the information in many fields of the run models or code scripts. User should build Intel® Extension for PyTorch\* with profiler support as default and enable this tool by a `with` statement before the code segment. | ||
|
||
## Use Case | ||
|
||
To use the legacy profiler tool, you need to build Intel® Extension for PyTorch\* from source or install it via prebuilt wheel. You also have various methods to disable this tool. | ||
|
||
### Build Tool | ||
|
||
The build option `BUILD_PROFILER` is switched on as default but you can switch it off via setting `BUILD_PROFILER=OFF` while building Intel® Extension for PyTorch\* from source. With `BUILD_PROFILER=OFF`, no profiler code will be compiled and all python scripts using profiler with XPU support will raise a runtime error to user. | ||
|
||
```bash | ||
[BUILD_PROFILER=ON] python setup.py install # build from source with profiler tool | ||
BUILD_PROFILER=OFF python setup.py install # build from source without profiler tool | ||
``` | ||
|
||
### Use Tool | ||
|
||
In your model script, write `with` statement to enable the legacy profiler tool ahead of your code snippets, as shown in the following example: | ||
|
||
```python | ||
# import all necessary libraries | ||
import torch | ||
import intel_extension_for_pytorch | ||
|
||
# these lines won't be profiled before enabling profiler tool | ||
input_tensor = torch.randn(1024, dtype=torch.float32, device='xpu:0') | ||
|
||
# enable legacy profiler tool with a `with` statement | ||
with torch.autograd.profiler_legacy.profile(use_xpu=True) as prof: | ||
# do what you want to profile here after the `with` statement with proper indent | ||
output_tensor_1 = torch.nonzero(input_tensor) | ||
output_tensor_2 = torch.unique(input_tensor) | ||
|
||
# print the result table formatted by the legacy profiler tool as your wish | ||
print(prof.key_averages().table()) | ||
``` | ||
|
||
There are a number of useful parameters defined in `torch.autograd.profiler_legacy.profile()`. Many of them are aligned with usages defined in PyTorch\*'s official profiler, such as `record_shapes`, a very useful parameter to control whether to record the shape of input tensors for each operator. To enable legacy profiler on XPU devices, pass `use_xpu=True`. For the usage of more parameters, please refer to [PyTorch\*'s tutorial page](https://pytorch.org/tutorials/recipes/recipes/profiler_recipe.html). | ||
|
||
### Disable Tool in Model Script | ||
|
||
To disable the legacy profiler tool temporarily in your model script, pass `enabled=False` to `torch.autograd.profiler_legacy.profile()`: | ||
|
||
```python | ||
with torch.autograd.profiler_legacy.profile(enabled=False, use_xpu=True) as prof: | ||
# as `enabled` is set to false, the profiler won't work on these lines of code | ||
output_tensor_1 = torch.nonzero(input_tensor) | ||
output_tensor_2 = torch.unique(input_tensor) | ||
|
||
# This print will raise an error to user as the profiler was disabled | ||
print(prof.key_averages().table()) | ||
``` | ||
|
||
### Results | ||
|
||
Using the script shown above in **Use Tool** part, you'll see the result table printed out to the console as below: | ||
|
||
![Legacy_profiler_result_1](../../images/profiler_legacy/Legacy_profiler_result_1.png) | ||
|
||
In this result, you can find several fields like: | ||
|
||
- `Name`: the name of run operators | ||
- `Self CPU %`, `Self CPU`: the time consumed by the operator itself at host excluded its children operator call. The column marked with percentage sign shows the propotion of time to total self cpu time. While an operator calls more than once in a run, the self cpu time may increase in this field. | ||
- `CPU total %`, `CPU total`: the time consumed by the operator at host included its children operator call. The column marked with percentasge sign shows the propotion of time to total cpu time. While an operator calls more than once in a run, the cpu time may increase in this field. | ||
- `CPU time avg`: the average time consumed by each once call of the operator at host. This average is calculated on the cpu total time. | ||
- `Self XPU`, `Self XPU %`: similar to `Self CPU (%)` but shows the time consumption on XPU devices. | ||
- `XPU total`: similar to `CPU total` but shows the time consumption on XPU devices. | ||
- `XPU time avg`: similar to `CPU time avg` but shows average time sonsumption on XPU devices. This average is calculated on the XPU total time. | ||
- `# of Calls`: number of call for each operators in a run. | ||
|
||
You can print result table in different styles, such as sort all called operators in reverse order via `print(prof.table(sort_by='id'))` like: | ||
|
||
![Legacy_profiler_result_2](../../images/profiler_legacy/Legacy_profiler_result_2.png) | ||
|
||
### Export to Chrome Trace | ||
|
||
You can export the result to a json file and then load it in the Chrome trace viewer (`chrome://tracing`) by add this line in your model script: | ||
|
||
```python | ||
prof.export_chrome_trace("trace_file.json") | ||
``` | ||
|
||
In Chrome trace viewer, you may find the result shows like: | ||
|
||
![Legacy_profiler_result_3](../../images/profiler_legacy/Legacy_profiler_result_3.png) | ||
|
||
For more example results, please refer to [PyTorch\*'s tutorial page](https://pytorch.org/tutorials/recipes/recipes/profiler_recipe.html). | ||
|
||
The legacy profiler tool will be deprecated from Intel® Extension for PyTorch* very soon. Please use [Kineto Supported Profiler Tool](./profiler_kineto.md) instead for profiling operators' executing time cost on Intel® GPU devices. |
Oops, something went wrong.