Skip to content

Commit 6c9c6a1

Browse files
tye1sramakinteljingxu10ZhaoqiongZ
authored
Doc and script changes for 2.1.40 (#4628)
* Update README.md and known_issues.md * correct image tag in README of dockerfile (#4638) * update miniconda to miniforge * update compile bundle to stop displaying torch-ccl version * remove optimizer_fusion_cpu and split sgd * fix advanced config not correctly displayed issue --------- Co-authored-by: Srikanth Ramakrishna <[email protected]> Co-authored-by: Jing Xu <[email protected]> Co-authored-by: Zheng, Zhaoqiong <[email protected]>
1 parent 189cdfc commit 6c9c6a1

File tree

8 files changed

+48
-19
lines changed

8 files changed

+48
-19
lines changed

docker/Dockerfile.compile

+6-6
Original file line numberDiff line numberDiff line change
@@ -37,17 +37,17 @@ RUN useradd -m -s /bin/bash ubuntu && \
3737
USER ubuntu
3838
WORKDIR /home/ubuntu
3939

40-
RUN curl -fsSL -v -o miniconda.sh -O https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh && \
41-
bash miniconda.sh -b -p ./miniconda3 && \
42-
rm miniconda.sh && \
43-
echo "source ~/miniconda3/bin/activate" >> ./.bashrc
40+
RUN curl -fsSL -v -o miniforge.sh -O https://github.com/conda-forge/miniforge/releases/download/24.1.2-0/Miniforge3-24.1.2-0-Linux-x86_64.sh && \
41+
bash miniforge.sh -b -p ./miniforge3 && \
42+
rm miniforge.sh && \
43+
echo "source ~/miniforge3/bin/activate" >> ./.bashrc
4444

4545
FROM base AS dev
4646
RUN bash /basekit_driver_install_helper.sh dev
4747
COPY --chown=ubuntu:ubuntu . ./intel-extension-for-pytorch/
4848
RUN cp ./intel-extension-for-pytorch/scripts/compile_bundle.sh ./ && \
4949
sed -i "s/VER_IPEX=.*/VER_IPEX=/" compile_bundle.sh
50-
RUN . ./miniconda3/bin/activate && \
50+
RUN . ./miniforge3/bin/activate && \
5151
conda create -y -n compile_py310 python=3.10 && conda activate compile_py310 && \
5252
bash compile_bundle.sh /opt/intel/oneapi/compiler/latest /opt/intel/oneapi/mkl/latest /opt/intel/oneapi/ccl/latest /opt/intel/oneapi/mpi/latest pvc,ats-m150,acm-g11 && \
5353
mkdir wheels && cp pytorch/dist/*.whl vision/dist/*.whl audio/dist/*.whl intel-extension-for-pytorch/dist/*.whl intel-extension-for-pytorch/ecological_libs/deepspeed/dist/*.whl ./wheels
@@ -60,7 +60,7 @@ RUN bash /basekit_driver_install_helper.sh runtime && \
6060
sudo rm /basekit_driver_install_helper.sh
6161
COPY --from=dev --chown=ubuntu:ubuntu /home/ubuntu/wheels ./wheels
6262
COPY --from=dev --chown=ubuntu:ubuntu /home/ubuntu/intel-extension-for-pytorch/tools/get_libstdcpp_lib.sh .
63-
RUN . ./miniconda3/bin/activate && \
63+
RUN . ./miniforge3/bin/activate && \
6464
conda create -y -n py310 python=3.10 && conda activate py310 && \
6565
conda install -y libpng libjpeg-turbo -c conda-forge && \
6666
python -m pip install ./wheels/*.whl && \

docker/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ export IMAGE_TYPE="xpu"
3434
To pull docker images use the following command:
3535

3636
```bash
37-
docker pull intel/intel-extension-for-pytorch:2.1.40-xpu-pip-base
37+
docker pull intel/intel-extension-for-pytorch:2.1.40-xpu
3838
```
3939
### Running container:
4040

docs/tutorials/features/advanced_configuration.md

-4
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@ The default settings for Intel® Extension for PyTorch\* are sufficient for most
88
The following build options are supported by Intel® Extension for PyTorch\*. Users who install Intel® Extension for PyTorch\* via source compilation could override the default configuration by explicitly setting a build option ON or OFF, and then build.
99

1010
| **Build Option** | **Default<br>Value** | **Description** |
11-
| ------ | ------ | ------ |
1211

1312
For above build options which can be configured to ON or OFF, users can configure them to 1 or 0 also, while ON equals to 1 and OFF equals to 0.
1413

@@ -17,13 +16,10 @@ For above build options which can be configured to ON or OFF, users can configur
1716
The following launch options are supported in Intel® Extension for PyTorch\*. Users who execute AI models on XPU could override the default configuration by explicitly setting the option value at runtime using environment variables, and then launch the execution.
1817

1918
| **Launch Option<br>CPU, GPU** | **Default<br>Value** | **Description** |
20-
| ------ | ------ | ------ |
2119

2220
| **Launch Option<br>GPU ONLY** | **Default<br>Value** | **Description** |
23-
| ------ | ------ | ------ |
2421

2522
| **Launch Option<br>Experimental** | **Default<br>Value** | **Description** |
26-
| ------ | ------ | ------ |
2723

2824
| **Distributed Option<br>GPU ONLY** | **Default<br>Value** | **Description** |
2925
| ------ | ------ | ------ |

docs/tutorials/known_issues.md

+4-3
Original file line numberDiff line numberDiff line change
@@ -39,9 +39,10 @@ Troubleshooting
3939
- **Problem**: Random bad termination after AI model convergence test (>24 hours) finishes.
4040
- **Cause**: This is a random issue when some AI model convergence test execution finishes. It is not user-friendly as the model execution ends ungracefully.
4141
- **Solution**: Kill the process after the convergence test finished, or use checkpoints to divide the convergence test into several phases and execute separately.
42-
- **Problem**: Random instability issues such as page fault or atomic access violation when executing LLM inference workloads on Intel® Data Center GPU Max series cards.
43-
- **Cause**: This issue is reported on LTS driver [803.29](https://dgpu-docs.intel.com/releases/LTS_803.29_20240131.html). The root cause is under investigation.
44-
- **Solution**: Use active rolling stable release driver [775.20](https://dgpu-docs.intel.com/releases/stable_775_20_20231219.html) or latest driver version to workaround.
42+
- **Problem**: Runtime error `munmap_chunk(): invalid pointer` when executing some scaling LLM workloads on Intel® Data Center GPU Max Series platform
43+
- **Cause**: Users targeting GPU use, must set the environment variable ‘FI_HMEM=system’ to disable GPU support in underlying libfabric as Intel® MPI Library 2021.13.1 will offload the GPU support instead. This avoids a potential bug in libfabric GPU initialization.
44+
- **Solution**: Set the environment variable ‘FI_HMEM=system’ to workaround this issue when encounter.
45+
4546
4647
## Library Dependencies
4748

docs/tutorials/releases.md

+32
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,38 @@
11
Releases
22
=============
33

4+
## 2.1.40+xpu
5+
6+
Intel® Extension for PyTorch\* v2.1.40+xpu is a minor release which supports Intel® GPU platforms (Intel® Data Center GPU Flex Series, Intel® Data Center GPU Max Series,Intel® Arc™ A-Series Graphics and Intel® Core™ Ultra Processors with Intel® Arc™ Graphics) based on PyTorch\* 2.1.0.
7+
8+
### Highlights
9+
10+
- Intel® oneAPI Base Toolkit 2024.2.1 compatibility
11+
- Intel® oneDNN v3.5 integration
12+
- Intel® oneCCL 2021.13.1 integration
13+
- Intel® Core™ Ultra Processors with Intel® Arc™ Graphics (MTL-H) support on Windows (Prototype)
14+
- Bug fixing and other optimization
15+
- Fix host memory leak [#4280](https://github.com/intel/intel-extension-for-pytorch/commit/5c252a1e34ccecc8e2e5d10ccc67f410ac7b87e2)
16+
- Fix LayerNorm issue for undefined grad_input [#4317](https://github.com/intel/intel-extension-for-pytorch/commit/619cd9f5c300a876455411bcacc470bd94c923be)
17+
- Replace FP64 device check method [#4354](https://github.com/intel/intel-extension-for-pytorch/commit/d60d45187b1dd891ec8aa2abc42eca8eda5cb242)
18+
- Fix online doc search issue [#4358](https://github.com/intel/intel-extension-for-pytorch/commit/2e957315fdad776617e24a3222afa55f54b51507)
19+
- Fix pdist unit test failure on client GPUs [#4361](https://github.com/intel/intel-extension-for-pytorch/commit/00f94497a94cf6d69ebba33ff95d8ab39113ecf4)
20+
- Remove primitive cache from conv fwd [#4429](https://github.com/intel/intel-extension-for-pytorch/commit/bb1c6e92d4d11faac5b6fc01b226d27950b86579)
21+
- Fix sdp bwd page fault with no grad bias [#4439](https://github.com/intel/intel-extension-for-pytorch/commit/d015f00011ad426af33bb970451331321417bcdb)
22+
- Fix implicit data conversion [#4463](https://github.com/intel/intel-extension-for-pytorch/commit/d6987649e58af0da4964175aed3286aef16c78c9)
23+
- Fix compiler version parsing issue [#4468](https://github.com/intel/intel-extension-for-pytorch/commit/50b2b5933b6df6632a18d76bdec46b638750dc48)
24+
- Fix irfft invalid descriptor [#4480](https://github.com/intel/intel-extension-for-pytorch/commit/3e60e87cf011b643cc0e72d82c10b28417061d97)
25+
- Change condition order to fix out-of-bound access in index [#4495](https://github.com/intel/intel-extension-for-pytorch/commit/8b74d6c5371ed0bd442279be42b0d454cb2b31b3)
26+
- Add parameter check in embedding bag [#4504](https://github.com/intel/intel-extension-for-pytorch/commit/57174797bab9de2647abb8fdbcda638b0c694e01)
27+
- Add the backward implementation for rms norm [#4527](https://github.com/intel/intel-extension-for-pytorch/commit/e4938e0a9cee15ffe2f8d205e0228c1842a5735c)
28+
- Fix attn_mask for sdpa beam_search [#4557](https://github.com/intel/intel-extension-for-pytorch/commit/80ed47655b003fa132ac264b3d3008c298865473)
29+
- Use data_ptr template instead of force data conversion [#4558](https://github.com/intel/intel-extension-for-pytorch/commit/eeb92d2f4c34f143fc76e409987543d42e68d065)
30+
- Workaround windows AOT image size over 2GB issue on Intel® Core™ Ultra Processors with Intel® Arc™ Graphics [#4407](https://github.com/intel/intel-extension-for-pytorch/commit/d7ebba7c94374bdd12883ffd45d6670b96029d11) [#4450](https://github.com/intel/intel-extension-for-pytorch/commit/550fd767b723bd9a1a799b05be5d8ce073e6faf7)
31+
32+
### Known Issues
33+
34+
Please refer to [Known Issues webpage](./known_issues.md).
35+
436
## 2.1.30+xpu
537

638
Intel® Extension for PyTorch\* v2.1.30+xpu is an update release which supports Intel® GPU platforms (Intel® Data Center GPU Flex Series, Intel® Data Center GPU Max Series and Intel® Arc™ A-Series Graphics) based on PyTorch\* 2.1.0.

docs/tutorials/technical_details.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ Optimizers are a key part of the training workloads. Intel® Extension for PyTor
1717
technical_details/optimizer_fusion_gpu
1818

1919

20-
For more detailed information, check `Optimizer Fusion on CPU <technical_details/optimizer_fusion_cpu.md>`_, `Optimizer Fusion on GPU <technical_details/optimizer_fusion_gpu.md>`_ and `Split SGD <technical_details/split_sgd.html>`_.
20+
For more detailed information, check `Optimizer Fusion on GPU <technical_details/optimizer_fusion_gpu.md>`_.
2121

2222
Ahead of Time Compilation (AOT) [GPU]
2323
-------------------------------------

scripts/build_doc.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -221,7 +221,7 @@ elif [[ ${DEVICE} == "gpu" ]]; then
221221
parse_example "../examples/gpu/inference/cpp/example-usm/CMakeLists.txt" ${MDEXAMPLE} "(marker_cppsdk_cmake_usm)" "cmake"
222222

223223
cp ${MDCONF} tutorials/features/advanced_configuration.md.bk
224-
sed -i "/^| [[:alnum:]_-]/d" ${MDCONF}
224+
#sed -i "/^| [[:alnum:]_-]/d" ${MDCONF}
225225
parse_build_options "../cmake/gpu/Options.cmake" ${MDCONF}
226226
parse_launch_options "../csrc/gpu/utils/Settings.cpp" ${MDCONF} "==========ALL=========="
227227
parse_launch_options "../csrc/gpu/utils/Settings.cpp" ${MDCONF} "==========GPU=========="

scripts/compile_bundle.sh

+3-3
Original file line numberDiff line numberDiff line change
@@ -338,8 +338,8 @@ if [ $((${MODE} & 0x02)) -ne 0 ]; then
338338
CMD="${CMD} import torchaudio; print(f'torchaudio_version: {torchaudio.__version__}');"
339339
fi
340340
CMD="${CMD} import intel_extension_for_pytorch as ipex; print(f'ipex_version: {ipex.__version__}');"
341-
if [ $((${MODE} & 0x01)) -ne 0 ]; then
342-
CMD="${CMD} import oneccl_bindings_for_pytorch as torch_ccl; print(f'torchccl_version: {torch_ccl.__version__}');"
343-
fi
341+
#if [ $((${MODE} & 0x01)) -ne 0 ]; then
342+
# CMD="${CMD} import oneccl_bindings_for_pytorch as torch_ccl; print(f'torchccl_version: {torch_ccl.__version__}');"
343+
#fi
344344
python -c "${CMD}"
345345

0 commit comments

Comments
 (0)