diff --git a/examples/deprecated/pytorch/nlp/huggingface_models/text-to-image/inference/README.md b/examples/deprecated/pytorch/nlp/huggingface_models/text-to-image/inference/README.md index cde82e0ef1f..45e69ca1868 100644 --- a/examples/deprecated/pytorch/nlp/huggingface_models/text-to-image/inference/README.md +++ b/examples/deprecated/pytorch/nlp/huggingface_models/text-to-image/inference/README.md @@ -9,7 +9,7 @@ The script ```run_sd.py``` is based on [runwayml/stable-diffusion-v1-5](https:// ``` conda install mkl mkl-include -y conda install jemalloc gperftools -c conda-forge -y -pip install torch==1.13.1 --extra-index-url https://download.pytorch.org/whl/cpu +pip install torch==1.13.1 --index-url https://download.pytorch.org/whl/cpu pip install intel_extension_for_pytorch==1.13.0 pip install -r requirements.txt ``` diff --git a/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/README.md b/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/README.md index e0d8e3a8312..6864ebb9b7c 100644 --- a/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/README.md +++ b/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/README.md @@ -10,7 +10,7 @@ python version requests equal or higher than 3.9 due to [text evaluation library ```bash pip install oneccl_bind_pt==2.4.0 --index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ -pip install -r requirements_cpu_woq.txt +pip install -r requirements_cpu_woq.txt --extra-index-url https://download.pytorch.org/whl/cpu ``` @@ -119,19 +119,14 @@ Pytorch and Intel-extension-for-pytorch version for intel GPU > 2.1 are required ```bash pip install -r requirements_GPU.txt pip install transformers==4.38.1 # llama use 4.38.1 -git clone https://github.com/intel/intel-extension-for-pytorch.git ipex-gpu -cd ipex-gpu -git submodule update --init --recursive -export USE_AOT_DEVLIST='pvc,ats-m150' -export BUILD_WITH_CPU=OFF +pip install oneccl_bind_pt==2.3.100+xpu --index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +pip install torch==2.3.1+cxx11.abi intel-extension-for-pytorch==2.3.110+xpu --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ export LD_LIBRARY_PATH=${CONDA_PREFIX}/lib/:$LD_LIBRARY_PATH export OCL_ICD_VENDORS=/etc/OpenCL/vendors export CCL_ROOT=${CONDA_PREFIX} source /opt/intel/oneapi/setvars.sh --force export LLM_ACC_TEST=1 - -python setup.py install ``` ## Run diff --git a/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_GPU.txt b/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_GPU.txt index 5cfe114895d..a59a55b13cd 100644 --- a/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_GPU.txt +++ b/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_GPU.txt @@ -4,8 +4,6 @@ diffusers peft protobuf sentencepiece != 0.1.92 ---extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ -torch==2.3.110 transformers bitsandbytes #baichuan transformers_stream_generator diff --git a/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_cpu_woq.txt b/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_cpu_woq.txt index be40cd8d967..dc2602a020c 100644 --- a/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_cpu_woq.txt +++ b/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/transformers/weight_only/text-generation/requirements_cpu_woq.txt @@ -3,8 +3,7 @@ datasets >= 2.0 peft protobuf sentencepiece != 0.1.92 ---extra-index-url https://download.pytorch.org/whl/cpu -torch==2.4.0+cpu +torch==2.4.0 transformers bitsandbytes #baichuan transformers_stream_generator @@ -15,5 +14,4 @@ lm-eval>=0.4.2 huggingface_hub numba tbb ---extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ intel-extension-for-pytorch==2.4.0