From 8006e59f39a72099484b86e9d3b2055d8e682db7 Mon Sep 17 00:00:00 2001 From: changwangss Date: Thu, 4 Jul 2024 19:37:50 -0700 Subject: [PATCH] fix optimum-intel Signed-off-by: changwangss --- .../pytorch/code-generation/quantization/Dockerfile-multiple | 2 +- .../pytorch/code-generation/quantization/requirements.txt | 2 +- .../pytorch/text-generation/quantization/requirements_GPU.txt | 2 +- .../pytorch/text-generation/quantization/requirements_sq.txt | 2 +- tests/requirements.txt | 2 +- 5 files changed, 5 insertions(+), 5 deletions(-) diff --git a/examples/huggingface/pytorch/code-generation/quantization/Dockerfile-multiple b/examples/huggingface/pytorch/code-generation/quantization/Dockerfile-multiple index a903d161614..d58375c38b4 100644 --- a/examples/huggingface/pytorch/code-generation/quantization/Dockerfile-multiple +++ b/examples/huggingface/pytorch/code-generation/quantization/Dockerfile-multiple @@ -61,7 +61,7 @@ ENV COMPOSE_DOCKER_CLI_BUILD=0 # Install torch and intel-extension-for-pytorch 2.1 RUN python3 -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu RUN python3 -m pip install intel-extension-for-pytorch intel-extension-for-transformers optimum -RUN python3 -m pip install git+https://github.com/huggingface/optimum-intel.git@f95dea1ae8966dee4d75d622e7b2468c514ba02d +RUN python3 -m pip install git+https://github.com/huggingface/optimum-intel.git@50d867c13b22c22eda451ddb67bddb8159670f85 RUN python3 -m pip install git+https://github.com/bigcode-project/bigcode-evaluation-harness@0d84db85f9ff971fa23a187a3347b7f59af288dc # Standard requirements diff --git a/examples/huggingface/pytorch/code-generation/quantization/requirements.txt b/examples/huggingface/pytorch/code-generation/quantization/requirements.txt index 5f02605d7d0..455eccd2b26 100644 --- a/examples/huggingface/pytorch/code-generation/quantization/requirements.txt +++ b/examples/huggingface/pytorch/code-generation/quantization/requirements.txt @@ -10,6 +10,6 @@ transformers >= 4.35.0 tiktoken #code_gen neural-compressor intel_extension_for_pytorch==2.3.0 -optimum-intel +git+https://github.com/huggingface/optimum-intel.git@50d867c13b22c22eda451ddb67bddb8159670f85 auto-round==0.2 git+https://github.com/bigcode-project/bigcode-evaluation-harness@094c7cc197d13a53c19303865e2056f1c7488ac1 diff --git a/examples/huggingface/pytorch/text-generation/quantization/requirements_GPU.txt b/examples/huggingface/pytorch/text-generation/quantization/requirements_GPU.txt index 65980710432..1b000e0c61b 100644 --- a/examples/huggingface/pytorch/text-generation/quantization/requirements_GPU.txt +++ b/examples/huggingface/pytorch/text-generation/quantization/requirements_GPU.txt @@ -7,7 +7,7 @@ sentencepiece != 0.1.92 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ torch==2.1.0a0 transformers -optimum-intel +git+https://github.com/huggingface/optimum-intel.git@50d867c13b22c22eda451ddb67bddb8159670f85 bitsandbytes #baichuan transformers_stream_generator tiktoken #qwen diff --git a/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt b/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt index 1ce834dda7b..02655339b5d 100644 --- a/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt +++ b/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt @@ -7,7 +7,7 @@ sentencepiece != 0.1.92 torch==2.3.0+cpu transformers==4.38.1 intel_extension_for_pytorch==2.3.0 -optimum-intel +git+https://github.com/huggingface/optimum-intel.git@50d867c13b22c22eda451ddb67bddb8159670f85 bitsandbytes #baichuan transformers_stream_generator tiktoken #qwen diff --git a/tests/requirements.txt b/tests/requirements.txt index 4ae2eb85d49..d2c2dca3f74 100644 --- a/tests/requirements.txt +++ b/tests/requirements.txt @@ -7,6 +7,7 @@ datasets==2.16.1 einops evaluate gguf +git+https://github.com/huggingface/optimum-intel.git@50d867c13b22c22eda451ddb67bddb8159670f85 git+https://github.com/intel/neural-compressor.git git+https://github.com/intel/neural-speed.git intel-extension-for-pytorch==2.3.0 @@ -16,7 +17,6 @@ mlflow nlpaug==1.1.9 onnx onnxruntime -optimum-intel peft==0.6.2 py-cpuinfo sacremoses