Skip to content
This repository has been archived by the owner on Oct 25, 2024. It is now read-only.

Commit

Permalink
Update to 2.3.0 (#1539)
Browse files Browse the repository at this point in the history
  • Loading branch information
VincyZhang authored May 13, 2024
1 parent 2af19c7 commit d54b10e
Show file tree
Hide file tree
Showing 67 changed files with 93 additions and 90 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -89,7 +89,7 @@ function main() {
apt-get install libsm6 libxext6 -y
wget http://nz2.archive.ubuntu.com/ubuntu/pool/main/o/openssl/libssl1.1_1.1.1f-1ubuntu2.19_amd64.deb
dpkg -i libssl1.1_1.1.1f-1ubuntu2.19_amd64.deb
python -m pip install --upgrade --force-reinstall torch==2.2.0
python -m pip install --upgrade --force-reinstall torch==2.3.0
python -m pip install fastapi==0.103.2
cd ${WORKING_DIR} || exit 1
echo "test on ${test_name}"
Expand Down
2 changes: 1 addition & 1 deletion docker/Dockerfile_chatbot
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers
RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi
WORKDIR /intel-extension-for-transformers

RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \
RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \
cd /intel-extension-for-transformers && pip install -r requirements.txt && \
python setup.py install && \
cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \
Expand Down
2 changes: 1 addition & 1 deletion docker/Dockerfile_code_gen
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ COPY . /app/intel-extension-for-transformers

RUN ${PYTHON} -m pip install -r requirements.txt --no-cache-dir -f https://download.pytorch.org/whl/cpu/torch_stable.html

RUN ${PYTHON} -m pip install torch==2.2.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html
RUN ${PYTHON} -m pip install torch==2.3.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html

RUN ${PYTHON} -m pip install -r examples/huggingface/pytorch/code-generation/quantization/requirements.txt --no-cache-dir

Expand Down
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.2.0+cpu
torch==2.3.0+cpu
transformers
numpy
sentencepiece
Expand Down
4 changes: 2 additions & 2 deletions examples/huggingface/neural_speed/requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ lm-eval
sentencepiece
gguf
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.2.0+cpu
torch==2.3.0+cpu
transformers
intel_extension_for_pytorch==2.2.0
intel_extension_for_pytorch==2.3.0
tiktoken
transformers_stream_generator
Original file line number Diff line number Diff line change
Expand Up @@ -4,12 +4,12 @@ datasets >= 2.0
protobuf
sentencepiece != 0.1.92
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.2.0+cpu
torch==2.3.0+cpu
peft==0.6.2
transformers >= 4.35.0
tiktoken #code_gen
neural-compressor
intel_extension_for_pytorch==2.2.0
intel_extension_for_pytorch==2.3.0
optimum-intel
auto-round
git+https://github.com/bigcode-project/bigcode-evaluation-harness@094c7cc197d13a53c19303865e2056f1c7488ac1
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torchvision==0.17.0
torch==2.3.0
torchvision==0.18.0
onnx>=1.12
onnxruntime==1.13.1
onnxoptimizer
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torchvision==0.17.0
torch==2.3.0
torchvision==0.18.0
onnx>=1.12
onnxruntime==1.13.1
evaluate
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
transformers
torch==2.2.0
torch==2.3.0
optimum
accelerate
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.12.1
optimum
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
transformers
datasets
torchprofile
torch==2.2.0
intel_extension_for_pytorch==2.2.0
torch==2.3.0
intel_extension_for_pytorch==2.3.0
accelerate
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
datasets >= 1.8.0
torch==2.2.0
torch==2.3.0
transformers
wandb
accelerate
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
accelerate
datasets
transformers
torch==2.2.0
torch==2.3.0
neural-compressor==2.0
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ sentencepiece != 0.1.92
rouge-score
nltk
py7zr
torch==2.2.0
torch==2.3.0
transformers
protobuf
schema
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
torch==2.2.0
torch==2.3.0
numpy
transformers
datasets
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
optimum
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
torch==2.2.0
torch==2.3.0
transformers
datasets
allennlp
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,5 +6,5 @@ sentencepiece
scipy
scikit-learn
protobuf
torch==2.2.0
torch==2.3.0
evaluate
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
accelerate
torch==2.2.0
torch==2.3.0
datasets >= 1.1.3
sentencepiece != 0.1.92
transformers
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,6 @@ accelerate
datasets >= 1.1.3
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
transformers
wandb
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,6 @@ accelerate
datasets >= 1.1.3
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
transformers
wandb
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,6 @@ accelerate
datasets >= 1.1.3
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
transformers
wandb
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,6 @@ accelerate
datasets >= 1.1.3
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
transformers
wandb
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx
onnxruntime==1.13.1
mteb==1.1.1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,8 +42,8 @@ cd examples/huggingface/pytorch/text-generation/quantization
pip install -r requirements.txt
pip install neural-compressor==2.5
pip install transformers==4.35.2
pip install torch==2.2.0+cpu --index-url https://download.pytorch.org/whl/cpu
pip install intel-extension-for-pytorch==2.2.0
pip install torch==2.3.0+cpu --index-url https://download.pytorch.org/whl/cpu
pip install intel-extension-for-pytorch==2.3.0
```

# Run Quantization and evaluate INT8 accuracy
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ peft
protobuf
sentencepiece != 0.1.92
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.2.0+cpu
torch==2.3.0+cpu
transformers
bitsandbytes #baichuan
transformers_stream_generator
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,9 @@ peft
protobuf
sentencepiece != 0.1.92
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.2.0+cpu
torch==2.3.0+cpu
transformers
intel_extension_for_pytorch==2.2.0
intel_extension_for_pytorch==2.3.0
git+https://github.com/huggingface/optimum.git@e38d40ad220a180213f99b1d93d0407a826c326d
optimum-intel
bitsandbytes #baichuan
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ accelerate
datasets >= 1.8.0
sentencepiece != 0.1.92
protobuf
torch==2.2.0
torch==2.3.0
onnx>=1.12
onnxruntime==1.13.1
diffusers==0.12.1
Expand Down
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
diffusers==0.4.1
accelerate
torch==2.2.0
torchvision==0.17.0
torch==2.3.0
torchvision==0.18.0
transformers
ftfy
tensorboard
modelcards
intel_extension_for_pytorch==2.2.0
intel_extension_for_pytorch==2.3.0
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
accelerate
seqeval
datasets >= 1.1.3
torch==2.2.0
torch==2.3.0
transformers
wandb
Original file line number Diff line number Diff line change
Expand Up @@ -4,5 +4,5 @@ sentencepiece != 0.1.92
protobuf
sacrebleu >= 1.4.12
py7zr
torch==2.2.0
torch==2.3.0
transformers
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,6 @@ protobuf<3.20
py-cpuinfo
sentencepiece
tiktoken
torch==2.2.0+cpu
torch==2.3.0+cpu
transformers
transformers_stream_generator
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers
RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi
WORKDIR /intel-extension-for-transformers

RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \
RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \
cd /intel-extension-for-transformers && pip install schema==0.7.5 numpy==1.26.4 && \
python setup.py install && \
cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers
RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi
WORKDIR /intel-extension-for-transformers

RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \
RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \
cd /intel-extension-for-transformers && pip install -r requirements.txt && python setup.py install && \
cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \
cd /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat && pip install -r requirements_cpu.txt && \
Expand Down
Loading

0 comments on commit d54b10e

Please sign in to comment.