File tree 17 files changed +2
-17
lines changed
docs/readthedocs/source/doc/LLM/Quickstart
17 files changed +2
-17
lines changed Original file line number Diff line number Diff line change @@ -381,7 +381,7 @@ jobs:
381
381
shell : bash
382
382
run : |
383
383
python -m pip uninstall datasets -y
384
- python -m pip install transformers==4.36.0 datasets peft==0.10.0 accelerate==0.23.0
384
+ python -m pip install transformers==4.36.0 datasets peft==0.10.0
385
385
python -m pip install bitsandbytes scipy
386
386
# Specific oneapi position on arc ut test machines
387
387
if [[ "$RUNNER_OS" == "Linux" ]]; then
Original file line number Diff line number Diff line change @@ -50,7 +50,6 @@ RUN mkdir -p /ipex_llm/data && mkdir -p /ipex_llm/model && \
50
50
# install huggingface dependencies
51
51
pip install datasets transformers==4.36.0 && \
52
52
pip install fire peft==0.10.0 && \
53
- pip install accelerate==0.23.0 && \
54
53
pip install bitsandbytes && \
55
54
# get qlora example code
56
55
cd /ipex_llm && \
Original file line number Diff line number Diff line change @@ -63,7 +63,6 @@ RUN mkdir -p /ipex_llm/data && mkdir -p /ipex_llm/model && \
63
63
# install huggingface dependencies
64
64
pip install datasets transformers==4.36.0 && \
65
65
pip install fire peft==0.10.0 && \
66
- pip install accelerate==0.23.0 && \
67
66
# install basic dependencies
68
67
apt-get update && apt-get install -y curl wget gpg gpg-agent && \
69
68
# Install Intel oneAPI keys.
Original file line number Diff line number Diff line change @@ -41,7 +41,7 @@ RUN wget -O- https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRO
41
41
rm -rf IPEX-LLM && \
42
42
# install transformers & peft dependencies
43
43
pip install transformers==4.36.0 && \
44
- pip install peft==0.10.0 datasets accelerate==0.23.0 && \
44
+ pip install peft==0.10.0 datasets && \
45
45
pip install bitsandbytes scipy fire && \
46
46
# Prepare accelerate config
47
47
mkdir -p /root/.cache/huggingface/accelerate && \
Original file line number Diff line number Diff line change @@ -216,7 +216,6 @@ pip install -e .
216
216
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
217
217
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
218
218
# install transformers etc
219
- pip install accelerate==0.23.0
220
219
# to avoid https://github.com/OpenAccess-AI-Collective/axolotl/issues/1544
221
220
pip install datasets==2.15.0
222
221
pip install transformers==4.37.0
Original file line number Diff line number Diff line change @@ -22,7 +22,6 @@ pip install --pre --upgrade ipex-llm[all] --extra-index-url https://download.pyt
22
22
pip install transformers==4.36.0
23
23
pip install peft==0.10.0
24
24
pip install datasets
25
- pip install accelerate==0.23.0
26
25
pip install bitsandbytes scipy
27
26
```
28
27
Original file line number Diff line number Diff line change @@ -10,7 +10,6 @@ conda activate llm
10
10
pip install --pre --upgrade ipex-llm[all]
11
11
pip install datasets transformers==4.36.0
12
12
pip install fire peft==0.10.0
13
- pip install accelerate==0.23.0
14
13
pip install bitsandbytes scipy
15
14
```
16
15
Original file line number Diff line number Diff line change @@ -19,7 +19,6 @@ conda activate llm
19
19
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
20
20
pip install transformers==4.36.0 datasets
21
21
pip install trl peft==0.10.0
22
- pip install accelerate==0.23.0
23
22
pip install bitsandbytes
24
23
```
25
24
Original file line number Diff line number Diff line change @@ -17,7 +17,6 @@ pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-exte
17
17
pip install transformers==4.36.0 datasets
18
18
pip install fire peft==0.10.0
19
19
pip install oneccl_bind_pt==2.1.100 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ # necessary to run distributed finetuning
20
- pip install accelerate==0.23.0
21
20
pip install bitsandbytes scipy
22
21
```
23
22
Original file line number Diff line number Diff line change @@ -13,7 +13,6 @@ conda create -n llm python=3.11
13
13
conda activate llm
14
14
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
15
15
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
16
- pip install accelerate==0.23.0
17
16
pip install bitsandbytes==0.43.0
18
17
pip install datasets==2.18.0
19
18
pip install --upgrade transformers==4.36.0
Original file line number Diff line number Diff line change @@ -15,7 +15,6 @@ pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-exte
15
15
pip install transformers==4.36.0 datasets
16
16
pip install fire peft==0.10.0
17
17
pip install oneccl_bind_pt==2.1.100 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ # necessary to run distributed finetuning
18
- pip install accelerate==0.23.0
19
18
pip install bitsandbytes scipy
20
19
```
21
20
Original file line number Diff line number Diff line change @@ -15,7 +15,6 @@ pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-exte
15
15
pip install transformers==4.36.0 datasets
16
16
pip install fire peft==0.10.0
17
17
pip install oneccl_bind_pt==2.1.100 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ # necessary to run distributed finetuning
18
- pip install accelerate==0.23.0
19
18
pip install bitsandbytes scipy
20
19
```
21
20
Original file line number Diff line number Diff line change @@ -18,7 +18,6 @@ pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-exte
18
18
pip install transformers==4.36.0 datasets
19
19
pip install fire peft==0.10.0
20
20
pip install oneccl_bind_pt==2.1.100 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ # necessary to run distributed finetuning
21
- pip install accelerate==0.23.0
22
21
pip install bitsandbytes scipy
23
22
# configures OneAPI environment variables
24
23
source /opt/intel/oneapi/setvars.sh # necessary to run before installing deepspeed
Original file line number Diff line number Diff line change @@ -19,7 +19,6 @@ conda activate llm
19
19
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
20
20
pip install transformers==4.36.0 datasets
21
21
pip install peft==0.10.0
22
- pip install accelerate==0.23.0
23
22
pip install bitsandbytes scipy
24
23
```
25
24
Original file line number Diff line number Diff line change @@ -19,7 +19,6 @@ conda activate llm
19
19
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
20
20
pip install transformers==4.36.0 datasets
21
21
pip install peft==0.10.0
22
- pip install accelerate==0.23.0
23
22
pip install bitsandbytes scipy trl
24
23
```
25
24
Original file line number Diff line number Diff line change @@ -15,7 +15,6 @@ pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-exte
15
15
pip install transformers==4.36.0 datasets
16
16
pip install fire peft==0.10.0
17
17
pip install oneccl_bind_pt==2.1.100 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ # necessary to run distributed finetuning
18
- pip install accelerate==0.23.0
19
18
pip install bitsandbytes scipy
20
19
```
21
20
Original file line number Diff line number Diff line change @@ -132,7 +132,6 @@ pip install -e .
132
132
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
133
133
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
134
134
# install transformers etc
135
- pip install accelerate==0.23.0
136
135
# to avoid https://github.com/OpenAccess-AI-Collective/axolotl/issues/1544
137
136
pip install datasets==2.15.0
138
137
pip install transformers==4.37.0
You can’t perform that action at this time.
0 commit comments