Skip to content

Commit e5b0b18

Browse files
committed
add new models in autotp workflow and enable gptj, baichuan
1 parent 4560381 commit e5b0b18

File tree

1 file changed

+90
-0
lines changed

1 file changed

+90
-0
lines changed

.github/workflows/autotp.yml

Lines changed: 90 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -126,3 +126,93 @@ jobs:
126126
cd DeepSpeedExamples/inference/huggingface/text-generation
127127
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
128128
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
129+
130+
- name: AutoTP test (EleutherAI/gpt-j-6b)
131+
run: |
132+
export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
133+
source oneCCL/build/_install/env/setvars.sh
134+
# modify MODEL to change the model name, other lines are the same
135+
export MODEL=EleutherAI/gpt-j-6b
136+
cd DeepSpeedExamples/inference/huggingface/text-generation
137+
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
138+
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
139+
140+
- name: AutoTP test (baichuan-inc/Baichuan-7B)
141+
run: |
142+
export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
143+
source oneCCL/build/_install/env/setvars.sh
144+
# modify MODEL to change the model name, other lines are the same
145+
export MODEL=baichuan-inc/Baichuan-7B
146+
cd DeepSpeedExamples/inference/huggingface/text-generation
147+
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
148+
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
149+
150+
# - name: AutoTP test (bigcode/starcoder)
151+
# run: |
152+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
153+
# source oneCCL/build/_install/env/setvars.sh
154+
# # modify MODEL to change the model name, other lines are the same
155+
# export MODEL=bigcode/starcoder
156+
# cd DeepSpeedExamples/inference/huggingface/text-generation
157+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
158+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
159+
#
160+
# - name: AutoTP test (tiiuae/falcon-7b)
161+
# run: |
162+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
163+
# source oneCCL/build/_install/env/setvars.sh
164+
# # modify MODEL to change the model name, other lines are the same
165+
# export MODEL=tiiuae/falcon-7b
166+
# cd DeepSpeedExamples/inference/huggingface/text-generation
167+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
168+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
169+
#
170+
# - name: AutoTP test (google/flan-t5-xl)
171+
# run: |
172+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
173+
# source oneCCL/build/_install/env/setvars.sh
174+
# # modify MODEL to change the model name, other lines are the same
175+
# export MODEL=google/flan-t5-xl
176+
# cd DeepSpeedExamples/inference/huggingface/text-generation
177+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
178+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
179+
#
180+
# - name: AutoTP test (mistralai/Mistral-7B-v0.1)
181+
# run: |
182+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
183+
# source oneCCL/build/_install/env/setvars.sh
184+
# # modify MODEL to change the model name, other lines are the same
185+
# export MODEL=mistralai/Mistral-7B-v0.1
186+
# cd DeepSpeedExamples/inference/huggingface/text-generation
187+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
188+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
189+
#
190+
# - name: AutoTP test (mosaicml/mpt-7b)
191+
# run: |
192+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
193+
# source oneCCL/build/_install/env/setvars.sh
194+
# # modify MODEL to change the model name, other lines are the same
195+
# export MODEL=mosaicml/mpt-7b
196+
# cd DeepSpeedExamples/inference/huggingface/text-generation
197+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
198+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
199+
#
200+
# - name: AutoTP test (meta-llama/Llama-2-7b-hf)
201+
# run: |
202+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
203+
# source oneCCL/build/_install/env/setvars.sh
204+
# # modify MODEL to change the model name, other lines are the same
205+
# export MODEL=meta-llama/Llama-2-7b-hf
206+
# cd DeepSpeedExamples/inference/huggingface/text-generation
207+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
208+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
209+
#
210+
# - name: AutoTP test (EleutherAI/gpt-neox-20b)
211+
# run: |
212+
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
213+
# source oneCCL/build/_install/env/setvars.sh
214+
# # modify MODEL to change the model name, other lines are the same
215+
# export MODEL=EleutherAI/gpt-neox-20
216+
# cd DeepSpeedExamples/inference/huggingface/text-generation
217+
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
218+
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1

0 commit comments

Comments
 (0)