@@ -126,3 +126,93 @@ jobs:
126
126
cd DeepSpeedExamples/inference/huggingface/text-generation
127
127
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
128
128
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
129
+
130
+ - name : AutoTP test (EleutherAI/gpt-j-6b)
131
+ run : |
132
+ export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
133
+ source oneCCL/build/_install/env/setvars.sh
134
+ # modify MODEL to change the model name, other lines are the same
135
+ export MODEL=EleutherAI/gpt-j-6b
136
+ cd DeepSpeedExamples/inference/huggingface/text-generation
137
+ deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
138
+ deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
139
+
140
+ - name : AutoTP test (baichuan-inc/Baichuan-7B)
141
+ run : |
142
+ export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
143
+ source oneCCL/build/_install/env/setvars.sh
144
+ # modify MODEL to change the model name, other lines are the same
145
+ export MODEL=baichuan-inc/Baichuan-7B
146
+ cd DeepSpeedExamples/inference/huggingface/text-generation
147
+ deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
148
+ deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
149
+
150
+ # - name: AutoTP test (bigcode/starcoder)
151
+ # run: |
152
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
153
+ # source oneCCL/build/_install/env/setvars.sh
154
+ # # modify MODEL to change the model name, other lines are the same
155
+ # export MODEL=bigcode/starcoder
156
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
157
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
158
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
159
+ #
160
+ # - name: AutoTP test (tiiuae/falcon-7b)
161
+ # run: |
162
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
163
+ # source oneCCL/build/_install/env/setvars.sh
164
+ # # modify MODEL to change the model name, other lines are the same
165
+ # export MODEL=tiiuae/falcon-7b
166
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
167
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
168
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
169
+ #
170
+ # - name: AutoTP test (google/flan-t5-xl)
171
+ # run: |
172
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
173
+ # source oneCCL/build/_install/env/setvars.sh
174
+ # # modify MODEL to change the model name, other lines are the same
175
+ # export MODEL=google/flan-t5-xl
176
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
177
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
178
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
179
+ #
180
+ # - name: AutoTP test (mistralai/Mistral-7B-v0.1)
181
+ # run: |
182
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
183
+ # source oneCCL/build/_install/env/setvars.sh
184
+ # # modify MODEL to change the model name, other lines are the same
185
+ # export MODEL=mistralai/Mistral-7B-v0.1
186
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
187
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
188
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
189
+ #
190
+ # - name: AutoTP test (mosaicml/mpt-7b)
191
+ # run: |
192
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
193
+ # source oneCCL/build/_install/env/setvars.sh
194
+ # # modify MODEL to change the model name, other lines are the same
195
+ # export MODEL=mosaicml/mpt-7b
196
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
197
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
198
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
199
+ #
200
+ # - name: AutoTP test (meta-llama/Llama-2-7b-hf)
201
+ # run: |
202
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
203
+ # source oneCCL/build/_install/env/setvars.sh
204
+ # # modify MODEL to change the model name, other lines are the same
205
+ # export MODEL=meta-llama/Llama-2-7b-hf
206
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
207
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
208
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
209
+ #
210
+ # - name: AutoTP test (EleutherAI/gpt-neox-20b)
211
+ # run: |
212
+ # export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
213
+ # source oneCCL/build/_install/env/setvars.sh
214
+ # # modify MODEL to change the model name, other lines are the same
215
+ # export MODEL=EleutherAI/gpt-neox-20
216
+ # cd DeepSpeedExamples/inference/huggingface/text-generation
217
+ # deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
218
+ # deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
0 commit comments