Skip to content

Commit cd44c56

Browse files
authored
More chagnes to mlperf setup to support llama2 70b and mixtral 8x7b (#185)
1 parent 8eb7c5c commit cd44c56

File tree

6 files changed

+26
-8
lines changed

6 files changed

+26
-8
lines changed

benchmarks/mlperf/main.py

Lines changed: 9 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@
2323

2424
import mlperf_loadgen as lg
2525

26-
_MLPERF_ID = "mixtral-8x7b"
26+
_MLPERF_ID = "llama2-70b"
2727

2828
sys.path.insert(0, os.getcwd())
2929

@@ -135,7 +135,12 @@ def get_args():
135135
'eg. {"tok_input_len": "tok_input_length"}'
136136
),
137137
)
138-
138+
parser.add_argument(
139+
"--mlperf-conf-id",
140+
type=str,
141+
default=_MLPERF_ID,
142+
help="When given overrides the default user.conf path",
143+
)
139144
args = parser.parse_args()
140145
return args
141146

@@ -156,8 +161,8 @@ def main():
156161
else:
157162
user_conf = args.user_conf
158163

159-
settings.FromConfig(args.mlperf_conf, _MLPERF_ID, args.scenario)
160-
settings.FromConfig(user_conf, _MLPERF_ID, args.scenario)
164+
settings.FromConfig(args.mlperf_conf, args.mlperf_conf_id, args.scenario)
165+
settings.FromConfig(user_conf, args.mlperf_conf_id, args.scenario)
161166
log.info("Mlperf config: %s", args.mlperf_conf)
162167
log.info("User config: %s", user_conf)
163168

benchmarks/mlperf/scripts/generate_server_accuracy_run.sh

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,9 @@
1414

1515
source run_utils.sh
1616

17+
export TOKENIZER_PATH=meta-llama/Llama-2-70b-chat-hf
18+
export DATASET_PREFIX=""
19+
export MODEL_ID="llama2-70b"
1720
DATASET_NAME=$(get_dataset_name ${DATASET_TYPE})
1821
export DATASET_PATH=${DATA_DISK_DIR}/${DATASET_NAME}.pkl
1922
export API_URL=${API_URL}
@@ -33,6 +36,7 @@ echo "OUTPUT_ACCURACY_JSON_PATH: ${OUTPUT_ACCURACY_JSON_PATH}"
3336
echo "USER_CONFIG: ${USER_CONFIG}"
3437

3538
mkdir -p ${OUTPUT_LOG_DIR} && cp ../${USER_CONFIG} ${OUTPUT_LOG_DIR}
39+
MIXTRAL_COLS_RENAME="{\"tok_input_len\": \"tok_input_length\", \"tok_ref_output_len\": \"tok_output_length\"}"
3640

3741
# Accuracy Run
3842
cd ../ && python3 main.py \
@@ -53,6 +57,8 @@ cd ../ && python3 main.py \
5357
--tokenizer-path ${TOKENIZER_PATH} \
5458
--log-interval ${LOG_INTERVAL} \
5559
--num-client-threads ${NUM_CLIENT_THREADS} \
60+
--mlperf-conf-id "${MODEL_ID}" \
61+
--rename-dataset-cols "${MIXTRAL_COLS_RENAME}" \
5662
--output-log-dir ${OUTPUT_LOG_DIR} 2>&1 | tee ${OUTPUT_LOG_DIR}/server_accuracy_log.log
5763

5864
# Eval Run

benchmarks/mlperf/scripts/generate_server_audit_run.sh

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,9 @@
1313
# limitations under the License.
1414

1515
source run_utils.sh
16-
16+
export TOKENIZER_PATH=meta-llama/Llama-2-70b-chat-hf
17+
export DATASET_PREFIX=""
18+
export MODEL_ID="llama2-70b"
1719
DATASET_NAME=$(get_dataset_name ${DATASET_TYPE})
1820
export DATASET_PATH=${DATA_DISK_DIR}/${DATASET_NAME}.pkl
1921
export API_URL=${API_URL}
@@ -50,4 +52,5 @@ cd ../ && python3 main.py \
5052
--tokenizer-path ${TOKENIZER_PATH} \
5153
--log-interval ${LOG_INTERVAL} \
5254
--num-client-threads ${NUM_CLIENT_THREADS} \
55+
--mlperf-conf-id "${MODEL_ID}" \
5356
--output-log-dir ${OUTPUT_LOG_DIR} 2>&1 | tee ${OUTPUT_LOG_DIR}/server_audit_log.log

benchmarks/mlperf/scripts/generate_server_performance_run.sh

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,9 @@
1414

1515
source run_utils.sh
1616

17+
export TOKENIZER_PATH=meta-llama/Llama-2-70b-chat-hf
18+
export DATASET_PREFIX=""
19+
export MODEL_ID="llama2-70b"
1720
DATASET_NAME=$(get_dataset_name ${DATASET_TYPE})
1821
export DATASET_PATH=${DATA_DISK_DIR}/${DATASET_NAME}.pkl
1922
export API_URL=${API_URL}
@@ -52,4 +55,5 @@ cd ../ && python3 main.py \
5255
--log-interval ${LOG_INTERVAL} \
5356
--num-client-threads ${NUM_CLIENT_THREADS} \
5457
--rename-dataset-cols "${MIXTRAL_COLS_RENAME}" \
58+
--mlperf-conf-id "${MODEL_ID}" \
5559
--output-log-dir ${OUTPUT_LOG_DIR} 2>&1 | tee ${OUTPUT_LOG_DIR}/server_performance_log.log

benchmarks/mlperf/scripts/run_utils.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515

1616
# Tokenizer
1717
# export TOKENIZER_PATH=meta-llama/Llama-2-70b-chat-hf
18-
export DATASET_PREFIX=mixtral
18+
export DATASET_PREFIX="mixtral-"
1919
export TOKENIZER_PATH=mistralai/Mixtral-8x7B-Instruct-v0.1
2020
export NUM_CLIENT_THREADS=${NUM_CLIENT_THREADS:=600}
2121

@@ -25,7 +25,7 @@ export LOADGEN_RUN_TIMESTAMP=$(TZ=America/Los_Angeles date +%Y%m%d%H%M%S%Z)
2525
get_dataset_name() {
2626
dataset_type=$1
2727
if [ ${dataset_type} = "full" ]
28-
then echo "${DATASET_PREFIX}-processed-data"
28+
then echo "${DATASET_PREFIX}processed-data"
2929
elif [ ${dataset_type} = "calibration" ]
3030
then echo "${DATASET_PREFIX}-processed-calibration-data"
3131
fi

benchmarks/mlperf/user.conf

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ mixtral-8x7b.Server.min_query_count = 15000
2222

2323
# These fields should be defined and overridden by user.conf.
2424
*.Offline.target_qps = 5.0
25-
llama2-70b.Server.target_qps = 1.0
25+
llama2-70b.Server.target_qps = 14.5
2626
mixtral-8x7b.Server.target_qps = 11.0
2727

2828

0 commit comments

Comments
 (0)