Skip to content

Commit a33cb4f

Browse files
committed
Fix max sequence padding
Use transformers trainer
1 parent b99e3d7 commit a33cb4f

File tree

4 files changed

+20
-3780
lines changed

4 files changed

+20
-3780
lines changed

ptuning/evaluate.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ torchrun --standalone --nnodes=1 --nproc-per-node=$NUM_GPUS main.py \
1010
--overwrite_cache \
1111
--prompt_column content \
1212
--response_column summary \
13-
--model_name_or_path chatglm2-6b \
13+
--model_name_or_path THUDM/chatglm2-6b \
1414
--ptuning_checkpoint ./output/$CHECKPOINT/checkpoint-$STEP \
1515
--output_dir ./output/$CHECKPOINT \
1616
--overwrite_output_dir \

ptuning/main.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -178,7 +178,7 @@ def preprocess_function_eval(examples):
178178
return model_inputs
179179

180180
def preprocess_function_train(examples):
181-
max_seq_length = data_args.max_source_length + data_args.max_target_length
181+
max_seq_length = data_args.max_source_length + data_args.max_target_length + 1
182182

183183
model_inputs = {
184184
"input_ids": [],
@@ -335,7 +335,7 @@ def compute_metrics(eval_preds):
335335
tokenizer=tokenizer,
336336
data_collator=data_collator,
337337
compute_metrics=compute_metrics if training_args.predict_with_generate else None,
338-
save_prefixencoder=model_args.pre_seq_len is not None
338+
save_changed=model_args.pre_seq_len is not None
339339
)
340340

341341
# Training

0 commit comments

Comments
 (0)