1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
| input_text = ds["test"][-1]["content"] inputs = tokenzier("文本摘要: \n" + input_text + tokenzier.mask_token, return_tensors="pt") inputs = tokenzier.build_inputs_for_generation(inputs, max_gen_length=64) inputs = inputs.to("cuda") output = model.generate(**inputs, max_new_tokens=64, eos_token_id=tokenzier.eop_token_id, do_sample=True) tokenzier.decode(output[0].tolist())
import torch model = model.eval() def predict_test(): predict = [] with torch.inference_mode(): for d in ds["test"]: inputs = tokenzier("文本摘要: \n" + d["content"] + tokenzier.mask_token, return_tensors="pt") inputs = tokenzier.build_inputs_for_generation(inputs, max_gen_length=64) inputs = inputs.to("cuda") output = model.generate(**inputs, max_new_tokens=64, eos_token_id=tokenzier.eop_token_id, do_sample=True) predict.append(tokenzier.decode(output[0].tolist()).split("<|startofpiece|>")[1].replace("<|endofpiece|>", "").strip()) return predict result = predict_test()
|