Skip to content

Commit

Permalink
bugfix: remove duplicate load_peft_model (#124)
Browse files Browse the repository at this point in the history
  • Loading branch information
HamidShojanazeri authored Aug 22, 2023
2 parents 03faba6 + fcc817e commit 971c079
Showing 1 changed file with 1 addition and 5 deletions.
6 changes: 1 addition & 5 deletions inference/inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -99,11 +99,7 @@ def main(
print("Skipping the inference as the prompt is not safe.")
sys.exit(1) # Exit the program with an error status

if peft_model:
model = load_peft_model(model, peft_model)

model.eval()
batch = tokenizer(user_prompt, padding='max_length', truncation=True,max_length=max_padding_length,return_tensors="pt")
batch = tokenizer(user_prompt, padding='max_length', truncation=True, max_length=max_padding_length, return_tensors="pt")

batch = {k: v.to("cuda") for k, v in batch.items()}
start = time.perf_counter()
Expand Down

0 comments on commit 971c079

Please sign in to comment.