playmak3r commited on
Commit
ad655c0
·
1 Parent(s): c893725

fix model output

Browse files
Files changed (1) hide show
  1. model.py +8 -2
model.py CHANGED
@@ -24,9 +24,15 @@ def run(
24
  add_generation_prompt=False,
25
  return_tensors="pt"
26
  )
 
 
27
 
28
- outputs = model.generate(tokenized_chat.to(model.device), max_new_tokens=2048)
29
- output_text = tokenizer.decode(outputs[0])
 
 
 
 
30
  return output_text
31
 
32
  if __name__ == "__main__":
 
24
  add_generation_prompt=False,
25
  return_tensors="pt"
26
  )
27
+ input_ids = tokenized_chat.to(model.device)
28
+ input_length = input_ids.shape[1]
29
 
30
+ outputs = model.generate(input_ids, max_new_tokens=2048)
31
+ # 2. Fatiamos o tensor: pegamos do [input_length:] até o fim
32
+ # Isso isola apenas os tokens novos gerados
33
+ generated_tokens = outputs[0][input_length:]
34
+
35
+ output_text = tokenizer.decode(generated_tokens, skip_special_tokens=True)
36
  return output_text
37
 
38
  if __name__ == "__main__":