fahmiaziz98 commited on
Commit
616d439
·
1 Parent(s): 073edba
Files changed (1) hide show
  1. app.py +14 -17
app.py CHANGED
@@ -175,25 +175,22 @@ class QwenReranker(RerankerModel):
175
  def _process_inputs(self, pairs: List[str]):
176
  """Process input pairs for Qwen model."""
177
  inputs = self.tokenizer(
178
- pairs,
179
- padding=False,
180
- truncation='longest_first',
181
- return_attention_mask=False,
182
- max_length=self.max_length - len(self.prefix_tokens) - len(self.suffix_tokens)
183
  )
184
-
 
185
  for i, ele in enumerate(inputs['input_ids']):
186
- inputs['input_ids'][i] = self.prefix_tokens + ele + self.suffix_tokens
187
-
188
- inputs = self.tokenizer.pad(
189
- inputs,
190
- padding=True,
191
- return_tensors="pt",
192
- max_length=self.max_length
193
- )
194
-
195
- for key in inputs:
196
- inputs[key] = inputs[key].to(self.model.device)
197
 
198
  return inputs
199
 
 
175
  def _process_inputs(self, pairs: List[str]):
176
  """Process input pairs for Qwen model."""
177
  inputs = self.tokenizer(
178
+ pairs,
179
+ padding="max_length", # ✅ pad langsung ke max_length
180
+ truncation=True, # ✅ aktifkan truncation
181
+ max_length=self.max_length - len(self.prefix_tokens) - len(self.suffix_tokens),
182
+ return_tensors="pt"
183
  )
184
+
185
+ # Tambahkan prefix & suffix tokens
186
  for i, ele in enumerate(inputs['input_ids']):
187
+ inputs['input_ids'][i] = torch.tensor(
188
+ self.prefix_tokens + ele.tolist() + self.suffix_tokens,
189
+ dtype=torch.long
190
+ )
191
+
192
+ inputs = {k: v.to(self.model.device) for k, v in inputs.items()}
193
+
 
 
 
 
194
 
195
  return inputs
196