fahmiaziz98 commited on
Commit
8136c86
·
1 Parent(s): 616d439
Files changed (1) hide show
  1. app.py +12 -16
app.py CHANGED
@@ -173,26 +173,22 @@ class QwenReranker(RerankerModel):
173
  )
174
 
175
  def _process_inputs(self, pairs: List[str]):
176
- """Process input pairs for Qwen model."""
 
 
 
 
 
177
  inputs = self.tokenizer(
178
- pairs,
179
- padding="max_length", # ✅ pad langsung ke max_length
180
- truncation=True, # ✅ aktifkan truncation
181
- max_length=self.max_length - len(self.prefix_tokens) - len(self.suffix_tokens),
182
  return_tensors="pt"
183
- )
184
-
185
- # Tambahkan prefix & suffix tokens
186
- for i, ele in enumerate(inputs['input_ids']):
187
- inputs['input_ids'][i] = torch.tensor(
188
- self.prefix_tokens + ele.tolist() + self.suffix_tokens,
189
- dtype=torch.long
190
- )
191
 
192
- inputs = {k: v.to(self.model.device) for k, v in inputs.items()}
193
-
194
-
195
  return inputs
 
196
 
197
  @torch.no_grad()
198
  def _compute_logits(self, inputs):
 
173
  )
174
 
175
  def _process_inputs(self, pairs: List[str]):
176
+ """Process input pairs for Qwen model with prefix and suffix."""
177
+ modified_pairs = [
178
+ self.tokenizer.decode(self.prefix_tokens) + text + self.tokenizer.decode(self.suffix_tokens)
179
+ for text in pairs
180
+ ]
181
+
182
  inputs = self.tokenizer(
183
+ modified_pairs,
184
+ padding="max_length",
185
+ truncation=True,
186
+ max_length=self.max_length,
187
  return_tensors="pt"
188
+ ).to(self.model.device)
 
 
 
 
 
 
 
189
 
 
 
 
190
  return inputs
191
+
192
 
193
  @torch.no_grad()
194
  def _compute_logits(self, inputs):