TiberiuCristianLeon commited on
Commit
fa48bf0
·
verified ·
1 Parent(s): 1e52bb5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -246,8 +246,8 @@ class Translators:
246
  def rosetta(self):
247
  model = AutoModelForCausalLM.from_pretrained(
248
  self.model_name,
249
- dtype=torch.float32, # bfloat16
250
- low_cpu_mem_usage=True,
251
  device_map="auto")
252
  tokenizer = AutoTokenizer.from_pretrained(self.model_name)
253
  context = {"tone": "Formal and precise"}
@@ -257,7 +257,6 @@ class Translators:
257
  {"role": "user", "content": self.input_text},
258
  ]
259
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
260
- print(prompt)
261
  inputs = tokenizer(prompt, return_tensors="pt").to(self.device)
262
  input_length = inputs["input_ids"].shape[1]
263
  model.eval()
 
246
  def rosetta(self):
247
  model = AutoModelForCausalLM.from_pretrained(
248
  self.model_name,
249
+ dtype=torch.bfloat16, # float32 slow
250
+ low_cpu_mem_usage=False, # True
251
  device_map="auto")
252
  tokenizer = AutoTokenizer.from_pretrained(self.model_name)
253
  context = {"tone": "Formal and precise"}
 
257
  {"role": "user", "content": self.input_text},
258
  ]
259
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
 
260
  inputs = tokenizer(prompt, return_tensors="pt").to(self.device)
261
  input_length = inputs["input_ids"].shape[1]
262
  model.eval()