Update app.py
Browse files
app.py
CHANGED
|
@@ -246,8 +246,8 @@ class Translators:
|
|
| 246 |
def rosetta(self):
|
| 247 |
model = AutoModelForCausalLM.from_pretrained(
|
| 248 |
self.model_name,
|
| 249 |
-
dtype=torch.
|
| 250 |
-
low_cpu_mem_usage=True
|
| 251 |
device_map="auto")
|
| 252 |
tokenizer = AutoTokenizer.from_pretrained(self.model_name)
|
| 253 |
context = {"tone": "Formal and precise"}
|
|
@@ -257,7 +257,6 @@ class Translators:
|
|
| 257 |
{"role": "user", "content": self.input_text},
|
| 258 |
]
|
| 259 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
| 260 |
-
print(prompt)
|
| 261 |
inputs = tokenizer(prompt, return_tensors="pt").to(self.device)
|
| 262 |
input_length = inputs["input_ids"].shape[1]
|
| 263 |
model.eval()
|
|
|
|
| 246 |
def rosetta(self):
|
| 247 |
model = AutoModelForCausalLM.from_pretrained(
|
| 248 |
self.model_name,
|
| 249 |
+
dtype=torch.bfloat16, # float32 slow
|
| 250 |
+
low_cpu_mem_usage=False, # True
|
| 251 |
device_map="auto")
|
| 252 |
tokenizer = AutoTokenizer.from_pretrained(self.model_name)
|
| 253 |
context = {"tone": "Formal and precise"}
|
|
|
|
| 257 |
{"role": "user", "content": self.input_text},
|
| 258 |
]
|
| 259 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
|
|
|
| 260 |
inputs = tokenizer(prompt, return_tensors="pt").to(self.device)
|
| 261 |
input_length = inputs["input_ids"].shape[1]
|
| 262 |
model.eval()
|