Update README.md
Browse files
README.md
CHANGED
|
@@ -179,7 +179,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 179 |
device_map="auto",
|
| 180 |
)
|
| 181 |
|
| 182 |
-
prompt = "Who would
|
| 183 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
| 184 |
|
| 185 |
outputs = model.generate(
|
|
@@ -205,7 +205,7 @@ sampling_params = SamplingParams(
|
|
| 205 |
max_tokens=32768,
|
| 206 |
)
|
| 207 |
|
| 208 |
-
prompt = "Who would
|
| 209 |
outputs = llm.generate(prompt, sampling_params)
|
| 210 |
print(outputs[0].outputs[0].text)
|
| 211 |
```
|
|
|
|
| 179 |
device_map="auto",
|
| 180 |
)
|
| 181 |
|
| 182 |
+
prompt = "Who would win in a fight - a dinosaur or a cow named MooMoo?"
|
| 183 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
| 184 |
|
| 185 |
outputs = model.generate(
|
|
|
|
| 205 |
max_tokens=32768,
|
| 206 |
)
|
| 207 |
|
| 208 |
+
prompt = "Who would win in a fight - a dinosaur or a cow named MooMoo?"
|
| 209 |
outputs = llm.generate(prompt, sampling_params)
|
| 210 |
print(outputs[0].outputs[0].text)
|
| 211 |
```
|