Text Generation
Transformers
Safetensors
PyTorch
English
llama
nvidia
llama-3
conversational
text-generation-inference
suhara commited on
Commit
e684ea3
·
verified ·
1 Parent(s): 0d1ddc8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -252,7 +252,7 @@ After launching a vLLM server, you can call the server with tool-call support us
252
  )
253
 
254
  >>> completion = client.chat.completions.create(
255
- model="Llama-Nemotron-Nano-v1.1",
256
  messages=[
257
  {"role": "system", "content": "detailed thinking on"},
258
  {"role": "user", "content": "My bill is $100. What will be the amount for 18% tip?"},
 
252
  )
253
 
254
  >>> completion = client.chat.completions.create(
255
+ model="Llama-Nemotron-Nano-4B-v1.1",
256
  messages=[
257
  {"role": "system", "content": "detailed thinking on"},
258
  {"role": "user", "content": "My bill is $100. What will be the amount for 18% tip?"},