Chris-Alexiuk commited on
Commit
d10ce65
·
verified ·
1 Parent(s): 42f62a4

Include expected behaviour for Reasoning ON

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -90,6 +90,7 @@ Llama-3.1-Nemotron-Nano-8B-v1 is a general purpose reasoning and chat model inte
90
  2. We recommend setting temperature to `0.6`, and Top P to `0.95` for Reasoning ON mode
91
  3. We recommend using greedy decoding for Reasoning OFF mode
92
  4. We have provided a list of prompts to use for evaluation for each benchmark where a specific template is required
 
93
 
94
  You can try this model out through the preview API, using this link: [Llama-3.1-Nemotron-Nano-8B-v1](https://build.nvidia.com/nvidia/llama-3_1-nemotron-nano-8b-v1).
95
 
 
90
  2. We recommend setting temperature to `0.6`, and Top P to `0.95` for Reasoning ON mode
91
  3. We recommend using greedy decoding for Reasoning OFF mode
92
  4. We have provided a list of prompts to use for evaluation for each benchmark where a specific template is required
93
+ 5. The model will include `<think></think>` if no reasoning was necessary in Reasoning ON model, this is expected behaviour
94
 
95
  You can try this model out through the preview API, using this link: [Llama-3.1-Nemotron-Nano-8B-v1](https://build.nvidia.com/nvidia/llama-3_1-nemotron-nano-8b-v1).
96