DoeyLLM commited on
Commit
c80c2b9
·
verified ·
1 Parent(s): c6bfe6c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -15
README.md CHANGED
@@ -1,3 +1,13 @@
 
 
 
 
 
 
 
 
 
 
1
  ## **Model Summary**
2
  This model is a fine-tuned version of **LLaMA 3.2-3B**, optimized using **LoRA (Low-Rank Adaptation)** on the [NVIDIA ChatQA-Training-Data](https://huggingface.co/datasets/nvidia/ChatQA-Training-Data). It is tailored for conversational AI, question answering, and other instruction-following tasks, with support for sequences up to 1024 tokens.
3
 
@@ -102,19 +112,8 @@ print(outputs[0]["generated_text"][-1])
102
 
103
  ## Responsibility & Safety
104
 
105
- As part of our Responsible release approach, we followed a three-pronged strategy to managing trust & safety risks:
106
 
107
- 1. Enable developers to deploy helpful, safe and flexible experiences for their target audience and for the use cases supported by Llama
108
- 2. Protect developers against adversarial users aiming to exploit Llama capabilities to potentially cause harm
109
- 3. Provide protections for the community to help prevent the misuse of our models
110
-
111
- ---
112
- license: apache-2.0
113
- datasets:
114
- - nvidia/ChatQA-Training-Data
115
- language:
116
- - en
117
- base_model:
118
- - meta-llama/Llama-3.2-3B-Instruct
119
- pipeline_tag: text-generation
120
- ---
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - nvidia/ChatQA-Training-Data
5
+ language:
6
+ - en
7
+ base_model:
8
+ - meta-llama/Llama-3.2-3B
9
+ pipeline_tag: text-generation
10
+ ---
11
  ## **Model Summary**
12
  This model is a fine-tuned version of **LLaMA 3.2-3B**, optimized using **LoRA (Low-Rank Adaptation)** on the [NVIDIA ChatQA-Training-Data](https://huggingface.co/datasets/nvidia/ChatQA-Training-Data). It is tailored for conversational AI, question answering, and other instruction-following tasks, with support for sequences up to 1024 tokens.
13
 
 
112
 
113
  ## Responsibility & Safety
114
 
115
+ As part of our responsible release strategy, we adopted a three-pronged approach to managing trust and safety risks:
116
 
117
+ Enable developers to deploy helpful, safe, and flexible experiences for their target audience and the use cases supported by the model.
118
+ Protect developers from adversarial users attempting to exploit the model’s capabilities to potentially cause harm.
119
+ Provide safeguards for the community to help prevent the misuse of the model.