DachengZhang commited on
Commit
5217b9f
·
verified ·
1 Parent(s): 9925e22

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -55,7 +55,7 @@ pipeline_tag: text-generation
55
  - Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
56
  - Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
57
  - The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
58
- - The long-chat version supports extremely long texts, extending up to 200K tokens.
59
  - The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
60
  <table style="border-collapse: collapse; width: 100%;">
61
  <tr>
@@ -71,7 +71,7 @@ pipeline_tag: text-generation
71
  - Orion-14B series models including:
72
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
73
  - **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
74
- - **Orion-14B-LongChat:** This model is optimized for long context lengths more than 200k tokens and demonstrates performance comparable to proprietary models on long context evaluation sets.
75
  - **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
76
  - **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
77
  - **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
 
55
  - Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
56
  - Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
57
  - The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
58
+ - The long-chat version supports extremely long texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
59
  - The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
60
  <table style="border-collapse: collapse; width: 100%;">
61
  <tr>
 
71
  - Orion-14B series models including:
72
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
73
  - **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
74
+ - **Orion-14B-LongChat:** The long-context version excels at handling extremely lengthy texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
75
  - **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
76
  - **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
77
  - **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.