MoC: Mixtures of Text Chunking Learners for Retrieval-Augmented Generation System

arXiv Paper Apache 2.0 License

The Meta-chunker-1.5B-60K was fully fine-tuned on the Qwen2.5-1.5B-Instruct utilizing 60K data entries from the CRUD_MASK.jsonl and WanJuan1_MASK.jsonl, which were prepared with GPT-4o and ERNIE-3.5-128K, respectively.

Downloads last month
2
Safetensors
Model size
1.54B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Robot2050/Meta-chunker-1.5B-60K

Quantizations
1 model