mergekit-uploader's picture
Upload folder using huggingface_hub
4cc0ada verified
|
raw
history blame
2.78 kB
metadata
base_model:
  - Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1
  - nvidia/AceMath-1.5B-Instruct
  - Qwen/Qwen2.5-1.5B-Instruct
  - bond005/meno-tiny-0.1
  - hotmailuser/QwenStock-1.7B
  - fblgit/miniclaus-qw1.5B-UNAMGS-GRPO
  - Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2
  - Sakalti/Saba1-1.8B
  - hotmailuser/QwenModelStock-1.8B
  - Sakalti/SJT-1.5B-Alpha
  - lalainy/ECE-PRYMMAL-YL-1B-SLERP-V3
  - nvidia/AceInstruct-1.5B
  - Sakalti/Saba1.5-1.5B
  - fblgit/miniclaus-qw1.5B-UNAMGS
library_name: transformers
tags:
  - mergekit
  - merge

merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the SCE merge method using Qwen/Qwen2.5-1.5B-Instruct as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: bond005/meno-tiny-0.1
  - model: fblgit/miniclaus-qw1.5B-UNAMGS
  - model: fblgit/miniclaus-qw1.5B-UNAMGS-GRPO
  - model: hotmailuser/QwenModelStock-1.8B
  - model: hotmailuser/QwenStock-1.7B
  - model: lalainy/ECE-PRYMMAL-YL-1B-SLERP-V3
  - model: nvidia/AceInstruct-1.5B
  - model: nvidia/AceMath-1.5B-Instruct
  - model: Qwen/Qwen2.5-1.5B-Instruct
  - model: Sakalti/Saba1.5-1.5B
  - model: Sakalti/Saba1-1.8B
  - model: Sakalti/SJT-1.5B-Alpha
  - model: Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1
  - model: Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2
tokenizer:
  source: base
merge_method: sce
base_model: Qwen/Qwen2.5-1.5B-Instruct
dtype: bfloat16
parameters:
  int8_mask: true