--- base_model: - bond005/meno-tiny-0.1 - Qwen/Qwen2.5-1.5B-Instruct library_name: transformers tags: - mergekit - merge license: apache-2.0 --- # **SpaceYL/ECE_Poirot** First model merged on the ECE intelligence Lab proprietary GPUs This model has been produced by: - **LALAIN Youri**, engineering student at French Engineering School ECE - **RAGE LILIAN**, engineering student at French Engineering School ECE Under the supervision of: - **Andre-Louis Rochet**, Lecturer at ECE, Co-founder at TW3 Partners - **Paul Lemaistre**, Lecturer at ECE, CTO at TW3 Partners - **Mohammed Mounir**, Solution Architect at Exaion - **Hervé Chibois**, Infrastructure Expert at Exaion - **Des Bontés Sonafouo**, Chef de projet IT at Omnes With the contribution of: - **ECE engineering school** as sponsor and financial contributor - **François STEPHAN** as director of ECE - **Gérard REUS** as acting director of iLAB ### Supervisory structure The iLab (intelligence Lab) is a structure created by the ECE and dedicated to artificial intelligence ### About ECE ECE, a multi-program, multi-campus, and multi-sector engineering school specializing in digital engineering, trains engineers and technology experts for the 21st century, capable of meeting the challenges of the dual digital and sustainable development revolutions. ### Models Merged The following models were included in the merge: * [bond005/meno-tiny-0.1](https://huggingface.co/bond005/meno-tiny-0.1) * [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: bond005/meno-tiny-0.1 layer_range: [0, 28] - model: Qwen/Qwen2.5-1.5B-Instruct layer_range: [0, 28] merge_method: slerp base_model: Qwen/Qwen2.5-1.5B-Instruct parameters: t: - filter: self_attn value: [0, 0.25, 0.5, 0.75, 1] - filter: mlp value: [1, 0.75, 0.5, 0.25, 0] - value: 0.5 dtype: bfloat16 ```