hanzlajavaid's picture

hanzlajavaid

hanzla

AI & ML interests

Direct Preference Optimization, Supervised Finetuning, Stable Diffusion

Recent Activity

updated a model about 12 hours ago
hanzla/Qwen2-0.5B-GRPO-summary_test_v4
updated a model about 14 hours ago
hanzla/Qwen2-0.5B-GRPO-summary_test_v3
published a model about 16 hours ago
hanzla/Qwen2-0.5B-GRPO-summary_test_v4
View all activity

Organizations

ZeroGPU Explorers's profile picture Journalists on Hugging Face's profile picture MLX Community's profile picture ModularityAI's profile picture Social Post Explorers's profile picture

hanzla's activity

New activity in hanzla/Falcon3-Mamba-R1-v0 22 days ago

Ollama support

1
#1 opened 23 days ago by
ayan4m1
posted an update 27 days ago
view post
Post
1994
Hi community,

Few days back, I posted about my ongoing research on making reasoning mamba models and I found great insights from the community.

Today, I am announcing an update to the model weights. With newer checkpoints, the Falcon3 Mamba R1 model now outperforms very large transformer based LLMs (including Gemini) for Formal Logic questions of MMLU. It scores 60% on formal logic which is considered a tough subset of questions in MMLU.

I would highly appreciate your insights and suggestions on this new checkpoint.

Model Repo: hanzla/Falcon3-Mamba-R1-v0

Chat space: hanzla/Falcon3MambaReasoner