AI & ML interests

Exploring Extreme Quantization techniques !

erikkaumย 
posted an update 12 days ago
view post
Post
2532
ZML just released a technical preview of their new Inference Engine: LLMD.

- Just 2.4GB container, which means fast startup times and efficient autoscaling
- Cross-Platform GPU Support: works on both NVIDIA and AMD GPUs.
- written in Zig

I just tried it out and deployed it on Hugging Face Inference Endpoints and wrote a quick guide ๐Ÿ‘‡ You can try it in like 5 minutes!

https://huggingface.co/blog/erikkaum/test-driving-llmd-inference-engine
  • 1 reply
ยท
erikkaumย 
posted an update 13 days ago
view post
Post
1980
We just released native support for @SGLang and @vllm-project in Inference Endpoints ๐Ÿ”ฅ

Inference Endpoints is becoming the central place where you deploy high performance Inference Engines.

And that provides the managed infra for it. Instead of spending weeks configuring infrastructure, managing servers, and debugging deployment issues, you can focus on what matters most: your AI model and your users ๐Ÿ™Œ
Abhaykoulย 
posted an update 15 days ago
view post
Post
2969
๐ŸŽ‰ Dhanishtha-2.0-preview-0725 is Now Live

The Intermediate Thinking Model just got even better.
With the new update, Dhanishtha is now sharper, smarter, and trained further on tool use

๐Ÿง  What Makes Dhanishtha Different?
Unlike standard COT models that give one-shot responses, Dhanishtha thinks in layers:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

HelpingAI/Dhanishtha-2.0-preview-0725
Parveshiiiiย 
posted an update 24 days ago
view post
Post
2644
๐Ÿง  Glimpses of AGI โ€” A Vision for All Humanity
What if AGI wasnโ€™t just a distant dreamโ€”but a blueprint already unfolding?

Iโ€™ve just published a deep dive called Glimpses of AGI, exploring how scalable intelligence, synthetic reasoning, and alignment strategies are paving a new path forward. This isnโ€™t your average tech commentaryโ€”itโ€™s a bold vision for conscious AI systems that reason, align, and adapt beyond narrow tasks.

๐Ÿ” Read it, upvote it if it sparks something, and letโ€™s ignite a collective conversation about the future of AGI.

https://huggingface.co/blog/Parveshiiii/glimpses-of-agi


Parveshiiiiย 
posted an update 26 days ago
view post
Post
2795
๐Ÿง  MathX-5M by XenArcAI โ€” Scalable Math Reasoning for Smarter LLMs

Introducing MathX-5M, a high-quality, instruction-tuned dataset built to supercharge mathematical reasoning in large language models. With 5 million rigorously filtered examples, it spans everything from basic arithmetic to advanced calculusโ€”curated from public sources and enhanced with synthetic data.

๐Ÿ” Key Highlights:
- Step-by-step reasoning with verified answers
- Covers algebra, geometry, calculus, logic, and more
- RL-validated correctness and multi-stage filtering
- Ideal for fine-tuning, benchmarking, and educational AI

๐Ÿ“‚ - XenArcAI/MathX-5M


  • 1 reply
ยท
Abhaykoulย 
posted an update 29 days ago
view post
Post
2972
๐ŸŽ‰ Dhanishtha 2.0 Preview is Now Open Source!

The world's first Intermediate Thinking Model is now available to everyone!

Dhanishtha 2.0 Preview brings revolutionary intermediate thinking capabilities to the open-source community. Unlike traditional reasoning models that think once, Dhanishtha can think, answer, rethink, answer again, and continue rethinking as needed using multiple blocks between responses.

๐Ÿš€ Key Features
- Intermediate thinking: Think โ†’ Answer โ†’ Rethink โ†’ Answer โ†’ Rethink if needed...
- Token efficient: Uses up to 79% fewer tokens than DeepSeek R1 on similar queries
- Transparent thinking: See the model's reasoning process in real-time
- Open source: Freely available for research and development


HelpingAI/Dhanishtha-2.0-preview
https://helpingai.co/chat
  • 1 reply
ยท
Abhaykoulย 
posted an update about 1 month ago
view post
Post
4357
Introducing Dhanishtha 2.0: World's first Intermediate Thinking Model

Dhanishtha 2.0 is the world's first LLM designed to think between the responses. Unlike other Reasoning LLMs, which think just once.

Dhanishtha can think, rethink, self-evaluate, and refine in between responses using multiple <think> blocks.
This technique makes it Hinghlt Token efficient it Uses up to 79% fewer tokens than DeepSeek R1
---

You can try our model from: https://helpingai.co/chat
Also, we're gonna Open-Source Dhanistha on July 1st.

---
For Devs:
๐Ÿ”‘ Get your API key at https://helpingai.co/dashboard
from HelpingAI import HAI  # pip install HelpingAI==1.1.1
from rich import print

hai = HAI(api_key="hl-***********************")

response = hai.chat.completions.create(
    model="Dhanishtha-2.0-preview",
    messages=[{"role": "user", "content": "What is the value of โˆซ0โˆž๐‘ฅ3/๐‘ฅโˆ’1๐‘‘๐‘ฅ ?"}],
    stream=True,
    hide_think=False # Hide or show models thinking
)

for chunk in response:
    print(chunk.choices[0].delta.content, end="", flush=True)
  • 2 replies
ยท
alielfilali01ย 
posted an update 3 months ago
alielfilali01ย 
posted an update 5 months ago
view post
Post
1063
๐Ÿšจ Arabic LLM Evaluation ๐Ÿšจ

Few models join the ranking of https://huggingface.co/spaces/inceptionai/AraGen-Leaderboard Today.

The new MistralAI model, Saba, is quite impressive, Top10 ! Well done @arthurmensch and team.

Sadly Mistral did not follow its strategy about public weights this time, we hope this changes soon and we get the model with a permissive license.

We added other Mistral models and apparently, we have been sleeping on mistralai/Mistral-Large-Instruct-2411 !

Another impressive model that joined the ranking today is ALLaM-AI/ALLaM-7B-Instruct-preview. After a long wait finally ALLaM is here and it is IMPRESSIVE given its size !

ALLaM is ranked on OALL/Open-Arabic-LLM-Leaderboard as well.
Abhaykoulย 
posted an update 6 months ago
view post
Post
4775
๐Ÿ”ฅ THE WAIT IS OVER... HAI-SER IS HERE! ๐Ÿ”ฅ

Yo fam, this ain't just another AI dropโ€” this is the FUTURE of emotional intelligence! ๐Ÿš€

Introducing HAI-SER, powered by Structured Emotional Reasoning (SER), the next-level AI that doesnโ€™t just understand your wordsโ€”it feels you, analyzes your emotions, and helps you navigate lifeโ€™s toughest moments. ๐Ÿ’ก

๐Ÿ’ฅ What makes HAI-SER a game-changer?
๐Ÿ”น Emotional Vibe Check โ€“ Gets the mood, energy, and whatโ€™s really going on ๐ŸŽญ
๐Ÿ”น Mind-State Analysis โ€“ Breaks down your thoughts, beliefs, and patterns ๐Ÿคฏ
๐Ÿ”น Root Cause Deep-Dive โ€“ Unpacks the WHY behind your emotions ๐Ÿ’ก
๐Ÿ”น Impact Check โ€“ Sees how itโ€™s affecting your life and mental health ๐Ÿ’”
๐Ÿ”น Safety Check โ€“ Prioritizes your well-being and crisis management ๐Ÿšจ
๐Ÿ”น Healing Game Plan โ€“ Custom strategies to help you bounce back ๐Ÿ’ช
๐Ÿ”น Growth Potential โ€“ Turns struggles into opportunities for self-improvement ๐Ÿ“ˆ
๐Ÿ”น How to Approach โ€“ Teaches you and others how to communicate and heal ๐Ÿค
๐Ÿ”น Personalized Response โ€“ Not just generic adviceโ€”real talk, tailored to YOU ๐Ÿ’ฏ

No more robotic AI responses. No more surface-level advice. HAI-SER gets deep, analyzing emotions with precision and giving real, actionable support.

This ainโ€™t just AIโ€”this is your digital therapist, life coach, and hype squad all in one. Whether itโ€™s mental health, career struggles, relationships, or personal growth, HAI-SER has your back.

๐Ÿš€ The future of emotionally intelligent AI is HERE.
Are you ready? ๐Ÿ”ฅ๐Ÿ’ฏ

HelpingAI/HAI-SER
ยท
alielfilali01ย 
posted an update 7 months ago
view post
Post
2145
3C3H AraGen Leaderboard welcomes today deepseek-ai/DeepSeek-V3 and 12 other models (including the late gpt-3.5 ๐Ÿ’€) to the ranking of best LLMs in Arabic !


Observations:
- DeepSeek-v3 ranked 3rd and only Open model among the top 5 !

- A 14B open model ( Qwen/Qwen2.5-14B-Instruct) outperforms gpt-3.5-turbo-0125 (from last year). This shows how much we came in advancing and supporting Arabic presence within the LLM ecosystem !

- Contrary to what observed in likelihood-acc leaderboards (like OALL/Open-Arabic-LLM-Leaderboard) further finetuned models like maldv/Qwentile2.5-32B-Instruct actually decreased the performance compared to the original model Qwen/Qwen2.5-32B-Instruct.
It's worth to note that the decrease is statiscally insignificant which imply that at best, the out-domain finetuning do not really hurts the model original capabilities acquired during pretraining.
Previous work addressed this (finetuning VS pretraining) but more investigation in this regard is required (any PhDs here ? This could be your question ...)


Check out the latest rankings: https://huggingface.co/spaces/inceptionai/AraGen-Leaderboard
alielfilali01ย 
posted an update 7 months ago
view post
Post
2057
~75% on the challenging GPQA with only 40M parameters ๐Ÿ”ฅ๐Ÿฅณ

GREAT ACHIEVEMENT ! Or is it ?

This new Work, "Data Laundering: Artificially Boosting Benchmark Results through Knowledge Distillation", take out the mystery about many models i personally suspected their results. Speacially on leaderboards other than the english one, Like the Open Arabic LLM Leaderbaord OALL/Open-Arabic-LLM-Leaderboard.

The authors of this work, first started by training a model on the GPQA data, which, unsurprisingly, led to the model achieving 100% performance.

Afterward, they trained what they referred to as a 'legitimate' model on legitimate data (MedMCQA). However, they introduced a distillation loss from the earlier, 'cheated' model.

What they discovered was fascinating: the knowledge of GPQA leaked through this distillation loss, even though the legitimate model was never explicitly trained on GPQA during this stage.

This raises important questions about the careful use of distillation in model training, especially when the training data is opaque. As they demonstrated, itโ€™s apparently possible to (intentionally or unintentionally) leak test data through this method.

Find out more: Data Laundering: Artificially Boosting Benchmark Results through Knowledge Distillation (2412.15255)
  • 1 reply
ยท
Abhaykoulย 
posted an update 7 months ago
view post
Post
2206
๐Ÿ”ฅ BIG ANNOUNCEMENT: THE HELPINGAI API IS LIVE! ๐Ÿ”ฅ

Yo, the moment youโ€™ve all been waiting for is here! ๐Ÿš€ The HelpingAI API is now LIVE and ready to level up your projects! ๐Ÿ”ฅ Weโ€™re bringing that next-level AI goodness straight to your fingertips. ๐Ÿ’ฏ

No more waitingโ€” itโ€™s time to build something epic! ๐Ÿ™Œ

From now on, you can integrate our cutting-edge AI models into your own applications, workflows, and everything in between. Whether youโ€™re a developer, a creator, or just someone looking to make some serious moves, this is your chance to unlock the full potential of emotional intelligence and adaptive AI.

Check out the docs ๐Ÿ”ฅ and letโ€™s get to work! ๐Ÿš€

๐Ÿ‘‰ Check out the docs and start building (https://helpingai.co/docs)
๐Ÿ‘‰ Visit the HelpingAI website (https://helpingai.co/)
ยท
alielfilali01ย 
posted an update 8 months ago
view post
Post
3551
Unpopular opinion: Open Source takes courage to do !

Not everyone is brave enough to release what they have done (the way they've done it) to the wild to be judged !
It really requires a high level of "knowing wth are you doing" ! It's kind of a super power !

Cheers to the heroes here who see this!
ยท
alielfilali01ย 
posted an update 8 months ago
view post
Post
1600
Apparently i forgot to put this here !

Well, this is a bit late but consider given our recent blog a read if you are interested in Evaluation.

You don't have to be into Arabic NLP in order to read it, the main contribution we are introducing is a new evaluation measure for NLG. We made the fisrt application of this measure on Arabic for now and we will be working with colleagues from the community to expand it to other languages.

Blog:
Rethinking LLM Evaluation with 3C3H: AraGen Benchmark and Leaderboard
https://huggingface.co/blog/leaderboard-3c3h-aragen

Space:
https://huggingface.co/spaces/inceptionai/AraGen-Leaderboard

Give it a read and let me know your thoughts ๐Ÿค—