AI & ML interests

vision , multimedia , gradio, accessibility & cool demos

NymboΒ 
posted an update 7 days ago
view post
Post
558
Anyone using Jan-v1-4B for local MCP-based web search, I highly recommend you try out Intelligent-Internet/II-Search-4B

Very impressed with this lil guy and it deserves more downloads. It's based on the original version of Qwen3-4B but find that it questions reality way less often. Jan-v1 seems to think that everything it sees is synthetic data and constantly gaslights me
ZennyKennyΒ 
posted an update 11 days ago
TonicΒ 
posted an update 21 days ago
hesamationΒ 
posted an update about 1 month ago
view post
Post
3288
longer context doesn't generate better responses. it can even hurt your llm/agent. 1M context window doesn't automatically make models smarter as it's not about the size; it's how you use it.

here are 4 types of context failure and why each one happens:

1. context poisoning: if hallucination finds its way into your context, the agent will rely on that false information to make its future moves. for example if the agent hallucinates about the "task description", all of its planning to solve the task would also be corrupt.

2. context distraction: when the context becomes too bloated, the model focuses too much on it rather than come up with novel ideas or to follow what it has learned during training. as Gemini 2.5 Pro technical report points out, as context grows significantly from 100K tokens, "the agent showed a tendency toward favoring repeating actions from its vast history rather than synthesizing novel plans".

3. context confusion: everyone lost it when MCPs became popular, it seemed like AGI was achieved. I suspected there is something wrong and there was: it's not just about providing tools, bloating the context with tool use derails the model from selecting the right one! even if you can fit all your tool metadata in the context, as their number grows, the model gets confused over which one to pick.

4. Context Clash: if you exchange conversation with a model step by step and provide information as you go along, chances are you get worse performance rather than providing all the useful information at once. one the model's context fills with wrong information, it's more difficult to guide it to embrace the right info. agents pull information from tools, documents, user queries, etc. and there is a chance that some of these information contradict each other, and it's not good new for agentic applications.

check this article by Drew Breunig for deeper read: https://www.dbreunig.com/2025/06/26/how-to-fix-your-context.html?ref=blog.langchain.com
  • 2 replies
Β·
AtAndDevΒ 
posted an update about 1 month ago
view post
Post
409
Qwen 3 Coder is a personal attack to k2, and I love it.
It achieves near SOTA on LCB while not having reasoning.
Finally people are understanding that reasoning isnt necessary for high benches...

Qwen ftw!

DECENTRALIZE DECENTRALIZE DECENTRALIZE
TonicΒ 
posted an update about 1 month ago
view post
Post
739
πŸ‘‹ Hey there folks,

just submitted my plugin idea to the G-Assist Plugin Hackathon by @nvidia . Check it out, it's a great way to use a local SLA model on a windows machine to easily and locally get things done ! https://github.com/NVIDIA/G-Assist
TonicΒ 
posted an update about 1 month ago
view post
Post
572
πŸ™‹πŸ»β€β™‚οΈ Hey there folks ,

Yesterday , Nvidia released a reasoning model that beats o3 on science, math and coding !

Today you can try it out here : Tonic/Nvidia-OpenReasoning

hope you like it !
TonicΒ 
posted an update about 1 month ago
view post
Post
3318
πŸ™‹πŸ»β€β™‚οΈ Normalize adding compute & runtime traces to your model cards
  • 2 replies
Β·
hesamationΒ 
posted an update about 1 month ago
view post
Post
5057
in case you didn’t know, Claude now has a developer training course with certificates,

this is better than anything you can find on Coursera.

covers Claude Code, MCP and its advanced topics and even more:

https://www.anthropic.com/learn/build-with-claude
TonicΒ 
posted an update about 2 months ago
view post
Post
506
Who's going to Raise Summit in Paris Tomorrow ?

If you're around , I would love to meet you :-)
NymboΒ 
posted an update about 2 months ago
view post
Post
2796
Anyone know how to reset Claude web's MCP config? I connected mine when the HF MCP first released with just the default example spaces added. I added lots of other MCP spaces but Claude.ai doesn't update the available tools... "Disconnecting" the HF integration does nothing, deleting it and adding it again does nothing.

Refreshing tools works fine in VS Code because I can manually restart it in mcp.json, but claude.ai has no such option. Anyone got any ideas?
Β·
hesamationΒ 
posted an update 2 months ago
view post
Post
2887
this repo is gold! a collection of LLM apps with multi-agents, MCP, RAG and so much more.

the best way to learn is by building, and this repo provides the blueprint.

Repo: https://github.com/Shubhamsaboo/awesome-llm-apps
TonicΒ 
posted an update 3 months ago
view post
Post
688
πŸ™‹πŸ»β€β™‚οΈ hey there folks ,

So every bio/med/chem meeting i go to i always the same questions "why are you sharing a gdrive link with me for this?" and "Do you have any plans to publish your model weights and datasets on huggingface?" and finally i got a good answer today which explains everything :

basically there is some kind of government censorship on this (usa, but i'm sure others too) and they are told they are not allowed as it is considered a "dataleak" which is illegal !!!!

this is terrible ! but the good news is that we can do something about it !

so there is this "call for opinions and comments" here from the NIH (usa) , and here we can make our opinion on this topic known : https://osp.od.nih.gov/comment-form-responsibly-developing-and-sharing-generative-artificial-intelligence-tools-using-nih-controlled-access-data/

kindly consider dropping your opinion and thoughts about this censorship of science , and share this post , link or thoughts widely .

Together maybe we can start to share data and model weights appropriately and openly in a good way πŸ™πŸ»πŸš€

cc. @cyrilzakka

AtAndDevΒ 
posted an update 3 months ago
view post
Post
2957
deepseek-ai/DeepSeek-R1-0528

This is the end
  • 1 reply
Β·
hesamationΒ 
posted an update 3 months ago
view post
Post
2757
I really like how this seven-stage pipeline was laid out in the Ultimate Guide to Fine-Tuning book.

It gives an overview, then goes into detail for each stage, even providing best practices.

It’s 115 pages on arxiv, definitely worth a read.

Check it out: https://arxiv.org/abs/2408.13296
TonicΒ 
posted an update 3 months ago
view post
Post
2543
πŸ™‹πŸ»β€β™‚οΈ Hey there folks ,

Yesterday the world's first "Learn to Vibe Code" application was released .

As vibe coding is the mainstream paradigm , so now the first educational app is there to support it .

You can try it out already :

https://vibe.takara.ai

and of course it's entirely open source, so i already made my issue and feature branch :-) πŸš€
hesamationΒ 
posted an update 3 months ago
hesamationΒ 
posted an update 3 months ago
view post
Post
3104
this book actually exists for free, β€œthe little book of deep learning”. best to refresh your mind about DL basics:
> foundations of machine learning
> how models train
> common layers (dropout, pooling…)
> basic intro to LLMs
actually optimized for mobile.

Book: https://fleuret.org/public/lbdl.pdf
NymboΒ 
posted an update 4 months ago
view post
Post
4091
Haven't seen this posted anywhere - Llama-3.3-8B-Instruct is available on the new Llama API. Is this a new model or did someone mislabel Llama-3.1-8B?
  • 1 reply
Β·
ZennyKennyΒ 
posted an update 4 months ago
view post
Post
948
Community! πŸ’‘πŸ’‘πŸ’‘

It's the last day to submit your datasets for the Reasoning Datasets Competition: https://www.bespokelabs.ai/blog/reasoning-datasets-competition

Here are my submissions:
- ZennyKenny/synthetic_vc_financial_decisions_reasoning_dataset
- ZennyKenny/cosa-benchmark-dataset
- ZennyKenny/tactical-military-reasoning-v.1.0
- ZennyKenny/tron-dataset-v.1.0

Have a look and drop a ❀️ or comment! Check out the entire collection of submissions here: https://huggingface.co/datasets?other=reasoning-datasets-competition