Victor Mustar's picture

Victor Mustar PRO

victor

AI & ML interests

Building the UX of this website

Recent Activity

Articles

Organizations

Hugging Face's profile picture Google's profile picture Safetensors's profile picture Competitions's profile picture 21 RNN's profile picture Spaces-explorers's profile picture Text Generation Inference's profile picture Spaces Examples's profile picture CVPR Demo Track's profile picture Hugging Chat's profile picture Webhooks Explorers (BETA)'s profile picture lora concepts library's profile picture Huggingface Projects's profile picture Scanned Tokens's profile picture hf admins's profile picture Hugging Face OSS Metrics's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture Core ML Projects's profile picture temp-org's profile picture Blog-explorers's profile picture Mustarz's profile picture Open LLM Leaderboard's profile picture Enterprise Explorers's profile picture The Collectionists's profile picture ZeroGPU Explorers's profile picture Hugging Face Tools's profile picture TstOrg141's profile picture Stable Video benchmark's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture LLHF's profile picture SLLHF's profile picture

victor's activity

reacted to merve's post with πŸ”₯ about 17 hours ago
view post
Post
679
supercharge your LLM apps with smolagents πŸ”₯

however cool your LLM is, without being agentic it can only go so far

enter smolagents: a new agent library by Hugging Face to make the LLM write code, do analysis and automate boring stuff!

Here's our blog for you to get started https://huggingface.co/blog/smolagents
reacted to cfahlgren1's post with πŸš€ about 20 hours ago
reacted to sequelbox's post with πŸ‘ about 20 hours ago
reacted to ivanfioravanti's post with πŸ‘ 3 days ago
view post
Post
1427
Probably most of you already knows this trick but just in case:
πŸ€” Unable to connect to Hugging Face Spaces Dev Mode through local Cursor? πŸ’‘ Don't worry there is an easy trick!

- right click Connect with VS Code
- copy link in your browser
- vscode://vscode-remote/...
- replace vscode with cursor and go
- cursor://vscode-remote/...
reacted to hexgrad's post with πŸ”₯ 5 days ago
view post
Post
3843
Merry Christmas! πŸŽ„ Open sourced a small TTS model at hexgrad/Kokoro-82M
  • 2 replies
Β·
reacted to merve's post with πŸ‘ 5 days ago
reacted to AdinaY's post with πŸš€πŸ”₯ 5 days ago
view post
Post
3410
The Chinese community is shipping 🚒

DeepSeek V3 (685 B MoE) has quietly released on the hub!
Base: deepseek-ai/DeepSeek-V3-Base
Instruct: deepseek-ai/DeepSeek-V3

Can’t wait to see what’s next!
  • 1 reply
Β·
reacted to vincentg64's post with πŸ”₯ 5 days ago
view post
Post
2172
LLM 2.0, RAG & Non-Standard Gen AI on GitHub https://mltblog.com/3DsyZSq

In this article, I share my latest Gen AI and LLM advances, featuring innovative approaches radically different from both standard AI and classical ML/NLP. The focus is on doing better with less, using efficient architectures, new algorithms and evaluation metrics. It originates from research that I started long ago. It gained significant momentum in the last two years. See background and history at https://mltblog.com/4g2sKTv.

OpenAI, Perplexity, Anthropic, Llama and others typically follow the trend and implement solutions very similar to mines within 3 to 6 months after I publish new milestones. For instance, multi-tokens, knowledge graph tokens, multi-indexes, real-time fine-tuning, mixtures of experts, LLM routers, small enterprise sub-LLMs, prompt distillation, relevancy scoring engine, deep contextual retrieval, optimum agentic chunking, and modern UI instead of the basic prompt box. I keep adding new features all the time, staying ahead of competition.

➑️ Read full article with links to GitHub, at https://mltblog.com/3DsyZSq
  • 1 reply
Β·
reacted to as-cle-bert's post with πŸ‘ 5 days ago
reacted to hexgrad's post with πŸ€— 5 days ago
view post
Post
3010
Tonight, Adam & Michael join the 82M Apache TTS model in hexgrad/Kokoro-82M
reacted to nicolay-r's post with ❀️ 5 days ago
view post
Post
2032
πŸ“’ Deligted to share the most recent milestone on quick deployment of Named Entity Recognition (NER) in Gen-AI powered systems.

Releasing the bulk-ner 0.25.0 which represent a tiny framework that would save you time for deploing NER with any model.

πŸ’Ž Why is this important? In the era of GenAI the handling out textual output might be challenging. Instead, recognizing named-entities via domain-oriented systems for your donwstream LLM would be preferable option.

πŸ“¦: https://pypi.org/project/bulk-ner/0.25.0/
🌟: https://github.com/nicolay-r/bulk-ner

I noticed that the direct adaptaion of the LM for NER would result in spending signifcant amount of time on formatting your texts according to the NER-model needs.
In particular:
1. Processing CONLL format with B-I-O tags from model outputs
2. Input trimming: long input content might not be completely fitted

To cope with these problems, in version 0.25.0 I made a huge steps forward by providing:
βœ… 🐍 Python API support: see screenshot below for a quick deployment (see screenshot below πŸ“Έ)
βœ… πŸͺΆ No-string: dependencies are now clear, so it is purely Python implementation for API calls.
βœ… πŸ‘Œ Simplified output formatting: we use lists to represent texts with inner lists that refer to annotated objects (see screenshot below πŸ“Έ)

πŸ“’ We have a colab for a quick start here (or screenshot for bash / Python API πŸ“Έ)
https://colab.research.google.com/github/nicolay-r/ner-service/blob/main/NER_annotation_service.ipynb

πŸ‘ The code for pipeline deployment is taken from the AREkit project:
https://github.com/nicolay-r/AREkit
reacted to Kseniase's post with πŸ‘ 9 days ago
view post
Post
2670
**15 Agentic Systems and Frameworks of 2024**

This year, we started our β€œAI Agents and Agentic Workflows” series (https://www.turingpost.com/t/AI-Agents) to explore everything about AI agents step by step: all the vocabulary, how they work, and how to build them.
The huge interest in this series and the large number of studies conducted on agents showed that it was one of the most popular and important themes of the year. In 2025, most likely, agents will reach new highs – we will be covering that for you. Now, let’s review the agentic systems that have emerged this year.

Here is a list of 15 agentic systems and frameworks of 2024:

1. GUI Agents: A Survey (2412.13501)

2. Large Language Models Orchestrating Structured Reasoning Achieve Kaggle Grandmaster Level (2411.03562)

3. The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery (2408.06292)

4. MALT: Improving Reasoning with Multi-Agent LLM Training (2412.01928)

5. Agent S: An Open Agentic Framework that Uses Computers Like a Human (2410.08164)

6. Automated Design of Agentic Systems (2408.08435)

7. AgentInstruct: Toward Generative Teaching with Agentic Flows (2407.03502)

8. AgentStore: Scalable Integration of Heterogeneous Agents As Specialized Generalist Computer Assistant (2410.18603)

9. WALL-E: World Alignment by Rule Learning Improves World Model-based LLM Agents (2410.07484)

10. Generative Agent Simulations of 1,000 People (2411.10109)

11. DynaSaur: Large Language Agents Beyond Predefined Actions (2411.01747)

12. PRefLexOR: Preference-based Recursive Language Modeling for Exploratory Optimization of Reasoning and Agentic Thinking (2410.12375)

13. Generative World Explorer (2411.11844)

14. Bel Esprit: Multi-Agent Framework for Building AI Model Pipelines (2412.14684)

15. AutoKaggle: A Multi-Agent Framework for Autonomous Data Science Competitions (2410.20424)

Thanks for reading Turing Post!
Subscribe to receive new posts straight into your inbox -> https://www.turingpost.com/subscribe
reacted to nroggendorff's post with πŸ‘€ 9 days ago
view post
Post
1295
Can we please do something about this? It makes everything I do so much harder, and because my local machine is so terrible, I am forced to test in production. This makes debugging so difficult.
nroggendorff/system-exit

cc @victor
  • 1 reply
Β·
reacted to anton-l's post with πŸ”₯ 12 days ago
view post
Post
2074
Introducing πŸ“π…π’π§πžπŒπšπ­π‘: the best public math pre-training dataset with 50B+ tokens!
HuggingFaceTB/finemath

Math remains challenging for LLMs and by training on FineMath we see considerable gains over other math datasets, especially on GSM8K and MATH.

We build the dataset by:
πŸ› οΈ carefully extracting math data from Common Crawl;
πŸ”Ž iteratively filtering and recalling high quality math pages using a classifier trained on synthetic annotations to identify math reasoning and deduction.

We conducted a series of ablations comparing the performance of Llama-3.2-3B-Base after continued pre-training on FineMath and observe notable gains compared to the baseline model and other public math datasets.

We hope this helps advance the performance of LLMs on math and reasoning! πŸš€
We’re also releasing all the ablation models as well as the evaluation code.

HuggingFaceTB/finemath-6763fb8f71b6439b653482c2
reacted to m-ric's post with πŸ”₯ 12 days ago
view post
Post
1766
After 6 years, BERT, the workhorse of encoder models, finally gets a replacement: π—ͺ𝗲𝗹𝗰𝗼𝗺𝗲 π— π—Όπ—±π—²π—Ώπ—»π—•π—˜π—₯𝗧! πŸ€—

We talk a lot about ✨Generative AI✨, meaning "Decoder version of the Transformers architecture", but this is only one of the ways to build LLMs: encoder models, that turn a sentence in a vector, are maybe even more widely used in industry than generative models.

The workhorse for this category has been BERT since its release in 2018 (that's prehistory for LLMs).

It's not a fancy 100B parameters supermodel (just a few hundred millions), but it's an excellent workhorse, kind of a Honda Civic for LLMs.

Many applications use BERT-family models - the top models in this category cumulate millions of downloads on the Hub.

➑️ Now a collaboration between Answer.AI and LightOn just introduced BERT's replacement: ModernBERT.

π—§π—Ÿ;𝗗π—₯:
πŸ›οΈ Architecture changes:
β‡’ First, standard modernizations:
- Rotary positional embeddings (RoPE)
- Replace GeLU with GeGLU,
- Use Flash Attention 2
✨ The team also introduced innovative techniques like alternating attention instead of full attention, and sequence packing to get rid of padding overhead.

πŸ₯‡ As a result, the model tops the game of encoder models:
It beats previous standard DeBERTaV3 for 1/5th the memory footprint, and runs 4x faster!

Read the blog post πŸ‘‰ https://huggingface.co/blog/modernbert
  • 1 reply
Β·
reacted to akhaliq's post with πŸ”₯ 12 days ago
view post
Post
3154
Google drops Gemini 2.0 Flash Thinking

a new experimental model that unlocks stronger reasoning capabilities and shows its thoughts. The model plans (with thoughts visible), can solve complex problems with Flash speeds, and more

now available in anychat, try it out: akhaliq/anychat
reacted to Lewdiculous's post with βž• 12 days ago
reacted to KnutJaegersberg's post with πŸ‘ 12 days ago
reacted to FranckAbgrall's post with πŸ”₯ 13 days ago
view post
Post
1091
πŸ†• It should now be easier to identify discussions or pull requests where repository owners are participating on HF, let us know it that helps πŸ’¬πŸ€—
  • 1 reply
Β·