kotol

company
Activity Feed

AI & ML interests

None defined yet.

Recent Activity

gv-hf's activity

XenovaĀ 
posted an update 3 days ago
view post
Post
4235
First project of 2025: Vision Transformer Explorer

I built a web app to interactively explore the self-attention maps produced by ViTs. This explains what the model is focusing on when making predictions, and provides insights into its inner workings! šŸ¤Æ

Try it out yourself! šŸ‘‡
webml-community/attention-visualization

Source code: https://github.com/huggingface/transformers.js-examples/tree/main/attention-visualization
merveĀ 
posted an update 4 days ago
view post
Post
3750
supercharge your LLM apps with smolagents šŸ”„

however cool your LLM is, without being agentic it can only go so far

enter smolagents: a new agent library by Hugging Face to make the LLM write code, do analysis and automate boring stuff!

Here's our blog for you to get started https://huggingface.co/blog/smolagents
merveĀ 
posted an update 10 days ago
XenovaĀ 
posted an update 16 days ago
view post
Post
2697
Introducing Moonshine Web: real-time speech recognition running 100% locally in your browser!
šŸš€ Faster and more accurate than Whisper
šŸ”’ Privacy-focused (no data leaves your device)
āš”ļø WebGPU accelerated (w/ WASM fallback)
šŸ”„ Powered by ONNX Runtime Web and Transformers.js

Demo: webml-community/moonshine-web
Source code: https://github.com/huggingface/transformers.js-examples/tree/main/moonshine-web
Ā·
merveĀ 
posted an update 17 days ago
view post
Post
2731
Aya by Cohere For AI can now see! šŸ‘€

C4AI community has built Maya 8B, a new open-source multilingual VLM built on SigLIP and Aya 8B šŸŒ± works on 8 languages! šŸ—£ļø

The authors extend Llava dataset using Aya's translation capabilities with 558k examples!
ry it here kkr5155/maya_demo

Dataset maya-multimodal/pretrain

Model maya-multimodal/maya šŸ‘
kudos @nahidalam and team
  • 1 reply
Ā·
merveĀ 
posted an update 18 days ago
view post
Post
3159
Apollo is a new family of open-source video language models by Meta, where 3B model outperforms most 7B models and 7B outperforms most 30B models šŸ§¶

āœØ the models come in 1.5B https://huggingface.co/Apollo-LMMs/Apollo-1_5B-t32, 3B https://huggingface.co/Apollo-LMMs/Apollo-3B-t32 and 7B https://huggingface.co/Apollo-LMMs/Apollo-7B-t32 with A2.0 license, based on Qwen1.5 & Qwen2
āœØ the authors also release a benchmark dataset https://huggingface.co/spaces/Apollo-LMMs/ApolloBench

The paper has a lot of experiments (they trained 84 models!) about what makes the video LMs work āÆļø

Try the demo for best setup here https://huggingface.co/spaces/Apollo-LMMs/Apollo-3B
they evaluate sampling strategies, scaling laws for models and datasets, video representation and more!
> The authors find out that whatever design decision was applied to small models also scale properly when the model and dataset are scaled šŸ“ˆ scaling dataset has diminishing returns for smaller models
> They evaluate frame sampling strategies, and find that FPS sampling is better than uniform sampling, and they find 8-32 tokens per frame optimal
> They also compare image encoders, they try a variation of models from shape optimized SigLIP to DINOv2
they find google/siglip-so400m-patch14-384 to be most powerful šŸ”„
> they also compare freezing different parts of models, training all stages with some frozen parts give the best yield

They eventually release three models, where Apollo-3B outperforms most 7B models and Apollo 7B outperforms 30B models šŸ”„
Ā·
merveĀ 
posted an update 23 days ago
view post
Post
1739
A complete RAG pipeline includes a reranker, which ranks the documents to find the best document šŸ““
Same goes for multimodal RAG, multimodal rerankers which we can integrate to multimodal RAG pipelines!
Learn how to build a complete multimodal RAG pipeline with vidore/colqwen2-v1.0 as retriever, lightonai/MonoQwen2-VL-v0.1 as reranker, Qwen/Qwen2-VL-7B-Instruct as VLM in this notebook that runs on a GPU as small as L4 šŸ”„ https://huggingface.co/learn/cookbook/multimodal_rag_using_document_retrieval_and_reranker_and_vlms
XenovaĀ 
posted an update 26 days ago
view post
Post
2961
Introducing TTS WebGPU: The first ever text-to-speech web app built with WebGPU acceleration! šŸ”„ High-quality and natural speech generation that runs 100% locally in your browser, powered by OuteTTS and Transformers.js. šŸ¤— Try it out yourself!

Demo: webml-community/text-to-speech-webgpu
Source code: https://github.com/huggingface/transformers.js-examples/tree/main/text-to-speech-webgpu
Model: onnx-community/OuteTTS-0.2-500M (ONNX), OuteAI/OuteTTS-0.2-500M (PyTorch)
merveĀ 
posted an update 27 days ago
view post
Post
5558
This week in open-source AI was insane šŸ¤  A small recapšŸ•ŗšŸ» merve/dec-6-releases-67545caebe9fc4776faac0a3

Multimodal šŸ–¼ļø
> Google shipped a PaliGemma 2, new iteration of PaliGemma with more sizes: 3B, 10B and 28B, with pre-trained and captioning variants šŸ‘
> OpenGVLab released InternVL2, seven new vision LMs in different sizes, with sota checkpoint with MIT license āœØ
> Qwen team at Alibaba released the base models of Qwen2VL models with 2B, 7B and 72B ckpts

LLMs šŸ’¬
> Meta released a new iteration of Llama 70B, Llama3.2-70B trained further
> EuroLLM-9B-Instruct is a new multilingual LLM for European languages with Apache 2.0 license šŸ”„
> Dataset: CohereForAI released GlobalMMLU, multilingual version of MMLU with 42 languages with Apache 2.0 license
> Dataset: QwQ-LongCoT-130K is a new dataset to train reasoning models
> Dataset: FineWeb2 just landed with multilinguality update! šŸ”„ nearly 8TB pretraining data in many languages!

Image/Video Generation šŸ–¼ļø
> Tencent released HunyuanVideo, a new photorealistic video generation model
> OminiControl is a new editing/control framework for image generation models like Flux

Audio šŸ”Š
> Indic-Parler-TTS is a new text2speech model made by community
merveĀ 
posted an update 28 days ago
view post
Post
1533
New InternVL drop with a state-of-the-art 78B vision language model with MIT license šŸ”„ https://huggingface.co/collections/OpenGVLab/internvl-25-673e1019b66e2218f68d7c1c
The release comes with seven new vision LMs based on InternViT 300M/6B and Qwen2.5 (0.5B, 3B, 32B, 72B) and InternLM2 (8B, 7B, 20B) in different sizes
78B model is of InternViT 6B and Qwen2.5-72B Instruct, can accomplish variety of tasks šŸ‘ Try here OpenGVLab/InternVL
ariG23498Ā 
posted an update 29 days ago

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498

Update README.md

#3 opened about 1 month ago by
ariG23498