Peter Szemraj PRO

pszemraj

AI & ML interests

metallic intuition

Recent Activity

updated a dataset 1 day ago
BEE-spoke-data/govdocs1-by-extension
updated a dataset 1 day ago
BEE-spoke-data/cosmopedia-v2-mincols
updated a dataset 1 day ago
pszemraj/cnn_dailymail-cleaned
View all activity

Organizations

Stacked Summaries's profile picture Analytics Club at ETH ZĆ¼rich's profile picture ml4pubmed's profile picture Gradio-Blocks-Party's profile picture OneFrame's profile picture postbot's profile picture Ontocord's M*DEL's profile picture BEEspoke Data's profile picture Sablo AI's profile picture memecaps's profile picture Hugging Face Discord Community's profile picture

pszemraj's activity

reacted to tomaarsen's post with šŸ”„ 3 days ago
view post
Post
2348
That didn't take long! Nomic AI has finetuned the new ModernBERT-base encoder model into a strong embedding model for search, classification, clustering and more!

Details:
šŸ¤– Based on ModernBERT-base with 149M parameters.
šŸ“Š Outperforms both nomic-embed-text-v1 and nomic-embed-text-v1.5 on MTEB!
šŸŽļø Immediate FA2 and unpacking support for super efficient inference.
šŸŖ† Trained with Matryoshka support, i.e. 2 valid output dimensionalities: 768 and 256.
āž”ļø Maximum sequence length of 8192 tokens!
2ļøāƒ£ Trained in 2 stages: unsupervised contrastive data -> high quality labeled datasets.
āž• Integrated in Sentence Transformers, Transformers, LangChain, LlamaIndex, Haystack, etc.
šŸ›ļø Apache 2.0 licensed: fully commercially permissible

Try it out here: nomic-ai/modernbert-embed-base

Very nice work by Zach Nussbaum and colleagues at Nomic AI.