Adam Molnar

lunarflu

AI & ML interests

join the Hugging Face discord! hf.co/discord/join

Recent Activity

View all activity

Organizations

Hugging Face's profile picture Society & Ethics's profile picture Gradio's profile picture Turing's Solutions's profile picture OpenDILab's profile picture Open-Source AI Meetup's profile picture lora concepts library's profile picture Huggingface Projects's profile picture Arabic Machine Learning 's profile picture Pseudo Lab's profile picture Keras Dreambooth Event's profile picture Hugging Face OSS Metrics's profile picture Ontocord's M*DEL's profile picture The Waifu Research Department's profile picture M.O.F.U.'s profile picture Blog-explorers's profile picture OpenSky's profile picture Hugging Face for Computer Vision's profile picture CyberHarem's profile picture Touhou AI Experimental Group (MOFU)'s profile picture huggingPartyParis's profile picture Multi๐Ÿค–Transformers's profile picture Team Tonic's profile picture That Time I got Reincarnated as a Hugging Face Organization's profile picture hf-qa-bot's profile picture Journalists on Hugging Face's profile picture MLX Community's profile picture the collabage patch's profile picture INNOVA AI's profile picture Narra's profile picture Social Post Explorers's profile picture Quasar Research's profile picture Hugging Face for Legal's profile picture Hugging Face Discord Community's profile picture HF-Study-Group's profile picture

lunarflu's activity

Reacted to AdinaY's post with ๐Ÿค— about 8 hours ago
view post
Post
769
Zhipu AI, the Chinese generative AI startup behind CogVideo, just launched their first productized AI Agent - AutoGLM ๐Ÿ”ฅ
๐Ÿ‘‰ https://agent.aminer.cn

With simple text or voice commands, it:
โœจ Simulates phone operations effortlessly
โœจ Autonomously handles 50+ step tasks
โœจ Seamlessly operates across apps

Powered by Zhipu's "Decoupled Interface" and "Self-Evolving Learning Framework" to achieve major performance gains in Phone Use and Web Browser Use!

Meanwhile, GLM4-Edge is now on Hugging Face hub๐Ÿš€
๐Ÿ‘‰ THUDM/glm-edge-6743283c5809de4a7b9e0b8b
Packed with advanced dialogue + multimodal models:
๐Ÿ“ฑ 1.5B / 2B models: Built for mobile & in-car systems
๐Ÿ’ป 4B / 5B models: Optimized for PCs
replied to m-ric's post about 8 hours ago
view reply

waiting for owo/uwu (new peak of models)

Reacted to m-ric's post with โค๏ธ about 8 hours ago
view post
Post
864
Single most important thing to do today: ๐—ด๐—ผ ๐˜๐—ฟ๐˜† ๐—ค๐˜„๐—ค ๐—ผ๐—ป ๐—›๐˜‚๐—ด๐—ด๐—ถ๐—ป๐—ด ๐—–๐—ต๐—ฎ๐˜!

๐Ÿ‘‰ https://huggingface.co/chat/models/Qwen/QwQ-32B-Preview
  • 2 replies
ยท
Reacted to julien-c's post with ๐Ÿค—๐Ÿ”ฅ about 8 hours ago
view post
Post
447
wow ๐Ÿ˜ฎ

INTELLECT-1 is the first collaboratively trained 10 billion parameter language model trained from scratch on 1 trillion tokens of English text and code.

PrimeIntellect/INTELLECT-1-Instruct
Reacted to burtenshaw's post with ๐Ÿ‘ about 9 hours ago
view post
Post
363
[SATURDAY ROUNDUP] โ˜•๏ธ๐Ÿง‘โ€๐ŸŽ“

In case you missed everything this week. Itโ€™s all about vision language models and image preference datasets. Here are the models and datasets you can use in your projects.

QWQ-32B-Preview is the first open weights model to reason like o1 with comparable performance. Itโ€™s large but is acing some of the hardest tasks.

https://bsky.app/profile/philschmid.bsky.social/post/3lbylz6nzqk25

SmolVLM is a vision implementation of the recently released SmolLM2. It uses the Idefics3 approach to add a vision encoder. The main difference being the smaller language model (8b > 1.7b) and more compression of images. This results in a model that is very accurate for its memory footprint.

https://huggingface.co/blog/smolvlm

ColSmolVLM is a vision embedding model based on SmolVLM using the Colbert approach from ColPali. This is shown to be great at document retrieval and everyone should test it out in their RAG setups.

https://huggingface.co/posts/merve/663466156074132

In an effort to build a FLUX level open source image generation model, the community is building a dataset of image preferences. The dataset is already open and the project is still running. Join in!

https://huggingface.co/posts/davidberenstein1957/405018978675827

TRL tutorial Drop - This week I dropped a load of tutorials on finetuning and aligning models with TRL. If youโ€™re upskilling in this space, you should check these out.

https://bsky.app/profile/benburtenshaw.bsky.social/post/3lbrc56ap3222
Reacted to DawnC's post with ๐Ÿ‘ about 9 hours ago
view post
Post
433
๐Ÿ’ก Curious about dog breeds? ๐Ÿ• Try PawMatchAI!
Itโ€™s a fun project I built to recognize dog breeds, recommend the perfect match for your lifestyle, and even compare breeds! ๐Ÿพ

๐Ÿ”Ž While itโ€™s not perfect, Iโ€™d love for you to give it a try and share your feedback!

๐Ÿ‘‰ check it out here !
DawnC/PawMatchAI

Your support means a lot as I continue learning and growing in my journey toward a career in AI. ๐Ÿ™Œ

Like it? Donโ€™t forget to click that thumbs up! ๐Ÿ‘
Reacted to abhishek's post with ๐Ÿค— about 9 hours ago
view post
Post
357
๐ŸŽ‰ SUPER BLACK FRIDAY DEAL ๐ŸŽ‰

Train almost any model on a variety of tasks such as llm finetuning, text classification/regression, summarization, question answering, image classification/regression, object detection, tabular data, etc for FREE using AutoTrain locally. ๐Ÿ”ฅ
https://github.com/huggingface/autotrain-advanced
Reacted to KnutJaegersberg's post with ๐Ÿ‘€ about 9 hours ago
Reacted to sequelbox's post with ๐Ÿ‘ 12 days ago
Reacted to reach-vb's post with ๐Ÿš€๐Ÿค—๐Ÿ‘๐Ÿ”ฅ 12 days ago
view post
Post
4166
What a brilliant week for Open Source AI!

Qwen 2.5 Coder by Alibaba - 0.5B / 1.5B / 3B / 7B / 14B/ 32B (Base + Instruct) Code generation LLMs, with 32B tackling giants like Gemnini 1.5 Pro, Claude Sonnet
Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f

LLM2CLIP from Microsoft - Leverage LLMs to train ultra-powerful CLIP models! Boosts performance over the previous SOTA by ~17%
microsoft/llm2clip-672323a266173cfa40b32d4c

Athene v2 Chat & Agent by NexusFlow - SoTA general LLM fine-tuned from Qwen 2.5 72B excels at Chat + Function Calling/ JSON/ Agents
Nexusflow/athene-v2-6735b85e505981a794fb02cc

Orca Agent Instruct by Microsoft - 1 million instruct pairs covering text editing, creative writing, coding, reading comprehension, etc - permissively licensed
microsoft/orca-agentinstruct-1M-v1

Ultravox by FixieAI - 70B/ 8B model approaching GPT4o level, pick any LLM, train an adapter with Whisper as Audio Encoder
reach-vb/ultravox-audio-language-model-release-67373b602af0a52b2a88ae71

JanusFlow 1.3 by DeepSeek - Next iteration of their Unified MultiModal LLM Janus with RectifiedFlow
deepseek-ai/JanusFlow-1.3B

Common Corpus by Pleais - 2,003,039,184,047 multilingual, commercially permissive and high quality tokens!
PleIAs/common_corpus

I'm sure I missed a lot, can't wait for the next week!

Put down in comments what I missed! ๐Ÿค—
Reacted to TuringsSolutions's post with ๐Ÿ‘€ 12 days ago
view post
Post
735
If I am correct and the LLM model changes the 'shape' of the data as it learns, then I should be able to track and utilize those shape changes as a backpropagation training mechanism, right? Well guess what, I can do that! Entropy, Sparsity, and Density, this is how I can measure the shape of the data the LLM model is creating. Nodes, Clusters, and Edges, these are the mechanisms within the neural network the LLM model updates as it learns these concepts. I measure the effects of these updates, via Entropy, Sparsity, and Density. Check out more in this video: https://youtu.be/jADTt5HHtiw
  • 2 replies
ยท
Reacted to erikkaum's post with ๐Ÿ‘€๐Ÿ”ฅ 12 days ago
view post
Post
1675
A while ago I started experimenting with compiling the Python interpreter to WASM.

To build a secure, fast, and lightweight sandbox for code execution โ€” ideal for running LLM-generated Python code.

- Send code simply as a POST request
- 1-2ms startup times

Hack away:
https://github.com/ErikKaum/runner
Reacted to AdinaY's post with ๐Ÿ‘€ 12 days ago
Reacted to sayakpaul's post with ๐Ÿš€โค๏ธ 12 days ago
view post
Post
2409
It's been a while we shipped native quantization support in diffusers ๐Ÿงจ

We currently support bistandbytes as the official backend but using others like torchao is already very simple.

This post is just a reminder of what's possible:

1. Loading a model with a quantization config
2. Saving a model with quantization config
3. Loading a pre-quantized model
4. enable_model_cpu_offload()
5. Training and loading LoRAs into quantized checkpoints

Docs:
https://huggingface.co/docs/diffusers/main/en/quantization/bitsandbytes
  • 1 reply
ยท