Adam Molnar

lunarflu

AI & ML interests

join the Hugging Face discord! hf.co/discord/join

Recent Activity

updated a Space about 7 hours ago
lunarflu/verification-bot
updated a Space about 9 hours ago
discord-community/LevelBot
View all activity

Organizations

lunarflu's activity

New activity in alpindale/two-million-bluesky-posts about 6 hours ago

🚩 Report: Legal issue(s)

8
#3 opened about 6 hours ago by terenceeden
updated a Space about 9 hours ago
upvoted an article 1 day ago
view article
Article

Let’s make a generation of amazing image generation models

By burtenshaw β€’
β€’ 29
New activity in discord-community/LevelBot 1 day ago

Suggestion Discussion

1
#25 opened 27 days ago by nroggendorff
Reacted to sequelbox's post with πŸ‘ 9 days ago
Reacted to reach-vb's post with πŸš€πŸ€—πŸ‘πŸ”₯ 9 days ago
view post
Post
4083
What a brilliant week for Open Source AI!

Qwen 2.5 Coder by Alibaba - 0.5B / 1.5B / 3B / 7B / 14B/ 32B (Base + Instruct) Code generation LLMs, with 32B tackling giants like Gemnini 1.5 Pro, Claude Sonnet
Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f

LLM2CLIP from Microsoft - Leverage LLMs to train ultra-powerful CLIP models! Boosts performance over the previous SOTA by ~17%
microsoft/llm2clip-672323a266173cfa40b32d4c

Athene v2 Chat & Agent by NexusFlow - SoTA general LLM fine-tuned from Qwen 2.5 72B excels at Chat + Function Calling/ JSON/ Agents
Nexusflow/athene-v2-6735b85e505981a794fb02cc

Orca Agent Instruct by Microsoft - 1 million instruct pairs covering text editing, creative writing, coding, reading comprehension, etc - permissively licensed
microsoft/orca-agentinstruct-1M-v1

Ultravox by FixieAI - 70B/ 8B model approaching GPT4o level, pick any LLM, train an adapter with Whisper as Audio Encoder
reach-vb/ultravox-audio-language-model-release-67373b602af0a52b2a88ae71

JanusFlow 1.3 by DeepSeek - Next iteration of their Unified MultiModal LLM Janus with RectifiedFlow
deepseek-ai/JanusFlow-1.3B

Common Corpus by Pleais - 2,003,039,184,047 multilingual, commercially permissive and high quality tokens!
PleIAs/common_corpus

I'm sure I missed a lot, can't wait for the next week!

Put down in comments what I missed! πŸ€—
Reacted to TuringsSolutions's post with πŸ‘€ 9 days ago
view post
Post
733
If I am correct and the LLM model changes the 'shape' of the data as it learns, then I should be able to track and utilize those shape changes as a backpropagation training mechanism, right? Well guess what, I can do that! Entropy, Sparsity, and Density, this is how I can measure the shape of the data the LLM model is creating. Nodes, Clusters, and Edges, these are the mechanisms within the neural network the LLM model updates as it learns these concepts. I measure the effects of these updates, via Entropy, Sparsity, and Density. Check out more in this video: https://youtu.be/jADTt5HHtiw
  • 2 replies
Β·
Reacted to erikkaum's post with πŸ‘€πŸ”₯ 9 days ago
view post
Post
1668
A while ago I started experimenting with compiling the Python interpreter to WASM.

To build a secure, fast, and lightweight sandbox for code execution β€” ideal for running LLM-generated Python code.

- Send code simply as a POST request
- 1-2ms startup times

Hack away:
https://github.com/ErikKaum/runner
Reacted to AdinaY's post with πŸ‘€ 9 days ago
Reacted to sayakpaul's post with πŸš€β€οΈ 9 days ago
view post
Post
2340
It's been a while we shipped native quantization support in diffusers 🧨

We currently support bistandbytes as the official backend but using others like torchao is already very simple.

This post is just a reminder of what's possible:

1. Loading a model with a quantization config
2. Saving a model with quantization config
3. Loading a pre-quantized model
4. enable_model_cpu_offload()
5. Training and loading LoRAs into quantized checkpoints

Docs:
https://huggingface.co/docs/diffusers/main/en/quantization/bitsandbytes
  • 1 reply
Β·