spooner

spooner2

AI & ML interests

None yet

Recent Activity

Organizations

None yet

spooner2's activity

Reacted to yongchanghao's post with ๐Ÿ”ฅ 27 days ago
view post
Post
3734
We just released a paper (NeuZip) that compresses VRAM in a lossless manner to run larger models. This should be particularly useful when VRAM is insufficient during training/inference. Specifically, we look inside each floating number and find that the exponents are highly compressible (as shown in the figure below).

Read more about the work at NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks (2410.20650)
Reacted to codelion's post with ๐Ÿš€ 7 months ago
view post
Post
1759
Happy to announce the open source framework to turbo charge devops called patchwork - https://github.com/patched-codes/patchwork

You can use it to build patchflows - workflows that use LLMs for software development tasks like bug fixing, pull request review, library migration and documentation.

Supports any LLM of your choice including our own MoE model - patched-codes/patched-mix-4x7B

Give it a try!
  • 2 replies
ยท