The Nobel Prize background for Hopfield and Hinton's work on neural networks is pure gold. It's a masterclass in explaining AI basics.
Key takeaways from the conclusion: - ML applications are expanding rapidly. We're still figuring out which will stick. - Ethical discussions are crucial as the tech develops. - Physics 🤝 AI: A two-way street of innovation.
Some mind-blowing AI applications in physics: - Discovering the Higgs particle - Cleaning up gravitational wave data - Hunting exoplanets - Predicting molecular structures - Designing better solar cells
We're just scratching the surface. The interplay between AI and physics is reshaping both fields.
Bonus: The illustrations accompanying the background document are really neat. (Credit: Johan Jarnestad/The Royal Swedish Academy of Sciences)
Meta AI vision has been cooking @facebook They shipped multiple models and demos for their papers at @ECCV🤗
Here's a compilation of my top picks: - Sapiens is family of foundation models for human-centric depth estimation, segmentation and more, all models have open weights and demos 👏
All models have their demos and even torchscript checkpoints! A collection of models and demos: facebook/sapiens-66d22047daa6402d565cb2fc - VFusion3D is state-of-the-art consistent 3D generation model from images
🧠 Stanford paper might be the key to OpenAI o1’s performance: What’s so effective about Chain of Thought? ⇒ it unlocks radically different sequential tasks!
💭 Reminder: A Chain of Thought (CoT) means that you instruct the model to “think step by step”. Often it’s literally just putting in the prompt “let’s think step by step.”
🤔 This method has been shown to be unreasonably effective to increase perf on benchmarks. However why it works so well remains unclear.
Here's the scoop: Transformers are amazing at parallel processing, but they've always struggled with tasks that require sequential reasoning.
⛔️ For instance if you ask them the result of 3^2^2^2^…, with 20 iterations, they’ll nearly always fail.
💡 Indeed, researchers prove mathematically, by assimilating transformers networks to logical circuits, that effectively they cannot solve sequential tasks that require more than a certain threshold of sequences.
But CoT enables sequential reasoning:
- 🧱 Each step in the CoT corresponds to simulating one operation in a complex circuit. - 🔄 This allows the transformer to "reset" the depth of intermediate outputs, overcoming previous limitations. - 🚀 Thus, with CoT, constant-depth transformers can now solve ANY problem computable by polynomial-size circuits! (That's a huge class of problems in computer science.) - 🔑 Transformers can now handle tricky tasks like iterated squares (computing 3^2^2^2^2) composed permutations and evaluating circuits - stuff that requires serial computation. - 📊 The improvement is especially dramatic for transformers with a limited depth. Empirical tests on four arithmetic problems showed massive accuracy gains with CoT on inherently serial tasks.
Main takeaway: Chain-of-thought isn't just a neat trick - it fundamentally expands what transformer models can do!
🔥🎭🌟 New Research Alert - HeadGAP (Avatars Collection)! 🌟🎭🔥 📄 Title: HeadGAP: Few-shot 3D Head Avatar via Generalizable Gaussian Priors 🔝
📝 Description: HeadGAP introduces a novel method for generating high-fidelity, animatable 3D head avatars from few-shot data, using Gaussian priors and dynamic part-based modelling for personalized and generalizable results.
🔥🎭🌟 New Research Alert - ECCV 2024 (Avatars Collection)! 🌟🎭🔥 📄 Title: MeshAvatar: Learning High-quality Triangular Human Avatars from Multi-view Videos 🔝
📝 Description: MeshAvatar is a novel pipeline that generates high-quality triangular human avatars from multi-view videos, enabling realistic editing and rendering through a mesh-based approach with physics-based decomposition.
👥 Authors: Yushuo Chen, Zerong Zheng, Zhe Li, Chao Xu, and Yebin Liu
Pro Tip - if you're a Firefox user, you can set up Hugging Chat as integrated AI Assistant, with contextual links to summarize or simplify any text - handy!