AI with Papers - Artificial Intelligence & Deep Learning
15.2K subscribers
135 photos
247 videos
13 files
1.31K links
All the AI with papers. Every day fresh updates about #DeepLearning, #MachineLearning, LLMs and #ComputerVision

Curated by Alessandro Ferrari | https://www.linkedin.com/in/visionarynet/

#artificialintelligence #machinelearning #ml #AI
Download Telegram
panohead_overview-min.gif
24.3 MB
🍥 PanoHead: 3D Full-Head Synthesis 🍥

👉#ByteDance (+UW-M) unveils PanoHead: 360◦ view-consistent portraits from a single-view image

😎Review https://t.ly/MrLNR
😎Paper arxiv.org/pdf/2303.13071.pdf
😎Project sizhean.github.io/panohead
😎Code github.com/sizhean/panohead
🔥74🤯3😱1
This media is not supported in your browser
VIEW IN TELEGRAM
🐤 MagicVideo-V2 announced! 🐤

👉#Bytedance announces a novel multi-stage pipeline capable of generating high-aesthetic videos from textual description

👉Review https://t.ly/zIq4v
👉Project https://lnkd.in/dKUrJPJd
👉Paper https://lnkd.in/dixnN-kU
🔥71👍1🥰1💩1
This media is not supported in your browser
VIEW IN TELEGRAM
🆔 Magic-Me: ID-Specific Video 🆔

👉#ByteDance VCD: with just a few images of a specific identity it can generate temporal consistent videos aligned with the given prompt

👉Review https://t.ly/qjJ2O
👉Paper arxiv.org/pdf/2402.09368.pdf
👉Project magic-me-webpage.github.io
👉Code github.com/Zhen-Dong/Magic-Me
6🥰1🤯1🤣1
This media is not supported in your browser
VIEW IN TELEGRAM
VoRA: Vision as LoRA

👉#ByteDance unveils Vision as LoRA (VoRA), a novel paradigm converting LLMs into Multimodal Large Language Models (MLLMs) by integrating vision-specific LoRA layers. All training data, codes, and model weights available💙

👉Review https://t.ly/guNVN
👉Paper arxiv.org/pdf/2503.20680
👉Repo github.com/Hon-Wong/VoRA
👉Project georgeluimmortal.github.io/vora-homepage.github.io/
👍157🤯4👏1
ezgif-8120c4563e81c3.mp4
510.6 KB
🥶 OmniHuman-1.5 🥶

👉#ByteDance proposes a novel framework designed to generate character animations that are not only physically plausible but also semantically coherent and expressive. Coherency with speech's rhythm, prosody and semantic content. Impressive results but no code 🥺

👉Review https://t.ly/CnRmX
👉Paper arxiv.org/pdf/2508.19209
👉Project omnihuman-lab.github.io/v1_5/
👉Repo 🥺
4🤯2👍1🔥1
This media is not supported in your browser
VIEW IN TELEGRAM
🐙Human-Centric Video Generation🐙

👉Tsinghua & #ByteDance unveil HuMo: a unified, human-centric video generation framework designed to produce HQ fine-grained, and controllable human videos from multimodal inputs: text prompt following, consistent subject preservation, synchronized audio-driven motion. Repo released under Apache2.0💙

👉Review https://t.ly/3S8Yb
👉Paper https://arxiv.org/pdf/2509.08519
👉Project https://phantom-video.github.io/HuMo/
👉Repo https://github.com/Phantom-video/HuMo
🔥8🤯32👏1