πΈAnother BRIXEL in the Wall πΈ
πBRIXEL allows the user to produce high-resolution feature maps using the DINOv3 backbone without requiring large amounts of compute. Repo releasedπ
πReview https://t.ly/fZPwC
πPaper arxiv.org/pdf/2511.05168
πRepo github.com/alexanderlappe/BRIXEL
πBRIXEL allows the user to produce high-resolution feature maps using the DINOv3 backbone without requiring large amounts of compute. Repo releasedπ
πReview https://t.ly/fZPwC
πPaper arxiv.org/pdf/2511.05168
πRepo github.com/alexanderlappe/BRIXEL
π€©7π€―3π₯2β€1π1π1
This media is not supported in your browser
VIEW IN TELEGRAM
πΌPixel-Dense EmbeddingπΌ
πFlowFeat is a novel high-resolution and multi-task feature representation that embeds a distribution of plausible apparent motions, or motion profiles. Repo available under π
πReview https://t.ly/aUx_U
πPaper arxiv.org/pdf/2511.07696
πProject tum-vision.github.io/flowfeat
πRepo github.com/tum-vision/flowfeat
πFlowFeat is a novel high-resolution and multi-task feature representation that embeds a distribution of plausible apparent motions, or motion profiles. Repo available under π
πReview https://t.ly/aUx_U
πPaper arxiv.org/pdf/2511.07696
πProject tum-vision.github.io/flowfeat
πRepo github.com/tum-vision/flowfeat
π₯5π3β€2
π¨ Announcement π¨
Iβve received numerous reports of people blatantly copying my content on LinkedIn just to get a few likes.
Let me be very clear: I put a great deal of time and effort into reviewing papers and creating original, meaningful content. Itβs disappointing to see professionals (some of whom are even members of this group or my connections) resorting to plagiarism instead of contributing their own ideas.
π Starting today, Iβll be removing these connections from LinkedIn and banning such individuals from this group.
π’ I also encourage everyone to report these cases whenever you come across them. Every single report helps stop this bad habit and keeps our community fair, respectful, and authentic.
Iβve received numerous reports of people blatantly copying my content on LinkedIn just to get a few likes.
Let me be very clear: I put a great deal of time and effort into reviewing papers and creating original, meaningful content. Itβs disappointing to see professionals (some of whom are even members of this group or my connections) resorting to plagiarism instead of contributing their own ideas.
π Starting today, Iβll be removing these connections from LinkedIn and banning such individuals from this group.
π’ I also encourage everyone to report these cases whenever you come across them. Every single report helps stop this bad habit and keeps our community fair, respectful, and authentic.
β€64π21π17π’1
This media is not supported in your browser
VIEW IN TELEGRAM
π© Foundational Humanoid π©
π#NVIDIA unveils SONIC a novel foundational model for high-precision teleoperation & interactive control capabilities (running, jumping, crawling) with natural human-like movements. Code announcedπ
πReview https://t.ly/_3wnt
πPaper https://lnkd.in/dctfShu8
πProject https://lnkd.in/d_inmA2p
π#NVIDIA unveils SONIC a novel foundational model for high-precision teleoperation & interactive control capabilities (running, jumping, crawling) with natural human-like movements. Code announcedπ
πReview https://t.ly/_3wnt
πPaper https://lnkd.in/dctfShu8
πProject https://lnkd.in/d_inmA2p
π€―9β€4π1π₯1
This media is not supported in your browser
VIEW IN TELEGRAM
π₯Depth Anything 3 is outπ₯
πByteDance unveils Depth Anything 3 (DA3), a model that predicts spatially consistent geometry from arbitrary visual inputs, with or without known camera poses. Repo under Apache 2.0π
πReview https://t.ly/AOPu7
πPaper arxiv.org/pdf/2511.10647
πProject https://lnkd.in/dnByyn2z
πRepo https://lnkd.in/daCVz_4a
πDemo https://lnkd.in/dKUZiJt
πByteDance unveils Depth Anything 3 (DA3), a model that predicts spatially consistent geometry from arbitrary visual inputs, with or without known camera poses. Repo under Apache 2.0π
πReview https://t.ly/AOPu7
πPaper arxiv.org/pdf/2511.10647
πProject https://lnkd.in/dnByyn2z
πRepo https://lnkd.in/daCVz_4a
πDemo https://lnkd.in/dKUZiJt
π₯18β€9π1π1
This media is not supported in your browser
VIEW IN TELEGRAM
π©οΈ It's "Time-to-Move" π©οΈ
πTechnion + Nvidia Time-to-Move (TTM) is a training-free, plug-and-play framework for motion- and appearance-controlled video generation with I2V diffusion models (Wan 2.2, CogVideoX, & Stable VD). Impressive results!
πReview https://t.ly/0pwXm
πPaper https://lnkd.in/dxD3uHYb
πProject https://lnkd.in/dcE5juyM
πRepo https://lnkd.in/dMMUjybJ
πTechnion + Nvidia Time-to-Move (TTM) is a training-free, plug-and-play framework for motion- and appearance-controlled video generation with I2V diffusion models (Wan 2.2, CogVideoX, & Stable VD). Impressive results!
πReview https://t.ly/0pwXm
πPaper https://lnkd.in/dxD3uHYb
πProject https://lnkd.in/dcE5juyM
πRepo https://lnkd.in/dMMUjybJ
1π2π₯2β€1
This media is not supported in your browser
VIEW IN TELEGRAM
β Multi-Shot Video Segmentation β
πFudan focuses on an underexplored task of multi-shot video object segmentation (MVOS). Benchmark and repo available (the extension part of SAM) under Apache 2.0π
πReview https://t.ly/WBW00
πPaper https://arxiv.org/pdf/2511.13715
πProject https://henghuiding.com/SAAS/
πRepo https://github.com/FudanCVL/SAAS
πFudan focuses on an underexplored task of multi-shot video object segmentation (MVOS). Benchmark and repo available (the extension part of SAM) under Apache 2.0π
πReview https://t.ly/WBW00
πPaper https://arxiv.org/pdf/2511.13715
πProject https://henghuiding.com/SAAS/
πRepo https://github.com/FudanCVL/SAAS
1π₯6β€2
This media is not supported in your browser
VIEW IN TELEGRAM
π₯ SAM 3/3D are OUT!! π₯
π#META released SAM 3, a unified model for detection, segmentation, tracking of objects in images & video using text, exemplar & visual prompts. Repo/Models under proprietary licenseπ
πReview https://t.ly/lnRZN
πPaper https://t.ly/5tq9N
πProject https://ai.meta.com/sam3/
πDemo: https://segment-anything.com
πRepo https://github.com/facebookresearch/sam3
π#META released SAM 3, a unified model for detection, segmentation, tracking of objects in images & video using text, exemplar & visual prompts. Repo/Models under proprietary licenseπ
πReview https://t.ly/lnRZN
πPaper https://t.ly/5tq9N
πProject https://ai.meta.com/sam3/
πDemo: https://segment-anything.com
πRepo https://github.com/facebookresearch/sam3
π₯22β€4π1
This media is not supported in your browser
VIEW IN TELEGRAM
π―Unwrapping of 3D Meshesπ―
πPartUV is a novel part-based UV unwrapping method for 3D meshes; it combines learned part priors with geometric cues to generate a compact set of part-aligned charts. Repo releasedπ
πReview https://t.ly/8dNIY
πPaper arxiv.org/pdf/2511.16659
πProject www.zhaoningwang.com/PartUV/
πRepo github.com/EricWang12/PartUV
πPartUV is a novel part-based UV unwrapping method for 3D meshes; it combines learned part priors with geometric cues to generate a compact set of part-aligned charts. Repo releasedπ
πReview https://t.ly/8dNIY
πPaper arxiv.org/pdf/2511.16659
πProject www.zhaoningwang.com/PartUV/
πRepo github.com/EricWang12/PartUV
β€14π2π₯1
π Upsample Anything π
πUpsample Anything, a novel universal, training-free up-sampler via lightweight test-time optimization. No code but it's a relevant paperπ
πReview https://t.ly/7LE6G
πPaper https://lnkd.in/dsUfdtih
πUpsample Anything, a novel universal, training-free up-sampler via lightweight test-time optimization. No code but it's a relevant paperπ
πReview https://t.ly/7LE6G
πPaper https://lnkd.in/dsUfdtih
π₯7β€3π2π1
This media is not supported in your browser
VIEW IN TELEGRAM
π¦Single Synthetic Image per Classπ¦
πMIT unveils Linear Gradient Matching (H/T Torralba), a novel method of distillation to use a single synthetic image per class for linear classifiers training (and more). Repo availableπ
πReview https://t.ly/dD3un
πPaper arxiv.org/pdf/2511.16674
πProject linear-gradient-matching.github.io/
πRepo github.com/GeorgeCazenavette/linear-gradient-matching
πMIT unveils Linear Gradient Matching (H/T Torralba), a novel method of distillation to use a single synthetic image per class for linear classifiers training (and more). Repo availableπ
πReview https://t.ly/dD3un
πPaper arxiv.org/pdf/2511.16674
πProject linear-gradient-matching.github.io/
πRepo github.com/GeorgeCazenavette/linear-gradient-matching
1β€6π₯2π1π1
This media is not supported in your browser
VIEW IN TELEGRAM
π§ͺ EfficientSAM3 is out π§ͺ
πBristol announces EfficientSAM3, a family of efficient models built on Progressive Hierarchical Distillation that transfers capability from SAM3 to lightweight students. Code coming (in sync with SAM3 release)π
πReview https://t.ly/bfXP2
πPaper arxiv.org/pdf/2511.15833
πProject simonzeng7108.github.io/efficientsam3/
πRepo github.com/SimonZeng7108/efficientsam3
πBristol announces EfficientSAM3, a family of efficient models built on Progressive Hierarchical Distillation that transfers capability from SAM3 to lightweight students. Code coming (in sync with SAM3 release)π
πReview https://t.ly/bfXP2
πPaper arxiv.org/pdf/2511.15833
πProject simonzeng7108.github.io/efficientsam3/
πRepo github.com/SimonZeng7108/efficientsam3
β€3π2π₯1π1
This media is not supported in your browser
VIEW IN TELEGRAM
π©οΈ Cloud4D in time π©οΈ
πCloud4D: physically-realistic 3D cloud fields using ground-based cameras at a 25 m spatial resolution and 5 s temporal resolution. Repo coming, Data releasedπ
πReview https://t.ly/w7Zly
πPaper arxiv.org/pdf/2511.19431
πProject cloud4d.jacob-lin.com/
πData https://drive.google.com/drive/folders/1QU_0kIUXIVt8h3uqygBeaF3Gvr_L5SdX?usp=drive_link
πRepo TBA
πCloud4D: physically-realistic 3D cloud fields using ground-based cameras at a 25 m spatial resolution and 5 s temporal resolution. Repo coming, Data releasedπ
πReview https://t.ly/w7Zly
πPaper arxiv.org/pdf/2511.19431
πProject cloud4d.jacob-lin.com/
πData https://drive.google.com/drive/folders/1QU_0kIUXIVt8h3uqygBeaF3Gvr_L5SdX?usp=drive_link
πRepo TBA
π₯5