This media is not supported in your browser
VIEW IN TELEGRAM
๐ฑPlanarTrack: Large Planar Tracking๐ฑ
๐PlanarTrack is a large-scale HQ and challenging benchmark for planar tracking: 1,150 sequences with 733K+ frames, including 1,000 short-term & 150 long-term videos. Repo & Dataset available๐
๐Review https://t.ly/mYNi7
๐Paper arxiv.org/pdf/2510.23368
๐Repo https://lnkd.in/edb3GMyT
๐Project https://lnkd.in/eC-hVB-U
๐Data https://lnkd.in/eew2j4tM
๐PlanarTrack is a large-scale HQ and challenging benchmark for planar tracking: 1,150 sequences with 733K+ frames, including 1,000 short-term & 150 long-term videos. Repo & Dataset available๐
๐Review https://t.ly/mYNi7
๐Paper arxiv.org/pdf/2510.23368
๐Repo https://lnkd.in/edb3GMyT
๐Project https://lnkd.in/eC-hVB-U
๐Data https://lnkd.in/eew2j4tM
๐ฅ11โค5๐2๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ขGenerative View Stitching ๐ข
๐GVS is a novel approach that enables collision-free camera-guided video generation for predefined trajectories, it's a non-autoregressive alternative to video length extrapolation. Full repo under MIT๐
๐Review https://t.ly/TiN_5
๐Paper https://arxiv.org/pdf/2510.24718
๐Project https://andrewsonga.github.io/gvs/
๐Repo github.com/andrewsonga/generative_view_stitching
๐GVS is a novel approach that enables collision-free camera-guided video generation for predefined trajectories, it's a non-autoregressive alternative to video length extrapolation. Full repo under MIT๐
๐Review https://t.ly/TiN_5
๐Paper https://arxiv.org/pdf/2510.24718
๐Project https://andrewsonga.github.io/gvs/
๐Repo github.com/andrewsonga/generative_view_stitching
๐ฅ10โค3๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ชTracking Object Transformations๐ช
๐"Track Any State": tracking objects through transformations while detecting/describing state changes. Repo & Dataset available under MIT๐
๐Review https://t.ly/NPyW4
๐Paper https://lnkd.in/d4pA3bXJ
๐Project https://lnkd.in/dgbNfCuj
๐Repo https://lnkd.in/dtVWq2z7
๐"Track Any State": tracking objects through transformations while detecting/describing state changes. Repo & Dataset available under MIT๐
๐Review https://t.ly/NPyW4
๐Paper https://lnkd.in/d4pA3bXJ
๐Project https://lnkd.in/dgbNfCuj
๐Repo https://lnkd.in/dtVWq2z7
๐ฅ20โค7๐คฏ3๐2๐1
๐ธAnother BRIXEL in the Wall ๐ธ
๐BRIXEL allows the user to produce high-resolution feature maps using the DINOv3 backbone without requiring large amounts of compute. Repo released๐
๐Review https://t.ly/fZPwC
๐Paper arxiv.org/pdf/2511.05168
๐Repo github.com/alexanderlappe/BRIXEL
๐BRIXEL allows the user to produce high-resolution feature maps using the DINOv3 backbone without requiring large amounts of compute. Repo released๐
๐Review https://t.ly/fZPwC
๐Paper arxiv.org/pdf/2511.05168
๐Repo github.com/alexanderlappe/BRIXEL
๐คฉ7๐คฏ3๐ฅ2โค1๐1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ผPixel-Dense Embedding๐ผ
๐FlowFeat is a novel high-resolution and multi-task feature representation that embeds a distribution of plausible apparent motions, or motion profiles. Repo available under ๐
๐Review https://t.ly/aUx_U
๐Paper arxiv.org/pdf/2511.07696
๐Project tum-vision.github.io/flowfeat
๐Repo github.com/tum-vision/flowfeat
๐FlowFeat is a novel high-resolution and multi-task feature representation that embeds a distribution of plausible apparent motions, or motion profiles. Repo available under ๐
๐Review https://t.ly/aUx_U
๐Paper arxiv.org/pdf/2511.07696
๐Project tum-vision.github.io/flowfeat
๐Repo github.com/tum-vision/flowfeat
๐ฅ5๐3โค2
๐จ Announcement ๐จ
Iโve received numerous reports of people blatantly copying my content on LinkedIn just to get a few likes.
Let me be very clear: I put a great deal of time and effort into reviewing papers and creating original, meaningful content. Itโs disappointing to see professionals (some of whom are even members of this group or my connections) resorting to plagiarism instead of contributing their own ideas.
๐ Starting today, Iโll be removing these connections from LinkedIn and banning such individuals from this group.
๐ข I also encourage everyone to report these cases whenever you come across them. Every single report helps stop this bad habit and keeps our community fair, respectful, and authentic.
Iโve received numerous reports of people blatantly copying my content on LinkedIn just to get a few likes.
Let me be very clear: I put a great deal of time and effort into reviewing papers and creating original, meaningful content. Itโs disappointing to see professionals (some of whom are even members of this group or my connections) resorting to plagiarism instead of contributing their own ideas.
๐ Starting today, Iโll be removing these connections from LinkedIn and banning such individuals from this group.
๐ข I also encourage everyone to report these cases whenever you come across them. Every single report helps stop this bad habit and keeps our community fair, respectful, and authentic.
โค64๐21๐17๐ข1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฉ Foundational Humanoid ๐ฉ
๐#NVIDIA unveils SONIC a novel foundational model for high-precision teleoperation & interactive control capabilities (running, jumping, crawling) with natural human-like movements. Code announced๐
๐Review https://t.ly/_3wnt
๐Paper https://lnkd.in/dctfShu8
๐Project https://lnkd.in/d_inmA2p
๐#NVIDIA unveils SONIC a novel foundational model for high-precision teleoperation & interactive control capabilities (running, jumping, crawling) with natural human-like movements. Code announced๐
๐Review https://t.ly/_3wnt
๐Paper https://lnkd.in/dctfShu8
๐Project https://lnkd.in/d_inmA2p
๐คฏ9โค4๐1๐ฅ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฅDepth Anything 3 is out๐ฅ
๐ByteDance unveils Depth Anything 3 (DA3), a model that predicts spatially consistent geometry from arbitrary visual inputs, with or without known camera poses. Repo under Apache 2.0๐
๐Review https://t.ly/AOPu7
๐Paper arxiv.org/pdf/2511.10647
๐Project https://lnkd.in/dnByyn2z
๐Repo https://lnkd.in/daCVz_4a
๐Demo https://lnkd.in/dKUZiJt
๐ByteDance unveils Depth Anything 3 (DA3), a model that predicts spatially consistent geometry from arbitrary visual inputs, with or without known camera poses. Repo under Apache 2.0๐
๐Review https://t.ly/AOPu7
๐Paper arxiv.org/pdf/2511.10647
๐Project https://lnkd.in/dnByyn2z
๐Repo https://lnkd.in/daCVz_4a
๐Demo https://lnkd.in/dKUZiJt
๐ฅ18โค9๐1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฉ๏ธ It's "Time-to-Move" ๐ฉ๏ธ
๐Technion + Nvidia Time-to-Move (TTM) is a training-free, plug-and-play framework for motion- and appearance-controlled video generation with I2V diffusion models (Wan 2.2, CogVideoX, & Stable VD). Impressive results!
๐Review https://t.ly/0pwXm
๐Paper https://lnkd.in/dxD3uHYb
๐Project https://lnkd.in/dcE5juyM
๐Repo https://lnkd.in/dMMUjybJ
๐Technion + Nvidia Time-to-Move (TTM) is a training-free, plug-and-play framework for motion- and appearance-controlled video generation with I2V diffusion models (Wan 2.2, CogVideoX, & Stable VD). Impressive results!
๐Review https://t.ly/0pwXm
๐Paper https://lnkd.in/dxD3uHYb
๐Project https://lnkd.in/dcE5juyM
๐Repo https://lnkd.in/dMMUjybJ
1๐2๐ฅ2โค1
This media is not supported in your browser
VIEW IN TELEGRAM
โ Multi-Shot Video Segmentation โ
๐Fudan focuses on an underexplored task of multi-shot video object segmentation (MVOS). Benchmark and repo available (the extension part of SAM) under Apache 2.0๐
๐Review https://t.ly/WBW00
๐Paper https://arxiv.org/pdf/2511.13715
๐Project https://henghuiding.com/SAAS/
๐Repo https://github.com/FudanCVL/SAAS
๐Fudan focuses on an underexplored task of multi-shot video object segmentation (MVOS). Benchmark and repo available (the extension part of SAM) under Apache 2.0๐
๐Review https://t.ly/WBW00
๐Paper https://arxiv.org/pdf/2511.13715
๐Project https://henghuiding.com/SAAS/
๐Repo https://github.com/FudanCVL/SAAS
1๐ฅ6โค2
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฅ SAM 3/3D are OUT!! ๐ฅ
๐#META released SAM 3, a unified model for detection, segmentation, tracking of objects in images & video using text, exemplar & visual prompts. Repo/Models under proprietary license๐
๐Review https://t.ly/lnRZN
๐Paper https://t.ly/5tq9N
๐Project https://ai.meta.com/sam3/
๐Demo: https://segment-anything.com
๐Repo https://github.com/facebookresearch/sam3
๐#META released SAM 3, a unified model for detection, segmentation, tracking of objects in images & video using text, exemplar & visual prompts. Repo/Models under proprietary license๐
๐Review https://t.ly/lnRZN
๐Paper https://t.ly/5tq9N
๐Project https://ai.meta.com/sam3/
๐Demo: https://segment-anything.com
๐Repo https://github.com/facebookresearch/sam3
๐ฅ22โค4๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฏUnwrapping of 3D Meshes๐ฏ
๐PartUV is a novel part-based UV unwrapping method for 3D meshes; it combines learned part priors with geometric cues to generate a compact set of part-aligned charts. Repo released๐
๐Review https://t.ly/8dNIY
๐Paper arxiv.org/pdf/2511.16659
๐Project www.zhaoningwang.com/PartUV/
๐Repo github.com/EricWang12/PartUV
๐PartUV is a novel part-based UV unwrapping method for 3D meshes; it combines learned part priors with geometric cues to generate a compact set of part-aligned charts. Repo released๐
๐Review https://t.ly/8dNIY
๐Paper arxiv.org/pdf/2511.16659
๐Project www.zhaoningwang.com/PartUV/
๐Repo github.com/EricWang12/PartUV
โค14๐2๐ฅ1
๐ Upsample Anything ๐
๐Upsample Anything, a novel universal, training-free up-sampler via lightweight test-time optimization. No code but it's a relevant paper๐
๐Review https://t.ly/7LE6G
๐Paper https://lnkd.in/dsUfdtih
๐Upsample Anything, a novel universal, training-free up-sampler via lightweight test-time optimization. No code but it's a relevant paper๐
๐Review https://t.ly/7LE6G
๐Paper https://lnkd.in/dsUfdtih
๐ฅ7โค3๐2๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฆSingle Synthetic Image per Class๐ฆ
๐MIT unveils Linear Gradient Matching (H/T Torralba), a novel method of distillation to use a single synthetic image per class for linear classifiers training (and more). Repo available๐
๐Review https://t.ly/dD3un
๐Paper arxiv.org/pdf/2511.16674
๐Project linear-gradient-matching.github.io/
๐Repo github.com/GeorgeCazenavette/linear-gradient-matching
๐MIT unveils Linear Gradient Matching (H/T Torralba), a novel method of distillation to use a single synthetic image per class for linear classifiers training (and more). Repo available๐
๐Review https://t.ly/dD3un
๐Paper arxiv.org/pdf/2511.16674
๐Project linear-gradient-matching.github.io/
๐Repo github.com/GeorgeCazenavette/linear-gradient-matching
1โค6๐ฅ2๐1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐งช EfficientSAM3 is out ๐งช
๐Bristol announces EfficientSAM3, a family of efficient models built on Progressive Hierarchical Distillation that transfers capability from SAM3 to lightweight students. Code coming (in sync with SAM3 release)๐
๐Review https://t.ly/bfXP2
๐Paper arxiv.org/pdf/2511.15833
๐Project simonzeng7108.github.io/efficientsam3/
๐Repo github.com/SimonZeng7108/efficientsam3
๐Bristol announces EfficientSAM3, a family of efficient models built on Progressive Hierarchical Distillation that transfers capability from SAM3 to lightweight students. Code coming (in sync with SAM3 release)๐
๐Review https://t.ly/bfXP2
๐Paper arxiv.org/pdf/2511.15833
๐Project simonzeng7108.github.io/efficientsam3/
๐Repo github.com/SimonZeng7108/efficientsam3
โค4๐2๐ฅ1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฉ๏ธ Cloud4D in time ๐ฉ๏ธ
๐Cloud4D: physically-realistic 3D cloud fields using ground-based cameras at a 25 m spatial resolution and 5 s temporal resolution. Repo coming, Data released๐
๐Review https://t.ly/w7Zly
๐Paper arxiv.org/pdf/2511.19431
๐Project cloud4d.jacob-lin.com/
๐Data https://drive.google.com/drive/folders/1QU_0kIUXIVt8h3uqygBeaF3Gvr_L5SdX?usp=drive_link
๐Repo TBA
๐Cloud4D: physically-realistic 3D cloud fields using ground-based cameras at a 25 m spatial resolution and 5 s temporal resolution. Repo coming, Data released๐
๐Review https://t.ly/w7Zly
๐Paper arxiv.org/pdf/2511.19431
๐Project cloud4d.jacob-lin.com/
๐Data https://drive.google.com/drive/folders/1QU_0kIUXIVt8h3uqygBeaF3Gvr_L5SdX?usp=drive_link
๐Repo TBA
๐ฅ7