β¨DINOv3
π Summary:
DINOv3 is a self-supervised vision model excelling across tasks. It scales datasets, prevents dense feature degradation via Gram anchoring, and uses post-hoc strategies for flexibility. This versatile foundation model outperforms specialized state of the art without fine-tuning.
πΉ Publication Date: Published on Aug 13
πΉ Paper Links:
β’ arXiv Page: https://huggingface.co/collections/facebook/dinov3
β’ PDF: https://arxiv.org/pdf/2508.10104
β’ Project Page: https://ai.meta.com/blog/dinov3-self-supervised-vision-model/
β’ Github: https://github.com/facebookresearch/dinov3
πΉ Models citing this paper:
β’ https://huggingface.co/facebook/dinov3-vit7b16-pretrain-lvd1689m
β’ https://huggingface.co/facebook/dinov3-vitb16-pretrain-lvd1689m
β’ https://huggingface.co/facebook/dinov3-vitl16-pretrain-lvd1689m
β¨ Datasets citing this paper:
β’ https://huggingface.co/datasets/zhuangzhe1229/test_dataset
β’ https://huggingface.co/datasets/simon123905/vitl
β¨ Spaces citing this paper:
β’ https://huggingface.co/spaces/atalaydenknalbant/DINOv3
β’ https://huggingface.co/spaces/manu02/DINOv3-Interactive-Patch-Cosine-Similarity
β’ https://huggingface.co/spaces/merve/dinov3-viz
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#DINOv3 #SelfSupervisedLearning #ComputerVision #FoundationModels #AI
π Summary:
DINOv3 is a self-supervised vision model excelling across tasks. It scales datasets, prevents dense feature degradation via Gram anchoring, and uses post-hoc strategies for flexibility. This versatile foundation model outperforms specialized state of the art without fine-tuning.
πΉ Publication Date: Published on Aug 13
πΉ Paper Links:
β’ arXiv Page: https://huggingface.co/collections/facebook/dinov3
β’ PDF: https://arxiv.org/pdf/2508.10104
β’ Project Page: https://ai.meta.com/blog/dinov3-self-supervised-vision-model/
β’ Github: https://github.com/facebookresearch/dinov3
πΉ Models citing this paper:
β’ https://huggingface.co/facebook/dinov3-vit7b16-pretrain-lvd1689m
β’ https://huggingface.co/facebook/dinov3-vitb16-pretrain-lvd1689m
β’ https://huggingface.co/facebook/dinov3-vitl16-pretrain-lvd1689m
β¨ Datasets citing this paper:
β’ https://huggingface.co/datasets/zhuangzhe1229/test_dataset
β’ https://huggingface.co/datasets/simon123905/vitl
β¨ Spaces citing this paper:
β’ https://huggingface.co/spaces/atalaydenknalbant/DINOv3
β’ https://huggingface.co/spaces/manu02/DINOv3-Interactive-Patch-Cosine-Similarity
β’ https://huggingface.co/spaces/merve/dinov3-viz
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#DINOv3 #SelfSupervisedLearning #ComputerVision #FoundationModels #AI
huggingface.co
DINOv3 - a facebook Collection
DINOv3: foundation models producing excellent dense features, outperforming SotA w/o fine-tuning - https://arxiv.org/abs/2508.10104
π€π§ Concerto: How Joint 2D-3D Self-Supervised Learning Is Redefining Spatial Intelligence
ποΈ 09 Nov 2025
π AI News & Trends
The world of artificial intelligence is rapidly evolving and self-supervised learning has become a driving force behind breakthroughs in computer vision and 3D scene understanding. Traditional supervised learning relies heavily on labeled datasets which are expensive and time-consuming to produce. Self-supervised learning, on the other hand, extracts meaningful patterns without manual labels allowing models to ...
#SelfSupervisedLearning #ComputerVision #3DSceneUnderstanding #SpatialIntelligence #AIResearch #DeepLearning
ποΈ 09 Nov 2025
π AI News & Trends
The world of artificial intelligence is rapidly evolving and self-supervised learning has become a driving force behind breakthroughs in computer vision and 3D scene understanding. Traditional supervised learning relies heavily on labeled datasets which are expensive and time-consuming to produce. Self-supervised learning, on the other hand, extracts meaningful patterns without manual labels allowing models to ...
#SelfSupervisedLearning #ComputerVision #3DSceneUnderstanding #SpatialIntelligence #AIResearch #DeepLearning
β¨VideoSSR: Video Self-Supervised Reinforcement Learning
π Summary:
VideoSSR is a novel self-supervised reinforcement learning framework that leverages intrinsic video information to generate high-quality training data. It uses three pretext tasks and the VideoSSR-30K dataset, improving MLLM performance across 17 benchmarks by over 5%.
πΉ Publication Date: Published on Nov 9
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.06281
β’ PDF: https://arxiv.org/pdf/2511.06281
β’ Project Page: https://github.com/lcqysl/VideoSSR
β’ Github: https://github.com/lcqysl/VideoSSR
πΉ Models citing this paper:
β’ https://huggingface.co/yhx12/VideoSSR
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#ReinforcementLearning #SelfSupervisedLearning #VideoAI #MachineLearning #DeepLearning
π Summary:
VideoSSR is a novel self-supervised reinforcement learning framework that leverages intrinsic video information to generate high-quality training data. It uses three pretext tasks and the VideoSSR-30K dataset, improving MLLM performance across 17 benchmarks by over 5%.
πΉ Publication Date: Published on Nov 9
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.06281
β’ PDF: https://arxiv.org/pdf/2511.06281
β’ Project Page: https://github.com/lcqysl/VideoSSR
β’ Github: https://github.com/lcqysl/VideoSSR
πΉ Models citing this paper:
β’ https://huggingface.co/yhx12/VideoSSR
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#ReinforcementLearning #SelfSupervisedLearning #VideoAI #MachineLearning #DeepLearning
β¨OlmoEarth: Stable Latent Image Modeling for Multimodal Earth Observation
π Summary:
OlmoEarth is a novel multimodal spatio-temporal foundation model for Earth observation data. It employs new self-supervised learning methods to achieve state-of-the-art performance on many tasks. It is deployed as a platform for non-profits and NGOs.
πΉ Publication Date: Published on Nov 17
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.13655
β’ PDF: https://arxiv.org/pdf/2511.13655
β’ Project Page: https://olmoearth.allenai.org/
β’ Github: https://github.com/allenai/olmoearth_pretrain
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#EarthObservation #FoundationModels #AI #RemoteSensing #SelfSupervisedLearning
π Summary:
OlmoEarth is a novel multimodal spatio-temporal foundation model for Earth observation data. It employs new self-supervised learning methods to achieve state-of-the-art performance on many tasks. It is deployed as a platform for non-profits and NGOs.
πΉ Publication Date: Published on Nov 17
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.13655
β’ PDF: https://arxiv.org/pdf/2511.13655
β’ Project Page: https://olmoearth.allenai.org/
β’ Github: https://github.com/allenai/olmoearth_pretrain
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#EarthObservation #FoundationModels #AI #RemoteSensing #SelfSupervisedLearning
β¨UnSAMv2: Self-Supervised Learning Enables Segment Anything at Any Granularity
π Summary:
UnSAMv2 enables continuous segmentation granularity control for the SAM model without human annotations. It uses self-supervised learning on unlabeled data to discover mask-granularity pairs and a novel control embedding. UnSAMv2 significantly enhances SAM-2s performance across various segmentati...
πΉ Publication Date: Published on Nov 17
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.13714
β’ PDF: https://arxiv.org/pdf/2511.13714
β’ Project Page: https://yujunwei04.github.io/UnSAMv2-Project-Page/
β’ Github: https://github.com/yujunwei04/UnSAMv2
β¨ Spaces citing this paper:
β’ https://huggingface.co/spaces/yujunwei04/UnSAMv2
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#AI #ComputerVision #SelfSupervisedLearning #ImageSegmentation #DeepLearning
π Summary:
UnSAMv2 enables continuous segmentation granularity control for the SAM model without human annotations. It uses self-supervised learning on unlabeled data to discover mask-granularity pairs and a novel control embedding. UnSAMv2 significantly enhances SAM-2s performance across various segmentati...
πΉ Publication Date: Published on Nov 17
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.13714
β’ PDF: https://arxiv.org/pdf/2511.13714
β’ Project Page: https://yujunwei04.github.io/UnSAMv2-Project-Page/
β’ Github: https://github.com/yujunwei04/UnSAMv2
β¨ Spaces citing this paper:
β’ https://huggingface.co/spaces/yujunwei04/UnSAMv2
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#AI #ComputerVision #SelfSupervisedLearning #ImageSegmentation #DeepLearning
β¨Ξ¦eat: Physically-Grounded Feature Representation
π Summary:
Ξ¦eat is a new self-supervised visual backbone that captures material identity like reflectance and mesostructure. It learns robust features invariant to external physical factors such as shape and lighting, promoting physics-aware perception.
πΉ Publication Date: Published on Nov 14
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.11270
β’ PDF: https://arxiv.org/pdf/2511.11270
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#ComputerVision #SelfSupervisedLearning #DeepLearning #FeatureLearning #PhysicsAwareAI
π Summary:
Ξ¦eat is a new self-supervised visual backbone that captures material identity like reflectance and mesostructure. It learns robust features invariant to external physical factors such as shape and lighting, promoting physics-aware perception.
πΉ Publication Date: Published on Nov 14
πΉ Paper Links:
β’ arXiv Page: https://arxiv.org/abs/2511.11270
β’ PDF: https://arxiv.org/pdf/2511.11270
==================================
For more data science resources:
β https://t.iss.one/DataScienceT
#ComputerVision #SelfSupervisedLearning #DeepLearning #FeatureLearning #PhysicsAwareAI