Data Science | Machine Learning with Python for Researchers
32.6K subscribers
3.31K photos
125 videos
23 files
3.52K links
ads: @HusseinSheikho

The Data Science and Python channel is for researchers and advanced programmers

Buy ads: https://telega.io/c/dataScienceT
Download Telegram
πŸ€–πŸ§  Skyvern: The Future of Browser Automation Powered by AI and Computer Vision

πŸ—“οΈ 16 Nov 2025
πŸ“š AI News & Trends

In today’s fast-evolving digital landscape, automation plays a crucial role in enhancing productivity, efficiency and innovation. Yet, traditional browser automation tools often struggle with complexity, maintenance and reliability. They rely heavily on DOM parsing, XPaths and rigid scripts that easily break when websites change their layout. Enter Skyvern, an open-source, AI-driven browser automation platform developed ...

#Skyvern #BrowserAutomation #AIDriven #ComputerVision #OpenSource #WebAutomation
❀‍πŸ”₯1❀1πŸ‘1
✨SpatialThinker: Reinforcing 3D Reasoning in Multimodal LLMs via Spatial Rewards

πŸ“ Summary:
SpatialThinker is a new 3D-aware MLLM that uses RL and dense spatial rewards to significantly improve spatial understanding. It integrates structured spatial grounding and multi-step reasoning, outperforming existing models and GPT-4o on spatial VQA and real-world benchmarks.

πŸ”Ή Publication Date: Published on Nov 10

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.07403
β€’ PDF: https://arxiv.org/pdf/2511.07403
β€’ Github: https://github.com/hunarbatra/SpatialThinker

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/OX-PIXL/SpatialThinker-3B
β€’ https://huggingface.co/OX-PIXL/SpatialThinker-7B

✨ Datasets citing this paper:
β€’ https://huggingface.co/datasets/OX-PIXL/STVQA-7K

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#MultimodalLLM #3DReasoning #ReinforcementLearning #AIResearch #ComputerVision
✨WEAVE: Unleashing and Benchmarking the In-context Interleaved Comprehension and Generation

πŸ“ Summary:
WEAVE introduces a suite with a large dataset and benchmark to assess multi-turn context-dependent image generation and editing in multimodal models. It enables new capabilities like visual memory in models while exposing current limitations in these complex tasks.

πŸ”Ή Publication Date: Published on Nov 14

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.11434
β€’ PDF: https://arxiv.org/pdf/2511.11434
β€’ Project Page: https://weichow23.github.io/weave/
β€’ Github: https://github.com/weichow23/weave

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#MultimodalAI #ImageGeneration #GenerativeAI #ComputerVision #AIResearch
✨RF-DETR: Neural Architecture Search for Real-Time Detection Transformers

πŸ“ Summary:
RF-DETR is a light-weight detection transformer leveraging weight-sharing NAS to optimize accuracy-latency tradeoffs across diverse datasets. It significantly outperforms prior state-of-the-art, being the first real-time detector to surpass 60 AP on COCO.

πŸ”Ή Publication Date: Published on Nov 12

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.09554
β€’ PDF: https://arxiv.org/pdf/2511.09554

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#ObjectDetection #ComputerVision #MachineLearning #NeuralArchitectureSearch #Transformers
✨TiViBench: Benchmarking Think-in-Video Reasoning for Video Generative Models

πŸ“ Summary:
TiViBench is a new benchmark assessing image-to-video models reasoning across four dimensions and 24 tasks. Commercial models show stronger reasoning potential. VideoTPO, a test-time strategy, significantly enhances performance, advancing reasoning in video generation.

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13704
β€’ PDF: https://arxiv.org/pdf/2511.13704
β€’ Project Page: https://haroldchen19.github.io/TiViBench-Page/
β€’ Github: https://haroldchen19.github.io/TiViBench-Page/

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#VideoGeneration #AIBenchmark #ComputerVision #DeepLearning #AIResearch
This media is not supported in your browser
VIEW IN TELEGRAM
✨PhysX-Anything: Simulation-Ready Physical 3D Assets from Single Image

πŸ“ Summary:
PhysX-Anything generates simulation-ready physical 3D assets from single images, crucial for embodied AI. It uses a novel VLM-based model and an efficient 3D representation, enabling direct use in robotic policy learning.

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13648
β€’ PDF: https://arxiv.org/pdf/2511.13648
β€’ Project Page: https://physx-anything.github.io/
β€’ Github: https://github.com/ziangcao0312/PhysX-Anything

✨ Datasets citing this paper:
β€’ https://huggingface.co/datasets/Caoza/PhysX-Mobility

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#EmbodiedAI #3DReconstruction #Robotics #ComputerVision #AIResearch
✨Part-X-MLLM: Part-aware 3D Multimodal Large Language Model

πŸ“ Summary:
Part-X-MLLM is a 3D multimodal large language model that unifies diverse 3D tasks by generating structured programs from RGB point clouds and language prompts. It outputs part-level data and edit commands, enabling state-of-the-art 3D generation and editing through one interface.

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13647
β€’ PDF: https://arxiv.org/pdf/2511.13647
β€’ Project Page: https://chunshi.wang/Part-X-MLLM/

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#3D #MLLM #GenerativeAI #ComputerVision #AIResearch
✨Back to Basics: Let Denoising Generative Models Denoise

πŸ“ Summary:
Denoising diffusion models should predict clean images directly, not noise, leveraging the data manifold assumption. The paper introduces JiT, a model using simple, large-patch Transformers that achieves competitive generative results on ImageNet.

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13720
β€’ PDF: https://arxiv.org/pdf/2511.13720
β€’ Github: https://github.com/LTH14/JiT

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#DiffusionModels #GenerativeAI #DeepLearning #ComputerVision #AIResearch
❀1
✨UnSAMv2: Self-Supervised Learning Enables Segment Anything at Any Granularity

πŸ“ Summary:
UnSAMv2 enables continuous segmentation granularity control for the SAM model without human annotations. It uses self-supervised learning on unlabeled data to discover mask-granularity pairs and a novel control embedding. UnSAMv2 significantly enhances SAM-2s performance across various segmentati...

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13714
β€’ PDF: https://arxiv.org/pdf/2511.13714
β€’ Project Page: https://yujunwei04.github.io/UnSAMv2-Project-Page/
β€’ Github: https://github.com/yujunwei04/UnSAMv2

✨ Spaces citing this paper:
β€’ https://huggingface.co/spaces/yujunwei04/UnSAMv2

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#AI #ComputerVision #SelfSupervisedLearning #ImageSegmentation #DeepLearning
✨Error-Driven Scene Editing for 3D Grounding in Large Language Models

πŸ“ Summary:
DEER-3D improves 3D LLM grounding by iteratively editing and retraining models. It diagnoses predicate-level errors, then generates targeted 3D scene edits as counterfactuals to enhance spatial understanding and accuracy.

πŸ”Ή Publication Date: Published on Nov 18

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.14086
β€’ PDF: https://arxiv.org/pdf/2511.14086
β€’ Github: https://github.com/zhangyuejoslin/Deer-3D

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#LLMs #3DGrounding #ComputerVision #DeepLearning #AIResearch
✨Orion: A Unified Visual Agent for Multimodal Perception, Advanced Visual Reasoning and Execution

πŸ“ Summary:
Orion is a visual agent framework that orchestrates specialized computer vision tools to execute complex visual workflows. It achieves competitive performance on benchmarks and enables autonomous, tool-driven visual reasoning.

πŸ”Ή Publication Date: Published on Nov 18

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.14210
β€’ PDF: https://arxiv.org/pdf/2511.14210

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#ComputerVision #AIagents #VisualReasoning #MultimodalAI #DeepLearning
✨A Style is Worth One Code: Unlocking Code-to-Style Image Generation with Discrete Style Space

πŸ“ Summary:
CoTyle introduces code-to-style image generation, creating consistent visual styles from numerical codes. It is the first open-source academic method for this task, using a discrete style codebook and a text-to-image diffusion model for diverse, reproducible styles.

πŸ”Ή Publication Date: Published on Nov 13

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.10555
β€’ PDF: https://arxiv.org/pdf/2511.10555
β€’ Project Page: https://Kwai-Kolors.github.io/CoTyle/
β€’ Github: https://github.com/Kwai-Kolors/CoTyle

✨ Spaces citing this paper:
β€’ https://huggingface.co/spaces/Kwai-Kolors/CoTyle

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#ImageGeneration #DiffusionModels #NeuralStyle #ComputerVision #DeepLearning
✨MVI-Bench: A Comprehensive Benchmark for Evaluating Robustness to Misleading Visual Inputs in LVLMs

πŸ“ Summary:
MVI-Bench introduces a new benchmark to evaluate Large Vision-Language Models robustness against misleading visual inputs. It utilizes a hierarchical taxonomy and a novel metric to uncover significant vulnerabilities in state-of-the-art LVLMs.

πŸ”Ή Publication Date: Published on Nov 18

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.14159
β€’ PDF: https://arxiv.org/pdf/2511.14159
β€’ Github: https://github.com/chenyil6/MVI-Bench

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#LVLMs #ComputerVision #AIrobustness #MachineLearning #AI
✨REVISOR: Beyond Textual Reflection, Towards Multimodal Introspective Reasoning in Long-Form Video Understanding

πŸ“ Summary:
Text-only self-reflection is insufficient for long-form video understanding. REVISOR is a new framework enabling MLLMs to perform multimodal introspective reflection across text and visual modalities. This significantly enhances reasoning for long videos without extra fine-tuning, achieving stron...

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13026
β€’ PDF: https://arxiv.org/pdf/2511.13026

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#MultimodalAI #VideoUnderstanding #MLLMs #AIResearch #ComputerVision
✨Φeat: Physically-Grounded Feature Representation

πŸ“ Summary:
Ξ¦eat is a new self-supervised visual backbone that captures material identity like reflectance and mesostructure. It learns robust features invariant to external physical factors such as shape and lighting, promoting physics-aware perception.

πŸ”Ή Publication Date: Published on Nov 14

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.11270
β€’ PDF: https://arxiv.org/pdf/2511.11270

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#ComputerVision #SelfSupervisedLearning #DeepLearning #FeatureLearning #PhysicsAwareAI
✨VIDEOP2R: Video Understanding from Perception to Reasoning

πŸ“ Summary:
VideoP2R is a novel reinforcement fine-tuning framework for video understanding. It separately models perception and reasoning processes, using a new CoT dataset and a process-aware RL algorithm. This approach achieves state-of-the-art results on video reasoning benchmarks.

πŸ”Ή Publication Date: Published on Nov 14

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.11113v1
β€’ PDF: https://arxiv.org/pdf/2511.11113

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#VideoUnderstanding #ReinforcementLearning #AIResearch #ComputerVision #Reasoning
✨Reasoning via Video: The First Evaluation of Video Models' Reasoning Abilities through Maze-Solving Tasks

πŸ“ Summary:
VR-Bench evaluates video models' spatial reasoning using maze-solving tasks. It demonstrates that video models excel in spatial perception and reasoning, outperforming VLMs, and benefit from diverse sampling during inference. These findings show the strong potential of reasoning via video for spa...

πŸ”Ή Publication Date: Published on Nov 19

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.15065
β€’ PDF: https://arxiv.org/pdf/2511.15065
β€’ Project Page: https://imyangc7.github.io/VRBench_Web/
β€’ Github: https://github.com/ImYangC7/VR-Bench

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#VideoModels #AIReasoning #SpatialAI #ComputerVision #MachineLearning
❀1
✨ARC-Chapter: Structuring Hour-Long Videos into Navigable Chapters and Hierarchical Summaries

πŸ“ Summary:
ARC-Chapter is a large-scale video chaptering model trained on millions of long video chapters, using a new bilingual and hierarchical dataset. It introduces a novel evaluation metric, GRACE, to better reflect real-world chaptering. The model achieves state-of-the-art performance and demonstrates...

πŸ”Ή Publication Date: Published on Nov 18

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.14349
β€’ PDF: https://arxiv.org/pdf/2511.14349
β€’ Project Page: https://arcchapter.github.io/index_en.html
β€’ Github: https://github.com/TencentARC/ARC-Chapter

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#VideoChaptering #AI #MachineLearning #VideoSummarization #ComputerVision
✨Medal S: Spatio-Textual Prompt Model for Medical Segmentation

πŸ“ Summary:
Medal S is a medical segmentation foundation model using spatio-textual prompts for efficient, high-accuracy multi-class segmentation across diverse modalities. It uniquely aligns volumetric prompts with text embeddings and processes masks in parallel, significantly outperforming prior methods.

πŸ”Ή Publication Date: Published on Nov 17

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2511.13001
β€’ PDF: https://arxiv.org/pdf/2511.13001
β€’ Github: https://github.com/yinghemedical/Medal-S

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/spc819/Medal-S-V1.0

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#MedicalSegmentation #FoundationModels #AI #DeepLearning #ComputerVision
✨OmniParser for Pure Vision Based GUI Agent

πŸ“ Summary:
OmniParser enhances GPT-4V's ability to act as a GUI agent by improving screen parsing. It identifies interactable icons and understands element semantics using specialized models. This significantly boosts GPT-4V's performance on benchmarks like ScreenSpot, Mind2Web, and AITW.

πŸ”Ή Publication Date: Published on Aug 1, 2024

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2408.00203
β€’ PDF: https://arxiv.org/pdf/2408.00203
β€’ Github: https://github.com/microsoft/omniparser

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/microsoft/OmniParser
β€’ https://huggingface.co/microsoft/OmniParser-v2.0
β€’ https://huggingface.co/banao-tech/OmniParser

✨ Datasets citing this paper:
β€’ https://huggingface.co/datasets/mlfoundations/Click-100k

✨ Spaces citing this paper:
β€’ https://huggingface.co/spaces/callmeumer/OmniParser-v2
β€’ https://huggingface.co/spaces/nofl/OmniParser-v2
β€’ https://huggingface.co/spaces/SheldonLe/OmniParser-v2

==================================

For more data science resources:
βœ“ https://t.iss.one/DataScienceT

#GUIagents #ComputerVision #GPT4V #AIagents #DeepLearning