Data Science | Machine Learning with Python for Researchers
32.5K subscribers
3.17K photos
112 videos
23 files
3.38K links
ads: @HusseinSheikho

The Data Science and Python channel is for researchers and advanced programmers

Buy ads: https://telega.io/c/dataScienceT
Download Telegram
5 minutes of work - 127,000$ profit!

Opened access to the Jay Welcome Club where the AI bot does all the work itself💻

Usually you pay crazy money to get into this club, but today access is free for everyone!

23,432% on deposit earned by club members in the last 6 months📈

Just follow Jay's trades and earn! 👇

https://t.iss.one/+mONXtEgVxtU5NmZl
🔹 Title:
Offline Reinforcement Learning from Datasets with Structured Non-Stationarity

🔹 Publication Date: Published on May 23, 2024

🔹 Abstract:
A novel Offline RL method uses Contrastive Predictive Coding to handle non-stationary transition and reward functions in datasets, outperforming baselines in various control tasks. AI-generated summary Current Reinforcement Learning (RL) is often limited by the large amount of data needed to learn a successful policy. Offline RL aims to solve this issue by using transitions collected by a different behavior policy. We address a novel Offline RL problem setting in which, while collecting the dataset, the transition and reward functions gradually change between episodes but stay constant within each episode. We propose a method based on Contrastive Predictive Coding that identifies this non-stationarity in the offline dataset, accounts for it when training a policy, and predicts it during evaluation. We analyze our proposed method and show that it performs well in simple continuous control tasks and challenging, high-dimensional locomotion tasks . We show that our method often achieves the oracle performance and performs better than baselines.

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2405.14114
• PDF: https://arxiv.org/pdf/2405.14114

🔹 Datasets citing this paper:
https://huggingface.co/datasets/johannesack/OfflineRLStructuredNonstationary

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
🔹 Title: MoHoBench: Assessing Honesty of Multimodal Large Language Models via Unanswerable Visual Questions

🔹 Publication Date:
Published on Jul 29

🔹 Abstract:
A systematic assessment of honesty in Multimodal Large Language Models (MLLMs) using a large-scale benchmark reveals that models often fail to appropriately refuse unanswerable visual questions, highlighting the need for multimodal honesty alignment methods. AI-generated summary Recently Multimodal Large Language Models ( MLLMs ) have achieved considerable advancements in vision-language tasks, yet produce potentially harmful or untrustworthy content. Despite substantial work investigating the trustworthiness of language models, MMLMs' capability to act honestly, especially when faced with visually unanswerable questions, remains largely underexplored. This work presents the first systematic assessment of honesty behaviors across various MLLMs . We ground honesty in models' response behaviors to unanswerable visual questions, define four representative types of such questions, and construct MoHoBench , a large-scale MMLM honest benchmark, consisting of 12k+ visual question samples , whose quality is guaranteed by multi-stage filtering and human verification. Using MoHoBench , we benchmarked the honesty of 28 popular MMLMs and conducted a comprehensive analysis. Our findings show that: (1) most models fail to appropriately refuse to answer when necessary, and (2) MMLMs' honesty is not solely a language modeling issue, but is deeply influenced by visual information, necessitating the development of dedicated methods for multimodal honesty alignment. Therefore, we implemented initial alignment methods using supervised and preference learning to improve honesty behavior, providing a foundation for future work on trustworthy MLLMs . Our data and code can be found at https://github.com/DSTTSD/ MoHoBench .

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2507.21503
• PDF: https://arxiv.org/pdf/2507.21503
• Github: https://github.com/DSTTSD/MoHoBench

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
1
🔹 Title: X-Omni: Reinforcement Learning Makes Discrete Autoregressive Image Generative Models Great Again

🔹 Publication Date:
Published on Jul 29

🔹 Abstract:
Reinforcement learning enhances discrete autoregressive modeling for image and language generation, achieving high-quality image generation and instruction-following capabilities. AI-generated summary Numerous efforts have been made to extend the ``next token prediction'' paradigm to visual contents, aiming to create a unified approach for both image generation and understanding. Nevertheless, attempts to generate images through autoregressive modeling with discrete tokens have been plagued by issues such as low visual fidelity , distorted outputs , and failure to adhere to complex instructions when rendering intricate details. These shortcomings are likely attributed to cumulative errors during autoregressive inference or information loss incurred during the discretization process. Probably due to this challenge, recent research has increasingly shifted toward jointly training image generation with diffusion objectives and language generation with autoregressive objectives, moving away from unified modeling approaches. In this work, we demonstrate that reinforcement learning can effectively mitigate artifacts and largely enhance the generation quality of a discrete autoregressive modeling method, thereby enabling seamless integration of image and language generation . Our framework comprises a semantic image tokenizer , a unified autoregressive model for both language and images, and an offline diffusion decoder for image generation , termed X-Omni . X-Omni achieves state-of-the-art performance in image generation tasks using a 7B language model, producing images with high aesthetic quality while exhibiting strong capabilities in following instructions and rendering long texts.

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2507.22058
• PDF: https://arxiv.org/pdf/2507.22058
• Project Page: https://x-omni-team.github.io
• Github: https://github.com/X-Omni-Team/X-Omni

🔹 Spaces citing this paper:
https://huggingface.co/spaces/zhangxiaosong18/X-Omni-En
https://huggingface.co/spaces/zhangxiaosong18/X-Omni-Zh
==================================

For more data science resources:
https://t.iss.one/DataScienceT
2
🔹 Title:
FedRand: Enhancing Privacy in Federated Learning with Randomized LoRA Subparameter Updates

🔹 Publication Date: Published on Mar 10

🔹 Abstract:
FedRand framework enhances data privacy in federated learning by keeping a subset of LoRA parameters private, reducing the risk of membership inference attacks while maintaining model accuracy. AI-generated summary Federated Learning (FL) is a widely used framework for training models in a decentralized manner, ensuring that the central server does not have direct access to data from local clients. However, this approach may still fail to fully preserve data privacy, as models from local clients are exposed to the central server during the aggregation process. This issue becomes even more critical when training vision-language models (VLMs) with FL, as VLMs can easily memorize training data instances, making them vulnerable to membership inference attacks (MIAs). To address this challenge, we propose the FedRand framework, which avoids disclosing the full set of client parameters. In this framework, each client randomly selects subparameters of Low-Rank Adaptation ( LoRA ) from the server and keeps the remaining counterparts of the LoRA weights as private parameters. After training both parameters on the client's private dataset, only the non-private client parameters are sent back to the server for aggregation. This approach mitigates the risk of exposing client-side VLM parameters, thereby enhancing data privacy. We empirically validate that FedRand improves robustness against MIAs compared to relevant baselines while achieving accuracy comparable to methods that communicate full LoRA parameters across several benchmark datasets.

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2503.07216
• PDF: https://arxiv.org/pdf/2503.07216

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
2
🔹 Title: CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning

🔹 Publication Date: Published on Jul 18

🔹 Abstract: CUDA-L1, an automated reinforcement learning framework, significantly improves CUDA optimization across various GPU architectures, achieving substantial speedups without human expertise. AI-generated summary The exponential growth in demand for GPU computing resources, driven by the rapid advancement of Large Language Models, has created an urgent need for automated CUDA optimization strategies. While recent advances in LLMs show promise for code generation, current SOTA models (e.g. R1, o1) achieve low success rates in improving CUDA speed. In this paper, we introduce CUDA-L1 , an automated reinforcement learning framework for CUDA optimization . CUDA-L1 achieves performance improvements on the CUDA optimization task: trained on NVIDIA A100 , it delivers an average speedup of x17.7 across all 250 CUDA kernels of KernelBench , with peak speedup s reaching x449. Furthermore, the model also demonstrates excellent portability across GPU architectures , achieving average speedup s of x17.8 on H100 , x19.0 on RTX 3090 , x16.5 on L40 , x14.7 on H800 , and x13.9 on H20 despite being optimized specifically for A100. Beyond these benchmark results, CUDA-L1 demonstrates several remarkable properties: 1) Discovers a variety of CUDA optimization techniques and learns to combine them strategically to achieve optimal performance; 2) Uncovers fundamental principles of CUDA optimization ; 3) Identifies non-obvious performance bottlenecks and rejects seemingly beneficial optimizations that harm performance. The capabilities of CUDA-L1 demonstrate that reinforcement learning can transform an initially poor-performing LLM into an effective CUDA optimizer through speedup -based reward signals alone, without human expertise or domain knowledge. More importantly, the trained RL model extend the acquired reasoning abilities to new kernels. This paradigm opens possibilities for automated optimization of CUDA operations, and holds promise to substantially promote GPU efficiency and alleviate the rising pressure on GPU computing resources.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.14111

• PDF: https://arxiv.org/pdf/2507.14111

🔹 Datasets citing this paper:
https://huggingface.co/datasets/deepreinforce-ai/CUDA-L1

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
3
Join our WhatsApp channel

There are dedicated resources only for WhatsApp users

https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
1
🚀 Become an Agentic AI Builder — Free 12‑Week Certification by Ready Tensor

Ready Tensor’s Agentic AI Developer Certification is a free, project first 12‑week program designed to help you build and deploy real-world agentic AI systems. You'll complete three portfolio-ready projects using tools like LangChain, LangGraph, and vector databases, while deploying production-ready agents with FastAPI or Streamlit.

The course focuses on developing autonomous AI agents that can plan, reason, use memory, and act safely in complex environments. Certification is earned not by watching lectures, but by building — each project is reviewed against rigorous standards.

You can start anytime, and new cohorts begin monthly. Ideal for developers and engineers ready to go beyond chat prompts and start building true agentic systems.

👉 Apply now: https://www.readytensor.ai/agentic-ai-cert/
🔹 Title:
Music Arena: Live Evaluation for Text-to-Music

🔹 Publication Date: Published on Jul 28

🔹 Abstract:
Music Arena provides a scalable, interactive platform for evaluating text-to-music models through user-generated preferences and detailed feedback. AI-generated summary We present Music Arena, an open platform for scalable human preference evaluation of text-to-music (TTM) models. Soliciting human preferences via listening studies is the gold standard for evaluation in TTM, but these studies are expensive to conduct and difficult to compare, as study protocols may differ across systems. Moreover, human preferences might help researchers align their TTM systems or improve automatic evaluation metrics, but an open and renewable source of preferences does not currently exist. We aim to fill these gaps by offering *live* evaluation for TTM. In Music Arena, real-world users input text prompts of their choosing and compare outputs from two TTM systems, and their preferences are used to compile a leaderboard. While Music Arena follows recent evaluation trends in other AI domains, we also design it with key features tailored to music: an LLM-based routing system to navigate the heterogeneous type signatures of TTM systems, and the collection of *detailed* preferences including listening data and natural language feedback . We also propose a rolling data release policy with user privacy guarantees , providing a renewable source of preference data and increasing platform transparency. Through its standardized evaluation protocol , transparent data access policies , and music-specific features, Music Arena not only addresses key challenges in the TTM ecosystem but also demonstrates how live evaluation can be thoughtfully adapted to unique characteristics of specific AI domains. Music Arena is available at: https://music-arena.org

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2507.20900
• PDF: https://arxiv.org/pdf/2507.20900
• Github: https://github.com/gclef-cmu/music-arena

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
2
🔹 Title:
Privacy-Aware Energy Consumption Modeling of Connected Battery Electric Vehicles using Federated Learning

🔹 Publication Date: Published on Dec 12, 2023

🔹 Abstract:
Federated Learning methods like FedAvg and FedPer improve BEV energy consumption prediction while protecting user privacy. AI-generated summary Battery Electric Vehicles (BEVs) are increasingly significant in modern cities due to their potential to reduce air pollution. Precise and real-time estimation of energy consumption for them is imperative for effective itinerary planning and optimizing vehicle systems, which can reduce driving range anxiety and decrease energy costs. As public awareness of data privacy increases, adopting approaches that safeguard data privacy in the context of BEV energy consumption modeling is crucial. Federated Learning ( FL ) is a promising solution mitigating the risk of exposing sensitive information to third parties by allowing local data to remain on devices and only sharing model updates with a central server. Our work investigates the potential of using FL methods, such as FedAvg , and FedPer , to improve BEV energy consumption prediction while maintaining user privacy. We conducted experiments using data from 10 BEVs under simulated real-world driving conditions. Our results demonstrate that the FedAvg - LSTM model achieved a reduction of up to 67.84\% in the MAE value of the prediction results. Furthermore, we explored various real-world scenarios and discussed how FL methods can be employed in those cases. Our findings show that FL methods can effectively improve the performance of BEV energy consumption prediction while maintaining user privacy.

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2312.07371
• PDF: https://arxiv.org/pdf/2312.07371

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
2
🔹 Title: Seed-Prover: Deep and Broad Reasoning for Automated Theorem Proving

🔹 Publication Date: Published on Jul 31

🔹 Abstract: Seed-Prover, a lemma-style reasoning model using Lean, achieves high performance in formal theorem proving and automated mathematical reasoning through iterative refinement and specialized geometry support. AI-generated summary LLMs have demonstrated strong mathematical reasoning abilities by leveraging reinforcement learning with long chain-of-thought , yet they continue to struggle with theorem proving due to the lack of clear supervision signals when solely using natural language. Dedicated domain-specific languages like Lean provide clear supervision via formal verification of proofs, enabling effective training through reinforcement learning . In this work, we propose Seed-Prover , a lemma-style whole-proof reasoning model. Seed-Prover can iteratively refine its proof based on Lean feedback, proved lemmas, and self-summarization. To solve IMO-level contest problems, we design three test-time inference strategies that enable both deep and broad reasoning. Seed-Prover proves 78.1% of formalized past IMO problems, saturates MiniF2F , and achieves over 50\% on PutnamBench , outperforming the previous state-of-the-art by a large margin. To address the lack of geometry support in Lean, we introduce a geometry reasoning engine Seed-Geometry , which outperforms previous formal geometry engines. We use these two systems to participate in IMO 2025 and fully prove 5 out of 6 problems. This work represents a significant advancement in automated mathematical reasoning , demonstrating the effectiveness of formal verification with long chain-of-thought reasoning.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.23726

• PDF: https://arxiv.org/pdf/2507.23726

• Github: https://github.com/ByteDance-Seed/Seed-Prover

🔹 Models citing this paper:
No models found

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
2
🔹 Title: RecGPT Technical Report

🔹 Publication Date: Published on Jul 30

🔹 Abstract: RecGPT integrates large language models into recommender systems to focus on user intent, improving content diversity and satisfaction while enhancing merchant and platform performance. AI-generated summary Recommender systems are among the most impactful applications of artificial intelligence, serving as critical infrastructure connecting users, merchants, and platforms. However, most current industrial systems remain heavily reliant on historical co-occurrence patterns and log-fitting objectives, i.e., optimizing for past user interactions without explicitly modeling user intent . This log-fitting approach often leads to overfitting to narrow historical preferences, failing to capture users' evolving and latent interests. As a result, it reinforces filter bubbles and long-tail phenomena, ultimately harming user experience and threatening the sustainability of the whole recommendation ecosystem. To address these challenges, we rethink the overall design paradigm of recommender systems and propose RecGPT, a next-generation framework that places user intent at the center of the recommendation pipeline. By integrating large language models ( LLMs ) into key stages of user interest mining , item retrieval , and explanation generation , RecGPT transforms log-fitting recommendation into an intent-centric process. To effectively align general-purpose LLMs to the above domain-specific recommendation tasks at scale, RecGPT incorporates a multi-stage training paradigm, which integrates reasoning-enhanced pre-alignment and self-training evolution , guided by a Human-LLM cooperative judge system. Currently, RecGPT has been fully deployed on the Taobao App. Online experiments demonstrate that RecGPT achieves consistent performance gains across stakeholders: users benefit from increased content diversity and satisfaction, merchants and the platform gain greater exposure and conversions. These comprehensive improvement results across all stakeholders validates that LLM-driven, intent-centric design can foster a more sustainable and mutually beneficial recommendation ecosystem.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.22879

• PDF: https://arxiv.org/pdf/2507.22879

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
2
🔹 Title: Beyond Linear Bottlenecks: Spline-Based Knowledge Distillation for Culturally Diverse Art Style Classification

🔹 Publication Date: Published on Jul 31

🔹 Abstract: Enhancing dual-teacher self-supervised frameworks with Kolmogorov-Arnold Networks improves art style classification by better modeling nonlinear feature correlations and disentangling complex style manifolds. AI-generated summary Art style classification remains a formidable challenge in computational aesthetics due to the scarcity of expertly labeled datasets and the intricate, often nonlinear interplay of stylistic elements. While recent dual-teacher self-supervised frameworks reduce reliance on labeled data, their linear projection layers and localized focus struggle to model global compositional context and complex style-feature interactions. We enhance the dual-teacher knowledge distillation framework to address these limitations by replacing conventional MLP projection and prediction heads with Kolmogorov-Arnold Networks ( KANs ). Our approach retains complementary guidance from two teacher networks, one emphasizing localized texture and brushstroke patterns, the other capturing broader stylistic hierarchies while leveraging KANs ' spline-based activations to model nonlinear feature correlations with mathematical precision. Experiments on WikiArt and Pandora18k demonstrate that our approach outperforms the base dual teacher architecture in Top-1 accuracy. Our findings highlight the importance of KANs in disentangling complex style manifolds , leading to better linear probe accuracy than MLP projection s.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.23436

• PDF: https://arxiv.org/pdf/2507.23436

• Project Page: https://huggingface.co/papers?q=MLP%20projection

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
2
🔹 Title:
4DSloMo: 4D Reconstruction for High Speed Scene with Asynchronous Capture

🔹 Publication Date: Published on Jul 7

🔹 Abstract:
A high-speed 4D capturing system using low FPS cameras with asynchronous capture and video-diffusion-based artifact correction enhances reconstruction quality. AI-generated summary Reconstructing fast-dynamic scenes from multi-view videos is crucial for high-speed motion analysis and realistic 4D reconstruction. However, the majority of 4D capture systems are limited to frame rates below 30 FPS (frames per second), and a direct 4D reconstruction of high-speed motion from low FPS input may lead to undesirable results. In this work, we propose a high-speed 4D capturing system only using low FPS cameras, through novel capturing and processing modules. On the capturing side, we propose an asynchronous capture scheme that increases the effective frame rate by staggering the start times of cameras. By grouping cameras and leveraging a base frame rate of 25 FPS, our method achieves an equivalent frame rate of 100-200 FPS without requiring specialized high-speed cameras. On processing side, we also propose a novel generative model to fix artifacts caused by 4D sparse-view reconstruction, as asynchrony reduces the number of viewpoints at each timestamp. Specifically, we propose to train a video-diffusion-based artifact-fix model for sparse 4D reconstruction, which refines missing details, maintains temporal consistency , and improves overall reconstruction quality. Experimental results demonstrate that our method significantly enhances high-speed 4D reconstruction compared to synchronous capture.

🔹 Links:
• arXiv Page: https://arxivexplained.com/papers/4dslomo-4d-reconstruction-for-high-speed-scene-with-asynchronous-capture
• PDF: https://arxiv.org/pdf/2507.05163
• Github: https://openimaginglab.github.io/4DSloMo/

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
5🔥1
This is a channel for a job opportunity
🔹 Title:
Using Degeneracy in the Loss Landscape for Mechanistic Interpretability

🔹 Publication Date: Published on May 17, 2024

🔹 Abstract:
Mechanistic interpretability of neural networks is enhanced by a new technique that identifies and exploits degenerate parameters through the Interaction Basis, leading to sparser and more interpretable network representations. AI-generated summary Mechanistic Interpretability aims to reverse engineer the algorithms implemented by neural networks by studying their weights and activations . An obstacle to reverse engineering neural networks is that many of the parameters inside a network are not involved in the computation being implemented by the network. These degenerate parameters may obfuscate internal structure. Singular learning theory teaches us that neural network parameterizations are biased towards being more degenerate, and parameterizations with more degeneracy are likely to generalize further. We identify 3 ways that network parameters can be degenerate: linear dependence between activations in a layer; linear dependence between gradients passed back to a layer; ReLUs which fire on the same subset of datapoints. We also present a heuristic argument that modular networks are likely to be more degenerate, and we develop a metric for identifying modules in a network that is based on this argument. We propose that if we can represent a neural network in a way that is invariant to reparameterizations that exploit the degeneracies, then this representation is likely to be more interpretable, and we provide some evidence that such a representation is likely to have sparser interactions. We introduce the Interaction Basis , a tractable technique to obtain a representation that is invariant to degeneracies from linear dependence of activations or Jacobians .

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2405.10927
• PDF: https://arxiv.org/pdf/2405.10927

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
1
🔹 Title:
Diffuman4D: 4D Consistent Human View Synthesis from Sparse-View Videos with Spatio-Temporal Diffusion Models

🔹 Publication Date: Published on Jul 17

🔹 Abstract:
A sliding iterative denoising process is proposed to enhance spatio-temporal consistency in 4D diffusion models for high-fidelity view synthesis from sparse-view videos. AI-generated summary This paper addresses the challenge of high-fidelity view synthesis of humans with sparse-view videos as input. Previous methods solve the issue of insufficient observation by leveraging 4D diffusion models to generate videos at novel viewpoints. However, the generated videos from these models often lack spatio-temporal consistency , thus degrading view synthesis quality. In this paper, we propose a novel sliding iterative denoising process to enhance the spatio-temporal consistency of the 4D diffusion model. Specifically, we define a latent grid in which each latent encodes the image , camera pose , and human pose for a certain viewpoint and timestamp, then alternately denoising the latent grid along spatial and temporal dimensions with a sliding window, and finally decode the videos at target viewpoints from the corresponding denoised latents. Through the iterative sliding, information flows sufficiently across the latent grid , allowing the diffusion model to obtain a large receptive field and thus enhance the 4D consistency of the output, while making the GPU memory consumption affordable. The experiments on the DNA-Rendering and ActorsHQ datasets demonstrate that our method is able to synthesize high-quality and consistent novel-view videos and significantly outperforms the existing approaches. See our project page for interactive demos and video results: https://diffuman4d.github.io/ .

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2507.13344
• PDF: https://arxiv.org/pdf/2507.13344
• Project Page: https://diffuman4d.github.io/
• Github: https://github.com/zju3dv/Diffuman4D

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
1
🔹 Title:
ForCenNet: Foreground-Centric Network for Document Image Rectification

🔹 Publication Date: Published on Jul 26

🔹 Abstract:
A Foreground-Centric Network for document image rectification improves state-of-the-art by effectively handling foreground elements and layout distortions. AI-generated summary Document image rectification aims to eliminate geometric deformation in photographed documents to facilitate text recognition. However, existing methods often neglect the significance of foreground elements, which provide essential geometric references and layout information for document image correction. In this paper, we introduce Foreground-Centric Network ( ForCenNet ) to eliminate geometric distortions in document images. Specifically, we initially propose a foreground-centric label generation method, which extracts detailed foreground elements from an undistorted image. Then we introduce a foreground-centric mask mechanism to enhance the distinction between readable and background regions. Furthermore, we design a curvature consistency loss to leverage the detailed foreground labels to help the model understand the distorted geometric distribution. Extensive experiments demonstrate that ForCenNet achieves new state-of-the-art on four real-world benchmarks, such as DocUNet, DIR300, WarpDoc, and DocReal. Quantitative analysis shows that the proposed method effectively undistorts layout elements, such as text lines and table borders. The resources for further comparison are provided at https://github.com/caipeng328/ ForCenNet .

🔹 Links:
• arXiv Page: https://arxiv.org/abs/2507.19804
• PDF: https://arxiv.org/pdf/2507.19804
• Github: https://github.com/caipeng328/ForCenNet

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:

https://t.iss.one/DataScienceT
1
🔹 Title: Seed-Prover: Deep and Broad Reasoning for Automated Theorem Proving

🔹 Publication Date: Published on Jul 31

🔹 Abstract: Seed-Prover, a lemma-style reasoning model using Lean, achieves high performance in formal theorem proving and automated mathematical reasoning through iterative refinement and specialized geometry support. AI-generated summary LLMs have demonstrated strong mathematical reasoning abilities by leveraging reinforcement learning with long chain-of-thought , yet they continue to struggle with theorem proving due to the lack of clear supervision signals when solely using natural language. Dedicated domain-specific languages like Lean provide clear supervision via formal verification of proofs, enabling effective training through reinforcement learning . In this work, we propose Seed-Prover , a lemma-style whole-proof reasoning model. Seed-Prover can iteratively refine its proof based on Lean feedback, proved lemmas, and self-summarization. To solve IMO-level contest problems, we design three test-time inference strategies that enable both deep and broad reasoning. Seed-Prover proves 78.1% of formalized past IMO problems, saturates MiniF2F , and achieves over 50\% on PutnamBench , outperforming the previous state-of-the-art by a large margin. To address the lack of geometry support in Lean, we introduce a geometry reasoning engine Seed-Geometry , which outperforms previous formal geometry engines. We use these two systems to participate in IMO 2025 and fully prove 5 out of 6 problems. This work represents a significant advancement in automated mathematical reasoning , demonstrating the effectiveness of formal verification with long chain-of-thought reasoning.

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.23726

• PDF: https://arxiv.org/pdf/2507.23726

• Github: https://github.com/ByteDance-Seed/Seed-Prover

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
3
🔹 Title: Phi-Ground Tech Report: Advancing Perception in GUI Grounding

🔹 Publication Date: Published on Jul 31

🔹 Abstract: The Phi-Ground model family achieves state-of-the-art performance in GUI grounding for multimodal reasoning models, improving accuracy across various benchmarks. AI-generated summary With the development of multimodal reasoning models , Computer Use Agents (CUAs), akin to Jarvis from "Iron Man", are becoming a reality. GUI grounding is a core component for CUAs to execute actual actions, similar to mechanical control in robotics, and it directly leads to the success or failure of the system. It determines actions such as clicking and typing, as well as related parameters like the coordinates for clicks. Current end-to-end grounding models still achieve less than 65\% accuracy on challenging benchmarks like ScreenSpot-pro and UI-Vision , indicating they are far from being ready for deployment. % , as a single misclick can result in unacceptable consequences. In this work, we conduct an empirical study on the training of grounding models, examining details from data collection to model training. Ultimately, we developed the Phi-Ground model family , which achieves state-of-the-art performance across all five grounding benchmarks for models under 10B parameters in agent settings. In the end-to-end model setting, our model still achieves SOTA results with scores of \textbf{43.2} on ScreenSpot-pro and \textbf{27.2} on UI-Vision . We believe that the various details discussed in this paper, along with our successes and failures, not only clarify the construction of grounding models but also benefit other perception tasks. Project homepage: https://zhangmiaosen2000.github.io/Phi-Ground/{https://zhangmiaosen2000.github.io/Phi-Ground/}

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2507.23779

• PDF: https://arxiv.org/pdf/2507.23779

• Project Page: https://zhangmiaosen2000.github.io/Phi-Ground/

• Github: https://github.com/zhangmiaosen2000/Phi-Ground

🔹 Datasets citing this paper:
No datasets found

🔹 Spaces citing this paper:
No spaces found
==================================

For more data science resources:
https://t.iss.one/DataScienceT
2