Data Science | Machine Learning with Python for Researchers
31.5K subscribers
1.6K photos
102 videos
22 files
1.87K links
Admin: @HusseinSheikho

The Data Science and Python channel is for researchers and advanced programmers

Buy ads: https://telega.io/c/dataScienceT
Download Telegram
This channels is for Programmers, Coders, Software Engineers.

0️⃣ Python
1️⃣ Data Science
2️⃣ Machine Learning
3️⃣ Data Visualization
4️⃣ Artificial Intelligence
5️⃣ Data Analysis
6️⃣ Statistics
7️⃣ Deep Learning
8️⃣ programming Languages

βœ… https://t.iss.one/addlist/8_rRW2scgfRhOTc0

βœ… https://t.iss.one/Codeprogrammer
Please open Telegram to view this post
VIEW IN TELEGRAM
❀3
Article Title:
Uni3C: Unifying Precisely 3D-Enhanced Camera and Human Motion Controls for Video Generation

Article Date: 21 Apr 2025

Article Description:
Camera and human motion controls have been extensively studied for video generation, but existing approaches typically address them separately, suffering from limited data with high-quality annotations for both aspects. To overcome this, we present Uni3C, a unified 3D-enhanced framework for precise control of both camera and human motion in video generation. Uni3C includes two key contributions. First, we propose a plug-and-play control module trained with a frozen video generative backbone, PCDController, which utilizes unprojected point clouds from monocular depth to achieve accurate camera control. By leveraging the strong 3D priors of point clouds and the powerful capacities of video foundational models, PCDController shows impressive generalization, performing well regardless of whether the inference backbone is frozen or fine-tuned. This flexibility enables different modules of Uni3C to be trained in specific domains, i.e., either camera control or human motion control, reducing the dependency on jointly annotated data. Second, we propose a jointly aligned 3D world guidance for the inference phase that seamlessly integrates both scenic point clouds and SMPL-X characters to unify the control signals for camera and human motion, respectively. Extensive experiments confirm that PCDController enjoys strong robustness in driving camera motion for fine-tuned backbones of video generation. Uni3C substantially outperforms competitors in both camera controllability and human motion quality. Additionally, we collect tailored validation sets featuring challenging camera movements and human actions to validate the effectiveness of our method.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2504.14899v1.pdf

GitHub:
β€’ https://github.com/ewrfcas/uni3c

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀7
30k 🌟
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘3❀1
If anyone wants you to donate stars, please donate by gifts, it will be better for us.

There are gifts at a price of 10 stars
❀3
Article Title:
Curie: Toward Rigorous and Automated Scientific Experimentation with AI Agents

Article Date: 22 Feb 2025

Article Description:
Scientific experimentation, a cornerstone of human progress, demands rigor in reliability, methodical control, and interpretability to yield meaningful results. Despite the growing capabilities of large language models (LLMs) in automating different aspects of the scientific process, automating rigorous experimentation remains a significant challenge. To address this gap, we propose Curie, an AI agent framework designed to embed rigor into the experimentation process through three key components: an intra-agent rigor module to enhance reliability, an inter-agent rigor module to maintain methodical control, and an experiment knowledge module to enhance interpretability. To evaluate Curie, we design a novel experimental benchmark composed of 46 questions across four computer science domains, derived from influential research papers, and widely adopted open-source projects. Compared to the strongest baseline tested, we achieve a 3.4$\times$ improvement in correctly answering experimental questions.Curie is open-sourced at https://github.com/Just-Curieous/Curie.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2502.16069v1.pdf

GitHub:
β€’ https://github.com/just-curieous/curie

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀4πŸ‘1πŸŽ‰1
Article Title:
Token Reduction Should Go Beyond Efficiency in Generative Models -- From Vision, Language to Multimodality

Article Date: 23 May 2025

Article Description:
In Transformer architectures, tokens\textemdash discrete units derived from raw data\textemdash are formed by segmenting inputs into fixed-length chunks. Each token is then mapped to an embedding, enabling parallel attention computations while preserving the input's essential information. Due to the quadratic computational complexity of transformer self-attention mechanisms, token reduction has primarily been used as an efficiency strategy. This is especially true in single vision and language domains, where it helps balance computational costs, memory usage, and inference latency. Despite these advances, this paper argues that token reduction should transcend its traditional efficiency-oriented role in the era of large generative models. Instead, we position it as a fundamental principle in generative modeling, critically influencing both model architecture and broader applications. Specifically, we contend that across vision, language, and multimodal systems, token reduction can: (i) facilitate deeper multimodal integration and alignment, (ii) mitigate "overthinking" and hallucinations, (iii) maintain coherence over long inputs, and (iv) enhance training stability, etc. We reframe token reduction as more than an efficiency measure. By doing so, we outline promising future directions, including algorithm design, reinforcement learning-guided token reduction, token optimization for in-context learning, and broader ML and scientific domains. We highlight its potential to drive new model architectures and learning strategies that improve robustness, increase interpretability, and better align with the objectives of generative modeling.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2505.18227v1.pdf

GitHub:
β€’ https://github.com/zlkong/awesome-token-compression-reduction

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀5πŸ‘1
Article Title:
Darwin Godel Machine: Open-Ended Evolution of Self-Improving Agents

Article Date: 29 May 2025

Article Description:
Today's AI systems have human-designed, fixed architectures and cannot autonomously and continuously improve themselves. The advance of AI could itself be automated. If done safely, that would accelerate AI development and allow us to reap its benefits much sooner. Meta-learning can automate the discovery of novel algorithms, but is limited by first-order improvements and the human design of a suitable search space. The G\"odel machine proposed a theoretical alternative: a self-improving AI that repeatedly modifies itself in a provably beneficial manner. Unfortunately, proving that most changes are net beneficial is impossible in practice. We introduce the Darwin G\"odel Machine (DGM), a self-improving system that iteratively modifies its own code (thereby also improving its ability to modify its own codebase) and empirically validates each change using coding benchmarks. Inspired by Darwinian evolution and open-endedness research, the DGM maintains an archive of generated coding agents. It grows the archive by sampling an agent from it and using a foundation model to create a new, interesting, version of the sampled agent. This open-ended exploration forms a growing tree of diverse, high-quality agents and allows the parallel exploration of many different paths through the search space. Empirically, the DGM automatically improves its coding capabilities (e.g., better code editing tools, long-context window management, peer-review mechanisms), increasing performance on SWE-bench from 20.0% to 50.0%, and on Polyglot from 14.2% to 30.7%. Furthermore, the DGM significantly outperforms baselines without self-improvement or open-ended exploration. All experiments were done with safety precautions (e.g., sandboxing, human oversight). The DGM is a significant step toward self-improving AI, capable of gathering its own stepping stones along paths that unfold into endless innovation.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2505.22954v1.pdf

GitHub:
β€’ https://github.com/jennyzzt/dgm

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀4
Article Title:
AReaL: A Large-Scale Asynchronous Reinforcement Learning System for Language Reasoning

Article Date: 30 May 2025

Article Description:
Reinforcement learning (RL) has become a dominant paradigm for training large language models (LLMs), particularly for reasoning tasks. Effective RL for LLMs requires massive parallelization and poses an urgent need for efficient training systems. Most existing large-scale RL systems for LLMs are synchronous, alternating generation and training in a batch setting where rollouts in each training batch are generated by the same model. This approach stabilizes RL training but suffers from severe system-level inefficiency: generation must wait until the longest output in the batch is completed before model updates, resulting in GPU underutilization. We present AReaL, a fully asynchronous RL system that completely decouples generation from training. Rollout workers in AReaL continuously generate new outputs without waiting, while training workers update the model whenever a batch of data is collected. AReaL also incorporates a collection of system-level optimizations, leading to substantially higher GPU utilization. To stabilize RL training, AReaL balances the workload of rollout and training workers to control data staleness, and adopts a staleness-enhanced PPO variant to better handle outdated training samples. Extensive experiments on math and code reasoning benchmarks show that AReaL achieves up to 2.77$\times$ training speedup compared to synchronous systems with the same number of GPUs and matched or improved final performance. The code of AReaL is available at https://github.com/inclusionAI/AReaL/.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2505.24298v2.pdf

GitHub:
β€’ https://github.com/inclusionai/areal

Datasets:
β€’ MATH
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀4πŸ™1
Article Title:
AutoAgent: A Fully-Automated and Zero-Code Framework for LLM Agents

Article Date: 9 Feb 2025

Article Description:
Large Language Model (LLM) Agents have demonstrated remarkable capabilities in task automation and intelligent decision-making, driving the widespread adoption of agent development frameworks such as LangChain and AutoGen. However, these frameworks predominantly serve developers with extensive technical expertise - a significant limitation considering that only 0.03 % of the global population possesses the necessary programming skills. This stark accessibility gap raises a fundamental question: Can we enable everyone, regardless of technical background, to build their own LLM agents using natural language alone? To address this challenge, we introduce AutoAgent-a Fully-Automated and highly Self-Developing framework that enables users to create and deploy LLM agents through Natural Language Alone. Operating as an autonomous Agent Operating System, AutoAgent comprises four key components: i) Agentic System Utilities, ii) LLM-powered Actionable Engine, iii) Self-Managing File System, and iv) Self-Play Agent Customization module. This lightweight yet powerful system enables efficient and dynamic creation and modification of tools, agents, and workflows without coding requirements or manual intervention. Beyond its code-free agent development capabilities, AutoAgent also serves as a versatile multi-agent system for General AI Assistants. Comprehensive evaluations on the GAIA benchmark demonstrate AutoAgent's effectiveness in generalist multi-agent tasks, surpassing existing state-of-the-art methods. Furthermore, AutoAgent's Retrieval-Augmented Generation (RAG)-related capabilities have shown consistently superior performance compared to many alternative LLM-based solutions.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2502.05957v2.pdf

GitHub:
β€’ https://github.com/hkuds/autoagent
β€’ https://github.com/hkuds/auto-deep-research

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀3πŸ‘1πŸ”₯1
Article Title:
SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond

Article Date: 26 May 2025

Article Description:
Recent advances such as OpenAI-o1 and DeepSeek R1 have demonstrated the potential of Reinforcement Learning (RL) to enhance reasoning abilities in Large Language Models (LLMs). While open-source replication efforts have primarily focused on mathematical and coding domains, methods and resources for developing general reasoning capabilities remain underexplored. This gap is partly due to the challenge of collecting diverse and verifiable reasoning data suitable for RL. We hypothesize that logical reasoning is critical for developing general reasoning capabilities, as logic forms a fundamental building block of reasoning. In this work, we present SynLogic, a data synthesis framework and dataset that generates diverse logical reasoning data at scale, encompassing 35 diverse logical reasoning tasks. The SynLogic approach enables controlled synthesis of data with adjustable difficulty and quantity. Importantly, all examples can be verified by simple rules, making them ideally suited for RL with verifiable rewards. In our experiments, we validate the effectiveness of RL training on the SynLogic dataset based on 7B and 32B models. SynLogic leads to state-of-the-art logical reasoning performance among open-source datasets, surpassing DeepSeek-R1-Distill-Qwen-32B by 6 points on BBEH. Furthermore, mixing SynLogic data with mathematical and coding tasks improves the training efficiency of these domains and significantly enhances reasoning generalization. Notably, our mixed training model outperforms DeepSeek-R1-Zero-Qwen-32B across multiple benchmarks. These findings position SynLogic as a valuable resource for advancing the broader reasoning capabilities of LLMs. We open-source both the data synthesis pipeline and the SynLogic dataset at https://github.com/MiniMax-AI/SynLogic.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2505.19641v1.pdf

GitHub:
β€’ https://github.com/minimax-ai/synlogic

Datasets:
β€’ MATH
β€’ BBH
β€’ GPQA
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀5
Article Title:
Code to Think, Think to Code: A Survey on Code-Enhanced Reasoning and Reasoning-Driven Code Intelligence in LLMs

Article Date: 26 Feb 2025

Article Description:
In large language models (LLMs), code and reasoning reinforce each other: code offers an abstract, modular, and logic-driven structure that supports reasoning, while reasoning translates high-level goals into smaller, executable steps that drive more advanced code intelligence. In this study, we examine how code serves as a structured medium for enhancing reasoning: it provides verifiable execution paths, enforces logical decomposition, and enables runtime validation. We also explore how improvements in reasoning have transformed code intelligence from basic completion to advanced capabilities, enabling models to address complex software engineering tasks through planning and debugging. Finally, we identify key challenges and propose future research directions to strengthen this synergy, ultimately improving LLM's performance in both areas.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2502.19411v1.pdf

GitHub:
β€’ https://github.com/dayuyang1999/awesome-code-reasoning

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀3
Article Title:
Advanced long-term earth system forecasting by learning the small-scale nature

Article Date: 26 May 2025

Article Description:
Reliable long-term forecast of Earth system dynamics is heavily hampered by instabilities in current AI models during extended autoregressive simulations. These failures often originate from inherent spectral bias, leading to inadequate representation of critical high-frequency, small-scale processes and subsequent uncontrolled error amplification. We present Triton, an AI framework designed to address this fundamental challenge. Inspired by increasing grids to explicitly resolve small scales in numerical models, Triton employs a hierarchical architecture processing information across multiple resolutions to mitigate spectral bias and explicitly model cross-scale dynamics. We demonstrate Triton's superior performance on challenging forecast tasks, achieving stable year-long global temperature forecasts, skillful Kuroshio eddy predictions till 120 days, and high-fidelity turbulence simulations preserving fine-scale structures all without external forcing, with significantly surpassing baseline AI models in long-term stability and accuracy. By effectively suppressing high-frequency error accumulation, Triton offers a promising pathway towards trustworthy AI-driven simulation for climate and earth system science.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2505.19432v1.pdf

GitHub:
β€’ https://github.com/easylearningscores/triton_ai4earth

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀5
Article Title:
OmniConsistency: Learning Style-Agnostic Consistency from Paired Stylization Data

Article Date: 24 May 2025

Article Description:
Diffusion models have advanced image stylization significantly, yet two core challenges persist: (1) maintaining consistent stylization in complex scenes, particularly identity, composition, and fine details, and (2) preventing style degradation in image-to-image pipelines with style LoRAs. GPT-4o's exceptional stylization consistency highlights the performance gap between open-source methods and proprietary models. To bridge this gap, we propose \textbf{OmniConsistency}, a universal consistency plugin leveraging large-scale Diffusion Transformers (DiTs). OmniConsistency contributes: (1) an in-context consistency learning framework trained on aligned image pairs for robust generalization; (2) a two-stage progressive learning strategy decoupling style learning from consistency preservation to mitigate style degradation; and (3) a fully plug-and-play design compatible with arbitrary style LoRAs under the Flux framework. Extensive experiments show that OmniConsistency significantly enhances visual coherence and aesthetic quality, achieving performance comparable to commercial state-of-the-art model GPT-4o.PDFAbstract

PDF Download Link:
https://arxiv.org/pdf/2505.18445v1.pdf

GitHub:
β€’ https://github.com/showlab/omniconsistency

Datasets:
β€’ No datasets information available
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀6
πŸ”Ή Title:
InstaFlow: One Step is Enough for High-Quality Diffusion-Based Text-to-Image Generation

πŸ”Ή Publication Date: Published on Sep 12, 2023

πŸ”Ή Abstract:
Rectified Flow is used to develop an ultra-fast one-step text-to-image generator named InstaFlow, achieving high image quality with significantly reduced inference time compared to existing methods. AI-generated summary Diffusion models have revolutionized text-to-image generation with its exceptional quality and creativity. However, its multi-step sampling process is known to be slow, often requiring tens of inference steps to obtain satisfactory results. Previous attempts to improve its sampling speed and reduce computational costs through distillation have been unsuccessful in achieving a functional one-step model. In this paper, we explore a recent method called Rectified Flow , which, thus far, has only been applied to small datasets. The core of Rectified Flow lies in its reflow procedure , which straightens the trajectories of probability flows , refines the coupling between noises and images, and facilitates the distillation process with student models. We propose a novel text-conditioned pipeline to turn Stable Diffusion (SD) into an ultra-fast one-step model, in which we find reflow plays a critical role in improving the assignment between noise and images. Leveraging our new pipeline, we create, to the best of our knowledge, the first one-step diffusion-based text-to-image generator with SD-level image quality, achieving an FID ( Frechet Inception Distance ) of 23.3 on MS COCO 2017-5k, surpassing the previous state-of-the-art technique, progressive distillation , by a significant margin (37.2 rightarrow 23.3 in FID ). By utilizing an expanded network with 1.7B parameters, we further improve the FID to 22.4. We call our one-step models InstaFlow . On MS COCO 2014-30k, InstaFlow yields an FID of 13.1 in just 0.09 second, the best in leq 0.1 second regime, outperforming the recent StyleGAN-T (13.9 in 0.1 second). Notably, the training of InstaFlow only costs 199 A100 GPU days . Project page:~https://github.com/gnobitab/ InstaFlow .

πŸ”Ή Links:
β€’ arXiv Page: https://arxiv.org/abs/2309.06380
β€’ PDF: https://arxiv.org/pdf/2309.06380

πŸ”Ή Datasets citing this paper:
β€’ https://huggingface.co/datasets/diffusers/community-pipelines-mirror

πŸ”Ή Spaces citing this paper:
β€’ https://huggingface.co/spaces/FlowChef/FlowChef-InstaFlow-Edit
β€’ https://huggingface.co/spaces/FlowChef/FlowChef-InstaFlow-InverseProblem-Inpainting
β€’ https://huggingface.co/spaces/XCLiu/InstaFlow
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀3
πŸ”Ή Title:
Mutarjim: Advancing Bidirectional Arabic-English Translation with a Small Language Model

πŸ”Ή Publication Date: Published on May 23

πŸ”Ή Abstract:
Mutarjim is a compact Arabic-English translation model that outperforms larger models on established benchmarks and achieves state-of-the-art performance on a new comprehensive Tarjama-25 benchmark. AI-generated summary We introduce Mutarjim, a compact yet powerful language model for bidirectional Arabic-English translation . While large-scale LLMs have shown impressive progress in natural language processing tasks, including machine translation, smaller models. Leveraging this insight, we developed Mutarjim based on Kuwain-1.5B , a language model tailored for both Arabic and English. Despite its modest size, Mutarjim outperforms much larger models on several established benchmarks, achieved through an optimized two-phase training approach and a carefully curated, high-quality training corpus .. Experimental results show that Mutarjim rivals models up to 20 times larger while significantly reducing computational costs and training requirements. We also introduce Tarjama-25 , a new benchmark designed to overcome limitations in existing Arabic-English benchmarking datasets, such as domain narrowness , short sentence lengths, and English-source bias . Tarjama-25 comprises 5,000 expert-reviewed sentence pairs and spans a wide range of domains, offering a more comprehensive and balanced evaluation framework. Notably, Mutarjim achieves state-of-the-art performance on the English-to-Arabic task in Tarjama-25 , surpassing even significantly larger and proprietary models like GPT-4 o mini. We publicly release Tarjama-25 to support future research and advance the evaluation of Arabic-English translation systems.

πŸ”Ή Links:
β€’ arXiv Page: https://arxiv.org/abs/2505.17894
β€’ PDF: https://arxiv.org/pdf/2505.17894

πŸ”Ή Datasets citing this paper:
β€’ https://huggingface.co/datasets/Misraj/Arabic-Image-Captioning_100M
β€’ https://huggingface.co/datasets/Misraj/Tarjama-25

πŸ”Ή Spaces citing this paper:
No spaces found
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀2
πŸ”Ή Title:
Will It Still Be True Tomorrow? Multilingual Evergreen Question Classification to Improve Trustworthy QA

πŸ”Ή Publication Date: Published on May 27

πŸ”Ή Abstract:
EverGreenQA, a multilingual QA dataset with evergreen labels, is introduced to benchmark LLMs on temporality encoding and assess their performance through verbalized judgments and uncertainty signals. AI-generated summary Large Language Models (LLMs) often hallucinate in question answering ( QA ) tasks. A key yet underexplored factor contributing to this is the temporality of questions -- whether they are evergreen (answers remain stable over time) or mutable (answers change). In this work, we introduce EverGreen QA , the first multilingual QA dataset with evergreen labels, supporting both evaluation and training. Using EverGreen QA , we benchmark 12 modern LLMs to assess whether they encode question temporality explicitly (via verbalized judgments) or implicitly (via uncertainty signals). We also train EG-E5 , a lightweight multilingual classifier that achieves SoTA performance on this task. Finally, we demonstrate the practical utility of evergreen classification across three applications: improving self-knowledge estimation , filtering QA datasets , and explaining GPT-4o retrieval behavior .

πŸ”Ή Links:
β€’ arXiv Page: https://arxiv.org/abs/2505.21115
β€’ PDF: https://arxiv.org/pdf/2505.21115
β€’ Github: https://github.com/s-nlp/Evergreen-classification

πŸ”Ή Datasets citing this paper:
β€’ https://huggingface.co/datasets/s-nlp/EverGreen-Multilingual

πŸ”Ή Spaces citing this paper:
No spaces found
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀4
πŸ”Ή Title:
Look Before You Leap: A GUI-Critic-R1 Model for Pre-Operative Error Diagnosis in GUI Automation

πŸ”Ή Publication Date: Published on Jun 5

πŸ”Ή Abstract:
A pre-operative critic mechanism with Suggestion-aware Gradient Relative Policy Optimization enhances the reliability of multimodal reasoning tasks in GUI automation. AI-generated summary In recent years, Multimodal Large Language Models (MLLMs) have been extensively utilized for multimodal reasoning tasks, including Graphical User Interface (GUI) automation. Unlike general offline multimodal tasks, GUI automation is executed in online interactive environments, necessitating step-by-step decision-making based on real-time status of the environment. This task has a lower tolerance for decision-making errors at each step, as any mistakes may cumulatively disrupt the process and potentially lead to irreversible outcomes like deletions or payments. To address these issues, we introduce a pre-operative critic mechanism that provides effective feedback prior to the actual execution, by reasoning about the potential outcome and correctness of actions. Specifically, we propose a Suggestion-aware Gradient Relative Policy Optimization (S-GRPO) strategy to construct our pre-operative critic model GUI-Critic-R1 , incorporating a novel suggestion reward to enhance the reliability of the model's feedback. Furthermore, we develop a reasoning-bootstrapping based data collection pipeline to create a GUI-Critic-Train and a GUI-Critic-Test , filling existing gaps in GUI critic data. Static experiments on the GUI-Critic-Test across both mobile and web domains reveal that our GUI-Critic-R1 offers significant advantages in critic accuracy compared to current MLLMs. Dynamic evaluation on GUI automation benchmark further highlights the effectiveness and superiority of our model, as evidenced by improved success rates and operational efficiency.

πŸ”Ή Links:
β€’ arXiv Page: https://arxiv.org/abs/2506.04614
β€’ PDF: https://arxiv.org/pdf/2506.04614
β€’ Github: https://github.com/X-PLUG/MobileAgent

πŸ”Ή Datasets citing this paper:
No datasets found

πŸ”Ή Spaces citing this paper:
No spaces found
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀4
πŸ™πŸ’Έ 500$ FOR THE FIRST 500 WHO JOIN THE CHANNEL! πŸ™πŸ’Έ

Join our channel today for free! Tomorrow it will cost 500$!

https://t.iss.one/+Y4vkzbTTshVhYTQ1

You can join at this link! πŸ‘†πŸ‘‡

https://t.iss.one/+Y4vkzbTTshVhYTQ1
πŸ”Ή Title:
Charting and Navigating Hugging Face's Model Atlas

πŸ”Ή Publication Date: Published on Mar 13

πŸ”Ή Abstract:
An atlas of Hugging Face's model repository provides visualizations and analysis, with methods for mapping undocumented areas based on structural priors. AI-generated summary As there are now millions of publicly available neural networks, searching and analyzing large model repositories becomes increasingly important. Navigating so many models requires an atlas, but as most models are poorly documented charting such an atlas is challenging. To explore the hidden potential of model repositories, we chart a preliminary atlas representing the documented fraction of Hugging Face. It provides stunning visualizations of the model landscape and evolution. We demonstrate several applications of this atlas including predicting model attributes (e.g., accuracy), and analyzing trends in computer vision models . However, as the current atlas remains incomplete, we propose a method for charting undocumented regions. Specifically, we identify high-confidence structural priors based on dominant real-world model training practices. Leveraging these priors, our approach enables accurate mapping of previously undocumented areas of the atlas. We publicly release our datasets, code, and interactive atlas .

πŸ”Ή Links:
β€’ arXiv Page: https://arxiv.org/abs/2503.10633
β€’ PDF: https://arxiv.org/pdf/2503.10633
β€’ Project Page: https://horwitz.ai/model-atlas
β€’ Github: https://github.com/eliahuhorwitz/Model-Atlas

πŸ”Ή Datasets citing this paper:
β€’ https://huggingface.co/datasets/Eliahu/ModelAtlasData

πŸ”Ή Spaces citing this paper:
β€’ https://huggingface.co/spaces/Eliahu/Model-Atlas
β€’ https://huggingface.co/spaces/Nymbo/Model-Atlas
==================================

For more data science resources:

βœ“ https://t.iss.one/DataScienceT
❀4πŸ‘3