Data Science | Machine Learning with Python for Researchers
31.5K subscribers
1.6K photos
102 videos
22 files
1.87K links
Admin: @HusseinSheikho

The Data Science and Python channel is for researchers and advanced programmers

Buy ads: https://telega.io/c/dataScienceT
Download Telegram
PiEEG kit - bioscience Lab in home for your Brain and Body

πŸ–₯ Github: https://github.com/pieeg-club/PiEEG_Kit

πŸ“• Paper: https://arxiv.org/abs/2503.13482

🌟 Methods: https://paperswithcode.com/task/eeg-1
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘4
FastCuRL: Curriculum Reinforcement Learning with Progressive Context Extension for Efficient Training R1-like Reasoning Models

πŸ–₯ Github: https://github.com/nick7nlp/FastCuRL

πŸ“• Paper: https://arxiv.org/abs/2503.17287v1

🌟 Tasks
: https://paperswithcode.com/task/language-modeling
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘5❀1
Greetings.
As part of our research, we want to write a review article in the field of pathology. Friends who are interested in the 2nd and 3rd places on this topic can participate.

βœ… Approximate start time: April 10th.

Journal: scientific reports https://www.nature.com/srep/

Price:
2: $400
3: $300

I will help with complete explanations and how to write each section.

@Raminmousa
@Machine_learn
@Paper4money
πŸ‘4❀1
InfiniteYou: Flexible Photo Recrafting While Preserving Your Identity

20 Mar 2025 Β· Liming Jiang, Qing Yan, Yumin Jia, Zichuan Liu, Hao Kang, Xin Lu Β·

Achieving flexible and high-fidelity identity-preserved image generation remains formidable, particularly with advanced Diffusion Transformers (DiTs) like FLUX. We introduce InfiniteYou (InfU), one of the earliest robust frameworks leveraging DiTs for this task. InfU addresses significant issues of existing methods, such as insufficient identity similarity, poor text-image alignment, and low generation quality and aesthetics. Central to InfU is InfuseNet, a component that injects identity features into the DiT base model via residual connections, enhancing identity similarity while maintaining generation capabilities. A multi-stage training strategy, including pretraining and supervised fine-tuning (SFT) with synthetic single-person-multiple-sample (SPMS) data, further improves text-image alignment, ameliorates image quality, and alleviates face copy-pasting. Extensive experiments demonstrate that InfU achieves state-of-the-art performance, surpassing existing baselines. In addition, the plug-and-play design of InfU ensures compatibility with various existing methods, offering a valuable contribution to the broader community.


Paper: https://arxiv.org/pdf/2503.16418v1.pdf

Code: https://github.com/bytedance/infiniteyou

Dataset: 10,000 People - Human Pose Recognition Data

https://t.iss.one/DataScienceT ⚠️
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘3❀1
LHM: Large Animatable Human Reconstruction Model from a Single Image in Seconds

13 Mar 2025 Β· Lingteng Qiu, Xiaodong Gu, Peihao Li, Qi Zuo, Weichao Shen, Junfei Zhang, Kejie Qiu, Weihao Yuan, GuanYing Chen, Zilong Dong, Liefeng Bo Β·

Animatable 3D human reconstruction from a single image is a challenging problem due to the ambiguity in decoupling geometry, appearance, and deformation. Recent advances in 3D human reconstruction mainly focus on static human modeling, and the reliance of using synthetic 3D scans for training limits their generalization ability. Conversely, optimization-based video methods achieve higher fidelity but demand controlled capture conditions and computationally intensive refinement processes. Motivated by the emergence of large reconstruction models for efficient static reconstruction, we propose LHM (Large Animatable Human Reconstruction Model) to infer high-fidelity avatars represented as 3D Gaussian splatting in a feed-forward pass. Our model leverages a multimodal transformer architecture to effectively encode the human body positional features and image features with attention mechanism, enabling detailed preservation of clothing geometry and texture. To further boost the face identity preservation and fine detail recovery, we propose a head feature pyramid encoding scheme to aggregate multi-scale features of the head regions. Extensive experiments demonstrate that our LHM generates plausible animatable human in seconds without post-processing for face and hands, outperforming existing methods in both reconstruction accuracy and generalization ability.


Paper: https://arxiv.org/pdf/2503.10625v1.pdf

Code: https://github.com/aigc3d/LHM

https://t.iss.one/DataScienceT ⚠️
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘4
Long-Context Autoregressive Video Modeling with Next-Frame Prediction

25 Mar 2025 Β· YuChao Gu, Weijia Mao, Mike Zheng Shou Β·

Long-context autoregressive modeling has significantly advanced language generation, but video generation still struggles to fully utilize extended temporal contexts. To investigate long-context video modeling, we introduce Frame AutoRegressive (FAR), a strong baseline for video autoregressive modeling. Just as language models learn causal dependencies between tokens (i.e., Token AR), FAR models temporal causal dependencies between continuous frames, achieving better convergence than Token AR and video diffusion transformers. Building on FAR, we observe that long-context vision modeling faces challenges due to visual redundancy. Existing RoPE lacks effective temporal decay for remote context and fails to extrapolate well to long video sequences. Additionally, training on long videos is computationally expensive, as vision tokens grow much faster than language tokens. To tackle these issues, we propose balancing locality and long-range dependency. We introduce FlexRoPE, an test-time technique that adds flexible temporal decay to RoPE, enabling extrapolation to 16x longer vision contexts. Furthermore, we propose long short-term context modeling, where a high-resolution short-term context window ensures fine-grained temporal consistency, while an unlimited long-term context window encodes long-range information using fewer tokens. With this approach, we can train on long video sequences with a manageable token context length. We demonstrate that FAR achieves state-of-the-art performance in both short- and long-video generation, providing a simple yet effective baseline for video autoregressive modeling.


Paper: https://arxiv.org/pdf/2503.19325v1.pdf

Code: https://github.com/showlab/FAR

Dataset: UCF101

Ranked #2 on Video Generation on UCF-101

https://t.iss.one/DataScienceT ⚠️
Please open Telegram to view this post
VIEW IN TELEGRAM
❀3πŸ‘3
Harnessing the Reasoning Economy: A Survey of Efficient Reasoning for Large Language Models

πŸ–₯ Github: https://github.com/devoallen/awesome-reasoning-economy-papers

πŸ“• Paper: https://arxiv.org/abs/2503.24377v1
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘4❀1
This channels is for Programmers, Coders, Software Engineers.

0️⃣ Python
1️⃣ Data Science
2️⃣ Machine Learning
3️⃣ Data Visualization
4️⃣ Artificial Intelligence
5️⃣ Data Analysis
6️⃣ Statistics
7️⃣ Deep Learning
8️⃣ programming Languages

βœ… https://t.iss.one/addlist/8_rRW2scgfRhOTc0

βœ… https://t.iss.one/Codeprogrammer
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘3❀1
πŸ™πŸ’Έ 500$ FOR THE FIRST 500 WHO JOIN THE CHANNEL! πŸ™πŸ’Έ

Join our channel today for free! Tomorrow it will cost 500$!

https://t.iss.one/+vhF2zNz5GBw3NTU1

You can join at this link! πŸ‘†πŸ‘‡

https://t.iss.one/+vhF2zNz5GBw3NTU1
πŸ‘3
Open Deep Search: Democratizing Search with Open-source Reasoning Agents

26 Mar 2025 Β· Salaheddin Alzubi, Creston Brooks, Purva Chiniya, Edoardo Contente, Chiara von Gerlach, Lucas Irwin, Yihan Jiang, Arda Kaz, Windsor Nguyen, Sewoong Oh, Himanshu Tyagi, Pramod Viswanath Β·

We introduce Open Deep Search (ODS) to close the increasing gap between the proprietary search AI solutions, such as Perplexity's Sonar Reasoning Pro and OpenAI's GPT-4o Search Preview, and their open-source counterparts. The main innovation introduced in ODS is to augment the reasoning capabilities of the latest open-source LLMs with reasoning agents that can judiciously use web search tools to answer queries. Concretely, ODS consists of two components that work with a base LLM chosen by the user: Open Search Tool and Open Reasoning Agent. Open Reasoning Agent interprets the given task and completes it by orchestrating a sequence of actions that includes calling tools, one of which is the Open Search Tool. Open Search Tool is a novel web search tool that outperforms proprietary counterparts. Together with powerful open-source reasoning LLMs, such as DeepSeek-R1, ODS nearly matches and sometimes surpasses the existing state-of-the-art baselines on two benchmarks: SimpleQA and FRAMES. For example, on the FRAMES evaluation benchmark, ODS improves the best existing baseline of the recently released GPT-4o Search Preview by 9.7% in accuracy. ODS is a general framework for seamlessly augmenting any LLMs -- for example, DeepSeek-R1 that achieves 82.4% on SimpleQA and 30.1% on FRAMES -- with search and reasoning capabilities to achieve state-of-the-art performance: 88.3% on SimpleQA and 75.3% on FRAMES.


Paper: https://arxiv.org/pdf/2503.20201v1.pdf

Code: https://github.com/sentient-agi/opendeepsearch

#DataScience #ArtificialIntelligence #MachineLearning #PythonProgramming #DeepLearning #LLM #AIResearch #BigData #NeuralNetworks #DataAnalytics #NLP #AutoML #DataVisualization #ScikitLearn #Pandas #NumPy #TensorFlow #AIethics #PredictiveModeling #GPUComputing #OpenSourceAI #DeepSeek #RAG #Agents #GPT4

https://t.iss.one/DataScienceT
πŸ‘4
Effect-driven interpretation: Functors for natural language composition

πŸ–₯ Github: https://github.com/UCSC-VLAA/MedReason

πŸ“• Paper: https://arxiv.org/abs/2504.00993v1

πŸ”— Tasks: https://paperswithcode.com/task/knowledge-graphs
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘2❀1
TripoSG: High-Fidelity 3D Shape Synthesis using Large-Scale Rectified Flow Models

10 Feb 2025 Β· Yangguang Li, Zi-Xin Zou, Zexiang Liu, Dehu Wang, Yuan Liang, Zhipeng Yu, Xingchao Liu, Yuan-Chen Guo, Ding Liang, Wanli Ouyang, Yan-Pei Cao Β·

Recent advancements in diffusion techniques have propelled image and video generation to unprecedented levels of quality, significantly accelerating the deployment and application of generative AI. However, 3D shape generation technology has so far lagged behind, constrained by limitations in 3D data scale, complexity of 3D data processing, and insufficient exploration of advanced techniques in the 3D domain. Current approaches to 3D shape generation face substantial challenges in terms of output quality, generalization capability, and alignment with input conditions. We present TripoSG, a new streamlined shape diffusion paradigm capable of generating high-fidelity 3D meshes with precise correspondence to input images. Specifically, we propose: 1) A large-scale rectified flow transformer for 3D shape generation, achieving state-of-the-art fidelity through training on extensive, high-quality data. 2) A hybrid supervised training strategy combining SDF, normal, and eikonal losses for 3D VAE, achieving high-quality 3D reconstruction performance. 3) A data processing pipeline to generate 2 million high-quality 3D samples, highlighting the crucial rules for data quality and quantity in training 3D generative models. Through comprehensive experiments, we have validated the effectiveness of each component in our new framework. The seamless integration of these parts has enabled TripoSG to achieve state-of-the-art performance in 3D shape generation. The resulting 3D shapes exhibit enhanced detail due to high-resolution capabilities and demonstrate exceptional fidelity to input images. Moreover, TripoSG demonstrates improved versatility in generating 3D models from diverse image styles and contents, showcasing strong generalization capabilities. To foster progress and innovation in the field of 3D generation, we will make our model publicly available.


Paper: https://arxiv.org/pdf/2502.06608v3.pdf

Codes:
https://github.com/VAST-AI-Research/TripoSG
https://github.com/tencent/flashvdm

Dataset: 100poisonMpts

#DataScience #ArtificialIntelligence #MachineLearning #PythonProgramming #DeepLearning #LLM #AIResearch #BigData #NeuralNetworks #DataAnalytics #NLP #AutoML #DataVisualization #ScikitLearn #Pandas #NumPy #TensorFlow #AIethics #PredictiveModeling #GPUComputing #OpenSourceAI #DeepSeek #RAG #Agents #GPT4

https://t.iss.one/DataScienceT
πŸ‘3
The latest and the most up-to-date cyber news will be presented on PPHM HACKER NEWS.
PPHM subscribers are the first people that receive firsthand cybernews and Tech news.

You won't miss any cyber news with us.


https://t.iss.one/pphm_HackerNews
πŸ‘3
Data Science | Machine Learning with Python for Researchers pinned Β«The latest and the most up-to-date cyber news will be presented on PPHM HACKER NEWS. PPHM subscribers are the first people that receive firsthand cybernews and Tech news. You won't miss any cyber news with us. https://t.iss.one/pphm_HackerNewsΒ»
Large Language Model Agent: A Survey on Methodology, Applications and Challenges

27 Mar 2025 Β· Junyu Luo, Weizhi Zhang, Ye Yuan, Yusheng Zhao, Junwei Yang, Yiyang Gu, Bohan Wu, Binqi Chen, Ziyue Qiao, Qingqing Long, RongCheng Tu, Xiao Luo, Wei Ju, Zhiping Xiao, Yifan Wang, Meng Xiao, Chenwu Liu, Jingyang Yuan, Shichang Zhang, Yiqiao Jin, Fan Zhang, Xian Wu, Hanqing Zhao, DaCheng Tao, Philip S. Yu, Ming Zhang

The era of intelligent agents is upon us, driven by revolutionary advancements in large language models. Large Language Model (LLM) agents, with goal-driven behaviors and dynamic adaptation capabilities, potentially represent a critical pathway toward artificial general intelligence. This survey systematically deconstructs LLM agent systems through a methodology-centered taxonomy, linking architectural foundations, collaboration mechanisms, and evolutionary pathways. We unify fragmented research threads by revealing fundamental connections between agent design principles and their emergent behaviors in complex environments. Our work provides a unified architectural perspective, examining how agents are constructed, how they collaborate, and how they evolve over time, while also addressing evaluation methodologies, tool applications, practical challenges, and diverse application domains. By surveying the latest developments in this rapidly evolving field, we offer researchers a structured taxonomy for understanding LLM agents and identify promising directions for future research. The collection is available at https://github.com/luo-junyu/Awesome-Agent-Papers.


Paper: https://arxiv.org/pdf/2503.21460v1.pdf

Code: https://github.com/luo-junyu/awesome-agent-papers

https://t.iss.one/DataScienceT βœ‰οΈ
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘5πŸ‘1
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘5
4 advanced attention mechanisms you should know:

β€’ Slim attention β€” 8Γ— less memory, 5Γ— faster generation by storing only K from KV pairs and recomputing V.

β€’ XAttention β€” 13.5Γ— speedup on long sequences via "looking" at the sum of values along diagonal lines in the attention matrix.

β€’ Kolmogorov-Arnold Attention, KArAt β€” Adaptable attention with learnable activation functions using KANs instead of softmax.

β€’ Multi-token attention (MTA) β€” Lets the model consider groups of nearby words together for smarter long-context handling.

Read the overview of them in our free article on
https://huggingface.co/blog/Kseniase/attentions

https://t.iss.one/DataScienceM 🌟
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘8