Open Deep Search: Democratizing Search with Open-source Reasoning Agents
26 Mar 2025 Β· Salaheddin Alzubi, Creston Brooks, Purva Chiniya, Edoardo Contente, Chiara von Gerlach, Lucas Irwin, Yihan Jiang, Arda Kaz, Windsor Nguyen, Sewoong Oh, Himanshu Tyagi, Pramod Viswanath Β·
Paper: https://arxiv.org/pdf/2503.20201v1.pdf
Code: https://github.com/sentient-agi/opendeepsearch
26 Mar 2025 Β· Salaheddin Alzubi, Creston Brooks, Purva Chiniya, Edoardo Contente, Chiara von Gerlach, Lucas Irwin, Yihan Jiang, Arda Kaz, Windsor Nguyen, Sewoong Oh, Himanshu Tyagi, Pramod Viswanath Β·
We introduce Open Deep Search (ODS) to close the increasing gap between the proprietary search AI solutions, such as Perplexity's Sonar Reasoning Pro and OpenAI's GPT-4o Search Preview, and their open-source counterparts. The main innovation introduced in ODS is to augment the reasoning capabilities of the latest open-source LLMs with reasoning agents that can judiciously use web search tools to answer queries. Concretely, ODS consists of two components that work with a base LLM chosen by the user: Open Search Tool and Open Reasoning Agent. Open Reasoning Agent interprets the given task and completes it by orchestrating a sequence of actions that includes calling tools, one of which is the Open Search Tool. Open Search Tool is a novel web search tool that outperforms proprietary counterparts. Together with powerful open-source reasoning LLMs, such as DeepSeek-R1, ODS nearly matches and sometimes surpasses the existing state-of-the-art baselines on two benchmarks: SimpleQA and FRAMES. For example, on the FRAMES evaluation benchmark, ODS improves the best existing baseline of the recently released GPT-4o Search Preview by 9.7% in accuracy. ODS is a general framework for seamlessly augmenting any LLMs -- for example, DeepSeek-R1 that achieves 82.4% on SimpleQA and 30.1% on FRAMES -- with search and reasoning capabilities to achieve state-of-the-art performance: 88.3% on SimpleQA and 75.3% on FRAMES.
Paper: https://arxiv.org/pdf/2503.20201v1.pdf
Code: https://github.com/sentient-agi/opendeepsearch
#DataScience #ArtificialIntelligence #MachineLearning #PythonProgramming #DeepLearning #LLM #AIResearch #BigData #NeuralNetworks #DataAnalytics #NLP #AutoML #DataVisualization #ScikitLearn #Pandas #NumPy #TensorFlow #AIethics #PredictiveModeling #GPUComputing #OpenSourceAI #DeepSeek #RAG #Agents #GPT4
https://t.iss.one/DataScienceT
π4
Effect-driven interpretation: Functors for natural language composition
π₯ Github: https://github.com/UCSC-VLAA/MedReason
π Paper: https://arxiv.org/abs/2504.00993v1
π Tasks: https://paperswithcode.com/task/knowledge-graphs
π Tasks: https://paperswithcode.com/task/knowledge-graphs
Please open Telegram to view this post
VIEW IN TELEGRAM
π2β€1
Forwarded from Python | Machine Learning | Coding | R
Explore the world of Data Science through Jupyter Notebooksβinsights, tutorials, and tools to boost your data journey. Code, analyze, and visualize smarter with every post.
https://t.iss.one/DataScienceN
https://t.iss.one/DataScienceN
Telegram
Data Science Jupyter Notebooks
Explore the world of Data Science through Jupyter Notebooksβinsights, tutorials, and tools to boost your data journey. Code, analyze, and visualize smarter with every post.
π2
TripoSG: High-Fidelity 3D Shape Synthesis using Large-Scale Rectified Flow Models
10 Feb 2025 Β· Yangguang Li, Zi-Xin Zou, Zexiang Liu, Dehu Wang, Yuan Liang, Zhipeng Yu, Xingchao Liu, Yuan-Chen Guo, Ding Liang, Wanli Ouyang, Yan-Pei Cao Β·
Paper: https://arxiv.org/pdf/2502.06608v3.pdf
Codes:
https://github.com/VAST-AI-Research/TripoSG
https://github.com/tencent/flashvdm
Dataset: 100poisonMpts
10 Feb 2025 Β· Yangguang Li, Zi-Xin Zou, Zexiang Liu, Dehu Wang, Yuan Liang, Zhipeng Yu, Xingchao Liu, Yuan-Chen Guo, Ding Liang, Wanli Ouyang, Yan-Pei Cao Β·
Recent advancements in diffusion techniques have propelled image and video generation to unprecedented levels of quality, significantly accelerating the deployment and application of generative AI. However, 3D shape generation technology has so far lagged behind, constrained by limitations in 3D data scale, complexity of 3D data processing, and insufficient exploration of advanced techniques in the 3D domain. Current approaches to 3D shape generation face substantial challenges in terms of output quality, generalization capability, and alignment with input conditions. We present TripoSG, a new streamlined shape diffusion paradigm capable of generating high-fidelity 3D meshes with precise correspondence to input images. Specifically, we propose: 1) A large-scale rectified flow transformer for 3D shape generation, achieving state-of-the-art fidelity through training on extensive, high-quality data. 2) A hybrid supervised training strategy combining SDF, normal, and eikonal losses for 3D VAE, achieving high-quality 3D reconstruction performance. 3) A data processing pipeline to generate 2 million high-quality 3D samples, highlighting the crucial rules for data quality and quantity in training 3D generative models. Through comprehensive experiments, we have validated the effectiveness of each component in our new framework. The seamless integration of these parts has enabled TripoSG to achieve state-of-the-art performance in 3D shape generation. The resulting 3D shapes exhibit enhanced detail due to high-resolution capabilities and demonstrate exceptional fidelity to input images. Moreover, TripoSG demonstrates improved versatility in generating 3D models from diverse image styles and contents, showcasing strong generalization capabilities. To foster progress and innovation in the field of 3D generation, we will make our model publicly available.
Paper: https://arxiv.org/pdf/2502.06608v3.pdf
Codes:
https://github.com/VAST-AI-Research/TripoSG
https://github.com/tencent/flashvdm
Dataset: 100poisonMpts
#DataScience #ArtificialIntelligence #MachineLearning #PythonProgramming #DeepLearning #LLM #AIResearch #BigData #NeuralNetworks #DataAnalytics #NLP #AutoML #DataVisualization #ScikitLearn #Pandas #NumPy #TensorFlow #AIethics #PredictiveModeling #GPUComputing #OpenSourceAI #DeepSeek #RAG #Agents #GPT4
https://t.iss.one/DataScienceT
π3
The latest and the most up-to-date cyber news will be presented on PPHM HACKER NEWS.
PPHM subscribers are the first people that receive firsthand cybernews and Tech news.
You won't miss any cyber news with us.
https://t.iss.one/pphm_HackerNews
PPHM subscribers are the first people that receive firsthand cybernews and Tech news.
You won't miss any cyber news with us.
https://t.iss.one/pphm_HackerNews
Telegram
PPHM HACKER NEWS
PPHM Hacker News is a reliable news outlet that brings you the latest and most credential cyber news.
Website:
https://pphmnews.com/
Website:
https://pphmnews.com/
π3
Data Science | Machine Learning with Python for Researchers pinned Β«The latest and the most up-to-date cyber news will be presented on PPHM HACKER NEWS. PPHM subscribers are the first people that receive firsthand cybernews and Tech news. You won't miss any cyber news with us. https://t.iss.one/pphm_HackerNewsΒ»
Large Language Model Agent: A Survey on Methodology, Applications and Challenges
Paper: https://arxiv.org/pdf/2503.21460v1.pdf
Code: https://github.com/luo-junyu/awesome-agent-papers
https://t.iss.one/DataScienceTβοΈ
27 Mar 2025 Β· Junyu Luo, Weizhi Zhang, Ye Yuan, Yusheng Zhao, Junwei Yang, Yiyang Gu, Bohan Wu, Binqi Chen, Ziyue Qiao, Qingqing Long, RongCheng Tu, Xiao Luo, Wei Ju, Zhiping Xiao, Yifan Wang, Meng Xiao, Chenwu Liu, Jingyang Yuan, Shichang Zhang, Yiqiao Jin, Fan Zhang, Xian Wu, Hanqing Zhao, DaCheng Tao, Philip S. Yu, Ming Zhang
The era of intelligent agents is upon us, driven by revolutionary advancements in large language models. Large Language Model (LLM) agents, with goal-driven behaviors and dynamic adaptation capabilities, potentially represent a critical pathway toward artificial general intelligence. This survey systematically deconstructs LLM agent systems through a methodology-centered taxonomy, linking architectural foundations, collaboration mechanisms, and evolutionary pathways. We unify fragmented research threads by revealing fundamental connections between agent design principles and their emergent behaviors in complex environments. Our work provides a unified architectural perspective, examining how agents are constructed, how they collaborate, and how they evolve over time, while also addressing evaluation methodologies, tool applications, practical challenges, and diverse application domains. By surveying the latest developments in this rapidly evolving field, we offer researchers a structured taxonomy for understanding LLM agents and identify promising directions for future research. The collection is available at https://github.com/luo-junyu/Awesome-Agent-Papers.
Paper: https://arxiv.org/pdf/2503.21460v1.pdf
Code: https://github.com/luo-junyu/awesome-agent-papers
https://t.iss.one/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
π5π1
This media is not supported in your browser
VIEW IN TELEGRAM
Crystal Generation with Space Group Informed Transformer
π₯ Github: https://github.com/deepmodeling/crystalformer
π Paper: https://arxiv.org/abs/2504.02367v1
π Dataset: https://paperswithcode.com/dataset/alex-20
π Dataset: https://paperswithcode.com/dataset/alex-20
Please open Telegram to view this post
VIEW IN TELEGRAM
π5
4 advanced attention mechanisms you should know:
β’ Slim attention β 8Γ less memory, 5Γ faster generation by storing only K from KV pairs and recomputing V.
β’ XAttention β 13.5Γ speedup on long sequences via "looking" at the sum of values along diagonal lines in the attention matrix.
β’ Kolmogorov-Arnold Attention, KArAt β Adaptable attention with learnable activation functions using KANs instead of softmax.
β’ Multi-token attention (MTA) β Lets the model consider groups of nearby words together for smarter long-context handling.
Read the overview of them in our free article on https://huggingface.co/blog/Kseniase/attentions
https://t.iss.one/DataScienceMπ
β’ Slim attention β 8Γ less memory, 5Γ faster generation by storing only K from KV pairs and recomputing V.
β’ XAttention β 13.5Γ speedup on long sequences via "looking" at the sum of values along diagonal lines in the attention matrix.
β’ Kolmogorov-Arnold Attention, KArAt β Adaptable attention with learnable activation functions using KANs instead of softmax.
β’ Multi-token attention (MTA) β Lets the model consider groups of nearby words together for smarter long-context handling.
Read the overview of them in our free article on https://huggingface.co/blog/Kseniase/attentions
https://t.iss.one/DataScienceM
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
π8
CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models
28 Mar 2025 Β· Zhihang Lin, Mingbao Lin, Yuan Xie, Rongrong Ji
Paper: https://arxiv.org/pdf/2503.22342v1.pdf
Code: https://github.com/lzhxmu/cppo
Datasets: GSM8K - MATH
https://t.iss.one/DataScienceTβ
28 Mar 2025 Β· Zhihang Lin, Mingbao Lin, Yuan Xie, Rongrong Ji
This paper introduces Completion Pruning Policy Optimization (CPPO) to accelerate the training of reasoning models based on Group Relative Policy Optimization (GRPO). GRPO, while effective, incurs high training costs due to the need for sampling multiple completions for each question. Our experiment and theoretical analysis reveals that the number of completions impacts model accuracy yet increases training time multiplicatively, and not all completions contribute equally to policy training -- their contribution depends on their relative advantage. To address these issues, we propose CPPO, which prunes completions with low absolute advantages, significantly reducing the number needed for gradient calculation and updates. Additionally, we introduce a dynamic completion allocation strategy to maximize GPU utilization by incorporating additional questions, further enhancing training efficiency. Experimental results demonstrate that CPPO achieves up to
speedup on GSM8K and on Math while preserving or even enhancing the accuracy compared to the original GRPO. We release our code at https://github.com/lzhxmu/CPPO.
Paper: https://arxiv.org/pdf/2503.22342v1.pdf
Code: https://github.com/lzhxmu/cppo
Datasets: GSM8K - MATH
https://t.iss.one/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
π6
This media is not supported in your browser
VIEW IN TELEGRAM
β½ VoRA: Vision as LoRA β½
#ByteDance introduces #VoRA (Vision as #LoRA) β a novel framework that transforms #LLMs into Multimodal Large Language Models (MLLMs) by integrating vision-specific LoRA layers.
All training data, source code, and model weights are openly available!
Key Resources:
Overview: https://t.ly/guNVN
Paper: arxiv.org/pdf/2503.20680
GitHub Repo: github.com/Hon-Wong/VoRA
Project Page: georgeluimmortal.github.io/vora-homepage.github.io
#ByteDance introduces #VoRA (Vision as #LoRA) β a novel framework that transforms #LLMs into Multimodal Large Language Models (MLLMs) by integrating vision-specific LoRA layers.
All training data, source code, and model weights are openly available!
Key Resources:
Overview: https://t.ly/guNVN
Paper: arxiv.org/pdf/2503.20680
GitHub Repo: github.com/Hon-Wong/VoRA
Project Page: georgeluimmortal.github.io/vora-homepage.github.io
π5β€1
This media is not supported in your browser
VIEW IN TELEGRAM
#SkyworkAI unveils #SkyReelsA2 β a controllable video generation framework that can assemble arbitrary visual elements (e.g., characters, objects, backgrounds) into fully synthesized videos from text prompts.
Code, models, and evaluation benchmark are all released!
π Resources:
Review: https://t.ly/MEjzL
Paper: https://arxiv.org/pdf/2504.02436
Project: https://skyworkai.github.io/skyreels-a2.github.io/
Repo: https://github.com/SkyworkAI/SkyReels-A2
#AI #VideoGeneration #Multimodal #GenerativeAI #SkyReels #OpenSource
https://t.iss.one/DataScienceT
Please open Telegram to view this post
VIEW IN TELEGRAM
π5
Forwarded from Python | Machine Learning | Coding | R
This channels is for Programmers, Coders, Software Engineers.
0οΈβ£ Python
1οΈβ£ Data Science
2οΈβ£ Machine Learning
3οΈβ£ Data Visualization
4οΈβ£ Artificial Intelligence
5οΈβ£ Data Analysis
6οΈβ£ Statistics
7οΈβ£ Deep Learning
8οΈβ£ programming Languages
β
https://t.iss.one/addlist/8_rRW2scgfRhOTc0
β
https://t.iss.one/Codeprogrammer
Please open Telegram to view this post
VIEW IN TELEGRAM
This media is not supported in your browser
VIEW IN TELEGRAM
Adding TTT layers into a pre-trained Transformer enables generating a one-minute clip from text storyboards.
Videos, code & annotations released
#AI #VideoGeneration #MachineLearning #DeepLearning #Transformers #TTT #GenerativeAI
Please open Telegram to view this post
VIEW IN TELEGRAM
π2
This media is not supported in your browser
VIEW IN TELEGRAM
#AI #DeepLearning #ComputerVision #YOLO #AttentionMechanism #OpenSource
Please open Telegram to view this post
VIEW IN TELEGRAM
π5β€1
ZClip: Adaptive Spike Mitigation for LLM Pre-Training
π₯ Github: https://github.com/bluorion-com/ZClip
π Paper: https://arxiv.org/abs/2504.02507v1
π Dataset: https://paperswithcode.com/dataset/hellaswag
π₯ Github: https://github.com/bluorion-com/ZClip
π Paper: https://arxiv.org/abs/2504.02507v1
π Dataset: https://paperswithcode.com/dataset/hellaswag
π4
Forwarded from Python | Machine Learning | Coding | R
Please open Telegram to view this post
VIEW IN TELEGRAM
π6
#DataScience #MachineLearning #DeepLearning #Python #AI #MLProjects #DataAnalysis #ExplainableAI #100DaysOfCode #TechEducation #MLInterviewPrep #NeuralNetworks #MathForML #Statistics #Coding #AIForEveryone #PythonForDataScience
Please open Telegram to view this post
VIEW IN TELEGRAM
π10β€2
d9ff625c-57ff-44d5-b57d-e5a30c4c0026.gif
120.6 KB
Zep: A Temporal Knowledge Graph Architecture for Agent Memory
20 Jan 2025 Β· Preston Rasmussen, Pavlo Paliychuk, Travis Beauvais, Jack Ryan, Daniel Chalef Β·
Paper: https://arxiv.org/pdf/2501.13956v1.pdf
Code: https://github.com/getzep/graphiti
β‘οΈ BEST DATA SCIENCE CHANNELS ON TELEGRAM π
20 Jan 2025 Β· Preston Rasmussen, Pavlo Paliychuk, Travis Beauvais, Jack Ryan, Daniel Chalef Β·
We introduce Zep, a novel memory layer service for AI agents that outperforms the current state-of-the-art system, MemGPT, in the Deep Memory Retrieval (DMR) benchmark. Additionally, Zep excels in more comprehensive and challenging evaluations than DMR that better reflect real-world enterprise use cases. While existing retrieval-augmented generation (RAG) frameworks for large language model (LLM)-based agents are limited to static document retrieval, enterprise applications demand dynamic knowledge integration from diverse sources including ongoing conversations and business data. Zep addresses this fundamental limitation through its core component Graphiti -- a temporally-aware knowledge graph engine that dynamically synthesizes both unstructured conversational data and structured business data while maintaining historical relationships. In the DMR benchmark, which the MemGPT team established as their primary evaluation metric, Zep demonstrates superior performance (94.8% vs 93.4%). Beyond DMR, Zep's capabilities are further validated through the more challenging LongMemEval benchmark, which better reflects enterprise use cases through complex temporal reasoning tasks. In this evaluation, Zep achieves substantial results with accuracy improvements of up to 18.5% while simultaneously reducing response latency by 90% compared to baseline implementations. These results are particularly pronounced in enterprise-critical tasks such as cross-session information synthesis and long-term context maintenance, demonstrating Zep's effectiveness for deployment in real-world applications.
Paper: https://arxiv.org/pdf/2501.13956v1.pdf
Code: https://github.com/getzep/graphiti
Please open Telegram to view this post
VIEW IN TELEGRAM
π5