✨RadAgent: A tool-using AI agent for stepwise interpretation of chest computed tomography
📝 Summary:
RadAgent, a tool-using AI agent, enhances chest CT report generation through interpretable step-by-step reasoning traces that improve clinical accuracy, robustness, and faithfulness compared to existi...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15231
• PDF: https://arxiv.org/pdf/2604.15231
• Project Page: https://rad-agent.github.io
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
RadAgent, a tool-using AI agent, enhances chest CT report generation through interpretable step-by-step reasoning traces that improve clinical accuracy, robustness, and faithfulness compared to existi...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15231
• PDF: https://arxiv.org/pdf/2604.15231
• Project Page: https://rad-agent.github.io
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
✨Boosting Visual Instruction Tuning with Self-Supervised Guidance
📝 Summary:
Multimodal models struggle with visual reasoning due to under-utilizing visual information during instruction tuning. This paper proposes augmenting instruction tuning with visually grounded self-supervised tasks expressed as natural language. This simple method significantly improves performance...
🔹 Publication Date: Published on Apr 14
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.12966
• PDF: https://arxiv.org/pdf/2604.12966
• Github: https://github.com/sirkosophia/V-GIFT
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
Multimodal models struggle with visual reasoning due to under-utilizing visual information during instruction tuning. This paper proposes augmenting instruction tuning with visually grounded self-supervised tasks expressed as natural language. This simple method significantly improves performance...
🔹 Publication Date: Published on Apr 14
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.12966
• PDF: https://arxiv.org/pdf/2604.12966
• Github: https://github.com/sirkosophia/V-GIFT
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
✨An Optimal Transport-driven Approach for Cultivating Latent Space in Online Incremental Learning
📝 Summary:
An online mixture model learning framework based on optimal transport theory addresses challenges in incremental learning with distributional shifts by enabling dynamic centroid updates and improving ...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2211.16780
• PDF: https://arxiv.org/pdf/2211.16780
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
An online mixture model learning framework based on optimal transport theory addresses challenges in incremental learning with distributional shifts by enabling dynamic centroid updates and improving ...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2211.16780
• PDF: https://arxiv.org/pdf/2211.16780
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
✨Three-Phase Transformer
📝 Summary:
The Three-Phase Transformer introduces a structural prior for decoder-only Transformers through channel partitioning and phase-respecting operations that stabilize training and improve convergence. AI...
🔹 Publication Date: Published on Apr 15
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14430
• PDF: https://arxiv.org/pdf/2604.14430
• Github: https://github.com/achelousace/three-phase-transformer
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
The Three-Phase Transformer introduces a structural prior for decoder-only Transformers through channel partitioning and phase-respecting operations that stabilize training and improve convergence. AI...
🔹 Publication Date: Published on Apr 15
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14430
• PDF: https://arxiv.org/pdf/2604.14430
• Github: https://github.com/achelousace/three-phase-transformer
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
✨Beyond Prompts: Unconditional 3D Inversion for Out-of-Distribution Shapes
📝 Summary:
State-of-the-art text-to-3D generative models suffer from latent sink traps where they lose sensitivity to text prompts, but a robust framework can overcome this by decoupling geometric representation...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14914
• PDF: https://arxiv.org/pdf/2604.14914
• Project Page: https://daidedou.sorpi.fr/publication/beyondprompts
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
State-of-the-art text-to-3D generative models suffer from latent sink traps where they lose sensitivity to text prompts, but a robust framework can overcome this by decoupling geometric representation...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14914
• PDF: https://arxiv.org/pdf/2604.14914
• Project Page: https://daidedou.sorpi.fr/publication/beyondprompts
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
❤1
✨Reinforcement Learning via Value Gradient Flow
📝 Summary:
Value Gradient Flow presents a scalable approach to behavior-regularized reinforcement learning by formulating it as an optimal transport problem solved through discrete gradient flow, enabling adapti...
🔹 Publication Date: Published on Apr 15
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14265
• PDF: https://arxiv.org/pdf/2604.14265
• Project Page: https://ryanxhr.github.io/vgf/
• Github: https://github.com/ryanxhr/vgf
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
Value Gradient Flow presents a scalable approach to behavior-regularized reinforcement learning by formulating it as an optimal transport problem solved through discrete gradient flow, enabling adapti...
🔹 Publication Date: Published on Apr 15
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14265
• PDF: https://arxiv.org/pdf/2604.14265
• Project Page: https://ryanxhr.github.io/vgf/
• Github: https://github.com/ryanxhr/vgf
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
👍3
This media is not supported in your browser
VIEW IN TELEGRAM
✨Envisioning the Future, One Step at a Time
📝 Summary:
Autoregressive diffusion models predict open-set future scene dynamics by modeling sparse point trajectories, enabling fast and scalable multi-modal motion prediction with physical plausibility. AI-ge...
🔹 Publication Date: Published on Apr 10
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.09527
• PDF: https://arxiv.org/pdf/2604.09527
• Project Page: https://compvis.github.io/myriad
• Github: https://github.com/compvis/myriad
🔹 Models citing this paper:
• https://huggingface.co/CompVis/myriad
✨ Datasets citing this paper:
• https://huggingface.co/datasets/CompVis/owm-95
• https://huggingface.co/datasets/CompVis/myriad-physics
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
Autoregressive diffusion models predict open-set future scene dynamics by modeling sparse point trajectories, enabling fast and scalable multi-modal motion prediction with physical plausibility. AI-ge...
🔹 Publication Date: Published on Apr 10
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.09527
• PDF: https://arxiv.org/pdf/2604.09527
• Project Page: https://compvis.github.io/myriad
• Github: https://github.com/compvis/myriad
🔹 Models citing this paper:
• https://huggingface.co/CompVis/myriad
✨ Datasets citing this paper:
• https://huggingface.co/datasets/CompVis/owm-95
• https://huggingface.co/datasets/CompVis/myriad-physics
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
❤1
This media is not supported in your browser
VIEW IN TELEGRAM
✨VideoFlexTok: Flexible-Length Coarse-to-Fine Video Tokenization
📝 Summary:
VideoFlexTok enables efficient video representation through variable-length token sequences that capture abstract information first, followed by fine-grained details, allowing for reduced computationa...
🔹 Publication Date: Published on Apr 14
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.12887
• PDF: https://arxiv.org/pdf/2604.12887
• Github: https://github.com/apple/ml-videoflextok
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
VideoFlexTok enables efficient video representation through variable-length token sequences that capture abstract information first, followed by fine-grained details, allowing for reduced computationa...
🔹 Publication Date: Published on Apr 14
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.12887
• PDF: https://arxiv.org/pdf/2604.12887
• Github: https://github.com/apple/ml-videoflextok
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
❤1
✨PersonaVLM: Long-Term Personalized Multimodal LLMs
📝 Summary:
PersonaVLM introduces a framework for long-term personalized multimodal LLMs. It remembers interactions, reasons multi-turn using retrieved memories, and aligns responses with evolving user personality. This novel method significantly outperforms baselines and GPT-4o on a new evaluation benchmark.
🔹 Publication Date: Published on Mar 20
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.13074
• PDF: https://arxiv.org/pdf/2604.13074
• Project Page: https://personavlm.github.io/
• Github: https://github.com/MiG-NJU/PersonaVLM
🔹 Models citing this paper:
• https://huggingface.co/ClareNie/PersonaVLM
✨ Datasets citing this paper:
• https://huggingface.co/datasets/ClareNie/Persona-MME
• https://huggingface.co/datasets/ClareNie/PersonaVLM-Dataset
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LLM #MultimodalAI #PersonalizedAI #AIResearch #MemoryAI
📝 Summary:
PersonaVLM introduces a framework for long-term personalized multimodal LLMs. It remembers interactions, reasons multi-turn using retrieved memories, and aligns responses with evolving user personality. This novel method significantly outperforms baselines and GPT-4o on a new evaluation benchmark.
🔹 Publication Date: Published on Mar 20
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.13074
• PDF: https://arxiv.org/pdf/2604.13074
• Project Page: https://personavlm.github.io/
• Github: https://github.com/MiG-NJU/PersonaVLM
🔹 Models citing this paper:
• https://huggingface.co/ClareNie/PersonaVLM
✨ Datasets citing this paper:
• https://huggingface.co/datasets/ClareNie/Persona-MME
• https://huggingface.co/datasets/ClareNie/PersonaVLM-Dataset
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LLM #MultimodalAI #PersonalizedAI #AIResearch #MemoryAI
✨VEFX-Bench: A Holistic Benchmark for Generic Video Editing and Visual Effects
📝 Summary:
VEFX-Bench offers a large human-annotated video editing dataset and VEFX-Reward, a specialized model for quality assessment. This benchmark allows standardized comparison, showing current models struggle with instruction following and edit locality.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16272
• PDF: https://arxiv.org/pdf/2604.16272
• Project Page: https://xiangbogaobarry.github.io/VEFX-Bench/
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#VideoEditing #VFX #AI #ComputerVision #Benchmarks
📝 Summary:
VEFX-Bench offers a large human-annotated video editing dataset and VEFX-Reward, a specialized model for quality assessment. This benchmark allows standardized comparison, showing current models struggle with instruction following and edit locality.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16272
• PDF: https://arxiv.org/pdf/2604.16272
• Project Page: https://xiangbogaobarry.github.io/VEFX-Bench/
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#VideoEditing #VFX #AI #ComputerVision #Benchmarks
✨Qwen3.5-Omni Technical Report
📝 Summary:
Qwen3.5-Omni is a large multimodal model excelling in audio-visual understanding and generation, achieving SOTA results across many benchmarks. It features a Hybrid Attention MoE architecture, introduces ARIA for improved speech synthesis, and exhibits a new Audio-Visual Vibe Coding capability.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15804
• PDF: https://arxiv.org/pdf/2604.15804
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#MultimodalAI #AIResearch #DeepLearning #GenerativeAI #SpeechSynthesis
📝 Summary:
Qwen3.5-Omni is a large multimodal model excelling in audio-visual understanding and generation, achieving SOTA results across many benchmarks. It features a Hybrid Attention MoE architecture, introduces ARIA for improved speech synthesis, and exhibits a new Audio-Visual Vibe Coding capability.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15804
• PDF: https://arxiv.org/pdf/2604.15804
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#MultimodalAI #AIResearch #DeepLearning #GenerativeAI #SpeechSynthesis
✨ArtifactNet: Detecting AI-Generated Music via Forensic Residual Physics
📝 Summary:
ArtifactNet detects AI-generated music by analyzing codec-specific artifacts in audio signals using a lightweight neural network and codec-aware training. It achieves superior performance and efficiency compared to existing methods, establishing forensic physics as a new detection paradigm.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16254
• PDF: https://arxiv.org/pdf/2604.16254
• Project Page: https://demo.intrect.io
🔹 Models citing this paper:
• https://huggingface.co/intrect/artifactnet
✨ Datasets citing this paper:
• https://huggingface.co/datasets/intrect/artifactbench
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #MachineLearning #AIMusic #DigitalForensics #AudioProcessing
📝 Summary:
ArtifactNet detects AI-generated music by analyzing codec-specific artifacts in audio signals using a lightweight neural network and codec-aware training. It achieves superior performance and efficiency compared to existing methods, establishing forensic physics as a new detection paradigm.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16254
• PDF: https://arxiv.org/pdf/2604.16254
• Project Page: https://demo.intrect.io
🔹 Models citing this paper:
• https://huggingface.co/intrect/artifactnet
✨ Datasets citing this paper:
• https://huggingface.co/datasets/intrect/artifactbench
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #MachineLearning #AIMusic #DigitalForensics #AudioProcessing
✨PRL-Bench: A Comprehensive Benchmark Evaluating LLMs' Capabilities in Frontier Physics Research
📝 Summary:
PRL-Bench is a new benchmark evaluating LLMs' end-to-end capabilities in theoretical and computational physics research. It uses 100 curated papers to assess exploration-oriented, long-horizon workflows. Current LLMs perform poorly, revealing a significant gap in autonomous scientific discovery.
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15411
• PDF: https://arxiv.org/pdf/2604.15411
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LLMs #PhysicsResearch #ScientificDiscovery #AI #Benchmarking
📝 Summary:
PRL-Bench is a new benchmark evaluating LLMs' end-to-end capabilities in theoretical and computational physics research. It uses 100 curated papers to assess exploration-oriented, long-horizon workflows. Current LLMs perform poorly, revealing a significant gap in autonomous scientific discovery.
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15411
• PDF: https://arxiv.org/pdf/2604.15411
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LLMs #PhysicsResearch #ScientificDiscovery #AI #Benchmarking
✨Cut Your Losses! Learning to Prune Paths Early for Efficient Parallel Reasoning
📝 Summary:
STOP is a systematic, learnable token-level path pruning method for Large Reasoning Models. It improves efficiency and accuracy, outperforming baselines and scaling across compute budgets to reduce futile paths in parallel reasoning.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16029
• PDF: https://arxiv.org/pdf/2604.16029
• Project Page: https://bijiaxihh.github.io/STOP/
• Github: https://github.com/bijiaxihh/STOP
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #LLM #MachineLearning #ParallelReasoning #ModelEfficiency
📝 Summary:
STOP is a systematic, learnable token-level path pruning method for Large Reasoning Models. It improves efficiency and accuracy, outperforming baselines and scaling across compute budgets to reduce futile paths in parallel reasoning.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16029
• PDF: https://arxiv.org/pdf/2604.16029
• Project Page: https://bijiaxihh.github.io/STOP/
• Github: https://github.com/bijiaxihh/STOP
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #LLM #MachineLearning #ParallelReasoning #ModelEfficiency
✨GTA-2: Benchmarking General Tool Agents from Atomic Tool-Use to Open-Ended Workflows
📝 Summary:
GTA-2 is a new benchmark for General Tool Agents, covering both atomic and real-world, open-ended workflows. It shows frontier models struggle significantly, especially on workflows. The study emphasizes that execution frameworks are crucial for performance, more so than just model capacity.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15715
• PDF: https://arxiv.org/pdf/2604.15715
• Github: https://github.com/open-compass/GTA
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AIAgents #BenchmarkingAI #LLMs #AIWorkflows #AIResearch
📝 Summary:
GTA-2 is a new benchmark for General Tool Agents, covering both atomic and real-world, open-ended workflows. It shows frontier models struggle significantly, especially on workflows. The study emphasizes that execution frameworks are crucial for performance, more so than just model capacity.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.15715
• PDF: https://arxiv.org/pdf/2604.15715
• Github: https://github.com/open-compass/GTA
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AIAgents #BenchmarkingAI #LLMs #AIWorkflows #AIResearch
✨Learning Adaptive Reasoning Paths for Efficient Visual Reasoning
📝 Summary:
Existing visual reasoning models often overthink, using redundant steps. AVR is an adaptive framework that dynamically chooses efficient reasoning formats. It reduces token usage by 50-90 percent while maintaining accuracy.
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14568
• PDF: https://arxiv.org/pdf/2604.14568
• Github: https://github.com/RunRiotComeOn/AVR
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#VisualReasoning #AI #MachineLearning #Efficiency #DeepLearning
📝 Summary:
Existing visual reasoning models often overthink, using redundant steps. AVR is an adaptive framework that dynamically chooses efficient reasoning formats. It reduces token usage by 50-90 percent while maintaining accuracy.
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.14568
• PDF: https://arxiv.org/pdf/2604.14568
• Github: https://github.com/RunRiotComeOn/AVR
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#VisualReasoning #AI #MachineLearning #Efficiency #DeepLearning
This media is not supported in your browser
VIEW IN TELEGRAM
✨Repurposing 3D Generative Model for Autoregressive Layout Generation
📝 Summary:
LaviGen is a 3D layout generation framework that repurposes 3D generative models. It uses an adapted 3D diffusion model for autoregressive generation, explicitly modeling geometric relations and physical constraints. This achieves superior, more plausible 3D layouts 65% faster than previous methods.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16299
• PDF: https://arxiv.org/pdf/2604.16299
• Project Page: https://fenghora.github.io/LaviGen-Page/
• Github: https://github.com/fenghora/LaviGen
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#3DGeneration #DiffusionModels #GenerativeAI #ComputerGraphics #DeepLearning
📝 Summary:
LaviGen is a 3D layout generation framework that repurposes 3D generative models. It uses an adapted 3D diffusion model for autoregressive generation, explicitly modeling geometric relations and physical constraints. This achieves superior, more plausible 3D layouts 65% faster than previous methods.
🔹 Publication Date: Published on Apr 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.16299
• PDF: https://arxiv.org/pdf/2604.16299
• Project Page: https://fenghora.github.io/LaviGen-Page/
• Github: https://github.com/fenghora/LaviGen
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#3DGeneration #DiffusionModels #GenerativeAI #ComputerGraphics #DeepLearning
✨Web Retrieval-Aware Chunking (W-RAC) for Efficient and Cost-Effective Retrieval-Augmented Generation Systems
📝 Summary:
Web Retrieval-Aware Chunking (W-RAC) introduces a cost-efficient framework for web document processing that reduces LLM token usage and hallucination risks through structured content representation an...
🔹 Publication Date: Published on Jan 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.04936
• PDF: https://arxiv.org/pdf/2604.04936
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
Web Retrieval-Aware Chunking (W-RAC) introduces a cost-efficient framework for web document processing that reduces LLM token usage and hallucination risks through structured content representation an...
🔹 Publication Date: Published on Jan 8
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2604.04936
• PDF: https://arxiv.org/pdf/2604.04936
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
✨Maximal Brain Damage Without Data or Optimization: Disrupting Neural Networks via Sign-Bit Flips
📝 Summary:
Deep neural networks exhibit catastrophic vulnerability to minimal parameter bit flips across multiple domains, which can be identified and mitigated through targeted protection strategies. AI-generat...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2502.07408
• PDF: https://arxiv.org/pdf/2502.07408
• Project Page: https://mkimhi.github.io/DNL/
• Github: https://github.com/IdoGalil/maximal-brain-damage
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
Deep neural networks exhibit catastrophic vulnerability to minimal parameter bit flips across multiple domains, which can be identified and mitigated through targeted protection strategies. AI-generat...
🔹 Publication Date: Published on Apr 16
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2502.07408
• PDF: https://arxiv.org/pdf/2502.07408
• Project Page: https://mkimhi.github.io/DNL/
• Github: https://github.com/IdoGalil/maximal-brain-damage
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
✨AccelOpt: A Self-Improving LLM Agentic System for AI Accelerator Kernel Optimization
📝 Summary:
AccelOpt is a self-improving LLM agentic system that autonomously optimizes kernels for AI accelerators using iterative generation and optimization memory, achieving significant throughput improvement...
🔹 Publication Date: Published on Apr 15
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.15915
• PDF: https://arxiv.org/pdf/2511.15915
• Project Page: https://ppl.stanford.edu/accelopt.html
• Github: https://github.com/zhang677/AccelOpt
🔹 Models citing this paper:
• https://huggingface.co/Genghan/sft-qwen-7b-instruct_GRPO_nki_pure_0920_cluster3
• https://huggingface.co/Genghan/deepseek-coder-33b-instruct_GRPO_nki_pure_0907_cluster1
• https://huggingface.co/Genghan/sft-deepseek-coder-33b-instruct_GRPO_nki_pure_0921_cluster4
✨ Datasets citing this paper:
• https://huggingface.co/datasets/Genghan/NKIBench
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
📝 Summary:
AccelOpt is a self-improving LLM agentic system that autonomously optimizes kernels for AI accelerators using iterative generation and optimization memory, achieving significant throughput improvement...
🔹 Publication Date: Published on Apr 15
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.15915
• PDF: https://arxiv.org/pdf/2511.15915
• Project Page: https://ppl.stanford.edu/accelopt.html
• Github: https://github.com/zhang677/AccelOpt
🔹 Models citing this paper:
• https://huggingface.co/Genghan/sft-qwen-7b-instruct_GRPO_nki_pure_0920_cluster3
• https://huggingface.co/Genghan/deepseek-coder-33b-instruct_GRPO_nki_pure_0907_cluster1
• https://huggingface.co/Genghan/sft-deepseek-coder-33b-instruct_GRPO_nki_pure_0921_cluster4
✨ Datasets citing this paper:
• https://huggingface.co/datasets/Genghan/NKIBench
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #DataScience #MachineLearning #HuggingFace #Research
arXiv.org
AccelOpt: A Self-Improving LLM Agentic System for AI Accelerator...
We present AccelOpt, a self-improving large language model (LLM) agentic system that autonomously optimizes kernels for emerging AI acclerators, eliminating the need for expert-provided...