✨Dense Motion Captioning
📝 Summary:
The paper introduces Dense Motion Captioning, a new task for 3D human motion understanding. It presents CompMo, a large dataset with complex, temporally annotated motions, and DEMO, a model combining a language model with a motion adapter to generate detailed, grounded captions.
🔹 Publication Date: Published on Nov 7
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.05369
• PDF: https://arxiv.org/pdf/2511.05369
• Project Page: https://xusy2333.com/demo/
• Github: https://github.com/41xu/DEMO
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#MotionCaptioning #3DMotion #ComputerVision #LanguageModels #AIResearch
📝 Summary:
The paper introduces Dense Motion Captioning, a new task for 3D human motion understanding. It presents CompMo, a large dataset with complex, temporally annotated motions, and DEMO, a model combining a language model with a motion adapter to generate detailed, grounded captions.
🔹 Publication Date: Published on Nov 7
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.05369
• PDF: https://arxiv.org/pdf/2511.05369
• Project Page: https://xusy2333.com/demo/
• Github: https://github.com/41xu/DEMO
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#MotionCaptioning #3DMotion #ComputerVision #LanguageModels #AIResearch
✨Llama-Embed-Nemotron-8B: A Universal Text Embedding Model for Multilingual and Cross-Lingual Tasks
📝 Summary:
Llama-Embed-Nemotron-8B is an open-source text embedding model achieving state-of-the-art performance, especially in multilingual tasks. Its success comes from a novel data mix and detailed ablation studies, making it a universal solution.
🔹 Publication Date: Published on Nov 10
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.07025
• PDF: https://arxiv.org/pdf/2511.07025
🔹 Models citing this paper:
• https://huggingface.co/nvidia/llama-embed-nemotron-8b
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#TextEmbeddings #MultilingualNLP #CrossLingual #LanguageModels #AIResearch
📝 Summary:
Llama-Embed-Nemotron-8B is an open-source text embedding model achieving state-of-the-art performance, especially in multilingual tasks. Its success comes from a novel data mix and detailed ablation studies, making it a universal solution.
🔹 Publication Date: Published on Nov 10
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.07025
• PDF: https://arxiv.org/pdf/2511.07025
🔹 Models citing this paper:
• https://huggingface.co/nvidia/llama-embed-nemotron-8b
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#TextEmbeddings #MultilingualNLP #CrossLingual #LanguageModels #AIResearch
✨Beyond Outlining: Heterogeneous Recursive Planning for Adaptive Long-form Writing with Language Models
📝 Summary:
This paper proposes an AI agent framework for adaptive long-form writing. It uses recursive task decomposition and dynamically integrates retrieval, reasoning, and composition, overcoming rigid outline-based methods. The framework consistently outperforms state-of-the-art approaches.
🔹 Publication Date: Published on Mar 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2503.08275
• PDF: https://arxiv.org/pdf/2503.08275
• Github: https://github.com/principia-ai/WriteHERE
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #LanguageModels #LongformWriting #NLP #GenerativeAI
📝 Summary:
This paper proposes an AI agent framework for adaptive long-form writing. It uses recursive task decomposition and dynamically integrates retrieval, reasoning, and composition, overcoming rigid outline-based methods. The framework consistently outperforms state-of-the-art approaches.
🔹 Publication Date: Published on Mar 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2503.08275
• PDF: https://arxiv.org/pdf/2503.08275
• Github: https://github.com/principia-ai/WriteHERE
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #LanguageModels #LongformWriting #NLP #GenerativeAI
❤1
✨AraLingBench A Human-Annotated Benchmark for Evaluating Arabic Linguistic Capabilities of Large Language Models
📝 Summary:
AraLingBench is a human-annotated benchmark evaluating Arabic LLM linguistic competence using expert-designed questions. It reveals models achieve surface proficiency but lack deep understanding, often relying on memorization rather than true comprehension.
🔹 Publication Date: Published on Nov 18
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.14295
• PDF: https://arxiv.org/pdf/2511.14295
✨ Datasets citing this paper:
• https://huggingface.co/datasets/hammh0a/AraLingBench
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#ArabicNLP #LLMEvaluation #AIResearch #LanguageModels #NLPBenchmarking
📝 Summary:
AraLingBench is a human-annotated benchmark evaluating Arabic LLM linguistic competence using expert-designed questions. It reveals models achieve surface proficiency but lack deep understanding, often relying on memorization rather than true comprehension.
🔹 Publication Date: Published on Nov 18
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.14295
• PDF: https://arxiv.org/pdf/2511.14295
✨ Datasets citing this paper:
• https://huggingface.co/datasets/hammh0a/AraLingBench
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#ArabicNLP #LLMEvaluation #AIResearch #LanguageModels #NLPBenchmarking
This media is not supported in your browser
VIEW IN TELEGRAM
✨Computer-Use Agents as Judges for Generative User Interface
📝 Summary:
This paper introduces a framework where Computer-Use Agents CUA act as judges for coding language models Coder to automatically design GUIs. The goal is to optimize interfaces for CUA efficiency and task solvability, rather than human aesthetics, using a new benchmark called AUI-Gym.
🔹 Publication Date: Published on Nov 19
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.15567
• PDF: https://arxiv.org/pdf/2511.15567
• Project Page: https://showlab.github.io/AUI/
• Github: https://github.com/showlab/AUI/
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AIAgents #GUIDesign #GenerativeAI #AIevaluation #LanguageModels
📝 Summary:
This paper introduces a framework where Computer-Use Agents CUA act as judges for coding language models Coder to automatically design GUIs. The goal is to optimize interfaces for CUA efficiency and task solvability, rather than human aesthetics, using a new benchmark called AUI-Gym.
🔹 Publication Date: Published on Nov 19
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.15567
• PDF: https://arxiv.org/pdf/2511.15567
• Project Page: https://showlab.github.io/AUI/
• Github: https://github.com/showlab/AUI/
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AIAgents #GUIDesign #GenerativeAI #AIevaluation #LanguageModels
✨AICC: Parse HTML Finer, Make Models Better -- A 7.3T AI-Ready Corpus Built by a Model-Based HTML Parser
📝 Summary:
This paper introduces MinerU-HTML, a novel language model-based HTML parser that semantically extracts web content, preserving structure better than heuristic methods. It constructs the 7.3T AICC corpus, demonstrating that models trained on AICC significantly outperform those from other parsers, ...
🔹 Publication Date: Published on Nov 20
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.16397
• PDF: https://arxiv.org/pdf/2511.16397
✨ Datasets citing this paper:
• https://huggingface.co/datasets/opendatalab/AICC
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #HTMLParsing #Corpus #LanguageModels #WebData
📝 Summary:
This paper introduces MinerU-HTML, a novel language model-based HTML parser that semantically extracts web content, preserving structure better than heuristic methods. It constructs the 7.3T AICC corpus, demonstrating that models trained on AICC significantly outperform those from other parsers, ...
🔹 Publication Date: Published on Nov 20
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.16397
• PDF: https://arxiv.org/pdf/2511.16397
✨ Datasets citing this paper:
• https://huggingface.co/datasets/opendatalab/AICC
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#AI #HTMLParsing #Corpus #LanguageModels #WebData
✨Beyond Multiple Choice: Verifiable OpenQA for Robust Vision-Language RFT
📝 Summary:
ReVeL converts multiple-choice questions to verifiable open-form questions to address unreliable MCQA metrics and answer guessing. This framework improves data efficiency and robustness for multimodal language models, revealing significant score inflation in MCQA benchmarks.
🔹 Publication Date: Published on Nov 21
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.17405
• PDF: https://arxiv.org/pdf/2511.17405
• Github: https://flageval-baai.github.io/ReVeL/
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#OpenQA #VisionLanguage #LanguageModels #AIEvaluation #MachineLearning
📝 Summary:
ReVeL converts multiple-choice questions to verifiable open-form questions to address unreliable MCQA metrics and answer guessing. This framework improves data efficiency and robustness for multimodal language models, revealing significant score inflation in MCQA benchmarks.
🔹 Publication Date: Published on Nov 21
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.17405
• PDF: https://arxiv.org/pdf/2511.17405
• Github: https://flageval-baai.github.io/ReVeL/
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#OpenQA #VisionLanguage #LanguageModels #AIEvaluation #MachineLearning
✨Xmodel-2.5: 1.3B Data-Efficient Reasoning SLM
📝 Summary:
Xmodel-2.5 is a 1.3B language model designed for efficient edge deployments. It uses maximal-update parameterization and a novel training curriculum that switches from AdamW to Muon, improving reasoning skills by 4.58% while maintaining efficiency.
🔹 Publication Date: Published on Nov 23
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.19496
• PDF: https://arxiv.org/pdf/2511.19496
• Github: https://github.com/XiaoduoAILab/Xmodel-2.5
🔹 Models citing this paper:
• https://huggingface.co/XiaoduoAILab/Xmodel-2.5
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#SLM #EdgeAI #LanguageModels #DeepLearning #ReasoningAI
📝 Summary:
Xmodel-2.5 is a 1.3B language model designed for efficient edge deployments. It uses maximal-update parameterization and a novel training curriculum that switches from AdamW to Muon, improving reasoning skills by 4.58% while maintaining efficiency.
🔹 Publication Date: Published on Nov 23
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.19496
• PDF: https://arxiv.org/pdf/2511.19496
• Github: https://github.com/XiaoduoAILab/Xmodel-2.5
🔹 Models citing this paper:
• https://huggingface.co/XiaoduoAILab/Xmodel-2.5
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#SLM #EdgeAI #LanguageModels #DeepLearning #ReasoningAI
❤1
✨Masks Can Be Distracting: On Context Comprehension in Diffusion Language Models
📝 Summary:
Masked Diffusion Language Models MDLMs show locality bias and poor context comprehension due to appended mask tokens acting as distractors. A mask-agnostic loss function was introduced. This function improves MDLM robustness by mitigating the masks distracting effect.
🔹 Publication Date: Published on Nov 26
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.21338
• PDF: https://arxiv.org/pdf/2511.21338
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LanguageModels #DiffusionModels #NLP #ContextComprehension #AIResearch
📝 Summary:
Masked Diffusion Language Models MDLMs show locality bias and poor context comprehension due to appended mask tokens acting as distractors. A mask-agnostic loss function was introduced. This function improves MDLM robustness by mitigating the masks distracting effect.
🔹 Publication Date: Published on Nov 26
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.21338
• PDF: https://arxiv.org/pdf/2511.21338
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LanguageModels #DiffusionModels #NLP #ContextComprehension #AIResearch
❤1
✨Scaling Behavior of Discrete Diffusion Language Models
📝 Summary:
Research on discrete diffusion language models DLMs shows their scaling behavior depends on noise type. Uniform diffusion is more parameter and data efficient than masked diffusion, making it promising for data-bound settings. A 10B parameter model confirmed this.
🔹 Publication Date: Published on Dec 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.10858
• PDF: https://arxiv.org/pdf/2512.10858
• Github: https://github.com/dvruette/gidd-easydel
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#DiffusionModels #LanguageModels #NLP #AIResearch #DeepLearning
📝 Summary:
Research on discrete diffusion language models DLMs shows their scaling behavior depends on noise type. Uniform diffusion is more parameter and data efficient than masked diffusion, making it promising for data-bound settings. A 10B parameter model confirmed this.
🔹 Publication Date: Published on Dec 11
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.10858
• PDF: https://arxiv.org/pdf/2512.10858
• Github: https://github.com/dvruette/gidd-easydel
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#DiffusionModels #LanguageModels #NLP #AIResearch #DeepLearning
❤1
✨Physics of Language Models: Part 4.1, Architecture Design and the Magic of Canon Layers
📝 Summary:
Canon layers are lightweight architectural components that enhance language model reasoning depth and breadth by promoting horizontal information flow. They improve performance across various architectures, validated in synthetic tasks and real-world pretraining.
🔹 Publication Date: Published on Dec 19
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.17351
• PDF: https://arxiv.org/pdf/2512.17351
• Project Page: https://physics.allen-zhu.com/part-4-architecture-design/part-4-1
• Github: https://github.com/facebookresearch/PhysicsLM4
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LanguageModels #LLM #AIArchitecture #DeepLearning #NLP
📝 Summary:
Canon layers are lightweight architectural components that enhance language model reasoning depth and breadth by promoting horizontal information flow. They improve performance across various architectures, validated in synthetic tasks and real-world pretraining.
🔹 Publication Date: Published on Dec 19
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.17351
• PDF: https://arxiv.org/pdf/2512.17351
• Project Page: https://physics.allen-zhu.com/part-4-architecture-design/part-4-1
• Github: https://github.com/facebookresearch/PhysicsLM4
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LanguageModels #LLM #AIArchitecture #DeepLearning #NLP
❤1
✨Bolmo: Byteifying the Next Generation of Language Models
📝 Summary:
Bolmo introduces competitive byte-level language models by efficiently converting existing subword models. This byteification overcomes subword limitations, matching performance with minimal training. Bolmo makes byte-level LMs practical.
🔹 Publication Date: Published on Dec 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.15586
• PDF: https://arxiv.org/pdf/2512.15586
🔹 Models citing this paper:
• https://huggingface.co/allenai/Bolmo-7B
• https://huggingface.co/allenai/Bolmo-1B
✨ Datasets citing this paper:
• https://huggingface.co/datasets/allenai/bolmo_mix
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LanguageModels #ByteLevelLMs #NLP #DeepLearning #AIResearch
📝 Summary:
Bolmo introduces competitive byte-level language models by efficiently converting existing subword models. This byteification overcomes subword limitations, matching performance with minimal training. Bolmo makes byte-level LMs practical.
🔹 Publication Date: Published on Dec 17
🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.15586
• PDF: https://arxiv.org/pdf/2512.15586
🔹 Models citing this paper:
• https://huggingface.co/allenai/Bolmo-7B
• https://huggingface.co/allenai/Bolmo-1B
✨ Datasets citing this paper:
• https://huggingface.co/datasets/allenai/bolmo_mix
==================================
For more data science resources:
✓ https://t.iss.one/DataScienceT
#LanguageModels #ByteLevelLMs #NLP #DeepLearning #AIResearch
❤1