ML Research Hub
32.8K subscribers
4.13K photos
244 videos
23 files
4.46K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B

📝 Summary:
VibeThinker-1.5B, a 1.5B-parameter model, uses the Spectrum-to-Signal Principle to achieve superior reasoning. It outperforms much larger models on math and coding benchmarks, proving small models can deliver advanced AI at low cost.

🔹 Publication Date: Published on Nov 9

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.06221
• PDF: https://arxiv.org/pdf/2511.06221
• Github: https://github.com/WeiboAI/VibeThinker

🔹 Models citing this paper:
https://huggingface.co/WeiboAI/VibeThinker-1.5B
https://huggingface.co/Mungert/VibeThinker-1.5B-GGUF

==================================

For more data science resources:
https://t.iss.one/DataScienceT

#SLM #AIReasoning #ModelOptimization #MachineLearning #EfficientAI
DiRL: An Efficient Post-Training Framework for Diffusion Language Models

📝 Summary:
DiRL is an efficient post-training framework for Diffusion Language Models, integrating online updates and introducing DiPO for unbiased policy optimization. It achieves state-of-the-art math performance for dLLMs, surpassing comparable models.

🔹 Publication Date: Published on Dec 23

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.22234
• PDF: https://arxiv.org/pdf/2512.22234
• Github: https://github.com/OpenMOSS/DiRL

🔹 Models citing this paper:
https://huggingface.co/OpenMOSS-Team/DiRL-8B-Instruct

==================================

For more data science resources:
https://t.iss.one/DataScienceT

#DiffusionModels #LLM #ModelOptimization #MachineLearning #AI
FlowBlending: Stage-Aware Multi-Model Sampling for Fast and High-Fidelity Video Generation

📝 Summary:
FlowBlending optimizes video generation by adapting model capacity to each stage. It uses large models for critical early and late timesteps, and small models for intermediate ones. This achieves faster inference and fewer FLOPs with no loss in large model fidelity.

🔹 Publication Date: Published on Dec 31, 2025

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.24724
• PDF: https://arxiv.org/pdf/2512.24724

==================================

For more data science resources:
https://t.iss.one/DataScienceT

#VideoGeneration #GenerativeAI #DeepLearning #AIResearch #ModelOptimization