ML Research Hub
32.9K subscribers
5.36K photos
335 videos
24 files
5.8K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
Nemotron-Flash: Towards Latency-Optimal Hybrid Small Language Models

📝 Summary:
This study optimizes small language models for real-device latency by identifying key architectural factors and efficient operators. It introduces Nemotron-Flash, a new family of hybrid SLMs that significantly improves accuracy, latency, and throughput compared to current models.

🔹 Publication Date: Published on Nov 24

🔹 Paper Links:
• arXiv Page: https://arxiv.org/pdf/2511.18890
• PDF: https://arxiv.org/pdf/2511.18890

🔹 Models citing this paper:
https://huggingface.co/nvidia/Nemotron-Flash-3B-Instruct
https://huggingface.co/nvidia/Nemotron-Flash-1B
https://huggingface.co/nvidia/Nemotron-Flash-3B

==================================

For more data science resources:
https://t.iss.one/DataScienceT

#SmallLanguageModels #LatencyOptimization #AI #DeepLearning #NLP
1
Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts

📝 Summary:
Nanbeige4.1-3B is a 3B-parameter model excelling in agentic behavior, code generation, and reasoning. It outperforms larger models through advanced reward modeling and training, demonstrating broad competence for a small language model.

🔹 Publication Date: Published on Feb 13

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2602.13367
• PDF: https://arxiv.org/pdf/2602.13367
• Project Page: https://huggingface.co/Nanbeige/Nanbeige4.1-3B

🔹 Models citing this paper:
https://huggingface.co/Nanbeige/Nanbeige4.1-3B

Spaces citing this paper:
https://huggingface.co/spaces/PioTio/AIMan

==================================

For more data science resources:
https://t.iss.one/DataScienceT

#LLM #AI #SmallLanguageModels #AgenticAI #CodeGeneration
1