Hugging Face (Twitter)
RT @Meituan_LongCat: π LongCat-Flash-Chat Launches!
β«οΈ 560B Total Params | 18.6B-31.3B Dynamic Activation
β«οΈ Trained on 20T Tokens | 100+ tokens/sec Inference
β«οΈ High Performance: TerminalBench 39.5 | ΟΒ²-Bench 67.7
π Model: https://huggingface.co/meituan-longcat/LongCat-Flash-Chat
π» Try Now: longcat.ai
RT @Meituan_LongCat: π LongCat-Flash-Chat Launches!
β«οΈ 560B Total Params | 18.6B-31.3B Dynamic Activation
β«οΈ Trained on 20T Tokens | 100+ tokens/sec Inference
β«οΈ High Performance: TerminalBench 39.5 | ΟΒ²-Bench 67.7
π Model: https://huggingface.co/meituan-longcat/LongCat-Flash-Chat
π» Try Now: longcat.ai
This media is not supported in your browser
VIEW IN TELEGRAM
Hugging Face (Twitter)
RT @HuggingPapers: ByteDance Seed and Stanford introduce Mixture of Contexts (MoC) for long video generation, tackling the memory bottleneck with a novel sparse attention routing module.
It enables minute-long consistent videos with short-video cost.
RT @HuggingPapers: ByteDance Seed and Stanford introduce Mixture of Contexts (MoC) for long video generation, tackling the memory bottleneck with a novel sparse attention routing module.
It enables minute-long consistent videos with short-video cost.
βHugging Face (Twitter)
RT @AdinaYakup: Hunyuan-MT-7B π₯ open translation model released by @TencentHunyuan
https://huggingface.co/collections/tencent/hunyuan-mt-68b42f76d473f82798882597
β¨ Supports 33 languages, including 5 ethnic minority languages in China π
β¨ Including a translation ensemble model: Chimera-7B
β¨ Full pipeline: pretrain > CPT > SFT > enhancement > ensemble refinement > SOTA performance at similar scale
RT @AdinaYakup: Hunyuan-MT-7B π₯ open translation model released by @TencentHunyuan
https://huggingface.co/collections/tencent/hunyuan-mt-68b42f76d473f82798882597
β¨ Supports 33 languages, including 5 ethnic minority languages in China π
β¨ Including a translation ensemble model: Chimera-7B
β¨ Full pipeline: pretrain > CPT > SFT > enhancement > ensemble refinement > SOTA performance at similar scale
huggingface.co
Hunyuan-MT - a tencent Collection
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
Hugging Face (Twitter)
RT @multimodalart: a mysterious new button appeared on the @huggingface Spaces Nano Banana app π
RT @multimodalart: a mysterious new button appeared on the @huggingface Spaces Nano Banana app π
Hugging Face (Twitter)
RT @reach_vb: that's a Chinese food delivery company absolutely mogging the competition https://twitter.com/reach_vb/status/1961833208737103997#m
RT @reach_vb: that's a Chinese food delivery company absolutely mogging the competition https://twitter.com/reach_vb/status/1961833208737103997#m
βHugging Face (Twitter)
RT @NLPingu: Hugging Faceγ§γε ¬ιγγΎγγγ
https://twitter.com/NLPingu/status/1962414807229034711#m
RT @NLPingu: Hugging Faceγ§γε ¬ιγγΎγγγ
https://twitter.com/NLPingu/status/1962414807229034711#m
huggingface.co
llm-jp/jgpqa Β· Datasets at Hugging Face
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
Hugging Face (Twitter)
RT @MaziyarPanahi: need your help! list your top 5 datasets on @huggingface for rl training with verified answers.
- math
- code
- everyday stuff
RT @MaziyarPanahi: need your help! list your top 5 datasets on @huggingface for rl training with verified answers.
- math
- code
- everyday stuff
Hugging Face (Twitter)
RT @MaziyarPanahi: 1/ shipping two synthetic med qa sets from @OpenMed_AI community, made by @mkurman88 (core contributor):
β’ med-synth qwen3-235b-a22b (2507)
β’ med-synth gemma 3 (27b-it)
datasets on @huggingface π
RT @MaziyarPanahi: 1/ shipping two synthetic med qa sets from @OpenMed_AI community, made by @mkurman88 (core contributor):
β’ med-synth qwen3-235b-a22b (2507)
β’ med-synth gemma 3 (27b-it)
datasets on @huggingface π
This media is not supported in your browser
VIEW IN TELEGRAM
Hugging Face (Twitter)
RT @reach_vb: BOOM! Microsoft just released an upgraded VibeVoice Large ~10B Text to Speech model - MIT licensed π₯
> Generate multi-speaker podcasts in minutes β‘
> Works blazingly fast on ZeroGPU with H200 (FREE)
Try it out today! https://twitter.com/reach_vb/status/1960064616278417826#m
RT @reach_vb: BOOM! Microsoft just released an upgraded VibeVoice Large ~10B Text to Speech model - MIT licensed π₯
> Generate multi-speaker podcasts in minutes β‘
> Works blazingly fast on ZeroGPU with H200 (FREE)
Try it out today! https://twitter.com/reach_vb/status/1960064616278417826#m
This media is not supported in your browser
VIEW IN TELEGRAM
Hugging Face (Twitter)
RT @ClementDelangue: If you think @Apple is not doing much in AI, you're getting blindsided by the chatbot hype and not paying enough attention!
They just released FastVLM and MobileCLIP2 on @huggingface. The models are up to 85x faster and 3.4x smaller than previous work, enabling real-time vision language model (VLM) applications! It can even do live video captioning 100% locally in your browser π€―π€―π€―
RT @ClementDelangue: If you think @Apple is not doing much in AI, you're getting blindsided by the chatbot hype and not paying enough attention!
They just released FastVLM and MobileCLIP2 on @huggingface. The models are up to 85x faster and 3.4x smaller than previous work, enabling real-time vision language model (VLM) applications! It can even do live video captioning 100% locally in your browser π€―π€―π€―
Hugging Face (Twitter)
RT @eliebakouch: Super excited to announce that our research team at @huggingface will be doing an AMA on r/LocalLLaMA.
Come ask any questions to the team behind SmolLM, FineWeb and more! And who knows, maybe thereβll be a shiny new release to talk about?
Thursday 4th September, 8AM-11AM PST π€
RT @eliebakouch: Super excited to announce that our research team at @huggingface will be doing an AMA on r/LocalLLaMA.
Come ask any questions to the team behind SmolLM, FineWeb and more! And who knows, maybe thereβll be a shiny new release to talk about?
Thursday 4th September, 8AM-11AM PST π€
This media is not supported in your browser
VIEW IN TELEGRAM
Hugging Face (Twitter)
RT @reach_vb: π¬ One prompt β a full video
GPT-5 + open models, stitched together with @OpenAI Codex + HF MCP Server π€―
RT @reach_vb: π¬ One prompt β a full video
GPT-5 + open models, stitched together with @OpenAI Codex + HF MCP Server π€―
Hugging Face (Twitter)
RT @RisingSayak: ZeroGPU on π€ HF Spaces enables anyone to build delightful ML demos, benefitting from powerful compute. But, due to its serverless nature, it is hard to optimize these demos.
That CHANGES today πͺ
Use AoT compilation to melt our ZeroGPU servers π₯
Details β¬οΈ
RT @RisingSayak: ZeroGPU on π€ HF Spaces enables anyone to build delightful ML demos, benefitting from powerful compute. But, due to its serverless nature, it is hard to optimize these demos.
That CHANGES today πͺ
Use AoT compilation to melt our ZeroGPU servers π₯
Details β¬οΈ
Hugging Face (Twitter)
RT @LoubnaBenAllal1: Our science team at @huggingface will be doing an AMA on r/LocalLLaMA tomorrow at 8AM PST (5PM CET). The team members behind SmolLM, SmolVLM, FineWeb, and more will be present to answer all your questions!
RT @LoubnaBenAllal1: Our science team at @huggingface will be doing an AMA on r/LocalLLaMA tomorrow at 8AM PST (5PM CET). The team members behind SmolLM, SmolVLM, FineWeb, and more will be present to answer all your questions!
Hugging Face (Twitter)
RT @Xianbao_QIAN: I'm very glad to see that the new translation model from @TencentHunyuan is now ranking the 3rd. It's a reminder that small domain tuned models are more valuable than they appears.
Agentic stack needs both large and small models. Large models can handle planning and leverage sub-agents based on lean models to perform a particular task. Small models are cheap, fast and fine-tunable. They're not the opposite of large models but the complement to it.
RT @Xianbao_QIAN: I'm very glad to see that the new translation model from @TencentHunyuan is now ranking the 3rd. It's a reminder that small domain tuned models are more valuable than they appears.
Agentic stack needs both large and small models. Large models can handle planning and leverage sub-agents based on lean models to perform a particular task. Small models are cheap, fast and fine-tunable. They're not the opposite of large models but the complement to it.
This media is not supported in your browser
VIEW IN TELEGRAM
Hugging Face (Twitter)
RT @multimodalart: we hacked Wan 2.2 and discovered that it does first and last frame filling, works out of the box on 𧨠diffusers
i've built an app for it on @huggingface Spaces (which is powering powering our nano banana video mode too π π¬)
RT @multimodalart: we hacked Wan 2.2 and discovered that it does first and last frame filling, works out of the box on 𧨠diffusers
i've built an app for it on @huggingface Spaces (which is powering powering our nano banana video mode too π π¬)
Hugging Face (Twitter)
RT @QGallouedec: sept 4
8-11 am pst
@huggingface science team AMA
reddit r/LocalLlama
π½
RT @QGallouedec: sept 4
8-11 am pst
@huggingface science team AMA
reddit r/LocalLlama
π½
Hugging Face (Twitter)
RT @moby763canary21: I'm really glad that people are using my @huggingface model. It's really cool to contribute to Open ML!
#ai #machinelearning #huggingface @ClementDelangue
RT @moby763canary21: I'm really glad that people are using my @huggingface model. It's really cool to contribute to Open ML!
#ai #machinelearning #huggingface @ClementDelangue
Hugging Face (Twitter)
RT @lhoestq: "we made uploads to @huggingface using @ApacheSpark much faster than to any other cloud storage"
Spark is faster with Xet on Hugging Face for editing & publishing AI datasets π₯
I explained how it works hereπ
PS: it's π€―
PS2: thumb up and subπππ€π€π€
https://www.youtube.com/watch?v=vmwxVfye8fA?si=hp6Z3a28N0-bmZHF&t=2179
RT @lhoestq: "we made uploads to @huggingface using @ApacheSpark much faster than to any other cloud storage"
Spark is faster with Xet on Hugging Face for editing & publishing AI datasets π₯
I explained how it works hereπ
PS: it's π€―
PS2: thumb up and subπππ€π€π€
https://www.youtube.com/watch?v=vmwxVfye8fA?si=hp6Z3a28N0-bmZHF&t=2179
Hugging Face (Twitter)
RT @lvwerra: The Hugging Face research team is doing an AMA on r/LocalLlaMa tomorrow! π
Join if you are interested in:
> How did we get into the field? We cover a broad range of backgrounds and paths!
> How can you do impactful things while being more limited in resources than other labs?
> How do we decide which projects to work on when so many things are exciting?
> How does a fully remote team in a high velocity field even work?
> What's the most exciting thing coming in the next few months?
> What's your favourite optimizer and why is it Adam?
> How does Hugging Face make money?π€«
Or whatever else you want to ask - it's an AMA!
RT @lvwerra: The Hugging Face research team is doing an AMA on r/LocalLlaMa tomorrow! π
Join if you are interested in:
> How did we get into the field? We cover a broad range of backgrounds and paths!
> How can you do impactful things while being more limited in resources than other labs?
> How do we decide which projects to work on when so many things are exciting?
> How does a fully remote team in a high velocity field even work?
> What's the most exciting thing coming in the next few months?
> What's your favourite optimizer and why is it Adam?
> How does Hugging Face make money?π€«
Or whatever else you want to ask - it's an AMA!