Published: August 8, 2025
140
567
4.5k

🚀 Qwen3-30B-A3B-2507 and Qwen3-235B-A22B-2507 now support ultra-long context—up to 1 million tokens! 🔧 Powered by: • Dual Chunk Attention (DCA) – A length extrapolation method that splits long sequences into manageable chunks while preserving global coherence. •

Image in tweet by Qwen

@Alibaba_Qwen @grok what is vllm?

@Alibaba_Qwen anyone know how much memory is needed to run this locally?

@Alibaba_Qwen can we please get a model with 300k native context instead?

@Alibaba_Qwen Is this another ROPE?

@Alibaba_Qwen So it can write a whole book all in one go?

@Alibaba_Qwen name is a bit short and easy to say don’t you think?

@Alibaba_Qwen Chatgp what? 🤣

@Alibaba_Qwen coming to qwen code or already there?

@Alibaba_Qwen Think AI was born polished? Read this. “Chatting with an Artificial Intelligence” shows the pre-hype texture of 2022: misses, leaps, and the first hints of voice. A rare primary source on how our conversations with machines began. #AI #ChatGPT https://www.amazon.com/dp/B0BP...

@Alibaba_Qwen open source strikes again. well done.

@Alibaba_Qwen You guys are just awesome. None of the AGI bs just practical stuff that actually works

@Alibaba_Qwen Wait, what? 🤯 Had no idea DCA even existed. Love these quick drop-ins. Now that’s what you call a truly open-sourced company.

@Alibaba_Qwen When are we going to get Qwen3 Omni series models?

@Alibaba_Qwen Boom! China is so gonna dominate the market in Open source - this is going to be awesome!

@Alibaba_Qwen @UnslothAI just got used to your model, guess Coder’s coming soon

@Alibaba_Qwen Damn china is rocking

@Alibaba_Qwen Are there any benchmark results for Qwen’s long context performance?

@Alibaba_Qwen Wow thanks! The east truly pushes open source.

@Alibaba_Qwen Awesome!

@Alibaba_Qwen 1M tokens? sheesh AIgust indeed.

@Alibaba_Qwen I think I will start looking at it

@Alibaba_Qwen This is amazing, thank you.

@Alibaba_Qwen magic ai labs do your thing.

@Alibaba_Qwen 1M token context?? guess i’ll start training for the ultramarathon too

@Alibaba_Qwen amazing.

@Alibaba_Qwen those guys just cant stop?

@Alibaba_Qwen Keep it coming. The same for coder?

@Alibaba_Qwen OpenAI in shambles

@Alibaba_Qwen qwen mogging again

@Alibaba_Qwen It just needs 1 TB RAM From the HF readme: To effectively process a 1 million token context, users will require approximately 1000 GB of total GPU memory. This accounts for model weights, KV-cache storage, and peak activation memory demands.

@Alibaba_Qwen what about the chat platform

@Alibaba_Qwen Great job, not like some gpt-ass

@Alibaba_Qwen 太好了,是 100 万上下文,番茄小说作者有救了!

@Alibaba_Qwen awesome wanna @Alibaba_Qwen to be the best at research also

@Alibaba_Qwen Amazing! 👏

Image in tweet by Qwen

Share this thread

Read on Twitter

View original thread

Navigate thread

1/37