
Qwen
Alibaba_Qwen
We're leveling up the game with our latest open-source models, Qwen2.5-1M ! 💥 Now supporting a 1 MILLION TOKEN CONTEXT LENGTH 🔥 Here's what’s new: 1️⃣ Open Models: Meet Qwen2.5-7B-Instruct-1M & Qwen2.5-14B-Instruct-1M —our first-ever models handling 1M-token contexts! 🤯 2️⃣ Lightning-Fast Inference Framework: We’ve fully open-sourced our inference framework based on vLLM , integrated with sparse attention methods. Experience 3x to 7x faster processing for 1M-token inputs! ⚡⚡ 3️⃣ Tech Deep Dive: Check out our detailed Technical Report for all the juicy details behind the Qwen2.5-1M series! 📊 📖 Technical Report: https://qianwen-res.oss-cn-bei... 📄 Blog: https://qwenlm.github.io/blog/... Experience Qwen2.5-1M live: 👉 Play with Qwen2.5-Turbo supporting 1M tokens in Qwen Chat (https://chat.qwenlm.ai) 👉 Try it on Huggingface (https://huggingface.co/collect... 👉 Or head over to Modelscope (https://modelscope.cn/collecti...