QWEN
30 days · UTC
Synchronizing with global intelligence nodes...
SWE-Bench Pro leaderboard: small gains at the top, big contexts, and mostly self-reported results
A new SWE-Bench Pro leaderboard shows top code models clustered around 0.55–0.58, with large contexts and self-reported scores. The updated [SWE-Benc...
Open models heat up: Tencent eyes OpenClaw, Qwen3.5-35B-A3B guide lands, Fireworks teases coding plan
Open-source LLM options are shifting as Tencent reportedly backs OpenClaw, a Qwen3.5-35B-A3B setup guide circulates, and Fireworks AI hints at a codin...
Stop starving your GPUs: make agent rollout a service
Separating I/O-heavy agent rollouts from GPU training nearly doubled coding-agent performance and fixed chronic GPU underutilization. An NVIDIA audit...
Runpod data: Qwen just passed Llama as the most-deployed self‑hosted LLM
Runpod’s latest platform data says Qwen has overtaken Llama as the top self-hosted LLM. According to Runpod’s report, more teams now spin up Qwen tha...
Google’s Gemini 3.1 Flash-Lite targets high-volume, low-latency workloads
Google released Gemini 3.1 Flash-Lite, a faster, cheaper model aimed at high-volume developer workloads and signaling a broader shift to lighter LLMs ...
Using third‑party LLM APIs in VS Code (Qwen via Together/DeepInfra)
A developer is replacing a flat-fee assistant with pay‑per‑use API models in VS Code, specifically Qwen Coder 2.5 via Together or DeepInfra, for occas...