Unveiling QwenLong-L1.5: A Post-Training Blueprint for Mastering Long-Context Reasoning and Memory Management Summary QwenLong-L1.5, built on Qwen3-30B-A3B-Thinking, excels in long-context reasoning through innovative post-training techniques. It features a data synthesis pipeline for multi-hop tasks, stabilized RL with task-balanced sampling and AEPO, and a memory framework for ultra-long inputs. Evaluations show a 9.9-point average gain, matching GPT-5 and Gemini-2.5-Pro levels. Have you ever wondered why large language models struggle with lengthy texts, often losing track of key details across thousands of words? Picture this: you’re sifting through a massive report, needing to connect dots from scattered evidence to form a coherent …