MSA: Memory Sparse Attention for Efficient End-to-End Memory Model Scaling to 100M Tokens Paper • 2603.23516 • Published 23 days ago • 26
khtsly/Qwen3.5-27B-Claude-4.6-Opus-Distilled-32k-GGUF Image-Text-to-Text • 27B • Updated 6 days ago • 4.93k • 2
khtsly/Qwen3.5-2B-Claude-4.6-Opus-Distilled-32k-GGUF Image-Text-to-Text • 2B • Updated 6 days ago • 2.5k • 1
khtsly/Qwen3.5-35B-A3B-Claude-4.6-Opus-Distilled-32k-GGUF Image-Text-to-Text • 35B • Updated 6 days ago • 5.06k • 5
khtsly/Qwen3.5-4B-Claude-4.6-Opus-Distilled-32k-GGUF Image-Text-to-Text • 4B • Updated 6 days ago • 6.14k • 5
khtsly/Qwen3.5-9B-Claude-4.6-Opus-Distilled-32k-GGUF Image-Text-to-Text • 9B • Updated 6 days ago • 10.5k • 5