{"href":"https://api.simplecast.com/oembed?url=https%3A%2F%2Fcontext-rot.simplecast.com%2Fepisodes%2Fthis-week-qwen-36-27b-deepseek-v4-gpt-55-apd1_lba","width":444,"version":"1.0","type":"rich","title":"This week: Qwen 3.6 27B, DeepSeek V4, & GPT 5.5","thumbnail_width":300,"thumbnail_url":"https://image.simplecastcdn.com/images/6bc44c50-3376-443d-a3df-4525e5207c89/bfef3627-f100-445f-81ca-93d5152dfb76/context-20rot-20cover-20-1.jpg","thumbnail_height":300,"provider_url":"https://simplecast.com","provider_name":"Simplecast","html":"<iframe src=\"https://player.simplecast.com/c0f332b0-7985-489f-9bc0-d45518152735\" height=\"200\" width=\"100%\" title=\"This week: Qwen 3.6 27B, DeepSeek V4, &amp; GPT 5.5\" frameborder=\"0\" scrolling=\"no\"></iframe>","height":200,"description":"Somewhere in the last seven days, a 27-billion-parameter model quietly beat a 397-billion-parameter model at its own game. Not close — beat it. On every major coding benchmark. The model doing the beating runs on a single consumer GPU. The model getting beaten costs fifteen times more to run. And the company that built both of them just... released the smaller one for free. That's the story we're starting with today. But honestly? It might not even be the wildest thing that happened this week."}