DeepSeek vs. OpenAI's OSS: A Tale of Two Open-Source Models

Two major players recently dropped new open-source models, but they represent two fundamentally different philosophies. OpenAI, the established leader, returned to the open-source scene with fanfare and its gpt-oss-20b model. Shortly after, the Chinese startup DeepSeek quietly released v3.1. While one was a media event, the other was a single tweet. The initial results from hands-on testing are starkly one-sided. Out-of-the-Box Performance: A Clear Winner When you evaluate a model as a tool to be used right now, the comparison is not even close. Across multiple practical tests, DeepSeek v3.1 consistently delivered superior results: ...

27 August, 2025 · 4 min · 654 words · Yury Akinin

OpenAI's GPT-OSS: A Major Step Back Towards 'Open'

OpenAI just made a significant move by releasing GPT-OSS, its first truly open-source large language model family since GPT-2. With a permissive Apache 2.0 license, this isn’t just a minor release; it’s a fundamental shift that puts real power back into the hands of developers. The family includes two Mixture-of-Experts (MoE) models, gpt-oss-20b and gpt-oss-120b, designed for high-performance inference with strong reasoning capabilities. Why This Is a Game-Changer For years, the most powerful models from OpenAI have been locked behind APIs. This meant dealing with rate limits, opaque pricing, and sending potentially sensitive data to a third party. GPT-OSS changes that equation entirely. ...

13 August, 2025 · 2 min · 418 words · Yury Akinin

Claude Sonnet 4's 1M Token Window: A Practical Take for Builders

Anthropic just announced a 5x context window increase for Claude Sonnet 4, pushing it to 1 million tokens. While big numbers in AI are common, this move has tangible, practical implications for those of us building complex systems. From my perspective, this isn’t just a quantitative leap; it’s a qualitative one that unlocks a new class of problems we can solve. Moving from File Analysis to System-Level Understanding The ability to load an entire codebase—over 75,000 lines with source files, tests, and docs—into a single prompt is a significant shift. Previously, AI code analysis was often limited to individual files or small modules. We could check for errors or refactor a specific function, but the AI lacked a holistic view. ...

13 August, 2025 · 3 min · 441 words · Yury Akinin

Claude Opus 4.1: A Focused Upgrade on Coding and a Measured Stance on Autonomy

Anthropic has released Claude Opus 4.1, an incremental but important update that sharpens its flagship model’s capabilities in specific, high-value areas: agentic tasks, real-world coding, and reasoning. This isn’t a complete overhaul, but a focused enhancement for professional and development use cases. Enhanced Coding and Reasoning The primary upgrade is in coding performance. Opus 4.1 achieves a 74.5% score on the SWE-bench Verified benchmark. Digging into the technical details, it solved an average of 18.4 problems on the hard subset, up from 16.6 for Claude Opus 4. ...

6 August, 2025 · 2 min · 330 words · Yury Akinin

Deep Research: From Information Hunter to Strategic Co-Pilot

Your Thought Process, Packaged Deep Research isn’t just another AI feature; it’s a fundamental shift toward an agent-based architecture. In this model, the LLM stops being a simple chatbot and becomes a co-author—an agent that independently searches, filters, validates, and structures information. What does this change? If you’re designing a business, a startup, or a product, you don’t have time to personally read 200 sources. Now, an AI agent does it for you. This frees you up to do the high-value work: to think, not just to search. ...

14 April, 2025 · 2 min · 421 words · Yury Akinin

DeepSeek-V3: A Quiet Release with Impressive Local Performance

DeepSeek has once again followed its “quiet release” strategy, making its new DeepSeek-V3-0324 model available on Hugging Face without any major announcements. Instead of marketing hype, they’ve simply delivered a solution for the community to evaluate. I tested the model locally on a Mac Studio equipped with an M3 Ultra chip and saw impressive performance, generating over 20 tokens per second. This marks a significant acceleration for running capable models on local hardware, making it a viable option for developers. ...

27 March, 2025 · 1 min · 113 words · Yury Akinin