Why AI Training Costs Millions: A Look at the 'Gigafactory of Compute'

I’m often asked which AI training project cost millions of dollars and two years of my life. People wonder: why is it so expensive? My usual answer is that it’s not particularly expensive—especially considering we don’t own our own hardware yet. Training AI has always been about massive data centers; that’s just the reality of the field. When you’re not immersed in it, the sheer scale can be hard to visualize. ...

9 May, 2025 · 2 min · 268 words · Yury Akinin

Diary of an AI Startup

This series of posts will be my way of documenting the journey of creating one of our team’s most ambitious products: the intelligent assistant, A.V.E.L.I.N. To give you some context, my development team and I are currently beta-testing the project within our Mozgii Ecosystem AI platform. Our primary focus is on A.V.E.L.I.N.—an intelligent personal assistant in Telegram built to handle both basic and complex tasks involving AI-powered search, processing, and analysis of information. ...

1 May, 2025 · 1 min · 211 words · Yury Akinin

Adhocracy in IT: The Operating System for Modern Startups

In traditional companies, everything is built on a clear hierarchy: decisions are made at the top and executed at the bottom. This approach might work in a stable environment, but for an IT startup, especially in AI, it stifles growth. IT companies need adhocracy: a management model where competence and results are valued more than titles. It’s about flexibility over bureaucracy and speed over approvals. The value of an idea is judged by its effectiveness, not by the position of its author. ...

28 April, 2025 · 1 min · 192 words · Yury Akinin

The Emerging Skill of the AI Era: Beyond Just Searching

From my experience working with neural networks, it’s become obvious how two people can interact with the same model and get radically different outcomes. This isn’t a minor variation—it signals a fundamental departure from the search engine paradigm. We have entered a new era of interacting with artificial intelligence. AI doesn’t just aggregate data; it selects and synthesizes relevant information in response to specific requests. This changes the very nature of how we engage with information. Where the key skill was once finding data in search engines, it is now the ability to correctly formulate requests to an AI. ...

25 April, 2025 · 1 min · 205 words · Yury Akinin

Why Sber and Yandex Lag Behind Global AI Leaders

I’m often asked why international AI models, like those from OpenAI, consistently outperform Russian counterparts such as GigaChat. To understand the gap, we need to look beyond the code and analyze the foundational, structural challenges. Here are the key factors limiting Russia’s position in the global AI race. 1. The Compute Bottleneck Effective AI development at scale depends on raw computational power. Since 2022, access to essential high-performance NVIDIA chips (like the A100 and H100) has been severed. Training a model on the scale of GPT-4 requires a cluster of over 10,000 GPUs—a resource capacity that simply doesn’t exist in Russia. For context, Sber’s most powerful supercomputer, Christofari Neo, operates at around 12 petaflops, making it 50 to 100 times less powerful than the world’s leading AI research centers. ...

24 April, 2025 · 2 min · 406 words · Yury Akinin

AI in Messengers: Moving Beyond Simple Bots

Messengers have long evolved past being simple tools for exchanging messages. Today, they are the space where our work, personal lives, news, and friends converge. The logical next step is to integrate an AI assistant into this environment—one that helps structure communication and interacts with you in a familiar interface, just like any other contact. A majority of the solutions on the market are bots running on simplified or free versions of GPT. They can generate answers, but often fail to provide high-quality, deep information. ...

22 April, 2025 · 2 min · 252 words · Yury Akinin

Beyond the Interface: 5 Key Differentiators of Modern AI Models

Users see a chat window. Sometimes voice, sometimes images. But behind this familiar interface lie radically different architectures and capabilities. Here are five key parameters that distinguish the top AI models in 2025: 1. Memory (Context Window) This defines how much information a model can retain within a single conversation. GPT-4o: 128k tokens (~300 pages of text) Claude 3 Opus & Gemini 2.5 Pro: Up to 1 million tokens (~2,000 pages) DeepSeek-VL Mini: ~8k tokens (~20 pages) More memory enables greater context and reduces hallucinations, but it also demands more powerful hardware. ...

19 April, 2025 · 2 min · 367 words · Yury Akinin

OpenAI's Codex CLI: A Quiet Win for Open-Source

OpenAI has released Codex CLI, an open-source AI agent for developers. This marks a quiet but significant victory for the open-source community. The tool allows developers to use natural language directly in the terminal—the agent interprets the request, then writes, executes, and tests the code. Most importantly, this entire process runs locally, without sending data to the cloud. With this release, the industry moves one step closer to a system that can independently understand, build, and deploy solutions. It underscores a critical point: the future isn’t just about choosing the right model, but about engineering the right architecture that connects thought → action. ...

17 April, 2025 · 2 min · 222 words · Yury Akinin

Models Are Tools, Not Events: The Real Meaning Behind GPT-4.1 and the End of GPT-4.5

Yesterday, OpenAI opened access to the GPT-4.1 API. It’s a refined version of their flagship model—faster and architecturally closer to the concept of ‘agents.’ In parallel, the company officially announced it is winding down GPT-4.5, its most resource-intensive model, due to its excessive complexity and support challenges. With GPT-4.5, it seems they hit an architectural dead end. We are at a point where models appear and disappear rapidly. They are becoming what they should be: tools, not landmark events. We have a growing catalog of specialized AIs: some calculate, others write code, plan tasks, or generate video. But the average user should not be expected to know and choose between every AI in existence. That paradigm defies the logic of good user experience. ...

15 April, 2025 · 2 min · 269 words · Yury Akinin

Deep Research: From Information Hunter to Strategic Co-Pilot

Your Thought Process, Packaged Deep Research isn’t just another AI feature; it’s a fundamental shift toward an agent-based architecture. In this model, the LLM stops being a simple chatbot and becomes a co-author—an agent that independently searches, filters, validates, and structures information. What does this change? If you’re designing a business, a startup, or a product, you don’t have time to personally read 200 sources. Now, an AI agent does it for you. This frees you up to do the high-value work: to think, not just to search. ...

14 April, 2025 · 2 min · 421 words · Yury Akinin