Cryptopolitan
2025-08-03 15:37:54

AI amplifies intelligence, but only if you’re already smart, says Balaji

Balaji says AI is polytheistic, not monotheistic, meaning there isn’t one super-intelligent system ruling over all. Instead, there are many strong AIs, each backed by different players. In his words: “We empirically observe polytheistic AI… rather than a single all-powerful model.” That wipes out the fantasy of one AGI turning the world into paperclips. What we have is a balance of power between many human/AI combinations, not one dominant force. He says AI right now only works “middle-to-middle.” It doesn’t handle full jobs from start to finish. You still need people at both ends; one to prompt the AI and another to check what it outputs. So all the real costs and effort have shifted to the edges: prompting and verifying. That’s where companies are now spending their money, even though AI speeds up the center of the process. AI makes you smarter, but only if you are already smart Balaji doesn’t call it artificial intelligence. He calls it amplified intelligence. Because the AI isn’t acting on its own, it’s not fully agentic, it doesn’t set long-term goals, and it can’t verify its own output. “You have to spend a lot of effort on prompting, verifying, and system integrating,” he said. So, how useful AI is depends on how smart you are. If you give it bad instructions, it gives you bad results. He also says AI doesn’t replace you, it just helps you do more jobs. With it, you can fake your way into being a passable UI designer or game animator. But don’t expect expert quality. AI makes you good enough to be average, not excellent. For real quality, you still need specialists. There’s another job it does take, and that’s the job of the last version of itself. Midjourney pushed Stable Diffusion out of the workflow. GPT-4 took GPT-3’s spot. As Balaji puts it, “AI doesn’t take your job, it takes the job of the previous AI.” Once companies create a space for AI in a workflow, like image creation or code generation, that space stays filled. It just gets handed off to the newer, better model. He also says AI is better at visuals than text. Easier for humans to judge a picture than to verify a wall of code or paragraphs of text. “User interfaces and images can easily be verified by the human eye,” Balaji says. With text, it’s slower and more costly for people to check the accuracy. Crypto limits what AI can and can’t do Balaji draws a line between how AI works and how crypto works. AI is probabilistic; it guesses based on patterns. But crypto is deterministic; it runs on hard, provable math. So crypto becomes a boundary that AI can’t easily cross. AI might break captchas, but it can’t fake a blockchain balance. “AI makes everything fake, but crypto makes it real again,” he says. AI might solve simple equations, but cryptographic equations still block it. There’s also already a version of killer AI out there. It’s drones. “Every country is pursuing it,” Balaji says. It’s not image generators or chatbots that pose the threat, it’s autonomous weapons. That’s the area where AI’s real-world impact is already lethal. He argues that AI is decentralizing, not centralizing. Right now, there are tons of AI companies, not just one or two giants. Small teams with good tools can do a lot. And open-source models are improving fast. So even without massive budgets, small groups can build strong AI systems. That breaks up power instead of concentrating it. Balaji also rejects the idea that more AI is always better. He says the ideal amount is not zero, not 100%. “0% AI is slow, but 100% AI is slop.” Real value lives in between. Too little AI means you’re behind. Too much, and quality falls apart. He compares it to the Laffer Curve, a concept in economics that says there’s a sweet spot between extremes. In his final argument , he lays out why today’s systems are constrained AIs, not godlike machines. He breaks that into four kinds of limits: Economic: Every API call costs money. Using AI at scale isn’t free. Mathematical: AI can’t solve chaotic or cryptographic problems. Practical: You still need humans to prompt and verify results. AI can’t complete the full task alone. Physical: AI doesn’t gather real-world data on its own. It can’t sense its environment or interpret it like people do. He ends by saying these limits might be removed later. It’s possible that future researchers can merge System 1 thinking (fast and intuitive, like AI) with System 2 thinking, which is more logical and careful, like traditional computing. But right now, that’s just theory. It’s still an open problem. There is no all-knowing AI. There are just tools (expensive, limited, competitive tools) that do what they’re told and need constant checking. KEY Difference Wire helps crypto brands break through and dominate headlines fast

获取加密通讯
阅读免责声明 : 此处提供的所有内容我们的网站,超链接网站,相关应用程序,论坛,博客,社交媒体帐户和其他平台(“网站”)仅供您提供一般信息,从第三方采购。 我们不对与我们的内容有任何形式的保证,包括但不限于准确性和更新性。 我们提供的内容中没有任何内容构成财务建议,法律建议或任何其他形式的建议,以满足您对任何目的的特定依赖。 任何使用或依赖我们的内容完全由您自行承担风险和自由裁量权。 在依赖它们之前,您应该进行自己的研究,审查,分析和验证我们的内容。 交易是一项高风险的活动,可能导致重大损失,因此请在做出任何决定之前咨询您的财务顾问。 我们网站上的任何内容均不构成招揽或要约