{"id":6051,"date":"2026-04-08T15:14:26","date_gmt":"2026-04-08T15:14:26","guid":{"rendered":"https:\/\/news678.top\/?p=6051"},"modified":"2026-04-08T15:14:26","modified_gmt":"2026-04-08T15:14:26","slug":"mustafa-suleyman-ai-development-wont-hit-a-wall-anytime-soon-heres-why","status":"publish","type":"post","link":"https:\/\/news678.top\/?p=6051","title":{"rendered":"Mustafa Suleyman: AI development won\u2019t hit a wall anytime soon\u2014here\u2019s why"},"content":{"rendered":"<p><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">We evolved for a linear world. If you walk for an hour, you cover a certain distance. Walk for two hours and you cover double that distance. This intuition served us well on the savannah. But it catastrophically fails when confronting AI and the core exponential trends at its heart.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">From the time I began work on AI in 2010 to now, the amount of training data that goes into frontier AI models has grown by a staggering 1 trillion times\u2014from roughly 10\u00b9\u2074 flops (floating-point operations\u201a the core unit of computation) for early systems to over 10\u00b2\u2076 flops for today\u2019s largest models. This is an explosion. Everything else in AI follows from this fact.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">The skeptics keep predicting walls. And they keep being wrong in the face of this epic generational compute ramp. Often, they point out that Moore\u2019s Law is slowing. They also mention a lack of data, or they cite limitations on energy.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\"><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"><\/span><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">But when you look at the combined forces driving this revolution, the exponential trend seems quite predictable. To understand why, it\u2019s worth looking at the complex and fast-moving reality beneath the headlines.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Think of AI training as a room full of people working calculators. For years, adding computational power meant adding more people with calculators to that room. Much of the time those workers sat idle, drumming their fingers on desks, waiting for the numbers to come through for their next calculation. Every pause was wasted potential. Today\u2019s revolution goes beyond more and better calculators (although it delivers those); it is actually about ensuring that all those calculators never stop, and that they work together as one.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Three advances are now converging to enable this. First, the basic calculators got faster. Nvidia\u2019s chips have delivered an eightfold increase in raw performance in just six years, from<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">312 teraflops in 2020<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> to<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">2,500 teraflops today<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">. Our own<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Maia 200<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> chip, launched this January, delivers 30% better performance per dollar than any other hardware in our fleet. Second, the numbers arrive faster thanks to a technology called HBM, or high bandwidth memory, which stacks chips vertically like tiny skyscrapers; the latest generation, HBM3, triples the bandwidth of its predecessor, feeding data to processors fast enough to keep them busy all the time. Third, the room of people with calculators became an office and then a whole campus or city. Technologies like<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">NVLink<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> and<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">InfiniBand<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> connect hundreds of thousands of GPUs into warehouse-size supercomputers that function as single cognitive entities. A few years ago this was impossible.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">These gains all come together to deliver dramatically more compute. Where training a language model took 167 minutes on eight GPUs in 2020, it now takes under four minutes on equivalent modern hardware. To put this in perspective: <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Moore\u2019s Law would predict <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">only about a 5x improvement over this period. We saw 50x. We\u2019ve gone from two GPUs training AlexNet, the image recognition model that kicked off the modern boom in deep learning in 2012, to over 100,000 GPUs in today\u2019s largest clusters, each one individually far more powerful than its predecessors.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Then there\u2019s the revolution in software. Research from<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Epoch AI<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> suggests that the compute required to reach a fixed performance level halves approximately every eight months, much faster than the traditional 18-to-24-month doubling of Moore\u2019s Law. The costs of serving some recent models have collapsed by a factor of up to 900 on an annualized basis. AI is becoming radically cheaper to deploy.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">The numbers for the near future are just as staggering. Consider that leading labs are growing capacity at nearly 4x annually. Since 2020, the compute used to train frontier models has grown<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">5x every year<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">. Global AI-relevant compute is forecast to hit 100 million H100-equivalents by 2027, a tenfold increase in three years. Put all this together and we\u2019re looking at something like another 1,000x in effective compute by the end of 2028. It\u2019s plausible that by 2030 we\u2019ll bring an additional<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">200 <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">gigawatts of compute online every year\u2014akin to the peak energy use of the UK, France, Germany, and Italy put together.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">What does all this get us? I believe it will drive the transition from chatbots to nearly human-level agents\u2014semiautonomous systems capable of writing code for days, carrying out weeks- and months-long projects, making calls, negotiating contracts, managing logistics. Forget basic assistants that answer questions. Think teams of AI workers that deliberate, collaborate, and execute. Right now we\u2019re only in the foothills of this transition, and the implications stretch far beyond tech. Every industry built on cognitive work will be transformed.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">The obvious constraint here is energy. A single refrigerator-size AI rack consumes 120 kilowatts, equivalent to 100 homes. But this hunger collides with another exponential: <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Solar costs<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> have fallen by a factor of nearly 100 over 50 years;<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> <\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">battery prices<\/span><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\"> have dropped 97% over three decades. There is a pathway to clean scaling coming into view.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">The capital is deployed. The engineering is delivering. The $100 billion clusters, the 10-gigawatt power draws, the warehouse-scale supercomputers &#8230; these are no longer science fiction. Ground is being broken for these projects now across the US and the world. As a result, we are heading toward true cognitive abundance. At Microsoft AI, this is the world our superintelligence lab is planning for and building.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><\/p>\n<p dir=\"ltr\" style=\"line-height: 1.806546;margin-top: 0pt;margin-bottom: 0pt\">\n<p><span style=\"font-size: 12pt;font-family: Arial, sans-serif;vertical-align: baseline;white-space: pre-wrap\">Skeptics accustomed to a linear world will continue predicting diminishing returns. They will continue being surprised. The compute explosion is the technological story of our time, full stop. And it is still only just beginning.<\/span><\/span><\/p>\n<p><span id=\"docs-internal-guid-5e5a2a70-7fff-a08f-f865-6f5020e43e8c\" style=\"white-space: normal\"><span style=\"font-size: 12pt;font-family: Arial, sans-serif;font-style: italic;vertical-align: baseline;white-space: pre-wrap\"><em>Mustafa Suleyman is CEO of Microsoft AI.<\/em><\/span><\/span><\/p>\n<p>#Mustafa #Suleyman #development #wont #hit #wall #anytime #soonheres<\/p>\n","protected":false},"excerpt":{"rendered":"<p>We evolved for a linear world. If you walk for an hour, you cover a&#8230;<\/p>\n","protected":false},"author":1,"featured_media":6052,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[7],"tags":[5783,5777,4061,5778,5779,1280,5781,5776,5784,5782,5780,3944,866],"class_list":["post-6051","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-stories","tag-anytime","tag-ceo-of-microsoft-ai","tag-development","tag-explains-why-ai-development-wont-hit-a-wall-anytime-soon-in-fact","tag-he-says","tag-hit","tag-mustafa","tag-mustafa-suleyman","tag-soonheres","tag-suleyman","tag-the-compute-explosion-is-the-technological-story-of-our-time-and-it-is-still-only-just-beginning","tag-wall","tag-wont"],"featured_image_urls":{"full":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman.png",1200,600,false],"thumbnail":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman-150x150.png",150,150,true],"medium":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman-300x150.png",300,150,true],"medium_large":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman-768x384.png",640,320,true],"large":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman-1024x512.png",640,320,true],"1536x1536":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman.png",1200,600,false],"2048x2048":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman.png",1200,600,false],"covernews-featured":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman-1024x512.png",1024,512,true],"covernews-medium":["https:\/\/news678.top\/wp-content\/uploads\/2026\/04\/Final-Header_Suleyman-540x340.png",540,340,true]},"author_info":{"display_name":"admin","author_link":"https:\/\/news678.top\/?author=1"},"category_info":"<a href=\"https:\/\/news678.top\/?cat=7\" rel=\"category\">Stories<\/a>","tag_info":"Stories","comment_count":"0","_links":{"self":[{"href":"https:\/\/news678.top\/index.php?rest_route=\/wp\/v2\/posts\/6051","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/news678.top\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/news678.top\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/news678.top\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/news678.top\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=6051"}],"version-history":[{"count":0,"href":"https:\/\/news678.top\/index.php?rest_route=\/wp\/v2\/posts\/6051\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/news678.top\/index.php?rest_route=\/wp\/v2\/media\/6052"}],"wp:attachment":[{"href":"https:\/\/news678.top\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=6051"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/news678.top\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=6051"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/news678.top\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=6051"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}