大多数人认为人工智能就是一个聊天机器人。我能理解。你打开 ChatGPT,让它帮你修改一封邮件,它就做到了。感觉就像魔法一样。你离开时以为自己已经明白了其中的原理。但这就像是在餐厅刷了一张信用卡,然后就觉得自己理解了 Visa 是怎么赚钱的一样。你使用了产品,但你没有看到背后的系统。
我去年大部分时间都在试图弄清楚人工智能真正的资金流向在哪里,而诚实的答案是:我花了相当尴尬的长时间才意识到自己一直在看错层级。我一直盯着 ChatGPT、Claude 和 Gemini 这些你可以直接接触到的东西,而与此同时,有 7000 亿美元正被悄悄投入到一些我甚至说不上名字的基础设施中。我从未听说过的芯片。带着听起来像是编出来的缩写的封装技术。冷却系统。发电厂。混凝土正在德克萨斯、爱荷华和海得拉巴被浇筑。一年前,我认识的人里几乎没有人在谈论这些事情。现在他们都在谈了。
这篇文章会很长。如果你现在没有时间,请先收藏,之后再回来阅读。我想带你走一遍整个 AI 的价值链,每一层,从为数据中心供电的电力,一直到你手机上的应用程序。我会用一种即使你一辈子都没读过年报也能理解的方式来讲解。当我使用专业术语时,我会解释它们。我会为每一个论断附上真实的数字。而且我也会坦诚地说明那些我仍然不太确定的部分,因为确实有一些。
让我们开始。
一 —— 五层蛋糕(以及为什么没人谈论最底下的四层)
人工智能是基础设施。就像互联网,就像电力,它需要工厂。
~ Jensen Huang
大多数人理解人工智能的方式大概是这样的:一台聪明的电脑回答问题。
这就像说互联网是“你看视频的地方”。从技术上讲并不错误,但它忽略了整个重点。Nvidia CEO 黄仁勋在2026年1月的达沃斯会议上将人工智能描述为一个五层系统。能源、芯片、云计算、模型、应用。他将整个系统称为“人类历史上最大的基础设施建设”。想一想这个词。基础设施。道路、电网、水系统。这些是让文明运转的东西,直到它们出问题时,没人会去想它们。人工智能正在变成这种东西。隐形、必不可少、并且极其昂贵。我要称之为“AI堆栈”。五个层次,一个叠加在另一个上面,每一层都为上一层提供支持,资金则在这几层之间流动。这里是我能给出的最简单版本:能源。你需要电力来驱动计算机。大量的电力。芯片。你需要专门的处理器来进行计算。这些可不是你笔记本电脑的“大脑”。云计算。你需要装满这些芯片的大型仓库,通过极快的网络连接。模型。你需要实际的AI软件,即从数据中学习模式的“大脑”。应用。你需要人们实际使用的产品。ChatGPT、谷歌搜索、你银行的欺诈检测系统。任何只关注第五层的关于人工智能的讨论,都错过了80%的全貌。以下是如果你是投资者、创始人或仅仅是想理解这个世界走向的人,你需要知道的重点:资金不会在这些层之间均匀流动。它会集中,复合。而现在,它正集中在大多数人没有关注的地方。II - 跟随资金流动(它并不在你想的地方)每个人的思维都集中在应用层。ChatGPT、Copilot、Claude、Perplexity。
你每天接触的是这些产品,所以感觉它们就是全部的故事。但有一件事几乎所有人都忽略了。到 2026 年,四大云计算公司(Amazon、Microsoft、Google 和 Meta)预计将投入大约 6500 亿到 7000 亿美元的资本支出。合计。在短短一年内。这大致相当于瑞士的 GDP。
而其中将近 75%,大约 4500 亿美元,将直接投入到 AI 基础设施中。不是用在聊天机器人上。不是用在应用程序上。而是用在建筑、芯片、电缆和冷却系统上。
在鸡尾酒会上没有人谈论这些东西。正因为如此,你才知道钱真正流向哪里。
因为想想看。在任何人使用 ChatGPT 之前,必须有人先建造一个和大型购物中心一样大的数据中心,在里面放入成千上万的专用处理器,用价格比大多数公司估值还高的网络设备把它们连接起来,然后为整个系统提供足以为一座小城市供电的电力。每天都是如此。
这就是第 1 层到第 3 层。那些看不见的层。真正的大规模资本正在这些层中部署。
“那 OpenAI 呢?他们不是在赚几十亿美元吗?”
确实如此。到 2025 年底,OpenAI 的年化经常性收入达到 200 亿美元,而一年前是 60 亿美元,再前一年是 20 亿美元。两年增长了 10 倍。历史上没有任何公司能在这样一个基础上实现如此快的收入扩张。
但问题在于:OpenAI 在 2025 年大约烧掉了 90 亿美元现金,并预计在 2026 年烧掉 170 亿美元现金。他们的推理成本(也就是当你向 AI 提问时实际运行 AI 的成本)在 2025 年达到 84 亿美元,预计在 2026 年将达到 141 亿美元。他们预计要到 2029 年或 2030 年才会实现正现金流。
那么,那些被烧掉的现金流向哪里?它沿着产业链向下流动。流向微软 Azure(OpenAI 在 2032 年前将其总收入的 20% 支付给微软)。流向英伟达购买芯片。流向建设和配备数据中心的公司。流向发电的电力公司。如果你盯着看足够久,会发现这几乎是一个循环。
微软投资 OpenAI。OpenAI 将这些钱花在 Azure 上。Azure 用收入购买更多的英伟达芯片。英伟达报告创纪录的收益。人人欢庆。而现金继续往下流。大部分用户处在产业链的顶端。大部分利润处在底端。这种断层就是整个论点。
这是 AI 价值链的第一课:收入向上流,资本向下流。
III - 你以前看过这个“电影”
全人类的问题都是工程问题,而工程问题可以被解决。
~ 巴克敏斯特·富勒
如果你想理解 AI 正在发生的事情,就研究一下 1880 到 1920 年间电力发生了什么。
当托马斯·爱迪生在 1882 年在曼哈顿珍珠街建造了第一座商业发电站时,人们认为电力只是个新奇事物。一种照亮房间的花哨方式。气灯用得好好的,谁还需要电呢?
不到 40 年,电力就重组了地球上的每一个行业:制造业、交通、通信、医疗、娱乐。
胜出的公司并不是那些发明电灯泡的公司,而是那些建造发电厂、铺设铜线、制造发电机的公司:通用电气、西屋电气、公用事业公司、铜矿商、建筑商。
同样的模式正在 AI 领域上演,只不过压缩到了几年而非几十年。
AI → 数据中心 → 芯片 → 原材料 → 能源
电力 → 工厂 → 机器 → 原材料 → 煤/水
这个箭头式的演进几乎完全相同。
而赢家,再一次,并不主要在应用层。
他们在基础设施层。
我把这称为基础设施引力(Infrastructure Gravity)。
每当一个新的计算平台出现时,最初的财富创造都发生在“镐和铲”上。
应用随后才会出现。
应用获得所有的媒体关注。
但基础设施拿走了所有的利润。
英伟达在2026财年(截至2026年1月)的全年营收达到 2159亿美元,同比增长 65%。
仅数据中心业务在最后一个季度就实现了 623亿美元 的收入,同比增长 75%。
这个单一业务板块现在占英伟达总收入的 超过91%。
想想看:
一家公司一个季度就做到 680亿美元 的收入,而其中 十分之九 的钱来自同一条业务线。
台积电,这家实际制造英伟达芯片(以及几乎所有其他公司芯片)的公司,在 2025年 捕获了全球晶圆代工市场 近70% 的份额,销售额达到 1225亿美元。
最接近的竞争对手三星只有 7.2%。
这种级别的统治力,甚至会让 标准石油(Standard Oil) 都感到不安。
基础设施总是最先获胜。
问题只在于,这个窗口会开放多久。
问任何人互联网革命是关于什么的,他们会说:
Google、Amazon 和 Facebook。
但如果问早期的钱真正是在哪里赚到的,答案是:
Cisco、Corning,以及那些铺设光纤的公司。
同一个故事。
不同的年代。
IV —— 没有人愿意听的那部分
股票市场是一个把钱从没有耐心的人手中转移到有耐心的人手中的装置。
~ Charlie Munger
我得坦白一件事。
当我第一次以投资者的身份开始关注 AI 时,我犯了一个大多数人都会犯的错误:我把目光放在了应用层。
我看到 ChatGPT 在增长。我看到 Anthropic 融资数十亿美元。我当时想:AI 公司会赢,所以就投资 AI 公司。
有三件事改变了我的想法。而且它们是按顺序发生的,这一点很重要,因为每一件都建立在前一件的基础上。
第一,我注意到几乎每一家“AI 公司”都在大量烧钱。OpenAI、Anthropic、Mistral、xAI,全都花得比赚得快。
这不是因为它们是糟糕的企业,而是因为计算成本是结构性的。
每当你向一个 AI 模型提问时,生成这个答案都会产生真实的成本。而且模型越聪明,需要的计算量就越多,这意味着运行成本也越高。
你认为的那些“AI 的赢家”,其实恰恰是花钱最多的公司。
第二,我意识到基础设施公司正在以一种我自 Google 早期以来从未见过的利润率疯狂赚钱。
NVIDIA 的毛利率徘徊在 75% 左右。TSMC 一边扩张产能,一边提高价格,因为需求远远超过供应。
这些公司没有“我们什么时候才能实现盈利”的问题。
它们面临的是“我们根本来不及建得更快”的问题。
这是两种完全不同的问题。
第三点,也是最让我不舒服的一点,我意识到自己一直是以消费者的视角,而不是工程师的视角来思考 AI。
消费者看到的是应用。
工程师看到的是技术栈。
一旦你看到了这层技术栈,就再也无法忽视它。
每一次 AI 的发布都会变成一次资本开支(capex)的公告。
每一次模型改进都会变成一张芯片订单。
每一个新功能都会变成一份数据中心租约。
整个行业开始看起来像一系列同心圆,越往中心走,利润就越集中。所以,也许你是一个一直关注AI模型的软件工程师,也许你是一个在300美元时买入英伟达的零售投资者,正在试图弄清楚接下来该怎么做。也许你是一个在印度的人,远距离观看整个革命,想知道这一切如何与自己的投资组合相关。(或者也许你是这三者的结合,这也是最有趣的立场。)无论你在哪里,原则都是一样的。消费者看到的是产品,投资者看到的是供应链。最优秀的投资者在产品发货之前就能看到供应链。当然,回头看这一切都显得井然有序,但当时并非如此。我花了几个星期来回反复思考这个问题。我不得不放弃很多来自SaaS时代的模式匹配思维,那时大部分价值集中在应用层。我一直想找到“下一个OpenAI”,而我应该关注的是OpenAI给谁开支票。AI在结构上与SaaS不同。计算需求如此庞大,以至于基础设施层在这个周期阶段捕获了更多的价值。理解技术堆栈会改变你解读每个头条的方式,改变你评估每个公司的方式,改变你如何分配资本。我将会写更多类似的内容,深入探讨投资、AI,以及财富如何真正流动背后的系统。如果你不想依赖算法来展示下一个机会,最好的方法是关注并开启通知。V - 投资者的地图:逐层解析 好了,内容有点长了,我会加快速度。
以下是AI技术栈每一层的详细拆解:每一层发生什么、关键参与者是谁,以及投资机会在哪里。请继续看下去。
第1层:能源
AI数据中心极其耗电。一次大型AI训练运行所消耗的电力,可能相当于一个小城镇一整年的用电量。预计到2026年,这些设施每年将消耗约90太瓦时的电力,大约是2022年水平的十倍。
这就形成了一个非常直接的投资逻辑:谁能够为数据中心生产、传输并提供稳定电力,谁就会受益。包括位于主要数据中心集群附近的核能、天然气和可再生能源公司;拥有多余电力容量的公用事业公司;以及建设电网基础设施的公司。
Jensen Huang在2025年10月直言不讳地表示:“数据中心自发电的发展速度可能会远快于把电力接入电网。”
一些公司已经在建设专门为数据中心配套的发电设施,直接连接到数据中心本身,完全绕过电网系统。这一点让我很惊讶。
这些科技公司实际上正在变成自己的公用事业供应商。
受益者包括:公用事业公司(尤其是拥有核电能力的公司)、独立电力生产商、制造变压器、开关设备以及其他电力基础设施的企业。
在印度,随着超大规模数据中心园区在亚洲不断扩张,电力设备和输电领域的公司也将从中受益。
第2层:芯片
这是大多数人听说过的一层,因为Nvidia。但这一层并不只是某一家公司那么简单。
芯片层本身也包含多个子层,每个子层都有不同的竞争结构。
在最顶层,是设计者——那些架构芯片的公司:Nvidia(GPU)、AMD(GPU 和 CPU)、Broadcom(定制 ASIC)、Qualcomm,以及越来越多的超大规模云厂商本身(Google 的 TPU、Amazon 的 Trainium、Microsoft 的 Maia)。
接下来是制造商。TSMC 在晶圆代工领域占据主导地位,全球代工市场份额接近 70%。Samsung 以 7.2% 的份额远远排在第二位。Intel 正在尝试重建其代工业务,但这是一个为期多年的项目,结果并没有保证。
然后是设备层,也就是那些制造“制造芯片机器”的公司。ASML 是地球上唯一一家生产制造最先进芯片所需的极紫外光(EUV)光刻机的公司。Applied Materials、Lam Research 和 Tokyo Electron 也与其并列处于这一层级。
再往下是存储(AI 模型需要海量的高带宽内存,而 SK Hynix、Samsung 和 Micron 是这个领域最关键的三家公司)以及封装(例如 TSMC 的 CoWoS 等先进芯片封装技术已经成为真正的瓶颈)。
这种集中度是我反复思考的一点。Nvidia 在 AI 数据中心 GPU 市场的份额估计达到 92%。TSMC 为 Nvidia、AMD、Broadcom、Qualcomm、Apple 以及几乎所有其他主要芯片设计公司制造芯片。ASML 则是全球唯一的 EUV 光刻机供应商。
一家公司负责设计。一家公司负责制造。一家公司负责制造制造设备。
这种程度的集中既是一个投资逻辑,也是一个地缘政治风险。而我认为,没有足够多的人同时认真思考这两个概念。
第三层:云和数据中心
这就是这些芯片所在的地方。
大规模的仓储级设施,内部布满服务器,通过高速网络互联,并由越来越复杂的热管理系统保持冷却(液冷正逐渐成为标准配置,而非可有可无)。市场被三大超大规模云服务商主导:亚马逊云服务(Amazon Web Services,市场份额31%)、微软 Azure(24%)和谷歌云(Google Cloud,11%)。甲骨文(Oracle)在这一领域也在积极扩张,目标是在2026年实现500亿美元的资本支出。但云层的深度远不止这些超大规模云服务商。富士康(鸿海)目前组装了全球约40%的 AI 服务器。Arista Networks 和 Credo Technology(其股票在2025年因高能效数据传输上涨了117%)构建了连接一切的网络基础设施。Vertiv 负责液冷。像 Equinix 和 Digital Realty 这样的数据中心房地产投资信托(REITs)拥有土地和建筑。甚至还有人负责浇筑混凝土。每一层都有自己的供应链。根据美国银行的估算,超大规模云服务商在2026年将把90%的运营现金流用于资本支出,而2025年这一比例为65%。摩根士丹利预计,这些公司今年将借款超过4,000亿美元来资助建设,比2025年的1,650亿美元多一倍还多。当我第一次看到这个数字时,简直目瞪口呆——单单一年就发行4,000亿美元债务,只是为了建造计算机仓库。
第四层:模型
这是“大脑”层。负责训练和构建实际 AI 模型的公司。知名企业包括:OpenAI(GPT 系列,年经常性收入超过200亿美元)、Anthropic(Claude,据报道到2026年初年化收入约190亿美元)、Google DeepMind(Gemini)、Meta AI(Llama,开源)、Mistral,以及 xAI(埃隆·马斯克的公司,开发 Grok)。这一层令我着迷,因为它既是最受关注的,也是最不盈利的。
OpenAI的收入增长速度前所未见,但到2026年,它将烧掉170亿美元现金。Anthropic增长迅速,但严重依赖大规模的融资轮(2026年初以1700亿美元估值筹集了50亿美元)。商业模式问题是结构性的:模型在增加计算资源支出的同时会变得更好,但这种支出增长的速度超过了收入的增长。这有点像经营一家餐厅,每道菜都需要比上一道更贵的食材,但顾客却期望价格保持不变。利润空间持续压缩。我不知道这种情况什么时候会改变,也许永远不会改变。对于投资者来说,这一层是高风险、高回报。大多数这些公司是私有的。你对公开市场的曝光主要通过托管这些公司的云服务提供商(微软拥有OpenAI的大量股份并在Azure上运行其计算)以及通过在训练过程中使用的芯片公司来实现。
第五层:应用层
这是你每天看到的层级。ChatGPT。由Gemini驱动的Google搜索。Office中的Microsoft Copilot。银行的AI欺诈检测。Netflix推荐系统。你手机的照片增强。应用层是最广泛且最拥挤的层级。成千上万的初创公司和老牌企业正在这里建设。它最终将成为总可寻址市场最大的层级(一些估计认为到2030年代初将超过2万亿美元),但目前它也是利润最薄且谁将获胜最不确定的层级。数据是这一层的区分因素。拥有独特、专有数据的公司将建立持久的竞争优势。Salesforce拥有企业CRM数据。Bloomberg拥有金融数据。Epic拥有医疗记录。
拥有这种专有数据壁垒的公司,可以以通用聊天机器人无法做到的方式对AI模型进行微调。对于投资者来说,应用层是最终带来最大收益的地方,但也是大多数资本会被消耗的地方。大多数AI初创公司将会失败。那些幸存下来的公司将会迅速增长。在未来3到5年,最佳回报可能是这样的:现在投资基础设施,之后投资应用程序。最聪明的资本已经根据这一趋势进行了布局。那些真正能够在第5层获胜的公司,是那些掌握别人无法获取的数据的公司。而且其中大多数公司现在甚至还不自称为AI公司。
VI - “但这难道不只是一个泡沫吗?”
投资者的主要问题,甚至他最大的敌人,很可能就是他自己。
~ 本杰明·格雷厄姆
让我直接谈谈这个最核心的问题。
“但是互联网泡沫破裂怎么办?这难道不是同样的事情吗?巨额的基础设施投入,没有利润,所有人都被炒作裹挟?”
这是个很公平的问题。它值得一个严肃的回答。
区别在这里。
在互联网泡沫时期,公司是在为尚未真正出现的需求建设基础设施。他们为仍然使用拨号上网的互联网用户建设光纤网络和网页服务器。
基础设施建好了,但需求要再过5到7年才真正出现,而这期间的一切都被清算了。
到了2026年,AI需求已经真实存在。
NVIDIA的芯片生产速度已经跟不上需求。
TSMC的先进封装产能已经全部售罄。
云计算租赁价格正在上涨,而不是下降。
仅在2025年3月至10月之间,OpenAI每周活跃用户就增加了4亿。
这些模型正在被使用。
算力正在被消耗。
客户正在付费。
这并不意味着没有风险。
风险是巨大的。
而且我思考这个问题的次数,比我愿意承认的还要多。
尤其是三个方面:
第一,资本错配。
2026年,各家公司在数据中心上的支出超过6500亿美元。
如果AI服务带来的收入增长不够快,无法证明这些支出的合理性,那么其中一些公司将面临严重的利润率压缩。
亚马逊今年的自由现金流甚至可能会变成负数。
那可是亚马逊。
几乎发明了云计算的那家公司。
第二,集中度风险。
AI供应链的集中程度危险地高。
TSMC生产了全球近70%的芯片。
ASML是EUV光刻机的唯一供应商。
NVIDIA设计了92%的AI数据中心GPU。
任何中断(地缘政治、自然灾害或竞争因素)都可能波及整个产业链。台湾新竹的一次地震,就可能让全球 AI 发展倒退数年。光是想想这个就应该让你感到不安。
第三,DeepSeek 的问题。2025 年 1 月,中国 AI 实验室 DeepSeek 发布了一款模型,其性能接近前沿水平,但训练成本仅为一小部分。这挑战了“投入越多就能获得更好 AI”的假设。如果开源且高效的模型继续缩小差距,基础设施支出论点的力度就会减弱。我并不认为 DeepSeek 彻底推翻了这一论点,但它引入了一个以前不存在的变量,而这样的变量不会轻易消失。
不过,这里是我不断回到的框架。麦肯锡估计,到 2030 年,全球数据中心累计投资可能达到 6.7 万亿美元。普华永道估计,到 2030 年,AI 可能为全球 GDP 贡献 15.7 万亿美元。IDC 预测,到 2030 年,AI 解决方案和服务将带来 22.3 万亿美元的累计影响。即便这些数字有 50% 的误差,我们仍在谈论自互联网以来最大的一次技术驱动经济变革。问题在于规模,而不是方向。
我不断听到有人说:“我对 AI 持怀疑态度”,仿佛这是一个立场。没问题,对模型持怀疑态度,对时间表持怀疑态度,但不要对供应链一无所知。这是两码事。前者是一种健康的智力态度,后者会让你亏钱。
五年后,赢得这一轮周期的名字会显得显而易见——总是这样。现在的游戏,是在所有人赶上之前先看清结构。
VII - 在正确的层级玩游戏
把 AI 想象成一个有五个层级的视频游戏,层层叠加。
一级(能源)是教程阶段。基础、朴素,几乎不可能输,只要你玩就行。低风险,稳定回报。可以把它看作是那些永远不会死、但总是付钱的任务发布者。
二级(芯片)是Boss战。最高的力量集中度,最高的利润率,但也是最容易受到破坏和地缘政治风险影响的层级。奖励巨大,但难度设置为困难。
三级(云)是多人服务器。每个人都在这里玩。超级规模公司是服务器管理员。他们从每一项中抽成。
四级(模型)是PvP竞技场。残酷的竞争,快速的创新,大多数玩家会被淘汰。只有最强大的才能生存。
五级(应用)是开放世界。无限的可能性,但没有保证的奖励。你必须自己寻找任务。
元策略很简单。你不必玩所有五个级别。大多数人试图玩第五级,因为它最显眼。聪明的资金在第二级和第三级上耕作,因为现在那里的经验值最高。你在这五个层级中的位置决定了你应该关注什么。
对于非技术人员:你不需要理解GPU是如何工作的。你需要理解的是,有人必须制造它们,有人必须安置它们,有人必须为它们提供电力。而这些“某些人”是上市公司,你可以阅读他们的季度财报。
对于技术人员,你已经知道模型在变得更好。你可能低估的是,物理限制(电力、散热、芯片封装)正迅速成为制约瓶颈。未来十年的AI进展将在工程领域胜出,而不是在架构论文中。
对于投资者而言,人工智能价值链是五个层级依次堆叠的,每个层级都有不同的风险回报特征、不同的时间跨度和不同的赢家。把“人工智能”当作一个单一行业来看,就像把“科技”在1998年当作一个单一行业来看一样。在“人工智能”内部,最好的结果和最差的结果之间的差距是巨大的。这种情况不会永远持续下去。随着时间的推移,基础设施的建设将会成熟,应用层将会整合,价值将会向上转移,就像互联网时代一样。亚马逊、谷歌和 Facebook(互联网时代的应用层)最终获得了比光纤电缆公司和服务器制造商更多的价值。但我们现在还没到那一步,人工智能还处于基础设施阶段,处于“铲子和锄头”的阶段,而这些“铲子和锄头”正在赚取大量财富。那些了解整个价值链的人将会在转变发生之前就预见到它们,其他人则会一次又一次地被真正的财富去向所惊讶。十年后,理解人工智能的价值链将会像理解资产负债表一样基础。了解这个价值链,映射各个层级,追踪资本流向,这才是游戏的规则。~ Anish Moonka
显示英文原文 / Show English Original
Most people think AI is a chatbot. I get it. You open ChatGPT, ask it to fix your email, and it does. Feels like magic. You walk away thinking you understand what's going on. But that's like swiping a credit card at a restaurant and thinking you understand how Visa makes money. You used the product. You didn't see the system. I spent the better part of last year trying to figure out where the real money in AI actually flows, and the honest answer is that it took me embarrassingly long to stop looking at the wrong layer. I kept staring at ChatGPT and Claude and Gemini, the stuff you can touch, while $700 billion was being quietly deployed into infrastructure I couldn't even name. Chips I'd never heard of. Packaging technologies with acronyms that sound made up. Cooling systems. Power plants. Concrete is being poured in Texas, Iowa, and Hyderabad. Nobody I know was talking about any of this a year ago. They are now. This article is going to be long. If you don't have time right now, bookmark it and come back. I want to walk through the entire AI value chain, every layer, from the electricity that powers the data centres to the app on your phone, and I want to do it in a way that makes sense even if you've never read an annual report in your life. I'll explain the jargon when I use it. I'm going to attach real numbers to every claim. And I'm going to be honest about the parts I'm still not sure about, because there are a few. Let's begin. I - The five-layer cake (and why nobody talks about the bottom four) AI is infrastructure. Just like the internet, just like electricity, it needs factories. ~ Jensen Huang The way most people understand AI goes something like this: a smart computer answers questions That's like saying the internet is "a place where you watch videos." Technically not wrong. But it misses the entire point. Jensen Huang, CEO of Nvidia, described AI at Davos in January 2026 as a five-layer system. Energy. Chips. Cloud. Models. Applications. He called the whole thing "the largest infrastructure buildout in human history." Think about that word for a second. Infrastructure. Roads. Power grids. Water systems. These are the things that make civilisation work, and nobody thinks about them until they break. AI is becoming that kind of thing. Invisible, essential, and enormously expensive to build. I call this the AI Stack. Five layers, stacked on top of each other, where each layer feeds the one above it, and the money flows both ways. Here's the simplest version I can give you: Energy. You need electricity to power the computers. Lots of it. Chips. You need specialised processors to do the math. These are not your laptop's brain. Cloud. You need massive warehouses full of these chips, connected by insanely fast networks. Models. You need the actual AI software, the "brain" that learns patterns from data. Applications. You need products people actually use. ChatGPT. Google Search. Your bank's fraud detector. Every conversation about AI that focuses only on Layer 5 is missing 80% of the picture. And here's the part that matters if you're an investor, a founder, or just someone trying to understand where the world is going: the money doesn't flow evenly across these layers. It concentrates. It compounds. And right now, it's concentrating in places most people aren't looking. II - Follow the money (it's not where you think) Everyone's mind goes to the application layer. ChatGPT. Copilot. Claude. Perplexity These are the products you touch, so they feel like the whole story. But here's what everyone misses. In 2026, the four biggest cloud companies (Amazon, Microsoft, Google, and Meta) are on track to spend somewhere between $650 billion and $700 billion on capital expenditures. Combined. In a single year. That is roughly equal to Switzerland's GDP. And almost 75% of it, about $450 billion, is going directly into AI infrastructure. Not into chatbots. Not into apps. Into buildings, chips, cables, and cooling systems. Nobody talks about this stuff at cocktail parties. That's how you know it's where the money is. Because think about it. Before anyone can use ChatGPT, someone has to build a data centre the size of a shopping mall, fill it with tens of thousands of specialised processors, connect them with networking equipment that costs more than most companies are worth, and then feed the whole thing enough electricity to power a small city. Every single day. That's Layer 1 through Layer 3. The invisible layers. The layers where the serious capital is being deployed. "But what about OpenAI? Aren't they making billions?" They are. OpenAI hit $20 billion in annualised recurring revenue by the end of 2025, up from $6 billion a year earlier and $2 billion the year before that. That's 10x growth in two years. No company in history has scaled revenue that fast from that base. But here's the catch. OpenAI burned roughly $9 billion in cash in 2025 and projects $17 billion in cash burn for 2026. Their inference costs (the cost of actually running the AI when you ask it a question) hit $8.4 billion in 2025 and are projected to reach $14.1 billion in 2026. They don't expect to turn cash-flow positive until 2029 or 2030 So where does that burned cash go? It flows downward through the stack. To Microsoft Azure (OpenAI pays Microsoft 20% of its total revenue through 2032). To Nvidia for chips. To the companies building and equipping data centres. To the power companies generating electricity. There's something almost circular about it if you stare at it long enough. Microsoft invests in OpenAI. OpenAI spends that money on Azure. Azure uses the revenue to buy more Nvidia chips. NVIDIA reports record earnings. Everyone celebrates. And the cash keeps flowing downhill. Most of the users are at the top of the stack. Most of the profit is at the bottom. That disconnect is the entire thesis. This is the first lesson of the AI value chain: revenue flows up, capital flows down. III - You've seen this movie before All of humanity's problems are engineering problems, and engineering problems can be solved. ~ Buckminster Fuller If you want to understand what's happening with AI, study what happened with electricity between 1880 and 1920. When Thomas Edison built the first commercial power station in 1882 on Pearl Street in Manhattan, people thought electricity was a novelty. A fancy way to light a room. Why would anyone need this when gas lamps worked perfectly well? Within 40 years, electricity had reorganised every industry on earth. Manufacturing. Transportation. Communication. Medicine. Entertainment. The companies that won weren't the ones that invented the lightbulb. They were the ones who built the power plants, laid the copper wire, and manufactured the generators. General Electric. Westinghouse. The utility companies. The copper miners. The builders. The same pattern is playing out with AI, just compressed into years instead of decades AI → data centres → chips → raw materials → energy Electricity → factories → machines → raw materials → coal/water The arrow progression is almost identical. And the winners, again, are not primarily at the application layer. They're at the infrastructure layer. I call this Infrastructure Gravity. Every time a new computing platform emerges, initial wealth creation happens in the picks and shovels. The applications come later. The applications get all the press. But the infrastructure gets all the margin. NVIDIA posted $215.9 billion in full-year revenue for fiscal 2026 (ending January 2026), up 65% from the prior year. Their data centre segment alone did $62.3 billion in the final quarter, growing 75% year-over-year. That single segment now represents over 91% of Nvidia's total revenue. Think about that: a company doing $68 billion in a quarter, and nine out of ten of those dollars come from one business line. TSMC, the company that physically manufactures Nvidia's chips (and nearly everyone else's), captured almost 70% of the global foundry market in 2025, with $122.5 billion in sales. Samsung, the nearest competitor, had 7.2%. That is dominance at a level that would make Standard Oil uncomfortable. The infrastructure always wins first. The question is how long the window stays open. Ask anyone what the internet revolution was about, and they'll say Google, Amazon, and Facebook. Ask where the early money was actually made, and the answer is Cisco, Corning, and the companies laying fibre. Same story. Different decade. IV - The part nobody wants to hear The stock market is a device for transferring money from the impatient to the patient. ~ Charlie Munger I'll be honest about something When I first started paying attention to AI as an investor, I made the same mistake most people make. I looked at the application layer. I saw ChatGPT growing. I saw Anthropic raising billions. I thought: the AI companies will win, so invest in AI companies. Three things changed my mind. And they happened in sequence, which is important because each one built on the last. First, I noticed that nearly every "AI company" was hemorrhaging cash. OpenAI, Anthropic, Mistral, xAI. All are burning faster than they earn. Not because they're bad businesses, but because the computing costs are structural. Every time you ask an AI model a question, it costs real money to generate that answer. And the smarter the model gets, the more compute it needs, which means it costs more to run. The companies you think of as "AI winners" are actually the ones spending the most. Second, I realised that the infrastructure companies were printing money at margins I hadn't seen since the early days of Google. NVIDIA's gross margins were hovering around 75%. TSMC was expanding capacity and raising prices simultaneously because demand so dramatically exceeded supply. These companies don't have a "when will we monetise" problem. They have a "we literally cannot build fast enough" problem. Those are very different problems to have. Third, and this was the uncomfortable one, I realised I'd been thinking about AI like a consumer rather than an engineer. The consumer sees the app. The engineer sees the stack. Once you see the stack, you can't unsee it. Every AI announcement becomes a capex announcement. Every model improvement becomes a chip order. Every new feature becomes a data centre lease The whole industry starts looking like a series of concentric circles, and the further toward the centre you go, the more concentrated the profits become. So, maybe you're a software engineer who's been following AI models. Maybe you're a retail investor who bought Nvidia at $300 and is trying to figure out what's next. Maybe you're someone in India watching this entire revolution from a distance, wondering how any of it connects to your portfolio. (Or maybe you're all three, which is the most interesting position to be in.) Wherever you are, the principle is the same. The consumer sees the product. The investor sees the supply chain. The best investors see the supply chain before the product even ships. Of course, this all sounds neat and tidy in hindsight. It wasn't. I spent weeks going back and forth on this. I had to unlearn a lot of pattern-matching from the SaaS era, where most of the value accrued at the application layer. I kept wanting to find the "next OpenAI" when I should have been looking at who OpenAI was writing cheques to. AI is structurally different from SaaS. The compute requirements are so massive that the infrastructure layer captures more value, at least in this phase of the cycle. Understanding the stack changes how you read every headline. It changes how you evaluate every company. It changes how you allocate capital. I'm going to be writing a lot more like this. Deep dives into investing, AI, and the systems behind how wealth actually moves. If you don't want to rely on the algorithm to show you the next one, the best move is to follow and turn on notifications. V - The investor's map: a layer-by-layer breakdown Okay, this is getting long, so I'm going to speed things up Here's the exact breakdown of each layer of the AI Stack, what happens there, who the important players are, and where the investment opportunities sit. Stick with me. Layer 1: Energy AI data centres are extraordinarily power-hungry. A single large AI training run can consume as much electricity as a small town uses in a year. These facilities are projected to consume around 90 terawatt-hours of electricity annually by 2026, roughly a tenfold increase from 2022 levels. This creates a straightforward investment thesis: whoever can generate, transmit, and deliver reliable power to data centres will benefit. Nuclear, natural gas, and renewable energy companies near major data centre clusters. Utility companies with excess capacity. Companies are building grid infrastructure. Jensen Huang said it plainly in October 2025: "Data centre self-generated power could move a lot faster than putting it on the grid." Companies are already building dedicated power generation attached directly to their data centres, bypassing the grid entirely. That part surprised me. These tech companies are basically becoming their own utility providers. Who benefits: utility companies (especially those with nuclear capacity), independent power producers, companies manufacturing transformers, switchgear, and other electrical infrastructure. In India, companies in the power equipment and transmission space stand to benefit as hyperscaler campuses expand across Asia. Layer 2: Chips This is the layer most people have heard of, because of Nvidia. But it's more complex than one company. The chip layer has its own sub-layers, each with a different competitive structure
At the top, you have the designers, the companies that architect the chips: Nvidia (GPUs), AMD (GPUs and CPUs), Broadcom (custom ASICs), Qualcomm, and increasingly the hyperscalers themselves (Google's TPUs, Amazon's Trainium, Microsoft's Maia). Then you have the manufacturers. TSMC dominates fabrication, with nearly 70% of the global foundry market share. Samsung is a distant second at 7.2%. Intel is trying to rebuild its foundry business, but that's a multi-year project with no guaranteed outcome. Then there's the equipment layer, the companies that make the machines that make the chips. ASML is the only company on earth that makes the extreme ultraviolet lithography machines required for the most advanced chips. Applied Materials, Lam Research, and Tokyo Electron sit alongside them. Below that, you've got memory (AI models need enormous amounts of high-bandwidth memory, and SK Hynix, Samsung, and Micron are the three players that matter) and packaging (advanced chip packaging like TSMC's CoWoS technology has become a genuine bottleneck). The concentration here is something I keep coming back to. Nvidia holds an estimated 92% share of the AI data centre GPU market. TSMC manufactures chips for Nvidia, AMD, Broadcom, Qualcomm, Apple, and nearly every other major chip designer. ASML is literally the only supplier of EUV lithography machines on the planet. One company designs. One company builds. One company makes the machine that builds. That level of concentration is both an investment thesis and a geopolitical risk. And I don't think enough people sit with both of those ideas at the same time. Layer 3: Cloud and data centres This is where the chips live Massive warehouse-scale facilities packed with servers, connected by high-speed networking, and kept cool by increasingly elaborate thermal management systems (liquid cooling is becoming standard, not a nice-to-have). The market is dominated by three hyperscalers: Amazon Web Services (31% market share), Microsoft Azure (24%), and Google Cloud (11%). Oracle is aggressively growing in this space too, with a $50 billion capex target for 2026. But the cloud layer goes much deeper than the hyperscalers. Foxconn (Hon Hai) now assembles about 40% of the world's AI servers. Arista Networks and Credo Technology (whose stock rose 117% in 2025 on energy-efficient data transfer) build the networking infrastructure that connects everything. Vertiv handles liquid cooling. Data centre REITs like Equinix and Digital Realty own the land and buildings. Someone even has to pour the concrete. Every layer has its own supply chain. The hyperscalers are spending 90% of their operating cash flow on capex in 2026, according to Bank of America estimates. That's up from 65% in 2025. Morgan Stanley expects these companies to borrow over $400 billion this year to fund the buildout, more than double 2025's $165 billion. That number stopped me in my tracks when I first read it. $400 billion in debt issuance in a single year, just to build computer warehouses. Layer 4: Models This is the "brain" layer. The companies that train and build the actual AI models. The big names: OpenAI (GPT series, $20B+ ARR), Anthropic (Claude, reportedly around $19B annualised revenue by early 2026), Google DeepMind (Gemini), Meta AI (Llama, open source), Mistral, and xAI (Elon Musk's company, building Grok). This layer fascinates me because it's simultaneously the most hyped and the most unprofitable OpenAI's revenue is growing at a pace we've never seen, but it's burning $17 billion in cash in 2026. Anthropic is growing fast but is heavily reliant on massive funding rounds ($5B at a reported $170B valuation in early 2026). The business model problem is structural: models get better when you spend more on compute, but that spending grows faster than revenue. It's a bit like running a restaurant where every dish requires more expensive ingredients than the last, but customers expect the price to stay the same. The margins stay compressed. I don't know when that changes. Maybe it doesn't. For investors, this layer is high risk, high potential reward. Most of these companies are private. Your public market exposure comes through the cloud providers who host them (Microsoft owns a large stake in OpenAI and runs its compute on Azure) and through the chip companies whose products get consumed during training. Layer 5: Applications This is the layer you see every day. ChatGPT. Google Search powered by Gemini. Microsoft Copilot in Office. AI-powered fraud detection at your bank. Netflix recommendations. Your phone's photo enhancement. The application layer is the widest and most crowded. Thousands of startups and incumbents are building here. It will eventually be the largest layer by total addressable market (some estimates put it above $2 trillion by the early 2030s), but right now it's also the layer with the thinnest margins and the greatest uncertainty about who will win. The differentiator at this layer is data. Companies with unique, proprietary data will build durable advantages. Salesforce has enterprise CRM data. Bloomberg has financial data. Epic has healthcare records Companies sitting on that kind of proprietary data moat can fine-tune AI models in ways that a generic chatbot cannot. For investors, the application layer is where the biggest upside eventually lives, but also where most of the capital will be destroyed. Most AI startups will fail. The ones that survive will compound aggressively. The best returns over the next 3 to 5 years probably look like this: infrastructure now, applications later. The smartest capital is already positioned accordingly. The companies that will actually win at Layer 5 are the ones sitting on data nobody else can get. And most of them don't even call themselves AI companies yet. VI - "But isn't this just a bubble?" The investor's chief problem, and even his worst enemy, is likely to be himself. ~ Benjamin Graham Let me address the elephant directly. "But what about the dot-com bust? Isn't this the same thing? Massive infrastructure spending, no profits, everyone caught up in hype?" Fair question. It deserves a serious answer. Here's the difference. During the dot-com era, companies were spending on infrastructure for demand that hadn't yet materialized. They were building fibre-optic networks and web servers for an internet audience that was still on dial-up. The infrastructure was built, demand didn't materialize for another 5 to 7 years, and everything in between got liquidated. By 2026, AI demand is already here. NVIDIA can't make chips fast enough. TSMC's advanced packaging capacity is sold out. Cloud computing rental prices are rising, not falling. OpenAI added 400 million weekly active users between March and October 2025 alone. The models are being used. The compute is being consumed. The customers are paying. That doesn't mean there's no risk. There is an enormous risk. And I think about it more than I'd like to admit. Three things in particular: First, capital misallocation. Companies are spending $650 billion+ on data centres in 2026. If the revenue from AI services doesn't materialise fast enough to justify that spend, some of these companies will face serious margin compression. Amazon's free cash flow may actually go negative this year. That's Amazon. The company that basically invented cloud computing. Second, concentration risk. The AI supply chain is dangerously concentrated. TSMC fabricates nearly 70% of the world's chips. ASML is the sole supplier of EUV machines. NVIDIA designs 92% of AI data centre GPUs Any disruption (geopolitical, natural disaster, or competitive) could ripple through the entire stack. A single earthquake in Hsinchu, Taiwan, could set back global AI development by years. That thought should make you uncomfortable. Third, the DeepSeek question. In January 2025, Chinese AI lab DeepSeek released a model that approached frontier performance at a fraction of the training cost. This challenged the assumption that greater spend automatically equates to better AI. If open-source and efficient models continue to close the gap, the infrastructure spending thesis weakens. I don't think DeepSeek killed the thesis. But it introduced a variable that wasn't there before, and variables like that don't just go away. But here's the frame I keep coming back to. McKinsey estimates that cumulative data centre investment could reach $6.7 trillion globally by 2030. PwC estimates AI could contribute $15.7 trillion to global GDP by 2030. IDC projects AI solutions and services will generate $22.3 trillion in cumulative impact by 2030. Even if those numbers are 50% wrong, we're still talking about the largest technology-driven economic shift since the internet. The question is about magnitude, not direction. I keep hearing people say "I'm sceptical about AI," as if it's a position. Fine. Be sceptical about the models. Be sceptical about the timelines. But don't be ignorant about the supply chain. Those are different things. One is a healthy intellectual posture. The other will cost you money. Five years from now, the names that won this cycle will feel obvious. They always do. The game right now is seeing the structure before everyone else catches up. VII - Play the game at the right layer Think about AI as a video game with five levels, stacked on top of each other Level 1 (Energy) is the tutorial stage. Essential, unglamorous, and almost impossible to lose if you play it. Low risk, steady returns. Think of these as the quest givers who never die but always pay. Level 2 (Chips) is the boss fight. Highest concentration of power, highest margins, but also the layer most exposed to disruption and geopolitical risk. The rewards are massive, but the difficulty is set to hard. Level 3 (Cloud) is the multiplayer server. Everyone plays here. The hyperscalers are the server administrators. They take a cut of everything. Level 4 (Models) is the PvP arena. Brutal competition, rapid innovation, and most players get eliminated. Only the best-equipped survive. Level 5 (Applications) is the open world. Infinite possibilities, but no guaranteed loot. You have to find your own quests. The meta-strategy is simple. You don't have to play all five levels. Most people try to play Level 5 because it's the most visible. The smart money is farming Level 2 and Level 3 because that's where the XP is highest right now. Where you are in the stack determines what you should focus on. For the non-technical: you don't need to understand how a GPU works. You need to understand that someone has to make them, someone has to house them, and someone has to power them. And those "someones" are publicly traded companies with quarterly earnings you can read. For the technical, you already know the models are getting better. What you might be underestimating is how fast the physical constraints (power, cooling, chip packaging) are becoming the binding bottleneck. The next decade of AI progress will be won or lost in engineering, not in architecture papers For the investor, the AI value chain is five trades stacked on top of each other, each with a different risk-reward profile, different time horizon, and different set of winners. Treating "AI" as a single sector is like treating "technology" as a single sector in 1998. The spread between the best and worst outcomes within "AI" is enormous. This won't last forever. At some point, the infrastructure buildout will mature. The application layer will consolidate. And the value will shift upward in the stack, just like it did with the internet. Amazon, Google, and Facebook (the application layer of the internet era) eventually captured more value than the fiber-optic cable companies and server manufacturers. But we're not there yet with AI. We're in the infrastructure phase. The picks-and-shovels phase. And the picks and shovels are printing money. The people who understand the full stack will see the transitions before they happen. Everyone else will be surprised, over and over, by where the money actually goes. In 10 years, understanding the AI stack will be as basic as understanding a balance sheet. Learn the stack. Map the layers. Follow the capital. That's the game. ~ Anish Moonka