注册并分享邀请链接,可获得视频播放与邀请奖励。

搜索结果 xAI
xAI 贴吧
一个关键词就是一个贴吧,路径全站唯一。
创建贴吧
用户
未找到
包含 xAI 的推特
xAI 上周裁员约 10 人,近期离职人员包括 Mistral AI 创始成员。 就在马斯克旗下企业获得以 600 亿美元收购 Cursor 这家代码初创公司的收购期权仅数周后,Cursor 已开始在 SpaceX 旗下 AI 部门逐步发挥影响力。知情人士称,随后 xAI 接连出现人员离职,上周五还进行了裁员。
显示更多
XAI:與Anthropic建立新的運算合作關係。
XAI: NEW COMPUTE PARTNERSHIP WITH ANTHROPIC
#xAI# 计划以 600 亿美元收购编程开发工具 #Cursor,或支付# 100 亿美元进行密切合作。Cursor 与 xAI 的合作对双方而言都有优势,xAI 可以借助 Cursor 杀入 AI 编程工具市场,Cursor 则可以获得充足资金与 Claude Code 以及 OpenAI Codex 竞争:
显示更多
为什么xAI要把数据中心Colossus1租给Authropic?这篇推文应该是分析最到位的,核心逻辑是: xAI目前总共持有大约55 万+个GPU(以H100等效性能为基础),而Colossus1(22 万个)仅占总可用容量的约40%,且是一个混合 H100/H200/GB200 的训练集群。这种混合集群并不是合适训练(不同代际GPU见通信延迟很大),但是非常适合推理(推理需要远没有那么紧密同步的 GPU 间通信)。 恰恰Authropic现在最需要推理算力,而且一家就能把Colossus1的算力全部吃掉。而且Anthropic 作为单一租户占用所有22万个 GPU,多租户下出现的网络交换抖动(意外延迟)消失了。双方的技术弱点最终几乎完美互补。 老马把完全基于 Blackwell 构建的数据中心Colossus 2留给自己,用以训练xAI下一代大模型。 租赁出较旧的、混合代的 Colossus 1。作为一个混合H100/H200/GB200的训练集群,Colossus 1只能实现 11% 的MFU(利用率)。然而,一旦它被移交给单一推理客户,这个资产就转变为一个现金流资产,以大约每 GPU 小时 2.60 美元的价格出租(GPU 类型租赁率的加权平均)。对于 xAI 来说,本来是训练的“地狱集群”,在重新部署用于推理时变成了“金鹅”,每年带来50–60 亿美元的收入。 将这 60 亿美元与 xAI 的损益表对比时,其分量就更清晰了。将 xAI 的 1Q26 净亏损年化,大约每年 60 亿美元的亏损。换句话说,向 Anthropic 租赁 Colossus 1 产生的 50–60 亿美元年度收入,几乎完美对冲了 xAI 的亏损数字 不得不说、这是一次完美的合作:Authropic获得急缺的推理算力; SpaceXAI获得能弥补其AI业务年度亏损的现金流。
显示更多
Why did xAI hand over a 220,000-GPU cluster to Anthropic? The technical backdrop to xAI's decision to hand Colossus 1 over to Anthropic in its entirety is more interesting than it appears. xAI deployed more than 220,000 NVIDIA GPUs at its Colossus 1 data center in Memphis. Of these, roughly 150,000 are estimated to be H100s, 50,000 H200s, and 20,000 GB200s. In other words, three different generations of silicon are mixed together inside a single cluster — a "heterogeneous architecture." For distributed training, however, this configuration is close to a disaster, according to engineers familiar with the setup. In distributed training, 100,000 GPUs must finish a single step simultaneously before the cluster can advance to the next one. Even if the GB200s finish their computation first, the remaining 99,999 chips have to wait for the slower H100s — or for any GPU that has hit a stack-related snag — to catch up. This is known as the straggler effect. The 11% GPU utilization rate (MFU: the share of theoretical FLOPs actually realized) at xAI recently reported by The Information can be read as the numerical fallout of this problem. It stands in stark contrast to the 40%-plus MFU figures achieved by Meta and Google. The problem runs deeper still. As discussed earlier, NVIDIA's NCCL has traditionally been optimized for a ring topology. It works beautifully at the 1,000–10,000 GPU scale, but once you push into the 100,000-unit range, the latency of data traversing the ring once around becomes punishingly long. GPUs need to churn through computations rapidly to keep MFU high, but while they sit waiting endlessly for data to arrive over the network fabric, more than half of the silicon falls into idle. Google sidestepped this bottleneck with its own custom topology (Google's OCS: Apollo/Palomar), but xAI, by my read, has not yet reached that stage. Layer Blackwell's (GB200) "power smoothing" issue on top, and the picture comes into focus. According to Zeeshan Patel, formerly in charge of multimodal pre-training at xAI, Blackwell GPUs draw power so aggressively that the chip itself includes a hardware feature for smoothing power delivery. xAI's existing software stack, however, was optimized for Hopper and does not understand the characteristics of the new hardware; when it imposes irregular loads on the chip, the silicon physically destructs — literally melts. That means the modeling stack must be rewritten from scratch, which in turn means scaling is far harder than most of us imagine. Pulling all of this together points to a single conclusion. xAI judged that training frontier models on Colossus 1 simply was not efficient enough to be worthwhile. It therefore moved its own training workloads wholesale onto Colossus 2, built as a 100% Blackwell homogeneous cluster. Colossus 1, on the other hand — whose mixed architecture is far less crippling for inference, which parallelizes more forgivingly — was leased in its entirety to an Anthropic that desperately needed inference capacity. Many observers point to what looks like a contradiction: Elon Musk poured enormous capital into building Colossus, only to hand the core asset over to a direct competitor in Anthropic. Others read it as xAI capitulating because it is a "middling frontier lab." But these are surface-level reads. Look at the numbers and a different picture emerges. xAI today holds roughly 550,000+ GPUs in total (on an H100-equivalent performance basis), and Colossus 1 (220,000 units) accounts for only about 40% of the total available capacity. Colossus 2 — built entirely on Blackwell — is already operational and continuing to expand. Elon kept the all-Blackwell homogeneous cluster (Colossus 2) for himself and leased out the older, mixed-generation Colossus 1. In other words, he handed the pain of rewriting the stack — the MFU-11% debacle — to Anthropic, while keeping his own focus on training the next generation of models. The real point, then, is this. Elon's objective appears to be positioning ahead of the SpaceXAI IPO at a $1.75 trillion valuation, currently floated for as early as June. The narrative SpaceXAI now needs is that xAI — long the "sore finger" — is not merely a research lab burning cash, but a business with a "neo-cloud" model in the mold of AWS, capable of leasing surplus assets at high yields. From a cost-of-capital perspective, an "AGI cash incinerator" is far less attractive to investors than a "data-center landlord generating cash." As noted above, the most important detail of the Colossus 1 lease is that it is for inference, not training. Unlike training, inference requires far less tightly synchronized inter-GPU communication. Even when the chips are heterogeneous, the workload parcels out cleanly across them in parallel. The straggler effect — the chief weakness of a mixed cluster — is essentially neutralized for inference workloads. Furthermore, with Anthropic occupying all 220,000 GPUs as a single tenant, the network-switch jitter (unanticipated latency) that arises under multi-tenancy disappears. The two sides' technical weaknesses end up complementing each other almost exactly. One insight follows. As a training cluster mixing H100/H200/GB200, Colossus 1 was an asset that could only deliver an MFU of 11%. The moment it was handed over to a single inference customer, however, that asset transformed into a cash-flow asset rented out at roughly $2.60 per GPU-hour (a weighted average of the lease rates across GPU types). For xAI, what was a "cluster from hell" for training has become a "golden goose" minting $5–6 billion in annual revenue when redeployed for inference. Elon's genius, I would argue, lies not in the model but in this asset-rotation structure. The weight of that $6 billion becomes clearer when set against xAI's income statement. Annualizing xAI's 1Q26 net loss yields roughly $6 billion in losses per year. The $5–6 billion in annual revenue generated by leasing Colossus 1 to Anthropic, in other words, almost perfectly hedges xAI's loss figure. This single deal effectively pulls xAI to break-even. Heading into the SpaceXAI IPO, this functions as a core line of financial defense. From a cost-of-capital standpoint, if the image shifts from "research lab burning cash" to "infrastructure tollgate stably printing $6 billion a year," the entire tone of the offering can change. (May 8, 2026, Mirae Asset Securities)
显示更多
0
20
194
27
转发到社区
与xAI合作得到算力支持后,Anthropic宣布: 1)将 Pro、Max 和 Team 计划中 Claude Code 的 5 小时速率限制额度翻倍; 2)取消 Pro 和 Max 计划在高峰时段对 Claude Code 的额度削减限制; 3)大幅提高 Opus 模型的 API 速率限制。
显示更多
0
26
29
1
转发到社区
是 xAI 的推理利用率太低了,以至于外租给 Claude code 使用了吗?
We’ve agreed to a partnership with @SpaceX that will substantially increase our compute capacity. This, along with our other recent compute deals, means that we’ve been able to increase our usage limits for Claude Code and the Claude API.
显示更多
0
17
50
1
转发到社区
马斯克把xAI的团队拆了,xAI并入SpaceX,同时将超算集群租给Anthropic 1. xAI解散:xAI并入SpaceX,成为其AI部门。 2. 算力交易:SpaceX将22万张GPU租给Anthropic,改写AI算力格局。 3. 对抗OpenAI:马斯克通过支持Anthropic,间接对抗OpenAI 从财务回报来看,他这个动作绝对是扭亏为盈,把烧钱的大模型训练,变成了挣钱的现金流生意。在AI领域能否挣到钱这件事上,当下来看,依然是做大模型的,不如卖算力的。
显示更多
0
40
66
4
转发到社区
马斯克要把 xAI 并入 SpaceX。 事实证明,老马的舒适区还是在搞硬件 or 搞供应链。 他的神奇在于可以在整合供应链的各个环节,使用(1)领导能力(2)钞能力(3) PUA 能力,快速完成他的目标。 几条供应环节劈里啪啦一整合,就攒出来一个牛逼的硬件,比如说特斯拉,比如说 SpaceX 火箭,再比如说 Colossus 这个号称全球最大的算力中心。 他在孟菲斯建的这个 Colossus 算力中心,全球最大之一。 里面的 20 万 GPU 算力显卡,一部分由英伟达提供,另一部分是马斯克强行从特斯拉挪用过来的,再加上远超工厂旧址原有的电力供应也是马斯克搞定的,号称 122 天就建了起来。这就是老马的舒适区。 但是显然,马斯克的技术栈不包括搞软件。 xAI 在 11 位联创全部出走之后,如今遗憾地被马斯克划到了 SpaceX 架构底下,变成了 SpaceXAI(这起名马斯克味儿也太浓了)。 其实,如果马斯克要是软件能搞好的话,他也不至于跟 OpenAI 在那么早期就闹崩了。 曾几何时,当时 OpenAI 成立就是为了对抗 Google 的 DeepMind,是真的想搞非盈利的 AI 组织。 我看路透社有篇报道说,甚至 Sam Altman 还提议说成立一个五人委员会,成员包括: 1. 他自己 2. 马斯克 3. 比尔·盖茨 4. eBay 老板 5. Facebook 联创 说实话,这阵容听起来科幻味儿真的拉满了。我如果将来有机会写小说的话,我一定会把这个桥段作为其中一个 if 线。 这个五人委员会有一种“硅谷版共济会”的感觉,你知道吧? 当然了,就目前这个 GPT 5.5 和哈基米 3.1 的表现来看,OpenAI 确实达到了击败谷歌的目的,只是可惜也并不 open,也走了 close 路线。 马斯克和 Sam 的官司直接在打这个点。 考虑到 OpenAI 更让他感到恶心,所以马斯克现在决定和 Anthropic 合作,把闲置算力提供给 Cluade 用。 要知道之前马斯克其实对 Anthropic 还是风言风语比较多的,例如管人家叫:厌世、虚伪、盗窃,甚至嘲笑那个菊花 Logo。 我看有报道说,Colossus 这个算力中心显卡的利用率也才 1/10,大部分都在空转,反正自己的 xAI 现在用户量也不大,配一个这么大的算力中心完全是浪费,所以说马斯克只能捏着鼻子把这些算力给 Anthropic 去使用。 敌人的敌人就是朋友,可能老马也是这么想的。 你发现了没,搞到最后,马斯克 AI 之旅里,硕果仅存居然不是 Grok,而是他快速攒出来的算力中心 Colossus。 不过神仙打架,我们这些普通用户居然得到了一些实惠?😂 因为有了 Colossus 这么大一个算力中心的支持,Claude 终于决定多挤点牙膏,比方说: 1. 5 小时用量翻倍 2. CC 上降低峰值限速 只是可怜了 Grok,现在卡在 4.3 阶段,上不上下不下的就很难受。 考虑到 Grok @grok 在推特上给我们带来了太多的欢乐,真的希望 Grok 能够在 SpaceX 里面,匹配一个能够真正引领它的强大的新团队,努力回到御四家的位置上🥲。
显示更多
马斯克把xAI将解散了🤡 以后xAI不再作为一家独立公司运营 与SpaceX合并 更名为SpaceXAI
0
131
100
0
转发到社区