The new generation AI supercomputing chip architecture has been officially released, achieving significant breakthroughs in performance metrics. Compared to the previous generation, inference phase costs have been reduced to one-tenth, marking a turning point for the economics of large-scale model deployment. At the same time, the number of GPUs required for training has been cut by 75%, meaning enterprises can accomplish the same computational tasks with less hardware. Energy efficiency has increased fivefold, significantly reducing power consumption and heat dissipation under the same computing power.



Innovations at the technical architecture level are equally impressive—this is the first time confidentiality computing capabilities have been achieved at the rack level. The interconnection bandwidth between GPUs has reached an astonishing 260 TB/s, a data flow rate sufficient to support ultra-large-scale parallel computing scenarios. The entire platform has been thoroughly redesigned, abandoning traditional cable hoses and fan solutions in favor of a more compact and efficient hardware organization. The core engine consists of six modular components, offering greater flexibility for customization and expansion. The release of this generation will undoubtedly reshape the cost structure and deployment methods of the AI computing market.
View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • 9
  • Repost
  • Share
Comment
0/400
BtcDailyResearchervip
· 01-09 14:58
One-tenth of the cost? The friends in the mining industry must be panicking now. --- 260TB/s bandwidth... This number makes my head spin, but it seems like layoffs are coming again. --- Efficiency increased fivefold? This is a godsend for electricity bill enthusiasts. Starting to consider changing chips. --- Another round of iteration and reshuffling. This speed is really hard to keep up with. --- GPU needs to be cut by 75%... Huh, will the stocks of graphics card manufacturers fall? --- Modular design sounds good, but I'm worried it's just another marketing gimmick. We need to see actual benchmark scores to believe it. --- If this thing is really as powerful as the promotion claims, the entire AI computing market landscape will change. --- Reducing costs to one-tenth is truly outrageous. It must be tough for those who bought chips early. --- First time hearing about rack-level confidential computing. Feels like we're about to get cut again. --- Accelerate iteration quickly. It seems like the tech stack needs to be updated every three months.
View OriginalReply0
AirdropHermitvip
· 01-09 10:02
Whoa, one-tenth of the cost? Is this real? Feels like it's going to explode this time. This can improve efficiency by five times. Big players will have to start buying in madly. 260TB/s... Just hearing this number sounds outrageous. Can it really be achieved? Inference costs are directly cut by one-tenth. Small businesses finally have a chance. 75% fewer GPUs with the same computing power—who can handle that? Rack-level secure computing—this architecture is quite well thought out. Missing out on this wave and not jumping on board might mean missing out!
View OriginalReply0
GateUser-6bc33122vip
· 01-07 00:38
One-tenth of the cost? Now big model startups really have a chance.
View OriginalReply0
RektHuntervip
· 01-06 15:51
Wow, reducing inference costs to one-tenth? Small businesses can now play with large models too. The previous monopoly on computing power is about to break. 260TB/s is incredible; communication between GPUs is so smooth... But can it really run stably? A 75% reduction in GPUs—what does that mean? The saved electricity and hardware costs... Never mind, I don't want to think about it. It's going to hype up again. If this thing really performs this well, the industry landscape will have to change.
View OriginalReply0
SandwichTradervip
· 01-06 15:50
One-tenth of the cost? Now large models are really going to compete intensely --- 260TB/s sounds great, but can the cooling really be handled? --- GPU cut by 75%, what does this mean? Small and medium-sized enterprises can finally play with AI? --- Both modular and confidential computing, this architecture doesn't seem that simple --- Fivefold increase in energy efficiency? So all that electricity was wasted before, haha --- Reshaping the cost structure, isn't it just to grab market share? Same old story --- Is 260TB/s real? With this speed, it can run anything, right? --- I believe one-tenth of the cost, but has the upstream hardware cost really decreased? --- Abandoning fan solutions, is the new cooling method reliable? Don't let there be problems again --- Finally, someone is working on costs. The previous solutions were ridiculously expensive
View OriginalReply0
tokenomics_truthervip
· 01-06 15:44
260 TB/s? That number sounds unbelievable, but if we can really cut the inference cost down to one-tenth, miners will have a chance.
View OriginalReply0
MEVictimvip
· 01-06 15:41
One-tenth of the cost? If that's true, it should have appeared long ago. Don't let it be just on paper again.
View OriginalReply0
OnchainArchaeologistvip
· 01-06 15:39
One-tenth of the cost? This means big model startups are no longer burning money, finally able to breathe GPU costs cut by 75%, is this really true... corporate expenses are directly halved 260 TB/s bandwidth is outrageous, data flow is no longer a bottleneck Fivefold improvement in energy efficiency, cooling is finally not so crazy, amazing Modular design is imaginative, with large customization potential in the future Inference costs reduced to one-tenth, this update truly rewrites the game rules
View OriginalReply0
  • Pin

Trade Crypto Anywhere Anytime
qrCode
Scan to download Gate App
Community
  • 简体中文
  • English
  • Tiếng Việt
  • 繁體中文
  • Español
  • Русский
  • Français (Afrique)
  • Português (Portugal)
  • Bahasa Indonesia
  • 日本語
  • بالعربية
  • Українська
  • Português (Brasil)