DeepSeek has released a new paper,Fulwa (2025) Hindi Short Film with co-founder Liang Wenfeng credited as a contributor, detailing how its latest large language model DeepSeek-V3 achieves efficient training and inference using only 2,048 H800 GPUs – significantly fewer than the tens of thousands typically required. The team attributes this efficiency to four key innovations: memory optimization through multi-head latent attention (MLA), computational savings via a Mixture-of-Experts (MoE) design with FP8 precision, communication improvements using a multi-plane network topology, and faster inference through multi-token prediction (MTP). With MLA, KV cache memory usage is cut to just 70KB per token, up to 1/7 that of competing models. MoE architecture activates only 37 billion of the model’s 671 billion parameters per forward pass, reducing training costs by 90% compared to dense models. FP8 training further halves compute and memory usage, with minimal accuracy tradeoff. Beyond the model, the paper also outlines five future directions for AI hardware design, advocating for tighter integration between software and hardware to address memory, compute, and networking bottlenecks. [36Kr, in Chinese]
(Editor: {typename type="name"/})
Watch Duty: Track the Los Angeles wildfires with this free app
Best iPad deal: Save $70 on 10th Gen Apple iPad
CES 2025: When you'll be able to buy the new Lenovo Legion Go S handheld
Outdoor speaker deal: Save $20 on the Soundcore Boom 2
Best Target Circle 360 deal: How to get free gift
X announces labels for parody accounts
CES 2025: The Sleepwave For Me Buds track your brainwaves while you sleep
NYT mini crossword answers for May 12, 2025
Help, I can't stop thinking about Suzie Toot's 'Woman's World' lip sync
Ireland fines TikTok $600 million for sharing user data with China
What's new to streaming this week? (Jan. 10, 2025)
接受PR>=1、BR>=1,流量相当,内容相关类链接。