Day 6 of #OpenSourceWeek: One More Thing – DeepSeek-V3/R1 Inference System Overview

deepseek刚刚发布了
Day 6 of #OpenSourceWeek: One More Thing – DeepSeek-V3/R1 Inference System Overview Optimized throughput and latency via: Cross-node EP-powered batch scaling Computation-communication overlap Load balancing Statistics of DeepSeek’s Online Service: 73.7k/14.8k input/output tokens per second per H800 node Cost profit margin 545% We hope this week’s insights offer value to the community and contribute to our shared AGI goals. Deep Dive: bit.ly/4ihZUiO

5 个赞

Over the past 24 hours (UTC+8 02/27/2025 12:00 PM to 02/28/2025 12:00 PM), the combined peak node occupancy for V3 and R1 inference services reached 278, with an average occupancy of 226.75 nodes (each node contains 8 H800 GPUs). Assuming the leasing cost of one H800 GPU is $2 per hour, the total daily cost amounts to $87,072.
If all tokens were billed at DeepSeek-R1’s pricing (*), the total daily revenue would be $562,027, with a cost profit margin of 545%.
成本没有想象的高,长期看好英伟达


省流

One more thing :laughing:

这个545%好像是因为显卡是他们自己的。有人用租赁h800每小时按2美元算,利润率也有85%左右

继续给 DeepSeek 点赞

文章中就是拿租的显卡计算的成本

:innocent:这样吗,我再看看,推上有人说他算的是84.5%

此话题已在最后回复的 30 天后被自动关闭。不再允许新回复。