TA的每日心情 | 开心 2020-4-8 10:45 |
---|
签到天数: 227 天 [LV.7]分神
|
& N# ^' x. {; L8 h$ S
在论文里,这是第3.2.2节的内容7 x1 s8 d/ l1 r" h4 _
3 L. n& T* J$ B7 o7 Z3.2.2. Efficient Implementation of Cross-Node All-to-All Communication
( h% `2 [& ^4 n! q j6 C( X& e [; QIn order to ensure sufficient computational performance for DualPipe, we customize efficient3 x# j* I. L5 {9 P7 r. f
cross-node all-to-all communication kernels (including dispatching and combining) to conserve
& d0 \! [3 L8 ?) W2 Q5 rthe number of SMs dedicated to communication. The implementation of the kernels is codesigned with the MoE gating algorithm and the network topology of our cluster. To be specific,1 C! O& r, |2 H( X: A. B
in our cluster, cross-node GPUs are fully interconnected with IB, and intra-node communications; }8 E+ p- Q6 b* {% z$ e
are handled via NVLink. NVLink offers a bandwidth of 160 GB/s, roughly 3.2 times that of IB' }/ g, [) D& W, Q6 A8 v& d
(50 GB/s). To effectively leverage the different bandwidths of IB and NVLink, we limit each
+ r4 v% T% @$ l9 n* n5 h: i$ o' m: Vtoken to be dispatched to at most 4 nodes, thereby reducing IB traffic. For each token, when its# s6 W* n3 v4 m
routing decision is made, it will first be transmitted via IB to the GPUs with the same in-node
$ s% l0 f2 z: y0 z! l3 oindex on its target nodes. Once it reaches the target nodes, we will endeavor to ensure that it is
' u+ Y3 F& P: O4 q# y: e" Iinstantaneously forwarded via NVLink to specific GPUs that host their target experts, without& j3 p+ Q) v3 e- J8 x& w
being blocked by subsequently arriving tokens. In this way, communications via IB and NVLink
0 ?3 k* E( T$ I; T: m& N R& iare fully overlapped, and each token can efficiently select an average of 3.2 experts per node8 c- ?: F! q8 r6 \8 f
without incurring additional overhead from NVLink. This implies that, although DeepSeek-V37 j' i/ _1 C; O' I3 W7 c
13+ a' e4 \- r) t8 u! J
selects only 8 routed experts in practice, it can scale up this number to a maximum of 13 experts! \2 O, W+ k; ]1 n: i8 s+ y2 o$ q
(4 nodes × 3.2 experts/node) while preserving the same communication cost. Overall, under
! M) T2 k5 }$ u, X) `+ Nsuch a communication strategy, only 20 SMs are sufficient to fully utilize the bandwidths of IB
! [* N# [* V/ g0 ~3 Yand NVLink.
4 |% o$ V* q( Z9 h( wIn detail, we employ the warp specialization technique (Bauer et al., 2014) and partition
0 X4 c, {4 y' j$ ` Z6 z20 SMs into 10 communication channels. During the dispatching process, (1) IB sending, (2)0 I4 L6 M4 v& c- D
IB-to-NVLink forwarding, and (3) NVLink receiving are handled by respective warps. The3 }* d8 D0 r \
number of warps allocated to each communication task is dynamically adjusted according to the: m8 ?5 i8 @- l) ^8 \, ~
actual workload across all SMs. Similarly, during the combining process, (1) NVLink sending,
! s$ F6 j- C. f* R$ S9 W! d(2) NVLink-to-IB forwarding and accumulation, and (3) IB receiving and accumulation are also/ x& d& q3 Z: x% m: w
handled by dynamically adjusted warps. In addition, both dispatching and combining kernels- `, h6 B( D. D2 a/ G
overlap with the computation stream, so we also consider their impact on other SM computation$ D. E9 S% {5 ]
kernels. Specifically, we employ customized PTX (Parallel Thread Execution) instructions and8 \. o% F0 K/ V3 [8 P% E2 ^3 {$ V
auto-tune the communication chunk size, which significantly reduces the use of the L2 cache
' {- q' C: n( `9 c$ Gand the interference to other SMs.! l7 J7 d0 P( @
& N! V- ]" }, s6 X* z
通俗一点说,就是为了实现高效的跨节点全面通信。解决的问题本质上和唐家山老师日志里说的双机对拷的场景差不多。一般来说单机多卡之间用nvlink,多机多卡之间依赖IB网络,但nvlink的速率是IB网络的速率的3.2倍,需要通过一些优化来实现更好的传输策略。这是一整套方案。, w1 K( @& V. v- R
+ J# I1 B; U7 u8 ^' ?; f m
我的理解,使用PTX在其中,是为了更精准的定制线程执行减少通信块分配传输之间的串扰。7 u8 m. K3 E ? L) J$ I5 J
' \. F, j- u0 v {0 h
目的不是为了绕cuda,反而是为了让cuda的效率更高。
5 g# a$ e2 g' W) x& u
0 b6 s( L2 w+ a& b w! r- q类比一下,就好比发现网卡驱动在对拷特定内存块的时候会和应用的线程执行出现串行导致效率降低,而绕开操作系统定义的与网卡驱动的接口,直接使用网卡支持的指令集进行了优化。 |
评分
-
查看全部评分
|