DeepEP is a communication library optimized for Mixture-of-Experts (MoE) and expert parallelism, providing high-throughput GPU kernels and low-latency operations. The library supports both intranode and internode communication, offering specialized kernels for asymmetric-domain bandwidth forwarding and low-latency inference decoding, with comprehensive support for FP8 and RDMA networks.

https://github.com/deepseek-ai/DeepEP

#gpucomputing #aiinfrastructure #networking #moearchitecture #performanceoptimization

Reply to this note

Please Login to reply.

Discussion

No replies yet.