-

@ renzume.
2025-02-25 05:35:20
DeepEP is a communication library optimized for Mixture-of-Experts (MoE) and expert parallelism, providing high-throughput GPU kernels and low-latency operations. The library supports both intranode and internode communication, offering specialized kernels for asymmetric-domain bandwidth forwarding and low-latency inference decoding, with comprehensive support for FP8 and RDMA networks.
https://github.com/deepseek-ai/DeepEP
#gpucomputing #aiinfrastructure #networking #moearchitecture #performanceoptimization