TY - GEN
T1 - Mitigating Inter-datacenter Incast with a Proxy
T2 - 24th ACM Workshop on Hot Topics in Networks, HotNets 2025
AU - Zhou, Anchengcheng
AU - Costic, Carter
AU - Hè, Hongyu
AU - Ghalayini, Ahmad
AU - Kabbani, Abdul
AU - Apostolaki, Maria
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/11/17
Y1 - 2025/11/17
N2 - Many-to-one communication (i.e., incast) is a long-standing challenge in networking with a wide range of proposed solutions. However, as incast-inducing applications today (e.g., storage, ML training) scale beyond a single datacenter, they introduce new challenges that current solutions do not handle. In particular, inter-datacenter links have orders of magnitude higher latency than intra-datacenter paths, lengthening the feedback loop that senders rely on to adjust their sending rates and drastically increasing incast completion times.To reduce inter-datacenter incast latency, we propose adding a proxy server in the sending datacenter to relay traffic between the senders and the receiver. Surprisingly, adding this extra hop reduces incast latency! The insight is that the added hop shifts the congestion point closer to the senders, shortening the feedback loop and allowing senders to converge quickly at a rate that fully utilizes the link while avoiding severe congestion. Motivated by preliminary results, we investigate low-overhead proxy designs and explore ways to expose the proxy as a broader optimization opportunity for application developers, cloud operators, and tenants.
AB - Many-to-one communication (i.e., incast) is a long-standing challenge in networking with a wide range of proposed solutions. However, as incast-inducing applications today (e.g., storage, ML training) scale beyond a single datacenter, they introduce new challenges that current solutions do not handle. In particular, inter-datacenter links have orders of magnitude higher latency than intra-datacenter paths, lengthening the feedback loop that senders rely on to adjust their sending rates and drastically increasing incast completion times.To reduce inter-datacenter incast latency, we propose adding a proxy server in the sending datacenter to relay traffic between the senders and the receiver. Surprisingly, adding this extra hop reduces incast latency! The insight is that the added hop shifts the congestion point closer to the senders, shortening the feedback loop and allowing senders to converge quickly at a rate that fully utilizes the link while avoiding severe congestion. Motivated by preliminary results, we investigate low-overhead proxy designs and explore ways to expose the proxy as a broader optimization opportunity for application developers, cloud operators, and tenants.
KW - congestion feedback
KW - datacenter networks
KW - inter-datacenter incast
KW - proxy
KW - routing
UR - https://www.scopus.com/pages/publications/105023648563
UR - https://www.scopus.com/pages/publications/105023648563#tab=citedBy
U2 - 10.1145/3772356.3772410
DO - 10.1145/3772356.3772410
M3 - Conference contribution
AN - SCOPUS:105023648563
T3 - HotNets 2025 - Proceedings of the 2025 24th ACM Workshop on Hot Topics in Networks
SP - 344
EP - 353
BT - HotNets 2025 - Proceedings of the 2025 24th ACM Workshop on Hot Topics in Networks
PB - Association for Computing Machinery, Inc
Y2 - 17 November 2025 through 18 November 2025
ER -