Distributed general function computation

Malak, Derya
CCDWN 2022, Invited Talk, 6th Caching, Computing and Delivery in Wireless Networks Workshop, co-located with the 20th International Symposium on Modeling and Optimization in Mobile, Ad hoc, and Wireless Networks (WiOpt 2022), 19-22 September 2022. Turin, Italy

Large-scale distributed computing systems, such as MapReduce, Spark, or distributed deep networks, are critical for parallelizing the execution of computational tasks. Nevertheless, a struggle between computation and communication complexity lies at the heart of distributed computing. There has been recently substantial effort to address this problem for a class of functions, such as distributed matrix multiplication, distributed gradient coding, linearly separable functions, etc. The optimal cost has been achieved under some constraints, based mainly on ideas of linear separability of the tasks and linear space intersections. Motivated by the same challenge, we propose a novel distributed computing framework where a master seeks to compute an arbitrary function of distributed datasets in an asymptotically lossless manner. Our approach exploits notions of distributed source and functional compression using characteristic graphs each source builds. These graphs have been widely utilized by Shannon, Korner, and ¨Witsenhausen to derive the rate lower bounds for computation, and later by Alon-Orlitsky, Orlitsky-Roche, Doshi-Shah-Medard, and Feizi-Medard, to resolve some well-known distributed coding and communication problems, allowing for lowered communication complexity and even for a) correlated data, b) a broad class of functions, and c) well-known topologies. The novelty of our approach lies in accurately capturing the communication-computation cost tradeoff by melding the notions of characteristic graphs and distributed placement, to provide a natural generalization of distributed linear function computation, thus elevating distributed gradient coding and distributed linear transform to the realm of distributed computing of any function. This approach is well suited to one-shot or multi-shot computations, under uniform or skewed data distributions, and for general placement models. In toy scenarios, we demonstrate gains up to %70 over fully distributed solutions and an approximation ratio of 2 within the optimal centralized rate. This work is joint with Prof. Petros Elia, Dr. Berksan Serbetci, and Federico Brunero in the Communication Systems Department at EURECOM.


Type:
Talk
City:
Turin
Date:
2022-09-19
Department:
Communication systems
Eurecom Ref:
7038
Copyright:
© 2022 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.
See also:

PERMALINK : https://www.eurecom.fr/publication/7038