TY - JOUR
T1 - Planter
T2 - Rapid Prototyping of In-Network Machine Learning Inference
AU - Zheng, Changgang
AU - Zang, Mingyuan
AU - Hong, Xinpeng
AU - Perreault, Liam
AU - Bensoussane, Riyad
AU - Vargaftik, Shay
AU - Ben-Itzhak, Yaniv
AU - Zilberman, Noa
N1 - Publisher Copyright:
© 2024 Copyright is held by the owner/author(s).
PY - 2024/8/6
Y1 - 2024/8/6
N2 - In-network machine learning inference provides high throughput and low latency. It is ideally located within the network, power efficient, and improves applications' performance. Despite its advantages, the bar to in-network machine learning research is high, requiring significant expertise in programmable data planes, in addition to knowledge of machine learning and the application area. Existing solutions are mostly one-time efforts, hard to reproduce, change, or port across platforms. In this paper, we present Planter: a modular and efficient open-source framework for rapid prototyping of in-network machine learning models across a range of platforms and pipeline architectures. By identifying general mapping methodologies for machine learning algorithms, Planter introduces new machine learning mappings and improves existing ones. It provides users with several example use cases and supports different datasets, and was already extended by users to new fields and applications. Our evaluation shows that Planter improves machine learning performance compared with previous model-tailored works, while significantly reducing resource consumption and co-existing with network functionality. Planter-supported algorithms run at line rate on unmodified commodity hardware, providing billions of inference decisions per second.
AB - In-network machine learning inference provides high throughput and low latency. It is ideally located within the network, power efficient, and improves applications' performance. Despite its advantages, the bar to in-network machine learning research is high, requiring significant expertise in programmable data planes, in addition to knowledge of machine learning and the application area. Existing solutions are mostly one-time efforts, hard to reproduce, change, or port across platforms. In this paper, we present Planter: a modular and efficient open-source framework for rapid prototyping of in-network machine learning models across a range of platforms and pipeline architectures. By identifying general mapping methodologies for machine learning algorithms, Planter introduces new machine learning mappings and improves existing ones. It provides users with several example use cases and supports different datasets, and was already extended by users to new fields and applications. Our evaluation shows that Planter improves machine learning performance compared with previous model-tailored works, while significantly reducing resource consumption and co-existing with network functionality. Planter-supported algorithms run at line rate on unmodified commodity hardware, providing billions of inference decisions per second.
KW - Dimension reduction
KW - In-network computing
KW - Machine learning
KW - Machine learning compilers
KW - Modular framework
KW - P4
KW - Programmable switches
U2 - 10.1145/3687230.3687232
DO - 10.1145/3687230.3687232
M3 - Journal article
AN - SCOPUS:85200714200
SN - 0146-4833
VL - 54
SP - 2
EP - 20
JO - Computer Communication Review
JF - Computer Communication Review
IS - 1
ER -