RESEARCH

SYSTEMS RESEARCH

ResiliNet: Failure-Resilient Inference in Distributed Neural Networks

September 1, 2020

Abstract

Techniques such as Federated Learning and Split Learning aim to train distributed deep learning models without sharing private data. In Split Learning, when a neural network is partitioned and distributed across physical nodes, failure of physical nodes causes the failure of the neural units that are placed on those nodes, which results in a significant performance drop. Current approaches focus on resiliency of training in distributed neural networks. However, resiliency of inference in distributed neural networks is less explored. We introduce ResiliNet, a scheme for making inference in distributed neural networks resilient to physical node failures. ResiliNet combines two concepts to provide resiliency: skip hyperconnection, a concept for skipping nodes in distributed neural networks similar to skip connection in resnets, and a novel technique called failout, which is introduced in this paper. Failout simulates physical node failure conditions during training using dropout, and is specifically designed to improve the resiliency of distributed neural networks. The results of the experiments and ablation studies using three datasets confirm the ability of ResiliNet to provide inference resiliency for distributed neural networks.

Download the Paper

AUTHORS

Written by

Ashkan Yousefpour

Brian Q. Nguyen

Siddartha Devic

Guanhua Wang

Aboudy Kreidieh

Hans Lobel

Alexandre M. Bayen

Jason P. Jue

Publisher

International Workshop on Federated Learning for User Privacy and Data Confidentiality (FL-ICML)

Research Topics

Systems Research

Related Publications

December 07, 2018

SYSTEMS RESEARCH

Rethinking floating point for deep learning | Facebook AI Research

Reducing hardware overhead of neural networks for faster or lower power inference and training is an active area of research. Uniform quantization using integer multiply-add has been thoroughly investigated, which requires learning many…

Jeff Johnson

December 07, 2018

June 22, 2015

SYSTEMS RESEARCH

NLP

Fast Convolutional Nets With fbfft: A GPU Performance Evaluation | Facebook AI Research

We examine the performance profile of Convolutional Neural Network training on the current generation of NVIDIA Graphics Processing Units. We introduce two new Fast Fourier Transform convolution implementations: one based on NVIDIA’s cuFFT…

Nicolas Vasilache, Jeff Johnson, Michael Mathieu, Soumith Chintala, Serkan Piantino, Yann LeCun

June 22, 2015

March 02, 2020

SYSTEMS RESEARCH

Federated Optimization in Heterogenous Networks | Facebook AI Research

Federated Learning is a distributed learning paradigm with two key challenges that differentiate it from traditional distributed optimization: (1) significant variability in terms of the systems characteristics on each device in the network…

Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, Virginia Smith

March 02, 2020

September 01, 2020

SYSTEMS RESEARCH

ResiliNet: Failure-Resilient Inference in Distributed Neural Networks

Techniques such as Federated Learning and Split Learning aim to train distributed deep learning models without sharing private data.…

Ashkan Yousefpour, Brian Q. Nguyen, Siddartha Devic, Guanhua Wang, Aboudy Kreidieh, Hans Lobel, Alexandre M. Bayen, Jason P. Jue

September 01, 2020

Help Us Pioneer The Future of AI

We share our open source frameworks, tools, libraries, and models for everything from research exploration to large-scale production deployment.