Probabilistic Fault Tolerance of Neural Networks in the Continuous Limit

Probabilistic Fault Tolerance of Neural Networks in the Continuous Limit

Regularizer for failing neurons

This work derives a closed-form probabilistic model of how neuron failures (permanent or transient) propagate through a deep neural network in the continuous-width limit. It uses this model to design a regulariser that, when added to standard training loss, encourages the network to learn weight distributions that are inherently robust to runtime neuron failures, without requiring fault injection during training.

Deep Neural Networks
Maturity
Support
C4DT
Inactive
Lab
Unknown

Distributed Computing Lab

Distributed Computing Lab
Rachid Guerraoui

Prof. Rachid Guerraoui

The Distributed Computing Lab focuses currently on Scalable Implementations of Cryptocurrencies, Byzantine fault tolerance and privacy in distributed machine learning, distributed algorithms making use of RDMA and NVRAM.

This page was last edited on 2024-03-22.