If dropout limits trainable depth, does critical initialisation still matter? A large-scale statistical analysis on ReLU networks.
Arnu PretoriusElan Van BiljonBenjamin van NiekerkRyan EloffMatthew ReynardSteven D. JamesBenjamin RosmanHerman KamperSteve KroonPublished in: Pattern Recognit. Lett. (2020)