DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks.
Jie FuHongyin LuoJiashi FengKian Hsiang LowTat-Seng ChuaPublished in: IJCAI (2016)
Keyphrases
- hyperparameters
- neural network
- model selection
- cross validation
- random sampling
- bayesian framework
- support vector
- closed form
- bayesian inference
- noise level
- em algorithm
- gaussian process
- gaussian processes
- back propagation
- artificial neural networks
- sample size
- maximum a posteriori
- prior information
- maximum likelihood
- incremental learning
- active learning
- missing values
- learning process
- decision trees
- feature selection
- genetic algorithm