DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks.
Jie FuHongyin LuoJiashi FengKian Hsiang LowTat-Seng ChuaPublished in: CoRR (2016)
Keyphrases
- hyperparameters
- neural network
- model selection
- cross validation
- closed form
- support vector
- random sampling
- bayesian inference
- em algorithm
- maximum likelihood
- gaussian process
- bayesian framework
- prior information
- sample size
- incremental learning
- incomplete data
- noise level
- back propagation
- gaussian processes
- computer vision
- machine learning
- maximum a posteriori
- missing values
- error rate
- higher order
- support vector machine
- active learning
- genetic algorithm