Source code to the paper Deep Q-Networks for Accelerating the Training of Deep Neural Networks
We are using Lua/Torch. The DQN component is mostly modified from DeepMind Atari DQN.
You might need to run install_dependencies.sh
first.
cd mnist_lr/;
cd mnist;
th train-on-mnist.lua; #get regression filter, save in ../save/
./run_gpu; #Start tune learning rate using dqn
#To get the test curve, run following command
cd mnist_lr/dqn/logs;
python paint_lr_episode.py;
python paint_lr_vs.py;
cd mnist_minibatch;
cd mnist;
th train-on-mnist.lua; #get regression filter, save in ../save/
./run_gpu; #Start select mini-batch using dqn
#To get the test curve, run following command
cd mnist_minibatch/dqn/logs;
python paint_mini_episode.py;
python paint_mini_vs.py;
- GPU device can be set in
run_gpu
wheregpu=0
- Learning rate can be set in
/ataricifar/dqn/cnnGameEnv.lua
, in thestep
function. - When to stop doing regression is in
/ataricifar/dqn/cnnGameEnv/lua
, in line 250
- Experiments on CIFAR-10
- Transfer learning: subset of CIFAR-10 to full CIFAR-10
- Visualization of the actions taken by the DQN. For example, show which categories have been used at every iteration.
@article{dqn-accelerate-dnn,
title={Deep Q-Networks for Accelerating the Training of Deep Neural Networks},
author={Fu, Jie and Lin, Zichuan and Liu, Miao and Leonard, Nicholas and Feng, Jiashi and Chua, Tat-Seng},
journal={arXiv preprint arXiv:1606.01467},
year={2016}
}
If you have any problems or suggestions, please contact me: jie.fu A~_~T u.nus.education