Skip to content

LIBOL/ODL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Online Deep Learning: Learning Deep Neural Networks on the Fly

An implementation of the Hedge Backpropagation(HBP) proposed in Online Deep Learning: Learning Deep Neural Networks on the Fly

@inproceedings{sahoo2018online,
  title     = {Online Deep Learning: Learning Deep Neural Networks on the Fly},
  author    = {Doyen Sahoo and Quang Pham and Jing Lu and Steven C. H. Hoi},
  booktitle = {Proceedings of the Twenty-Seventh International Joint Conference on
               Artificial Intelligence, {IJCAI-18}},
  publisher = {International Joint Conferences on Artificial Intelligence Organization},             
  pages     = {2660--2666},
  year      = {2018},
  month     = {7},
  doi       = {10.24963/ijcai.2018/369},
  url       = {https://doi.org/10.24963/ijcai.2018/369},
}

Link to publication

Requirements and Installation

  • Theano 0.8.2
  • Keras 1.2.1

To install HBP, you need to replace the Keras's keras/engine/training.py file with our modified training.py. this doesn't affect normal projects that don't use HBP. Note that as the current HBP implementation only supports Keras 1.

Experiments

  • To run HBP on the sample Higgs dataset, first download the data:
wget -O data/higgs.mat https://www.dropbox.com/s/fvqnhe34cf0mlz9/higgs_100k.mat
  • To train HBP, run:
cd src/hbp
python main.py -c hb19.yaml
  • To train other baseline models, run:
cd src/baselines
./run.sh

Data sets

The data used in our experiments are available at https://drive.google.com/drive/folders/1fNZHK2NYbgfz27PPdSSA6lZTkoFakH28?usp=sharing

Train HBP on your own data

We provide a sample script in src/train.py to train HBP on a new dataset. Feel free to modify the code to suit your experiments.

Pytorch implementation

This is an independent pytorch implementation, please note that this is unofficial and not yet tested by us.