Skip to content

maximd33/kaggle-dogs-vs-cats-solution

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

37 Commits
 
 
 
 
 
 

Repository files navigation

Solutions for https://www.kaggle.com/c/dogs-vs-cats competition.

WORK IN PROGRESS

1.learning_from_scratch is folder with models that were trained in NVIDIA DIGITS with Caffe backend.

Here is table with results, but hyperparameters of neural nets can drastically affect accuracy, so if you know how to improve results let me know.

Name Acc. test Acc. val. Train time Forward pass time On disk model size Year Paper
AlexNet 93.65% - 35m - 227.5Mb 2012 link
SqeezeNet v1.1 92.46% - ~2h - 2.9Mb 2016 link
GoogLeNet 94.62% - 50m - 41.3Mb 2014 link
VGG-16 96.51% - 5h20m - 537.1Mb 2014 link
VGG-19 97.42% - 25h50m - 558.3Mb 2014 link
Network-In-Network 93.22% - ~2h - 26.3Mb 2014 link

All network models were trained for 30 epochs, but batch size, learning rate, etc. can vary. Also note for example for VGG-19 model sufficient accuracy(~97%) achived after 6 epochs, so in real life you need to train network shorter\longer than 30 epochs. Test accuracy measured on train-test split 80%-20%, maybe to achive better acuracy we can train on all available data. Validation accuracy measured on Kaggle leaderboard.

More models to test:

1. Train ResNet from scratch.
https://github.com/beniz/deepdetect/tree/master/templates/caffe
https://github.com/jay-mahadeokar/pynetbuilder/tree/master/models/imagenet
https://github.com/ducha-aiki/caffenet-benchmark

resnet
try this
https://github.com/lfrdm/Masterarbeit/blob/master/train_val-residual-small.prototxt
from thread
https://github.com/KaimingHe/deep-residual-networks/issues/6

2.finetuning is folder with models that were finetuned from models trained on ImageNet.

Name Acc. test Acc. val.
AlexNet 97.98% -
SqeezeNet v1.1 98.87% -
GoogLeNet 99.58% -
VGG-16 99.40% -
VGG-19 - -
Network-In-Network 98.49% -

TODO: data augmentation:

1. Create python layer
http://www.andrewjanowczyk.com/real-time-data-augmentation-using-nvidia-digits-python-layer/
Try to make it like in torch wrapper 
https://github.com/NVIDIA/DIGITS/issues/1034

Tested on system with following configuration:

Ubuntu version:

`lsb_release -a`

Ubuntu 14.04.4 LTS

`uname -a`

Linux myuser-computer 3.19.0-61-generic #69~14.04.1-Ubuntu SMP Thu Jun 9 09:09:13 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux

gcc version:

`gcc --version`

gcc (Ubuntu 4.8.4-2ubuntu1~14.04.3) 4.8.4

DIGITS version:

`./digits-devserver --version`

4.1-dev

Caffe version:

`git status`

branch caffe-0.15

`git log -n 1`

commit e638c0b1cb19afff50d830ce87cc1898f18568fd
Author: Sergei Nikolaev <[email protected]>
Date:   Wed Aug 31 14:32:28 2016 -0700
Mark 0.15.13

CPU:

`cat /proc/cpuinfo | grep "model name"`

Intel(R) Core(TM)2 Duo CPU     E8500  @ 3.16GHz

GPU:

`nvidia-smi`

+-----------------------------------------------------------------------------+
| NVIDIA-SMI 367.44                 Driver Version: 367.44                    |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|===============================+======================+======================|
|   0  GeForce GTX 1070    On   | 0000:01:00.0      On |                  N/A |
| 27%   38C    P8    10W / 151W |    150MiB /  8108MiB |      0%      Default |
+-------------------------------+----------------------+----------------------+
~~~



About

Kaggle dogs vs cats solution

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published