forked from jeremysalwen/DeepFire
-
Notifications
You must be signed in to change notification settings - Fork 0
Deep Learning Library for OpenCL/CUDA using Arrayfire
License
UIKit0/DeepFire
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This is a Deep Learning library using ArrayFire. Hopes to cover: Every variety of neural networks (feedforward, recursive, autoencoders, etc) Maybe some stuff like RBMS? As many training algorithms as possible (SGD and variations, HF and variations, linesearch techniques, minfunc) Not complete though. Please leave any feedback you have in the issue tracker (does not have to be a bug or feature request). -Jeremy Notes on implementation: Data is passed between layers as an af::array. Doing sample-wise processing is too slow, so everything works with batches. So we use the first dimension of the array to represent the different samples in the batch. Since the input and output size of layers should be independent of the batch size, we use a dim4 to represent the dimension of the data, but the first dimension is just ignored. TODO: ***Implement dropout/dropconnect in a way that is fast, generally applicable, but not overly ubiquitious. On a neural network, it doesn't take any understanding of the structure to apply dropconnect to the weights for training. For vanilla dropout, it does take some understanding. For both of them, you can also be /almost/ oblivous to how it was trained at test time. You still need to scale the inputs down by 1/2 though (or whatever the dropout factor was). However dropconnect introduces a new way of calculating the output values that supposedly gets better results by doing a better job of approximating the "averaging of networks" model. So essentially the dropout code has to be decently coupled with the network structure. However, this coupling would naively mean calculating the mask values independently from each other. I think this is bad for performance reasons (it's a lot of random values we are generating, so we want to do them all at once in a matrix). The original dropconnect folks even generated them in bitpacked form, although I don't know how that will work with ArrayFire without a custom kernel. We also want things to be simple to use if you don't enable dropout. Also, dropout is a form of regularization. Should it fit into a similar interface to regularizers like L2 norm?? ***What about somehow displaying something while training. That would be pretty cool, but I don't know if it's worth adding hooks to the optimizers. Perhaps the optimizer is itself responsible for the display of information, as it knows what information should be displayed.
About
Deep Learning Library for OpenCL/CUDA using Arrayfire
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published
Languages
- C++ 98.2%
- Makefile 1.8%