Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

thoughts on better neural net #28

Closed
andrewljohnson opened this issue May 10, 2016 · 3 comments
Closed

thoughts on better neural net #28

andrewljohnson opened this issue May 10, 2016 · 3 comments

Comments

@andrewljohnson
Copy link
Contributor

andrewljohnson commented May 10, 2016

The current neural net uses something really made for handwritten digit classification (MNIST).

We could do stuff like use Alexnet instead, or implement the neural nets described in Mnih/Hinton too. Literature also describes using a sequence of pre and post processing neural nets, where you can fill in gaps in road networks.

Expanding on these vague comments, there is a whole body of literature about how to use CNNs, RNNs, global topology, lidar elevation data, and much more to improve the accuracy of satellite imagery label. We should be able to get above 90% on the pixel level, just using semi-local RGB data, and push past that with multiple neural nets, more data, and other documented improvements in the last 2-3 years since Mnih's thesis.

See the README for a list of readings.

@andrewljohnson andrewljohnson changed the title Thoughts on Better Neural Net thoughts on better neural net May 10, 2016
@andrewljohnson
Copy link
Contributor Author

@lacker says to make a simpler neural net that works better with the given amount of IR training data, so the analysis overfits less.

Then expand the layers/data.

Don't just use some complicated neural net unsuited to the task.

@lacker
Copy link

lacker commented May 10, 2016

Yeah, I think you want to get to a state where "it works, just could be better" before scaling up. That way you can figure out how long you need to train, and keep params in a state where you don't overfit. I would try something real simple, like one convolution layer plus one softmax layer, and see if you can make it "work". By "work" I mean significantly better than random, like 80% accuracy, and not much of an overfitting problem.

@andrewljohnson
Copy link
Contributor Author

I think we're now in a state of "works with the simplest network." A single relu or single convolutional relu with a maxpool.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants