Skip to content
This repository has been archived by the owner on Nov 3, 2022. It is now read-only.

additional datasets - coco + voc2012 #47

Open
ahundt opened this issue Mar 17, 2017 · 10 comments
Open

additional datasets - coco + voc2012 #47

ahundt opened this issue Mar 17, 2017 · 10 comments

Comments

@ahundt
Copy link
Collaborator

ahundt commented Mar 17, 2017

I'm working on some code that loads ms coco and pascal voc for segmentation, but it would currently involve steps that depend on several other python libraries, including tf only steps. It could be used to create examples for the DenseNetFCN and #46 segmentation models.

I'm creating an example that trains on these keras-contrib models but with the dependency limitations mentioned above, and I can modify it for the keras-contrib code structure. However, this brings up the following questions:

  • Would keras-contrib accept a contribution with those limitations?
  • What would be the procedure?
  • Is someone interested in adapting for more general operation?

If it makes the most sense for me to just keep everything separate, that's okay.

@patyork
Copy link
Contributor

patyork commented Mar 18, 2017

What are the external libraries (and what are they used for)? Also, what are the steps that can only be done in TF?

@ahundt
Copy link
Collaborator Author

ahundt commented Mar 20, 2017

I'm sure they can be done with other tools, they are currently implemented with TF and would need to be rewritten to work with numpy arrays. Actually it probably could be done fairly easily without TF I've been using it a bit more since the previous post.

Pascal VOC 2012 + Berkeley additional labeling

  • pascal_voc.py
  • skimage for loading matlab matrices and converting to png

Extra for coco:

@ahundt
Copy link
Collaborator Author

ahundt commented Apr 19, 2017

I've made useful progress towards Pascal VOC and MS COCO in other repositories, which I plan to merge into keras-contrib when ready. Details are in https://github.com/ahundt/Keras-FCN/tree/densenet_atrous

One issue is that with more complex datasets it makes sense to have several manual utilities & external dependencies may be required.

How can the manual utility + dependency needs of large datasets be reconciled with the need to avoid dependencies when possible with Keras?

How should a single dataset dependency, like pycocotools, be handled?

@ahundt
Copy link
Collaborator Author

ahundt commented Apr 26, 2017

#80 adds pascal_voc support

@ahundt
Copy link
Collaborator Author

ahundt commented Apr 26, 2017

#81 adds ms coco support

@ahundt
Copy link
Collaborator Author

ahundt commented May 4, 2017

Some improvments to #81 should be made based on https://github.com/PavlosMelissinos/enet-keras/blob/master/src/data/datasets.py

@tarvaina
Copy link
Contributor

The Pascal VOC installation failed on Python 3. I submitted pull request #114 to fix the problem.

@ahundt
Copy link
Collaborator Author

ahundt commented Jun 21, 2017

Thanks!

@ahundt
Copy link
Collaborator Author

ahundt commented Jun 21, 2017

I'm also working on extending it for running via TFRecords but that may be a while before the PR is ready.

@ahundt
Copy link
Collaborator Author

ahundt commented Jun 21, 2017

If you might be interested in incorporating it the TFRecord code to be manually moved into keras-contrib is at warmspringwinds/tf-image-segmentation#25 and keras-team/keras#6928 adds support to Keras

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants