Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement initial basic crawler for extracting structred relations from PO.DAAC #1

Closed
lewismc opened this issue Jan 28, 2017 · 6 comments
Assignees
Milestone

Comments

@lewismc
Copy link
Member

lewismc commented Jan 28, 2017

As discussed at our first meeting, we should use the Any23 basic-crawler to implement a lightweight crawler which will run Any23 extractions over Webpages.
I will link a WIP implementation which we can build on.

@lewismc lewismc self-assigned this Jan 28, 2017
@lewismc lewismc added this to the 0.1 milestone Jan 28, 2017
@Yongyao
Copy link
Collaborator

Yongyao commented Jan 29, 2017

@lewismc , it just came to my mind that the results that Any23 will get is like "DOI:XXX, shortName:xxx,......", which is pretty much the same as metadata, if I understand correctly. People would argue that why don't we just use PO.DAAC web service to do it.

@lewismc
Copy link
Member Author

lewismc commented Jan 29, 2017

The reason being is that initial, first order extractions are just the beginning for ESKG :)
There is plenty more work to do. This is only s start @Yongyao there is loads more to do.

@lewismc
Copy link
Member Author

lewismc commented Jan 29, 2017

Lets walk before we can run!

@lewismc
Copy link
Member Author

lewismc commented Feb 1, 2017

Another commit made and the new crawler command line looks as follows

usage: ESKGCrawler
    --maxDepth <mDepth>          Max allowed crawler depth.
    --maxPages <mPages>          Max number of pages before interrupting
                                 crawl.
    --numCrawlers <nCrawlers>    Sets the number of crawlers.
    --pageFilter <filter>        Regex used to filter out page URLs during
                                 crawling.
    --politenessDelay <pDelay>   Politeness delay in milliseconds.
    --seedUrl <seed>             An individual seed URL used to bootstrap
                                 the crawl.
    --storageFolder <storage>    Folder used to store crawler temporary
                                 data.

The only parameter which is mandatory is --seedUrl

@Yongyao
Copy link
Collaborator

Yongyao commented Feb 1, 2017

Great. Just realized the basic crawler is based on crawler4j.

@lewismc
Copy link
Member Author

lewismc commented Feb 1, 2017

ack

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants