A web crawler can be used to browse the World Wide Web, typically for the purpose of Web indexing, as in many existing search engines, such as Google, Bing, Yandex, DuckDuckGo, Qwant, and so on.
This software was developed as an assignment during the 2020/21 university year at the CERI, Avignon University (France), by the following students:
- Abdelhakim RASFI
- Hacene SADOUDI
- Youssef ABIDAR
- imane HACEN
- Mohamed Kharchouf
The source code is organized as follows... <list the folders/packages, explain their role>
Here is the procedure to install this software :
- Download java sdk library
- Do that
- etc.
In order to use the software, you must...
- Do this
- Do that
- etc.
The project wiki (put a hyperlink) contains detailed instructions regarding how to use the web crawler.
The project relies on the following libraries:
- xxxxx : this library was used to...
- yyyyy: ...
During development, we use the following bibliographic resources:
- Webpage x: it explains the rules of
robots.txt
. - Book xxxx: it describes how to implement the
PageRank
algorithm. - etc.
- etc.