Çatalhöyük is an UNESCO World Heritage Site located in modern day Turkey. This site has been the subject of the Çatalhöyük Research Project, in which archaeologists have collected over 150,000 photographs richly detailing buildings, artifacts, and ways of life of this civilization.
As described on the project's website, "Çatalhöyük has been the subject of investigation for more than 50 years. Researchers from around the world have travelled to the site over the past half-century to study its vast landscape of buildings, remarkable ways of life, and its many exquisite works of art and craft. Since 1993, the Çatalhöyük Research Project has recruited an international group of specialists to pioneer new archaeological, conservation and curatorial methods on and off site. Simultaneously, it aims to advance our understandings of human life in the past."
Our goal is to use computer vision to help enrich the digital resources produced by the Çatalhöyük Research Project, in keeping with the project's mission to create a living archive.
A large number of photos of the dig site contain whiteboards with text describing the photo's contents, and often the contents of subsequent photos. Our first project is to detect photos with whiteboards, then extract the text from these whiteboards and use optical character recognition to digitize the information.
Many of the photos in the Çatalhöyük database are already labeled with metadata. To make the database more readily searchable, we aim to expand these tags as thoroughly as possible to entire archive.
Creating Conda environment
- Make sure you have Anaconda or Miniconda
- Download repo and
cdinto it. Run
conda env create --file=environment.yml.
- Add CUDA to the environment if you have a GPU.
- (Optional) Add interpreter to PyCharm
- Go to
Preferences > Project > Project Interpreter- Click the gear icon, then 'Add' - Select 'Conda Environment' and 'Existing Environment' - Click the three dots, then find the interpreter. Should be somewhere like
- Create and activate conda environment as described above.
python train.py --name=NAME --batch_size=128 --data_dir=data/wb130k.
- Locate your trained model in
python predict.py --ckpt_path=ckpts/NAME/best.pth.tar --data_dir=data/wb130k/ --phase=test --name=wb130k_test --gpu_ids=0,1,2,3 --batch_size=256 --prob_threshold=0.4.
cdinto this project's root directory.
tmux new -s tbfor a new
source activate resto set up the virtual environment.
tensorboard --logdir=. --port=5678.
dto detach from the
tmuxsession. Later run
tmux a -t trainto re-attach.
- (Local) Run
ssh -N -f -L localhost:1234:localhost:5678 <remote_host>.
- (Local) In a web browser, go to
+ args: Command-line arg parsing + ckpts: Holds model checkpoints + data: Placeholder for CIFAR dataset + data_loader: Wraps CIFAR data loader + logger: Logs training info to the console and TensorBoard + logs: Holds logs produced by the logger + optim: Optimizer and learning rate scheduler + saver: Saves and loads model checkpoints + scripts: Scripts for miscellaneous tasks + util: utility functions - train.py: Training - test.py: Test inference