Skip to content
A collaborative web portal for NTNU students
Python HTML CSS JavaScript Other
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
.circleci
config
dataporten
examiner
kokekunster
scripts
semesterpage
static
templates
tmp
.coveragerc
.editorconfig
.gitignore
.landscape.yml
.stylelintrc.json
.travis.yml
Dockerfile
LICENSE
README.md
conftest.py
docker-compose.yml
manage.py
package-lock.json
requirements.in
requirements.txt
tox.ini

README.md

A Link Portal for NTNU students Build Status Coverage Status

A link portal written in Django. The site features:

  • Responsive design using CSS flex box
  • An admin dashboard for adding and editing links, courses, semesters, and so on
  • Authentication using NTNU username and password, made possible by Dataporten
  • Automatic retrieval of courses which the student is enrolled in

Contributing

Run the following commands in order to set up the project locally

Download the project

git clone https://github.com/JakobGM/WikiLinks.git wikilinks
pip install -r requirements.txt

Installing dependencies

Tesseract and tesserocr

WikiLinks uses tesserocr for character recognition of PDF files. Follow the up-to-date instructions here, which at the time of this writing boils down to:

$ apt-get install tesseract-ocr libtesseract-dev libleptonica-dev

Take care that you install version 4 of tesseract, you can check the installed version by running:

tesseract --version

If you had to upgrade tesseract to 4.0, you might have to reinstall tesserocr without cache:

pip install --no-cache-dir tesserocr

Or else, you might get the following error message:

ImportError: libtesseract.so.3: cannot open shared object file:
No such file or directory

If python crashes on import tesserocr with the following error message:

!strcmp(locale, "C"):Error:Assert failed:in file baseapi.cpp, line 209

Set the following environment variables:

export LC_ALL=C
export PYTHONIOENCODING="UTF-8"

This issue is tracked here.

Ghostscript

WikiLinks must also convert .pdf files to .tiff files before passing them on to tesserocr. Ghostscript is used for this purpose, and must be installed in order to enable OCR functionality. You can check if it is installed by running

$ which gs

pdftotext

Lastly, you must install the dependencies of pdftotext. This utility extracts text from PDFs that already contain text overlay information, so no OCR is needed. See instructions here.

Install a development database

Install Postgres.app from here. It is important to add the command line utilities to your path. Run ./manage.py migrate.

Load a production data dump

Run psql -d kokudb -U postgres in order to enter the postgresql shell within the Wikilinks database.

CREATE ROLE koku_user;
GRANT ALL PRIVILEGES ON DATABASE koku_db TO koku_user;

Receive the db and media dump from @JakobGM and put them in the tmp folder within the project folder. Run the following commands:

./manage.py dbrestore
./manage.py mediarestore

You can run the test suite like this:

py.test

And you can run CSS linting like so:

stylelint "**/*.css"

Production:

Necessary cronjob for automatic backups

SHELL=/bin/bash
*/5 * * * * source /home/jakobgm/kokekunster/koku_env/bin/activate && set -a && source /home/jakobgm/.pam_environment && set +a && /home/jakobgm/kokekunster/koku_env/bin/python /home/jakobgm/kokekunster/manage.py runcrons &> /home/jakobgm/kokekunster/backup.log

Manual retrieval of production data dumps

Media directory

Copy the media directory out from the Django docker container.

docker cp -r <django_image_id>:/media <vps_target_directory>

Copy the media directory to your local media directory.

scp wikilinks.no/<vps_target_directory> /path/to/wikilinks/repo/media

Postgres database

Create a production database dump in production.

docker exec 03ee pg_dump -U postgres postgres -O -x -F t > sql_dump.tar

Copy the dump to your local machine.

scp -r wikilinks.no:/path/to/sql_dump.tar .

Delete and restore the Wikilinks database to make sure it is clean.

dropdb koku_db -U postgres
createdb koku_db -U postgres

Remove the public schema from the database, as the dump will create it.

DROP SCHEMA public;

Temporarily grant superuser privileges to the database user.

ALTER ROLE koku_user WITH SUPERUSER;

Import the database.

pg_restore -d koku_db sql_dump.tar -U koku_user --role=koku_user --no-owner --single-transaction --exit-on-error

Remove superuser privileges from the database user.

ALTER ROLE koku_user WITH NOSUPERUSER;
You can’t perform that action at this time.