Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Upgrade get_dataset.tokenize() to multiprocessing #24

Open
wants to merge 2 commits into
base: master
Choose a base branch
from

Commits on Aug 20, 2019

  1. Upgrade get_dataset.tokenize() to multiprocessing ability

    get_dataset.tokenize() is to slow on a single CPU. Therefore it is upgraded to multiprocessing by implementing the
    multiprocessing target function worker_tokenize(args_list). Additionally a multiprocessing debug logger mp_logger
    was added together with logger.debug() and mp_logger.debug() message to track progress in the python console.
    DrStoop committed Aug 20, 2019
    Configuration menu
    Copy the full SHA
    e78fcb6 View commit details
    Browse the repository at this point in the history

Commits on Oct 2, 2019

  1. Configuration menu
    Copy the full SHA
    a0631a2 View commit details
    Browse the repository at this point in the history