Permalink
Commits on Oct 11, 2018
  1. Stop module update (#840)

    diyclassics authored and kylepjohnson committed Oct 11, 2018
    * Update backoff.py
    
    Fix lemmatizer order
    
    * Add inc_values return option
    
    * Add test for inc_value return; clean up test class
    
    * Fix values tuple for zou
    
    * Update tests
    
    * Fix value for Zou
    
    * Add CorpusStoplist for Classical Chinese
    
    * Comment out redundant code
    
    * Update stop.py
    
    * Add Latin stop submodule
    
    * Update tests for stop
    
    * A/B module style
    
    * Update tests
    
    * Update stop module structure
    
    * Update module structure
    
    * Remove preprocessing setup from base Stoplist classes
    
    * Remove preprocessing setup from base Stoplist classes
    
    * Update __main__ test
    
    * Fix punctuation preprocessing
    
    * Add type check to main stop method
    
    * Remove redundant StringStoplist class
    
    * Rename base class, subclasses
    
    * Update tests
    
    * Update stop files
    
    * Clean up latin file v. module
    
    * Fix tests
    
    * Fix Latin stop import
    
    * Update docs for Stop module
Commits on Apr 26, 2018
  1. Merge pull request #731 from clovell/latin-lemmatizer

    diyclassics committed Apr 26, 2018
    change when lemmatizer is defined
Commits on Apr 24, 2018
  1. Merge branch 'master' into latin-lemmatizer

    diyclassics committed Apr 24, 2018
Commits on Feb 22, 2018
  1. Build stoplists from document collections (#686)

    diyclassics authored and kylepjohnson committed Feb 22, 2018
    * Add different bases to corpus-based stop module
    
    * Update .coveragerc
    
    * Update stop tests
    
    * Update target lists
    
    * Add frequency basis to stop
    
    * Update corpus stop for basis frequency
    
    * Add numpy, sklearn, and scipy to requirements
    
    * Fix pip install name
    
    * Update stops; increase test coverage
    
    * Remove set union
    
    * Add tests for checking package imports
    
    * Remove old test strings
    
    * Cleanup test
    
    * Fix to have correct license
    
    * Add else to basis check
    
    * Add docstring
    
    * Add option for inc_counts for string
    
    * Clean up stop.py
    
    * Update docs
    
    * Cleanup stoplist docs
    
    * Fix typo in stoplist docs
    
    * Fix typo
    
    * Add tf-idf to CorpusStoplist bases
    
    * Add test for tfidf; fix test for zou
Commits on Feb 8, 2018
  1. Fixes #630 Clean up word tokenizer module (#631)

    diyclassics authored and kylepjohnson committed Feb 8, 2018
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Update files
    
    * Reset master
    
    * Reset master
    
    * rm whitespace
    
    * Add default Greek tokenizer
    
    * Comment on word tokenizer
    
    * Cleanup order of languages/functions, alphabetical
    
    * Update docs for Greek word tokenizer
    
    * Add unittest for Greek word tokenizer
    
    * Remove encoding declaration
    
    * Clean up imports
    
    * Clean up author
    
    * Comment on missing authors
    
    * Delete unused imports
  2. Fixes Travis problem #657 (#662)

    diyclassics authored and kylepjohnson committed Feb 8, 2018
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Update files
    
    * Reset master
    
    * Reset master
    
    * rm whitespace
    
    * Add init
    
    * Update travis yml
Commits on Feb 1, 2018
  1. Add Swadesh lists for classical languages to cltk.corpus (#642)

    diyclassics authored and kylepjohnson committed Feb 1, 2018
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Update files
    
    * Reset master
    
    * Reset master
    
    * rm whitespace
    
    * Add swadesh lists for Latin, Greek; built class for generating lists
    
    * Update tests
    
    * Fix spacing
    
    * Update docs
Commits on Jan 28, 2018
  1. Fixes #628 Greek word tokenizer added (#629)

    diyclassics authored and kylepjohnson committed Jan 28, 2018
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Update files
    
    * Reset master
    
    * Reset master
    
    * rm whitespace
    
    * Add default Greek tokenizer
    
    * Comment on word tokenizer
    
    * Cleanup order of languages/functions, alphabetical
    
    * Update docs for Greek word tokenizer
    
    * Add unittest for Greek word tokenizer
Commits on Nov 9, 2017
  1. Stops update (#600)

    diyclassics authored and kylepjohnson committed Nov 9, 2017
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Reset master
    
    * Reset master
    
    * Add new general stop module
    
    * Add test for sting-based stoplist builder
    
    * Fix import in test
Commits on Jul 21, 2017
  1. Regex lemmatizer update (#565)

    diyclassics authored and kylepjohnson committed Jul 21, 2017
    * Refactor RegexpLemmatizer
    
    * Update Regexp Lemmatizer test
    
    * Update Regexp Lemmatizer test
    
    * Refactor RomanNumeralLemmatizer
    
    * Add test for BackoffLatinLemmatizer; fix test coverage in general
Commits on May 26, 2017
  1. Fixes #531, re: exclamation in sentence tokenizer (#533)

    diyclassics authored and kylepjohnson committed May 26, 2017
    * Add line tokenizer with tests
    
    * Fix line tokenizer return
    
    * Update docs to include line tokenizer
    
    * Fix code block in docs
    
    * Fixes #531, adds exclamation point to external Latin punctuation; updates tests
  2. Update Backoff Lemmatizer Docs (#532)

    diyclassics authored and kylepjohnson committed May 26, 2017
    * Add line tokenizer with tests
    
    * Fix line tokenizer return
    
    * Update docs to include line tokenizer
    
    * Fix code block in docs
    
    * Update Latin docs
  3. Add line tokenizer to CLTK (#530)

    diyclassics authored and kylepjohnson committed May 26, 2017
    * Add line tokenizer with tests
    
    * Fix line tokenizer return
    
    * Update docs to include line tokenizer
    
    * Fix code block in docs
Commits on Feb 6, 2017
  1. Add seed parameter to default BackoffLemmatizer (#443)

    diyclassics authored and kylepjohnson committed Feb 6, 2017
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Add helper function 'tokenize' to sentence tokenizer
    
    * Add custom word/sentence tokenizers to Latinlibrary 'book' loader
    
    * Add alternative sentence tokenizer for latinlibrary loader if CLTK_DATA is not present
    
    * Add GSoC Lemmatize files
    
    * Minor updates to lookup files, tests for main file
    
    * Improved docs for backoff.py; more consistent naming conventions throughout
    
    * More general updates to GSoC lemmatizer code and models.
    
    * Add test module for lemmatizer
    
    * Add more tests for base lemmatizer classes
    
    * Remove test_distance_sentences temporarily
    
    * Added a readme with GSoC 2016 info
    
    * Fixed markdown link
    
    * Update readme.md
    
    * do light cleanup
    
    * Fixes #2, adds English words to exceptions list for Latin tokenizer
    
    * Remove * import for nltk tag module; load old lemma model from cltk_data as pickle.
    
    * Deleted old_model; moved to cltk_data
    
    * Renamed ModelLemmatizer to TrainLemmatizer tests
    
    * Make all model imports check cltk_data
    
    * Fixed default settings--mostly preloading regex patterns--for various lemmatizers; modified tests accordingly
    
    * Updated docstrings for backoff.py
    
    * Moved model files out of module
    
    * Fix bad print statement
    
    * Clean up Greek lemmatize module
    
    * Remove extra readme files
    
    * Comment out TrigramPOSLemmatizer
    
    * Cleanup backoff.py; add test for BigramPOS lemmatizer
    
    * Remove function for old version of lemmatizer
    
    * Updated tests
    
    * Fixed typo in test_lemmatize imports
    
    * Fixes #417: Check token against full regex before iterating in Latin backoff lemmatizer
    
    * Updated docs for Latin lemmatzier
    
    * Add seed parameter so that Lemmatizer runs can be replicable
Commits on Nov 10, 2016
  1. Add docs for Backoff Latin lemmatizer (#419)

    diyclassics authored and kylepjohnson committed Nov 10, 2016
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Add helper function 'tokenize' to sentence tokenizer
    
    * Add custom word/sentence tokenizers to Latinlibrary 'book' loader
    
    * Add alternative sentence tokenizer for latinlibrary loader if CLTK_DATA is not present
    
    * Add GSoC Lemmatize files
    
    * Minor updates to lookup files, tests for main file
    
    * Improved docs for backoff.py; more consistent naming conventions throughout
    
    * More general updates to GSoC lemmatizer code and models.
    
    * Add test module for lemmatizer
    
    * Add more tests for base lemmatizer classes
    
    * Remove test_distance_sentences temporarily
    
    * Added a readme with GSoC 2016 info
    
    * Fixed markdown link
    
    * Update readme.md
    
    * do light cleanup
    
    * Fixes #2, adds English words to exceptions list for Latin tokenizer
    
    * Remove * import for nltk tag module; load old lemma model from cltk_data as pickle.
    
    * Deleted old_model; moved to cltk_data
    
    * Renamed ModelLemmatizer to TrainLemmatizer tests
    
    * Make all model imports check cltk_data
    
    * Fixed default settings--mostly preloading regex patterns--for various lemmatizers; modified tests accordingly
    
    * Updated docstrings for backoff.py
    
    * Moved model files out of module
    
    * Fix bad print statement
    
    * Clean up Greek lemmatize module
    
    * Remove extra readme files
    
    * Comment out TrigramPOSLemmatizer
    
    * Cleanup backoff.py; add test for BigramPOS lemmatizer
    
    * Remove function for old version of lemmatizer
    
    * Updated tests
    
    * Fixed typo in test_lemmatize imports
    
    * Fixes #417: Check token against full regex before iterating in Latin backoff lemmatizer
    
    * Updated docs for Latin lemmatzier
Commits on Nov 9, 2016
  1. Fixes #417: Check token against full regex before iterating in Latin (#…

    diyclassics authored and kylepjohnson committed Nov 9, 2016
    …418)
    
    * Fixes #2, adds English words to exceptions list for Latin tokenizer
    
    * Fixes #417: Check token against full regex before iterating in Latin backoff lemmatizer
Commits on Oct 29, 2016
  1. Closes #136; Beta version of Latin backoff lemmatizer (#395)

    diyclassics authored and kylepjohnson committed Oct 29, 2016
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Add helper function 'tokenize' to sentence tokenizer
    
    * Add custom word/sentence tokenizers to Latinlibrary 'book' loader
    
    * Add alternative sentence tokenizer for latinlibrary loader if CLTK_DATA is not present
    
    * Add GSoC Lemmatize files
    
    * Minor updates to lookup files, tests for main file
    
    * Improved docs for backoff.py; more consistent naming conventions throughout
    
    * More general updates to GSoC lemmatizer code and models.
    
    * Add test module for lemmatizer
    
    * Add more tests for base lemmatizer classes
    
    * Remove test_distance_sentences temporarily
    
    * Added a readme with GSoC 2016 info
    
    * Fixed markdown link
    
    * Update readme.md
    
    * do light cleanup
    
    * Remove * import for nltk tag module; load old lemma model from cltk_data as pickle.
    
    * Deleted old_model; moved to cltk_data
    
    * Renamed ModelLemmatizer to TrainLemmatizer tests
    
    * Make all model imports check cltk_data
    
    * Fixed default settings--mostly preloading regex patterns--for various lemmatizers; modified tests accordingly
    
    * Updated docstrings for backoff.py
    
    * Moved model files out of module
    
    * Fix bad print statement
    
    * Clean up Greek lemmatize module
    
    * Remove extra readme files
    
    * Comment out TrigramPOSLemmatizer
    
    * Cleanup backoff.py; add test for BigramPOS lemmatizer
    
    * Remove function for old version of lemmatizer
    
    * Updated tests
    
    * Fixed typo in test_lemmatize imports
Commits on Sep 30, 2016
  1. Added 'hymen' to list of -n exceptions for Latin tokenizer (#390)

    diyclassics authored and kylepjohnson committed Sep 30, 2016
Commits on Aug 29, 2016
  1. Fixes #2, adds English words to exceptions list for Latin tokenizer (#…

    diyclassics authored and kylepjohnson committed Aug 29, 2016
  2. Fix #366; fix case of cltk_data (#371)

    diyclassics authored and kylepjohnson committed Aug 29, 2016
Commits on Aug 5, 2016
  1. Fixes #345 (& fixes #297) re: errors in Latinlibrary corpus loading m…

    diyclassics authored and kylepjohnson committed Aug 5, 2016
    …ethods (#347)
    
    * Check that tokens exist before handling them in Latin word tokenizer
    
    * Add helper function 'tokenize' to sentence tokenizer
    
    * Add custom word/sentence tokenizers to Latinlibrary 'book' loader
    
    * Add alternative sentence tokenizer for latinlibrary loader if CLTK_DATA is not present
Commits on Jun 9, 2016
  1. Updates to Latin word tokenizer (#311)

    diyclassics authored and kylepjohnson committed Jun 9, 2016
    * Add feature to load Latin corpora directly using PlaintextCorpusReader
    
    Using the Latin Library corpus as a test case, this feature allows
    you to refer to a corpus and load it with PlaintextCorpusReader using
    the following syntax:
    
    from cltk.corpus.latin import latinlibrary
    
    The code checks to make sure that a corpus is installed in the
    main CLTK_DATA and raises an error if it is not there.
    
    * Added missing comma to list 'cum_inclusions'
    
    * Stop tokenizing 'nec'
    
    This is based on the discussion here: PerseusDL/treebank_data#8. Until this larger issue is resolved, it seems best to leave the form 'nec' as is.
    
    * Added hyphen before tokenized enclitics
    
    Looking at the Perseus NLP data, many of the tokenized enclitics
    are distinguished using a hyphen, e.g. "-que". This brings it more
    in line with that dataset. It also better distinguishes "ne" from "-ne".
    
    * Add 'neque' to que_exceptions
    
    Like 'nec', 'neque' should no longer be separated as two tokens to
    bring the tokenizer more in line with Perseus NLP data (see
    previous commit).
    
    * Revert "Stop tokenizing 'nec'"
    
    This reverts commit 1e80159.
    
    * Added hyphen before tokenized enclitics
    
    Looking at the Perseus NLP data, many of the tokenized enclitics
    are distinguished using a hyphen, e.g. "-que". This brings it more
    in line with that dataset. It also better distinguishes "ne" from "-ne".
    
    * Reversed order of tokenized enclitics
    
    To better follow Perseus NLP practice, enclitics are now tokenized
    in the order in which they appear, e.g.
    
    'virumque' > ['virum', '-que']
    
    See PerseusDL/treebank_data#8
    
    * Stop tokenizing 'nec'
    
    This is based on the discussion here: PerseusDL/treebank_data#8. Until this larger issue is resolved, it seems best to leave the form 'nec' as is.
    
    * Better handle case for enclitic tokenization
    
    * Updated test_tokenizer.py to reflect recent changes to the Latin tokenizer
    
    * Rewrote "-cum" handling
    
    Tokenization for "-cum" compounds, e.g. mecum, is now done through
    regex replacement on the original string rather than by
    iterating over and check all of the tokens. More efficient,
    easier to read.
    
    Includes a function to maintain case of original after
    replacement.
    
    * Handle exceptions in Latin tokenizer at string level
    
    Following the logic of 'cum' compound handling, exceptional cases
    are now handled through regex replacement at the string level.
    This is much more efficient than the multiple list comprehensions
    currently in use and much easier to read/maintain. It also now
    correctly handles 'similist' and 'qualist'.
    
    * Moved list of latin exceptions to a separate file
    
    * Make tokenizer split final period; update test.
    
    * Fixed typo in previous commit that will make Sanskrit tokenizer test fail
    
    * Updated tokenizer to use local function instead of NLTK's word_tokenize
    
    * Updated tokenizer to use local function instead of NLTK's word_tokenize
    
    * Updates to Latin tokenizer
    
    A few changes:
    - Most significant: Special handling for Latin moved into its own
        function. Makes the general tokenizer code much easier to read
        and makes effort to avoid the clutter than will arise from
        separate exceptions for each language.
    - Latin tokenizer now splits on sentences before splitting on words.
        This allows:
            - Better handling of '-ne' enclitic which can now be tested
                only on the sentence initial position.
            - Custom handling of Latin abbreviations. The test case
                included here are the praenomina; e.g. sentences will no
                longer incorrectly split on the name "Cn."
Commits on Jun 3, 2016
  1. Updates to tokenizer (#308)

    diyclassics authored and kylepjohnson committed Jun 3, 2016
    * Add feature to load Latin corpora directly using PlaintextCorpusReader
    
    Using the Latin Library corpus as a test case, this feature allows
    you to refer to a corpus and load it with PlaintextCorpusReader using
    the following syntax:
    
    from cltk.corpus.latin import latinlibrary
    
    The code checks to make sure that a corpus is installed in the
    main CLTK_DATA and raises an error if it is not there.
    
    * Added missing comma to list 'cum_inclusions'
    
    * Stop tokenizing 'nec'
    
    This is based on the discussion here: PerseusDL/treebank_data#8. Until this larger issue is resolved, it seems best to leave the form 'nec' as is.
    
    * Added hyphen before tokenized enclitics
    
    Looking at the Perseus NLP data, many of the tokenized enclitics
    are distinguished using a hyphen, e.g. "-que". This brings it more
    in line with that dataset. It also better distinguishes "ne" from "-ne".
    
    * Add 'neque' to que_exceptions
    
    Like 'nec', 'neque' should no longer be separated as two tokens to
    bring the tokenizer more in line with Perseus NLP data (see
    previous commit).
    
    * Revert "Stop tokenizing 'nec'"
    
    This reverts commit 1e80159.
    
    * Added hyphen before tokenized enclitics
    
    Looking at the Perseus NLP data, many of the tokenized enclitics
    are distinguished using a hyphen, e.g. "-que". This brings it more
    in line with that dataset. It also better distinguishes "ne" from "-ne".
    
    * Reversed order of tokenized enclitics
    
    To better follow Perseus NLP practice, enclitics are now tokenized
    in the order in which they appear, e.g.
    
    'virumque' > ['virum', '-que']
    
    See PerseusDL/treebank_data#8
    
    * Stop tokenizing 'nec'
    
    This is based on the discussion here: PerseusDL/treebank_data#8. Until this larger issue is resolved, it seems best to leave the form 'nec' as is.
    
    * Better handle case for enclitic tokenization
    
    * Updated test_tokenizer.py to reflect recent changes to the Latin tokenizer
    
    * Rewrote "-cum" handling
    
    Tokenization for "-cum" compounds, e.g. mecum, is now done through
    regex replacement on the original string rather than by
    iterating over and check all of the tokens. More efficient,
    easier to read.
    
    Includes a function to maintain case of original after
    replacement.
    
    * Handle exceptions in Latin tokenizer at string level
    
    Following the logic of 'cum' compound handling, exceptional cases
    are now handled through regex replacement at the string level.
    This is much more efficient than the multiple list comprehensions
    currently in use and much easier to read/maintain. It also now
    correctly handles 'similist' and 'qualist'.
    
    * Moved list of latin exceptions to a separate file
    
    * Make tokenizer split final period; update test.
    
    * Fixed typo in previous commit that will make Sanskrit tokenizer test fail
    
    * Updated tokenizer to use local function instead of NLTK's word_tokenize
    
    * Updated tokenizer to use local function instead of NLTK's word_tokenize
Commits on Jun 1, 2016
  1. Updates to Latin tokenizer (#303)

    diyclassics authored and kylepjohnson committed Jun 1, 2016
    * Add feature to load Latin corpora directly using PlaintextCorpusReader
    
    Using the Latin Library corpus as a test case, this feature allows
    you to refer to a corpus and load it with PlaintextCorpusReader using
    the following syntax:
    
    from cltk.corpus.latin import latinlibrary
    
    The code checks to make sure that a corpus is installed in the
    main CLTK_DATA and raises an error if it is not there.
    
    * Added missing comma to list 'cum_inclusions'
    
    * Stop tokenizing 'nec'
    
    This is based on the discussion here: PerseusDL/treebank_data#8. Until this larger issue is resolved, it seems best to leave the form 'nec' as is.
    
    * Added hyphen before tokenized enclitics
    
    Looking at the Perseus NLP data, many of the tokenized enclitics
    are distinguished using a hyphen, e.g. "-que". This brings it more
    in line with that dataset. It also better distinguishes "ne" from "-ne".
    
    * Add 'neque' to que_exceptions
    
    Like 'nec', 'neque' should no longer be separated as two tokens to
    bring the tokenizer more in line with Perseus NLP data (see
    previous commit).
    
    * Revert "Stop tokenizing 'nec'"
    
    This reverts commit 1e80159.
    
    * Added hyphen before tokenized enclitics
    
    Looking at the Perseus NLP data, many of the tokenized enclitics
    are distinguished using a hyphen, e.g. "-que". This brings it more
    in line with that dataset. It also better distinguishes "ne" from "-ne".
    
    * Reversed order of tokenized enclitics
    
    To better follow Perseus NLP practice, enclitics are now tokenized
    in the order in which they appear, e.g.
    
    'virumque' > ['virum', '-que']
    
    See PerseusDL/treebank_data#8
    
    * Stop tokenizing 'nec'
    
    This is based on the discussion here: PerseusDL/treebank_data#8. Until this larger issue is resolved, it seems best to leave the form 'nec' as is.
    
    * Better handle case for enclitic tokenization
    
    * Updated test_tokenizer.py to reflect recent changes to the Latin tokenizer
    
    * Rewrote "-cum" handling
    
    Tokenization for "-cum" compounds, e.g. mecum, is now done through
    regex replacement on the original string rather than by
    iterating over and check all of the tokens. More efficient,
    easier to read.
    
    Includes a function to maintain case of original after
    replacement.
    
    * Handle exceptions in Latin tokenizer at string level
    
    Following the logic of 'cum' compound handling, exceptional cases
    are now handled through regex replacement at the string level.
    This is much more efficient than the multiple list comprehensions
    currently in use and much easier to read/maintain. It also now
    correctly handles 'similist' and 'qualist'.
Commits on May 29, 2016
  1. Add feature to load Latin corpora directly using PlaintextCorpusReader (

    diyclassics authored and kylepjohnson committed May 29, 2016
    #296)
    
    Using the Latin Library corpus as a test case, this feature allows
    you to refer to a corpus and load it with PlaintextCorpusReader using
    the following syntax:
    
    from cltk.corpus.latin import latinlibrary
    
    The code checks to make sure that a corpus is installed in the
    main CLTK_DATA and raises an error if it is not there.
Commits on Mar 15, 2016
  1. Fixed indentation; missing assignment

    diyclassics committed Mar 15, 2016
Commits on Mar 14, 2016
  1. Update Latin tokenizer to handle '-n'

    diyclassics committed Mar 14, 2016
    This is an update on handling the 'ne' enclitic where the final 'e' is
    dropped. E.g. 'cenavin' > ['cenavi', 'ne']
    
    It also handles (as exceptions though) the forms 'satin', 'scin', and 'ain'
    where the 's' is elided as well. E.g. 'satin' > ['satis', 'ne'].
    
    Test updated.
Commits on Mar 11, 2016
  1. Merge branch 'master' of https://github.com/diyclassics/cltk

    diyclassics committed Mar 11, 2016
Commits on Mar 10, 2016
  1. Merge branch 'master' of https://github.com/cltk/cltk

    diyclassics committed Mar 10, 2016
  2. Update Latin tokenizer to handle more contractions

    diyclassics committed Mar 10, 2016
    The contractions covered in this update are:
        'sodes' > ['si', 'audes']
        'sultis' > ['si', 'vultis']
    
    They are handled (for now) as exceptions and split using list
    comprehensions (like 'nec' in previous update).
    
    Test is also updated.
  3. Merge branch 'master' of https://github.com/cltk/cltk

    diyclassics committed Mar 10, 2016
  4. Update Latin tokenizer to handle more contractions of 'est'

    diyclassics committed Mar 10, 2016
    Last version of the Latin tokenizer only handled contracts of 'est' with
    preceding words ending in 'm' (e.g. 'verumst' > ['verum', 'est']. This
    update handles more (but not all!) 'est' contractions, including the
    most common ones (e.g. 'opust', 'necessest')
    
    Test for Latin word tokenizer updated as well.
Commits on Feb 26, 2016
  1. Merge branch 'master' into master

    diyclassics committed Feb 26, 2016
  2. Merge branch 'master' of https://github.com/diyclassics/cltk

    diyclassics committed Feb 26, 2016