Releases: entelecheia/lexikanon
Releases · entelecheia/lexikanon
v0.6.5
Fix
- dependencies: Update python and hyfi versions (
d55c830
)
v0.6.4
Fix
- tests: Update HyFI method calls and behavior (
c55189a
)
- Makefile: Enable trust flag for copier command (
1717647
)
v0.6.3
Fix
- dependencies: Upgrade hyfi to 1.29.8 (
7747ddb
)
v0.6.2
Fix
- book: Update library description, add Zenodo link (
b55f063
)
Documentation
- readme: Add Zenodo DOI badge and link, update library description (
5513505
)
v0.6.1
Fix
- tokenizers: Add model validator after method (
cfd8a06
)
- normalizer: Change unescape_html type to Union[bool, str] (
d7e024f
)
- tokenizer: Add formal_en normalizer to nltk config (
b65d01a
)
- stopwords: Add verbose condition to logging (
a7f168b
)
- lexikanon: Adjust NLTKTagger for tagsets and default tag (
54ba18f
)
v0.6.0
Feature
- tokenizer: Add additional postags (
fe95c3e
)
- tokenizer: Add additional postags to nltk config (
a35f684
)
- tokenizers/nltk: Add language support, improve tagset flexibility, download universal_tagset (
82e2514
)
- lexikanon: Add new nltk_universal configuration file (
2c7880f
)
- tokenizer: Add nltk_universal configuration (
e68c5e1
)
- tokenizers: Add MecabTagger and NLTKTagger (
b766fb3
)
- tokenizer/tagger: Implement NLTKTagger (
4f2d945
)
Fix
- tokenizer: Add punctuation postags to mecab.yaml (
1624dbd
)
- tokenizers: Adjust tokenizer base configurations (
c96bfbf
)
- MecabTagger: Correct config_group path (
9b74064
)
v0.5.2
Fix
- dependencies: Upgrade hyfi to 1.20.0 (
1b6c402
)
v0.5.1
Fix
- stopwords: Simplify loading and accessing stopwords (
b4628f9
)
v0.5.0
Feature
- lexikanon: Add find_similar_docs_by_clustering configuration (
51ddf56
)
- lexikanon: Add find_similar_docs_by_clustering configuration (
e2122cf
)
- lexikanon: Add similarity.py for document similarity analysis (
b1fc21b
)
- pyproject.toml: Add scikit-learn dependency (
7dd0014
)
v0.4.3
Fix
- tokenizers: Add Tokenizer to lexikanon tokenizers (
66712af
)
- lexikanon: Change tokenizer_config_name to tokenizer (
dc014bc
)
- lexikanon/pipe/tokenize: Enhance tokenizer function to support string or dict types (
58e32f9
)