forked from pielstroem/Topics
-
Notifications
You must be signed in to change notification settings - Fork 13
/
preprocessing.py
executable file
·1034 lines (846 loc) · 42.7 KB
/
preprocessing.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Preprocessing Text Data, Creating Matrices and Cleaning Corpora
***************************************************************
Functions of this module are for **preprocessing purpose**. You can read text \
files, `tokenize <https://en.wikipedia.org/wiki/Tokenization_(lexical_analysis)>`_ \
and segment documents (if a document is chunked into smaller segments, each segment \
counts as one document), create and read `document-term matrices <https://en.wikipedia.org/wiki/Document-term_matrix>`_, \
determine and remove features. Recurrent variable names are based on the following \
conventions:
* ``corpus`` means an iterable containing at least one ``document``.
* ``document`` means one single string containing all characters of a text \
file, including whitespaces, punctuations, numbers, etc.
* ``dkpro_document`` means a pandas DataFrame containing tokens and additional \
information, e.g. *part-of-speech tags* or *lemmas*, produced by `DARIAH-DKPro-Wrapper <https://github.com/DARIAH-DE/DARIAH-DKPro-Wrapper>`_.
* ``tokenized_corpus`` means an iterable containing at least one ``tokenized_document`` \
or ``dkpro_document``.
* ``tokenized_document`` means an iterable containing tokens of a ``document``.
* ``document_labels`` means an iterable containing names of each ``document`` \
and must have as much elements as ``corpus`` or ``tokenized_corpus`` does.
* ``document_term_matrix`` means either a pandas DataFrame with rows corresponding to \
``document_labels`` and columns to types (distinct tokens in the corpus), whose \
values are token frequencies, or a pandas DataFrame with a MultiIndex \
and only one column corresponding to word frequencies. The first column of the \
MultiIndex corresponds to a document ID (based on ``document_labels``) and the \
second column to a type ID. The first variant is designed for small and the \
second for large corpora.
* ``token2id`` means a dictionary containing a token as key and an unique identifier \
as key, e.g. ``{'first_document': 0, 'second_document': 1}``.
Contents
********
* :func:`add_token2id` adds a token to a ``document_ids`` or ``type_ids`` dictionary \
and assigns an unique identifier.
* :func:`create_document_term_matrix()` creates a document-term matrix, for either \
small or large corpora.
* :func:`filter_pos_tags()` filters a ``dkpro_document`` by specific \
*part-of-speech tags* and returns either tokens or, if available, lemmas.
* :func:`find_hapax_legomena()` determines *hapax legomena* based on frequencies \
of a ``document_term_matrix``.
* :func:`find_stopwords()` determines *most frequent words* based on frequencies \
of a ``document_term_matrix``.
* :func:`read_document_term_matrix()` reads a document-term matrix from a CSV file.
* :func:`read_from_pathlist()` reads one or multiple files based on a pathlist.
* :func:`read_matrix_market_file()` reads a `Matrix Market <http://math.nist.gov/MatrixMarket/formats.html#MMformat>`_ \
file for `Gensim <https://radimrehurek.com/gensim/>`_.
* :func:`read_model()` reads a LDA model.
* :func:`read_token2id()` reads a ``document_ids`` or ``type_ids`` dictionary \
from a CSV file.
* :func:`remove_features()` removes features from a ``document_term_matrix``.
* :func:`segment()` is a wrapper for :func:`segment_fuzzy()` and segments a \
``tokenized_document`` into segments of a certain number of tokens, respecting existing chunks.
* :func:`segment_fuzzy()` segments a ``tokenized_document``, tolerating existing \
chunks (like paragraphs).
* :func:`split_paragraphs()` splits a ``document`` or ``dkpro_document`` by paragraphs.
* :func:`tokenize()` tokenizes a ``document`` based on a Unicode regular expression.
"""
from collections import Counter, defaultdict
import csv
from itertools import chain
from gensim.corpora import MmCorpus
import os
from lxml import etree
import numpy as np
import pandas as pd
import pickle
import regex
import logging
log = logging.getLogger(__name__)
log.addHandler(logging.NullHandler())
logging.basicConfig(level=logging.WARNING,
format='%(levelname)s %(name)s: %(message)s')
def add_token2id(token, token2id):
"""Adds token to token2id dictionary.
With this function you can append a ``token`` to an existing ``token2id`` \
dictionary. If ``token2id`` has *x* elements with *n* identifiers, ``token`` \
will be element *x + 1* with identifier *n + 1*.
Args:
token (str): Token.
token2id (dict): A dictionary with tokens as keys and identifiers as values.
Returns:
An extended token2id dictionary.
Raises:
ValueError, if ``token`` has alread an ID in ``token2id``.
Example:
>>> token = 'example'
>>> token2id = {'text': 0}
>>> len(add_token2id(token, token2id)) == 2
True
"""
if token in token2id.keys():
raise KeyError("{} has already an ID in token2id. Access its value with token2id[token].".format(token))
token2id[token] = len(token2id) + 1
return token2id
def create_document_term_matrix(tokenized_corpus, document_labels, large_corpus=False):
"""Creates a document-term matrix.
With this function you can create a document-term-matrix where rows \
correspond to documents in the collection and columns correspond to terms. \
Use the function :func:`read_from_pathlist()` to read and :func:`tokenize()` \
to tokenize your text files.
Args:
tokenized_corpus (list): Tokenized corpus as an iterable containing one
or more iterables containing tokens.
document_labels (list): Name or label of each text file.
large_corpus (bool, optional): Set to True, if ``tokenized_corpus`` is
very large. Defaults to False.
Returns:
Document-term matrix as pandas DataFrame.
Example:
>>> tokenized_corpus = [['this', 'is', 'document', 'one'], ['this', 'is', 'document', 'two']]
>>> document_labels = ['document_one', 'document_two']
>>> create_document_term_matrix(tokenized_corpus, document_labels) #doctest: +NORMALIZE_WHITESPACE
this is document two one
document_one 1.0 1.0 1.0 0.0 1.0
document_two 1.0 1.0 1.0 1.0 0.0
>>> document_term_matrix, document_ids, type_ids = create_document_term_matrix(tokenized_corpus, document_labels, True)
>>> isinstance(document_term_matrix, pd.DataFrame) and isinstance(document_ids, dict) and isinstance(type_ids, dict)
True
"""
if large_corpus:
return _create_large_corpus_model(tokenized_corpus, document_labels)
else:
return _create_small_corpus_model(tokenized_corpus, document_labels)
def filter_pos_tags(dkpro_document, pos_tags=['ADJ', 'V', 'NN'], lemma=True):
"""Gets tokens or lemmas respectively of selected POS-tags from pandas DataFrame.
With this function you can filter `DARIAH-DKPro-Wrapper <https://github.com/DARIAH-DE/DARIAH-DKPro-Wrapper>`_ \
output. Commit a list of POS-tags to get specific tokens (if ``lemma`` False) \
or lemmas (if ``lemma`` True).
Use the function :func:`read_from_pathlist()` to read CSV files.
Args:
dkpro_document (pandas.DataFrame): DARIAH-DKPro-Wrapper output.
pos_tags (list, optional): List of desired POS-tags. Defaults
to ``['ADJ', 'V', 'NN']``.
lemma (bool, optional): If True, lemmas will be selected, otherwise tokens.
Defaults to True.
Yields:
A pandas DataFrame containing tokens or lemmas.
Example:
>>> dkpro_document = pd.DataFrame({'CPOS': ['ART', 'V', 'ART', 'NN'],
... 'Token': ['this', 'was', 'a', 'document'],
... 'Lemma': ['this', 'is', 'a', 'document']})
>>> list(filter_pos_tags(dkpro_document)) #doctest: +NORMALIZE_WHITESPACE
[1 is
3 document
Name: Lemma, dtype: object]
"""
tokenized_document = dkpro_document[dkpro_document['CPOS'].isin(pos_tags)]
if lemma:
log.info("Selecting {} lemmas ...".format(pos_tags))
yield tokenized_document['Lemma']
else:
log.info("Selecting {} tokens ...".format(pos_tags))
yield tokenized_document['Token']
def find_hapax_legomena(document_term_matrix, type_ids=None):
"""Creates a list with hapax legommena.
With this function you can determine *hapax legomena* for each document. \
Use the function :func:`create_document_term_matrix()` to create a \
document-term matrix.
Args:
document_term_matrix (pandas.DataFrame): A document-term matrix.
type_ids (dict): A dictionary with types as key and identifiers as values.
If ``document_term_matrix`` is designed for large corpora, you have
to commit ``type_ids``, too.
Returns:
Hapax legomena in a list.
Example:
>>> document_labels = ['document']
>>> tokenized_corpus = [['hapax', 'stopword', 'stopword']]
>>> document_term_matrix = create_document_term_matrix(tokenized_corpus, document_labels)
>>> find_hapax_legomena(document_term_matrix)
['hapax']
>>> document_term_matrix, _, type_ids = create_document_term_matrix(tokenized_corpus, document_labels, large_corpus=True)
>>> find_hapax_legomena(document_term_matrix, type_ids)
['hapax']
"""
log.info("Determining hapax legomena ...")
if isinstance(document_term_matrix.index, pd.MultiIndex):
log.debug("Large corpus model ...")
return _hapax_legomena_large_corpus_model(document_term_matrix, type_ids)
else:
log.debug("Small corpus model ...")
return document_term_matrix.loc[:, document_term_matrix.max() == 1].columns.tolist()
def find_stopwords(document_term_matrix, most_frequent_tokens=100, type_ids=None):
"""Creates a list with stopword based on most frequent tokens.
With this function you can determine *most frequent tokens*, also known as \
*stopwords*. First, you have to translate your corpus into a document-term \
matrix.
Use the function :func:`create_document_term_matrix()` to create a \
document-term matrix.
Args:
document_term_matrix (pandas.DataFrame): A document-term matrix.
most_frequent_tokens (int, optional): Treshold for most frequent tokens.
type_ids (dict): If ``document_term_matrix`` is designed for large corpora,
you have to commit ``type_ids``, too.
Returns:
Most frequent tokens in a list.
Example:
>>> document_labels = ['document']
>>> tokenized_corpus = [['hapax', 'stopword', 'stopword']]
>>> document_term_matrix = create_document_term_matrix(tokenized_corpus, document_labels)
>>> find_stopwords(document_term_matrix, 1)
['stopword']
>>> document_term_matrix, _, type_ids = create_document_term_matrix(tokenized_corpus, document_labels, large_corpus=True)
>>> find_stopwords(document_term_matrix, 1, type_ids)
['stopword']
"""
log.info("Determining stopwords ...")
if isinstance(document_term_matrix.index, pd.MultiIndex):
log.debug("Large corpus model ...")
return _stopwords_large_corpus_model(document_term_matrix, type_ids, most_frequent_tokens)
else:
log.debug("Small corpus model ...")
return document_term_matrix.iloc[:, :most_frequent_tokens].columns.tolist()
def read_document_term_matrix(filepath):
"""Reads a document-term matrix from CSV file.
With this function you can read a CSV file containing a document-term \
matrix.
Use the function :func:`create_document_term_matrix()` to create a document-term \
matrix.
Args:
filepath (str): Path to CSV file.
Returns:
A document-term matrix as pandas DataFrame.
Example:
>>> import tempfile
>>> with tempfile.NamedTemporaryFile(suffix='.csv') as tmpfile:
... tmpfile.write(b'this,is,an,example,text\\ndocument,1,0,1,0,1') and True
... tmpfile.flush()
... read_document_term_matrix(tmpfile.name) #doctest: +NORMALIZE_WHITESPACE
True
this is an example text
document 1 0 1 0 1
>>> with tempfile.NamedTemporaryFile(suffix='.csv') as tmpfile:
... tmpfile.write(b'document_id,type_id,0\\n1,1,1') and True
... tmpfile.flush()
... read_document_term_matrix(tmpfile.name) #doctest: +NORMALIZE_WHITESPACE
True
0
document_id type_id
1 1 1
"""
document_term_matrix = pd.read_csv(filepath)
if 'document_id' and 'type_id' in document_term_matrix:
return document_term_matrix.set_index(['document_id', 'type_id'])
else:
return document_term_matrix
def read_from_pathlist(pathlist, file_format=None, xpath_expression='//tei:text', sep='\t', columns=None):
"""Reads text files based on a pathlist.
With this function you can read multiple file formats:
* Plain text files (``.txt``).
* TEI XML files (``.xml``).
* CSV files (``.csv``), e.g. produced by `DARIAH-DKPro-Wrapper <https://github.com/DARIAH-DE/DARIAH-DKPro-Wrapper>`_.
The argument ``pathlist`` is an iterable of full or relative paths. In case of \
CSV files, you have the ability to select specific columns via ``columns``. \
If there are multiple file formats in ``pathlist``, do not specify ``file_format`` \
and file extensions will be considered.
Args:
pathlist (list): One or more paths to text files.
file_format (str, optional): Format of the files. Possible values are
``text``, ``xml`` and ``csv`. If None, file extensions will be considered.
Defaults to None.
xpath_expression (str, optional): XPath expressions to match part of the
XML file. Defaults to ``//tei:text``.
sep (str, optional): Separator of CSV file. Defaults to ``'\\t'``
columns (list, optional): Column name or names for CSV files. If None, the
whole file will be processed. Defaults to None.
Yields:
A ``document`` as str or, in case of a CSV file, a ``dkpro_document`` as a pandas DataFrame.
Raises:
ValueError, if ``file_format`` is not supported.
Example:
>>> import tempfile
>>> with tempfile.NamedTemporaryFile(suffix='.txt') as first:
... pathlist = []
... first.write(b"This is the first example.") and True
... first.flush()
... pathlist.append(first.name)
... with tempfile.NamedTemporaryFile(suffix='.txt') as second:
... second.write(b"This is the second example.") and True
... second.flush()
... pathlist.append(second.name)
... list(read_from_pathlist(pathlist, 'text'))
True
True
['This is the first example.', 'This is the second example.']
"""
log.info("Reading {} files ...".format(len(pathlist)))
for n, file in enumerate(pathlist):
log.debug("File #{}".format(n))
_, extension = os.path.splitext(file)
if file_format == 'text' or extension == '.txt':
yield _read_txt(file)
elif file_format == 'xml' or extension == '.xml':
yield _read_xml(file, xpath_expression)
elif file_format == 'csv' or extension == '.csv':
yield _read_csv(file, sep, csv_columns)
else:
if file_format is None:
log.error("Skipping {}, because the file format {} is not supported.".format(file, extension))
pass
else:
raise ValueError("Unable to read {}, because the file format {} is not supported.".format(file, file_format))
def read_matrix_market_file(filepath):
"""Reads a Matrix Market file for Gensim.
With this function you can read a Matrix Market file to process it with \
`Gensim <https://radimrehurek.com/gensim/>`_.
Args:
filepath (str): Path to Matrix Market file.
Returns:
Matrix Market model for Gensim.
"""
if os.path.splitext(filepath)[1] is not '.mm':
raise ValueError("The file {} is not a Matrix Market file.".format(filepath))
return MmCorpus(filepath)
def read_model(filepath):
"""Reads a LDA model.
With this function you can read a LDA model, if it was saved using :module:`pickle`.
If you want to read MALLET models, you have to specify a parameter of the
function :func:`create_mallet_model()`.
Args:
filepath (str): Path to LDA model, e.g. ``/home/models/model.pickle``.
Returns:
A LDA model.
Example:
>>> import lda
>>> import gensim
>>> import tempfile
>>> a = lda.LDA
>>> with tempfile.NamedTemporaryFile(suffix='.pickle') as tmpfile:
... pickle.dump(a, tmpfile, protocol=pickle.HIGHEST_PROTOCOL)
... tmpfile.flush()
... read_model(tmpfile.name) == a
True
>>> a = gensim.models.LdaModel
>>> with tempfile.NamedTemporaryFile(suffix='.pickle') as tmpfile:
... pickle.dump(a, tmpfile, protocol=pickle.HIGHEST_PROTOCOL)
... tmpfile.flush()
... read_model(tmpfile.name) == a
True
"""
with open(filepath, 'rb') as model:
return pickle.load(model)
def read_token2id(filepath):
"""Reads a token2id dictionary from CSV file.
With this function you can read a CSV-file containing a document or type dictionary.
Args:
filepath (str): Path to CSV file.
Returns:
A dictionary.
Example:
>>> import tempfile
>>> with tempfile.NamedTemporaryFile(suffix='.csv') as tmpfile:
... tmpfile.write(b"0,this\\n1,is\\n2,an\\n3,example") and True
... tmpfile.flush()
... read_token2id(tmpfile.name)
True
{0: 'this', 1: 'is', 2: 'an', 3: 'example'}
"""
dictionary = pd.read_csv(filepath, header=None)
dictionary.index = dictionary[0]
dictionary = dictionary[1]
return dictionary.to_dict()
def remove_features(features, document_term_matrix=None, tokenized_corpus=None, type_ids=None):
"""Removes features based on a list of tokens.
With this function you can clean your corpus (either a document-term matrix \
or a ``tokenized_corpus``) from *stopwords* and *hapax legomena*.
Use the function :func:`create_document_term_matrix()` or :func:`tokenize` to \
create a document-term matrix or to tokenize your corpus, respectively.
Args:
features (list): A list of tokens.
document_term_matrix (pandas.DataFrame, optional): A document-term matrix.
tokenized_corpus (list, optional): An iterable of one or more ``tokenized_document``.
type_ids (dict, optional): A dictionary with types as key and identifiers as values.
Returns:
A clean document-term matrix as pandas DataFrame or ``tokenized_corpus`` as list.
Example:
>>> document_labels = ['document']
>>> tokenized_corpus = [['this', 'is', 'a', 'document']]
>>> document_term_matrix = create_document_term_matrix(tokenized_corpus, document_labels)
>>> features = ['this']
>>> remove_features(features, document_term_matrix) #doctest: +NORMALIZE_WHITESPACE
is document a
document 1.0 1.0 1.0
>>> document_term_matrix, _, type_ids = create_document_term_matrix(tokenized_corpus, document_labels, large_corpus=True)
>>> len(remove_features(features, document_term_matrix, type_ids=type_ids))
3
>>> list(remove_features(features, tokenized_corpus=tokenized_corpus))
[['is', 'a', 'document']]
"""
log.info("Removing features ...")
if document_term_matrix is not None and tokenized_corpus is None:
if isinstance(document_term_matrix.index, pd.MultiIndex):
return _remove_features_from_large_corpus_model(document_term_matrix, type_ids, features)
else:
return _remove_features_from_small_corpus_model(document_term_matrix, features)
elif document_term_matrix is None and tokenized_corpus is not None:
return _remove_features_from_tokenized_corpus(tokenized_corpus, features)
else:
raise ValueError("Commit either document-term matrix or tokenized_corpus.")
def segment(document, segment_size=1000, tolerance=0, chunker=None,
tokenizer=None, flatten_chunks=True, materialize=True):
"""Segments a document into segments of about ``segment_size`` tokens, respecting existing chunks.
Consider you have a document. You wish to split the document into \
segments of about 1000 tokens, but you prefer to keep paragraphs together \
if this does not increase or decrease the token size by more than 5%.
This is a convenience wrapper around :func:`segment_fuzzy()`.
Args:
document (list): The document to process. This is an iterable of
chunks, each of which is an iterable of tokens.
segment_size (int): The target size of each segment, in tokens. Defaults
to 1000.
tolerance (float, optional): How much may the actual segment size differ from
the segment_size? If ``0 < tolerance < 1``, this is interpreted as a
fraction of the segment_size, otherwise it is interpreted as an
absolute number. If ``tolerance < 0``, chunks are never split apart.
Defaults to None.
chunker (callable, optional): A one-argument function that cuts the document into
chunks. If this is present, it is called on the given document.
Defaults to None.
tokenizer (callable, optional): A one-argument function that tokenizes each chunk.
Defaults to None.
flatten_chunks (bool, optional): If True, undo the effect of the chunker by
chaining the chunks in each segment, thus each segment consists of
tokens. This can also be a one-argument function in order to
customize the un-chunking. Defaults to True.
materialize (bool, optional): If True, materializes the segments. Defaults to True.
Example:
>>> segment([['This', 'is', 'the', 'first', 'chunk'],
... ['this', 'is', 'the', 'second', 'chunk']], 2) #doctest: +NORMALIZE_WHITESPACE
[['This', 'is'],
['the', 'first'],
['chunk', 'this'],
['is', 'the'],
['second', 'chunk']]
"""
if chunker is not None:
document = chunker(document)
if tokenizer is not None:
document = map(tokenizer, document)
segments = segment_fuzzy(document, segment_size, tolerance)
if flatten_chunks:
if not callable(flatten_chunks):
def flatten_chunks(segment):
return list(chain.from_iterable(segment))
segments = map(flatten_chunks, segments)
if materialize:
segments = list(segments)
return segments
def segment_fuzzy(document, segment_size=5000, tolerance=0.05):
"""Segments a document, tolerating existing chunks (like paragraphs).
Consider you have a ``document``. You wish to split the ``document`` into \
segments of about 1000 tokens, but you prefer to keep paragraphs together \
if this does not increase or decrease the token size by more than 5%.
Args:
document (list): The document to process. This is an iterable of
chunks, each of which is an iterable of tokens.
segment_size (int, optional): The target length of each segment in tokens.
Defaults to 5000.
tolerance (float, optional): How much may the actual segment size differ from
the ``segment_size``? If ``0 < tolerance < 1``, this is interpreted as a
fraction of the ``segment_size``, otherwise it is interpreted as an
absolute number. If ``tolerance < 0``, chunks are never split apart.
Defaults to 0.05.
Yields:
Segments. Each segment is a list of chunks, each chunk is a list of
tokens.
Example:
>>> list(segment_fuzzy([['This', 'is', 'the', 'first', 'chunk'],
... ['this', 'is', 'the', 'second', 'chunk']], 2)) #doctest: +NORMALIZE_WHITESPACE
[[['This', 'is']],
[['the', 'first']],
[['chunk'], ['this']],
[['is', 'the']],
[['second', 'chunk']]]
"""
if tolerance > 0 and tolerance < 1:
tolerance = round(segment_size * tolerance)
current_segment = []
current_size = 0
carry = None
doc_iter = iter(document)
try:
while True:
chunk = list(carry if carry else next(doc_iter))
carry = None
current_segment.append(chunk)
current_size += len(chunk)
if current_size >= segment_size:
too_long = current_size - segment_size
too_short = segment_size - (current_size - len(chunk))
if tolerance >= 0 and min(too_long, too_short) > tolerance:
chunk_part0 = chunk[:-too_long]
carry = chunk[-too_long:]
current_segment[-1] = chunk_part0
elif too_long >= too_short:
carry = current_segment.pop()
yield current_segment
current_segment = []
current_size = 0
except StopIteration:
pass
if current_segment:
yield current_segment
def split_paragraphs(document, sep=regex.compile(r'\n')):
"""Splits the given document by paragraphs.
With this function you can split a document by paragraphs. In case of a \
document as str, you also have the ability to select a certain regular \
expression to split the document.
Use the function :func:`read_from_pathlist()` to read files.
Args:
document Union(str, pandas.DataFrame): Document text or DARIAH-DKPro-Wrapper output.
sep (regex.Regex, optional): Separator indicating a paragraph.
Returns:
A list of paragraphs.
Example:
>>> document = "First paragraph\\nsecond paragraph."
>>> split_paragraphs(document)
['First paragraph', 'second paragraph.']
>>> dkpro_document = pd.DataFrame({'Token': ['first', 'paragraph', 'second', 'paragraph', '.'],
... 'ParagraphId': [1, 1, 2, 2, 2]})
>>> split_paragraphs(dkpro_document)[0] #doctest: +NORMALIZE_WHITESPACE
Token
ParagraphId
1 first
1 paragraph
"""
if isinstance(document, str):
if not hasattr(sep, 'match'):
sep = regex.compile(sep)
splitted_document = sep.split(document)
return list(filter(str.strip, splitted_document)) #remove elements containing only whitespaces
elif isinstance(document, pd.DataFrame):
grouped_document = document.set_index('ParagraphId').groupby(level=0)
return [paragraphs for _, paragraphs in grouped_document]
def tokenize(document, pattern=r'\p{L}+\p{P}?\p{L}+', lower=True):
"""Tokenizes with Unicode regular expressions.
With this function you can tokenize a ``document`` with a regular expression. \
You also have the ability to commit your own regular expression. The default \
expression is ``\p{Letter}+\p{Punctuation}?\p{Letter}+``, which means one or \
more letters, followed by one or no punctuation, followed by one or more \
letters. So, one letter words will not match. In case you want to lower \
all tokens, set the argument ``lower`` to True (it is by default).
Use the functions :func:`read_from_pathlist()` to read your text files.
Args:
document (str): Document text.
pattern (str, optional): Regular expression to match tokens.
lower (boolean, optional): If True, lowers all characters. Defaults to True.
Yields:
All matching tokens in the ``document``.
Example:
>>> list(tokenize("This is 1 example text."))
['this', 'is', 'example', 'text']
"""
log.info("Tokenizing document ...")
if lower:
log.debug("Lowering all characters ...")
document = document.lower()
compiled_pattern = regex.compile(pattern)
tokenized_document = compiled_pattern.finditer(document)
for match in tokenized_document:
yield match.group()
def _create_bag_of_words(document_labels, tokenized_corpus):
"""Creates a bag-of-words model.
This private function is wrapped in :func:`_create_large_corpus_model()`. The \
first level consists of the document label as key, and the dictionary \
of counts as value. The second level consists of type ID as key, and the \
count of types in document pairs as value.
Args:
document_labels (list): Iterable of document labels.
tokenized_corpus (list): Tokenized corpus as an iterable
containing one or more iterables containing tokens.
Returns:
A bag-of-words model as dictionary of dictionaries, document IDs and type IDs.
Example:
>>> document_labels = ['exampletext']
>>> tokenized_corpus = [['this', 'is', 'an', 'example', 'text']]
>>> bag_of_words, document_ids, type_ids = _create_bag_of_words(document_labels, tokenized_corpus)
>>> isinstance(bag_of_words, dict) and isinstance(document_ids, dict) and isinstance(type_ids, dict)
True
"""
document_ids = _token2id(document_labels)
type_ids = _token2id(tokenized_corpus)
bag_of_words = defaultdict(dict)
for document_label, tokenized_document in zip(document_labels, tokenized_corpus):
bag_of_words[document_label] = Counter([type_ids[token] for token in tokenized_document])
return {document_ids[id_]: doc for id_, doc in bag_of_words.items()}, document_ids, type_ids
def _create_large_corpus_model(tokenized_corpus, document_labels):
"""Creates a document-term matrix for large corpora.
This private function is wrapped in :func:`create_document_term_matrix()` and \
creates a pandas DataFrame containing document and type IDs as MultiIndex \
and type frequencies as values representing the counts of tokens for each \
token in each document.
Args:
tokenized_corpus (list): Tokenized corpus as an iterable
containing one or more iterables containing tokens.
document_labels (list): Iterable of document labels.
Returns:
A document-term matrix as pandas DataFrame, ``document_ids`` and ``type_ids``.
Todo:
* Make the whole function faster.
Example:
>>> tokenized_corpus = [['this', 'is', 'document', 'one'], ['this', 'is', 'document', 'two']]
>>> document_labels = ['document_one', 'document_two']
>>> document_term_matrix, document_ids, type_ids = _create_large_corpus_model(tokenized_corpus, document_labels)
>>> isinstance(document_term_matrix, pd.DataFrame) and isinstance(document_ids, dict) and isinstance(type_ids, dict)
True
"""
bag_of_words, document_ids, type_ids = _create_bag_of_words(document_labels, tokenized_corpus)
multi_index = _create_multi_index(bag_of_words)
document_term_matrix = pd.DataFrame(np.zeros((len(multi_index), 1), dtype=int), index=multi_index)
index_iterator = multi_index.groupby(multi_index.get_level_values('document_id'))
for document_id in range(1, len(multi_index.levels[0]) + 1):
for type_id in [val[1] for val in index_iterator[document_id]]:
document_term_matrix.set_value((document_id, type_id), 0, int(bag_of_words[document_id][type_id]))
return document_term_matrix, document_ids, type_ids
def _create_multi_index(bag_of_words):
"""Creates a MultiIndex for a pandas DataFrame.
This private function is wrapped in :func:`_create_large_corpus_model()`.
Args:
bag_of_words (dict): A bag-of-words model of ``{document_id: {type_id: frequency}}``.
Returns:
Pandas MultiIndex.
Example:
>>> bag_of_words = {1: {1: 2, 2: 3, 3: 4}}
>>> _create_multi_index(bag_of_words)
MultiIndex(levels=[[1], [1, 2, 3]],
labels=[[0, 0, 0], [0, 1, 2]],
names=['document_id', 'type_id'])
"""
tuples = []
for document_id in range(1, len(bag_of_words) + 1):
if len(bag_of_words[document_id]) == 0:
tuples.append((document_id, 0))
for type_id in bag_of_words[document_id]:
tuples.append((document_id, type_id))
return pd.MultiIndex.from_tuples(tuples, names=['document_id', 'type_id'])
def _create_small_corpus_model(tokenized_corpus, document_labels):
"""Creates a document-term matrix for small corpora.
This private function is wrapped in :func:`create_document_term_matrix()`.
Args:
tokenized_corpus (list): Tokenized corpus as an iterable
containing one or more iterables containing tokens.
document_labels (list): Name or label of each text file.
Returns:
Document-term matrix as pandas DataFrame.
Example:
>>> tokenized_corpus = [['this', 'is', 'document', 'one'], ['this', 'is', 'document', 'two']]
>>> document_labels = ['document_one', 'document_two']
>>> _create_small_corpus_model(tokenized_corpus, document_labels) #doctest: +NORMALIZE_WHITESPACE
this is document two one
document_one 1.0 1.0 1.0 0.0 1.0
document_two 1.0 1.0 1.0 1.0 0.0
"""
log.info("Creating document-term matrix for small corpus ...")
document_term_matrix = pd.DataFrame()
for tokenized_document, document_label in zip(tokenized_corpus, document_labels):
log.debug("Updating {} in document-term matrix ...".format(document_label))
current_document = pd.Series(Counter(tokenized_document))
current_document.name = document_label
document_term_matrix = document_term_matrix.append(current_document)
document_term_matrix = document_term_matrix.loc[:, document_term_matrix.sum().sort_values(ascending=False).index]
return document_term_matrix.fillna(0)
def _hapax_legomena_large_corpus_model(document_term_matrix, type_ids):
"""Determines hapax legomena in large corpus model.
This private function is wrapped in :func:`find_hapax_legomena()`.
Args:
document_term_matrix (pandas.DataFrame): A document-term matrix.
type_ids (dict): A dictionary with types as key and identifiers as values.
Returns:
Hapax legomena in a list.
Example:
>>> document_labels = ['document']
>>> tokenized_corpus = [['hapax', 'stopword', 'stopword']]
>>> document_term_matrix, _, type_ids = create_document_term_matrix(tokenized_corpus, document_labels, large_corpus=True)
>>> find_hapax_legomena(document_term_matrix, type_ids)
['hapax']
"""
id2type = {id_: type_ for type_, id_ in type_ids.items()}
document_term_matrix_collapsed = document_term_matrix.groupby(document_term_matrix.index.get_level_values('type_id')).sum()
hapax_legomena = document_term_matrix_collapsed.loc[document_term_matrix_collapsed[0] == 1]
return [id2type[token] for token in hapax_legomena.index.get_level_values('type_id')]
def _read_csv(filepath, sep, columns):
"""Reads a CSV file based on its path.
This private function is wrapped in `read_from_pathlist()`.
Args:
filepath (str): Path to CSV file.
sep (str): Separator of CSV file.
columns (list): Column names for the CSV file. If None, the whole file will be processed.
Returns:
A ``dkpro_document`` as pandas DataFrame with additional information,
e.g. lemmas or POS-tags.
Example:
>>> import tempfile
>>> with tempfile.NamedTemporaryFile(suffix='.csv') as tmpfile:
... tmpfile.write(b"Token,POS\\nThis,ART\\nis,V\\na,ART\\nCSV,NN\\nexample,NN\\n.,PUNC") and True
... tmpfile.flush()
... _read_csv(tmpfile.name, ',', ['Token']) #doctest: +NORMALIZE_WHITESPACE
True
Token
0 This
1 is
2 a
3 CSV
4 example
5 .
"""
log.debug("Reading columns {} of {} ...".format(columns, filepath))
if columns is None:
log.warning("No column names were specified or do not match. The whole file will be processed.")
return pd.read_csv(filepath, sep=sep, quoting=csv.QUOTE_NONE, usecols=columns)
def _read_txt(filepath):
"""Reads a plain text file based on its path.
This private function is wrapped in `read_from_pathlist()`.
Args:
filepath (str): Path to plain text file.
Returns:
A ``document`` as str.
Example:
>>> import tempfile
>>> with tempfile.NamedTemporaryFile(suffix='.txt') as tmpfile:
... tmpfile.write(b"This is a plain text example.") and True
... tmpfile.flush()
... _read_txt(tmpfile.name)
True
'This is a plain text example.'
"""
log.debug("Reading {} ...".format(filepath))
with open(filepath, 'r', encoding='utf-8') as document:
return document.read()
def _read_xml(filepath, xpath_expression):
"""Reads a TEI XML file based on its path.
This private function is wrapped in `read_from_pathlist()`.
Args:
filepath (str): Path to XML file.
xpath_expression (str): XPath expressions to match part of the XML file.
Returns:
Either a ``document`` as str or a list of all parts of the ``document``,
e. g. chapters of a novel.
Example:
>>> import tempfile
>>> with tempfile.NamedTemporaryFile(suffix='.xml') as tmpfile:
... tmpfile.write(b"<text>This is a XML example.</text>") and True
... tmpfile.flush()
... _read_xml(tmpfile.name, '//text')
True
'This is a XML example.'
"""
log.debug("Reading {} matching part or parts of {} ...".format(xpath_expression, filepath))
ns = dict(tei='http://www.tei-c.org/ns/1.0')
tree = etree.parse(filepath)
document = [''.join(element.xpath('.//text()')) for element in tree.xpath(xpath_expression, namespaces=ns)]
if len(document) == 1:
return document[0]
else:
return document
def _remove_features_from_large_corpus_model(document_term_matrix, type_ids, features):
"""Removes features from large corpus model.
This private function is wrapped in :func:`remove_features()`.
Args:
document_term_matrix (pandas.DataFrame): A document-term matrix.
type_ids (dict): A dictionary with types as key and identifiers as values.
features (list): A list of tokens.
Returns:
A clean document-term matrix as pandas DataFrame.
Example:
>>> document_labels = ['document']
>>> tokenized_corpus = [['token', 'stopword', 'stopword']]
>>> document_term_matrix, _, type_ids = create_document_term_matrix(tokenized_corpus, document_labels, large_corpus=True)
>>> len(_remove_features_from_large_corpus_model(document_term_matrix, type_ids, ['token']))
1
"""
features = [token for token in set(type_ids.keys()) if token in features]
return document_term_matrix.drop([type_ids[token] for token in features], level='type_id')
def _remove_features_from_small_corpus_model(document_term_matrix, features):
"""Removes features from small corpus model.
This private function is wrapped in :func:`remove_features()`.
Args:
document_term_matrix (pandas.DataFrame): A document-term matrix.
features (list): A list of tokens.
Returns:
A clean document-term matrix as pandas DataFrame.
Example:
>>> document_labels = ['document']
>>> tokenized_corpus = [['token', 'stopword', 'stopword']]
>>> document_term_matrix = create_document_term_matrix(tokenized_corpus, document_labels)
>>> _remove_features_from_small_corpus_model(document_term_matrix, ['token']) #doctest: +NORMALIZE_WHITESPACE
stopword
document 2.0
"""
features = [token for token in features if token in document_term_matrix.columns]
return document_term_matrix.drop(features, axis=1)
def _remove_features_from_tokenized_corpus(tokenized_corpus, features):
"""Removes features from a tokenized corpus.
This private function is wrapped in :func:`remove_features()`.
Args:
tokenized_corpus (list): The tokenized corpus to process. This is an iterable of
documents, each of which is an iterable of tokens.
features (list): A list of tokens.
Yields:
A clean tokenized corpus as list.
Example:
>>> tokenized_corpus = [['token', 'stopword', 'stopword']]
>>> list(_remove_features_from_tokenized_corpus(tokenized_corpus, ['stopword']))
[['token']]
"""
tokenized_corpus_arr = np.array(tokenized_corpus)
features_arr = np.array(features)
indices = np.where(np.in1d(tokenized_corpus_arr, features_arr))
yield np.delete(tokenized_corpus_arr, indices).tolist()
def _stopwords_large_corpus_model(document_term_matrix, type_ids, most_frequent_tokens):
"""Determines stopwords in large corpus model.
This private function is wrapped in :func:`find_stopwords()`.
Args:
document_term_matrix (pandas.DataFrame): A document-term matrix.
type_ids (dict): A dictionary with types as key and identifiers as values.
most_frequent_tokens (int, optional): Treshold for most frequent tokens.
Returns:
Most frequent tokens in a list.