-
Notifications
You must be signed in to change notification settings - Fork 0
/
Tokenize_friso
33 lines (31 loc) · 1.18 KB
/
Tokenize_friso
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
#Apr28 modified tokenize
#May14 tokenize2: double check new sogou dict
#for every line
import os
#from threading import Thread
#Tried to use subprocess to talk to friso
#now use another tokenizer smallseg
from gensim import corpora, models, similarities
from smallseg import SEG
seg = SEG()
all_user_folder = "../timelines/"
for user_name in os.listdir(all_user_folder):
user_folder = os.path.join(all_user_folder, user_name)
# print "join folder"
if os.path.isdir(user_folder):
# print "is dir"
user_all_text = user_name + '_text'
user_tokenized_text = user_name + '_tokenized'
print "corpus is:"
print user_folder + user_all_text
with open(os.path.join(user_folder, user_all_text),'r') as user_text_corpus:
# print "in corpus"
for line in user_text_corpus:
print line
#every line of tweets has a list of words wlist
wlist = seg.cut(line)
wlist.reverse()
print wlist
tmp = " ".join(wlist)
fout = open(os.path.join("../UserTokenized_2", user_tokenized_text),'a')
fout.write(tmp.encode('utf-8'))