{"payload":{"header_redesign_enabled":false,"results":[{"id":"701887502","archived":false,"color":"#3572A5","followers":17,"has_funding_file":false,"hl_name":"samzshi0529/HanziNLP","hl_trunc_description":"A NLP package for Chinese text:Preprocessing, Tokenization, Chinese Fonts, Word Embeddings, Text Similarity and Sentiment Analysis 轻量级中文自…","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":701887502,"name":"HanziNLP","owner_id":60007017,"owner_login":"samzshi0529","updated_at":"2024-03-27T00:34:57.361Z","has_issues":true}},"sponsorable":false,"topics":["nlp","machine-learning","sentiment-analysis","chinese","hanzi","nlp-machine-learning","tokenization","hanzi-pinyin","chinese-tokenizer"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":89,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Asamzshi0529%252FHanziNLP%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/samzshi0529/HanziNLP/star":{"post":"6kuFJLB65dr4dxOqQWU2aE5-zFBclaBUgA1_H6W4C1wih_gmkDbe9jipMyo4aSorHCeRTGoCy2l2RlEx6fbhAQ"},"/samzshi0529/HanziNLP/unstar":{"post":"o_hzi_9QdCaH4dZW8skFBIK-wCm9qGqRMbC1jvRtlj5qNSIDHM4y9MlmriovflPv4OAWXxaRHW8Zd9mcz1gGMQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"TabOFD3TJoBlVpeW3N9QMHJH9HoEwgXttURabajMSzF8rYm2vX7x97Z1F_Ep6ybGVzgyfifFo25aEnws5kMmew"}}},"title":"Repository search results"}