Skip to content

Swift Package Manager wrapper around MeCab, a tokenizer / morphological analyzer for Japanese

License

Notifications You must be signed in to change notification settings

shinjukunian/Mecab-Swift

Repository files navigation

Swift

Mecab-Swift

This package provides a Swift wrapper around MeCab https://taku910.github.io/mecab/, a part-of-speech and morphological analyzer for Japanese. MeCab can tokenize Japanese text, provide readings for words containing Kanji characters as well as part-of-speech annotation of the tokens. This package is used in Furiganify https://apps.apple.com/us/app/furiganify/id1151320968?mt=12 and FuriganaPDF https://apps.apple.com/us/app/furigana-pdf/id1516570722.

Installation

Using the Swift package manager. Simply add

.package(url: "https://github.com/shinjukunian/Mecab-Swift", .branch("master"))

to dependencies in your Package.swift file or add Mecab-Swift via Xcode as a package dependency using https://github.com/shinjukunian/Mecab-Swift as the URL. Mecab-Swift contains the following targets:

  • Dictionary: This package provides protocols, i.e. DictionaryProviding, that can be used to use other dictionaries with Mecab-Swift
  • Mecab-Swift: The package that provides the core functionality, i.e. tokenization and tagging
  • IPADic: This package wraps the IPADic dictionary ready to use for Mecab and provides a sample implementation of DictionaryProviding.
  • StringTools: Various tools for handling Japanese text and a wrapper around CFStringTokenizer, which provides some of the functionality of Mecab on Apple platforms
  • CharacterFilter: Character lists of Japanese Kanji characters by school year. Useful for formatting Furigana annotations.

Usage

Mecab-Swift requires dictionary files to work. This package includes the IPADic dictionary (https://github.com/taku910/mecab/tree/master/mecab-ipadic), which is quite old. A number of dictionaries compatible with Mecab are available on the internet. To use a dictionary, you have to tell Mecab-Swift how to interpret the information returned from the tokenizer. This is achieved by conforming to the DictionaryProviding protocol, see the IPADic target for reference.

Mecab-Swift provides a playground that illustrates some use cases.

import IPADic
import Mecab_Swift

Using a brief text

let text = "蜂蜜は熊の大好物です。"

Instantiate the tokenizer with IPADic

let ipadic=IPADic()
let ipadicTokenizer = try Tokenizer(dictionary: ipadic)

To get the tokens, we can use

let ipadicTokens=ipadicTokenizer.tokenize(text: text, transliteration: .hiragana)
//[Base: 蜂蜜, reading: はちみつ, POS: noun, Base: は, reading: は, POS: particle, Base: 熊, reading: くま, POS: noun, Base: の, reading: の, POS: particle, Base: 大, reading: だい, POS: prefix, Base: 好物, reading: こうぶつ, POS: noun, Base: です, reading: です, POS: unknown, Base: 。, reading: 。, POS: symbol]

We can get all nouns in the sentence

let nouns=ipadicFurigana.filter {$0.partOfSpeech == .noun}.map {$0.base}
print("The nouns in \"\(text)\" are \(ListFormatter().string(from: nouns) ?? "")")
//The nouns in "蜂蜜は熊の大好物です。" are 蜂蜜, 熊, and 好物

We can use the tokens to convert the the text to hiragana:

let hiraganized = ipadicTokens.map{$0.reading}.joined()
//はちみつはくまのだいこうぶつです。

or to Romaji

let romajiTokens=ipadicTokenizer.tokenize(text: text, transliteration: .romaji)
let romanized = romajiTokens.map{$0.reading}.joined(separator: " ")
//hachimitsu ha kuma no dai kōbutsu desu 。

We can compare this to the output of the system tokenizer

let system=Tokenizer.systemTokenizer
let systemTokens=system.tokenize(text: text)
//[Base: 蜂蜜, reading: はちみつ, POS: unknown, Base: は, reading: は, POS: unknown, Base: 熊, reading: くま, POS: unknown, Base: の, reading: の, POS: unknown, Base: 大, reading: おお, POS: unknown, Base: 好物, reading: こうぶつ, POS: unknown, Base: です, reading: です, POS: unknown, Base: 。, reading: 。, POS: unknown]
//no part-of-speech annotation here
let hiragana=systemTokens.map {$0.reading}.joined())
//はちみつはくまのおおこうぶつです。
//close, but wrong

One key application is Kanji-to-Kana conversion, e.g. for Furigana annotations. This can be achieved by

let longerText=text + "でも鮭もよく食べます。"
let furigana=ipadicTokenizer.furiganaAnnotations(for: longerText, transliteration: .hiragana, options: [.kanjiOnly])
//[はちみつ, Index(_rawBits: 1)..<Index(_rawBits: 393217), くま, Index(_rawBits: 589825)..<Index(_rawBits: 786433), だい, Index(_rawBits: 983041)..<Index(_rawBits: 1179649), こうぶつ, Index(_rawBits: 1179649)..<Index(_rawBits: 1572865), さけ, Index(_rawBits: 2555905)..<Index(_rawBits: 2752513), た, Index(_rawBits: 3342337)..<Index(_rawBits: 3539713)]

This returns an array of FuriganaAnnotation. The .kanjiOnly option, which is the default, omits furigana for okurigana. FuriganaAnnotations can easily be converted to CTRubyAnnotations for display with CoreText.

Mecab also provides deinflected (lemmatized) forms of Japanese verbs.

let lemmatized = ipadicTokenizer.tokenize(text: "でも鮭もよく食べます。")
    .filter {$0.partOfSpeech == .verb}
    .map {$0.dictionaryForm}
//["食べる"]

On Apple platforms, tokenization is also provided by the NaturalLanguage framework. We can compare the output

let text="でも鮭もよく食べます。"
let NLtokenizer=NLTokenizer(unit: .word)
NLtokenizer.string=text
let NLtokens=NLtokenizer.tokens(for: text.startIndex..<text.endIndex).map{text[$0]}
//["で", "も", "鮭", "も", "よく", "食べ", "ます"]

As of iOS14, part-of-speech tagging and lemmatization appear to be unavailable for Japanese.

Licence

MIT for Mecab-Swift

Mecab and the dictionaries come with their own licence.

About

Swift Package Manager wrapper around MeCab, a tokenizer / morphological analyzer for Japanese

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages