Skip to content

ayush1999/TextAnalysis.jl

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TextAnalysis.jl

Build Status TextAnalysis TextAnalysis

Preface

This manual is designed to get you started doing text analysis in Julia. It assumes that you already familiar with the basic methods of text analysis.

Outline

Installation

The TextAnalysis package can be installed using Julia's package manager:

Pkg.add("TextAnalysis")

Getting Started

In all of the examples that follow, we'll assume that you have the TextAnalysis package fully loaded. This means that we think you've implicily typed

using TextAnalysis

before every snippet of code.

Creating Documents

The basic unit of text analysis is a document. The TextAnalysis package allows one to work with documents stored in a variety of formats:

  • FileDocument: A document represented using a plain text file on disk
  • StringDocument: A document represented using a UTF8 String stored in RAM
  • TokenDocument: A document represented as a sequence of UTF8 tokens
  • NGramDocument: A document represented as a bag of n-grams, which are UTF8 n-grams that map to counts

These format represent a hierarchy: you can always move down the hierachy, but can generally not move up the hierachy. A FileDocument can easily become a StringDocument, but an NGramDocument cannot easily become a FileDocument.

Creating any of the four basic types of documents is very easy:

str = "To be or not to be..."
sd = StringDocument(str)

pathname = "/usr/share/dict/words"
fd = FileDocument(pathname)

my_tokens = String["To", "be", "or", "not", "to", "be..."]
td = TokenDocument(my_tokens)

my_ngrams = Dict{String, Int}("To" => 1, "be" => 2,
                                "or" => 1, "not" => 1,
                                "to" => 1, "be..." => 1)
ngd = NGramDocument(my_ngrams)

For every type of document except a FileDocument, you can also construct a new document by simply passing in a string of text:

sd = StringDocument("To be or not to be...")
td = TokenDocument("To be or not to be...")
ngd = NGramDocument("To be or not to be...")

The system will automatically perform tokenization or n-gramization in order to produce the required data. Unfortunately, FileDocument's cannot be constructed this way because filenames are themselves strings. It would cause chaos if filenames were treated as the text contents of a document.

That said, there is one way around this restriction: you can use the generic Document() constructor function, which will guess at the type of the inputs and construct the appropriate type of document object:

Document("To be or not to be...")
Document("/usr/share/dict/words")
Document(String["To", "be", "or", "not", "to", "be..."])
Document(Dict{String, Int}("a" => 1, "b" => 3))

This constructor is very convenient for working in the REPL, but should be avoided in permanent code because, unlike the other constructors, the return type of the Document function cannot be known at compile-time.

Basic Functions for Working with Documents

Once you've created a document object, you can work with it in many ways. The most obvious thing is to access its text using the text() function:

text(sd)

This function works without warnings on StringDocument's and FileDocument's. For TokenDocument's it is not possible to know if the text can be reconstructed perfectly, so calling text(TokenDocument("This is text")) will produce a warning message before returning an approximate reconstruction of the text as it existed before tokenization. It is entirely impossible to reconstruct the text of an NGramDocument, so text(NGramDocument("This is text")) raises an error.

Instead of working with the text itself, you can work with the tokens or n-grams of a document using the tokens() and ngrams() functions:

tokens(sd)
ngrams(sd)

By default the ngrams() function produces unigrams. If you would like to produce bigrams or trigrams, you can specify that directly using a numeric argument to the ngrams() function:

ngrams(sd, 2)

If you have a NGramDocument, you can determine whether an NGramDocument contains unigrams, bigrams or a higher-order representation using the ngram_complexity() function:

ngram_complexity(ngd)

This information is not available for other types of Document objects because it is possible to produce any level of complexity when constructing n-grams from raw text or tokens.

Document Metadata

In addition to methods for manipulating the representation of the text of a document, every document object also stores basic metadata about itself, including the following pieces of information:

  • language(): What language is the document in? Defaults to EnglishLanguage, a Language type defined by the Languages package.
  • name(): What is the name of the document? Defaults to "Unnamed Document".
  • author(): Who wrote the document? Defaults to "Unknown Author".
  • timestamp(): When was the document written? Defaults to "Unknown Time".

Try these functions out on a StringDocument to see how the defaults work in practice:

language(sd)
name(sd)
author(sd)
timestamp(sd)

If you need reset these fields, you can use the mutating versions of the same functions:

language!(sd, Languages.SpanishLanguage)
name!(sd, "El Cid")
author!(sd, "Desconocido")
timestamp!(sd, "Desconocido")

Preprocessing Documents

Having easy access to the text of a document and its metadata is very important, but most text analysis tasks require some amount of preprocessing.

At a minimum, your text source may contain corrupt characters. You can remove these using the remove_corrupt_utf8!() function:

remove_corrupt_utf8!(sd)

Alternatively, you may want to edit the text to remove items that are hard to process automatically. For example, our sample text sentence taken from Hamlet has three periods that we might like to discard. We can remove this kind of punctuation using the remove_punctuation!() function:

remove_punctuation!(sd)

Like punctuation, numbers and case distinctions are often easier removed than dealt with. To remove numbers or case distinctions, use the remove_numbers!() and remove_case!() functions:

remove_numbers!(sd)
remove_case!(sd)

At times you'll want to remove specific words from a document like a person's name. To do that, use the remove_words!() function:

sd = StringDocument("Lear is mad")
remove_words!(sd, ["Lear"])

At other times, you'll want to remove whole classes of words. To make this easier, we can use several classes of basic words defined by the Languages.jl package:

  • Articles: "a", "an", "the"
  • Indefinite Articles: "a", "an"
  • Definite Articles: "the"
  • Prepositions: "across", "around", "before", ...
  • Pronouns: "I", "you", "he", "she", ...
  • Stop Words: "all", "almost", "alone", ...

These special classes can all be removed using specially-named functions:

  • remove_articles!()
  • remove_indefinite_articles!()
  • remove_definite_articles!()
  • remove_prepositions!()
  • remove_pronouns!()
  • remove_stop_words!()

These functions use words lists, so they are capable of working for many different languages without change:

remove_articles!(sd)
remove_indefinite_articles!(sd)
remove_definite_articles!(sd)
remove_prepositions!(sd)
remove_pronouns!(sd)
remove_stop_words!(sd)

In addition to removing words, it is also common to take words that are closely related like "dog" and "dogs" and stem them in order to produce a smaller set of words for analysis. We can do this using the stem!() function:

stem!(sd)

Creating a Corpus

Working with isolated documents gets boring quickly. We typically want to work with a collection of documents. We represent collections of documents using the Corpus type:

crps = Corpus(Any[StringDocument("Document 1"),
                  StringDocument("Document 2")])

Standardizing a Corpus

A Corpus may contain many different types of documents:

crps = Corpus(Any[StringDocument("Document 1"),
                  TokenDocument("Document 2"),
                  NGramDocument("Document 3")])

It is generally more convenient to standardize all of the documents in a corpus using a single type. This can be done using the standardize! function:

standardize!(crps, NGramDocument)

After this step, you can check that the corpus only contains NGramDocument's:

crps

Processing a Corpus

We can apply the same sort of preprocessing steps that are defined for individual documents to an entire corpus at once:

crps = Corpus(Any[StringDocument("Document 1"),
                  StringDocument("Document 2")])
remove_punctuation!(crps)

These operations are run on each document in the corpus individually.

Corpus Statistics

Often we wish to think broadly about properties of an entire corpus at once. In particular, we want to work with two constructs:

  • Lexicon: The lexicon of a corpus consists of all the terms that occur in any document in the corpus. The lexical frequency of a term tells us how often a term occurs across all of the documents. Often the most interesting words in a document are those words whose frequency within a document is higher than their frequency in the corpus as a whole.
  • Inverse Index: If we are interested in a specific term, we often want to know which documents in a corpus contain that term. The inverse index tells us this and therefore provides a simplistic sort of search algorithm.

Because computations involving the lexicon can take a long time, a Corpus's default lexicon is blank:

lexicon(crps)

In order to work with the lexicon, you have to update it and then access it:

update_lexicon!(crps)
lexicon(crps)

But once this work is done, you can easier address lots of interesting questions about a corpus:

lexical_frequency(crps, "Summer")
lexical_frequency(crps, "Document")

Like the lexicon, the inverse index for a corpus is blank by default:

inverse_index(crps)

Again, you need to update it before you can work with it:

update_inverse_index!(crps)
inverse_index(crps)

But once you've updated the inverse index, you can easily search the entire corpus:

crps["Document"]
crps["1"]
crps["Summer"]

Converting a DataFrame from a Corpus

Sometimes we want to apply non-text specific data analysis operations to a corpus. The easiest way to do this is to convert a Corpus object into a DataFrame:

convert(DataFrame, crps)

Creating a Document Term Matrix

Often we want to represent documents as a matrix of word counts so that we can apply linear algebra operations and statistical techniques. Before we do this, we need to update the lexicon:

update_lexicon!(crps)
m = DocumentTermMatrix(crps)

A DocumentTermMatrix object is a special type. If you would like to use a simple sparse matrix, call dtm() on this object:

dtm(m)

If you would like to use a dense matrix instead, you can pass this as an argument to the dtm function:

dtm(m, :dense)

Creating Individual Rows of a Document Term Matrix

In many cases, we don't need the entire document term matrix at once: we can make do with just a single row. You can get this using the dtv function. Because individual's document do not have a lexicon associated with them, we have to pass in a lexicon as an additional argument:

dtv(crps[1], lexicon(crps))

The Hash Trick

The need to create a lexicon before we can construct a document term matrix is often prohibitive. We can often employ a trick that has come to be called the "Hash Trick" in which we replace terms with their hashed valued using a hash function that outputs integers from 1 to N. To construct such a hash function, you can use the TextHashFunction(N) constructor:

h = TextHashFunction(10)

You can see how this function maps strings to numbers by calling the index_hash function:

index_hash("a", h)
index_hash("b", h)

Using a text hash function, we can represent a document as a vector with N entries by calling the hash_dtv function:

hash_dtv(crps[1], h)

This can be done for a corpus as a whole to construct a DTM without defining a lexicon in advance:

hash_dtm(crps, h)

Every corpus has a hash function built-in, so this function can be called using just one argument:

hash_dtm(crps)

Moreover, if you do not specify a hash function for just one row of the hash DTM, a default hash function will be constructed for you:

hash_dtv(crps[1])

TF-IDF

In many cases, raw word counts are not appropriate for use because:

  • (A) Some documents are longer than other documents
  • (B) Some words are more frequent than other words

You can work around this by performing TF-IDF on a DocumentTermMatrix:

m = DocumentTermMatrix(crps)
tf_idf(m)

As you can see, TF-IDF has the effect of inserting 0's into the columns of words that occur in all documents. This is a useful way to avoid having to remove those words during preprocessing.

LSA: Latent Semantic Analysis

Often we want to think about documents from the perspective of semantic content. One standard approach to doing this is to perform Latent Semantic Analysis or LSA on the corpus. You can do this using the lsa function:

lsa(crps)

LDA: Latent Dirichlet Allocation

Another way to get a handle on the semantic content of a corpus is to use Latent Dirichlet Allocation:

m = DocumentTermMatrix(crps)
k = 2            # number of topics
iteration = 1000 # number of gibbs sampling iterations
alpha = 0.1      # hyper parameter
beta = 0.1       # hyber parameter
l = lda(m, k, iteration, alpha, beta) # l is k x word matrix.
                                      # value is probablity of occurrence of a word in a topic.

Extended Usage Example

To show you how text analysis might work in practice, we're going to work with a text corpus composed of political speeches from American presidents given as part of the State of the Union Address tradition.

using TextAnalysis, DimensionalityReduction, Clustering

crps = DirectoryCorpus("sotu")

standardize!(crps, StringDocument)

crps = Corpus(crps[1:30])

remove_case!(crps)
remove_punctuation!(crps)

update_lexicon!(crps)
update_inverse_index!(crps)

crps["freedom"]

m = DocumentTermMatrix(crps)

D = dtm(m, :dense)

T = tf_idf(D)

cl = kmeans(T, 5)

Sentiment Analysis

This package also provides sentiment analysis functionality. It uses a pretrained model, which was trained using Keras on the Keras IMDb sentiment analysis dataset. This model was then imported into Flux using Keras.jl.

model = TextAnalysis.SentimentAnalyser()

model("i had a great time")
1

model("it was a pretty boring movie")
0

About

Julia package for text analysis

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Julia 100.0%