Skip to content

Commit

Permalink
Merge branch 'develop'
Browse files Browse the repository at this point in the history
  • Loading branch information
amaiya committed Apr 15, 2020
2 parents df2043a + 25fc828 commit b92f9ff
Show file tree
Hide file tree
Showing 13 changed files with 1,140 additions and 6 deletions.
14 changes: 14 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,20 @@ Most recent releases are shown at the top. Each release shows:
- **Fixed**: Bug fixes that don't change documented behaviour


## 0.14.0 (2020-04-15)

### New:
- support for building Question-Answering systems
- `textutils` now contains `paragraph_tokenize` function

### Changed
- N/A

### Fixed:
- resolved import issue with `textutils.sent_tokenize'



## 0.13.2 (2020-04-09)

### New:
Expand Down
7 changes: 4 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,8 @@


### News and Announcements
- **2020-04-15:**
- ***ktrain*** **v0.14.x is released** and now includes support for **open-domain question-answering**. See the [example QA notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/develop/examples/text/question_answering_with_bert.ipynb).
- **2020-04-09:**
- ***ktrain*** **v0.13.x is released** and includes support for:
- **link prediction** using graph neural networks - [see example link prediction notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/develop/examples/graphs/cora_link_prediction-GraphSAGE.ipynb) on citation prediction
Expand All @@ -33,8 +35,6 @@ model = txt.sequence_tagger('bilstm-bert', preproc, bert_model='monologg/biobert
learner = ktrain.get_learner(model, train_data=trn, val_data=val, batch_size=128)
learner.fit(0.01, 1, cycle_len=5)
```
- **2020-03-18:**
- ***ktrain*** **v0.11.x is released** and includes various fixes and enhancements to sequence-tagging including abilty to easily use non-English pretrained word embeddings covering 157 languages (e.g., [Dutch NER](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/master/examples/text/CoNLL2002_Dutch-BiLSTM.ipynb))
----

### Overview
Expand All @@ -50,7 +50,8 @@ learner.fit(0.01, 1, cycle_len=5)
- **Unsupervised Topic Modeling** with [LDA](http://www.jmlr.org/papers/volume3/blei03a/blei03a.pdf) <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/master/examples/text/20newsgroups-topic_modeling.ipynb)]</sup></sub>
- **Document Similarity with One-Class Learning**: given some documents of interest, find and score new documents that are semantically similar to them using [One-Class Text Classification](https://en.wikipedia.org/wiki/One-class_classification) <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/master/examples/text/20newsgroups-document_similarity_scorer.ipynb)]</sup></sub>
- **Document Recommendation Engine**: given text from a sample document, recommend documents that are thematically-related to it from a larger corpus <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/master/examples/text/20newsgroups-recommendation_engine.ipynb)]</sup></sub>
- **Text Summarization**: text summarization with a pretrained BART model - no training required <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/develop/examples/text/text_summarization_with_bart.ipynb)]</sup></sub>
- **Text Summarization**: summarize long documents with a pretrained BART model - no training required <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/develop/examples/text/text_summarization_with_bart.ipynb)]</sup></sub>
- **Open-Domain Question-Answering**: ask a large text corpus questions and receive exact answers <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/develop/examples/text/question_answering_with_bert.ipynb)]</sup></sub>
- `vision` data:
- **image classification** (e.g., [ResNet](https://arxiv.org/abs/1512.03385), [Wide ResNet](https://arxiv.org/abs/1605.07146), [Inception](https://www.cs.unc.edu/~wliu/papers/GoogLeNet.pdf)) <sub><sup>[[example notebook](https://nbviewer.jupyter.org/github/amaiya/ktrain/blob/master/examples/vision/dogs_vs_cats-ResNet50.ipynb)]</sup></sub>
- `graph` data:
Expand Down
2 changes: 2 additions & 0 deletions examples/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ This directory contains various example notebooks using *ktrain*. The directory
- [document recommender system](#docrec): given text from a sample document, recommend documents that are semantically similar to it from a larger corpus
- [Shallow NLP](#shallownlp): a small collection of miscellaneous text utilities amenable to being used on machines with only a CPU available (no GPU required)
- [Text Summarization](#bart): an example of text summarization using a pretrained BART model
- [Open-Domain Question-Answering](#textqa): ask questions to a large text corpus and receive exact candidate answers
- `vision`:
- [image classification](#imageclass): models for image datasetsimage classification examples using various models and datasets
- `graphs`:
Expand Down Expand Up @@ -122,6 +123,7 @@ The objective of the CoNLL2003 task is to classify sequences of words as belongi
- [20newsgroups-recommendation_engine.ipynb](https://github.com/amaiya/ktrain/tree/master/examples/text): given text from a sample document, recommend documents that are semantically similar to it from a larger corpus

### <a name="bart"></a>Text Summarization with pretrained BART: [text_summarization_with_bart.ipynb](https://github.com/amaiya/ktrain/tree/master/examples/text)
### <a name="textqa"></a>Open-Domain Question-Answering: [question_answering_with_bert.ipynb](https://github.com/amaiya/ktrain/tree/master/examples/text)


## Vision Data
Expand Down

0 comments on commit b92f9ff

Please sign in to comment.