Skip to content

Commit

Permalink
Add initial analysis notebook and readme
Browse files Browse the repository at this point in the history
  • Loading branch information
mitchell committed Apr 24, 2017
1 parent df4b0cf commit 89c4a74
Show file tree
Hide file tree
Showing 3 changed files with 705 additions and 0 deletions.
2 changes: 2 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -87,3 +87,5 @@ ENV/

# Rope project settings
.ropeproject

data/
26 changes: 26 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,2 +1,28 @@
# KaggleQuoraQuestionSimilarity
Github repo for kaggle quora question similarity problem

# Data
The data size for this compoetition is too large to host in GitHub directly.
All notebooks in this repo expect the following data to be available on your machine, in the following directories:

data/
* sample_submission.csv - from the [competition website](https://www.kaggle.com/c/quora-question-pairs/data)
* test.csv - from the [competition website](https://www.kaggle.com/c/quora-question-pairs/data)
* train.csv - from the [competition website](https://www.kaggle.com/c/quora-question-pairs/data)
* wikitext-103/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/)
* wiki.test.tokens
* wiki.valid.tokens
* wiki.train.tokens
* wikitext-2/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/)
* wiki.test.tokens
* wiki.train.tokens
* wiki.valid.tokens
* wikitext-103-raw/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/)
* wiki.test.raw
* wiki.valid.raw
* wiki.train.raw
* wikitext-2-raw/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/)
* wiki.test.raw
* wiki.train.raw
* wiki.valid.raw

0 comments on commit 89c4a74

Please sign in to comment.