-
Notifications
You must be signed in to change notification settings - Fork 5
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Add initial analysis notebook and readme
- Loading branch information
mitchell
committed
Apr 24, 2017
1 parent
df4b0cf
commit 89c4a74
Showing
3 changed files
with
705 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -87,3 +87,5 @@ ENV/ | |
|
||
# Rope project settings | ||
.ropeproject | ||
|
||
data/ |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,2 +1,28 @@ | ||
# KaggleQuoraQuestionSimilarity | ||
Github repo for kaggle quora question similarity problem | ||
|
||
# Data | ||
The data size for this compoetition is too large to host in GitHub directly. | ||
All notebooks in this repo expect the following data to be available on your machine, in the following directories: | ||
|
||
data/ | ||
* sample_submission.csv - from the [competition website](https://www.kaggle.com/c/quora-question-pairs/data) | ||
* test.csv - from the [competition website](https://www.kaggle.com/c/quora-question-pairs/data) | ||
* train.csv - from the [competition website](https://www.kaggle.com/c/quora-question-pairs/data) | ||
* wikitext-103/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/) | ||
* wiki.test.tokens | ||
* wiki.valid.tokens | ||
* wiki.train.tokens | ||
* wikitext-2/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/) | ||
* wiki.test.tokens | ||
* wiki.train.tokens | ||
* wiki.valid.tokens | ||
* wikitext-103-raw/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/) | ||
* wiki.test.raw | ||
* wiki.valid.raw | ||
* wiki.train.raw | ||
* wikitext-2-raw/ - from the [wikitext dataset](https://metamind.io/research/the-wikitext-long-term-dependency-language-modeling-dataset/) | ||
* wiki.test.raw | ||
* wiki.train.raw | ||
* wiki.valid.raw | ||
|
Oops, something went wrong.