Skip to content

saneem89/DavanagereWorkshop

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 

Repository files navigation

DavanagereWorkshop

Workshop on Deep Learning and its applications. Speakers - Saneem Chemmengath, Sneha Mondal, and Abhirut Gupta.

Contact Information

Pre-requisites

  1. Please download the following datasets -

You will need to create a Kaggle account, and accept the terms of the competition. (Follow the instructions from the Download All button.

  1. You will need Python for the workshop, we suggest installing Anaconda for Python 3.7 Installer links for different OS can be found at -
Mac - https://repo.anaconda.com/archive/Anaconda3-2018.12-MacOSX-x86_64.pkg
Windows - https://repo.anaconda.com/archive/Anaconda3-2018.12-Windows-x86_64.exe
Linux - https://repo.anaconda.com/archive/Anaconda3-2018.12-Linux-x86_64.sh and power machines - https://repo.anaconda.com/archive/Anaconda3-2018.12-Linux-ppc64le.sh

Anaconda installation guide is at - https://docs.anaconda.com/anaconda/install/

  1. After installing anaconda, you can install the following packages

The link for each package (https://anaconda.org/anaconda/nltk) contains instructions on how to install it. For example -

conda install -c anaconda nltk
  1. Download the spacy English model -
python -m spacy download en

and stopwords from nltk -

python -m nltk.downloader stopwords

Also download the IMDB sentiment dataset, and pre-trained Glove word embeddings on torchtext with the following python code -

import torch
from torchtext import data
from torchtext import datasets

TEXT = data.Field(tokenize='spacy')
LABEL = data.LabelField(dtype=torch.float)

train_data, test_data = datasets.IMDB.splits(TEXT, LABEL)
TEXT.build_vocab(train_data, max_size=25000, vectors="glove.6B.100d")
LABEL.build_vocab(train_data)

This is 84 MB and 862 MB respectively.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published