Skip to content

Kaggle competition: Test your analytics skills by predicting which New York Times blog articles will be the most popular

Notifications You must be signed in to change notification settings

datapolymath/PredictingNYT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Course

The Analytics Edge
MITx, edX

Description

Many blog articles are published each day, and the New York Times has to decide which articles should be featured. In this competition, we challenge you to develop an analytics model that will help the New York Times understand the features of a blog post that make it popular.

Data

The data provided for this competition is split into two files:

  • NYTimesBlogTrain.csv = the training data set. It consists of 6532 articles.
  • NYTimesBlogTest.csv = the testing data set. It consists of 1870 articles.

Variable Descriptions

The dependent variable in this problem is the variable Popular, which labels if an article had 25 or more comments in its online comment section (equal to 1 if it did, and 0 if it did not). The dependent variable is provided in the training data set, but not the testing dataset. The independent variables consist of 8 pieces of article data available at the time of publication, and a unique identifier:

  • NewsDesk = the New York Times desk that produced the story (Business, Culture, Foreign, etc.)
  • SectionName = the section the article appeared in (Opinion, Arts, Technology, etc.)
  • SubsectionName = the subsection the article appeared in (Education, Small Business, Room for Debate, etc.)
  • Headline = the title of the article
  • Snippet = a small portion of the article text
  • Abstract = a summary of the blog article, written by the New York Times
  • WordCount = the number of words in the article
  • PubDate = the publication date, in the format "Year-Month-Day Hour:Minute:Second"
  • UniqueID = a unique identifier for each article

Machine Learning Method

I found that Random Forest technique with an n-gram text classification produced the best results in terms of the predictive accuracy.

Results

Preliminary results: 144th position out of 2923 competitors (top 5%)
Final results: 184th position out of 2923 competitors (top 6.5%)

About

Kaggle competition: Test your analytics skills by predicting which New York Times blog articles will be the most popular

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages