Another module down - you're almost half way there!
All that remains in Phase 2 is to put your newfound data science skills to use with a large project!
In this project description, we will cover:
- Project Overview: the project goal, audience, and dataset
- Deliverables: the specific items you are required to produce for this project
- Grading: how your project will be scored
- Getting Started: guidance for how to begin working
For this project, you will use multiple linear regression modeling to analyze house sales in a northwestern county.
It is up to you to define a stakeholder and business problem appropriate to this dataset.
If you are struggling to define a stakeholder, we recommend you complete a project for a real estate agency that helps homeowners buy and/or sell homes. A business problem you could focus on for this stakeholder is the need to provide advice to homeowners about how home renovations might increase the estimated value of their homes, and by what amount.
This project uses the King County House Sales dataset, which can be found in kc_house_data.csv
in the data folder in this assignment's GitHub repository. The description of the column names can be found in column_names.md
in the same folder. As with most real world data sets, the column names are not perfectly described, so you'll have to do some research or use your best judgment if you have questions about what the data means.
It is up to you to decide what data from this dataset to use and how to use it. If you are feeling overwhelmed or behind, we recommend you ignore some or all of the following features:
date
view
sqft_above
sqft_basement
yr_renovated
address
lat
long
-
Your goal in regression modeling is to yield findings to support relevant recommendations. Those findings should include a metric describing overall model performance as well as at least two regression model coefficients. As you explore the data and refine your stakeholder and business problem definitions, make sure you are also thinking about how a linear regression model adds value to your analysis. "The assignment was to use linear regression" is not an acceptable answer! You can also use additional statistical techniques other than linear regression, so long as you clearly explain why you are using each technique.
-
You should demonstrate an iterative approach to modeling. This means that you must build multiple models. Begin with a basic model, evaluate it, and then provide justification for and proceed to a new model. After you finish refining your models, you should provide 1-3 paragraphs in the notebook discussing your final model.
-
Data visualization and analysis are still very important. In Phase 1, your project stopped earlier in the CRISP-DM process. Now you are going a step further, to modeling. Data visualization and analysis will help you build better models and tell a better story to your stakeholders.
There are three deliverables for this project:
- A non-technical presentation
- A Jupyter Notebook
- A GitHub repository
The deliverables requirements are almost the same as in the Phase 1 Project, and you can review those extended descriptions here.
Recall that the non-technical presentation is a slide deck presenting your analysis to business stakeholders, and should be presented live as well as submitted in PDF form on Canvas.
We recommend that you follow this structure, although the slide titles should be specific to your project:
- Beginning
- Overview
- Business and Data Understanding
- Middle
- Modeling
- Regression Results
- End
- Recommendations
- Next Steps
- Thank you
Make sure that your discussion of modeling and regression results is geared towards a non-technical audience! Assume that their prior knowledge of regression modeling is minimal. You don't need to explain how linear regression works, but you should explain why linear regression is useful for the problem context. Make sure you translate any metrics or coefficients into their plain language implications.
The graded elements for the non-technical presentation are the same as in Phase 1.
Recall that the Jupyter Notebook is a notebook that uses Python and Markdown to present your analysis to a data science audience. You will submit the notebook in PDF format on Canvas as well as in .ipynb
format in your GitHub repository.
The graded elements for the Jupyter Notebook are:
- Business Understanding
- Data Understanding
- Data Preparation
- Modeling
- Regression Results
- Code Quality
Recall that the GitHub repository is the cloud-hosted directory containing all of your project files as well as their version history.
The requirements are the same as in Phase 1, except for the required sections in the README.md
.
For this project, the README.md
file should contain:
- Overview
- Business and Data Understanding
- Explain your stakeholder audience here
- Modeling
- Regression Results
- Conclusion
Just like in Phase 1, the README.md
file should be the bridge between your non technical presentation and the Jupyter Notebook. It should not contain the code used to develop your analysis, but should provide a more in-depth explanation of your methodology and analysis than what is described in your presentation slides.
To pass this project, you must pass each project rubric objective. The project rubric objectives for Phase 2 are:
- Attention to Detail
- Statistical Communication
- Data Preparation Fundamentals
- Linear Modeling
Just like in Phase 1, this rubric objective is based on your completion of checklist items. In Phase 2, you need to complete 70% (7 out of 10) or more of the checklist elements in order to pass the Attention to Detail objective.
NOTE THAT THE PASSING BAR IS HIGHER IN PHASE 2 THAN IT WAS IN PHASE 1!
The standard will increase with each Phase, until you will be required to complete all elements to pass Phase 5 (Capstone).
80% or more of the project checklist items are complete
70% of the project checklist items are complete
60% of the project checklist items are complete
50% or fewer of the project checklist items are complete
Recall that communication is one of the key data science "soft skills". In Phase 2, we are specifically focused on Statistical Communication. We define Statistical Communication as:
Communicating results of statistical analyses to diverse audiences via writing and live presentation
Note that this is the same as in Phase 1, except we are replacing "basic data analysis" with "statistical analyses".
High-quality Statistical Communication includes rationale, results, limitations, and recommendations:
- Rationale: Explaining why you are using statistical analyses rather than basic data analysis
- For example, why are you using regression coefficients rather than just a graph?
- What about the problem or data is suitable for this form of analysis?
- For a data science audience, this includes your reasoning for the changes you applied while iterating between models.
- Results: Describing the overall model metrics and feature coefficients
- You need at least one overall model metric (e.g. r-squared or RMSE) and at least two feature coefficients.
- For a business audience, make sure you connect any metrics to real-world implications. You do not need to get into the details of how linear regression works.
- For a data science audience, you don't need to explain what a metric is, but make sure you explain why you chose that particular one.
- Limitations: Identifying the limitations and/or uncertainty present in your analysis
- This could include p-values/alpha values, confidence intervals, assumptions of linear regression, missing data, etc.
- In general, this should be more in-depth for a data science audience and more surface-level for a business audience.
- Recommendations: Interpreting the model results and limitations in the context of the business problem
- What should stakeholders do with this information?
Communicates the rationale, results, limitations, and specific recommendations of statistical analyses
See above for extended explanations of these terms.
Successfully communicates the results of statistical analyses without any major errors
The minimum requirement is to communicate the results, meaning at least one overall model metric (e.g. r-squared or RMSE) as well as at least two feature coefficients. See the Approaching Objective section for an explanation of what a "major error" means.
Communicates the results of statistical analyses with at least one major error
A major error means that some aspect of your explanation is fundamentally incorrect. For example, if a feature coefficient is negative and you say that an increase in that feature results in an increase of the target, that would be a major error. Another example would be if you say that the feature with the highest coefficient is the "most statistically significant" while ignoring the p-value. One more example would be reporting a coefficient that is not statistically significant, rather than saying "no statistically significant linear relationship was found"
"If a coefficient's t-statistic is not significant, don't interpret it at all. You can't be sure that the value of the corresponding parameter in the underlying regression model isn't really zero." DeVeaux, Velleman, and Bock (2012), Stats: Data and Models, 3rd edition, pg. 801. Check out this website for extensive additional examples of mistakes using statistics.
The easiest way to avoid making a major error is to have someone double-check your work. Reach out to peers on Slack and ask them to confirm whether your interpretation makes sense!
Does not communicate the results of statistical analyses
It is not sufficient to just display the entire results summary. You need to pull out at least one overall model metric (e.g. r-squared, RMSE) and at least two feature coefficients, and explain what those numbers mean.
We define this objective as:
Applying appropriate preprocessing and feature engineering steps to tabular data in preparation for statistical modeling
The two most important components of preprocessing for the Phase 2 project are:
- Handling Missing Values: Missing values may be present in the features you want to use, either encoded as
NaN
or as some other value such as"?"
. Before you can build a linear regression model, make sure you identify and address any missing values using techniques such as dropping or replacing data. - Handling Non-Numeric Data: A linear regression model needs all of the features to be numeric, not categorical. For this project, be sure to pick at least one non-numeric feature and try including it in a model. You can identify that a feature is currently non-numeric if the type is
object
when you run.info()
on your dataframe. Once you have identified the non-numeric features, address them using techniques such as ordinal or one-hot (dummy) encoding.
There is no single correct way to handle either of these situations! Use your best judgement to decide what to do, and be sure to explain your rationale in the Markdown of your notebook.
Feature engineering is encouraged but not required for this project.
Goes above and beyond with data preparation, such as feature engineering or merging in outside datasets
One example of feature engineering could be using the
date
feature to create a new feature calledseason
, which represents whether the home was sold in Spring, Summer, Fall, or Winter.
One example of merging in outside datasets could be finding data based on ZIP Code, such as household income or walkability, and joining that data with the provided CSV.
Successfully prepares data for modeling, including converting at least one non-numeric feature into ordinal or binary data and handling missing data as needed
As a reminder, you can identify the non-numeric features by calling
.info()
on the dataframe and looking for typeobject
.
Your final model does not necessarily need to include any features that were originally non-numeric, but you need to demonstrate your ability to handle this type of data.
Prepares some data successfully, but is unable to utilize non-numeric data
If you simply subset the dataframe to only columns with type
int64
orfloat64
, your model will run, but you will not pass this objective.
Does not prepare data for modeling
According to Kaggle's 2020 State of Data Science and Machine Learning Survey, linear and logistic regression are the most popular machine learning algorithms, used by 83.7% of data scientists. They are small, fast models compared to some of the models you will learn later, but have limitations in the kinds of relationships they are able to learn.
In this project you are required to use linear regression as the primary statistical analysis, although you are free to use additional statistical techniques as appropriate.
Goes above and beyond in the modeling process, such as recursive feature selection
Successfully builds a baseline model as well as at least one iterated model, and correctly extracts insights from a final model without any major errors
We are looking for you to (1) create a baseline model, (2) iterate on that model, making adjustments that are supported by regression theory or by descriptive analysis of the data, and (3) select a final model and report on its metrics and coefficients
Ideally you would include written justifications for each model iteration, but at minimum the iterations must be justifiable
For an explanation of "major errors", see the description below
Builds multiple models with at least one major error
The number one major error to avoid is including the target as one of your features. For example, if the target is
price
you should NOT make a "price per square foot" feature, because that feature would not be available if you didn't already know the price.
Other examples of major errors include: using a target other than
price
, attempting only simple linear regression (not multiple linear regression), dropping multiple one-hot encoded columns without explaining the resulting baseline, or using a unique identifier (id
in this dataset) as a feature.
Does not build multiple linear regression models
Please start by reviewing the contents of this project description. If you have any questions, please ask your instructor ASAP.
Next, you will need to complete the Project Proposal which must be reviewed by your instructor before you can continue with the project.
Here are some suggestions for creating your GitHub repository:
- Fork the Phase 2 Project Repository, clone it locally, and work in the
student.ipynb
file. Make sure to also add and commit a PDF of your presentation to your repository with a file name ofpresentation.pdf
. - Or, create a new repository from scratch by going to github.com/new and copying the data files from the Phase 2 Project Repository into your new repository.
- Recall that you can refer to the Phase 1 Project Template as an example structure
- This option will result in the most professional-looking portfolio repository, but can be more complicated to use. So if you are getting stuck with this option, try forking the project repository instead
This is your first modeling project! Take what you have learned in Phase 2 to create a project with a more sophisticated analysis than you completed in Phase 1. You will build on these skills as we move into the predictive machine learning mindset in Phase 3. You've got this!