Skip to content
This repository has been archived by the owner on Feb 24, 2022. It is now read-only.

Fixes #159, #178, #184 and #185 #179

Closed
wants to merge 25 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
25 commits
Select commit Hold shift + click to select a range
6698d47
Fixed typos
pedropb Dec 27, 2016
4edd7f7
Merge branch 'master' of https://github.com/udacity/machine-learning
pedropb Jan 9, 2017
6210782
Fixing typo on question 5 "evironment"
pedropb Jan 9, 2017
a08e782
Fixed typo in Question 5: "evironment"
pedropb Jan 10, 2017
b717946
Fixed typo in comments "award" -> "reward"
pedropb Jan 10, 2017
b974174
Fixing typo on Question 3:
pedropb Jan 10, 2017
ac14925
Fixed typo
pedropb Jan 11, 2017
757e2b7
Fixed typo "propogating" -> "propagating"
pedropb Jan 15, 2017
b818504
Fixed #159
pedropb Jan 15, 2017
f851b61
Fixed #185
pedropb Jan 15, 2017
76906ba
Fixed #185
pedropb Jan 16, 2017
085fae7
Added Implementation #1 Structure
pedropb Mar 14, 2017
68a5c42
Added Extract Data section.
pedropb Mar 14, 2017
0454d59
Removed reference to idx2numpy.
pedropb Mar 14, 2017
3d5b74e
Finished Extract Data
pedropb Mar 14, 2017
f693228
Improving on Explore Data (partial)
pedropb Mar 16, 2017
bdafca1
Finished exploring and preprocessing data
pedropb Mar 19, 2017
c3b5436
Finished Preprocessing data and Batch Generator
pedropb Mar 22, 2017
d0a6672
Fixed batch generator
pedropb Mar 22, 2017
e226a57
Added one hot encoding
pedropb Mar 24, 2017
b5348e6
Started model architecture.
pedropb Mar 24, 2017
e77a9b5
Bump
pedropb Mar 24, 2017
c0f85cc
Merge branch 'master' of https://github.com/udacity/machine-learning
pedropb Mar 24, 2017
de14b44
Finished preprocessing and adding gitignore
pedropb Mar 24, 2017
36c509b
Trial and error to achieve 50% acc
pedropb Mar 28, 2017
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
9 changes: 5 additions & 4 deletions projects/customer_segments/customer_segments.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,7 @@
"metadata": {},
"source": [
"### Question 2\n",
"*Which feature did you attempt to predict? What was the reported prediction score? Is this feature is necessary for identifying customers' spending habits?* \n",
"*Which feature did you attempt to predict? What was the reported prediction score? Is this feature necessary for identifying customers' spending habits?* \n",
"**Hint:** The coefficient of determination, `R^2`, is scored between 0 and 1, with 1 being a perfect fit. A negative `R^2` implies the model fails to fit the data."
]
},
Expand Down Expand Up @@ -387,7 +387,7 @@
"source": [
"### Question 5\n",
"*How much variance in the data is explained* ***in total*** *by the first and second principal component? What about the first four principal components? Using the visualization provided above, discuss what the first four dimensions best represent in terms of customer spending.* \n",
"**Hint:** A positive increase in a specific dimension corresponds with an *increase* of the *positive-weighted* features and a *decrease* of the *negative-weighted* features. The rate of increase or decrease is based on the indivdual feature weights."
"**Hint:** A positive increase in a specific dimension corresponds with an *increase* of the *positive-weighted* features and a *decrease* of the *negative-weighted* features. The rate of increase or decrease is based on the individual feature weights."
]
},
{
Expand Down Expand Up @@ -422,7 +422,7 @@
"metadata": {},
"source": [
"### Implementation: Dimensionality Reduction\n",
"When using principal component analysis, one of the main goals is to reduce the dimensionality of the data — in effect, reducing the complexity of the problem. Dimensionality reduction comes at a cost: Fewer dimensions used implies less of the total variance in the data is being explained. Because of this, the *cumulative explained variance ratio* is extremely important for knowing how many dimensions are necessary for the problem. Additionally, if a signifiant amount of variance is explained by only two or three dimensions, the reduced data can be visualized afterwards.\n",
"When using principal component analysis, one of the main goals is to reduce the dimensionality of the data — in effect, reducing the complexity of the problem. Dimensionality reduction comes at a cost: Fewer dimensions used implies less of the total variance in the data is being explained. Because of this, the *cumulative explained variance ratio* is extremely important for knowing how many dimensions are necessary for the problem. Additionally, if a significant amount of variance is explained by only two or three dimensions, the reduced data can be visualized afterwards.\n",
"\n",
"In the code block below, you will need to implement the following:\n",
" - Assign the results of fitting PCA in two dimensions with `good_data` to `pca`.\n",
Expand Down Expand Up @@ -786,8 +786,9 @@
}
],
"metadata": {
"anaconda-cloud": {},
"kernelspec": {
"display_name": "Python 2",
"display_name": "Python [default]",
"language": "python",
"name": "python2"
},
Expand Down
Loading