I am using a open source platform to showcase my work and motivate others to work along with me.
XGBoost Link1 Link2 How to use XGBoost with Python
Pytorch Udacity Introduction Lesson 2 (Videos 1 to 25 complete)
Study about Accuracy Paradox
Classification Algorithms study: KNN, SVC, K-medoid
Studied About Apriori Algorithm Association rule, How is it different from collaborative filtering? Studied about example of market basket analysis.
Apriori Introduction Apriori vs Collaborative filtering
Studied about Spectral Clustering
Studying Natural Language Processing: CFG, CNF, CYK CYK tells whether a given sentence can be generated from a given Content free grammer given.
Chomsky Normal Form is conversion from CFG to CNF. In CNF we have productions of form. A-> BC or A->epsilon
Cluster Algorithm KMeans, Heirarchical Clustering
Google Crash Course ML
Worked on Neural Style transfer Project and Watched PyTorch Udacity (Lecture 2) Working on Ben10 dataset
PyTorch Udacity Lecture 2 continue
Pytorch Udacity Lecture 2 continue
Evaluation metric for Classification
-
Jaccard Index: JI = |Intersection| / |Union|
- JI close to 1 means more similarity
- JI close to 0 means less similarity
-
F1-Score = 2* Precision * Recall/ (Precision + Recall)
- 1 is Best and 0 is Worst
-
Log Loss: Output of Class Label is Probability instead of categorical.
- Log Loss measures the Performance of a classifier where the predicted output is a probability value between 0 or 1.
- Log Loss calculated by Log loss equation.
- Log Loss = (y * log(y predicted)) + (1-y) * (log(1 - (y predicted)))
- Average Log Loss = -1/n * summation((y * log(y predicted)) + (1-y) * (log(1 - (y predicted))))
- Lower Log Loss means Best Model and Higher Log Loss value means Poor Model.
Working on Car Dataset Question.
- Shuffle rows of Dataset
np.random.shuffle(DataFrame.values)
- Concat two dataframes
- df1
- df2
- frames = [df1,df2]
- result = pd.concat(frames,axis=1)
- Rename Columns in Pandas
df.rename(columns={'A':'a'},inplace=True)
- Worked on ZigWheels dataset
- How qcut works?
pd.qcut(dataset, precision=3, labels=['low','med','high'])
- mean_squared_log_error
- average_precision_score works on y_true binary and y_scores continous
Udacity PyTorch Lecture 2, neural network finished.
Udacity Talk on PyTorch Lecture 3 finished.
- Single Layer Neural network
- features = torc.rand((1,5)) # createda (1,5) shape tensor
- Method 1 :y = activation(torch.sum(features*weights)+bias)
- Method 2: weights = weights.view(5,1) # used to reshape a tensor vector
- y = activation(torch.mm(features,weights)+bias)
Lecture 4 Started
Lecture 4 Continue
Lecture 4 Almost finished
Learned how to Save Weights of a Trained Model.
Finished Lecture 4 and Started lecture 5 CNN chapter start watched videos till Video 14.
PyTorch Project on Google Colab Started.
R programming Decision Tree, PCA, NaiveBayes, Linear Regression
Artcile on Feature Selection
Difference B/W Covariance and Correlation
Why is feature selection Important?
- Training time increases exponentially with number if features.
- Models have increasing risk of overfitting with increasing number of features.
Feature selection Techniques
- Filter methods
- Wrapper methods
- Embedded methods
Filter method considers the relationship b/w features and the target variable to compute the importance of features.
Wrapper Methods generate models with a subsets of feature and gauge their model performances.
Feature selection by insights provided by some Machine Learning Model.
Loss Functions:- Loss Functions
- 3 Methods: Elbow Method, Average Silhouette method, Gap statistic method
- Resource Link2
- Gap statistic
- Overfitting causes: Too many features, High epochs training with validation loss, many hidden layers, Good Performance in training set and poor generalization in test set.
- What is it?
- Conceptual Explaination
- Deep learning and overfitting
- Mainly used for matrix decomposition
- Mostly used in Recommendation systems.
- SVD link1
- SVD link2
- SVD MIT tutorial
- BEST explaination SVD
- Working and calculation Video
- Learning about Ranking Problems with MCDA or MCDM.
- Understanding how MCDM works with Research Paper LINK
- Studied WSM, WPM, AHP, ELECTRE, TOPSIS and MOORA
- Basics of MCDA Youtube Video