Week 1 |
1 |
Artificial Intelligence, Introduction to Machine Learning, Its implications with real-world examples, Concept of self-learning |
|
2 |
Types of learning{ Supervised learning (Classification, Regression), Unsupervised learning, Reinforcement learning) |
Week 2 |
3 |
Statistical Analysis of data (Descriptive & Predictive), Data Collection(Numerical data, Categorical data) |
|
4 |
Data Preprocessing and its techniques (Data cleaning, Data integration, Data Argumentation, Data reduction, Data transformation) |
Week 3 |
5 |
Implementation of data preprocessing using python, Introduction to python libraries |
|
6 |
Feature extraction intuition--PCA |
Week 4 |
7 |
Regression problem Linear Regression (linear equation, slop of the line, relationship between attributes, intercept, ordinary least square, residual error) |
|
8 |
Implementation of simple linear regression Evaluate the relation between attributes using plotting. |
Week 5 |
9 |
Multiple linear regression (Dummy variable, multicollinearity, dummy variable trap, building a model using backward elimination) |
|
10 |
Implementation of multiple linear regression using stat library in python. |
Week 6 |
11 |
Polynomial Regression and implementation(Degree of polynomial) |
|
12 |
Logistic Regression intuition and real-world example |
Week 7 |
13 |
Supervised Algorithm-Naïve Bayes (Conditional probability, Bayes theorem) Derive theorem mathematically |
|
14 |
Decision Tree (Entropy, information gain) Mathematical implementation using dataset on python |
Week 8 |
1 hours |
Mid Term |
Week 9 |
15 |
Support vector machine (linearly separable data, non-linearly separable data) Linear SVM implementation using example |
|
16 |
Kernel function intuition (RBF kernel function, Sigmoid function) Implementation of SVM |
Week 10 |
17 |
K-nearest neighbor intuition and solved example using real-world dataset. |
|
18 |
Implementation of K-NN |
Week 11 |
19 |
MID EXAMINATIONS |
|
20 |
MID EXAMINATIONS |
Week 12 |
21 |
Unsupervised learning, Clustering (k-mean clustering intuition- k mean++, the elbow method implementation using example) |
|
22 |
Implementation of K-mean EM Algorithm/DBSCAN intuition |
Week 13 |
23 |
Ensemble learning (Bagging, Boosting) Bagging( Bootstrap aggregation, row sampling with replacement), Random forest |
|
24 |
Bias-Variance trade off, Boosting (Adaboost) |
Week 14 |
25 |
Stacking Optimization algorithm (Gradient Descent, Stochastic Gradient Descent ) |
|
26 |
Evaluation Matrices, Reinforcement learning ( Tuning model complexity) |
Week 15 |
27 |
Natural language processing (Bag of words, stemming, lemmatization) |
|
28 |
Implementation of NLP using dataset using python |
Week 16 |
29 |
Neural network (how brain works, perceptron, the neuron) |
|
30 |
Activation function (SDG) Back propagation intuition and numerical |
Week 17 |
2 hours |
Final Term |