Python implementations of some of the fundamental Machine Learning models and algorithms from scratch.
The purpose of this project is not to produce as optimized and computationally efficient algorithms as possible but rather to present the inner workings of them in a transparent way. The reason the project uses scikit-learn is to evaluate the implementations on sklearn.datasets.
Feel free to reach out if you can think of ways to expand this project.
$ python setup.py install
$ python mlfromscratch/supervised_learning/regression.py
Figure: Polynomial ridge regression of temperature data measured in
Linköping, Sweden 2016.
$ python mlfromscratch/supervised_learning/neural_network.py
Figure: Classification of the digit dataset using CNN.
$ python mlfromscratch/unsupervised_learning/dbscan.py
Figure: Clustering of the moons dataset using DBSCAN.
$ python mlfromscratch/unsupervised_learning/apriori.py
+-------------+
| Apriori |
+-------------+
Minimum Support: 0.25
Minimum Confidence: 0.8
Transactions:
[1, 2, 3, 4]
[1, 2, 4]
[1, 2]
[2, 3, 4]
[2, 3]
[3, 4]
[2, 4]
Frequent Itemsets:
[1, 2, 3, 4, [1, 2], [1, 4], [2, 3], [2, 4], [3, 4], [1, 2, 4], [2, 3, 4]]
Rules:
1 -> 2 (support: 0.43, confidence: 1.0)
4 -> 2 (support: 0.57, confidence: 0.8)
[1, 4] -> 2 (support: 0.29, confidence: 1.0)
- Adaboost
- Bayesian Regression
- Decision Tree
- Deep Learning
- Layers
- Activation Layer
- Average Pooling Layer
- Constant Padding Layer
- Convolutional Layer
- Dropout Layer
- Flatten Layer
- Fully-Connected (Dense) Layer
- Max Pooling Layer
- Zero Padding Layer
- Model Types
- Convolutional Neural Network
- Multilayer Perceptron
- Layers
- Gradient Boosting
- K Nearest Neighbors
- Linear Discriminant Analysis
- Linear Regression
- Logistic Regression
- Multi-class Linear Discriminant Analysis
- Naive Bayes
- Perceptron
- Polynomial Regression
- Random Forest
- Ridge Regression
- Support Vector Machine
- XGBoost