ML from Scratch

Explore my collection of machine learning and AI implementations built from scratch using pure Python and NumPy. These projects demonstrate deep understanding of core algorithms without relying on high-level libraries (Except for visualization and evaluation).

Implementation Statistics

19+
Algorithms Implemented
5
Papers Implemented
100%
From Scratch
Python
& NumPy Only

Machine Learning Models

Linear Regression
Linear Regression using gradient descent with methods for training, prediction, scoring (R²), and cost visualization.
Logistic Regression
Binary logistic regression implemented from scratch using gradient descent with prediction, scoring, and 2D decision boundary plotting.
Multimodal Logistic Regression
Advanced logistic regression handling multiple data modalities with early, late, and intermediate fusion using gradient descent.
K-Nearest Neighbors (KNN)
K-Nearest Neighbors classifier implemented from scratch with multiple distance metrics, cross-validation for k selection, and decision boundary visualization.
K-Means Clustering
K‑Means clustering from scratch with random or k‑means++ initialization plus inertia, silhouette, elbow diagnostics, and rich visualization utilities.
DBSCAN
Density-Based Spatial Clustering of Applications with Noise implemented from scratch with batch processing, visualization, and parameter optimization.
Neural Network
Multi-layer neural network implemented from scratch with customizable architectures, activation functions, regularization, and comprehensive visualization capabilities.
Naive Bayes
Gaussian and Multinomial Naive Bayes classifiers implemented from scratch with visualization capabilities for feature distributions and decision boundaries.
Decision Tree
Decision Tree classifier implemented from scratch with support for Gini/Entropy splitting, feature importances, visualization, and boundary plotting.
Support Vector Machine
Support Vector Machine implemented from scratch using Sequential Minimal Optimization with multiple kernel functions and visualization capabilities.
Random Forest
Random Forest classifier implemented from scratch with bootstrap sampling, feature importance analysis, and visualization capabilities.
Gradient Boosting Regressor
Gradient Boosting Regressor implemented from scratch using decision trees as weak learners and mean initialization.
XGBoost
Extreme gradient boosting implementation from scratch with gradient boosting, regularization, subsampling, early stopping, and feature importance analysis.
AdaBoost
Adaptive boosting algorithm that combines weak learners with adaptive sample weighting.
More implementations coming soon! Including Computer Vision models, Reinforcement Learning algorithms, and Advanced Optimizers.