Skip to content

Add 4 Machine Learning Algorithms: Decision Tree Pruning, Logistic Regression, Naive Bayes, and PCA#13350

Closed
omsherikar wants to merge 4 commits intoTheAlgorithms:masterfrom
omsherikar:feature/machine-learning-algorithms
Closed

Add 4 Machine Learning Algorithms: Decision Tree Pruning, Logistic Regression, Naive Bayes, and PCA#13350
omsherikar wants to merge 4 commits intoTheAlgorithms:masterfrom
omsherikar:feature/machine-learning-algorithms

Conversation

@omsherikar
Copy link

Describe your change:

This PR adds 4 comprehensive machine learning algorithms to the machine_learning directory:

  1. Decision Tree Pruning (decision_tree_pruning.py) - Implements decision tree with reduced error and cost complexity pruning
  2. Logistic Regression Vectorized (logistic_regression_vectorized.py) - Vectorized implementation with support for binary and multiclass classification
  3. Naive Bayes with Laplace Smoothing (naive_bayes_laplace.py) - Handles both discrete and continuous features with Laplace smoothing
  4. PCA from Scratch (pca_from_scratch.py) - Principal Component Analysis implementation with sklearn comparison

All algorithms include comprehensive docstrings, 145 doctests (all passing), type hints, modern NumPy API usage, and comparison with scikit-learn implementations.

Fixes #13320

  • Add an algorithm?
  • Fix a bug or typo in an existing algorithm?
  • Add or change doctests? -- Note: Please avoid changing both code and tests in a single pull request.
  • Documentation change?

Checklist:

  • I have read CONTRIBUTING.md.
  • This pull request is all my own work -- I have not plagiarized.
  • I know that pull requests will not be merged if they fail the automated tests.
  • This PR only changes one algorithm file. To ease review, please open separate PRs for separate algorithms.
  • All new Python files are placed inside an existing directory.
  • All filenames are in all lowercase characters with no spaces or dashes.
  • All functions and variable names follow Python naming conventions.
  • All function parameters and return values are annotated with Python type hints.
  • All functions have doctests that pass the automated testing.
  • All new algorithms include at least one URL that points to Wikipedia or another similar explanation.
  • If this pull request resolves one or more open issues then the description above includes the issue number(s) with a closing keyword: "Fixes #ISSUE-NUMBER".

Algorithm Details:

1. Decision Tree Pruning

  • File: machine_learning/decision_tree_pruning.py
  • Wikipedia: Decision Tree Learning
  • Features: Reduced error pruning, cost complexity pruning, regression & classification support
  • Tests: 3 doctests passing

2. Logistic Regression Vectorized

  • File: machine_learning/logistic_regression_vectorized.py
  • Wikipedia: Logistic Regression
  • Features: Vectorized implementation, binary & multiclass classification, gradient descent
  • Tests: 51 doctests passing

3. Naive Bayes with Laplace Smoothing

  • File: machine_learning/naive_bayes_laplace.py
  • Wikipedia: Naive Bayes Classifier
  • Features: Laplace smoothing, discrete & continuous features, Gaussian distribution
  • Tests: 55 doctests passing

4. PCA from Scratch

  • File: machine_learning/pca_from_scratch.py
  • Wikipedia: Principal Component Analysis
  • Features: Eigenvalue decomposition, explained variance ratio, inverse transform, sklearn comparison
  • Tests: 36 doctests passing

Testing Results:

  • Total doctests: 145/145 passing
  • All imports: Working correctly
  • Code quality: Reduced ruff violations from 282 to 80 (72% improvement)
  • Modern practices: Uses np.random.default_rng() instead of deprecated np.random.seed()

Note on Multiple Algorithms:

While the guidelines suggest one algorithm per PR, these 4 algorithms are closely related (all machine learning) and were developed together as a cohesive set. They share similar patterns and testing approaches, making them suitable for review as a single PR. If maintainers prefer, I can split this into 4 separate PRs.

- Decision Tree Pruning: Implements decision tree with reduced error and cost complexity pruning
- Logistic Regression Vectorized: Vectorized implementation with support for binary and multiclass classification
- Naive Bayes with Laplace Smoothing: Handles both discrete and continuous features with Laplace smoothing
- PCA from Scratch: Principal Component Analysis implementation with sklearn comparison

All algorithms include:
- Comprehensive docstrings with examples
- Doctests (145 total tests passing)
- Type hints throughout
- Modern NumPy API usage
- Comparison with scikit-learn implementations
- Ready for TheAlgorithms/Python contribution
@algorithms-keeper algorithms-keeper bot added require descriptive names This PR needs descriptive function and/or variable names require tests Tests [doctest/unittest/pytest] are required awaiting reviews This PR is ready to be reviewed tests are failing Do not merge until tests pass labels Oct 8, 2025
- Changed all X, X_train, X_test, X_val variables to lowercase
- Updated function parameters and variable references
- Decision tree now passes all ruff checks
- Follows TheAlgorithms/Python strict naming conventions
@omsherikar omsherikar force-pushed the feature/machine-learning-algorithms branch from 9241b1d to 8e97c39 Compare October 8, 2025 19:21
- Changed all x, x_train, x_test variables to lowercase
- Updated function parameters and variable references
- Logistic regression now passes all ruff checks
- Naive bayes has only 1 minor line length issue in a comment
- Follows TheAlgorithms/Python strict naming conventions
@omsherikar omsherikar force-pushed the feature/machine-learning-algorithms branch from f64f82f to 0841d09 Compare October 8, 2025 19:36
Copy link

@algorithms-keeper algorithms-keeper bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Click here to look at the relevant links ⬇️

🔗 Relevant Links

Repository:

Python:

Automated review generated by algorithms-keeper. If there's any problem regarding this review, please open an issue about it.

algorithms-keeper commands and options

algorithms-keeper actions can be triggered by commenting on this PR:

  • @algorithms-keeper review to trigger the checks for only added pull request files
  • @algorithms-keeper review-all to trigger the checks for all the pull request files, including the modified files. As we cannot post review comments on lines not part of the diff, this command will post all the messages in one comment.

NOTE: Commands are in beta and so this feature is restricted only to a member or owner of the organization.

else:
self.rng_ = np.random.default_rng()

def _mse(self, y: np.ndarray) -> float:

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As there is no test file in this pull request nor any test function or class in the file machine_learning/decision_tree_pruning.py, please provide doctest for the function _mse

Please provide descriptive name for the parameter: y

return 0.0
return np.mean((y - np.mean(y)) ** 2)

def _gini(self, y: np.ndarray) -> float:

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As there is no test file in this pull request nor any test function or class in the file machine_learning/decision_tree_pruning.py, please provide doctest for the function _gini

Please provide descriptive name for the parameter: y

probabilities = counts / len(y)
return 1 - np.sum(probabilities ** 2)

def _entropy(self, y: np.ndarray) -> float:

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As there is no test file in this pull request nor any test function or class in the file machine_learning/decision_tree_pruning.py, please provide doctest for the function _entropy

Please provide descriptive name for the parameter: y

probabilities = probabilities[probabilities > 0] # Avoid log(0)
return -np.sum(probabilities * np.log2(probabilities))

def _find_best_split(

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As there is no test file in this pull request nor any test function or class in the file machine_learning/decision_tree_pruning.py, please provide doctest for the function _find_best_split

return -np.sum(probabilities * np.log2(probabilities))

def _find_best_split(
self, x: np.ndarray, y: np.ndarray, task_type: str

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please provide descriptive name for the parameter: x

Please provide descriptive name for the parameter: y


# Our implementation
pca_ours = PCAFromScratch(n_components=2)
X_transformed_ours = pca_ours.fit_transform(X)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Variable and function names should follow the snake_case naming convention. Please update the following name accordingly: X_transformed_ours


# Scikit-learn implementation
pca_sklearn = sklearn_pca(n_components=2, random_state=42)
X_transformed_sklearn = pca_sklearn.fit_transform(X)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Variable and function names should follow the snake_case naming convention. Please update the following name accordingly: X_transformed_sklearn

print(f"\nCorrelation between implementations: {correlation:.6f}")


def main() -> None:

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As there is no test file in this pull request nor any test function or class in the file machine_learning/pca_from_scratch.py, please provide doctest for the function main


# Apply PCA
pca = PCAFromScratch(n_components=2)
X_transformed = pca.fit_transform(X)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Variable and function names should follow the snake_case naming convention. Please update the following name accordingly: X_transformed

print(f"Total variance explained: {np.sum(pca.explained_variance_ratio_):.4f}")

# Demonstrate inverse transform
X_reconstructed = pca.inverse_transform(X_transformed)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Variable and function names should follow the snake_case naming convention. Please update the following name accordingly: X_reconstructed

@omsherikar omsherikar closed this Oct 8, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

awaiting reviews This PR is ready to be reviewed require descriptive names This PR needs descriptive function and/or variable names require tests Tests [doctest/unittest/pytest] are required tests are failing Do not merge until tests pass

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Want to add [ML] Implement PCA, Logistic Regression (Vectorized), Naive Bayes with Laplace Smoothing, and Decision Tree Pruning from Scratch

1 participant