diff --git a/.coveragerc b/.coveragerc
deleted file mode 100644
index f7e6eb212bc8..000000000000
--- a/.coveragerc
+++ /dev/null
@@ -1,4 +0,0 @@
-[report]
-sort = Cover
-omit =
- .env/*
diff --git a/.devcontainer/Dockerfile b/.devcontainer/Dockerfile
new file mode 100644
index 000000000000..a0bd05f47ec8
--- /dev/null
+++ b/.devcontainer/Dockerfile
@@ -0,0 +1,8 @@
+# https://github.com/microsoft/vscode-dev-containers/blob/main/containers/python-3/README.md
+ARG VARIANT=3.13-bookworm
+FROM mcr.microsoft.com/vscode/devcontainers/python:${VARIANT}
+COPY requirements.txt /tmp/pip-tmp/
+RUN python3 -m pip install --upgrade pip \
+ && python3 -m pip install --no-cache-dir install -r /tmp/pip-tmp/requirements.txt \
+ && pipx install pre-commit ruff \
+ && pre-commit install
diff --git a/.devcontainer/README.md b/.devcontainer/README.md
new file mode 100644
index 000000000000..ec3cdb61de7a
--- /dev/null
+++ b/.devcontainer/README.md
@@ -0,0 +1 @@
+https://code.visualstudio.com/docs/devcontainers/tutorial
diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json
new file mode 100644
index 000000000000..e23263f5b9de
--- /dev/null
+++ b/.devcontainer/devcontainer.json
@@ -0,0 +1,42 @@
+{
+ "name": "Python 3",
+ "build": {
+ "dockerfile": "Dockerfile",
+ "context": "..",
+ "args": {
+ // Update 'VARIANT' to pick a Python version: 3, 3.11, 3.10, 3.9, 3.8
+ // Append -bullseye or -buster to pin to an OS version.
+ // Use -bullseye variants on local on arm64/Apple Silicon.
+ "VARIANT": "3.13-bookworm",
+ }
+ },
+
+ // Configure tool-specific properties.
+ "customizations": {
+ // Configure properties specific to VS Code.
+ "vscode": {
+ // Set *default* container specific settings.json values on container create.
+ "settings": {
+ "python.defaultInterpreterPath": "/usr/local/bin/python",
+ "python.linting.enabled": true,
+ "python.formatting.blackPath": "/usr/local/py-utils/bin/black",
+ "python.linting.mypyPath": "/usr/local/py-utils/bin/mypy"
+ },
+
+ // Add the IDs of extensions you want installed when the container is created.
+ "extensions": [
+ "ms-python.python",
+ "ms-python.vscode-pylance"
+ ]
+ }
+ },
+
+ // Use 'forwardPorts' to make a list of ports inside the container available locally.
+ // "forwardPorts": [],
+
+ // Use 'postCreateCommand' to run commands after the container is created.
+ // "postCreateCommand": "pip3 install --user -r requirements.txt",
+
+ // Comment out to connect as root instead. More info: https://aka.ms/vscode-remote/containers/non-root.
+ "remoteUser": "vscode"
+}
diff --git a/.github/CODEOWNERS b/.github/CODEOWNERS
index 260b9704eda7..3cc25d1bae1c 100644
--- a/.github/CODEOWNERS
+++ b/.github/CODEOWNERS
@@ -7,9 +7,7 @@
# Order is important. The last matching pattern has the most precedence.
-/.* @cclauss @dhruvmanila
-
-# /arithmetic_analysis/
+/.* @cclauss
# /backtracking/
@@ -21,21 +19,21 @@
# /cellular_automata/
-# /ciphers/ @cclauss # TODO: Uncomment this line after Hacktoberfest
+# /ciphers/
# /compression/
# /computer_vision/
-# /conversions/ @cclauss # TODO: Uncomment this line after Hacktoberfest
+# /conversions/
-# /data_structures/ @cclauss # TODO: Uncomment this line after Hacktoberfest
+# /data_structures/
-/digital_image_processing/ @mateuszz0000
+# /digital_image_processing/
# /divide_and_conquer/
-/dynamic_programming/ @Kush1101
+# /dynamic_programming/
# /file_transfer/
@@ -59,7 +57,7 @@
# /machine_learning/
-/maths/ @Kush1101
+# /maths/
# /matrix/
@@ -67,9 +65,9 @@
# /neural_network/
-# /other/ @cclauss # TODO: Uncomment this line after Hacktoberfest
+# /other/
-/project_euler/ @dhruvmanila @Kush1101
+# /project_euler/
# /quantum/
@@ -79,9 +77,9 @@
# /searches/
-/sorts/ @mateuszz0000
+# /sorts/
-# /strings/ @cclauss # TODO: Uncomment this line after Hacktoberfest
+# /strings/
# /traversals/
diff --git a/.github/ISSUE_TEMPLATE/bug_report.yml b/.github/ISSUE_TEMPLATE/bug_report.yml
new file mode 100644
index 000000000000..4ccdb52cad24
--- /dev/null
+++ b/.github/ISSUE_TEMPLATE/bug_report.yml
@@ -0,0 +1,54 @@
+name: Bug report
+description: Create a bug report to help us address errors in the repository
+labels: [bug]
+body:
+ - type: markdown
+ attributes:
+ value: >
+ Before requesting please search [existing issues](https://github.com/TheAlgorithms/Python/labels/bug).
+ Usage questions such as "How do I...?" belong on the
+ [Discord](https://discord.gg/c7MnfGFGa6) and will be closed.
+
+ - type: input
+ attributes:
+ label: "Repository commit"
+ description: >
+ The commit hash for `TheAlgorithms/Python` repository. You can get this
+ by running the command `git rev-parse HEAD` locally.
+ placeholder: "a0b0f414ae134aa1772d33bb930e5a960f9979e8"
+ validations:
+ required: true
+
+ - type: input
+ attributes:
+ label: "Python version (python --version)"
+ placeholder: "Python 3.10.7"
+ validations:
+ required: true
+
+ - type: textarea
+ attributes:
+ label: "Dependencies version (pip freeze)"
+ description: >
+ This is the output of the command `pip freeze --all`. Note that the
+ actual output might be different as compared to the placeholder text.
+ placeholder: |
+ appnope==0.1.3
+ asttokens==2.0.8
+ backcall==0.2.0
+ ...
+ validations:
+ required: true
+
+ - type: textarea
+ attributes:
+ label: "Expected behavior"
+ description: "Describe the behavior you expect. May include images or videos."
+ validations:
+ required: true
+
+ - type: textarea
+ attributes:
+ label: "Actual behavior"
+ validations:
+ required: true
diff --git a/.github/ISSUE_TEMPLATE/config.yml b/.github/ISSUE_TEMPLATE/config.yml
new file mode 100644
index 000000000000..62019bb08938
--- /dev/null
+++ b/.github/ISSUE_TEMPLATE/config.yml
@@ -0,0 +1,5 @@
+blank_issues_enabled: false
+contact_links:
+ - name: Discord community
+ url: https://discord.gg/c7MnfGFGa6
+ about: Have any questions or need any help? Please contact us via Discord
diff --git a/.github/ISSUE_TEMPLATE/feature_request.yml b/.github/ISSUE_TEMPLATE/feature_request.yml
new file mode 100644
index 000000000000..20823bd58ab1
--- /dev/null
+++ b/.github/ISSUE_TEMPLATE/feature_request.yml
@@ -0,0 +1,19 @@
+name: Feature request
+description: Suggest features, propose improvements, discuss new ideas.
+labels: [enhancement]
+body:
+ - type: markdown
+ attributes:
+ value: >
+ Before requesting please search [existing issues](https://github.com/TheAlgorithms/Python/labels/enhancement).
+ Do not create issues to implement new algorithms as these will be closed.
+ Usage questions such as "How do I...?" belong on the
+ [Discord](https://discord.gg/c7MnfGFGa6) and will be closed.
+
+ - type: textarea
+ attributes:
+ label: "Feature description"
+ description: >
+ This could include new topics or improving any existing implementations.
+ validations:
+ required: true
diff --git a/.github/ISSUE_TEMPLATE/other.yml b/.github/ISSUE_TEMPLATE/other.yml
new file mode 100644
index 000000000000..44d6ff541506
--- /dev/null
+++ b/.github/ISSUE_TEMPLATE/other.yml
@@ -0,0 +1,19 @@
+name: Other
+description: Use this for any other issues. PLEASE do not create blank issues
+labels: ["awaiting triage"]
+body:
+ - type: textarea
+ id: issuedescription
+ attributes:
+ label: What would you like to share?
+ description: Provide a clear and concise explanation of your issue.
+ validations:
+ required: true
+
+ - type: textarea
+ id: extrainfo
+ attributes:
+ label: Additional information
+ description: Is there anything else we should know about this issue?
+ validations:
+ required: false
diff --git a/.github/dependabot.yml b/.github/dependabot.yml
new file mode 100644
index 000000000000..15e494ec867e
--- /dev/null
+++ b/.github/dependabot.yml
@@ -0,0 +1,8 @@
+# Keep GitHub Actions up to date with Dependabot...
+# https://docs.github.com/en/code-security/dependabot/working-with-dependabot/keeping-your-actions-up-to-date-with-dependabot
+version: 2
+updates:
+ - package-ecosystem: "github-actions"
+ directory: "/"
+ schedule:
+ interval: "daily"
diff --git a/.github/pull_request_template.md b/.github/pull_request_template.md
index 103ecf7c288a..e2ae0966cda5 100644
--- a/.github/pull_request_template.md
+++ b/.github/pull_request_template.md
@@ -1,12 +1,13 @@
-### **Describe your change:**
+### Describe your change:
* [ ] Add an algorithm?
* [ ] Fix a bug or typo in an existing algorithm?
+* [ ] Add or change doctests? -- Note: Please avoid changing both code and tests in a single pull request.
* [ ] Documentation change?
-### **Checklist:**
+### Checklist:
* [ ] I have read [CONTRIBUTING.md](https://github.com/TheAlgorithms/Python/blob/master/CONTRIBUTING.md).
* [ ] This pull request is all my own work -- I have not plagiarized.
* [ ] I know that pull requests will not be merged if they fail the automated tests.
@@ -16,5 +17,5 @@
* [ ] All functions and variable names follow Python naming conventions.
* [ ] All function parameters and return values are annotated with Python [type hints](https://docs.python.org/3/library/typing.html).
* [ ] All functions have [doctests](https://docs.python.org/3/library/doctest.html) that pass the automated testing.
-* [ ] All new algorithms have a URL in its comments that points to Wikipedia or other similar explanation.
-* [ ] If this pull request resolves one or more open issues then the commit message contains `Fixes: #{$ISSUE_NO}`.
+* [ ] All new algorithms include at least one URL that points to Wikipedia or another similar explanation.
+* [ ] If this pull request resolves one or more open issues then the description above includes the issue number(s) with a [closing keyword](https://docs.github.com/en/issues/tracking-your-work-with-issues/linking-a-pull-request-to-an-issue): "Fixes #ISSUE-NUMBER".
diff --git a/.github/stale.yml b/.github/stale.yml
index ba6fd155d7a3..0939e1f223ff 100644
--- a/.github/stale.yml
+++ b/.github/stale.yml
@@ -27,7 +27,7 @@ exemptAssignees: false
staleLabel: stale
# Limit the number of actions per hour, from 1-30. Default is 30
-limitPerRun: 30
+limitPerRun: 5
# Comment to post when removing the stale label.
# unmarkComment: >
@@ -45,7 +45,7 @@ pulls:
closeComment: >
Please reopen this pull request once you commit the changes requested
or make improvements on the code. If this is not the case and you need
- some help, feel free to seek help from our [Gitter](https://gitter.im/TheAlgorithms)
+ some help, feel free to seek help from our [Gitter](https://gitter.im/TheAlgorithms/community)
or ping one of the reviewers. Thank you for your contributions!
issues:
@@ -59,5 +59,5 @@ issues:
closeComment: >
Please reopen this issue once you add more information and updates here.
If this is not the case and you need some help, feel free to seek help
- from our [Gitter](https://gitter.im/TheAlgorithms) or ping one of the
+ from our [Gitter](https://gitter.im/TheAlgorithms/community) or ping one of the
reviewers. Thank you for your contributions!
diff --git a/.github/workflows/auto_close_empty_issues.yml b/.github/workflows/auto_close_empty_issues.yml
deleted file mode 100644
index a6334d6ade32..000000000000
--- a/.github/workflows/auto_close_empty_issues.yml
+++ /dev/null
@@ -1,20 +0,0 @@
-# GitHub Action that uses close-issue auto-close empty issues after they are opened.
-# If the issue body text is empty the Action auto-closes it and sends a notification.
-# Otherwise if the issue body is not empty, it does nothing and the issue remains open.
-# https://github.com/marketplace/actions/close-issue
-
-name: auto_close_empty_issues
-on:
- issues:
- types: [opened]
-jobs:
- check-issue-body-not-empty:
- runs-on: ubuntu-latest
- steps:
- - if: github.event.issue.body == 0
- name: Close Issue
- uses: peter-evans/close-issue@v1
- with:
- comment: |
- Issue body must contain content.
- Auto-closing this issue.
diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml
index 01ac9aea7a7c..62829b2b45a5 100644
--- a/.github/workflows/build.yml
+++ b/.github/workflows/build.yml
@@ -9,19 +9,28 @@ jobs:
build:
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v2
- - uses: actions/setup-python@v2
+ - uses: actions/checkout@v4
+ - uses: astral-sh/setup-uv@v5
with:
- python-version: "3.8"
- - uses: actions/cache@v2
+ enable-cache: true
+ cache-dependency-glob: uv.lock
+ - uses: actions/setup-python@v5
with:
- path: ~/.cache/pip
- key: ${{ runner.os }}-pip-${{ hashFiles('requirements.txt') }}
- - name: Install dependencies
- run: |
- python -m pip install --upgrade pip setuptools six
- python -m pip install pytest-cov -r requirements.txt
+ python-version: 3.13
+ allow-prereleases: true
+ - run: uv sync --group=test
- name: Run tests
- run: pytest --doctest-modules --ignore=project_euler/ --cov-report=term-missing:skip-covered --cov=. .
+ # TODO: #8818 Re-enable quantum tests
+ run: uv run pytest
+ --ignore=computer_vision/cnn_classification.py
+ --ignore=docs/conf.py
+ --ignore=dynamic_programming/k_means_clustering_tensorflow.py
+ --ignore=machine_learning/lstm/lstm_prediction.py
+ --ignore=neural_network/input_data.py
+ --ignore=project_euler/
+ --ignore=quantum/q_fourier_transform.py
+ --ignore=scripts/validate_solutions.py
+ --cov-report=term-missing:skip-covered
+ --cov=. .
- if: ${{ success() }}
run: scripts/build_directory_md.py 2>&1 | tee DIRECTORY.md
diff --git a/.github/workflows/directory_writer.yml b/.github/workflows/directory_writer.yml
index be8154a32696..55d89f455a25 100644
--- a/.github/workflows/directory_writer.yml
+++ b/.github/workflows/directory_writer.yml
@@ -6,13 +6,17 @@ jobs:
build:
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v1 # v1, NOT v2
- - uses: actions/setup-python@v2
+ - uses: actions/checkout@v4
+ with:
+ fetch-depth: 0
+ - uses: actions/setup-python@v5
+ with:
+ python-version: 3.x
- name: Write DIRECTORY.md
run: |
scripts/build_directory_md.py 2>&1 | tee DIRECTORY.md
- git config --global user.name github-actions
- git config --global user.email '${GITHUB_ACTOR}@users.noreply.github.com'
+ git config --global user.name "$GITHUB_ACTOR"
+ git config --global user.email "$GITHUB_ACTOR@users.noreply.github.com"
git remote set-url origin https://x-access-token:${{ secrets.GITHUB_TOKEN }}@github.com/$GITHUB_REPOSITORY
- name: Update DIRECTORY.md
run: |
diff --git a/.github/workflows/pre-commit.yml b/.github/workflows/pre-commit.yml
deleted file mode 100644
index 96175cfecea5..000000000000
--- a/.github/workflows/pre-commit.yml
+++ /dev/null
@@ -1,21 +0,0 @@
-name: pre-commit
-
-on: [push, pull_request]
-
-jobs:
- pre-commit:
- runs-on: ubuntu-latest
- steps:
- - uses: actions/checkout@v2
- - uses: actions/cache@v2
- with:
- path: |
- ~/.cache/pre-commit
- ~/.cache/pip
- key: ${{ runner.os }}-pre-commit-${{ hashFiles('.pre-commit-config.yaml') }}
- - uses: actions/setup-python@v2
- - name: Install pre-commit
- run: |
- python -m pip install --upgrade pip
- python -m pip install --upgrade pre-commit
- - run: pre-commit run --verbose --all-files --show-diff-on-failure
diff --git a/.github/workflows/project_euler.yml b/.github/workflows/project_euler.yml
index e8b011af20a6..8d51ad8850cf 100644
--- a/.github/workflows/project_euler.yml
+++ b/.github/workflows/project_euler.yml
@@ -1,31 +1,35 @@
on:
pull_request:
- # only check if a file is changed within the project_euler directory and related files
+ # Run only if a file is changed within the project_euler directory and related files
paths:
- - 'project_euler/**'
- - '.github/workflows/project_euler.yml'
- - 'scripts/validate_solutions.py'
+ - "project_euler/**"
+ - ".github/workflows/project_euler.yml"
+ - "scripts/validate_solutions.py"
+ schedule:
+ - cron: "0 0 * * *" # Run everyday
-name: 'Project Euler'
+name: "Project Euler"
jobs:
project-euler:
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v2
- - uses: actions/setup-python@v2
- - name: Install pytest and pytest-cov
- run: |
- python -m pip install --upgrade pip
- python -m pip install --upgrade pytest pytest-cov
- - run: pytest --doctest-modules --cov-report=term-missing:skip-covered --cov=project_euler/ project_euler/
+ - uses: actions/checkout@v4
+ - uses: astral-sh/setup-uv@v5
+ - uses: actions/setup-python@v5
+ with:
+ python-version: 3.x
+ - run: uv sync --group=euler-validate --group=test
+ - run: uv run pytest --doctest-modules --cov-report=term-missing:skip-covered --cov=project_euler/ project_euler/
validate-solutions:
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v2
- - uses: actions/setup-python@v2
- - name: Install pytest
- run: |
- python -m pip install --upgrade pip
- python -m pip install --upgrade pytest
- - run: pytest scripts/validate_solutions.py
+ - uses: actions/checkout@v4
+ - uses: astral-sh/setup-uv@v5
+ - uses: actions/setup-python@v5
+ with:
+ python-version: 3.x
+ - run: uv sync --group=euler-validate --group=test
+ - run: uv run pytest scripts/validate_solutions.py
+ env:
+ GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
diff --git a/.github/workflows/ruff.yml b/.github/workflows/ruff.yml
new file mode 100644
index 000000000000..cfe127b3521f
--- /dev/null
+++ b/.github/workflows/ruff.yml
@@ -0,0 +1,16 @@
+# https://beta.ruff.rs
+name: ruff
+on:
+ push:
+ branches:
+ - master
+ pull_request:
+ branches:
+ - master
+jobs:
+ ruff:
+ runs-on: ubuntu-latest
+ steps:
+ - uses: actions/checkout@v4
+ - uses: astral-sh/setup-uv@v5
+ - run: uvx ruff check --output-format=github .
diff --git a/.github/workflows/sphinx.yml b/.github/workflows/sphinx.yml
new file mode 100644
index 000000000000..16ff284a74f2
--- /dev/null
+++ b/.github/workflows/sphinx.yml
@@ -0,0 +1,50 @@
+name: sphinx
+
+on:
+ # Triggers the workflow on push or pull request events but only for the "master" branch
+ push:
+ branches: ["master"]
+ pull_request:
+ branches: ["master"]
+ # Or manually from the Actions tab
+ workflow_dispatch:
+
+# Sets permissions of the GITHUB_TOKEN to allow deployment to GitHub Pages
+permissions:
+ contents: read
+ pages: write
+ id-token: write
+
+# Allow only one concurrent deployment, skipping runs queued between the run in-progress and latest queued.
+# However, do NOT cancel in-progress runs as we want to allow these production deployments to complete.
+concurrency:
+ group: "pages"
+ cancel-in-progress: false
+
+jobs:
+ build_docs:
+ runs-on: ubuntu-24.04-arm
+ steps:
+ - uses: actions/checkout@v4
+ - uses: astral-sh/setup-uv@v5
+ - uses: actions/setup-python@v5
+ with:
+ python-version: 3.13
+ allow-prereleases: true
+ - run: uv sync --group=docs
+ - uses: actions/configure-pages@v5
+ - run: uv run sphinx-build -c docs . docs/_build/html
+ - uses: actions/upload-pages-artifact@v3
+ with:
+ path: docs/_build/html
+
+ deploy_docs:
+ environment:
+ name: github-pages
+ url: ${{ steps.deployment.outputs.page_url }}
+ if: github.event_name != 'pull_request'
+ needs: build_docs
+ runs-on: ubuntu-latest
+ steps:
+ - uses: actions/deploy-pages@v4
+ id: deployment
diff --git a/.github/workflows/stale.yml b/.github/workflows/stale.yml
deleted file mode 100644
index 42353d233a29..000000000000
--- a/.github/workflows/stale.yml
+++ /dev/null
@@ -1,31 +0,0 @@
-# name: Mark/Close stale issues and pull requests
-# on:
-# schedule:
-# - cron: "0 * * * *" # Run every hour
-# jobs:
-# stale:
-# runs-on: ubuntu-latest
-# steps:
-# - uses: actions/stale@v3.0.13
-# with:
-# repo-token: ${{ secrets.GITHUB_TOKEN }}
-# days-before-stale: 30
-# days-before-close: 7
-# stale-issue-message: >
-# This issue has been automatically marked as stale because it has not had
-# recent activity. It will be closed if no further activity occurs. Thank you
-# for your contributions.
-# close-issue-message: >
-# Please reopen this issue once you add more information and updates here.
-# If this is not the case and you need some help, feel free to seek help
-# from our [Gitter](https://gitter.im/TheAlgorithms) or ping one of the
-# reviewers. Thank you for your contributions!
-# stale-pr-message: >
-# This pull request has been automatically marked as stale because it has not had
-# recent activity. It will be closed if no further activity occurs. Thank you
-# for your contributions.
-# close-pr-message: >
-# Please reopen this pull request once you commit the changes requested
-# or make improvements on the code. If this is not the case and you need
-# some help, feel free to seek help from our [Gitter](https://gitter.im/TheAlgorithms)
-# or ping one of the reviewers. Thank you for your contributions!
diff --git a/.gitignore b/.gitignore
index 574cdf312836..baea84b8d1f1 100644
--- a/.gitignore
+++ b/.gitignore
@@ -107,3 +107,4 @@ venv.bak/
.idea
.try
.vscode/
+.vs/
diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
index a3288e1c5eef..0fc8b2b14e07 100644
--- a/.pre-commit-config.yaml
+++ b/.pre-commit-config.yaml
@@ -1,56 +1,38 @@
repos:
- repo: https://github.com/pre-commit/pre-commit-hooks
- rev: v3.2.0
+ rev: v5.0.0
hooks:
- id: check-executables-have-shebangs
+ - id: check-toml
- id: check-yaml
- id: end-of-file-fixer
types: [python]
- id: trailing-whitespace
- exclude: |
- (?x)^(
- data_structures/heap/binomial_heap.py
- )$
- id: requirements-txt-fixer
- - repo: https://github.com/psf/black
- rev: stable
+
+ - repo: https://github.com/MarcoGorelli/auto-walrus
+ rev: 0.3.4
hooks:
- - id: black
- - repo: https://github.com/PyCQA/isort
- rev: 5.5.3
+ - id: auto-walrus
+
+ - repo: https://github.com/astral-sh/ruff-pre-commit
+ rev: v0.11.2
hooks:
- - id: isort
- args:
- - --profile=black
- - repo: https://gitlab.com/pycqa/flake8
- rev: 3.8.3
- hooks:
- - id: flake8
- args:
- - --ignore=E203,W503
- - --max-complexity=25
- - --max-line-length=88
-# FIXME: fix mypy errors and then uncomment this
-# - repo: https://github.com/pre-commit/mirrors-mypy
-# rev: v0.782
-# hooks:
-# - id: mypy
-# args:
-# - --ignore-missing-imports
+ - id: ruff
+ - id: ruff-format
+
- repo: https://github.com/codespell-project/codespell
- rev: v1.17.1
+ rev: v2.4.1
hooks:
- id: codespell
- args:
- - --ignore-words-list=ans,fo,followings,hist,iff,mater,secant,som,tim
- - --skip="./.*,./other/dictionary.txt,./other/words,./project_euler/problem_022/p022_names.txt"
- - --quiet-level=2
- exclude: |
- (?x)^(
- other/dictionary.txt |
- other/words |
- project_euler/problem_022/p022_names.txt
- )$
+ additional_dependencies:
+ - tomli
+
+ - repo: https://github.com/tox-dev/pyproject-fmt
+ rev: "v2.5.1"
+ hooks:
+ - id: pyproject-fmt
+
- repo: local
hooks:
- id: validate-filenames
@@ -58,3 +40,25 @@ repos:
entry: ./scripts/validate_filenames.py
language: script
pass_filenames: false
+
+ - repo: https://github.com/abravalheri/validate-pyproject
+ rev: v0.24.1
+ hooks:
+ - id: validate-pyproject
+
+ - repo: https://github.com/pre-commit/mirrors-mypy
+ rev: v1.15.0
+ hooks:
+ - id: mypy
+ args:
+ - --explicit-package-bases
+ - --ignore-missing-imports
+ - --install-types # See mirrors-mypy README.md
+ - --non-interactive
+ additional_dependencies: [types-requests]
+
+ - repo: https://github.com/pre-commit/mirrors-prettier
+ rev: "v4.0.0-alpha.8"
+ hooks:
+ - id: prettier
+ types_or: [toml, yaml]
diff --git a/.vscode/settings.json b/.vscode/settings.json
new file mode 100644
index 000000000000..ef16fa1aa7ac
--- /dev/null
+++ b/.vscode/settings.json
@@ -0,0 +1,5 @@
+{
+ "githubPullRequests.ignoredPullRequestBranches": [
+ "master"
+ ]
+}
diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md
index eedcb0250169..3df39f95b784 100644
--- a/CONTRIBUTING.md
+++ b/CONTRIBUTING.md
@@ -2,30 +2,41 @@
## Before contributing
-Welcome to [TheAlgorithms/Python](https://github.com/TheAlgorithms/Python)! Before sending your pull requests, make sure that you **read the whole guidelines**. If you have any doubt on the contributing guide, please feel free to [state it clearly in an issue](https://github.com/TheAlgorithms/Python/issues/new) or ask the community in [Gitter](https://gitter.im/TheAlgorithms).
+Welcome to [TheAlgorithms/Python](https://github.com/TheAlgorithms/Python)! Before submitting your pull requests, please ensure that you __read the whole guidelines__. If you have any doubts about the contributing guide, please feel free to [state it clearly in an issue](https://github.com/TheAlgorithms/Python/issues/new) or ask the community on [Gitter](https://gitter.im/TheAlgorithms/community).
## Contributing
### Contributor
-We are very happy that you consider implementing algorithms and data structure for others! This repository is referenced and used by learners from all over the globe. Being one of our contributors, you agree and confirm that:
+We are delighted that you are considering implementing algorithms and data structures for others! This repository is referenced and used by learners from all over the globe. By being one of our contributors, you agree and confirm that:
-- You did your work - no plagiarism allowed
+- You did your work - no plagiarism allowed.
- Any plagiarized work will not be merged.
-- Your work will be distributed under [MIT License](LICENSE.md) once your pull request is merged
-- You submitted work fulfils or mostly fulfils our styles and standards
+- Your work will be distributed under [MIT License](LICENSE.md) once your pull request is merged.
+- Your submitted work fulfills or mostly fulfills our styles and standards.
-**New implementation** is welcome! For example, new solutions for a problem, different representations for a graph data structure or algorithm designs with different complexity but **identical implementation** of an existing implementation is not allowed. Please check whether the solution is already implemented or not before submitting your pull request.
+__New implementation__ is welcome! For example, new solutions for a problem, different representations for a graph data structure or algorithm designs with different complexity, but __identical implementation__ of an existing implementation is not allowed. Please check whether the solution is already implemented or not before submitting your pull request.
-**Improving comments** and **writing proper tests** are also highly welcome.
+__Improving comments__ and __writing proper tests__ are also highly welcome.
### Contribution
We appreciate any contribution, from fixing a grammar mistake in a comment to implementing complex algorithms. Please read this section if you are contributing your work.
-Your contribution will be tested by our [automated testing on Travis CI](https://travis-ci.org/TheAlgorithms/Python/pull_requests) to save time and mental energy. After you have submitted your pull request, you should see the Travis tests start to run at the bottom of your submission page. If those tests fail, then click on the ___details___ button try to read through the Travis output to understand the failure. If you do not understand, please leave a comment on your submission page and a community member will try to help.
+Your contribution will be tested by our [automated testing on GitHub Actions](https://github.com/TheAlgorithms/Python/actions) to save time and mental energy. After you have submitted your pull request, you should see the GitHub Actions tests start to run at the bottom of your submission page. If those tests fail, then click on the ___details___ button to read through the GitHub Actions output to understand the failure. If you do not understand, please leave a comment on your submission page and a community member will try to help.
-Please help us keep our issue list small by adding fixes: #{$ISSUE_NO} to the commit message of pull requests that resolve open issues. GitHub will use this tag to auto close the issue when the PR is merged.
+#### Issues
+
+If you are interested in resolving an [open issue](https://github.com/TheAlgorithms/Python/issues), simply make a pull request with your proposed fix. __We do not assign issues in this repo__ so please do not ask for permission to work on an issue.
+
+__Do not__ create an issue to contribute an algorithm. Please submit a pull request instead.
+
+Please help us keep our issue list small by adding `Fixes #{$ISSUE_NUMBER}` to the description of pull requests that resolve open issues.
+For example, if your pull request fixes issue #10, then please add the following to its description:
+```
+Fixes #10
+```
+GitHub will use this tag to [auto-close the issue](https://docs.github.com/en/issues/tracking-your-work-with-issues/linking-a-pull-request-to-an-issue) if and when the PR is merged.
#### What is an Algorithm?
@@ -33,7 +44,7 @@ An Algorithm is one or more functions (or classes) that:
* take one or more inputs,
* perform some internal calculations or data manipulations,
* return one or more outputs,
-* have minimal side effects (Ex. print(), plot(), read(), write()).
+* have minimal side effects (Ex. `print()`, `plot()`, `read()`, `write()`).
Algorithms should be packaged in a way that would make it easy for readers to put them into larger programs.
@@ -42,21 +53,34 @@ Algorithms should:
* use Python naming conventions and intuitive variable names to ease comprehension
* be flexible to take different input values
* have Python type hints for their input parameters and return values
-* raise Python exceptions (ValueError, etc.) on erroneous input values
+* raise Python exceptions (`ValueError`, etc.) on erroneous input values
* have docstrings with clear explanations and/or URLs to source materials
* contain doctests that test both valid and erroneous input values
* return all calculation results instead of printing or plotting them
-Algorithms in this repo should not be how-to examples for existing Python packages. Instead, they should perform internal calculations or manipulations to convert input values into different output values. Those calculations or manipulations can use data types, classes, or functions of existing Python packages but each algorithm in this repo should add unique value.
+Algorithms in this repo should not be how-to examples for existing Python packages. Instead, they should perform internal calculations or manipulations to convert input values into different output values. Those calculations or manipulations can use data types, classes, or functions of existing Python packages but each algorithm in this repo should add unique value.
+
+#### Pre-commit plugin
+Use [pre-commit](https://pre-commit.com/#installation) to automatically format your code to match our coding style:
+
+```bash
+python3 -m pip install pre-commit # only required the first time
+pre-commit install
+```
+That's it! The plugin will run every time you commit any changes. If there are any errors found during the run, fix them and commit those changes. You can even run the plugin manually on all files:
+
+```bash
+pre-commit run --all-files --show-diff-on-failure
+```
#### Coding Style
We want your work to be readable by others; therefore, we encourage you to note the following:
-- Please write in Python 3.7+. For instance: __print()__ is a function in Python 3 so __print "Hello"__ will _not_ work but __print("Hello")__ will.
-- Please focus hard on naming of functions, classes, and variables. Help your reader by using __descriptive names__ that can help you to remove redundant comments.
- - Single letter variable names are _old school_ so please avoid them unless their life only spans a few lines.
- - Expand acronyms because __gcd()__ is hard to understand but __greatest_common_divisor()__ is not.
+- Please write in Python 3.13+. For instance: `print()` is a function in Python 3 so `print "Hello"` will *not* work but `print("Hello")` will.
+- Please focus hard on the naming of functions, classes, and variables. Help your reader by using __descriptive names__ that can help you to remove redundant comments.
+ - Single letter variable names are *old school* so please avoid them unless their life only spans a few lines.
+ - Expand acronyms because `gcd()` is hard to understand but `greatest_common_divisor()` is not.
- Please follow the [Python Naming Conventions](https://pep8.org/#prescriptive-naming-conventions) so variable_names and function_names should be lower_case, CONSTANTS in UPPERCASE, ClassNames should be CamelCase, etc.
- We encourage the use of Python [f-strings](https://realpython.com/python-f-strings/#f-strings-a-new-and-improved-way-to-format-strings-in-python) where they make the code easier to read.
@@ -64,15 +88,15 @@ We want your work to be readable by others; therefore, we encourage you to note
- Please consider running [__psf/black__](https://github.com/python/black) on your Python file(s) before submitting your pull request. This is not yet a requirement but it does make your code more readable and automatically aligns it with much of [PEP 8](https://www.python.org/dev/peps/pep-0008/). There are other code formatters (autopep8, yapf) but the __black__ formatter is now hosted by the Python Software Foundation. To use it,
```bash
- pip3 install black # only required the first time
+ python3 -m pip install black # only required the first time
black .
```
-- All submissions will need to pass the test __flake8 . --ignore=E203,W503 --max-line-length=88__ before they will be accepted so if possible, try this test locally on your Python file(s) before submitting your pull request.
+- All submissions will need to pass the test `ruff .` before they will be accepted so if possible, try this test locally on your Python file(s) before submitting your pull request.
```bash
- pip3 install flake8 # only required the first time
- flake8 . --ignore=E203,W503 --max-line-length=88 --show-source
+ python3 -m pip install ruff # only required the first time
+ ruff check
```
- Original code submission require docstrings or comments to describe your work.
@@ -89,7 +113,7 @@ We want your work to be readable by others; therefore, we encourage you to note
This is too trivial. Comments are expected to be explanatory. For comments, you can write them above, on or below a line of code, as long as you are consistent within the same piece of code.
- We encourage you to put docstrings inside your functions but please pay attention to indentation of docstrings. The following is a good example:
+ We encourage you to put docstrings inside your functions but please pay attention to the indentation of docstrings. The following is a good example:
```python
def sum_ab(a, b):
@@ -121,7 +145,7 @@ We want your work to be readable by others; therefore, we encourage you to note
python3 -m doctest -v my_submission.py
```
- The use of the Python builtin __input()__ function is **not** encouraged:
+ The use of the Python built-in `input()` function is __not__ encouraged:
```python
input('Enter your input:')
@@ -129,7 +153,7 @@ We want your work to be readable by others; therefore, we encourage you to note
input = eval(input("Enter your input: "))
```
- However, if your code uses __input()__ then we encourage you to gracefully deal with leading and trailing whitespace in user input by adding __.strip()__ as in:
+ However, if your code uses `input()` then we encourage you to gracefully deal with leading and trailing whitespace in user input by adding `.strip()` as in:
```python
starting_value = int(input("Please enter a starting value: ").strip())
@@ -142,10 +166,12 @@ We want your work to be readable by others; therefore, we encourage you to note
return a + b
```
+ Instructions on how to install mypy can be found [here](https://github.com/python/mypy). Please use the command `mypy --ignore-missing-imports .` to test all files or `mypy --ignore-missing-imports path/to/file.py` to test a specific file.
+
- [__List comprehensions and generators__](https://docs.python.org/3/tutorial/datastructures.html#list-comprehensions) are preferred over the use of `lambda`, `map`, `filter`, `reduce` but the important thing is to demonstrate the power of Python in code that is easy to read and maintain.
- Avoid importing external libraries for basic algorithms. Only use those libraries for complicated algorithms.
-- If you need a third party module that is not in the file __requirements.txt__, please add it to that file as part of your submission.
+- If you need a third-party module that is not in the file __requirements.txt__, please add it to that file as part of your submission.
#### Other Requirements for Submissions
- If you are submitting code in the `project_euler/` directory, please also read [the dedicated Guideline](https://github.com/TheAlgorithms/Python/blob/master/project_euler/README.md) before contributing to our Project Euler library.
@@ -155,13 +181,13 @@ We want your work to be readable by others; therefore, we encourage you to note
- If possible, follow the standard *within* the folder you are submitting to.
- If you have modified/added code work, make sure the code compiles before submitting.
- If you have modified/added documentation work, ensure your language is concise and contains no grammar errors.
-- Do not update the README.md or DIRECTORY.md file which will be periodically autogenerated by our Travis CI processes.
+- Do not update the README.md or DIRECTORY.md file which will be periodically autogenerated by our GitHub Actions processes.
- Add a corresponding explanation to [Algorithms-Explanation](https://github.com/TheAlgorithms/Algorithms-Explanation) (Optional but recommended).
-- All submissions will be tested with [__mypy__](http://www.mypy-lang.org) so we encourage to add [__Python type hints__](https://docs.python.org/3/library/typing.html) where it makes sense to do so.
+- All submissions will be tested with [__mypy__](http://www.mypy-lang.org) so we encourage you to add [__Python type hints__](https://docs.python.org/3/library/typing.html) where it makes sense to do so.
- Most importantly,
- - **Be consistent in the use of these guidelines when submitting.**
- - **Join** [Gitter](https://gitter.im/TheAlgorithms) **now!**
+ - __Be consistent in the use of these guidelines when submitting.__
+ - __Join__ us on [Discord](https://discord.com/invite/c7MnfGFGa6) and [Gitter](https://gitter.im/TheAlgorithms/community) __now!__
- Happy coding!
Writer [@poyea](https://github.com/poyea), Jun 2019.
diff --git a/DIRECTORY.md b/DIRECTORY.md
index cd8f6fb8578c..1c02c191bd14 100644
--- a/DIRECTORY.md
+++ b/DIRECTORY.md
@@ -1,873 +1,1379 @@
-## Arithmetic Analysis
- * [Bisection](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/bisection.py)
- * [Gaussian Elimination](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/gaussian_elimination.py)
- * [In Static Equilibrium](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/in_static_equilibrium.py)
- * [Intersection](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/intersection.py)
- * [Lu Decomposition](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/lu_decomposition.py)
- * [Newton Forward Interpolation](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/newton_forward_interpolation.py)
- * [Newton Method](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/newton_method.py)
- * [Newton Raphson](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/newton_raphson.py)
- * [Secant Method](https://github.com/TheAlgorithms/Python/blob/master/arithmetic_analysis/secant_method.py)
+## Audio Filters
+ * [Butterworth Filter](audio_filters/butterworth_filter.py)
+ * [Iir Filter](audio_filters/iir_filter.py)
+ * [Show Response](audio_filters/show_response.py)
## Backtracking
- * [All Combinations](https://github.com/TheAlgorithms/Python/blob/master/backtracking/all_combinations.py)
- * [All Permutations](https://github.com/TheAlgorithms/Python/blob/master/backtracking/all_permutations.py)
- * [All Subsequences](https://github.com/TheAlgorithms/Python/blob/master/backtracking/all_subsequences.py)
- * [Coloring](https://github.com/TheAlgorithms/Python/blob/master/backtracking/coloring.py)
- * [Hamiltonian Cycle](https://github.com/TheAlgorithms/Python/blob/master/backtracking/hamiltonian_cycle.py)
- * [Knight Tour](https://github.com/TheAlgorithms/Python/blob/master/backtracking/knight_tour.py)
- * [Minimax](https://github.com/TheAlgorithms/Python/blob/master/backtracking/minimax.py)
- * [N Queens](https://github.com/TheAlgorithms/Python/blob/master/backtracking/n_queens.py)
- * [N Queens Math](https://github.com/TheAlgorithms/Python/blob/master/backtracking/n_queens_math.py)
- * [Rat In Maze](https://github.com/TheAlgorithms/Python/blob/master/backtracking/rat_in_maze.py)
- * [Sudoku](https://github.com/TheAlgorithms/Python/blob/master/backtracking/sudoku.py)
- * [Sum Of Subsets](https://github.com/TheAlgorithms/Python/blob/master/backtracking/sum_of_subsets.py)
+ * [All Combinations](backtracking/all_combinations.py)
+ * [All Permutations](backtracking/all_permutations.py)
+ * [All Subsequences](backtracking/all_subsequences.py)
+ * [Coloring](backtracking/coloring.py)
+ * [Combination Sum](backtracking/combination_sum.py)
+ * [Crossword Puzzle Solver](backtracking/crossword_puzzle_solver.py)
+ * [Generate Parentheses](backtracking/generate_parentheses.py)
+ * [Hamiltonian Cycle](backtracking/hamiltonian_cycle.py)
+ * [Knight Tour](backtracking/knight_tour.py)
+ * [Match Word Pattern](backtracking/match_word_pattern.py)
+ * [Minimax](backtracking/minimax.py)
+ * [N Queens](backtracking/n_queens.py)
+ * [N Queens Math](backtracking/n_queens_math.py)
+ * [Power Sum](backtracking/power_sum.py)
+ * [Rat In Maze](backtracking/rat_in_maze.py)
+ * [Sudoku](backtracking/sudoku.py)
+ * [Sum Of Subsets](backtracking/sum_of_subsets.py)
+ * [Word Break](backtracking/word_break.py)
+ * [Word Ladder](backtracking/word_ladder.py)
+ * [Word Search](backtracking/word_search.py)
## Bit Manipulation
- * [Binary And Operator](https://github.com/TheAlgorithms/Python/blob/master/bit_manipulation/binary_and_operator.py)
- * [Binary Or Operator](https://github.com/TheAlgorithms/Python/blob/master/bit_manipulation/binary_or_operator.py)
- * [Binary Xor Operator](https://github.com/TheAlgorithms/Python/blob/master/bit_manipulation/binary_xor_operator.py)
- * [Single Bit Manipulation Operations](https://github.com/TheAlgorithms/Python/blob/master/bit_manipulation/single_bit_manipulation_operations.py)
+ * [Binary And Operator](bit_manipulation/binary_and_operator.py)
+ * [Binary Coded Decimal](bit_manipulation/binary_coded_decimal.py)
+ * [Binary Count Setbits](bit_manipulation/binary_count_setbits.py)
+ * [Binary Count Trailing Zeros](bit_manipulation/binary_count_trailing_zeros.py)
+ * [Binary Or Operator](bit_manipulation/binary_or_operator.py)
+ * [Binary Shifts](bit_manipulation/binary_shifts.py)
+ * [Binary Twos Complement](bit_manipulation/binary_twos_complement.py)
+ * [Binary Xor Operator](bit_manipulation/binary_xor_operator.py)
+ * [Bitwise Addition Recursive](bit_manipulation/bitwise_addition_recursive.py)
+ * [Count 1S Brian Kernighan Method](bit_manipulation/count_1s_brian_kernighan_method.py)
+ * [Count Number Of One Bits](bit_manipulation/count_number_of_one_bits.py)
+ * [Excess 3 Code](bit_manipulation/excess_3_code.py)
+ * [Find Previous Power Of Two](bit_manipulation/find_previous_power_of_two.py)
+ * [Gray Code Sequence](bit_manipulation/gray_code_sequence.py)
+ * [Highest Set Bit](bit_manipulation/highest_set_bit.py)
+ * [Index Of Rightmost Set Bit](bit_manipulation/index_of_rightmost_set_bit.py)
+ * [Is Even](bit_manipulation/is_even.py)
+ * [Is Power Of Two](bit_manipulation/is_power_of_two.py)
+ * [Largest Pow Of Two Le Num](bit_manipulation/largest_pow_of_two_le_num.py)
+ * [Missing Number](bit_manipulation/missing_number.py)
+ * [Numbers Different Signs](bit_manipulation/numbers_different_signs.py)
+ * [Power Of 4](bit_manipulation/power_of_4.py)
+ * [Reverse Bits](bit_manipulation/reverse_bits.py)
+ * [Single Bit Manipulation Operations](bit_manipulation/single_bit_manipulation_operations.py)
+ * [Swap All Odd And Even Bits](bit_manipulation/swap_all_odd_and_even_bits.py)
## Blockchain
- * [Chinese Remainder Theorem](https://github.com/TheAlgorithms/Python/blob/master/blockchain/chinese_remainder_theorem.py)
- * [Diophantine Equation](https://github.com/TheAlgorithms/Python/blob/master/blockchain/diophantine_equation.py)
- * [Modular Division](https://github.com/TheAlgorithms/Python/blob/master/blockchain/modular_division.py)
+ * [Diophantine Equation](blockchain/diophantine_equation.py)
## Boolean Algebra
- * [Quine Mc Cluskey](https://github.com/TheAlgorithms/Python/blob/master/boolean_algebra/quine_mc_cluskey.py)
+ * [And Gate](boolean_algebra/and_gate.py)
+ * [Imply Gate](boolean_algebra/imply_gate.py)
+ * [Karnaugh Map Simplification](boolean_algebra/karnaugh_map_simplification.py)
+ * [Multiplexer](boolean_algebra/multiplexer.py)
+ * [Nand Gate](boolean_algebra/nand_gate.py)
+ * [Nimply Gate](boolean_algebra/nimply_gate.py)
+ * [Nor Gate](boolean_algebra/nor_gate.py)
+ * [Not Gate](boolean_algebra/not_gate.py)
+ * [Or Gate](boolean_algebra/or_gate.py)
+ * [Quine Mc Cluskey](boolean_algebra/quine_mc_cluskey.py)
+ * [Xnor Gate](boolean_algebra/xnor_gate.py)
+ * [Xor Gate](boolean_algebra/xor_gate.py)
## Cellular Automata
- * [One Dimensional](https://github.com/TheAlgorithms/Python/blob/master/cellular_automata/one_dimensional.py)
+ * [Conways Game Of Life](cellular_automata/conways_game_of_life.py)
+ * [Game Of Life](cellular_automata/game_of_life.py)
+ * [Langtons Ant](cellular_automata/langtons_ant.py)
+ * [Nagel Schrekenberg](cellular_automata/nagel_schrekenberg.py)
+ * [One Dimensional](cellular_automata/one_dimensional.py)
+ * [Wa Tor](cellular_automata/wa_tor.py)
## Ciphers
- * [A1Z26](https://github.com/TheAlgorithms/Python/blob/master/ciphers/a1z26.py)
- * [Affine Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/affine_cipher.py)
- * [Atbash](https://github.com/TheAlgorithms/Python/blob/master/ciphers/atbash.py)
- * [Base16](https://github.com/TheAlgorithms/Python/blob/master/ciphers/base16.py)
- * [Base32](https://github.com/TheAlgorithms/Python/blob/master/ciphers/base32.py)
- * [Base64 Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/base64_cipher.py)
- * [Base85](https://github.com/TheAlgorithms/Python/blob/master/ciphers/base85.py)
- * [Beaufort Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/beaufort_cipher.py)
- * [Brute Force Caesar Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/brute_force_caesar_cipher.py)
- * [Caesar Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/caesar_cipher.py)
- * [Cryptomath Module](https://github.com/TheAlgorithms/Python/blob/master/ciphers/cryptomath_module.py)
- * [Decrypt Caesar With Chi Squared](https://github.com/TheAlgorithms/Python/blob/master/ciphers/decrypt_caesar_with_chi_squared.py)
- * [Deterministic Miller Rabin](https://github.com/TheAlgorithms/Python/blob/master/ciphers/deterministic_miller_rabin.py)
- * [Diffie](https://github.com/TheAlgorithms/Python/blob/master/ciphers/diffie.py)
- * [Elgamal Key Generator](https://github.com/TheAlgorithms/Python/blob/master/ciphers/elgamal_key_generator.py)
- * [Enigma Machine2](https://github.com/TheAlgorithms/Python/blob/master/ciphers/enigma_machine2.py)
- * [Hill Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/hill_cipher.py)
- * [Mixed Keyword Cypher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/mixed_keyword_cypher.py)
- * [Mono Alphabetic Ciphers](https://github.com/TheAlgorithms/Python/blob/master/ciphers/mono_alphabetic_ciphers.py)
- * [Morse Code Implementation](https://github.com/TheAlgorithms/Python/blob/master/ciphers/morse_code_implementation.py)
- * [Onepad Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/onepad_cipher.py)
- * [Playfair Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/playfair_cipher.py)
- * [Porta Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/porta_cipher.py)
- * [Rabin Miller](https://github.com/TheAlgorithms/Python/blob/master/ciphers/rabin_miller.py)
- * [Rail Fence Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/rail_fence_cipher.py)
- * [Rot13](https://github.com/TheAlgorithms/Python/blob/master/ciphers/rot13.py)
- * [Rsa Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/rsa_cipher.py)
- * [Rsa Factorization](https://github.com/TheAlgorithms/Python/blob/master/ciphers/rsa_factorization.py)
- * [Rsa Key Generator](https://github.com/TheAlgorithms/Python/blob/master/ciphers/rsa_key_generator.py)
- * [Shuffled Shift Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/shuffled_shift_cipher.py)
- * [Simple Keyword Cypher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/simple_keyword_cypher.py)
- * [Simple Substitution Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/simple_substitution_cipher.py)
- * [Trafid Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/trafid_cipher.py)
- * [Transposition Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/transposition_cipher.py)
- * [Transposition Cipher Encrypt Decrypt File](https://github.com/TheAlgorithms/Python/blob/master/ciphers/transposition_cipher_encrypt_decrypt_file.py)
- * [Vigenere Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/vigenere_cipher.py)
- * [Xor Cipher](https://github.com/TheAlgorithms/Python/blob/master/ciphers/xor_cipher.py)
+ * [A1Z26](ciphers/a1z26.py)
+ * [Affine Cipher](ciphers/affine_cipher.py)
+ * [Atbash](ciphers/atbash.py)
+ * [Autokey](ciphers/autokey.py)
+ * [Baconian Cipher](ciphers/baconian_cipher.py)
+ * [Base16](ciphers/base16.py)
+ * [Base32](ciphers/base32.py)
+ * [Base64 Cipher](ciphers/base64_cipher.py)
+ * [Base85](ciphers/base85.py)
+ * [Beaufort Cipher](ciphers/beaufort_cipher.py)
+ * [Bifid](ciphers/bifid.py)
+ * [Brute Force Caesar Cipher](ciphers/brute_force_caesar_cipher.py)
+ * [Caesar Cipher](ciphers/caesar_cipher.py)
+ * [Cryptomath Module](ciphers/cryptomath_module.py)
+ * [Decrypt Caesar With Chi Squared](ciphers/decrypt_caesar_with_chi_squared.py)
+ * [Deterministic Miller Rabin](ciphers/deterministic_miller_rabin.py)
+ * [Diffie](ciphers/diffie.py)
+ * [Diffie Hellman](ciphers/diffie_hellman.py)
+ * [Elgamal Key Generator](ciphers/elgamal_key_generator.py)
+ * [Enigma Machine2](ciphers/enigma_machine2.py)
+ * [Fractionated Morse Cipher](ciphers/fractionated_morse_cipher.py)
+ * [Gronsfeld Cipher](ciphers/gronsfeld_cipher.py)
+ * [Hill Cipher](ciphers/hill_cipher.py)
+ * [Mixed Keyword Cypher](ciphers/mixed_keyword_cypher.py)
+ * [Mono Alphabetic Ciphers](ciphers/mono_alphabetic_ciphers.py)
+ * [Morse Code](ciphers/morse_code.py)
+ * [Onepad Cipher](ciphers/onepad_cipher.py)
+ * [Permutation Cipher](ciphers/permutation_cipher.py)
+ * [Playfair Cipher](ciphers/playfair_cipher.py)
+ * [Polybius](ciphers/polybius.py)
+ * [Porta Cipher](ciphers/porta_cipher.py)
+ * [Rabin Miller](ciphers/rabin_miller.py)
+ * [Rail Fence Cipher](ciphers/rail_fence_cipher.py)
+ * [Rot13](ciphers/rot13.py)
+ * [Rsa Cipher](ciphers/rsa_cipher.py)
+ * [Rsa Factorization](ciphers/rsa_factorization.py)
+ * [Rsa Key Generator](ciphers/rsa_key_generator.py)
+ * [Running Key Cipher](ciphers/running_key_cipher.py)
+ * [Shuffled Shift Cipher](ciphers/shuffled_shift_cipher.py)
+ * [Simple Keyword Cypher](ciphers/simple_keyword_cypher.py)
+ * [Simple Substitution Cipher](ciphers/simple_substitution_cipher.py)
+ * [Transposition Cipher](ciphers/transposition_cipher.py)
+ * [Transposition Cipher Encrypt Decrypt File](ciphers/transposition_cipher_encrypt_decrypt_file.py)
+ * [Trifid Cipher](ciphers/trifid_cipher.py)
+ * [Vernam Cipher](ciphers/vernam_cipher.py)
+ * [Vigenere Cipher](ciphers/vigenere_cipher.py)
+ * [Xor Cipher](ciphers/xor_cipher.py)
## Compression
- * [Burrows Wheeler](https://github.com/TheAlgorithms/Python/blob/master/compression/burrows_wheeler.py)
- * [Huffman](https://github.com/TheAlgorithms/Python/blob/master/compression/huffman.py)
- * [Lempel Ziv](https://github.com/TheAlgorithms/Python/blob/master/compression/lempel_ziv.py)
- * [Lempel Ziv Decompress](https://github.com/TheAlgorithms/Python/blob/master/compression/lempel_ziv_decompress.py)
- * [Peak Signal To Noise Ratio](https://github.com/TheAlgorithms/Python/blob/master/compression/peak_signal_to_noise_ratio.py)
+ * [Burrows Wheeler](compression/burrows_wheeler.py)
+ * [Huffman](compression/huffman.py)
+ * [Lempel Ziv](compression/lempel_ziv.py)
+ * [Lempel Ziv Decompress](compression/lempel_ziv_decompress.py)
+ * [Lz77](compression/lz77.py)
+ * [Peak Signal To Noise Ratio](compression/peak_signal_to_noise_ratio.py)
+ * [Run Length Encoding](compression/run_length_encoding.py)
## Computer Vision
- * [Harriscorner](https://github.com/TheAlgorithms/Python/blob/master/computer_vision/harriscorner.py)
- * [Meanthreshold](https://github.com/TheAlgorithms/Python/blob/master/computer_vision/meanthreshold.py)
+ * [Cnn Classification](computer_vision/cnn_classification.py)
+ * [Flip Augmentation](computer_vision/flip_augmentation.py)
+ * [Haralick Descriptors](computer_vision/haralick_descriptors.py)
+ * [Harris Corner](computer_vision/harris_corner.py)
+ * [Horn Schunck](computer_vision/horn_schunck.py)
+ * [Intensity Based Segmentation](computer_vision/intensity_based_segmentation.py)
+ * [Mean Threshold](computer_vision/mean_threshold.py)
+ * [Mosaic Augmentation](computer_vision/mosaic_augmentation.py)
+ * [Pooling Functions](computer_vision/pooling_functions.py)
## Conversions
- * [Binary To Decimal](https://github.com/TheAlgorithms/Python/blob/master/conversions/binary_to_decimal.py)
- * [Binary To Octal](https://github.com/TheAlgorithms/Python/blob/master/conversions/binary_to_octal.py)
- * [Decimal To Any](https://github.com/TheAlgorithms/Python/blob/master/conversions/decimal_to_any.py)
- * [Decimal To Binary](https://github.com/TheAlgorithms/Python/blob/master/conversions/decimal_to_binary.py)
- * [Decimal To Binary Recursion](https://github.com/TheAlgorithms/Python/blob/master/conversions/decimal_to_binary_recursion.py)
- * [Decimal To Hexadecimal](https://github.com/TheAlgorithms/Python/blob/master/conversions/decimal_to_hexadecimal.py)
- * [Decimal To Octal](https://github.com/TheAlgorithms/Python/blob/master/conversions/decimal_to_octal.py)
- * [Hexadecimal To Decimal](https://github.com/TheAlgorithms/Python/blob/master/conversions/hexadecimal_to_decimal.py)
- * [Molecular Chemistry](https://github.com/TheAlgorithms/Python/blob/master/conversions/molecular_chemistry.py)
- * [Prefix Conversions](https://github.com/TheAlgorithms/Python/blob/master/conversions/prefix_conversions.py)
- * [Roman To Integer](https://github.com/TheAlgorithms/Python/blob/master/conversions/roman_to_integer.py)
- * [Temperature Conversions](https://github.com/TheAlgorithms/Python/blob/master/conversions/temperature_conversions.py)
+ * [Astronomical Length Scale Conversion](conversions/astronomical_length_scale_conversion.py)
+ * [Binary To Decimal](conversions/binary_to_decimal.py)
+ * [Binary To Hexadecimal](conversions/binary_to_hexadecimal.py)
+ * [Binary To Octal](conversions/binary_to_octal.py)
+ * [Convert Number To Words](conversions/convert_number_to_words.py)
+ * [Decimal To Any](conversions/decimal_to_any.py)
+ * [Decimal To Binary](conversions/decimal_to_binary.py)
+ * [Decimal To Hexadecimal](conversions/decimal_to_hexadecimal.py)
+ * [Decimal To Octal](conversions/decimal_to_octal.py)
+ * [Energy Conversions](conversions/energy_conversions.py)
+ * [Excel Title To Column](conversions/excel_title_to_column.py)
+ * [Hex To Bin](conversions/hex_to_bin.py)
+ * [Hexadecimal To Decimal](conversions/hexadecimal_to_decimal.py)
+ * [Ipv4 Conversion](conversions/ipv4_conversion.py)
+ * [Length Conversion](conversions/length_conversion.py)
+ * [Molecular Chemistry](conversions/molecular_chemistry.py)
+ * [Octal To Binary](conversions/octal_to_binary.py)
+ * [Octal To Decimal](conversions/octal_to_decimal.py)
+ * [Octal To Hexadecimal](conversions/octal_to_hexadecimal.py)
+ * [Prefix Conversions](conversions/prefix_conversions.py)
+ * [Prefix Conversions String](conversions/prefix_conversions_string.py)
+ * [Pressure Conversions](conversions/pressure_conversions.py)
+ * [Rectangular To Polar](conversions/rectangular_to_polar.py)
+ * [Rgb Cmyk Conversion](conversions/rgb_cmyk_conversion.py)
+ * [Rgb Hsv Conversion](conversions/rgb_hsv_conversion.py)
+ * [Roman Numerals](conversions/roman_numerals.py)
+ * [Speed Conversions](conversions/speed_conversions.py)
+ * [Temperature Conversions](conversions/temperature_conversions.py)
+ * [Time Conversions](conversions/time_conversions.py)
+ * [Volume Conversions](conversions/volume_conversions.py)
+ * [Weight Conversion](conversions/weight_conversion.py)
## Data Structures
+ * Arrays
+ * [Equilibrium Index In Array](data_structures/arrays/equilibrium_index_in_array.py)
+ * [Find Triplets With 0 Sum](data_structures/arrays/find_triplets_with_0_sum.py)
+ * [Index 2D Array In 1D](data_structures/arrays/index_2d_array_in_1d.py)
+ * [Kth Largest Element](data_structures/arrays/kth_largest_element.py)
+ * [Median Two Array](data_structures/arrays/median_two_array.py)
+ * [Monotonic Array](data_structures/arrays/monotonic_array.py)
+ * [Pairs With Given Sum](data_structures/arrays/pairs_with_given_sum.py)
+ * [Permutations](data_structures/arrays/permutations.py)
+ * [Prefix Sum](data_structures/arrays/prefix_sum.py)
+ * [Product Sum](data_structures/arrays/product_sum.py)
+ * [Sparse Table](data_structures/arrays/sparse_table.py)
+ * [Sudoku Solver](data_structures/arrays/sudoku_solver.py)
* Binary Tree
- * [Avl Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/avl_tree.py)
- * [Basic Binary Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/basic_binary_tree.py)
- * [Binary Search Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/binary_search_tree.py)
- * [Binary Search Tree Recursive](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/binary_search_tree_recursive.py)
- * [Binary Tree Mirror](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/binary_tree_mirror.py)
- * [Binary Tree Traversals](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/binary_tree_traversals.py)
- * [Fenwick Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/fenwick_tree.py)
- * [Lazy Segment Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/lazy_segment_tree.py)
- * [Lowest Common Ancestor](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/lowest_common_ancestor.py)
- * [Non Recursive Segment Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/non_recursive_segment_tree.py)
- * [Number Of Possible Binary Trees](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/number_of_possible_binary_trees.py)
- * [Red Black Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/red_black_tree.py)
- * [Segment Tree](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/segment_tree.py)
- * [Segment Tree Other](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/segment_tree_other.py)
- * [Treap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/binary_tree/treap.py)
+ * [Avl Tree](data_structures/binary_tree/avl_tree.py)
+ * [Basic Binary Tree](data_structures/binary_tree/basic_binary_tree.py)
+ * [Binary Search Tree](data_structures/binary_tree/binary_search_tree.py)
+ * [Binary Search Tree Recursive](data_structures/binary_tree/binary_search_tree_recursive.py)
+ * [Binary Tree Mirror](data_structures/binary_tree/binary_tree_mirror.py)
+ * [Binary Tree Node Sum](data_structures/binary_tree/binary_tree_node_sum.py)
+ * [Binary Tree Path Sum](data_structures/binary_tree/binary_tree_path_sum.py)
+ * [Binary Tree Traversals](data_structures/binary_tree/binary_tree_traversals.py)
+ * [Diameter Of Binary Tree](data_structures/binary_tree/diameter_of_binary_tree.py)
+ * [Diff Views Of Binary Tree](data_structures/binary_tree/diff_views_of_binary_tree.py)
+ * [Distribute Coins](data_structures/binary_tree/distribute_coins.py)
+ * [Fenwick Tree](data_structures/binary_tree/fenwick_tree.py)
+ * [Flatten Binarytree To Linkedlist](data_structures/binary_tree/flatten_binarytree_to_linkedlist.py)
+ * [Floor And Ceiling](data_structures/binary_tree/floor_and_ceiling.py)
+ * [Inorder Tree Traversal 2022](data_structures/binary_tree/inorder_tree_traversal_2022.py)
+ * [Is Sorted](data_structures/binary_tree/is_sorted.py)
+ * [Is Sum Tree](data_structures/binary_tree/is_sum_tree.py)
+ * [Lazy Segment Tree](data_structures/binary_tree/lazy_segment_tree.py)
+ * [Lowest Common Ancestor](data_structures/binary_tree/lowest_common_ancestor.py)
+ * [Maximum Fenwick Tree](data_structures/binary_tree/maximum_fenwick_tree.py)
+ * [Maximum Sum Bst](data_structures/binary_tree/maximum_sum_bst.py)
+ * [Merge Two Binary Trees](data_structures/binary_tree/merge_two_binary_trees.py)
+ * [Mirror Binary Tree](data_structures/binary_tree/mirror_binary_tree.py)
+ * [Non Recursive Segment Tree](data_structures/binary_tree/non_recursive_segment_tree.py)
+ * [Number Of Possible Binary Trees](data_structures/binary_tree/number_of_possible_binary_trees.py)
+ * [Red Black Tree](data_structures/binary_tree/red_black_tree.py)
+ * [Segment Tree](data_structures/binary_tree/segment_tree.py)
+ * [Segment Tree Other](data_structures/binary_tree/segment_tree_other.py)
+ * [Serialize Deserialize Binary Tree](data_structures/binary_tree/serialize_deserialize_binary_tree.py)
+ * [Symmetric Tree](data_structures/binary_tree/symmetric_tree.py)
+ * [Treap](data_structures/binary_tree/treap.py)
+ * [Wavelet Tree](data_structures/binary_tree/wavelet_tree.py)
* Disjoint Set
- * [Alternate Disjoint Set](https://github.com/TheAlgorithms/Python/blob/master/data_structures/disjoint_set/alternate_disjoint_set.py)
- * [Disjoint Set](https://github.com/TheAlgorithms/Python/blob/master/data_structures/disjoint_set/disjoint_set.py)
+ * [Alternate Disjoint Set](data_structures/disjoint_set/alternate_disjoint_set.py)
+ * [Disjoint Set](data_structures/disjoint_set/disjoint_set.py)
* Hashing
- * [Double Hash](https://github.com/TheAlgorithms/Python/blob/master/data_structures/hashing/double_hash.py)
- * [Hash Table](https://github.com/TheAlgorithms/Python/blob/master/data_structures/hashing/hash_table.py)
- * [Hash Table With Linked List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/hashing/hash_table_with_linked_list.py)
+ * [Bloom Filter](data_structures/hashing/bloom_filter.py)
+ * [Double Hash](data_structures/hashing/double_hash.py)
+ * [Hash Map](data_structures/hashing/hash_map.py)
+ * [Hash Table](data_structures/hashing/hash_table.py)
+ * [Hash Table With Linked List](data_structures/hashing/hash_table_with_linked_list.py)
* Number Theory
- * [Prime Numbers](https://github.com/TheAlgorithms/Python/blob/master/data_structures/hashing/number_theory/prime_numbers.py)
- * [Quadratic Probing](https://github.com/TheAlgorithms/Python/blob/master/data_structures/hashing/quadratic_probing.py)
+ * [Prime Numbers](data_structures/hashing/number_theory/prime_numbers.py)
+ * [Quadratic Probing](data_structures/hashing/quadratic_probing.py)
+ * Tests
+ * [Test Hash Map](data_structures/hashing/tests/test_hash_map.py)
* Heap
- * [Binomial Heap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/binomial_heap.py)
- * [Heap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/heap.py)
- * [Heap Generic](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/heap_generic.py)
- * [Max Heap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/max_heap.py)
- * [Min Heap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/min_heap.py)
- * [Randomized Heap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/randomized_heap.py)
- * [Skew Heap](https://github.com/TheAlgorithms/Python/blob/master/data_structures/heap/skew_heap.py)
+ * [Binomial Heap](data_structures/heap/binomial_heap.py)
+ * [Heap](data_structures/heap/heap.py)
+ * [Heap Generic](data_structures/heap/heap_generic.py)
+ * [Max Heap](data_structures/heap/max_heap.py)
+ * [Min Heap](data_structures/heap/min_heap.py)
+ * [Randomized Heap](data_structures/heap/randomized_heap.py)
+ * [Skew Heap](data_structures/heap/skew_heap.py)
+ * Kd Tree
+ * [Build Kdtree](data_structures/kd_tree/build_kdtree.py)
+ * Example
+ * [Example Usage](data_structures/kd_tree/example/example_usage.py)
+ * [Hypercube Points](data_structures/kd_tree/example/hypercube_points.py)
+ * [Kd Node](data_structures/kd_tree/kd_node.py)
+ * [Nearest Neighbour Search](data_structures/kd_tree/nearest_neighbour_search.py)
+ * Tests
+ * [Test Kdtree](data_structures/kd_tree/tests/test_kdtree.py)
* Linked List
- * [Circular Linked List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/circular_linked_list.py)
- * [Deque Doubly](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/deque_doubly.py)
- * [Doubly Linked List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/doubly_linked_list.py)
- * [Doubly Linked List Two](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/doubly_linked_list_two.py)
- * [From Sequence](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/from_sequence.py)
- * [Has Loop](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/has_loop.py)
- * [Is Palindrome](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/is_palindrome.py)
- * [Middle Element Of Linked List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/middle_element_of_linked_list.py)
- * [Print Reverse](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/print_reverse.py)
- * [Singly Linked List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/singly_linked_list.py)
- * [Skip List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/skip_list.py)
- * [Swap Nodes](https://github.com/TheAlgorithms/Python/blob/master/data_structures/linked_list/swap_nodes.py)
- * Queue
- * [Circular Queue](https://github.com/TheAlgorithms/Python/blob/master/data_structures/queue/circular_queue.py)
- * [Double Ended Queue](https://github.com/TheAlgorithms/Python/blob/master/data_structures/queue/double_ended_queue.py)
- * [Linked Queue](https://github.com/TheAlgorithms/Python/blob/master/data_structures/queue/linked_queue.py)
- * [Priority Queue Using List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/queue/priority_queue_using_list.py)
- * [Queue On List](https://github.com/TheAlgorithms/Python/blob/master/data_structures/queue/queue_on_list.py)
- * [Queue On Pseudo Stack](https://github.com/TheAlgorithms/Python/blob/master/data_structures/queue/queue_on_pseudo_stack.py)
+ * [Circular Linked List](data_structures/linked_list/circular_linked_list.py)
+ * [Deque Doubly](data_structures/linked_list/deque_doubly.py)
+ * [Doubly Linked List](data_structures/linked_list/doubly_linked_list.py)
+ * [Doubly Linked List Two](data_structures/linked_list/doubly_linked_list_two.py)
+ * [Floyds Cycle Detection](data_structures/linked_list/floyds_cycle_detection.py)
+ * [From Sequence](data_structures/linked_list/from_sequence.py)
+ * [Has Loop](data_structures/linked_list/has_loop.py)
+ * [Is Palindrome](data_structures/linked_list/is_palindrome.py)
+ * [Merge Two Lists](data_structures/linked_list/merge_two_lists.py)
+ * [Middle Element Of Linked List](data_structures/linked_list/middle_element_of_linked_list.py)
+ * [Print Reverse](data_structures/linked_list/print_reverse.py)
+ * [Reverse K Group](data_structures/linked_list/reverse_k_group.py)
+ * [Rotate To The Right](data_structures/linked_list/rotate_to_the_right.py)
+ * [Singly Linked List](data_structures/linked_list/singly_linked_list.py)
+ * [Skip List](data_structures/linked_list/skip_list.py)
+ * [Swap Nodes](data_structures/linked_list/swap_nodes.py)
+ * Queues
+ * [Circular Queue](data_structures/queues/circular_queue.py)
+ * [Circular Queue Linked List](data_structures/queues/circular_queue_linked_list.py)
+ * [Double Ended Queue](data_structures/queues/double_ended_queue.py)
+ * [Linked Queue](data_structures/queues/linked_queue.py)
+ * [Priority Queue Using List](data_structures/queues/priority_queue_using_list.py)
+ * [Queue By List](data_structures/queues/queue_by_list.py)
+ * [Queue By Two Stacks](data_structures/queues/queue_by_two_stacks.py)
+ * [Queue On Pseudo Stack](data_structures/queues/queue_on_pseudo_stack.py)
* Stacks
- * [Balanced Parentheses](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/balanced_parentheses.py)
- * [Dijkstras Two Stack Algorithm](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/dijkstras_two_stack_algorithm.py)
- * [Evaluate Postfix Notations](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/evaluate_postfix_notations.py)
- * [Infix To Postfix Conversion](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/infix_to_postfix_conversion.py)
- * [Infix To Prefix Conversion](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/infix_to_prefix_conversion.py)
- * [Linked Stack](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/linked_stack.py)
- * [Next Greater Element](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/next_greater_element.py)
- * [Postfix Evaluation](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/postfix_evaluation.py)
- * [Prefix Evaluation](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/prefix_evaluation.py)
- * [Stack](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/stack.py)
- * [Stack Using Dll](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/stack_using_dll.py)
- * [Stock Span Problem](https://github.com/TheAlgorithms/Python/blob/master/data_structures/stacks/stock_span_problem.py)
+ * [Balanced Parentheses](data_structures/stacks/balanced_parentheses.py)
+ * [Dijkstras Two Stack Algorithm](data_structures/stacks/dijkstras_two_stack_algorithm.py)
+ * [Infix To Postfix Conversion](data_structures/stacks/infix_to_postfix_conversion.py)
+ * [Infix To Prefix Conversion](data_structures/stacks/infix_to_prefix_conversion.py)
+ * [Largest Rectangle Histogram](data_structures/stacks/largest_rectangle_histogram.py)
+ * [Lexicographical Numbers](data_structures/stacks/lexicographical_numbers.py)
+ * [Next Greater Element](data_structures/stacks/next_greater_element.py)
+ * [Postfix Evaluation](data_structures/stacks/postfix_evaluation.py)
+ * [Prefix Evaluation](data_structures/stacks/prefix_evaluation.py)
+ * [Stack](data_structures/stacks/stack.py)
+ * [Stack Using Two Queues](data_structures/stacks/stack_using_two_queues.py)
+ * [Stack With Doubly Linked List](data_structures/stacks/stack_with_doubly_linked_list.py)
+ * [Stack With Singly Linked List](data_structures/stacks/stack_with_singly_linked_list.py)
+ * [Stock Span Problem](data_structures/stacks/stock_span_problem.py)
+ * Suffix Tree
+ * Example
+ * [Example Usage](data_structures/suffix_tree/example/example_usage.py)
+ * [Suffix Tree](data_structures/suffix_tree/suffix_tree.py)
+ * [Suffix Tree Node](data_structures/suffix_tree/suffix_tree_node.py)
+ * Tests
+ * [Test Suffix Tree](data_structures/suffix_tree/tests/test_suffix_tree.py)
* Trie
- * [Trie](https://github.com/TheAlgorithms/Python/blob/master/data_structures/trie/trie.py)
+ * [Radix Tree](data_structures/trie/radix_tree.py)
+ * [Trie](data_structures/trie/trie.py)
## Digital Image Processing
- * [Change Brightness](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/change_brightness.py)
- * [Change Contrast](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/change_contrast.py)
- * [Convert To Negative](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/convert_to_negative.py)
+ * [Change Brightness](digital_image_processing/change_brightness.py)
+ * [Change Contrast](digital_image_processing/change_contrast.py)
+ * [Convert To Negative](digital_image_processing/convert_to_negative.py)
* Dithering
- * [Burkes](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/dithering/burkes.py)
+ * [Burkes](digital_image_processing/dithering/burkes.py)
* Edge Detection
- * [Canny](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/edge_detection/canny.py)
+ * [Canny](digital_image_processing/edge_detection/canny.py)
* Filters
- * [Bilateral Filter](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/filters/bilateral_filter.py)
- * [Convolve](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/filters/convolve.py)
- * [Gaussian Filter](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/filters/gaussian_filter.py)
- * [Median Filter](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/filters/median_filter.py)
- * [Sobel Filter](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/filters/sobel_filter.py)
+ * [Bilateral Filter](digital_image_processing/filters/bilateral_filter.py)
+ * [Convolve](digital_image_processing/filters/convolve.py)
+ * [Gabor Filter](digital_image_processing/filters/gabor_filter.py)
+ * [Gaussian Filter](digital_image_processing/filters/gaussian_filter.py)
+ * [Laplacian Filter](digital_image_processing/filters/laplacian_filter.py)
+ * [Local Binary Pattern](digital_image_processing/filters/local_binary_pattern.py)
+ * [Median Filter](digital_image_processing/filters/median_filter.py)
+ * [Sobel Filter](digital_image_processing/filters/sobel_filter.py)
* Histogram Equalization
- * [Histogram Stretch](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/histogram_equalization/histogram_stretch.py)
- * [Index Calculation](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/index_calculation.py)
+ * [Histogram Stretch](digital_image_processing/histogram_equalization/histogram_stretch.py)
+ * [Index Calculation](digital_image_processing/index_calculation.py)
+ * Morphological Operations
+ * [Dilation Operation](digital_image_processing/morphological_operations/dilation_operation.py)
+ * [Erosion Operation](digital_image_processing/morphological_operations/erosion_operation.py)
* Resize
- * [Resize](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/resize/resize.py)
+ * [Resize](digital_image_processing/resize/resize.py)
* Rotation
- * [Rotation](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/rotation/rotation.py)
- * [Sepia](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/sepia.py)
- * [Test Digital Image Processing](https://github.com/TheAlgorithms/Python/blob/master/digital_image_processing/test_digital_image_processing.py)
+ * [Rotation](digital_image_processing/rotation/rotation.py)
+ * [Sepia](digital_image_processing/sepia.py)
+ * [Test Digital Image Processing](digital_image_processing/test_digital_image_processing.py)
## Divide And Conquer
- * [Closest Pair Of Points](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/closest_pair_of_points.py)
- * [Convex Hull](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/convex_hull.py)
- * [Heaps Algorithm](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/heaps_algorithm.py)
- * [Heaps Algorithm Iterative](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/heaps_algorithm_iterative.py)
- * [Inversions](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/inversions.py)
- * [Kth Order Statistic](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/kth_order_statistic.py)
- * [Max Subarray Sum](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/max_subarray_sum.py)
- * [Mergesort](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/mergesort.py)
- * [Peak](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/peak.py)
- * [Power](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/power.py)
- * [Strassen Matrix Multiplication](https://github.com/TheAlgorithms/Python/blob/master/divide_and_conquer/strassen_matrix_multiplication.py)
+ * [Closest Pair Of Points](divide_and_conquer/closest_pair_of_points.py)
+ * [Convex Hull](divide_and_conquer/convex_hull.py)
+ * [Heaps Algorithm](divide_and_conquer/heaps_algorithm.py)
+ * [Heaps Algorithm Iterative](divide_and_conquer/heaps_algorithm_iterative.py)
+ * [Inversions](divide_and_conquer/inversions.py)
+ * [Kth Order Statistic](divide_and_conquer/kth_order_statistic.py)
+ * [Max Difference Pair](divide_and_conquer/max_difference_pair.py)
+ * [Max Subarray](divide_and_conquer/max_subarray.py)
+ * [Mergesort](divide_and_conquer/mergesort.py)
+ * [Peak](divide_and_conquer/peak.py)
+ * [Power](divide_and_conquer/power.py)
+ * [Strassen Matrix Multiplication](divide_and_conquer/strassen_matrix_multiplication.py)
+
+## Docs
+ * [Conf](docs/conf.py)
## Dynamic Programming
- * [Abbreviation](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/abbreviation.py)
- * [Bitmask](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/bitmask.py)
- * [Climbing Stairs](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/climbing_stairs.py)
- * [Coin Change](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/coin_change.py)
- * [Edit Distance](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/edit_distance.py)
- * [Factorial](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/factorial.py)
- * [Fast Fibonacci](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/fast_fibonacci.py)
- * [Fibonacci](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/fibonacci.py)
- * [Floyd Warshall](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/floyd_warshall.py)
- * [Fractional Knapsack](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/fractional_knapsack.py)
- * [Fractional Knapsack 2](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/fractional_knapsack_2.py)
- * [Integer Partition](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/integer_partition.py)
- * [Iterating Through Submasks](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/iterating_through_submasks.py)
- * [Knapsack](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/knapsack.py)
- * [Longest Common Subsequence](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/longest_common_subsequence.py)
- * [Longest Increasing Subsequence](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/longest_increasing_subsequence.py)
- * [Longest Increasing Subsequence O(Nlogn)](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/longest_increasing_subsequence_o(nlogn).py)
- * [Longest Sub Array](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/longest_sub_array.py)
- * [Matrix Chain Order](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/matrix_chain_order.py)
- * [Max Non Adjacent Sum](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/max_non_adjacent_sum.py)
- * [Max Sub Array](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/max_sub_array.py)
- * [Max Sum Contiguous Subsequence](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/max_sum_contiguous_subsequence.py)
- * [Minimum Cost Path](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/minimum_cost_path.py)
- * [Minimum Partition](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/minimum_partition.py)
- * [Optimal Binary Search Tree](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/optimal_binary_search_tree.py)
- * [Rod Cutting](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/rod_cutting.py)
- * [Subset Generation](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/subset_generation.py)
- * [Sum Of Subset](https://github.com/TheAlgorithms/Python/blob/master/dynamic_programming/sum_of_subset.py)
+ * [Abbreviation](dynamic_programming/abbreviation.py)
+ * [All Construct](dynamic_programming/all_construct.py)
+ * [Bitmask](dynamic_programming/bitmask.py)
+ * [Catalan Numbers](dynamic_programming/catalan_numbers.py)
+ * [Climbing Stairs](dynamic_programming/climbing_stairs.py)
+ * [Combination Sum Iv](dynamic_programming/combination_sum_iv.py)
+ * [Edit Distance](dynamic_programming/edit_distance.py)
+ * [Factorial](dynamic_programming/factorial.py)
+ * [Fast Fibonacci](dynamic_programming/fast_fibonacci.py)
+ * [Fibonacci](dynamic_programming/fibonacci.py)
+ * [Fizz Buzz](dynamic_programming/fizz_buzz.py)
+ * [Floyd Warshall](dynamic_programming/floyd_warshall.py)
+ * [Integer Partition](dynamic_programming/integer_partition.py)
+ * [Iterating Through Submasks](dynamic_programming/iterating_through_submasks.py)
+ * [K Means Clustering Tensorflow](dynamic_programming/k_means_clustering_tensorflow.py)
+ * [Knapsack](dynamic_programming/knapsack.py)
+ * [Largest Divisible Subset](dynamic_programming/largest_divisible_subset.py)
+ * [Longest Common Subsequence](dynamic_programming/longest_common_subsequence.py)
+ * [Longest Common Substring](dynamic_programming/longest_common_substring.py)
+ * [Longest Increasing Subsequence](dynamic_programming/longest_increasing_subsequence.py)
+ * [Longest Increasing Subsequence Iterative](dynamic_programming/longest_increasing_subsequence_iterative.py)
+ * [Longest Increasing Subsequence O Nlogn](dynamic_programming/longest_increasing_subsequence_o_nlogn.py)
+ * [Longest Palindromic Subsequence](dynamic_programming/longest_palindromic_subsequence.py)
+ * [Matrix Chain Multiplication](dynamic_programming/matrix_chain_multiplication.py)
+ * [Matrix Chain Order](dynamic_programming/matrix_chain_order.py)
+ * [Max Non Adjacent Sum](dynamic_programming/max_non_adjacent_sum.py)
+ * [Max Product Subarray](dynamic_programming/max_product_subarray.py)
+ * [Max Subarray Sum](dynamic_programming/max_subarray_sum.py)
+ * [Min Distance Up Bottom](dynamic_programming/min_distance_up_bottom.py)
+ * [Minimum Coin Change](dynamic_programming/minimum_coin_change.py)
+ * [Minimum Cost Path](dynamic_programming/minimum_cost_path.py)
+ * [Minimum Partition](dynamic_programming/minimum_partition.py)
+ * [Minimum Size Subarray Sum](dynamic_programming/minimum_size_subarray_sum.py)
+ * [Minimum Squares To Represent A Number](dynamic_programming/minimum_squares_to_represent_a_number.py)
+ * [Minimum Steps To One](dynamic_programming/minimum_steps_to_one.py)
+ * [Minimum Tickets Cost](dynamic_programming/minimum_tickets_cost.py)
+ * [Optimal Binary Search Tree](dynamic_programming/optimal_binary_search_tree.py)
+ * [Palindrome Partitioning](dynamic_programming/palindrome_partitioning.py)
+ * [Range Sum Query](dynamic_programming/range_sum_query.py)
+ * [Regex Match](dynamic_programming/regex_match.py)
+ * [Rod Cutting](dynamic_programming/rod_cutting.py)
+ * [Smith Waterman](dynamic_programming/smith_waterman.py)
+ * [Subset Generation](dynamic_programming/subset_generation.py)
+ * [Sum Of Subset](dynamic_programming/sum_of_subset.py)
+ * [Trapped Water](dynamic_programming/trapped_water.py)
+ * [Tribonacci](dynamic_programming/tribonacci.py)
+ * [Viterbi](dynamic_programming/viterbi.py)
+ * [Wildcard Matching](dynamic_programming/wildcard_matching.py)
+ * [Word Break](dynamic_programming/word_break.py)
+
+## Electronics
+ * [Apparent Power](electronics/apparent_power.py)
+ * [Builtin Voltage](electronics/builtin_voltage.py)
+ * [Capacitor Equivalence](electronics/capacitor_equivalence.py)
+ * [Carrier Concentration](electronics/carrier_concentration.py)
+ * [Charging Capacitor](electronics/charging_capacitor.py)
+ * [Charging Inductor](electronics/charging_inductor.py)
+ * [Circular Convolution](electronics/circular_convolution.py)
+ * [Coulombs Law](electronics/coulombs_law.py)
+ * [Electric Conductivity](electronics/electric_conductivity.py)
+ * [Electric Power](electronics/electric_power.py)
+ * [Electrical Impedance](electronics/electrical_impedance.py)
+ * [Ic 555 Timer](electronics/ic_555_timer.py)
+ * [Ind Reactance](electronics/ind_reactance.py)
+ * [Ohms Law](electronics/ohms_law.py)
+ * [Real And Reactive Power](electronics/real_and_reactive_power.py)
+ * [Resistor Color Code](electronics/resistor_color_code.py)
+ * [Resistor Equivalence](electronics/resistor_equivalence.py)
+ * [Resonant Frequency](electronics/resonant_frequency.py)
+ * [Wheatstone Bridge](electronics/wheatstone_bridge.py)
## File Transfer
- * [Receive File](https://github.com/TheAlgorithms/Python/blob/master/file_transfer/receive_file.py)
- * [Send File](https://github.com/TheAlgorithms/Python/blob/master/file_transfer/send_file.py)
+ * [Receive File](file_transfer/receive_file.py)
+ * [Send File](file_transfer/send_file.py)
* Tests
- * [Test Send File](https://github.com/TheAlgorithms/Python/blob/master/file_transfer/tests/test_send_file.py)
+ * [Test Send File](file_transfer/tests/test_send_file.py)
+
+## Financial
+ * [Equated Monthly Installments](financial/equated_monthly_installments.py)
+ * [Exponential Moving Average](financial/exponential_moving_average.py)
+ * [Interest](financial/interest.py)
+ * [Present Value](financial/present_value.py)
+ * [Price Plus Tax](financial/price_plus_tax.py)
+ * [Simple Moving Average](financial/simple_moving_average.py)
+
+## Fractals
+ * [Julia Sets](fractals/julia_sets.py)
+ * [Koch Snowflake](fractals/koch_snowflake.py)
+ * [Mandelbrot](fractals/mandelbrot.py)
+ * [Sierpinski Triangle](fractals/sierpinski_triangle.py)
+ * [Vicsek](fractals/vicsek.py)
## Fuzzy Logic
- * [Fuzzy Operations](https://github.com/TheAlgorithms/Python/blob/master/fuzzy_logic/fuzzy_operations.py)
+ * [Fuzzy Operations](fuzzy_logic/fuzzy_operations.py)
## Genetic Algorithm
- * [Basic String](https://github.com/TheAlgorithms/Python/blob/master/genetic_algorithm/basic_string.py)
+ * [Basic String](genetic_algorithm/basic_string.py)
## Geodesy
- * [Haversine Distance](https://github.com/TheAlgorithms/Python/blob/master/geodesy/haversine_distance.py)
- * [Lamberts Ellipsoidal Distance](https://github.com/TheAlgorithms/Python/blob/master/geodesy/lamberts_ellipsoidal_distance.py)
+ * [Haversine Distance](geodesy/haversine_distance.py)
+ * [Lamberts Ellipsoidal Distance](geodesy/lamberts_ellipsoidal_distance.py)
+
+## Geometry
+ * [Geometry](geometry/geometry.py)
## Graphics
- * [Bezier Curve](https://github.com/TheAlgorithms/Python/blob/master/graphics/bezier_curve.py)
- * [Vector3 For 2D Rendering](https://github.com/TheAlgorithms/Python/blob/master/graphics/vector3_for_2d_rendering.py)
+ * [Bezier Curve](graphics/bezier_curve.py)
+ * [Butterfly Pattern](graphics/butterfly_pattern.py)
+ * [Digital Differential Analyzer Line](graphics/digital_differential_analyzer_line.py)
+ * [Vector3 For 2D Rendering](graphics/vector3_for_2d_rendering.py)
## Graphs
- * [A Star](https://github.com/TheAlgorithms/Python/blob/master/graphs/a_star.py)
- * [Articulation Points](https://github.com/TheAlgorithms/Python/blob/master/graphs/articulation_points.py)
- * [Basic Graphs](https://github.com/TheAlgorithms/Python/blob/master/graphs/basic_graphs.py)
- * [Bellman Ford](https://github.com/TheAlgorithms/Python/blob/master/graphs/bellman_ford.py)
- * [Bfs Shortest Path](https://github.com/TheAlgorithms/Python/blob/master/graphs/bfs_shortest_path.py)
- * [Bfs Zero One Shortest Path](https://github.com/TheAlgorithms/Python/blob/master/graphs/bfs_zero_one_shortest_path.py)
- * [Bidirectional A Star](https://github.com/TheAlgorithms/Python/blob/master/graphs/bidirectional_a_star.py)
- * [Bidirectional Breadth First Search](https://github.com/TheAlgorithms/Python/blob/master/graphs/bidirectional_breadth_first_search.py)
- * [Breadth First Search](https://github.com/TheAlgorithms/Python/blob/master/graphs/breadth_first_search.py)
- * [Breadth First Search 2](https://github.com/TheAlgorithms/Python/blob/master/graphs/breadth_first_search_2.py)
- * [Breadth First Search Shortest Path](https://github.com/TheAlgorithms/Python/blob/master/graphs/breadth_first_search_shortest_path.py)
- * [Check Bipartite Graph Bfs](https://github.com/TheAlgorithms/Python/blob/master/graphs/check_bipartite_graph_bfs.py)
- * [Check Bipartite Graph Dfs](https://github.com/TheAlgorithms/Python/blob/master/graphs/check_bipartite_graph_dfs.py)
- * [Connected Components](https://github.com/TheAlgorithms/Python/blob/master/graphs/connected_components.py)
- * [Depth First Search](https://github.com/TheAlgorithms/Python/blob/master/graphs/depth_first_search.py)
- * [Depth First Search 2](https://github.com/TheAlgorithms/Python/blob/master/graphs/depth_first_search_2.py)
- * [Dijkstra](https://github.com/TheAlgorithms/Python/blob/master/graphs/dijkstra.py)
- * [Dijkstra 2](https://github.com/TheAlgorithms/Python/blob/master/graphs/dijkstra_2.py)
- * [Dijkstra Algorithm](https://github.com/TheAlgorithms/Python/blob/master/graphs/dijkstra_algorithm.py)
- * [Dinic](https://github.com/TheAlgorithms/Python/blob/master/graphs/dinic.py)
- * [Directed And Undirected (Weighted) Graph](https://github.com/TheAlgorithms/Python/blob/master/graphs/directed_and_undirected_(weighted)_graph.py)
- * [Edmonds Karp Multiple Source And Sink](https://github.com/TheAlgorithms/Python/blob/master/graphs/edmonds_karp_multiple_source_and_sink.py)
- * [Eulerian Path And Circuit For Undirected Graph](https://github.com/TheAlgorithms/Python/blob/master/graphs/eulerian_path_and_circuit_for_undirected_graph.py)
- * [Even Tree](https://github.com/TheAlgorithms/Python/blob/master/graphs/even_tree.py)
- * [Finding Bridges](https://github.com/TheAlgorithms/Python/blob/master/graphs/finding_bridges.py)
- * [Frequent Pattern Graph Miner](https://github.com/TheAlgorithms/Python/blob/master/graphs/frequent_pattern_graph_miner.py)
- * [G Topological Sort](https://github.com/TheAlgorithms/Python/blob/master/graphs/g_topological_sort.py)
- * [Gale Shapley Bigraph](https://github.com/TheAlgorithms/Python/blob/master/graphs/gale_shapley_bigraph.py)
- * [Graph List](https://github.com/TheAlgorithms/Python/blob/master/graphs/graph_list.py)
- * [Graph Matrix](https://github.com/TheAlgorithms/Python/blob/master/graphs/graph_matrix.py)
- * [Graphs Floyd Warshall](https://github.com/TheAlgorithms/Python/blob/master/graphs/graphs_floyd_warshall.py)
- * [Greedy Best First](https://github.com/TheAlgorithms/Python/blob/master/graphs/greedy_best_first.py)
- * [Kahns Algorithm Long](https://github.com/TheAlgorithms/Python/blob/master/graphs/kahns_algorithm_long.py)
- * [Kahns Algorithm Topo](https://github.com/TheAlgorithms/Python/blob/master/graphs/kahns_algorithm_topo.py)
- * [Karger](https://github.com/TheAlgorithms/Python/blob/master/graphs/karger.py)
- * [Minimum Spanning Tree Boruvka](https://github.com/TheAlgorithms/Python/blob/master/graphs/minimum_spanning_tree_boruvka.py)
- * [Minimum Spanning Tree Kruskal](https://github.com/TheAlgorithms/Python/blob/master/graphs/minimum_spanning_tree_kruskal.py)
- * [Minimum Spanning Tree Kruskal2](https://github.com/TheAlgorithms/Python/blob/master/graphs/minimum_spanning_tree_kruskal2.py)
- * [Minimum Spanning Tree Prims](https://github.com/TheAlgorithms/Python/blob/master/graphs/minimum_spanning_tree_prims.py)
- * [Minimum Spanning Tree Prims2](https://github.com/TheAlgorithms/Python/blob/master/graphs/minimum_spanning_tree_prims2.py)
- * [Multi Heuristic Astar](https://github.com/TheAlgorithms/Python/blob/master/graphs/multi_heuristic_astar.py)
- * [Page Rank](https://github.com/TheAlgorithms/Python/blob/master/graphs/page_rank.py)
- * [Prim](https://github.com/TheAlgorithms/Python/blob/master/graphs/prim.py)
- * [Scc Kosaraju](https://github.com/TheAlgorithms/Python/blob/master/graphs/scc_kosaraju.py)
- * [Strongly Connected Components](https://github.com/TheAlgorithms/Python/blob/master/graphs/strongly_connected_components.py)
- * [Tarjans Scc](https://github.com/TheAlgorithms/Python/blob/master/graphs/tarjans_scc.py)
+ * [A Star](graphs/a_star.py)
+ * [Ant Colony Optimization Algorithms](graphs/ant_colony_optimization_algorithms.py)
+ * [Articulation Points](graphs/articulation_points.py)
+ * [Basic Graphs](graphs/basic_graphs.py)
+ * [Bellman Ford](graphs/bellman_ford.py)
+ * [Bi Directional Dijkstra](graphs/bi_directional_dijkstra.py)
+ * [Bidirectional A Star](graphs/bidirectional_a_star.py)
+ * [Bidirectional Breadth First Search](graphs/bidirectional_breadth_first_search.py)
+ * [Boruvka](graphs/boruvka.py)
+ * [Breadth First Search](graphs/breadth_first_search.py)
+ * [Breadth First Search 2](graphs/breadth_first_search_2.py)
+ * [Breadth First Search Shortest Path](graphs/breadth_first_search_shortest_path.py)
+ * [Breadth First Search Shortest Path 2](graphs/breadth_first_search_shortest_path_2.py)
+ * [Breadth First Search Zero One Shortest Path](graphs/breadth_first_search_zero_one_shortest_path.py)
+ * [Check Bipatrite](graphs/check_bipatrite.py)
+ * [Check Cycle](graphs/check_cycle.py)
+ * [Connected Components](graphs/connected_components.py)
+ * [Deep Clone Graph](graphs/deep_clone_graph.py)
+ * [Depth First Search](graphs/depth_first_search.py)
+ * [Depth First Search 2](graphs/depth_first_search_2.py)
+ * [Dijkstra](graphs/dijkstra.py)
+ * [Dijkstra 2](graphs/dijkstra_2.py)
+ * [Dijkstra Algorithm](graphs/dijkstra_algorithm.py)
+ * [Dijkstra Alternate](graphs/dijkstra_alternate.py)
+ * [Dijkstra Binary Grid](graphs/dijkstra_binary_grid.py)
+ * [Dinic](graphs/dinic.py)
+ * [Directed And Undirected Weighted Graph](graphs/directed_and_undirected_weighted_graph.py)
+ * [Edmonds Karp Multiple Source And Sink](graphs/edmonds_karp_multiple_source_and_sink.py)
+ * [Eulerian Path And Circuit For Undirected Graph](graphs/eulerian_path_and_circuit_for_undirected_graph.py)
+ * [Even Tree](graphs/even_tree.py)
+ * [Finding Bridges](graphs/finding_bridges.py)
+ * [Frequent Pattern Graph Miner](graphs/frequent_pattern_graph_miner.py)
+ * [G Topological Sort](graphs/g_topological_sort.py)
+ * [Gale Shapley Bigraph](graphs/gale_shapley_bigraph.py)
+ * [Graph Adjacency List](graphs/graph_adjacency_list.py)
+ * [Graph Adjacency Matrix](graphs/graph_adjacency_matrix.py)
+ * [Graph List](graphs/graph_list.py)
+ * [Graphs Floyd Warshall](graphs/graphs_floyd_warshall.py)
+ * [Greedy Best First](graphs/greedy_best_first.py)
+ * [Greedy Min Vertex Cover](graphs/greedy_min_vertex_cover.py)
+ * [Kahns Algorithm Long](graphs/kahns_algorithm_long.py)
+ * [Kahns Algorithm Topo](graphs/kahns_algorithm_topo.py)
+ * [Karger](graphs/karger.py)
+ * [Lanczos Eigenvectors](graphs/lanczos_eigenvectors.py)
+ * [Markov Chain](graphs/markov_chain.py)
+ * [Matching Min Vertex Cover](graphs/matching_min_vertex_cover.py)
+ * [Minimum Path Sum](graphs/minimum_path_sum.py)
+ * [Minimum Spanning Tree Boruvka](graphs/minimum_spanning_tree_boruvka.py)
+ * [Minimum Spanning Tree Kruskal](graphs/minimum_spanning_tree_kruskal.py)
+ * [Minimum Spanning Tree Kruskal2](graphs/minimum_spanning_tree_kruskal2.py)
+ * [Minimum Spanning Tree Prims](graphs/minimum_spanning_tree_prims.py)
+ * [Minimum Spanning Tree Prims2](graphs/minimum_spanning_tree_prims2.py)
+ * [Multi Heuristic Astar](graphs/multi_heuristic_astar.py)
+ * [Page Rank](graphs/page_rank.py)
+ * [Prim](graphs/prim.py)
+ * [Random Graph Generator](graphs/random_graph_generator.py)
+ * [Scc Kosaraju](graphs/scc_kosaraju.py)
+ * [Strongly Connected Components](graphs/strongly_connected_components.py)
+ * [Tarjans Scc](graphs/tarjans_scc.py)
* Tests
- * [Test Min Spanning Tree Kruskal](https://github.com/TheAlgorithms/Python/blob/master/graphs/tests/test_min_spanning_tree_kruskal.py)
- * [Test Min Spanning Tree Prim](https://github.com/TheAlgorithms/Python/blob/master/graphs/tests/test_min_spanning_tree_prim.py)
+ * [Test Min Spanning Tree Kruskal](graphs/tests/test_min_spanning_tree_kruskal.py)
+ * [Test Min Spanning Tree Prim](graphs/tests/test_min_spanning_tree_prim.py)
-## Greedy Method
- * [Greedy Knapsack](https://github.com/TheAlgorithms/Python/blob/master/greedy_method/greedy_knapsack.py)
- * [Test Knapsack](https://github.com/TheAlgorithms/Python/blob/master/greedy_method/test_knapsack.py)
+## Greedy Methods
+ * [Best Time To Buy And Sell Stock](greedy_methods/best_time_to_buy_and_sell_stock.py)
+ * [Fractional Cover Problem](greedy_methods/fractional_cover_problem.py)
+ * [Fractional Knapsack](greedy_methods/fractional_knapsack.py)
+ * [Fractional Knapsack 2](greedy_methods/fractional_knapsack_2.py)
+ * [Gas Station](greedy_methods/gas_station.py)
+ * [Minimum Coin Change](greedy_methods/minimum_coin_change.py)
+ * [Minimum Waiting Time](greedy_methods/minimum_waiting_time.py)
+ * [Optimal Merge Pattern](greedy_methods/optimal_merge_pattern.py)
+ * [Smallest Range](greedy_methods/smallest_range.py)
## Hashes
- * [Adler32](https://github.com/TheAlgorithms/Python/blob/master/hashes/adler32.py)
- * [Chaos Machine](https://github.com/TheAlgorithms/Python/blob/master/hashes/chaos_machine.py)
- * [Djb2](https://github.com/TheAlgorithms/Python/blob/master/hashes/djb2.py)
- * [Enigma Machine](https://github.com/TheAlgorithms/Python/blob/master/hashes/enigma_machine.py)
- * [Hamming Code](https://github.com/TheAlgorithms/Python/blob/master/hashes/hamming_code.py)
- * [Md5](https://github.com/TheAlgorithms/Python/blob/master/hashes/md5.py)
- * [Sdbm](https://github.com/TheAlgorithms/Python/blob/master/hashes/sdbm.py)
- * [Sha1](https://github.com/TheAlgorithms/Python/blob/master/hashes/sha1.py)
+ * [Adler32](hashes/adler32.py)
+ * [Chaos Machine](hashes/chaos_machine.py)
+ * [Djb2](hashes/djb2.py)
+ * [Elf](hashes/elf.py)
+ * [Enigma Machine](hashes/enigma_machine.py)
+ * [Fletcher16](hashes/fletcher16.py)
+ * [Hamming Code](hashes/hamming_code.py)
+ * [Luhn](hashes/luhn.py)
+ * [Md5](hashes/md5.py)
+ * [Sdbm](hashes/sdbm.py)
+ * [Sha1](hashes/sha1.py)
+ * [Sha256](hashes/sha256.py)
## Knapsack
- * [Knapsack](https://github.com/TheAlgorithms/Python/blob/master/knapsack/knapsack.py)
- * [Test Knapsack](https://github.com/TheAlgorithms/Python/blob/master/knapsack/test_knapsack.py)
+ * [Greedy Knapsack](knapsack/greedy_knapsack.py)
+ * [Knapsack](knapsack/knapsack.py)
+ * [Recursive Approach Knapsack](knapsack/recursive_approach_knapsack.py)
+ * Tests
+ * [Test Greedy Knapsack](knapsack/tests/test_greedy_knapsack.py)
+ * [Test Knapsack](knapsack/tests/test_knapsack.py)
## Linear Algebra
+ * [Gaussian Elimination](linear_algebra/gaussian_elimination.py)
+ * [Jacobi Iteration Method](linear_algebra/jacobi_iteration_method.py)
+ * [Lu Decomposition](linear_algebra/lu_decomposition.py)
* Src
- * [Lib](https://github.com/TheAlgorithms/Python/blob/master/linear_algebra/src/lib.py)
- * [Polynom For Points](https://github.com/TheAlgorithms/Python/blob/master/linear_algebra/src/polynom_for_points.py)
- * [Power Iteration](https://github.com/TheAlgorithms/Python/blob/master/linear_algebra/src/power_iteration.py)
- * [Rayleigh Quotient](https://github.com/TheAlgorithms/Python/blob/master/linear_algebra/src/rayleigh_quotient.py)
- * [Test Linear Algebra](https://github.com/TheAlgorithms/Python/blob/master/linear_algebra/src/test_linear_algebra.py)
- * [Transformations 2D](https://github.com/TheAlgorithms/Python/blob/master/linear_algebra/src/transformations_2d.py)
+ * [Conjugate Gradient](linear_algebra/src/conjugate_gradient.py)
+ * [Gaussian Elimination Pivoting](linear_algebra/src/gaussian_elimination_pivoting.py)
+ * [Lib](linear_algebra/src/lib.py)
+ * [Polynom For Points](linear_algebra/src/polynom_for_points.py)
+ * [Power Iteration](linear_algebra/src/power_iteration.py)
+ * [Rank Of Matrix](linear_algebra/src/rank_of_matrix.py)
+ * [Rayleigh Quotient](linear_algebra/src/rayleigh_quotient.py)
+ * [Schur Complement](linear_algebra/src/schur_complement.py)
+ * [Test Linear Algebra](linear_algebra/src/test_linear_algebra.py)
+ * [Transformations 2D](linear_algebra/src/transformations_2d.py)
+
+## Linear Programming
+ * [Simplex](linear_programming/simplex.py)
## Machine Learning
- * [Astar](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/astar.py)
- * [Data Transformations](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/data_transformations.py)
- * [Decision Tree](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/decision_tree.py)
+ * [Apriori Algorithm](machine_learning/apriori_algorithm.py)
+ * [Astar](machine_learning/astar.py)
+ * [Automatic Differentiation](machine_learning/automatic_differentiation.py)
+ * [Data Transformations](machine_learning/data_transformations.py)
+ * [Decision Tree](machine_learning/decision_tree.py)
+ * [Dimensionality Reduction](machine_learning/dimensionality_reduction.py)
* Forecasting
- * [Run](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/forecasting/run.py)
- * [Gaussian Naive Bayes](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/gaussian_naive_bayes.py)
- * [Gradient Boosting Regressor](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/gradient_boosting_regressor.py)
- * [Gradient Descent](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/gradient_descent.py)
- * [K Means Clust](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/k_means_clust.py)
- * [K Nearest Neighbours](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/k_nearest_neighbours.py)
- * [Knn Sklearn](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/knn_sklearn.py)
- * [Linear Discriminant Analysis](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/linear_discriminant_analysis.py)
- * [Linear Regression](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/linear_regression.py)
- * [Logistic Regression](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/logistic_regression.py)
+ * [Run](machine_learning/forecasting/run.py)
+ * [Frequent Pattern Growth](machine_learning/frequent_pattern_growth.py)
+ * [Gradient Boosting Classifier](machine_learning/gradient_boosting_classifier.py)
+ * [Gradient Descent](machine_learning/gradient_descent.py)
+ * [K Means Clust](machine_learning/k_means_clust.py)
+ * [K Nearest Neighbours](machine_learning/k_nearest_neighbours.py)
+ * [Linear Discriminant Analysis](machine_learning/linear_discriminant_analysis.py)
+ * [Linear Regression](machine_learning/linear_regression.py)
+ * Local Weighted Learning
+ * [Local Weighted Learning](machine_learning/local_weighted_learning/local_weighted_learning.py)
+ * [Logistic Regression](machine_learning/logistic_regression.py)
+ * [Loss Functions](machine_learning/loss_functions.py)
* Lstm
- * [Lstm Prediction](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/lstm/lstm_prediction.py)
- * [Multilayer Perceptron Classifier](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/multilayer_perceptron_classifier.py)
- * [Polymonial Regression](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/polymonial_regression.py)
- * [Random Forest Classifier](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/random_forest_classifier.py)
- * [Random Forest Regressor](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/random_forest_regressor.py)
- * [Scoring Functions](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/scoring_functions.py)
- * [Sequential Minimum Optimization](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/sequential_minimum_optimization.py)
- * [Similarity Search](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/similarity_search.py)
- * [Support Vector Machines](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/support_vector_machines.py)
- * [Word Frequency Functions](https://github.com/TheAlgorithms/Python/blob/master/machine_learning/word_frequency_functions.py)
+ * [Lstm Prediction](machine_learning/lstm/lstm_prediction.py)
+ * [Mfcc](machine_learning/mfcc.py)
+ * [Multilayer Perceptron Classifier](machine_learning/multilayer_perceptron_classifier.py)
+ * [Polynomial Regression](machine_learning/polynomial_regression.py)
+ * [Principle Component Analysis](machine_learning/principle_component_analysis.py)
+ * [Scoring Functions](machine_learning/scoring_functions.py)
+ * [Self Organizing Map](machine_learning/self_organizing_map.py)
+ * [Sequential Minimum Optimization](machine_learning/sequential_minimum_optimization.py)
+ * [Similarity Search](machine_learning/similarity_search.py)
+ * [Support Vector Machines](machine_learning/support_vector_machines.py)
+ * [Word Frequency Functions](machine_learning/word_frequency_functions.py)
+ * [Xgboost Classifier](machine_learning/xgboost_classifier.py)
+ * [Xgboost Regressor](machine_learning/xgboost_regressor.py)
## Maths
- * [3N Plus 1](https://github.com/TheAlgorithms/Python/blob/master/maths/3n_plus_1.py)
- * [Abs](https://github.com/TheAlgorithms/Python/blob/master/maths/abs.py)
- * [Abs Max](https://github.com/TheAlgorithms/Python/blob/master/maths/abs_max.py)
- * [Abs Min](https://github.com/TheAlgorithms/Python/blob/master/maths/abs_min.py)
- * [Add](https://github.com/TheAlgorithms/Python/blob/master/maths/add.py)
- * [Aliquot Sum](https://github.com/TheAlgorithms/Python/blob/master/maths/aliquot_sum.py)
- * [Allocation Number](https://github.com/TheAlgorithms/Python/blob/master/maths/allocation_number.py)
- * [Area](https://github.com/TheAlgorithms/Python/blob/master/maths/area.py)
- * [Area Under Curve](https://github.com/TheAlgorithms/Python/blob/master/maths/area_under_curve.py)
- * [Armstrong Numbers](https://github.com/TheAlgorithms/Python/blob/master/maths/armstrong_numbers.py)
- * [Average Mean](https://github.com/TheAlgorithms/Python/blob/master/maths/average_mean.py)
- * [Average Median](https://github.com/TheAlgorithms/Python/blob/master/maths/average_median.py)
- * [Average Mode](https://github.com/TheAlgorithms/Python/blob/master/maths/average_mode.py)
- * [Bailey Borwein Plouffe](https://github.com/TheAlgorithms/Python/blob/master/maths/bailey_borwein_plouffe.py)
- * [Basic Maths](https://github.com/TheAlgorithms/Python/blob/master/maths/basic_maths.py)
- * [Binary Exp Mod](https://github.com/TheAlgorithms/Python/blob/master/maths/binary_exp_mod.py)
- * [Binary Exponentiation](https://github.com/TheAlgorithms/Python/blob/master/maths/binary_exponentiation.py)
- * [Binomial Coefficient](https://github.com/TheAlgorithms/Python/blob/master/maths/binomial_coefficient.py)
- * [Binomial Distribution](https://github.com/TheAlgorithms/Python/blob/master/maths/binomial_distribution.py)
- * [Bisection](https://github.com/TheAlgorithms/Python/blob/master/maths/bisection.py)
- * [Ceil](https://github.com/TheAlgorithms/Python/blob/master/maths/ceil.py)
- * [Chudnovsky Algorithm](https://github.com/TheAlgorithms/Python/blob/master/maths/chudnovsky_algorithm.py)
- * [Collatz Sequence](https://github.com/TheAlgorithms/Python/blob/master/maths/collatz_sequence.py)
- * [Combinations](https://github.com/TheAlgorithms/Python/blob/master/maths/combinations.py)
- * [Decimal Isolate](https://github.com/TheAlgorithms/Python/blob/master/maths/decimal_isolate.py)
- * [Entropy](https://github.com/TheAlgorithms/Python/blob/master/maths/entropy.py)
- * [Eulers Totient](https://github.com/TheAlgorithms/Python/blob/master/maths/eulers_totient.py)
- * [Explicit Euler](https://github.com/TheAlgorithms/Python/blob/master/maths/explicit_euler.py)
- * [Extended Euclidean Algorithm](https://github.com/TheAlgorithms/Python/blob/master/maths/extended_euclidean_algorithm.py)
- * [Factorial Iterative](https://github.com/TheAlgorithms/Python/blob/master/maths/factorial_iterative.py)
- * [Factorial Python](https://github.com/TheAlgorithms/Python/blob/master/maths/factorial_python.py)
- * [Factorial Recursive](https://github.com/TheAlgorithms/Python/blob/master/maths/factorial_recursive.py)
- * [Factors](https://github.com/TheAlgorithms/Python/blob/master/maths/factors.py)
- * [Fermat Little Theorem](https://github.com/TheAlgorithms/Python/blob/master/maths/fermat_little_theorem.py)
- * [Fibonacci](https://github.com/TheAlgorithms/Python/blob/master/maths/fibonacci.py)
- * [Fibonacci Sequence Recursion](https://github.com/TheAlgorithms/Python/blob/master/maths/fibonacci_sequence_recursion.py)
- * [Find Max](https://github.com/TheAlgorithms/Python/blob/master/maths/find_max.py)
- * [Find Max Recursion](https://github.com/TheAlgorithms/Python/blob/master/maths/find_max_recursion.py)
- * [Find Min](https://github.com/TheAlgorithms/Python/blob/master/maths/find_min.py)
- * [Find Min Recursion](https://github.com/TheAlgorithms/Python/blob/master/maths/find_min_recursion.py)
- * [Floor](https://github.com/TheAlgorithms/Python/blob/master/maths/floor.py)
- * [Gamma](https://github.com/TheAlgorithms/Python/blob/master/maths/gamma.py)
- * [Gaussian](https://github.com/TheAlgorithms/Python/blob/master/maths/gaussian.py)
- * [Greatest Common Divisor](https://github.com/TheAlgorithms/Python/blob/master/maths/greatest_common_divisor.py)
- * [Hardy Ramanujanalgo](https://github.com/TheAlgorithms/Python/blob/master/maths/hardy_ramanujanalgo.py)
- * [Is Square Free](https://github.com/TheAlgorithms/Python/blob/master/maths/is_square_free.py)
- * [Jaccard Similarity](https://github.com/TheAlgorithms/Python/blob/master/maths/jaccard_similarity.py)
- * [Kadanes](https://github.com/TheAlgorithms/Python/blob/master/maths/kadanes.py)
- * [Karatsuba](https://github.com/TheAlgorithms/Python/blob/master/maths/karatsuba.py)
- * [Krishnamurthy Number](https://github.com/TheAlgorithms/Python/blob/master/maths/krishnamurthy_number.py)
- * [Kth Lexicographic Permutation](https://github.com/TheAlgorithms/Python/blob/master/maths/kth_lexicographic_permutation.py)
- * [Largest Of Very Large Numbers](https://github.com/TheAlgorithms/Python/blob/master/maths/largest_of_very_large_numbers.py)
- * [Least Common Multiple](https://github.com/TheAlgorithms/Python/blob/master/maths/least_common_multiple.py)
- * [Line Length](https://github.com/TheAlgorithms/Python/blob/master/maths/line_length.py)
- * [Lucas Lehmer Primality Test](https://github.com/TheAlgorithms/Python/blob/master/maths/lucas_lehmer_primality_test.py)
- * [Lucas Series](https://github.com/TheAlgorithms/Python/blob/master/maths/lucas_series.py)
- * [Matrix Exponentiation](https://github.com/TheAlgorithms/Python/blob/master/maths/matrix_exponentiation.py)
- * [Miller Rabin](https://github.com/TheAlgorithms/Python/blob/master/maths/miller_rabin.py)
- * [Mobius Function](https://github.com/TheAlgorithms/Python/blob/master/maths/mobius_function.py)
- * [Modular Exponential](https://github.com/TheAlgorithms/Python/blob/master/maths/modular_exponential.py)
- * [Monte Carlo](https://github.com/TheAlgorithms/Python/blob/master/maths/monte_carlo.py)
- * [Monte Carlo Dice](https://github.com/TheAlgorithms/Python/blob/master/maths/monte_carlo_dice.py)
- * [Newton Raphson](https://github.com/TheAlgorithms/Python/blob/master/maths/newton_raphson.py)
- * [Number Of Digits](https://github.com/TheAlgorithms/Python/blob/master/maths/number_of_digits.py)
- * [Numerical Integration](https://github.com/TheAlgorithms/Python/blob/master/maths/numerical_integration.py)
- * [Perfect Cube](https://github.com/TheAlgorithms/Python/blob/master/maths/perfect_cube.py)
- * [Perfect Number](https://github.com/TheAlgorithms/Python/blob/master/maths/perfect_number.py)
- * [Perfect Square](https://github.com/TheAlgorithms/Python/blob/master/maths/perfect_square.py)
- * [Pi Monte Carlo Estimation](https://github.com/TheAlgorithms/Python/blob/master/maths/pi_monte_carlo_estimation.py)
- * [Polynomial Evaluation](https://github.com/TheAlgorithms/Python/blob/master/maths/polynomial_evaluation.py)
- * [Power Using Recursion](https://github.com/TheAlgorithms/Python/blob/master/maths/power_using_recursion.py)
- * [Prime Check](https://github.com/TheAlgorithms/Python/blob/master/maths/prime_check.py)
- * [Prime Factors](https://github.com/TheAlgorithms/Python/blob/master/maths/prime_factors.py)
- * [Prime Numbers](https://github.com/TheAlgorithms/Python/blob/master/maths/prime_numbers.py)
- * [Prime Sieve Eratosthenes](https://github.com/TheAlgorithms/Python/blob/master/maths/prime_sieve_eratosthenes.py)
- * [Pythagoras](https://github.com/TheAlgorithms/Python/blob/master/maths/pythagoras.py)
- * [Qr Decomposition](https://github.com/TheAlgorithms/Python/blob/master/maths/qr_decomposition.py)
- * [Quadratic Equations Complex Numbers](https://github.com/TheAlgorithms/Python/blob/master/maths/quadratic_equations_complex_numbers.py)
- * [Radians](https://github.com/TheAlgorithms/Python/blob/master/maths/radians.py)
- * [Radix2 Fft](https://github.com/TheAlgorithms/Python/blob/master/maths/radix2_fft.py)
- * [Relu](https://github.com/TheAlgorithms/Python/blob/master/maths/relu.py)
- * [Runge Kutta](https://github.com/TheAlgorithms/Python/blob/master/maths/runge_kutta.py)
- * [Segmented Sieve](https://github.com/TheAlgorithms/Python/blob/master/maths/segmented_sieve.py)
+ * [Abs](maths/abs.py)
+ * [Addition Without Arithmetic](maths/addition_without_arithmetic.py)
+ * [Aliquot Sum](maths/aliquot_sum.py)
+ * [Allocation Number](maths/allocation_number.py)
+ * [Arc Length](maths/arc_length.py)
+ * [Area](maths/area.py)
+ * [Area Under Curve](maths/area_under_curve.py)
+ * [Average Absolute Deviation](maths/average_absolute_deviation.py)
+ * [Average Mean](maths/average_mean.py)
+ * [Average Median](maths/average_median.py)
+ * [Average Mode](maths/average_mode.py)
+ * [Bailey Borwein Plouffe](maths/bailey_borwein_plouffe.py)
+ * [Base Neg2 Conversion](maths/base_neg2_conversion.py)
+ * [Basic Maths](maths/basic_maths.py)
+ * [Binary Exponentiation](maths/binary_exponentiation.py)
+ * [Binary Multiplication](maths/binary_multiplication.py)
+ * [Binomial Coefficient](maths/binomial_coefficient.py)
+ * [Binomial Distribution](maths/binomial_distribution.py)
+ * [Ceil](maths/ceil.py)
+ * [Chebyshev Distance](maths/chebyshev_distance.py)
+ * [Check Polygon](maths/check_polygon.py)
+ * [Chinese Remainder Theorem](maths/chinese_remainder_theorem.py)
+ * [Chudnovsky Algorithm](maths/chudnovsky_algorithm.py)
+ * [Collatz Sequence](maths/collatz_sequence.py)
+ * [Combinations](maths/combinations.py)
+ * [Continued Fraction](maths/continued_fraction.py)
+ * [Decimal Isolate](maths/decimal_isolate.py)
+ * [Decimal To Fraction](maths/decimal_to_fraction.py)
+ * [Dodecahedron](maths/dodecahedron.py)
+ * [Double Factorial](maths/double_factorial.py)
+ * [Dual Number Automatic Differentiation](maths/dual_number_automatic_differentiation.py)
+ * [Entropy](maths/entropy.py)
+ * [Euclidean Distance](maths/euclidean_distance.py)
+ * [Euler Method](maths/euler_method.py)
+ * [Euler Modified](maths/euler_modified.py)
+ * [Eulers Totient](maths/eulers_totient.py)
+ * [Extended Euclidean Algorithm](maths/extended_euclidean_algorithm.py)
+ * [Factorial](maths/factorial.py)
+ * [Factors](maths/factors.py)
+ * [Fast Inverse Sqrt](maths/fast_inverse_sqrt.py)
+ * [Fermat Little Theorem](maths/fermat_little_theorem.py)
+ * [Fibonacci](maths/fibonacci.py)
+ * [Find Max](maths/find_max.py)
+ * [Find Min](maths/find_min.py)
+ * [Floor](maths/floor.py)
+ * [Gamma](maths/gamma.py)
+ * [Gaussian](maths/gaussian.py)
+ * [Gcd Of N Numbers](maths/gcd_of_n_numbers.py)
+ * [Geometric Mean](maths/geometric_mean.py)
+ * [Germain Primes](maths/germain_primes.py)
+ * [Greatest Common Divisor](maths/greatest_common_divisor.py)
+ * [Hardy Ramanujanalgo](maths/hardy_ramanujanalgo.py)
+ * [Integer Square Root](maths/integer_square_root.py)
+ * [Interquartile Range](maths/interquartile_range.py)
+ * [Is Int Palindrome](maths/is_int_palindrome.py)
+ * [Is Ip V4 Address Valid](maths/is_ip_v4_address_valid.py)
+ * [Is Square Free](maths/is_square_free.py)
+ * [Jaccard Similarity](maths/jaccard_similarity.py)
+ * [Joint Probability Distribution](maths/joint_probability_distribution.py)
+ * [Josephus Problem](maths/josephus_problem.py)
+ * [Juggler Sequence](maths/juggler_sequence.py)
+ * [Karatsuba](maths/karatsuba.py)
+ * [Kth Lexicographic Permutation](maths/kth_lexicographic_permutation.py)
+ * [Largest Of Very Large Numbers](maths/largest_of_very_large_numbers.py)
+ * [Least Common Multiple](maths/least_common_multiple.py)
+ * [Line Length](maths/line_length.py)
+ * [Liouville Lambda](maths/liouville_lambda.py)
+ * [Lucas Lehmer Primality Test](maths/lucas_lehmer_primality_test.py)
+ * [Lucas Series](maths/lucas_series.py)
+ * [Maclaurin Series](maths/maclaurin_series.py)
+ * [Manhattan Distance](maths/manhattan_distance.py)
+ * [Matrix Exponentiation](maths/matrix_exponentiation.py)
+ * [Max Sum Sliding Window](maths/max_sum_sliding_window.py)
+ * [Minkowski Distance](maths/minkowski_distance.py)
+ * [Mobius Function](maths/mobius_function.py)
+ * [Modular Division](maths/modular_division.py)
+ * [Modular Exponential](maths/modular_exponential.py)
+ * [Monte Carlo](maths/monte_carlo.py)
+ * [Monte Carlo Dice](maths/monte_carlo_dice.py)
+ * [Number Of Digits](maths/number_of_digits.py)
+ * Numerical Analysis
+ * [Adams Bashforth](maths/numerical_analysis/adams_bashforth.py)
+ * [Bisection](maths/numerical_analysis/bisection.py)
+ * [Bisection 2](maths/numerical_analysis/bisection_2.py)
+ * [Integration By Simpson Approx](maths/numerical_analysis/integration_by_simpson_approx.py)
+ * [Intersection](maths/numerical_analysis/intersection.py)
+ * [Nevilles Method](maths/numerical_analysis/nevilles_method.py)
+ * [Newton Forward Interpolation](maths/numerical_analysis/newton_forward_interpolation.py)
+ * [Newton Raphson](maths/numerical_analysis/newton_raphson.py)
+ * [Numerical Integration](maths/numerical_analysis/numerical_integration.py)
+ * [Proper Fractions](maths/numerical_analysis/proper_fractions.py)
+ * [Runge Kutta](maths/numerical_analysis/runge_kutta.py)
+ * [Runge Kutta Fehlberg 45](maths/numerical_analysis/runge_kutta_fehlberg_45.py)
+ * [Runge Kutta Gills](maths/numerical_analysis/runge_kutta_gills.py)
+ * [Secant Method](maths/numerical_analysis/secant_method.py)
+ * [Simpson Rule](maths/numerical_analysis/simpson_rule.py)
+ * [Square Root](maths/numerical_analysis/square_root.py)
+ * [Odd Sieve](maths/odd_sieve.py)
+ * [Perfect Cube](maths/perfect_cube.py)
+ * [Perfect Number](maths/perfect_number.py)
+ * [Perfect Square](maths/perfect_square.py)
+ * [Persistence](maths/persistence.py)
+ * [Pi Generator](maths/pi_generator.py)
+ * [Pi Monte Carlo Estimation](maths/pi_monte_carlo_estimation.py)
+ * [Points Are Collinear 3D](maths/points_are_collinear_3d.py)
+ * [Pollard Rho](maths/pollard_rho.py)
+ * [Polynomial Evaluation](maths/polynomial_evaluation.py)
+ * Polynomials
+ * [Single Indeterminate Operations](maths/polynomials/single_indeterminate_operations.py)
+ * [Power Using Recursion](maths/power_using_recursion.py)
+ * [Prime Check](maths/prime_check.py)
+ * [Prime Factors](maths/prime_factors.py)
+ * [Prime Numbers](maths/prime_numbers.py)
+ * [Prime Sieve Eratosthenes](maths/prime_sieve_eratosthenes.py)
+ * [Primelib](maths/primelib.py)
+ * [Print Multiplication Table](maths/print_multiplication_table.py)
+ * [Pythagoras](maths/pythagoras.py)
+ * [Qr Decomposition](maths/qr_decomposition.py)
+ * [Quadratic Equations Complex Numbers](maths/quadratic_equations_complex_numbers.py)
+ * [Radians](maths/radians.py)
+ * [Radix2 Fft](maths/radix2_fft.py)
+ * [Remove Digit](maths/remove_digit.py)
+ * [Segmented Sieve](maths/segmented_sieve.py)
* Series
- * [Geometric Series](https://github.com/TheAlgorithms/Python/blob/master/maths/series/geometric_series.py)
- * [Harmonic Series](https://github.com/TheAlgorithms/Python/blob/master/maths/series/harmonic_series.py)
- * [P Series](https://github.com/TheAlgorithms/Python/blob/master/maths/series/p_series.py)
- * [Sieve Of Eratosthenes](https://github.com/TheAlgorithms/Python/blob/master/maths/sieve_of_eratosthenes.py)
- * [Sigmoid](https://github.com/TheAlgorithms/Python/blob/master/maths/sigmoid.py)
- * [Simpson Rule](https://github.com/TheAlgorithms/Python/blob/master/maths/simpson_rule.py)
- * [Softmax](https://github.com/TheAlgorithms/Python/blob/master/maths/softmax.py)
- * [Square Root](https://github.com/TheAlgorithms/Python/blob/master/maths/square_root.py)
- * [Sum Of Arithmetic Series](https://github.com/TheAlgorithms/Python/blob/master/maths/sum_of_arithmetic_series.py)
- * [Sum Of Digits](https://github.com/TheAlgorithms/Python/blob/master/maths/sum_of_digits.py)
- * [Sum Of Geometric Progression](https://github.com/TheAlgorithms/Python/blob/master/maths/sum_of_geometric_progression.py)
- * [Test Prime Check](https://github.com/TheAlgorithms/Python/blob/master/maths/test_prime_check.py)
- * [Trapezoidal Rule](https://github.com/TheAlgorithms/Python/blob/master/maths/trapezoidal_rule.py)
- * [Ugly Numbers](https://github.com/TheAlgorithms/Python/blob/master/maths/ugly_numbers.py)
- * [Volume](https://github.com/TheAlgorithms/Python/blob/master/maths/volume.py)
- * [Zellers Congruence](https://github.com/TheAlgorithms/Python/blob/master/maths/zellers_congruence.py)
+ * [Arithmetic](maths/series/arithmetic.py)
+ * [Geometric](maths/series/geometric.py)
+ * [Geometric Series](maths/series/geometric_series.py)
+ * [Harmonic](maths/series/harmonic.py)
+ * [Harmonic Series](maths/series/harmonic_series.py)
+ * [Hexagonal Numbers](maths/series/hexagonal_numbers.py)
+ * [P Series](maths/series/p_series.py)
+ * [Sieve Of Eratosthenes](maths/sieve_of_eratosthenes.py)
+ * [Sigmoid](maths/sigmoid.py)
+ * [Signum](maths/signum.py)
+ * [Simultaneous Linear Equation Solver](maths/simultaneous_linear_equation_solver.py)
+ * [Sin](maths/sin.py)
+ * [Sock Merchant](maths/sock_merchant.py)
+ * [Softmax](maths/softmax.py)
+ * [Solovay Strassen Primality Test](maths/solovay_strassen_primality_test.py)
+ * [Spearman Rank Correlation Coefficient](maths/spearman_rank_correlation_coefficient.py)
+ * Special Numbers
+ * [Armstrong Numbers](maths/special_numbers/armstrong_numbers.py)
+ * [Automorphic Number](maths/special_numbers/automorphic_number.py)
+ * [Bell Numbers](maths/special_numbers/bell_numbers.py)
+ * [Carmichael Number](maths/special_numbers/carmichael_number.py)
+ * [Catalan Number](maths/special_numbers/catalan_number.py)
+ * [Hamming Numbers](maths/special_numbers/hamming_numbers.py)
+ * [Happy Number](maths/special_numbers/happy_number.py)
+ * [Harshad Numbers](maths/special_numbers/harshad_numbers.py)
+ * [Hexagonal Number](maths/special_numbers/hexagonal_number.py)
+ * [Krishnamurthy Number](maths/special_numbers/krishnamurthy_number.py)
+ * [Perfect Number](maths/special_numbers/perfect_number.py)
+ * [Polygonal Numbers](maths/special_numbers/polygonal_numbers.py)
+ * [Pronic Number](maths/special_numbers/pronic_number.py)
+ * [Proth Number](maths/special_numbers/proth_number.py)
+ * [Triangular Numbers](maths/special_numbers/triangular_numbers.py)
+ * [Ugly Numbers](maths/special_numbers/ugly_numbers.py)
+ * [Weird Number](maths/special_numbers/weird_number.py)
+ * [Sum Of Arithmetic Series](maths/sum_of_arithmetic_series.py)
+ * [Sum Of Digits](maths/sum_of_digits.py)
+ * [Sum Of Geometric Progression](maths/sum_of_geometric_progression.py)
+ * [Sum Of Harmonic Series](maths/sum_of_harmonic_series.py)
+ * [Sumset](maths/sumset.py)
+ * [Sylvester Sequence](maths/sylvester_sequence.py)
+ * [Tanh](maths/tanh.py)
+ * [Test Prime Check](maths/test_prime_check.py)
+ * [Three Sum](maths/three_sum.py)
+ * [Trapezoidal Rule](maths/trapezoidal_rule.py)
+ * [Triplet Sum](maths/triplet_sum.py)
+ * [Twin Prime](maths/twin_prime.py)
+ * [Two Pointer](maths/two_pointer.py)
+ * [Two Sum](maths/two_sum.py)
+ * [Volume](maths/volume.py)
+ * [Zellers Congruence](maths/zellers_congruence.py)
## Matrix
- * [Count Islands In Matrix](https://github.com/TheAlgorithms/Python/blob/master/matrix/count_islands_in_matrix.py)
- * [Inverse Of Matrix](https://github.com/TheAlgorithms/Python/blob/master/matrix/inverse_of_matrix.py)
- * [Matrix Class](https://github.com/TheAlgorithms/Python/blob/master/matrix/matrix_class.py)
- * [Matrix Operation](https://github.com/TheAlgorithms/Python/blob/master/matrix/matrix_operation.py)
- * [Nth Fibonacci Using Matrix Exponentiation](https://github.com/TheAlgorithms/Python/blob/master/matrix/nth_fibonacci_using_matrix_exponentiation.py)
- * [Rotate Matrix](https://github.com/TheAlgorithms/Python/blob/master/matrix/rotate_matrix.py)
- * [Searching In Sorted Matrix](https://github.com/TheAlgorithms/Python/blob/master/matrix/searching_in_sorted_matrix.py)
- * [Sherman Morrison](https://github.com/TheAlgorithms/Python/blob/master/matrix/sherman_morrison.py)
- * [Spiral Print](https://github.com/TheAlgorithms/Python/blob/master/matrix/spiral_print.py)
+ * [Binary Search Matrix](matrix/binary_search_matrix.py)
+ * [Count Islands In Matrix](matrix/count_islands_in_matrix.py)
+ * [Count Negative Numbers In Sorted Matrix](matrix/count_negative_numbers_in_sorted_matrix.py)
+ * [Count Paths](matrix/count_paths.py)
+ * [Cramers Rule 2X2](matrix/cramers_rule_2x2.py)
+ * [Inverse Of Matrix](matrix/inverse_of_matrix.py)
+ * [Largest Square Area In Matrix](matrix/largest_square_area_in_matrix.py)
+ * [Matrix Based Game](matrix/matrix_based_game.py)
+ * [Matrix Class](matrix/matrix_class.py)
+ * [Matrix Equalization](matrix/matrix_equalization.py)
+ * [Matrix Multiplication Recursion](matrix/matrix_multiplication_recursion.py)
+ * [Matrix Operation](matrix/matrix_operation.py)
+ * [Max Area Of Island](matrix/max_area_of_island.py)
+ * [Median Matrix](matrix/median_matrix.py)
+ * [Nth Fibonacci Using Matrix Exponentiation](matrix/nth_fibonacci_using_matrix_exponentiation.py)
+ * [Pascal Triangle](matrix/pascal_triangle.py)
+ * [Rotate Matrix](matrix/rotate_matrix.py)
+ * [Searching In Sorted Matrix](matrix/searching_in_sorted_matrix.py)
+ * [Sherman Morrison](matrix/sherman_morrison.py)
+ * [Spiral Print](matrix/spiral_print.py)
* Tests
- * [Test Matrix Operation](https://github.com/TheAlgorithms/Python/blob/master/matrix/tests/test_matrix_operation.py)
+ * [Test Matrix Operation](matrix/tests/test_matrix_operation.py)
+ * [Validate Sudoku Board](matrix/validate_sudoku_board.py)
## Networking Flow
- * [Ford Fulkerson](https://github.com/TheAlgorithms/Python/blob/master/networking_flow/ford_fulkerson.py)
- * [Minimum Cut](https://github.com/TheAlgorithms/Python/blob/master/networking_flow/minimum_cut.py)
+ * [Ford Fulkerson](networking_flow/ford_fulkerson.py)
+ * [Minimum Cut](networking_flow/minimum_cut.py)
## Neural Network
- * [Back Propagation Neural Network](https://github.com/TheAlgorithms/Python/blob/master/neural_network/back_propagation_neural_network.py)
- * [Convolution Neural Network](https://github.com/TheAlgorithms/Python/blob/master/neural_network/convolution_neural_network.py)
- * [Perceptron](https://github.com/TheAlgorithms/Python/blob/master/neural_network/perceptron.py)
+ * Activation Functions
+ * [Binary Step](neural_network/activation_functions/binary_step.py)
+ * [Exponential Linear Unit](neural_network/activation_functions/exponential_linear_unit.py)
+ * [Gaussian Error Linear Unit](neural_network/activation_functions/gaussian_error_linear_unit.py)
+ * [Leaky Rectified Linear Unit](neural_network/activation_functions/leaky_rectified_linear_unit.py)
+ * [Mish](neural_network/activation_functions/mish.py)
+ * [Rectified Linear Unit](neural_network/activation_functions/rectified_linear_unit.py)
+ * [Scaled Exponential Linear Unit](neural_network/activation_functions/scaled_exponential_linear_unit.py)
+ * [Soboleva Modified Hyperbolic Tangent](neural_network/activation_functions/soboleva_modified_hyperbolic_tangent.py)
+ * [Softplus](neural_network/activation_functions/softplus.py)
+ * [Squareplus](neural_network/activation_functions/squareplus.py)
+ * [Swish](neural_network/activation_functions/swish.py)
+ * [Back Propagation Neural Network](neural_network/back_propagation_neural_network.py)
+ * [Convolution Neural Network](neural_network/convolution_neural_network.py)
+ * [Input Data](neural_network/input_data.py)
+ * [Simple Neural Network](neural_network/simple_neural_network.py)
+ * [Two Hidden Layers Neural Network](neural_network/two_hidden_layers_neural_network.py)
## Other
- * [Activity Selection](https://github.com/TheAlgorithms/Python/blob/master/other/activity_selection.py)
- * [Anagrams](https://github.com/TheAlgorithms/Python/blob/master/other/anagrams.py)
- * [Autocomplete Using Trie](https://github.com/TheAlgorithms/Python/blob/master/other/autocomplete_using_trie.py)
- * [Binary Exponentiation](https://github.com/TheAlgorithms/Python/blob/master/other/binary_exponentiation.py)
- * [Binary Exponentiation 2](https://github.com/TheAlgorithms/Python/blob/master/other/binary_exponentiation_2.py)
- * [Davis–Putnam–Logemann–Loveland](https://github.com/TheAlgorithms/Python/blob/master/other/davis–putnam–logemann–loveland.py)
- * [Detecting English Programmatically](https://github.com/TheAlgorithms/Python/blob/master/other/detecting_english_programmatically.py)
- * [Dijkstra Bankers Algorithm](https://github.com/TheAlgorithms/Python/blob/master/other/dijkstra_bankers_algorithm.py)
- * [Doomsday](https://github.com/TheAlgorithms/Python/blob/master/other/doomsday.py)
- * [Euclidean Gcd](https://github.com/TheAlgorithms/Python/blob/master/other/euclidean_gcd.py)
- * [Fischer Yates Shuffle](https://github.com/TheAlgorithms/Python/blob/master/other/fischer_yates_shuffle.py)
- * [Frequency Finder](https://github.com/TheAlgorithms/Python/blob/master/other/frequency_finder.py)
- * [Game Of Life](https://github.com/TheAlgorithms/Python/blob/master/other/game_of_life.py)
- * [Gauss Easter](https://github.com/TheAlgorithms/Python/blob/master/other/gauss_easter.py)
- * [Greedy](https://github.com/TheAlgorithms/Python/blob/master/other/greedy.py)
- * [Integeration By Simpson Approx](https://github.com/TheAlgorithms/Python/blob/master/other/integeration_by_simpson_approx.py)
- * [Largest Subarray Sum](https://github.com/TheAlgorithms/Python/blob/master/other/largest_subarray_sum.py)
- * [Least Recently Used](https://github.com/TheAlgorithms/Python/blob/master/other/least_recently_used.py)
- * [Lfu Cache](https://github.com/TheAlgorithms/Python/blob/master/other/lfu_cache.py)
- * [Linear Congruential Generator](https://github.com/TheAlgorithms/Python/blob/master/other/linear_congruential_generator.py)
- * [Lru Cache](https://github.com/TheAlgorithms/Python/blob/master/other/lru_cache.py)
- * [Magicdiamondpattern](https://github.com/TheAlgorithms/Python/blob/master/other/magicdiamondpattern.py)
- * [Markov Chain](https://github.com/TheAlgorithms/Python/blob/master/other/markov_chain.py)
- * [Max Sum Sliding Window](https://github.com/TheAlgorithms/Python/blob/master/other/max_sum_sliding_window.py)
- * [Median Of Two Arrays](https://github.com/TheAlgorithms/Python/blob/master/other/median_of_two_arrays.py)
- * [Nested Brackets](https://github.com/TheAlgorithms/Python/blob/master/other/nested_brackets.py)
- * [Palindrome](https://github.com/TheAlgorithms/Python/blob/master/other/palindrome.py)
- * [Password Generator](https://github.com/TheAlgorithms/Python/blob/master/other/password_generator.py)
- * [Primelib](https://github.com/TheAlgorithms/Python/blob/master/other/primelib.py)
- * [Scoring Algorithm](https://github.com/TheAlgorithms/Python/blob/master/other/scoring_algorithm.py)
- * [Sdes](https://github.com/TheAlgorithms/Python/blob/master/other/sdes.py)
- * [Sierpinski Triangle](https://github.com/TheAlgorithms/Python/blob/master/other/sierpinski_triangle.py)
- * [Tower Of Hanoi](https://github.com/TheAlgorithms/Python/blob/master/other/tower_of_hanoi.py)
- * [Triplet Sum](https://github.com/TheAlgorithms/Python/blob/master/other/triplet_sum.py)
- * [Two Pointer](https://github.com/TheAlgorithms/Python/blob/master/other/two_pointer.py)
- * [Two Sum](https://github.com/TheAlgorithms/Python/blob/master/other/two_sum.py)
- * [Word Patterns](https://github.com/TheAlgorithms/Python/blob/master/other/word_patterns.py)
+ * [Activity Selection](other/activity_selection.py)
+ * [Alternative List Arrange](other/alternative_list_arrange.py)
+ * [Bankers Algorithm](other/bankers_algorithm.py)
+ * [Davis Putnam Logemann Loveland](other/davis_putnam_logemann_loveland.py)
+ * [Doomsday](other/doomsday.py)
+ * [Fischer Yates Shuffle](other/fischer_yates_shuffle.py)
+ * [Gauss Easter](other/gauss_easter.py)
+ * [Graham Scan](other/graham_scan.py)
+ * [Greedy](other/greedy.py)
+ * [Guess The Number Search](other/guess_the_number_search.py)
+ * [H Index](other/h_index.py)
+ * [Least Recently Used](other/least_recently_used.py)
+ * [Lfu Cache](other/lfu_cache.py)
+ * [Linear Congruential Generator](other/linear_congruential_generator.py)
+ * [Lru Cache](other/lru_cache.py)
+ * [Magicdiamondpattern](other/magicdiamondpattern.py)
+ * [Majority Vote Algorithm](other/majority_vote_algorithm.py)
+ * [Maximum Subsequence](other/maximum_subsequence.py)
+ * [Nested Brackets](other/nested_brackets.py)
+ * [Number Container System](other/number_container_system.py)
+ * [Password](other/password.py)
+ * [Quine](other/quine.py)
+ * [Scoring Algorithm](other/scoring_algorithm.py)
+ * [Sdes](other/sdes.py)
+ * [Tower Of Hanoi](other/tower_of_hanoi.py)
+ * [Word Search](other/word_search.py)
+
+## Physics
+ * [Altitude Pressure](physics/altitude_pressure.py)
+ * [Archimedes Principle Of Buoyant Force](physics/archimedes_principle_of_buoyant_force.py)
+ * [Basic Orbital Capture](physics/basic_orbital_capture.py)
+ * [Casimir Effect](physics/casimir_effect.py)
+ * [Center Of Mass](physics/center_of_mass.py)
+ * [Centripetal Force](physics/centripetal_force.py)
+ * [Coulombs Law](physics/coulombs_law.py)
+ * [Doppler Frequency](physics/doppler_frequency.py)
+ * [Grahams Law](physics/grahams_law.py)
+ * [Horizontal Projectile Motion](physics/horizontal_projectile_motion.py)
+ * [Hubble Parameter](physics/hubble_parameter.py)
+ * [Ideal Gas Law](physics/ideal_gas_law.py)
+ * [In Static Equilibrium](physics/in_static_equilibrium.py)
+ * [Kinetic Energy](physics/kinetic_energy.py)
+ * [Lens Formulae](physics/lens_formulae.py)
+ * [Lorentz Transformation Four Vector](physics/lorentz_transformation_four_vector.py)
+ * [Malus Law](physics/malus_law.py)
+ * [Mass Energy Equivalence](physics/mass_energy_equivalence.py)
+ * [Mirror Formulae](physics/mirror_formulae.py)
+ * [N Body Simulation](physics/n_body_simulation.py)
+ * [Newtons Law Of Gravitation](physics/newtons_law_of_gravitation.py)
+ * [Newtons Second Law Of Motion](physics/newtons_second_law_of_motion.py)
+ * [Period Of Pendulum](physics/period_of_pendulum.py)
+ * [Photoelectric Effect](physics/photoelectric_effect.py)
+ * [Potential Energy](physics/potential_energy.py)
+ * [Rainfall Intensity](physics/rainfall_intensity.py)
+ * [Reynolds Number](physics/reynolds_number.py)
+ * [Rms Speed Of Molecule](physics/rms_speed_of_molecule.py)
+ * [Shear Stress](physics/shear_stress.py)
+ * [Speed Of Sound](physics/speed_of_sound.py)
+ * [Speeds Of Gas Molecules](physics/speeds_of_gas_molecules.py)
+ * [Terminal Velocity](physics/terminal_velocity.py)
## Project Euler
* Problem 001
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol3.py)
- * [Sol4](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol4.py)
- * [Sol5](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol5.py)
- * [Sol6](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol6.py)
- * [Sol7](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_001/sol7.py)
+ * [Sol1](project_euler/problem_001/sol1.py)
+ * [Sol2](project_euler/problem_001/sol2.py)
+ * [Sol3](project_euler/problem_001/sol3.py)
+ * [Sol4](project_euler/problem_001/sol4.py)
+ * [Sol5](project_euler/problem_001/sol5.py)
+ * [Sol6](project_euler/problem_001/sol6.py)
+ * [Sol7](project_euler/problem_001/sol7.py)
* Problem 002
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_002/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_002/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_002/sol3.py)
- * [Sol4](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_002/sol4.py)
- * [Sol5](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_002/sol5.py)
+ * [Sol1](project_euler/problem_002/sol1.py)
+ * [Sol2](project_euler/problem_002/sol2.py)
+ * [Sol3](project_euler/problem_002/sol3.py)
+ * [Sol4](project_euler/problem_002/sol4.py)
+ * [Sol5](project_euler/problem_002/sol5.py)
* Problem 003
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_003/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_003/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_003/sol3.py)
+ * [Sol1](project_euler/problem_003/sol1.py)
+ * [Sol2](project_euler/problem_003/sol2.py)
+ * [Sol3](project_euler/problem_003/sol3.py)
* Problem 004
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_004/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_004/sol2.py)
+ * [Sol1](project_euler/problem_004/sol1.py)
+ * [Sol2](project_euler/problem_004/sol2.py)
* Problem 005
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_005/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_005/sol2.py)
+ * [Sol1](project_euler/problem_005/sol1.py)
+ * [Sol2](project_euler/problem_005/sol2.py)
* Problem 006
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_006/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_006/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_006/sol3.py)
- * [Sol4](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_006/sol4.py)
+ * [Sol1](project_euler/problem_006/sol1.py)
+ * [Sol2](project_euler/problem_006/sol2.py)
+ * [Sol3](project_euler/problem_006/sol3.py)
+ * [Sol4](project_euler/problem_006/sol4.py)
* Problem 007
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_007/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_007/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_007/sol3.py)
+ * [Sol1](project_euler/problem_007/sol1.py)
+ * [Sol2](project_euler/problem_007/sol2.py)
+ * [Sol3](project_euler/problem_007/sol3.py)
* Problem 008
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_008/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_008/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_008/sol3.py)
+ * [Sol1](project_euler/problem_008/sol1.py)
+ * [Sol2](project_euler/problem_008/sol2.py)
+ * [Sol3](project_euler/problem_008/sol3.py)
* Problem 009
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_009/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_009/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_009/sol3.py)
+ * [Sol1](project_euler/problem_009/sol1.py)
+ * [Sol2](project_euler/problem_009/sol2.py)
+ * [Sol3](project_euler/problem_009/sol3.py)
* Problem 010
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_010/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_010/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_010/sol3.py)
+ * [Sol1](project_euler/problem_010/sol1.py)
+ * [Sol2](project_euler/problem_010/sol2.py)
+ * [Sol3](project_euler/problem_010/sol3.py)
* Problem 011
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_011/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_011/sol2.py)
+ * [Sol1](project_euler/problem_011/sol1.py)
+ * [Sol2](project_euler/problem_011/sol2.py)
* Problem 012
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_012/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_012/sol2.py)
+ * [Sol1](project_euler/problem_012/sol1.py)
+ * [Sol2](project_euler/problem_012/sol2.py)
* Problem 013
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_013/sol1.py)
+ * [Sol1](project_euler/problem_013/sol1.py)
* Problem 014
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_014/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_014/sol2.py)
+ * [Sol1](project_euler/problem_014/sol1.py)
+ * [Sol2](project_euler/problem_014/sol2.py)
* Problem 015
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_015/sol1.py)
+ * [Sol1](project_euler/problem_015/sol1.py)
* Problem 016
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_016/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_016/sol2.py)
+ * [Sol1](project_euler/problem_016/sol1.py)
+ * [Sol2](project_euler/problem_016/sol2.py)
* Problem 017
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_017/sol1.py)
+ * [Sol1](project_euler/problem_017/sol1.py)
* Problem 018
- * [Solution](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_018/solution.py)
+ * [Solution](project_euler/problem_018/solution.py)
* Problem 019
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_019/sol1.py)
+ * [Sol1](project_euler/problem_019/sol1.py)
* Problem 020
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_020/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_020/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_020/sol3.py)
- * [Sol4](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_020/sol4.py)
+ * [Sol1](project_euler/problem_020/sol1.py)
+ * [Sol2](project_euler/problem_020/sol2.py)
+ * [Sol3](project_euler/problem_020/sol3.py)
+ * [Sol4](project_euler/problem_020/sol4.py)
* Problem 021
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_021/sol1.py)
+ * [Sol1](project_euler/problem_021/sol1.py)
* Problem 022
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_022/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_022/sol2.py)
+ * [Sol1](project_euler/problem_022/sol1.py)
+ * [Sol2](project_euler/problem_022/sol2.py)
* Problem 023
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_023/sol1.py)
+ * [Sol1](project_euler/problem_023/sol1.py)
* Problem 024
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_024/sol1.py)
+ * [Sol1](project_euler/problem_024/sol1.py)
* Problem 025
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_025/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_025/sol2.py)
- * [Sol3](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_025/sol3.py)
+ * [Sol1](project_euler/problem_025/sol1.py)
+ * [Sol2](project_euler/problem_025/sol2.py)
+ * [Sol3](project_euler/problem_025/sol3.py)
* Problem 026
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_026/sol1.py)
+ * [Sol1](project_euler/problem_026/sol1.py)
* Problem 027
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_027/sol1.py)
+ * [Sol1](project_euler/problem_027/sol1.py)
* Problem 028
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_028/sol1.py)
+ * [Sol1](project_euler/problem_028/sol1.py)
* Problem 029
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_029/sol1.py)
+ * [Sol1](project_euler/problem_029/sol1.py)
* Problem 030
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_030/sol1.py)
+ * [Sol1](project_euler/problem_030/sol1.py)
* Problem 031
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_031/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_031/sol2.py)
+ * [Sol1](project_euler/problem_031/sol1.py)
+ * [Sol2](project_euler/problem_031/sol2.py)
* Problem 032
- * [Sol32](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_032/sol32.py)
+ * [Sol32](project_euler/problem_032/sol32.py)
* Problem 033
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_033/sol1.py)
+ * [Sol1](project_euler/problem_033/sol1.py)
* Problem 034
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_034/sol1.py)
+ * [Sol1](project_euler/problem_034/sol1.py)
* Problem 035
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_035/sol1.py)
+ * [Sol1](project_euler/problem_035/sol1.py)
* Problem 036
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_036/sol1.py)
+ * [Sol1](project_euler/problem_036/sol1.py)
* Problem 037
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_037/sol1.py)
+ * [Sol1](project_euler/problem_037/sol1.py)
* Problem 038
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_038/sol1.py)
+ * [Sol1](project_euler/problem_038/sol1.py)
* Problem 039
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_039/sol1.py)
+ * [Sol1](project_euler/problem_039/sol1.py)
* Problem 040
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_040/sol1.py)
+ * [Sol1](project_euler/problem_040/sol1.py)
* Problem 041
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_041/sol1.py)
+ * [Sol1](project_euler/problem_041/sol1.py)
* Problem 042
- * [Solution42](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_042/solution42.py)
+ * [Solution42](project_euler/problem_042/solution42.py)
* Problem 043
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_043/sol1.py)
+ * [Sol1](project_euler/problem_043/sol1.py)
* Problem 044
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_044/sol1.py)
+ * [Sol1](project_euler/problem_044/sol1.py)
* Problem 045
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_045/sol1.py)
+ * [Sol1](project_euler/problem_045/sol1.py)
* Problem 046
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_046/sol1.py)
+ * [Sol1](project_euler/problem_046/sol1.py)
* Problem 047
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_047/sol1.py)
+ * [Sol1](project_euler/problem_047/sol1.py)
* Problem 048
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_048/sol1.py)
+ * [Sol1](project_euler/problem_048/sol1.py)
* Problem 049
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_049/sol1.py)
+ * [Sol1](project_euler/problem_049/sol1.py)
* Problem 050
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_050/sol1.py)
+ * [Sol1](project_euler/problem_050/sol1.py)
* Problem 051
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_051/sol1.py)
+ * [Sol1](project_euler/problem_051/sol1.py)
* Problem 052
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_052/sol1.py)
+ * [Sol1](project_euler/problem_052/sol1.py)
* Problem 053
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_053/sol1.py)
+ * [Sol1](project_euler/problem_053/sol1.py)
* Problem 054
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_054/sol1.py)
- * [Test Poker Hand](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_054/test_poker_hand.py)
+ * [Sol1](project_euler/problem_054/sol1.py)
+ * [Test Poker Hand](project_euler/problem_054/test_poker_hand.py)
* Problem 055
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_055/sol1.py)
+ * [Sol1](project_euler/problem_055/sol1.py)
* Problem 056
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_056/sol1.py)
+ * [Sol1](project_euler/problem_056/sol1.py)
* Problem 057
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_057/sol1.py)
+ * [Sol1](project_euler/problem_057/sol1.py)
* Problem 058
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_058/sol1.py)
+ * [Sol1](project_euler/problem_058/sol1.py)
+ * Problem 059
+ * [Sol1](project_euler/problem_059/sol1.py)
* Problem 062
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_062/sol1.py)
+ * [Sol1](project_euler/problem_062/sol1.py)
* Problem 063
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_063/sol1.py)
+ * [Sol1](project_euler/problem_063/sol1.py)
* Problem 064
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_064/sol1.py)
+ * [Sol1](project_euler/problem_064/sol1.py)
* Problem 065
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_065/sol1.py)
+ * [Sol1](project_euler/problem_065/sol1.py)
* Problem 067
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_067/sol1.py)
+ * [Sol1](project_euler/problem_067/sol1.py)
+ * [Sol2](project_euler/problem_067/sol2.py)
+ * Problem 068
+ * [Sol1](project_euler/problem_068/sol1.py)
* Problem 069
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_069/sol1.py)
+ * [Sol1](project_euler/problem_069/sol1.py)
+ * Problem 070
+ * [Sol1](project_euler/problem_070/sol1.py)
* Problem 071
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_071/sol1.py)
+ * [Sol1](project_euler/problem_071/sol1.py)
* Problem 072
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_072/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_072/sol2.py)
+ * [Sol1](project_euler/problem_072/sol1.py)
+ * [Sol2](project_euler/problem_072/sol2.py)
+ * Problem 073
+ * [Sol1](project_euler/problem_073/sol1.py)
* Problem 074
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_074/sol1.py)
- * [Sol2](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_074/sol2.py)
+ * [Sol1](project_euler/problem_074/sol1.py)
+ * [Sol2](project_euler/problem_074/sol2.py)
* Problem 075
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_075/sol1.py)
+ * [Sol1](project_euler/problem_075/sol1.py)
* Problem 076
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_076/sol1.py)
+ * [Sol1](project_euler/problem_076/sol1.py)
+ * Problem 077
+ * [Sol1](project_euler/problem_077/sol1.py)
+ * Problem 078
+ * [Sol1](project_euler/problem_078/sol1.py)
+ * Problem 079
+ * [Sol1](project_euler/problem_079/sol1.py)
* Problem 080
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_080/sol1.py)
+ * [Sol1](project_euler/problem_080/sol1.py)
* Problem 081
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_081/sol1.py)
+ * [Sol1](project_euler/problem_081/sol1.py)
+ * Problem 082
+ * [Sol1](project_euler/problem_082/sol1.py)
+ * Problem 085
+ * [Sol1](project_euler/problem_085/sol1.py)
+ * Problem 086
+ * [Sol1](project_euler/problem_086/sol1.py)
* Problem 087
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_087/sol1.py)
+ * [Sol1](project_euler/problem_087/sol1.py)
+ * Problem 089
+ * [Sol1](project_euler/problem_089/sol1.py)
* Problem 091
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_091/sol1.py)
+ * [Sol1](project_euler/problem_091/sol1.py)
+ * Problem 092
+ * [Sol1](project_euler/problem_092/sol1.py)
+ * Problem 094
+ * [Sol1](project_euler/problem_094/sol1.py)
* Problem 097
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_097/sol1.py)
+ * [Sol1](project_euler/problem_097/sol1.py)
* Problem 099
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_099/sol1.py)
+ * [Sol1](project_euler/problem_099/sol1.py)
+ * Problem 100
+ * [Sol1](project_euler/problem_100/sol1.py)
+ * Problem 101
+ * [Sol1](project_euler/problem_101/sol1.py)
+ * Problem 102
+ * [Sol1](project_euler/problem_102/sol1.py)
+ * Problem 104
+ * [Sol1](project_euler/problem_104/sol1.py)
+ * Problem 107
+ * [Sol1](project_euler/problem_107/sol1.py)
+ * Problem 109
+ * [Sol1](project_euler/problem_109/sol1.py)
* Problem 112
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_112/sol1.py)
+ * [Sol1](project_euler/problem_112/sol1.py)
* Problem 113
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_113/sol1.py)
+ * [Sol1](project_euler/problem_113/sol1.py)
+ * Problem 114
+ * [Sol1](project_euler/problem_114/sol1.py)
+ * Problem 115
+ * [Sol1](project_euler/problem_115/sol1.py)
+ * Problem 116
+ * [Sol1](project_euler/problem_116/sol1.py)
+ * Problem 117
+ * [Sol1](project_euler/problem_117/sol1.py)
* Problem 119
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_119/sol1.py)
+ * [Sol1](project_euler/problem_119/sol1.py)
* Problem 120
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_120/sol1.py)
+ * [Sol1](project_euler/problem_120/sol1.py)
+ * Problem 121
+ * [Sol1](project_euler/problem_121/sol1.py)
* Problem 123
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_123/sol1.py)
+ * [Sol1](project_euler/problem_123/sol1.py)
* Problem 125
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_125/sol1.py)
+ * [Sol1](project_euler/problem_125/sol1.py)
+ * Problem 129
+ * [Sol1](project_euler/problem_129/sol1.py)
+ * Problem 131
+ * [Sol1](project_euler/problem_131/sol1.py)
+ * Problem 135
+ * [Sol1](project_euler/problem_135/sol1.py)
+ * Problem 144
+ * [Sol1](project_euler/problem_144/sol1.py)
+ * Problem 145
+ * [Sol1](project_euler/problem_145/sol1.py)
* Problem 173
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_173/sol1.py)
+ * [Sol1](project_euler/problem_173/sol1.py)
* Problem 174
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_174/sol1.py)
+ * [Sol1](project_euler/problem_174/sol1.py)
+ * Problem 180
+ * [Sol1](project_euler/problem_180/sol1.py)
+ * Problem 187
+ * [Sol1](project_euler/problem_187/sol1.py)
+ * Problem 188
+ * [Sol1](project_euler/problem_188/sol1.py)
* Problem 191
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_191/sol1.py)
+ * [Sol1](project_euler/problem_191/sol1.py)
* Problem 203
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_203/sol1.py)
+ * [Sol1](project_euler/problem_203/sol1.py)
+ * Problem 205
+ * [Sol1](project_euler/problem_205/sol1.py)
* Problem 206
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_206/sol1.py)
+ * [Sol1](project_euler/problem_206/sol1.py)
* Problem 207
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_207/sol1.py)
+ * [Sol1](project_euler/problem_207/sol1.py)
* Problem 234
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_234/sol1.py)
+ * [Sol1](project_euler/problem_234/sol1.py)
* Problem 301
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_301/sol1.py)
+ * [Sol1](project_euler/problem_301/sol1.py)
+ * Problem 493
+ * [Sol1](project_euler/problem_493/sol1.py)
* Problem 551
- * [Sol1](https://github.com/TheAlgorithms/Python/blob/master/project_euler/problem_551/sol1.py)
+ * [Sol1](project_euler/problem_551/sol1.py)
+ * Problem 587
+ * [Sol1](project_euler/problem_587/sol1.py)
+ * Problem 686
+ * [Sol1](project_euler/problem_686/sol1.py)
+ * Problem 800
+ * [Sol1](project_euler/problem_800/sol1.py)
## Quantum
- * [Deutsch Jozsa](https://github.com/TheAlgorithms/Python/blob/master/quantum/deutsch_jozsa.py)
- * [Half Adder](https://github.com/TheAlgorithms/Python/blob/master/quantum/half_adder.py)
- * [Not Gate](https://github.com/TheAlgorithms/Python/blob/master/quantum/not_gate.py)
- * [Quantum Entanglement](https://github.com/TheAlgorithms/Python/blob/master/quantum/quantum_entanglement.py)
- * [Ripple Adder Classic](https://github.com/TheAlgorithms/Python/blob/master/quantum/ripple_adder_classic.py)
- * [Single Qubit Measure](https://github.com/TheAlgorithms/Python/blob/master/quantum/single_qubit_measure.py)
+ * [Q Fourier Transform](quantum/q_fourier_transform.py)
## Scheduling
- * [First Come First Served](https://github.com/TheAlgorithms/Python/blob/master/scheduling/first_come_first_served.py)
- * [Round Robin](https://github.com/TheAlgorithms/Python/blob/master/scheduling/round_robin.py)
- * [Shortest Job First](https://github.com/TheAlgorithms/Python/blob/master/scheduling/shortest_job_first.py)
+ * [First Come First Served](scheduling/first_come_first_served.py)
+ * [Highest Response Ratio Next](scheduling/highest_response_ratio_next.py)
+ * [Job Sequence With Deadline](scheduling/job_sequence_with_deadline.py)
+ * [Job Sequencing With Deadline](scheduling/job_sequencing_with_deadline.py)
+ * [Multi Level Feedback Queue](scheduling/multi_level_feedback_queue.py)
+ * [Non Preemptive Shortest Job First](scheduling/non_preemptive_shortest_job_first.py)
+ * [Round Robin](scheduling/round_robin.py)
+ * [Shortest Job First](scheduling/shortest_job_first.py)
## Searches
- * [Binary Search](https://github.com/TheAlgorithms/Python/blob/master/searches/binary_search.py)
- * [Double Linear Search](https://github.com/TheAlgorithms/Python/blob/master/searches/double_linear_search.py)
- * [Double Linear Search Recursion](https://github.com/TheAlgorithms/Python/blob/master/searches/double_linear_search_recursion.py)
- * [Fibonacci Search](https://github.com/TheAlgorithms/Python/blob/master/searches/fibonacci_search.py)
- * [Hill Climbing](https://github.com/TheAlgorithms/Python/blob/master/searches/hill_climbing.py)
- * [Interpolation Search](https://github.com/TheAlgorithms/Python/blob/master/searches/interpolation_search.py)
- * [Jump Search](https://github.com/TheAlgorithms/Python/blob/master/searches/jump_search.py)
- * [Linear Search](https://github.com/TheAlgorithms/Python/blob/master/searches/linear_search.py)
- * [Quick Select](https://github.com/TheAlgorithms/Python/blob/master/searches/quick_select.py)
- * [Sentinel Linear Search](https://github.com/TheAlgorithms/Python/blob/master/searches/sentinel_linear_search.py)
- * [Simple Binary Search](https://github.com/TheAlgorithms/Python/blob/master/searches/simple_binary_search.py)
- * [Simulated Annealing](https://github.com/TheAlgorithms/Python/blob/master/searches/simulated_annealing.py)
- * [Tabu Search](https://github.com/TheAlgorithms/Python/blob/master/searches/tabu_search.py)
- * [Ternary Search](https://github.com/TheAlgorithms/Python/blob/master/searches/ternary_search.py)
+ * [Binary Search](searches/binary_search.py)
+ * [Binary Tree Traversal](searches/binary_tree_traversal.py)
+ * [Double Linear Search](searches/double_linear_search.py)
+ * [Double Linear Search Recursion](searches/double_linear_search_recursion.py)
+ * [Exponential Search](searches/exponential_search.py)
+ * [Fibonacci Search](searches/fibonacci_search.py)
+ * [Hill Climbing](searches/hill_climbing.py)
+ * [Interpolation Search](searches/interpolation_search.py)
+ * [Jump Search](searches/jump_search.py)
+ * [Linear Search](searches/linear_search.py)
+ * [Median Of Medians](searches/median_of_medians.py)
+ * [Quick Select](searches/quick_select.py)
+ * [Sentinel Linear Search](searches/sentinel_linear_search.py)
+ * [Simple Binary Search](searches/simple_binary_search.py)
+ * [Simulated Annealing](searches/simulated_annealing.py)
+ * [Tabu Search](searches/tabu_search.py)
+ * [Ternary Search](searches/ternary_search.py)
## Sorts
- * [Bead Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/bead_sort.py)
- * [Bitonic Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/bitonic_sort.py)
- * [Bogo Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/bogo_sort.py)
- * [Bubble Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/bubble_sort.py)
- * [Bucket Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/bucket_sort.py)
- * [Cocktail Shaker Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/cocktail_shaker_sort.py)
- * [Comb Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/comb_sort.py)
- * [Counting Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/counting_sort.py)
- * [Cycle Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/cycle_sort.py)
- * [Double Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/double_sort.py)
- * [External Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/external_sort.py)
- * [Gnome Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/gnome_sort.py)
- * [Heap Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/heap_sort.py)
- * [Insertion Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/insertion_sort.py)
- * [Iterative Merge Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/iterative_merge_sort.py)
- * [Merge Insertion Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/merge_insertion_sort.py)
- * [Merge Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/merge_sort.py)
- * [Natural Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/natural_sort.py)
- * [Odd Even Transposition Parallel](https://github.com/TheAlgorithms/Python/blob/master/sorts/odd_even_transposition_parallel.py)
- * [Odd Even Transposition Single Threaded](https://github.com/TheAlgorithms/Python/blob/master/sorts/odd_even_transposition_single_threaded.py)
- * [Pancake Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/pancake_sort.py)
- * [Patience Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/patience_sort.py)
- * [Pigeon Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/pigeon_sort.py)
- * [Pigeonhole Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/pigeonhole_sort.py)
- * [Quick Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/quick_sort.py)
- * [Quick Sort 3 Partition](https://github.com/TheAlgorithms/Python/blob/master/sorts/quick_sort_3_partition.py)
- * [Radix Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/radix_sort.py)
- * [Random Normal Distribution Quicksort](https://github.com/TheAlgorithms/Python/blob/master/sorts/random_normal_distribution_quicksort.py)
- * [Random Pivot Quick Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/random_pivot_quick_sort.py)
- * [Recursive Bubble Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/recursive_bubble_sort.py)
- * [Recursive Insertion Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/recursive_insertion_sort.py)
- * [Recursive Quick Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/recursive_quick_sort.py)
- * [Selection Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/selection_sort.py)
- * [Shell Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/shell_sort.py)
- * [Stooge Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/stooge_sort.py)
- * [Strand Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/strand_sort.py)
- * [Tim Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/tim_sort.py)
- * [Topological Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/topological_sort.py)
- * [Tree Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/tree_sort.py)
- * [Unknown Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/unknown_sort.py)
- * [Wiggle Sort](https://github.com/TheAlgorithms/Python/blob/master/sorts/wiggle_sort.py)
+ * [Bead Sort](sorts/bead_sort.py)
+ * [Binary Insertion Sort](sorts/binary_insertion_sort.py)
+ * [Bitonic Sort](sorts/bitonic_sort.py)
+ * [Bogo Sort](sorts/bogo_sort.py)
+ * [Bubble Sort](sorts/bubble_sort.py)
+ * [Bucket Sort](sorts/bucket_sort.py)
+ * [Circle Sort](sorts/circle_sort.py)
+ * [Cocktail Shaker Sort](sorts/cocktail_shaker_sort.py)
+ * [Comb Sort](sorts/comb_sort.py)
+ * [Counting Sort](sorts/counting_sort.py)
+ * [Cycle Sort](sorts/cycle_sort.py)
+ * [Double Sort](sorts/double_sort.py)
+ * [Dutch National Flag Sort](sorts/dutch_national_flag_sort.py)
+ * [Exchange Sort](sorts/exchange_sort.py)
+ * [External Sort](sorts/external_sort.py)
+ * [Gnome Sort](sorts/gnome_sort.py)
+ * [Heap Sort](sorts/heap_sort.py)
+ * [Insertion Sort](sorts/insertion_sort.py)
+ * [Intro Sort](sorts/intro_sort.py)
+ * [Iterative Merge Sort](sorts/iterative_merge_sort.py)
+ * [Merge Insertion Sort](sorts/merge_insertion_sort.py)
+ * [Merge Sort](sorts/merge_sort.py)
+ * [Msd Radix Sort](sorts/msd_radix_sort.py)
+ * [Natural Sort](sorts/natural_sort.py)
+ * [Odd Even Sort](sorts/odd_even_sort.py)
+ * [Odd Even Transposition Parallel](sorts/odd_even_transposition_parallel.py)
+ * [Odd Even Transposition Single Threaded](sorts/odd_even_transposition_single_threaded.py)
+ * [Pancake Sort](sorts/pancake_sort.py)
+ * [Patience Sort](sorts/patience_sort.py)
+ * [Pigeon Sort](sorts/pigeon_sort.py)
+ * [Pigeonhole Sort](sorts/pigeonhole_sort.py)
+ * [Quick Sort](sorts/quick_sort.py)
+ * [Quick Sort 3 Partition](sorts/quick_sort_3_partition.py)
+ * [Radix Sort](sorts/radix_sort.py)
+ * [Recursive Insertion Sort](sorts/recursive_insertion_sort.py)
+ * [Recursive Mergesort Array](sorts/recursive_mergesort_array.py)
+ * [Recursive Quick Sort](sorts/recursive_quick_sort.py)
+ * [Selection Sort](sorts/selection_sort.py)
+ * [Shell Sort](sorts/shell_sort.py)
+ * [Shrink Shell Sort](sorts/shrink_shell_sort.py)
+ * [Slowsort](sorts/slowsort.py)
+ * [Stooge Sort](sorts/stooge_sort.py)
+ * [Strand Sort](sorts/strand_sort.py)
+ * [Tim Sort](sorts/tim_sort.py)
+ * [Topological Sort](sorts/topological_sort.py)
+ * [Tree Sort](sorts/tree_sort.py)
+ * [Unknown Sort](sorts/unknown_sort.py)
+ * [Wiggle Sort](sorts/wiggle_sort.py)
## Strings
- * [Aho Corasick](https://github.com/TheAlgorithms/Python/blob/master/strings/aho_corasick.py)
- * [Boyer Moore Search](https://github.com/TheAlgorithms/Python/blob/master/strings/boyer_moore_search.py)
- * [Can String Be Rearranged As Palindrome](https://github.com/TheAlgorithms/Python/blob/master/strings/can_string_be_rearranged_as_palindrome.py)
- * [Capitalize](https://github.com/TheAlgorithms/Python/blob/master/strings/capitalize.py)
- * [Check Anagrams](https://github.com/TheAlgorithms/Python/blob/master/strings/check_anagrams.py)
- * [Check Pangram](https://github.com/TheAlgorithms/Python/blob/master/strings/check_pangram.py)
- * [Is Palindrome](https://github.com/TheAlgorithms/Python/blob/master/strings/is_palindrome.py)
- * [Jaro Winkler](https://github.com/TheAlgorithms/Python/blob/master/strings/jaro_winkler.py)
- * [Knuth Morris Pratt](https://github.com/TheAlgorithms/Python/blob/master/strings/knuth_morris_pratt.py)
- * [Levenshtein Distance](https://github.com/TheAlgorithms/Python/blob/master/strings/levenshtein_distance.py)
- * [Lower](https://github.com/TheAlgorithms/Python/blob/master/strings/lower.py)
- * [Manacher](https://github.com/TheAlgorithms/Python/blob/master/strings/manacher.py)
- * [Min Cost String Conversion](https://github.com/TheAlgorithms/Python/blob/master/strings/min_cost_string_conversion.py)
- * [Naive String Search](https://github.com/TheAlgorithms/Python/blob/master/strings/naive_string_search.py)
- * [Prefix Function](https://github.com/TheAlgorithms/Python/blob/master/strings/prefix_function.py)
- * [Rabin Karp](https://github.com/TheAlgorithms/Python/blob/master/strings/rabin_karp.py)
- * [Remove Duplicate](https://github.com/TheAlgorithms/Python/blob/master/strings/remove_duplicate.py)
- * [Reverse Letters](https://github.com/TheAlgorithms/Python/blob/master/strings/reverse_letters.py)
- * [Reverse Words](https://github.com/TheAlgorithms/Python/blob/master/strings/reverse_words.py)
- * [Split](https://github.com/TheAlgorithms/Python/blob/master/strings/split.py)
- * [Swap Case](https://github.com/TheAlgorithms/Python/blob/master/strings/swap_case.py)
- * [Upper](https://github.com/TheAlgorithms/Python/blob/master/strings/upper.py)
- * [Word Occurrence](https://github.com/TheAlgorithms/Python/blob/master/strings/word_occurrence.py)
- * [Z Function](https://github.com/TheAlgorithms/Python/blob/master/strings/z_function.py)
-
-## Traversals
- * [Binary Tree Traversals](https://github.com/TheAlgorithms/Python/blob/master/traversals/binary_tree_traversals.py)
+ * [Aho Corasick](strings/aho_corasick.py)
+ * [Alternative String Arrange](strings/alternative_string_arrange.py)
+ * [Anagrams](strings/anagrams.py)
+ * [Autocomplete Using Trie](strings/autocomplete_using_trie.py)
+ * [Barcode Validator](strings/barcode_validator.py)
+ * [Bitap String Match](strings/bitap_string_match.py)
+ * [Boyer Moore Search](strings/boyer_moore_search.py)
+ * [Camel Case To Snake Case](strings/camel_case_to_snake_case.py)
+ * [Can String Be Rearranged As Palindrome](strings/can_string_be_rearranged_as_palindrome.py)
+ * [Capitalize](strings/capitalize.py)
+ * [Check Anagrams](strings/check_anagrams.py)
+ * [Count Vowels](strings/count_vowels.py)
+ * [Credit Card Validator](strings/credit_card_validator.py)
+ * [Damerau Levenshtein Distance](strings/damerau_levenshtein_distance.py)
+ * [Detecting English Programmatically](strings/detecting_english_programmatically.py)
+ * [Dna](strings/dna.py)
+ * [Edit Distance](strings/edit_distance.py)
+ * [Frequency Finder](strings/frequency_finder.py)
+ * [Hamming Distance](strings/hamming_distance.py)
+ * [Indian Phone Validator](strings/indian_phone_validator.py)
+ * [Is Contains Unique Chars](strings/is_contains_unique_chars.py)
+ * [Is Isogram](strings/is_isogram.py)
+ * [Is Pangram](strings/is_pangram.py)
+ * [Is Polish National Id](strings/is_polish_national_id.py)
+ * [Is Spain National Id](strings/is_spain_national_id.py)
+ * [Is Srilankan Phone Number](strings/is_srilankan_phone_number.py)
+ * [Is Valid Email Address](strings/is_valid_email_address.py)
+ * [Jaro Winkler](strings/jaro_winkler.py)
+ * [Join](strings/join.py)
+ * [Knuth Morris Pratt](strings/knuth_morris_pratt.py)
+ * [Levenshtein Distance](strings/levenshtein_distance.py)
+ * [Lower](strings/lower.py)
+ * [Manacher](strings/manacher.py)
+ * [Min Cost String Conversion](strings/min_cost_string_conversion.py)
+ * [Naive String Search](strings/naive_string_search.py)
+ * [Ngram](strings/ngram.py)
+ * [Palindrome](strings/palindrome.py)
+ * [Pig Latin](strings/pig_latin.py)
+ * [Prefix Function](strings/prefix_function.py)
+ * [Rabin Karp](strings/rabin_karp.py)
+ * [Remove Duplicate](strings/remove_duplicate.py)
+ * [Reverse Letters](strings/reverse_letters.py)
+ * [Reverse Words](strings/reverse_words.py)
+ * [Snake Case To Camel Pascal Case](strings/snake_case_to_camel_pascal_case.py)
+ * [Split](strings/split.py)
+ * [String Switch Case](strings/string_switch_case.py)
+ * [Strip](strings/strip.py)
+ * [Text Justification](strings/text_justification.py)
+ * [Title](strings/title.py)
+ * [Top K Frequent Words](strings/top_k_frequent_words.py)
+ * [Upper](strings/upper.py)
+ * [Wave String](strings/wave_string.py)
+ * [Wildcard Pattern Matching](strings/wildcard_pattern_matching.py)
+ * [Word Occurrence](strings/word_occurrence.py)
+ * [Word Patterns](strings/word_patterns.py)
+ * [Z Function](strings/z_function.py)
## Web Programming
- * [Co2 Emission](https://github.com/TheAlgorithms/Python/blob/master/web_programming/co2_emission.py)
- * [Covid Stats Via Xpath](https://github.com/TheAlgorithms/Python/blob/master/web_programming/covid_stats_via_xpath.py)
- * [Crawl Google Results](https://github.com/TheAlgorithms/Python/blob/master/web_programming/crawl_google_results.py)
- * [Crawl Google Scholar Citation](https://github.com/TheAlgorithms/Python/blob/master/web_programming/crawl_google_scholar_citation.py)
- * [Current Stock Price](https://github.com/TheAlgorithms/Python/blob/master/web_programming/current_stock_price.py)
- * [Current Weather](https://github.com/TheAlgorithms/Python/blob/master/web_programming/current_weather.py)
- * [Daily Horoscope](https://github.com/TheAlgorithms/Python/blob/master/web_programming/daily_horoscope.py)
- * [Emails From Url](https://github.com/TheAlgorithms/Python/blob/master/web_programming/emails_from_url.py)
- * [Fetch Bbc News](https://github.com/TheAlgorithms/Python/blob/master/web_programming/fetch_bbc_news.py)
- * [Fetch Github Info](https://github.com/TheAlgorithms/Python/blob/master/web_programming/fetch_github_info.py)
- * [Fetch Jobs](https://github.com/TheAlgorithms/Python/blob/master/web_programming/fetch_jobs.py)
- * [Get Imdb Top 250 Movies Csv](https://github.com/TheAlgorithms/Python/blob/master/web_programming/get_imdb_top_250_movies_csv.py)
- * [Get Imdbtop](https://github.com/TheAlgorithms/Python/blob/master/web_programming/get_imdbtop.py)
- * [Instagram Crawler](https://github.com/TheAlgorithms/Python/blob/master/web_programming/instagram_crawler.py)
- * [Recaptcha Verification](https://github.com/TheAlgorithms/Python/blob/master/web_programming/recaptcha_verification.py)
- * [Slack Message](https://github.com/TheAlgorithms/Python/blob/master/web_programming/slack_message.py)
- * [World Covid19 Stats](https://github.com/TheAlgorithms/Python/blob/master/web_programming/world_covid19_stats.py)
+ * [Co2 Emission](web_programming/co2_emission.py)
+ * [Covid Stats Via Xpath](web_programming/covid_stats_via_xpath.py)
+ * [Crawl Google Results](web_programming/crawl_google_results.py)
+ * [Crawl Google Scholar Citation](web_programming/crawl_google_scholar_citation.py)
+ * [Currency Converter](web_programming/currency_converter.py)
+ * [Current Stock Price](web_programming/current_stock_price.py)
+ * [Current Weather](web_programming/current_weather.py)
+ * [Daily Horoscope](web_programming/daily_horoscope.py)
+ * [Download Images From Google Query](web_programming/download_images_from_google_query.py)
+ * [Emails From Url](web_programming/emails_from_url.py)
+ * [Fetch Anime And Play](web_programming/fetch_anime_and_play.py)
+ * [Fetch Bbc News](web_programming/fetch_bbc_news.py)
+ * [Fetch Github Info](web_programming/fetch_github_info.py)
+ * [Fetch Jobs](web_programming/fetch_jobs.py)
+ * [Fetch Quotes](web_programming/fetch_quotes.py)
+ * [Fetch Well Rx Price](web_programming/fetch_well_rx_price.py)
+ * [Get Amazon Product Data](web_programming/get_amazon_product_data.py)
+ * [Get Imdb Top 250 Movies Csv](web_programming/get_imdb_top_250_movies_csv.py)
+ * [Get Ip Geolocation](web_programming/get_ip_geolocation.py)
+ * [Get Top Billionaires](web_programming/get_top_billionaires.py)
+ * [Get Top Hn Posts](web_programming/get_top_hn_posts.py)
+ * [Giphy](web_programming/giphy.py)
+ * [Instagram Crawler](web_programming/instagram_crawler.py)
+ * [Instagram Pic](web_programming/instagram_pic.py)
+ * [Instagram Video](web_programming/instagram_video.py)
+ * [Nasa Data](web_programming/nasa_data.py)
+ * [Open Google Results](web_programming/open_google_results.py)
+ * [Random Anime Character](web_programming/random_anime_character.py)
+ * [Recaptcha Verification](web_programming/recaptcha_verification.py)
+ * [Reddit](web_programming/reddit.py)
+ * [Search Books By Isbn](web_programming/search_books_by_isbn.py)
+ * [Slack Message](web_programming/slack_message.py)
+ * [Test Fetch Github Info](web_programming/test_fetch_github_info.py)
+ * [World Covid19 Stats](web_programming/world_covid19_stats.py)
diff --git a/LICENSE.md b/LICENSE.md
index 3b7951527ab3..de631c3ef333 100644
--- a/LICENSE.md
+++ b/LICENSE.md
@@ -1,6 +1,6 @@
-MIT License
+## MIT License
-Copyright (c) 2020 The Algorithms
+Copyright (c) 2016-2022 TheAlgorithms and contributors
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
diff --git a/README.md b/README.md
index f81031b53ebb..d8eba4e016fa 100644
--- a/README.md
+++ b/README.md
@@ -1,27 +1,49 @@
-# The Algorithms - Python
-[](https://gitpod.io/#https://github.com/TheAlgorithms/Python)
-[](https://gitter.im/TheAlgorithms)
-[](https://github.com/TheAlgorithms/Python/actions)
-[](https://lgtm.com/projects/g/TheAlgorithms/Python/alerts)
-[](https://github.com/TheAlgorithms/Python/blob/master/CONTRIBUTING.md)
-[](https://www.paypal.me/TheAlgorithms/100)
-
-[](https://github.com/pre-commit/pre-commit)
-[](https://github.com/psf/black)
-
-
-### All algorithms implemented in Python (for education)
-
-These implementations are for learning purposes only. Therefore they may be less efficient than the implementations in the Python standard library.
-
-## Contribution Guidelines
-
-Read our [Contribution Guidelines](CONTRIBUTING.md) before you contribute.
-
-## Community Channel
-
-We're on [Gitter](https://gitter.im/TheAlgorithms)! Please join us.
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+

+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
All algorithms implemented in Python - for education
+
+
+Implementations are for learning purposes only. They may be less efficient than the implementations in the Python standard library. Use them at your discretion.
+
+## Getting Started
+
+Read through our [Contribution Guidelines](CONTRIBUTING.md) before you contribute.
+
+## Community Channels
+
+We are on [Discord](https://the-algorithms.com/discord) and [Gitter](https://gitter.im/TheAlgorithms/community)! Community channels are a great way for you to ask questions and get help. Please join us!
## List of Algorithms
-See our [directory](DIRECTORY.md).
+See our [directory](DIRECTORY.md) for easier navigation and a better overview of the project.
diff --git a/arithmetic_analysis/lu_decomposition.py b/arithmetic_analysis/lu_decomposition.py
deleted file mode 100644
index 763ba60f32b7..000000000000
--- a/arithmetic_analysis/lu_decomposition.py
+++ /dev/null
@@ -1,34 +0,0 @@
-"""Lower-Upper (LU) Decomposition."""
-
-# lower–upper (LU) decomposition - https://en.wikipedia.org/wiki/LU_decomposition
-import numpy
-
-
-def LUDecompose(table):
- # Table that contains our data
- # Table has to be a square array so we need to check first
- rows, columns = numpy.shape(table)
- L = numpy.zeros((rows, columns))
- U = numpy.zeros((rows, columns))
- if rows != columns:
- return []
- for i in range(columns):
- for j in range(i):
- sum = 0
- for k in range(j):
- sum += L[i][k] * U[k][j]
- L[i][j] = (table[i][j] - sum) / U[j][j]
- L[i][i] = 1
- for j in range(i, columns):
- sum1 = 0
- for k in range(i):
- sum1 += L[i][k] * U[k][j]
- U[i][j] = table[i][j] - sum1
- return L, U
-
-
-if __name__ == "__main__":
- matrix = numpy.array([[2, -2, 1], [0, 1, 2], [5, 3, 1]])
- L, U = LUDecompose(matrix)
- print(L)
- print(U)
diff --git a/arithmetic_analysis/newton_method.py b/arithmetic_analysis/newton_method.py
deleted file mode 100644
index a9a94372671e..000000000000
--- a/arithmetic_analysis/newton_method.py
+++ /dev/null
@@ -1,54 +0,0 @@
-"""Newton's Method."""
-
-# Newton's Method - https://en.wikipedia.org/wiki/Newton%27s_method
-from typing import Callable
-
-RealFunc = Callable[[float], float] # type alias for a real -> real function
-
-
-# function is the f(x) and derivative is the f'(x)
-def newton(
- function: RealFunc,
- derivative: RealFunc,
- starting_int: int,
-) -> float:
- """
- >>> newton(lambda x: x ** 3 - 2 * x - 5, lambda x: 3 * x ** 2 - 2, 3)
- 2.0945514815423474
- >>> newton(lambda x: x ** 3 - 1, lambda x: 3 * x ** 2, -2)
- 1.0
- >>> newton(lambda x: x ** 3 - 1, lambda x: 3 * x ** 2, -4)
- 1.0000000000000102
- >>> import math
- >>> newton(math.sin, math.cos, 1)
- 0.0
- >>> newton(math.sin, math.cos, 2)
- 3.141592653589793
- >>> newton(math.cos, lambda x: -math.sin(x), 2)
- 1.5707963267948966
- >>> newton(math.cos, lambda x: -math.sin(x), 0)
- Traceback (most recent call last):
- ...
- ZeroDivisionError: Could not find root
- """
- prev_guess = float(starting_int)
- while True:
- try:
- next_guess = prev_guess - function(prev_guess) / derivative(prev_guess)
- except ZeroDivisionError:
- raise ZeroDivisionError("Could not find root") from None
- if abs(prev_guess - next_guess) < 10 ** -5:
- return next_guess
- prev_guess = next_guess
-
-
-def f(x: float) -> float:
- return (x ** 3) - (2 * x) - 5
-
-
-def f1(x: float) -> float:
- return 3 * (x ** 2) - 2
-
-
-if __name__ == "__main__":
- print(newton(f, f1, 3))
diff --git a/arithmetic_analysis/newton_raphson.py b/arithmetic_analysis/newton_raphson.py
deleted file mode 100644
index 948759a09a2a..000000000000
--- a/arithmetic_analysis/newton_raphson.py
+++ /dev/null
@@ -1,40 +0,0 @@
-# Implementing Newton Raphson method in Python
-# Author: Syed Haseeb Shah (github.com/QuantumNovice)
-# The Newton-Raphson method (also known as Newton's method) is a way to
-# quickly find a good approximation for the root of a real-valued function
-from decimal import Decimal
-from math import * # noqa: F401, F403
-
-from sympy import diff
-
-
-def newton_raphson(func: str, a: int, precision: int = 10 ** -10) -> float:
- """Finds root from the point 'a' onwards by Newton-Raphson method
- >>> newton_raphson("sin(x)", 2)
- 3.1415926536808043
- >>> newton_raphson("x**2 - 5*x +2", 0.4)
- 0.4384471871911695
- >>> newton_raphson("x**2 - 5", 0.1)
- 2.23606797749979
- >>> newton_raphson("log(x)- 1", 2)
- 2.718281828458938
- """
- x = a
- while True:
- x = Decimal(x) - (Decimal(eval(func)) / Decimal(eval(str(diff(func)))))
- # This number dictates the accuracy of the answer
- if abs(eval(func)) < precision:
- return float(x)
-
-
-# Let's Execute
-if __name__ == "__main__":
- # Find root of trigonometric function
- # Find value of pi
- print(f"The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}")
- # Find root of polynomial
- print(f"The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}")
- # Find Square Root of 5
- print(f"The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}")
- # Exponential Roots
- print(f"The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}")
diff --git a/arithmetic_analysis/secant_method.py b/arithmetic_analysis/secant_method.py
deleted file mode 100644
index b05d44c627d8..000000000000
--- a/arithmetic_analysis/secant_method.py
+++ /dev/null
@@ -1,28 +0,0 @@
-# Implementing Secant method in Python
-# Author: dimgrichr
-
-
-from math import exp
-
-
-def f(x):
- """
- >>> f(5)
- 39.98652410600183
- """
- return 8 * x - 2 * exp(-x)
-
-
-def SecantMethod(lower_bound, upper_bound, repeats):
- """
- >>> SecantMethod(1, 3, 2)
- 0.2139409276214589
- """
- x0 = lower_bound
- x1 = upper_bound
- for i in range(0, repeats):
- x0, x1 = x1, x1 - (f(x1) * (x1 - x0)) / (f(x1) - f(x0))
- return x1
-
-
-print(f"The solution is: {SecantMethod(1, 3, 2)}")
diff --git a/audio_filters/README.md b/audio_filters/README.md
new file mode 100644
index 000000000000..4419bd8bdbf9
--- /dev/null
+++ b/audio_filters/README.md
@@ -0,0 +1,9 @@
+# Audio Filter
+
+Audio filters work on the frequency of an audio signal to attenuate unwanted frequency and amplify wanted ones.
+They are used within anything related to sound, whether it is radio communication or a hi-fi system.
+
+*
+*
+*
+*
diff --git a/arithmetic_analysis/__init__.py b/audio_filters/__init__.py
similarity index 100%
rename from arithmetic_analysis/__init__.py
rename to audio_filters/__init__.py
diff --git a/audio_filters/butterworth_filter.py b/audio_filters/butterworth_filter.py
new file mode 100644
index 000000000000..4e6ea1b18fb4
--- /dev/null
+++ b/audio_filters/butterworth_filter.py
@@ -0,0 +1,234 @@
+from math import cos, sin, sqrt, tau
+
+from audio_filters.iir_filter import IIRFilter
+
+"""
+Create 2nd-order IIR filters with Butterworth design.
+
+Code based on https://webaudio.github.io/Audio-EQ-Cookbook/audio-eq-cookbook.html
+Alternatively you can use scipy.signal.butter, which should yield the same results.
+"""
+
+
+def make_lowpass(
+ frequency: int,
+ samplerate: int,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates a low-pass filter
+
+ >>> filter = make_lowpass(1000, 48000)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [1.0922959556412573, -1.9828897227476208, 0.9077040443587427, 0.004277569313094809,
+ 0.008555138626189618, 0.004277569313094809]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+
+ b0 = (1 - _cos) / 2
+ b1 = 1 - _cos
+
+ a0 = 1 + alpha
+ a1 = -2 * _cos
+ a2 = 1 - alpha
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([a0, a1, a2], [b0, b1, b0])
+ return filt
+
+
+def make_highpass(
+ frequency: int,
+ samplerate: int,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates a high-pass filter
+
+ >>> filter = make_highpass(1000, 48000)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [1.0922959556412573, -1.9828897227476208, 0.9077040443587427, 0.9957224306869052,
+ -1.9914448613738105, 0.9957224306869052]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+
+ b0 = (1 + _cos) / 2
+ b1 = -1 - _cos
+
+ a0 = 1 + alpha
+ a1 = -2 * _cos
+ a2 = 1 - alpha
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([a0, a1, a2], [b0, b1, b0])
+ return filt
+
+
+def make_bandpass(
+ frequency: int,
+ samplerate: int,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates a band-pass filter
+
+ >>> filter = make_bandpass(1000, 48000)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [1.0922959556412573, -1.9828897227476208, 0.9077040443587427, 0.06526309611002579,
+ 0, -0.06526309611002579]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+
+ b0 = _sin / 2
+ b1 = 0
+ b2 = -b0
+
+ a0 = 1 + alpha
+ a1 = -2 * _cos
+ a2 = 1 - alpha
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([a0, a1, a2], [b0, b1, b2])
+ return filt
+
+
+def make_allpass(
+ frequency: int,
+ samplerate: int,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates an all-pass filter
+
+ >>> filter = make_allpass(1000, 48000)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [1.0922959556412573, -1.9828897227476208, 0.9077040443587427, 0.9077040443587427,
+ -1.9828897227476208, 1.0922959556412573]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+
+ b0 = 1 - alpha
+ b1 = -2 * _cos
+ b2 = 1 + alpha
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([b2, b1, b0], [b0, b1, b2])
+ return filt
+
+
+def make_peak(
+ frequency: int,
+ samplerate: int,
+ gain_db: float,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates a peak filter
+
+ >>> filter = make_peak(1000, 48000, 6)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [1.0653405327119334, -1.9828897227476208, 0.9346594672880666, 1.1303715025601122,
+ -1.9828897227476208, 0.8696284974398878]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+ big_a = 10 ** (gain_db / 40)
+
+ b0 = 1 + alpha * big_a
+ b1 = -2 * _cos
+ b2 = 1 - alpha * big_a
+ a0 = 1 + alpha / big_a
+ a1 = -2 * _cos
+ a2 = 1 - alpha / big_a
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([a0, a1, a2], [b0, b1, b2])
+ return filt
+
+
+def make_lowshelf(
+ frequency: int,
+ samplerate: int,
+ gain_db: float,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates a low-shelf filter
+
+ >>> filter = make_lowshelf(1000, 48000, 6)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [3.0409336710888786, -5.608870992220748, 2.602157875636628, 3.139954022810743,
+ -5.591841778072785, 2.5201667380627257]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+ big_a = 10 ** (gain_db / 40)
+ pmc = (big_a + 1) - (big_a - 1) * _cos
+ ppmc = (big_a + 1) + (big_a - 1) * _cos
+ mpc = (big_a - 1) - (big_a + 1) * _cos
+ pmpc = (big_a - 1) + (big_a + 1) * _cos
+ aa2 = 2 * sqrt(big_a) * alpha
+
+ b0 = big_a * (pmc + aa2)
+ b1 = 2 * big_a * mpc
+ b2 = big_a * (pmc - aa2)
+ a0 = ppmc + aa2
+ a1 = -2 * pmpc
+ a2 = ppmc - aa2
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([a0, a1, a2], [b0, b1, b2])
+ return filt
+
+
+def make_highshelf(
+ frequency: int,
+ samplerate: int,
+ gain_db: float,
+ q_factor: float = 1 / sqrt(2),
+) -> IIRFilter:
+ """
+ Creates a high-shelf filter
+
+ >>> filter = make_highshelf(1000, 48000, 6)
+ >>> filter.a_coeffs + filter.b_coeffs # doctest: +NORMALIZE_WHITESPACE
+ [2.2229172136088806, -3.9587208137297303, 1.7841414181566304, 4.295432981120543,
+ -7.922740859457287, 3.6756456963725253]
+ """
+ w0 = tau * frequency / samplerate
+ _sin = sin(w0)
+ _cos = cos(w0)
+ alpha = _sin / (2 * q_factor)
+ big_a = 10 ** (gain_db / 40)
+ pmc = (big_a + 1) - (big_a - 1) * _cos
+ ppmc = (big_a + 1) + (big_a - 1) * _cos
+ mpc = (big_a - 1) - (big_a + 1) * _cos
+ pmpc = (big_a - 1) + (big_a + 1) * _cos
+ aa2 = 2 * sqrt(big_a) * alpha
+
+ b0 = big_a * (ppmc + aa2)
+ b1 = -2 * big_a * pmpc
+ b2 = big_a * (ppmc - aa2)
+ a0 = pmc + aa2
+ a1 = 2 * mpc
+ a2 = pmc - aa2
+
+ filt = IIRFilter(2)
+ filt.set_coefficients([a0, a1, a2], [b0, b1, b2])
+ return filt
diff --git a/audio_filters/equal_loudness_filter.py.broken.txt b/audio_filters/equal_loudness_filter.py.broken.txt
new file mode 100644
index 000000000000..88cba8533cf7
--- /dev/null
+++ b/audio_filters/equal_loudness_filter.py.broken.txt
@@ -0,0 +1,61 @@
+from json import loads
+from pathlib import Path
+
+import numpy as np
+from yulewalker import yulewalk
+
+from audio_filters.butterworth_filter import make_highpass
+from audio_filters.iir_filter import IIRFilter
+
+data = loads((Path(__file__).resolve().parent / "loudness_curve.json").read_text())
+
+
+class EqualLoudnessFilter:
+ r"""
+ An equal-loudness filter which compensates for the human ear's non-linear response
+ to sound.
+ This filter corrects this by cascading a yulewalk filter and a butterworth filter.
+
+ Designed for use with samplerate of 44.1kHz and above. If you're using a lower
+ samplerate, use with caution.
+
+ Code based on matlab implementation at https://bit.ly/3eqh2HU
+ (url shortened for ruff)
+
+ Target curve: https://i.imgur.com/3g2VfaM.png
+ Yulewalk response: https://i.imgur.com/J9LnJ4C.png
+ Butterworth and overall response: https://i.imgur.com/3g2VfaM.png
+
+ Images and original matlab implementation by David Robinson, 2001
+ """
+
+ def __init__(self, samplerate: int = 44100) -> None:
+ self.yulewalk_filter = IIRFilter(10)
+ self.butterworth_filter = make_highpass(150, samplerate)
+
+ # pad the data to nyquist
+ curve_freqs = np.array(data["frequencies"] + [max(20000.0, samplerate / 2)])
+ curve_gains = np.array(data["gains"] + [140])
+
+ # Convert to angular frequency
+ freqs_normalized = curve_freqs / samplerate * 2
+ # Invert the curve and normalize to 0dB
+ gains_normalized = np.power(10, (np.min(curve_gains) - curve_gains) / 20)
+
+ # Scipy's `yulewalk` function is a stub, so we're using the
+ # `yulewalker` library instead.
+ # This function computes the coefficients using a least-squares
+ # fit to the specified curve.
+ ya, yb = yulewalk(10, freqs_normalized, gains_normalized)
+ self.yulewalk_filter.set_coefficients(ya, yb)
+
+ def process(self, sample: float) -> float:
+ """
+ Process a single sample through both filters
+
+ >>> filt = EqualLoudnessFilter()
+ >>> filt.process(0.0)
+ 0.0
+ """
+ tmp = self.yulewalk_filter.process(sample)
+ return self.butterworth_filter.process(tmp)
diff --git a/audio_filters/iir_filter.py b/audio_filters/iir_filter.py
new file mode 100644
index 000000000000..fa3e6c54b33f
--- /dev/null
+++ b/audio_filters/iir_filter.py
@@ -0,0 +1,100 @@
+from __future__ import annotations
+
+
+class IIRFilter:
+ r"""
+ N-Order IIR filter
+ Assumes working with float samples normalized on [-1, 1]
+
+ ---
+
+ Implementation details:
+ Based on the 2nd-order function from
+ https://en.wikipedia.org/wiki/Digital_biquad_filter,
+ this generalized N-order function was made.
+
+ Using the following transfer function
+ .. math:: H(z)=\frac{b_{0}+b_{1}z^{-1}+b_{2}z^{-2}+...+b_{k}z^{-k}}
+ {a_{0}+a_{1}z^{-1}+a_{2}z^{-2}+...+a_{k}z^{-k}}
+
+ we can rewrite this to
+ .. math:: y[n]={\frac{1}{a_{0}}}
+ \left(\left(b_{0}x[n]+b_{1}x[n-1]+b_{2}x[n-2]+...+b_{k}x[n-k]\right)-
+ \left(a_{1}y[n-1]+a_{2}y[n-2]+...+a_{k}y[n-k]\right)\right)
+ """
+
+ def __init__(self, order: int) -> None:
+ self.order = order
+
+ # a_{0} ... a_{k}
+ self.a_coeffs = [1.0] + [0.0] * order
+ # b_{0} ... b_{k}
+ self.b_coeffs = [1.0] + [0.0] * order
+
+ # x[n-1] ... x[n-k]
+ self.input_history = [0.0] * self.order
+ # y[n-1] ... y[n-k]
+ self.output_history = [0.0] * self.order
+
+ def set_coefficients(self, a_coeffs: list[float], b_coeffs: list[float]) -> None:
+ """
+ Set the coefficients for the IIR filter.
+ These should both be of size `order` + 1.
+ :math:`a_0` may be left out, and it will use 1.0 as default value.
+
+ This method works well with scipy's filter design functions
+
+ >>> # Make a 2nd-order 1000Hz butterworth lowpass filter
+ >>> import scipy.signal
+ >>> b_coeffs, a_coeffs = scipy.signal.butter(2, 1000,
+ ... btype='lowpass',
+ ... fs=48000)
+ >>> filt = IIRFilter(2)
+ >>> filt.set_coefficients(a_coeffs, b_coeffs)
+ """
+ if len(a_coeffs) < self.order:
+ a_coeffs = [1.0, *a_coeffs]
+
+ if len(a_coeffs) != self.order + 1:
+ msg = (
+ f"Expected a_coeffs to have {self.order + 1} elements "
+ f"for {self.order}-order filter, got {len(a_coeffs)}"
+ )
+ raise ValueError(msg)
+
+ if len(b_coeffs) != self.order + 1:
+ msg = (
+ f"Expected b_coeffs to have {self.order + 1} elements "
+ f"for {self.order}-order filter, got {len(a_coeffs)}"
+ )
+ raise ValueError(msg)
+
+ self.a_coeffs = a_coeffs
+ self.b_coeffs = b_coeffs
+
+ def process(self, sample: float) -> float:
+ """
+ Calculate :math:`y[n]`
+
+ >>> filt = IIRFilter(2)
+ >>> filt.process(0)
+ 0.0
+ """
+ result = 0.0
+
+ # Start at index 1 and do index 0 at the end.
+ for i in range(1, self.order + 1):
+ result += (
+ self.b_coeffs[i] * self.input_history[i - 1]
+ - self.a_coeffs[i] * self.output_history[i - 1]
+ )
+
+ result = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0]
+
+ self.input_history[1:] = self.input_history[:-1]
+ self.output_history[1:] = self.output_history[:-1]
+
+ self.input_history[0] = sample
+ self.output_history[0] = result
+
+ return result
diff --git a/audio_filters/loudness_curve.json b/audio_filters/loudness_curve.json
new file mode 100644
index 000000000000..fc066a0810fc
--- /dev/null
+++ b/audio_filters/loudness_curve.json
@@ -0,0 +1,76 @@
+{
+ "_comment": "The following is a representative average of the Equal Loudness Contours as measured by Robinson and Dadson, 1956",
+ "_doi": "10.1088/0508-3443/7/5/302",
+ "frequencies": [
+ 0,
+ 20,
+ 30,
+ 40,
+ 50,
+ 60,
+ 70,
+ 80,
+ 90,
+ 100,
+ 200,
+ 300,
+ 400,
+ 500,
+ 600,
+ 700,
+ 800,
+ 900,
+ 1000,
+ 1500,
+ 2000,
+ 2500,
+ 3000,
+ 3700,
+ 4000,
+ 5000,
+ 6000,
+ 7000,
+ 8000,
+ 9000,
+ 10000,
+ 12000,
+ 15000,
+ 20000
+ ],
+ "gains": [
+ 120,
+ 113,
+ 103,
+ 97,
+ 93,
+ 91,
+ 89,
+ 87,
+ 86,
+ 85,
+ 78,
+ 76,
+ 76,
+ 76,
+ 76,
+ 77,
+ 78,
+ 79.5,
+ 80,
+ 79,
+ 77,
+ 74,
+ 71.5,
+ 70,
+ 70.5,
+ 74,
+ 79,
+ 84,
+ 86,
+ 86,
+ 85,
+ 95,
+ 110,
+ 125
+ ]
+}
diff --git a/audio_filters/show_response.py b/audio_filters/show_response.py
new file mode 100644
index 000000000000..f9c9537c047c
--- /dev/null
+++ b/audio_filters/show_response.py
@@ -0,0 +1,95 @@
+from __future__ import annotations
+
+from abc import abstractmethod
+from math import pi
+from typing import Protocol
+
+import matplotlib.pyplot as plt
+import numpy as np
+
+
+class FilterType(Protocol):
+ @abstractmethod
+ def process(self, sample: float) -> float:
+ """
+ Calculate y[n]
+
+ >>> issubclass(FilterType, Protocol)
+ True
+ """
+
+
+def get_bounds(
+ fft_results: np.ndarray, samplerate: int
+) -> tuple[int | float, int | float]:
+ """
+ Get bounds for printing fft results
+
+ >>> import numpy
+ >>> array = numpy.linspace(-20.0, 20.0, 1000)
+ >>> get_bounds(array, 1000)
+ (-20, 20)
+ """
+ lowest = min([-20, np.min(fft_results[1 : samplerate // 2 - 1])])
+ highest = max([20, np.max(fft_results[1 : samplerate // 2 - 1])])
+ return lowest, highest
+
+
+def show_frequency_response(filter_type: FilterType, samplerate: int) -> None:
+ """
+ Show frequency response of a filter
+
+ >>> from audio_filters.iir_filter import IIRFilter
+ >>> filt = IIRFilter(4)
+ >>> show_frequency_response(filt, 48000)
+ """
+
+ size = 512
+ inputs = [1] + [0] * (size - 1)
+ outputs = [filter_type.process(item) for item in inputs]
+
+ filler = [0] * (samplerate - size) # zero-padding
+ outputs += filler
+ fft_out = np.abs(np.fft.fft(outputs))
+ fft_db = 20 * np.log10(fft_out)
+
+ # Frequencies on log scale from 24 to nyquist frequency
+ plt.xlim(24, samplerate / 2 - 1)
+ plt.xlabel("Frequency (Hz)")
+ plt.xscale("log")
+
+ # Display within reasonable bounds
+ bounds = get_bounds(fft_db, samplerate)
+ plt.ylim(max([-80, bounds[0]]), min([80, bounds[1]]))
+ plt.ylabel("Gain (dB)")
+
+ plt.plot(fft_db)
+ plt.show()
+
+
+def show_phase_response(filter_type: FilterType, samplerate: int) -> None:
+ """
+ Show phase response of a filter
+
+ >>> from audio_filters.iir_filter import IIRFilter
+ >>> filt = IIRFilter(4)
+ >>> show_phase_response(filt, 48000)
+ """
+
+ size = 512
+ inputs = [1] + [0] * (size - 1)
+ outputs = [filter_type.process(item) for item in inputs]
+
+ filler = [0] * (samplerate - size) # zero-padding
+ outputs += filler
+ fft_out = np.angle(np.fft.fft(outputs))
+
+ # Frequencies on log scale from 24 to nyquist frequency
+ plt.xlim(24, samplerate / 2 - 1)
+ plt.xlabel("Frequency (Hz)")
+ plt.xscale("log")
+
+ plt.ylim(-2 * pi, 2 * pi)
+ plt.ylabel("Phase shift (Radians)")
+ plt.plot(np.unwrap(fft_out, -2 * pi))
+ plt.show()
diff --git a/backtracking/README.md b/backtracking/README.md
new file mode 100644
index 000000000000..d4975dfb5ad7
--- /dev/null
+++ b/backtracking/README.md
@@ -0,0 +1,8 @@
+# Backtracking
+
+Backtracking is a way to speed up the search process by removing candidates when they can't be the solution of a problem.
+
+*
+*
+*
+*
diff --git a/backtracking/all_combinations.py b/backtracking/all_combinations.py
index 854dc5198422..1d15c6263e14 100644
--- a/backtracking/all_combinations.py
+++ b/backtracking/all_combinations.py
@@ -1,22 +1,95 @@
"""
- In this problem, we want to determine all possible combinations of k
- numbers out of 1 ... n. We use backtracking to solve this problem.
- Time complexity: O(C(n,k)) which is O(n choose k) = O((n!/(k! * (n - k)!)))
+In this problem, we want to determine all possible combinations of k
+numbers out of 1 ... n. We use backtracking to solve this problem.
+
+Time complexity: O(C(n,k)) which is O(n choose k) = O((n!/(k! * (n - k)!))),
"""
+from __future__ import annotations
+
+from itertools import combinations
+
+
+def combination_lists(n: int, k: int) -> list[list[int]]:
+ """
+ Generates all possible combinations of k numbers out of 1 ... n using itertools.
+
+ >>> combination_lists(n=4, k=2)
+ [[1, 2], [1, 3], [1, 4], [2, 3], [2, 4], [3, 4]]
+ """
+ return [list(x) for x in combinations(range(1, n + 1), k)]
+
-def generate_all_combinations(n: int, k: int) -> [[int]]:
+def generate_all_combinations(n: int, k: int) -> list[list[int]]:
"""
+ Generates all possible combinations of k numbers out of 1 ... n using backtracking.
+
>>> generate_all_combinations(n=4, k=2)
[[1, 2], [1, 3], [1, 4], [2, 3], [2, 4], [3, 4]]
+ >>> generate_all_combinations(n=0, k=0)
+ [[]]
+ >>> generate_all_combinations(n=10, k=-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: k must not be negative
+ >>> generate_all_combinations(n=-1, k=10)
+ Traceback (most recent call last):
+ ...
+ ValueError: n must not be negative
+ >>> generate_all_combinations(n=5, k=4)
+ [[1, 2, 3, 4], [1, 2, 3, 5], [1, 2, 4, 5], [1, 3, 4, 5], [2, 3, 4, 5]]
+ >>> generate_all_combinations(n=3, k=3)
+ [[1, 2, 3]]
+ >>> generate_all_combinations(n=3, k=1)
+ [[1], [2], [3]]
+ >>> generate_all_combinations(n=1, k=0)
+ [[]]
+ >>> generate_all_combinations(n=1, k=1)
+ [[1]]
+ >>> from itertools import combinations
+ >>> all(generate_all_combinations(n, k) == combination_lists(n, k)
+ ... for n in range(1, 6) for k in range(1, 6))
+ True
"""
+ if k < 0:
+ raise ValueError("k must not be negative")
+ if n < 0:
+ raise ValueError("n must not be negative")
- result = []
+ result: list[list[int]] = []
create_all_state(1, n, k, [], result)
return result
-def create_all_state(increment, total_number, level, current_list, total_list):
+def create_all_state(
+ increment: int,
+ total_number: int,
+ level: int,
+ current_list: list[int],
+ total_list: list[list[int]],
+) -> None:
+ """
+ Helper function to recursively build all combinations.
+
+ >>> create_all_state(1, 4, 2, [], result := [])
+ >>> result
+ [[1, 2], [1, 3], [1, 4], [2, 3], [2, 4], [3, 4]]
+ >>> create_all_state(1, 3, 3, [], result := [])
+ >>> result
+ [[1, 2, 3]]
+ >>> create_all_state(2, 2, 1, [1], result := [])
+ >>> result
+ [[1, 2]]
+ >>> create_all_state(1, 0, 0, [], result := [])
+ >>> result
+ [[]]
+ >>> create_all_state(1, 4, 0, [1, 2], result := [])
+ >>> result
+ [[1, 2]]
+ >>> create_all_state(5, 4, 2, [1, 2], result := [])
+ >>> result
+ []
+ """
if level == 0:
total_list.append(current_list[:])
return
@@ -27,13 +100,17 @@ def create_all_state(increment, total_number, level, current_list, total_list):
current_list.pop()
-def print_all_state(total_list):
- for i in total_list:
- print(*i)
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ print(generate_all_combinations(n=4, k=2))
+ tests = ((n, k) for n in range(1, 5) for k in range(1, 5))
+ for n, k in tests:
+ print(n, k, generate_all_combinations(n, k) == combination_lists(n, k))
+ print("Benchmark:")
+ from timeit import timeit
-if __name__ == "__main__":
- n = 4
- k = 2
- total_list = generate_all_combinations(n, k)
- print_all_state(total_list)
+ for func in ("combination_lists", "generate_all_combinations"):
+ print(f"{func:>25}(): {timeit(f'{func}(n=4, k = 2)', globals=globals())}")
diff --git a/backtracking/all_permutations.py b/backtracking/all_permutations.py
index 5244fef97f93..f376e6fa0945 100644
--- a/backtracking/all_permutations.py
+++ b/backtracking/all_permutations.py
@@ -1,21 +1,64 @@
"""
- In this problem, we want to determine all possible permutations
- of the given sequence. We use backtracking to solve this problem.
+In this problem, we want to determine all possible permutations
+of the given sequence. We use backtracking to solve this problem.
- Time complexity: O(n! * n),
- where n denotes the length of the given sequence.
+Time complexity: O(n! * n),
+where n denotes the length of the given sequence.
"""
+from __future__ import annotations
-def generate_all_permutations(sequence):
+
+def generate_all_permutations(sequence: list[int | str]) -> None:
create_state_space_tree(sequence, [], 0, [0 for i in range(len(sequence))])
-def create_state_space_tree(sequence, current_sequence, index, index_used):
+def create_state_space_tree(
+ sequence: list[int | str],
+ current_sequence: list[int | str],
+ index: int,
+ index_used: list[int],
+) -> None:
"""
Creates a state space tree to iterate through each branch using DFS.
We know that each state has exactly len(sequence) - index children.
It terminates when it reaches the end of the given sequence.
+
+ :param sequence: The input sequence for which permutations are generated.
+ :param current_sequence: The current permutation being built.
+ :param index: The current index in the sequence.
+ :param index_used: list to track which elements are used in permutation.
+
+ Example 1:
+ >>> sequence = [1, 2, 3]
+ >>> current_sequence = []
+ >>> index_used = [False, False, False]
+ >>> create_state_space_tree(sequence, current_sequence, 0, index_used)
+ [1, 2, 3]
+ [1, 3, 2]
+ [2, 1, 3]
+ [2, 3, 1]
+ [3, 1, 2]
+ [3, 2, 1]
+
+ Example 2:
+ >>> sequence = ["A", "B", "C"]
+ >>> current_sequence = []
+ >>> index_used = [False, False, False]
+ >>> create_state_space_tree(sequence, current_sequence, 0, index_used)
+ ['A', 'B', 'C']
+ ['A', 'C', 'B']
+ ['B', 'A', 'C']
+ ['B', 'C', 'A']
+ ['C', 'A', 'B']
+ ['C', 'B', 'A']
+
+ Example 3:
+ >>> sequence = [1]
+ >>> current_sequence = []
+ >>> index_used = [False]
+ >>> create_state_space_tree(sequence, current_sequence, 0, index_used)
+ [1]
"""
if index == len(sequence):
@@ -38,8 +81,8 @@ def create_state_space_tree(sequence, current_sequence, index, index_used):
sequence = list(map(int, input().split()))
"""
-sequence = [3, 1, 2, 4]
+sequence: list[int | str] = [3, 1, 2, 4]
generate_all_permutations(sequence)
-sequence = ["A", "B", "C"]
-generate_all_permutations(sequence)
+sequence_2: list[int | str] = ["A", "B", "C"]
+generate_all_permutations(sequence_2)
diff --git a/backtracking/all_subsequences.py b/backtracking/all_subsequences.py
index 9086e3a3d659..18696054eb7e 100644
--- a/backtracking/all_subsequences.py
+++ b/backtracking/all_subsequences.py
@@ -1,25 +1,77 @@
-from typing import Any, List
-
"""
- In this problem, we want to determine all possible subsequences
- of the given sequence. We use backtracking to solve this problem.
+In this problem, we want to determine all possible subsequences
+of the given sequence. We use backtracking to solve this problem.
- Time complexity: O(2^n),
- where n denotes the length of the given sequence.
+Time complexity: O(2^n),
+where n denotes the length of the given sequence.
"""
+from __future__ import annotations
+
+from typing import Any
+
-def generate_all_subsequences(sequence: List[Any]) -> None:
+def generate_all_subsequences(sequence: list[Any]) -> None:
create_state_space_tree(sequence, [], 0)
def create_state_space_tree(
- sequence: List[Any], current_subsequence: List[Any], index: int
+ sequence: list[Any], current_subsequence: list[Any], index: int
) -> None:
"""
Creates a state space tree to iterate through each branch using DFS.
We know that each state has exactly two children.
It terminates when it reaches the end of the given sequence.
+
+ :param sequence: The input sequence for which subsequences are generated.
+ :param current_subsequence: The current subsequence being built.
+ :param index: The current index in the sequence.
+
+ Example:
+ >>> sequence = [3, 2, 1]
+ >>> current_subsequence = []
+ >>> create_state_space_tree(sequence, current_subsequence, 0)
+ []
+ [1]
+ [2]
+ [2, 1]
+ [3]
+ [3, 1]
+ [3, 2]
+ [3, 2, 1]
+
+ >>> sequence = ["A", "B"]
+ >>> current_subsequence = []
+ >>> create_state_space_tree(sequence, current_subsequence, 0)
+ []
+ ['B']
+ ['A']
+ ['A', 'B']
+
+ >>> sequence = []
+ >>> current_subsequence = []
+ >>> create_state_space_tree(sequence, current_subsequence, 0)
+ []
+
+ >>> sequence = [1, 2, 3, 4]
+ >>> current_subsequence = []
+ >>> create_state_space_tree(sequence, current_subsequence, 0)
+ []
+ [4]
+ [3]
+ [3, 4]
+ [2]
+ [2, 4]
+ [2, 3]
+ [2, 3, 4]
+ [1]
+ [1, 4]
+ [1, 3]
+ [1, 3, 4]
+ [1, 2]
+ [1, 2, 4]
+ [1, 2, 3]
+ [1, 2, 3, 4]
"""
if index == len(sequence):
@@ -32,15 +84,10 @@ def create_state_space_tree(
current_subsequence.pop()
-"""
-remove the comment to take an input from the user
-
-print("Enter the elements")
-sequence = list(map(int, input().split()))
-"""
-
-sequence = [3, 1, 2, 4]
-generate_all_subsequences(sequence)
+if __name__ == "__main__":
+ seq: list[Any] = [1, 2, 3]
+ generate_all_subsequences(seq)
-sequence = ["A", "B", "C"]
-generate_all_subsequences(sequence)
+ seq.clear()
+ seq.extend(["A", "B", "C"])
+ generate_all_subsequences(seq)
diff --git a/backtracking/coloring.py b/backtracking/coloring.py
index ceaffe3fae76..f10cdbcf9d26 100644
--- a/backtracking/coloring.py
+++ b/backtracking/coloring.py
@@ -1,20 +1,19 @@
"""
- Graph Coloring also called "m coloring problem"
- consists of coloring given graph with at most m colors
- such that no adjacent vertices are assigned same color
+Graph Coloring also called "m coloring problem"
+consists of coloring a given graph with at most m colors
+such that no adjacent vertices are assigned the same color
- Wikipedia: https://en.wikipedia.org/wiki/Graph_coloring
+Wikipedia: https://en.wikipedia.org/wiki/Graph_coloring
"""
-from __future__ import annotations
def valid_coloring(
neighbours: list[int], colored_vertices: list[int], color: int
) -> bool:
"""
- For each neighbour check if coloring constraint is satisfied
+ For each neighbour check if the coloring constraint is satisfied
If any of the neighbours fail the constraint return False
- If all neighbours validate constraint return True
+ If all neighbours validate the constraint return True
>>> neighbours = [0,1,0,1,0]
>>> colored_vertices = [0, 2, 1, 2, 0]
@@ -42,14 +41,14 @@ def util_color(
Base Case:
1. Check if coloring is complete
- 1.1 If complete return True (meaning that we successfully colored graph)
+ 1.1 If complete return True (meaning that we successfully colored the graph)
Recursive Step:
- 2. Itterates over each color:
- Check if current coloring is valid:
+ 2. Iterates over each color:
+ Check if the current coloring is valid:
2.1. Color given vertex
- 2.2. Do recursive call check if this coloring leads to solving problem
- 2.4. if current coloring leads to solution return
+ 2.2. Do recursive call, check if this coloring leads to a solution
+ 2.4. if current coloring leads to a solution return
2.5. Uncolor given vertex
>>> graph = [[0, 1, 0, 0, 0],
diff --git a/backtracking/combination_sum.py b/backtracking/combination_sum.py
new file mode 100644
index 000000000000..3c6ed81f44f0
--- /dev/null
+++ b/backtracking/combination_sum.py
@@ -0,0 +1,66 @@
+"""
+In the Combination Sum problem, we are given a list consisting of distinct integers.
+We need to find all the combinations whose sum equals to target given.
+We can use an element more than one.
+
+Time complexity(Average Case): O(n!)
+
+Constraints:
+1 <= candidates.length <= 30
+2 <= candidates[i] <= 40
+All elements of candidates are distinct.
+1 <= target <= 40
+"""
+
+
+def backtrack(
+ candidates: list, path: list, answer: list, target: int, previous_index: int
+) -> None:
+ """
+ A recursive function that searches for possible combinations. Backtracks in case
+ of a bigger current combination value than the target value.
+
+ Parameters
+ ----------
+ previous_index: Last index from the previous search
+ target: The value we need to obtain by summing our integers in the path list.
+ answer: A list of possible combinations
+ path: Current combination
+ candidates: A list of integers we can use.
+ """
+ if target == 0:
+ answer.append(path.copy())
+ else:
+ for index in range(previous_index, len(candidates)):
+ if target >= candidates[index]:
+ path.append(candidates[index])
+ backtrack(candidates, path, answer, target - candidates[index], index)
+ path.pop(len(path) - 1)
+
+
+def combination_sum(candidates: list, target: int) -> list:
+ """
+ >>> combination_sum([2, 3, 5], 8)
+ [[2, 2, 2, 2], [2, 3, 3], [3, 5]]
+ >>> combination_sum([2, 3, 6, 7], 7)
+ [[2, 2, 3], [7]]
+ >>> combination_sum([-8, 2.3, 0], 1)
+ Traceback (most recent call last):
+ ...
+ RecursionError: maximum recursion depth exceeded
+ """
+ path = [] # type: list[int]
+ answer = [] # type: list[int]
+ backtrack(candidates, path, answer, target, 0)
+ return answer
+
+
+def main() -> None:
+ print(combination_sum([-8, 2.3, 0], 1))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ main()
diff --git a/backtracking/crossword_puzzle_solver.py b/backtracking/crossword_puzzle_solver.py
new file mode 100644
index 000000000000..e702c7e52153
--- /dev/null
+++ b/backtracking/crossword_puzzle_solver.py
@@ -0,0 +1,131 @@
+# https://www.geeksforgeeks.org/solve-crossword-puzzle/
+
+
+def is_valid(
+ puzzle: list[list[str]], word: str, row: int, col: int, vertical: bool
+) -> bool:
+ """
+ Check if a word can be placed at the given position.
+
+ >>> puzzle = [
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', '']
+ ... ]
+ >>> is_valid(puzzle, 'word', 0, 0, True)
+ True
+ >>> puzzle = [
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', '']
+ ... ]
+ >>> is_valid(puzzle, 'word', 0, 0, False)
+ True
+ """
+ for i in range(len(word)):
+ if vertical:
+ if row + i >= len(puzzle) or puzzle[row + i][col] != "":
+ return False
+ elif col + i >= len(puzzle[0]) or puzzle[row][col + i] != "":
+ return False
+ return True
+
+
+def place_word(
+ puzzle: list[list[str]], word: str, row: int, col: int, vertical: bool
+) -> None:
+ """
+ Place a word at the given position.
+
+ >>> puzzle = [
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', '']
+ ... ]
+ >>> place_word(puzzle, 'word', 0, 0, True)
+ >>> puzzle
+ [['w', '', '', ''], ['o', '', '', ''], ['r', '', '', ''], ['d', '', '', '']]
+ """
+ for i, char in enumerate(word):
+ if vertical:
+ puzzle[row + i][col] = char
+ else:
+ puzzle[row][col + i] = char
+
+
+def remove_word(
+ puzzle: list[list[str]], word: str, row: int, col: int, vertical: bool
+) -> None:
+ """
+ Remove a word from the given position.
+
+ >>> puzzle = [
+ ... ['w', '', '', ''],
+ ... ['o', '', '', ''],
+ ... ['r', '', '', ''],
+ ... ['d', '', '', '']
+ ... ]
+ >>> remove_word(puzzle, 'word', 0, 0, True)
+ >>> puzzle
+ [['', '', '', ''], ['', '', '', ''], ['', '', '', ''], ['', '', '', '']]
+ """
+ for i in range(len(word)):
+ if vertical:
+ puzzle[row + i][col] = ""
+ else:
+ puzzle[row][col + i] = ""
+
+
+def solve_crossword(puzzle: list[list[str]], words: list[str]) -> bool:
+ """
+ Solve the crossword puzzle using backtracking.
+
+ >>> puzzle = [
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', '']
+ ... ]
+
+ >>> words = ['word', 'four', 'more', 'last']
+ >>> solve_crossword(puzzle, words)
+ True
+ >>> puzzle = [
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', ''],
+ ... ['', '', '', '']
+ ... ]
+ >>> words = ['word', 'four', 'more', 'paragraphs']
+ >>> solve_crossword(puzzle, words)
+ False
+ """
+ for row in range(len(puzzle)):
+ for col in range(len(puzzle[0])):
+ if puzzle[row][col] == "":
+ for word in words:
+ for vertical in [True, False]:
+ if is_valid(puzzle, word, row, col, vertical):
+ place_word(puzzle, word, row, col, vertical)
+ words.remove(word)
+ if solve_crossword(puzzle, words):
+ return True
+ words.append(word)
+ remove_word(puzzle, word, row, col, vertical)
+ return False
+ return True
+
+
+if __name__ == "__main__":
+ PUZZLE = [[""] * 3 for _ in range(3)]
+ WORDS = ["cat", "dog", "car"]
+
+ if solve_crossword(PUZZLE, WORDS):
+ print("Solution found:")
+ for row in PUZZLE:
+ print(" ".join(row))
+ else:
+ print("No solution found:")
diff --git a/backtracking/generate_parentheses.py b/backtracking/generate_parentheses.py
new file mode 100644
index 000000000000..18c21e2a9b51
--- /dev/null
+++ b/backtracking/generate_parentheses.py
@@ -0,0 +1,77 @@
+"""
+author: Aayush Soni
+Given n pairs of parentheses, write a function to generate all
+combinations of well-formed parentheses.
+Input: n = 2
+Output: ["(())","()()"]
+Leetcode link: https://leetcode.com/problems/generate-parentheses/description/
+"""
+
+
+def backtrack(
+ partial: str, open_count: int, close_count: int, n: int, result: list[str]
+) -> None:
+ """
+ Generate valid combinations of balanced parentheses using recursion.
+
+ :param partial: A string representing the current combination.
+ :param open_count: An integer representing the count of open parentheses.
+ :param close_count: An integer representing the count of close parentheses.
+ :param n: An integer representing the total number of pairs.
+ :param result: A list to store valid combinations.
+ :return: None
+
+ This function uses recursion to explore all possible combinations,
+ ensuring that at each step, the parentheses remain balanced.
+
+ Example:
+ >>> result = []
+ >>> backtrack("", 0, 0, 2, result)
+ >>> result
+ ['(())', '()()']
+ """
+ if len(partial) == 2 * n:
+ # When the combination is complete, add it to the result.
+ result.append(partial)
+ return
+
+ if open_count < n:
+ # If we can add an open parenthesis, do so, and recurse.
+ backtrack(partial + "(", open_count + 1, close_count, n, result)
+
+ if close_count < open_count:
+ # If we can add a close parenthesis (it won't make the combination invalid),
+ # do so, and recurse.
+ backtrack(partial + ")", open_count, close_count + 1, n, result)
+
+
+def generate_parenthesis(n: int) -> list[str]:
+ """
+ Generate valid combinations of balanced parentheses for a given n.
+
+ :param n: An integer representing the number of pairs of parentheses.
+ :return: A list of strings with valid combinations.
+
+ This function uses a recursive approach to generate the combinations.
+
+ Time Complexity: O(2^(2n)) - In the worst case, we have 2^(2n) combinations.
+ Space Complexity: O(n) - where 'n' is the number of pairs.
+
+ Example 1:
+ >>> generate_parenthesis(3)
+ ['((()))', '(()())', '(())()', '()(())', '()()()']
+
+ Example 2:
+ >>> generate_parenthesis(1)
+ ['()']
+ """
+
+ result: list[str] = []
+ backtrack("", 0, 0, n, result)
+ return result
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/backtracking/hamiltonian_cycle.py b/backtracking/hamiltonian_cycle.py
index bf15cce4aca4..f6e4212e47f4 100644
--- a/backtracking/hamiltonian_cycle.py
+++ b/backtracking/hamiltonian_cycle.py
@@ -1,12 +1,11 @@
"""
- A Hamiltonian cycle (Hamiltonian circuit) is a graph cycle
- through a graph that visits each node exactly once.
- Determining whether such paths and cycles exist in graphs
- is the 'Hamiltonian path problem', which is NP-complete.
+A Hamiltonian cycle (Hamiltonian circuit) is a graph cycle
+through a graph that visits each node exactly once.
+Determining whether such paths and cycles exist in graphs
+is the 'Hamiltonian path problem', which is NP-complete.
- Wikipedia: https://en.wikipedia.org/wiki/Hamiltonian_path
+Wikipedia: https://en.wikipedia.org/wiki/Hamiltonian_path
"""
-from __future__ import annotations
def valid_connection(
@@ -16,8 +15,8 @@ def valid_connection(
Checks whether it is possible to add next into path by validating 2 statements
1. There should be path between current and next vertex
2. Next vertex should not be in path
- If both validations succeeds we return True saying that it is possible to connect
- this vertices either we return False
+ If both validations succeed we return True, saying that it is possible to connect
+ this vertices, otherwise we return False
Case 1:Use exact graph as in main function, with initialized values
>>> graph = [[0, 1, 0, 1, 0],
@@ -72,7 +71,7 @@ def util_hamilton_cycle(graph: list[list[int]], path: list[int], curr_ind: int)
>>> curr_ind = 1
>>> util_hamilton_cycle(graph, path, curr_ind)
True
- >>> print(path)
+ >>> path
[0, 1, 2, 4, 3, 0]
Case 2: Use exact graph as in previous case, but in the properties taken from
@@ -86,7 +85,7 @@ def util_hamilton_cycle(graph: list[list[int]], path: list[int], curr_ind: int)
>>> curr_ind = 3
>>> util_hamilton_cycle(graph, path, curr_ind)
True
- >>> print(path)
+ >>> path
[0, 1, 2, 4, 3, 0]
"""
@@ -96,10 +95,10 @@ def util_hamilton_cycle(graph: list[list[int]], path: list[int], curr_ind: int)
return graph[path[curr_ind - 1]][path[0]] == 1
# Recursive Step
- for next in range(0, len(graph)):
- if valid_connection(graph, next, curr_ind, path):
+ for next_ver in range(len(graph)):
+ if valid_connection(graph, next_ver, curr_ind, path):
# Insert current vertex into path as next transition
- path[curr_ind] = next
+ path[curr_ind] = next_ver
# Validate created path
if util_hamilton_cycle(graph, path, curr_ind + 1):
return True
diff --git a/backtracking/knight_tour.py b/backtracking/knight_tour.py
index 2413ba468838..8906aaa1094c 100644
--- a/backtracking/knight_tour.py
+++ b/backtracking/knight_tour.py
@@ -3,7 +3,7 @@
from __future__ import annotations
-def get_valid_pos(position: tuple[int], n: int) -> list[tuple[int]]:
+def get_valid_pos(position: tuple[int, int], n: int) -> list[tuple[int, int]]:
"""
Find all the valid positions a knight can move to from the current position.
@@ -24,10 +24,10 @@ def get_valid_pos(position: tuple[int], n: int) -> list[tuple[int]]:
]
permissible_positions = []
- for position in positions:
- y_test, x_test = position
+ for inner_position in positions:
+ y_test, x_test = inner_position
if 0 <= y_test < n and 0 <= x_test < n:
- permissible_positions.append(position)
+ permissible_positions.append(inner_position)
return permissible_positions
@@ -46,7 +46,9 @@ def is_complete(board: list[list[int]]) -> bool:
return not any(elem == 0 for row in board for elem in row)
-def open_knight_tour_helper(board: list[list[int]], pos: tuple[int], curr: int) -> bool:
+def open_knight_tour_helper(
+ board: list[list[int]], pos: tuple[int, int], curr: int
+) -> bool:
"""
Helper function to solve knight tour problem.
"""
@@ -76,8 +78,8 @@ def open_knight_tour(n: int) -> list[list[int]]:
>>> open_knight_tour(2)
Traceback (most recent call last):
- ...
- ValueError: Open Kight Tour cannot be performed on a board of size 2
+ ...
+ ValueError: Open Knight Tour cannot be performed on a board of size 2
"""
board = [[0 for i in range(n)] for j in range(n)]
@@ -89,7 +91,8 @@ def open_knight_tour(n: int) -> list[list[int]]:
return board
board[i][j] = 0
- raise ValueError(f"Open Kight Tour cannot be performed on a board of size {n}")
+ msg = f"Open Knight Tour cannot be performed on a board of size {n}"
+ raise ValueError(msg)
if __name__ == "__main__":
diff --git a/backtracking/match_word_pattern.py b/backtracking/match_word_pattern.py
new file mode 100644
index 000000000000..bfa9b1354d51
--- /dev/null
+++ b/backtracking/match_word_pattern.py
@@ -0,0 +1,61 @@
+def match_word_pattern(pattern: str, input_string: str) -> bool:
+ """
+ Determine if a given pattern matches a string using backtracking.
+
+ pattern: The pattern to match.
+ input_string: The string to match against the pattern.
+ return: True if the pattern matches the string, False otherwise.
+
+ >>> match_word_pattern("aba", "GraphTreesGraph")
+ True
+
+ >>> match_word_pattern("xyx", "PythonRubyPython")
+ True
+
+ >>> match_word_pattern("GG", "PythonJavaPython")
+ False
+ """
+
+ def backtrack(pattern_index: int, str_index: int) -> bool:
+ """
+ >>> backtrack(0, 0)
+ True
+
+ >>> backtrack(0, 1)
+ True
+
+ >>> backtrack(0, 4)
+ False
+ """
+ if pattern_index == len(pattern) and str_index == len(input_string):
+ return True
+ if pattern_index == len(pattern) or str_index == len(input_string):
+ return False
+ char = pattern[pattern_index]
+ if char in pattern_map:
+ mapped_str = pattern_map[char]
+ if input_string.startswith(mapped_str, str_index):
+ return backtrack(pattern_index + 1, str_index + len(mapped_str))
+ else:
+ return False
+ for end in range(str_index + 1, len(input_string) + 1):
+ substr = input_string[str_index:end]
+ if substr in str_map:
+ continue
+ pattern_map[char] = substr
+ str_map[substr] = char
+ if backtrack(pattern_index + 1, end):
+ return True
+ del pattern_map[char]
+ del str_map[substr]
+ return False
+
+ pattern_map: dict[str, str] = {}
+ str_map: dict[str, str] = {}
+ return backtrack(0, 0)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/backtracking/minimax.py b/backtracking/minimax.py
index 91188090c899..4eef90b75483 100644
--- a/backtracking/minimax.py
+++ b/backtracking/minimax.py
@@ -1,20 +1,38 @@
+"""
+Minimax helps to achieve maximum score in a game by checking all possible moves
+depth is current depth in game tree.
+
+nodeIndex is index of current node in scores[].
+if move is of maximizer return true else false
+leaves of game tree is stored in scores[]
+height is maximum height of Game tree
+"""
+
from __future__ import annotations
import math
-""" Minimax helps to achieve maximum score in a game by checking all possible moves
- depth is current depth in game tree.
- nodeIndex is index of current node in scores[].
- if move is of maximizer return true else false
- leaves of game tree is stored in scores[]
- height is maximum height of Game tree
-"""
-
def minimax(
depth: int, node_index: int, is_max: bool, scores: list[int], height: float
) -> int:
"""
+ This function implements the minimax algorithm, which helps achieve the optimal
+ score for a player in a two-player game by checking all possible moves.
+ If the player is the maximizer, then the score is maximized.
+ If the player is the minimizer, then the score is minimized.
+
+ Parameters:
+ - depth: Current depth in the game tree.
+ - node_index: Index of the current node in the scores list.
+ - is_max: A boolean indicating whether the current move
+ is for the maximizer (True) or minimizer (False).
+ - scores: A list containing the scores of the leaves of the game tree.
+ - height: The maximum height of the game tree.
+
+ Returns:
+ - An integer representing the optimal score for the current player.
+
>>> import math
>>> scores = [90, 23, 6, 33, 21, 65, 123, 34423]
>>> height = math.log(len(scores), 2)
@@ -32,36 +50,40 @@ def minimax(
>>> height = math.log(len(scores), 2)
>>> minimax(0, 0, True, scores, height)
12
- >>> minimax('1', 2, True, [], 2 )
- Traceback (most recent call last):
- ...
- TypeError: '<' not supported between instances of 'str' and 'int'
"""
if depth < 0:
raise ValueError("Depth cannot be less than 0")
-
if len(scores) == 0:
raise ValueError("Scores cannot be empty")
+ # Base case: If the current depth equals the height of the tree,
+ # return the score of the current node.
if depth == height:
return scores[node_index]
+ # If it's the maximizer's turn, choose the maximum score
+ # between the two possible moves.
if is_max:
return max(
minimax(depth + 1, node_index * 2, False, scores, height),
minimax(depth + 1, node_index * 2 + 1, False, scores, height),
)
+ # If it's the minimizer's turn, choose the minimum score
+ # between the two possible moves.
return min(
minimax(depth + 1, node_index * 2, True, scores, height),
minimax(depth + 1, node_index * 2 + 1, True, scores, height),
)
-def main():
+def main() -> None:
+ # Sample scores and height calculation
scores = [90, 23, 6, 33, 21, 65, 123, 34423]
height = math.log(len(scores), 2)
+
+ # Calculate and print the optimal value using the minimax algorithm
print("Optimal value : ", end="")
print(minimax(0, 0, True, scores, height))
diff --git a/backtracking/n_queens.py b/backtracking/n_queens.py
index ca7beb830bba..d10181f319b3 100644
--- a/backtracking/n_queens.py
+++ b/backtracking/n_queens.py
@@ -1,88 +1,101 @@
"""
- The nqueens problem is of placing N queens on a N * N
- chess board such that no queen can attack any other queens placed
- on that chess board.
- This means that one queen cannot have any other queen on its horizontal, vertical and
- diagonal lines.
+The nqueens problem is of placing N queens on a N * N
+chess board such that no queen can attack any other queens placed
+on that chess board.
+This means that one queen cannot have any other queen on its horizontal, vertical and
+diagonal lines.
"""
+
+from __future__ import annotations
+
solution = []
-def isSafe(board, row, column):
+def is_safe(board: list[list[int]], row: int, column: int) -> bool:
"""
This function returns a boolean value True if it is safe to place a queen there
considering the current state of the board.
- Parameters :
- board(2D matrix) : board
- row ,column : coordinates of the cell on a board
+ Parameters:
+ board (2D matrix): The chessboard
+ row, column: Coordinates of the cell on the board
- Returns :
+ Returns:
Boolean Value
+ >>> is_safe([[0, 0, 0], [0, 0, 0], [0, 0, 0]], 1, 1)
+ True
+ >>> is_safe([[0, 1, 0], [0, 0, 0], [0, 0, 0]], 1, 1)
+ False
+ >>> is_safe([[1, 0, 0], [0, 0, 0], [0, 0, 0]], 1, 1)
+ False
+ >>> is_safe([[0, 0, 1], [0, 0, 0], [0, 0, 0]], 1, 1)
+ False
"""
- for i in range(len(board)):
- if board[row][i] == 1:
- return False
- for i in range(len(board)):
- if board[i][column] == 1:
- return False
- for i, j in zip(range(row, -1, -1), range(column, -1, -1)):
- if board[i][j] == 1:
- return False
- for i, j in zip(range(row, -1, -1), range(column, len(board))):
- if board[i][j] == 1:
- return False
- return True
-
-
-def solve(board, row):
+
+ n = len(board) # Size of the board
+
+ # Check if there is any queen in the same upper column,
+ # left upper diagonal and right upper diagonal
+ return (
+ all(board[i][j] != 1 for i, j in zip(range(row), [column] * row))
+ and all(
+ board[i][j] != 1
+ for i, j in zip(range(row - 1, -1, -1), range(column - 1, -1, -1))
+ )
+ and all(
+ board[i][j] != 1
+ for i, j in zip(range(row - 1, -1, -1), range(column + 1, n))
+ )
+ )
+
+
+def solve(board: list[list[int]], row: int) -> bool:
"""
- It creates a state space tree and calls the safe function until it receives a
- False Boolean and terminates that branch and backtracks to the next
+ This function creates a state space tree and calls the safe function until it
+ receives a False Boolean and terminates that branch and backtracks to the next
possible solution branch.
"""
if row >= len(board):
"""
- If the row number exceeds N we have board with a successful combination
+ If the row number exceeds N, we have a board with a successful combination
and that combination is appended to the solution list and the board is printed.
-
"""
solution.append(board)
printboard(board)
print()
- return
+ return True
for i in range(len(board)):
"""
- For every row it iterates through each column to check if it is feasible to
+ For every row, it iterates through each column to check if it is feasible to
place a queen there.
- If all the combinations for that particular branch are successful the board is
+ If all the combinations for that particular branch are successful, the board is
reinitialized for the next possible combination.
"""
- if isSafe(board, row, i):
+ if is_safe(board, row, i):
board[row][i] = 1
solve(board, row + 1)
board[row][i] = 0
return False
-def printboard(board):
+def printboard(board: list[list[int]]) -> None:
"""
Prints the boards that have a successful combination.
"""
for i in range(len(board)):
for j in range(len(board)):
if board[i][j] == 1:
- print("Q", end=" ")
+ print("Q", end=" ") # Queen is present
else:
- print(".", end=" ")
+ print(".", end=" ") # Empty cell
print()
-# n=int(input("The no. of queens"))
+# Number of queens (e.g., n=8 for an 8x8 board)
n = 8
board = [[0 for i in range(n)] for j in range(n)]
solve(board, 0)
-print("The total no. of solutions are :", len(solution))
+print("The total number of solutions are:", len(solution))
diff --git a/backtracking/n_queens_math.py b/backtracking/n_queens_math.py
index 811611971616..287d1f090373 100644
--- a/backtracking/n_queens_math.py
+++ b/backtracking/n_queens_math.py
@@ -1,7 +1,7 @@
r"""
Problem:
-The n queens problem is of placing N queens on a N * N chess board such that no queen
+The n queens problem is: placing N queens on a N * N chess board such that no queen
can attack any other queens placed on that chess board. This means that one queen
cannot have any other queen on its horizontal, vertical and diagonal lines.
@@ -31,7 +31,7 @@
other we know that at least the queens can't attack each other in horizontal and
vertical.
-At this point we have that halfway completed and we will treat the chessboard as a
+At this point we have it halfway completed and we will treat the chessboard as a
Cartesian plane. Hereinafter we are going to remember basic math, so in the school we
learned this formula:
@@ -47,7 +47,7 @@
See::
https://www.enotes.com/homework-help/write-equation-line-that-hits-origin-45-degree-1474860
-Then we have this another formula:
+Then we have this other formula:
Slope intercept:
@@ -59,7 +59,7 @@
y - mx = b
-And like we already have the m values for the angles 45º and 135º, this formula would
+And since we already have the m values for the angles 45º and 135º, this formula would
look like this:
45º: y - (1)x = b
@@ -71,10 +71,11 @@
y = row
x = column
-Applying this two formulas we can check if a queen in some position is being attacked
+Applying these two formulas we can check if a queen in some position is being attacked
for another one or vice versa.
"""
+
from __future__ import annotations
@@ -94,40 +95,32 @@ def depth_first_search(
['. . Q . ', 'Q . . . ', '. . . Q ', '. Q . . ']
"""
- """ Get next row in the current board (possible_board) to fill it with a queen """
+ # Get next row in the current board (possible_board) to fill it with a queen
row = len(possible_board)
- """
- If row is equal to the size of the board it means there are a queen in each row in
- the current board (possible_board)
- """
+ # If row is equal to the size of the board it means there are a queen in each row in
+ # the current board (possible_board)
if row == n:
- """
- We convert the variable possible_board that looks like this: [1, 3, 0, 2] to
- this: ['. Q . . ', '. . . Q ', 'Q . . . ', '. . Q . ']
- """
- possible_board = [". " * i + "Q " + ". " * (n - 1 - i) for i in possible_board]
- boards.append(possible_board)
+ # We convert the variable possible_board that looks like this: [1, 3, 0, 2] to
+ # this: ['. Q . . ', '. . . Q ', 'Q . . . ', '. . Q . ']
+ boards.append([". " * i + "Q " + ". " * (n - 1 - i) for i in possible_board])
return
- """ We iterate each column in the row to find all possible results in each row """
+ # We iterate each column in the row to find all possible results in each row
for col in range(n):
-
- """
- We apply that we learned previously. First we check that in the current board
- (possible_board) there are not other same value because if there is it means
- that there are a collision in vertical. Then we apply the two formulas we
- learned before:
-
- 45º: y - x = b or 45: row - col = b
- 135º: y + x = b or row + col = b.
-
- And we verify if the results of this two formulas not exist in their variables
- respectively. (diagonal_right_collisions, diagonal_left_collisions)
-
- If any or these are True it means there is a collision so we continue to the
- next value in the for loop.
- """
+ # We apply that we learned previously. First we check that in the current board
+ # (possible_board) there are not other same value because if there is it means
+ # that there are a collision in vertical. Then we apply the two formulas we
+ # learned before:
+ #
+ # 45º: y - x = b or 45: row - col = b
+ # 135º: y + x = b or row + col = b.
+ #
+ # And we verify if the results of this two formulas not exist in their variables
+ # respectively. (diagonal_right_collisions, diagonal_left_collisions)
+ #
+ # If any or these are True it means there is a collision so we continue to the
+ # next value in the for loop.
if (
col in possible_board
or row - col in diagonal_right_collisions
@@ -135,21 +128,21 @@ def depth_first_search(
):
continue
- """ If it is False we call dfs function again and we update the inputs """
+ # If it is False we call dfs function again and we update the inputs
depth_first_search(
- possible_board + [col],
- diagonal_right_collisions + [row - col],
- diagonal_left_collisions + [row + col],
+ [*possible_board, col],
+ [*diagonal_right_collisions, row - col],
+ [*diagonal_left_collisions, row + col],
boards,
n,
)
def n_queens_solution(n: int) -> None:
- boards = []
+ boards: list[list[str]] = []
depth_first_search([], [], [], boards, n)
- """ Print all the boards """
+ # Print all the boards
for board in boards:
for column in board:
print(column)
diff --git a/backtracking/power_sum.py b/backtracking/power_sum.py
new file mode 100644
index 000000000000..ee2eac426ec7
--- /dev/null
+++ b/backtracking/power_sum.py
@@ -0,0 +1,91 @@
+"""
+Problem source: https://www.hackerrank.com/challenges/the-power-sum/problem
+Find the number of ways that a given integer X, can be expressed as the sum
+of the Nth powers of unique, natural numbers. For example, if X=13 and N=2.
+We have to find all combinations of unique squares adding up to 13.
+The only solution is 2^2+3^2. Constraints: 1<=X<=1000, 2<=N<=10.
+"""
+
+
+def backtrack(
+ needed_sum: int,
+ power: int,
+ current_number: int,
+ current_sum: int,
+ solutions_count: int,
+) -> tuple[int, int]:
+ """
+ >>> backtrack(13, 2, 1, 0, 0)
+ (0, 1)
+ >>> backtrack(10, 2, 1, 0, 0)
+ (0, 1)
+ >>> backtrack(10, 3, 1, 0, 0)
+ (0, 0)
+ >>> backtrack(20, 2, 1, 0, 0)
+ (0, 1)
+ >>> backtrack(15, 10, 1, 0, 0)
+ (0, 0)
+ >>> backtrack(16, 2, 1, 0, 0)
+ (0, 1)
+ >>> backtrack(20, 1, 1, 0, 0)
+ (0, 64)
+ """
+ if current_sum == needed_sum:
+ # If the sum of the powers is equal to needed_sum, then we have a solution.
+ solutions_count += 1
+ return current_sum, solutions_count
+
+ i_to_n = current_number**power
+ if current_sum + i_to_n <= needed_sum:
+ # If the sum of the powers is less than needed_sum, then continue adding powers.
+ current_sum += i_to_n
+ current_sum, solutions_count = backtrack(
+ needed_sum, power, current_number + 1, current_sum, solutions_count
+ )
+ current_sum -= i_to_n
+ if i_to_n < needed_sum:
+ # If the power of i is less than needed_sum, then try with the next power.
+ current_sum, solutions_count = backtrack(
+ needed_sum, power, current_number + 1, current_sum, solutions_count
+ )
+ return current_sum, solutions_count
+
+
+def solve(needed_sum: int, power: int) -> int:
+ """
+ >>> solve(13, 2)
+ 1
+ >>> solve(10, 2)
+ 1
+ >>> solve(10, 3)
+ 0
+ >>> solve(20, 2)
+ 1
+ >>> solve(15, 10)
+ 0
+ >>> solve(16, 2)
+ 1
+ >>> solve(20, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid input
+ needed_sum must be between 1 and 1000, power between 2 and 10.
+ >>> solve(-10, 5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid input
+ needed_sum must be between 1 and 1000, power between 2 and 10.
+ """
+ if not (1 <= needed_sum <= 1000 and 2 <= power <= 10):
+ raise ValueError(
+ "Invalid input\n"
+ "needed_sum must be between 1 and 1000, power between 2 and 10."
+ )
+
+ return backtrack(needed_sum, power, 1, 0, 0)[1] # Return the solutions_count
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/backtracking/rat_in_maze.py b/backtracking/rat_in_maze.py
index 788aeac13c09..626c83cb4a15 100644
--- a/backtracking/rat_in_maze.py
+++ b/backtracking/rat_in_maze.py
@@ -1,114 +1,197 @@
-def solve_maze(maze: list) -> bool:
+from __future__ import annotations
+
+
+def solve_maze(
+ maze: list[list[int]],
+ source_row: int,
+ source_column: int,
+ destination_row: int,
+ destination_column: int,
+) -> list[list[int]]:
"""
This method solves the "rat in maze" problem.
- In this problem we have some n by n matrix, a start point and an end point.
- We want to go from the start to the end. In this matrix zeroes represent walls
- and ones paths we can use.
Parameters :
- maze(2D matrix) : maze
+ - maze: A two dimensional matrix of zeros and ones.
+ - source_row: The row index of the starting point.
+ - source_column: The column index of the starting point.
+ - destination_row: The row index of the destination point.
+ - destination_column: The column index of the destination point.
Returns:
- Return: True if the maze has a solution or False if it does not.
+ - solution: A 2D matrix representing the solution path if it exists.
+ Raises:
+ - ValueError: If no solution exists or if the source or
+ destination coordinates are invalid.
+ Description:
+ This method navigates through a maze represented as an n by n matrix,
+ starting from a specified source cell and
+ aiming to reach a destination cell.
+ The maze consists of walls (1s) and open paths (0s).
+ By providing custom row and column values, the source and destination
+ cells can be adjusted.
>>> maze = [[0, 1, 0, 1, 1],
... [0, 0, 0, 0, 0],
... [1, 0, 1, 0, 1],
... [0, 0, 1, 0, 0],
... [1, 0, 0, 1, 0]]
- >>> solve_maze(maze)
- [1, 0, 0, 0, 0]
- [1, 1, 1, 1, 0]
- [0, 0, 0, 1, 0]
- [0, 0, 0, 1, 1]
- [0, 0, 0, 0, 1]
- True
+ >>> solve_maze(maze,0,0,len(maze)-1,len(maze)-1) # doctest: +NORMALIZE_WHITESPACE
+ [[0, 1, 1, 1, 1],
+ [0, 0, 0, 0, 1],
+ [1, 1, 1, 0, 1],
+ [1, 1, 1, 0, 0],
+ [1, 1, 1, 1, 0]]
+
+ Note:
+ In the output maze, the zeros (0s) represent one of the possible
+ paths from the source to the destination.
>>> maze = [[0, 1, 0, 1, 1],
... [0, 0, 0, 0, 0],
... [0, 0, 0, 0, 1],
... [0, 0, 0, 0, 0],
... [0, 0, 0, 0, 0]]
- >>> solve_maze(maze)
- [1, 0, 0, 0, 0]
- [1, 0, 0, 0, 0]
- [1, 0, 0, 0, 0]
- [1, 0, 0, 0, 0]
- [1, 1, 1, 1, 1]
- True
+ >>> solve_maze(maze,0,0,len(maze)-1,len(maze)-1) # doctest: +NORMALIZE_WHITESPACE
+ [[0, 1, 1, 1, 1],
+ [0, 1, 1, 1, 1],
+ [0, 1, 1, 1, 1],
+ [0, 1, 1, 1, 1],
+ [0, 0, 0, 0, 0]]
>>> maze = [[0, 0, 0],
... [0, 1, 0],
... [1, 0, 0]]
- >>> solve_maze(maze)
- [1, 1, 1]
- [0, 0, 1]
- [0, 0, 1]
- True
+ >>> solve_maze(maze,0,0,len(maze)-1,len(maze)-1) # doctest: +NORMALIZE_WHITESPACE
+ [[0, 0, 0],
+ [1, 1, 0],
+ [1, 1, 0]]
- >>> maze = [[0, 1, 0],
+ >>> maze = [[1, 0, 0],
... [0, 1, 0],
... [1, 0, 0]]
- >>> solve_maze(maze)
- No solution exists!
- False
+ >>> solve_maze(maze,0,1,len(maze)-1,len(maze)-1) # doctest: +NORMALIZE_WHITESPACE
+ [[1, 0, 0],
+ [1, 1, 0],
+ [1, 1, 0]]
+
+ >>> maze = [[1, 1, 0, 0, 1, 0, 0, 1],
+ ... [1, 0, 1, 0, 0, 1, 1, 1],
+ ... [0, 1, 0, 1, 0, 0, 1, 0],
+ ... [1, 1, 1, 0, 0, 1, 0, 1],
+ ... [0, 1, 0, 0, 1, 0, 1, 1],
+ ... [0, 0, 0, 1, 1, 1, 0, 1],
+ ... [0, 1, 0, 1, 0, 1, 1, 1],
+ ... [1, 1, 0, 0, 0, 0, 0, 1]]
+ >>> solve_maze(maze,0,2,len(maze)-1,2) # doctest: +NORMALIZE_WHITESPACE
+ [[1, 1, 0, 0, 1, 1, 1, 1],
+ [1, 1, 1, 0, 0, 1, 1, 1],
+ [1, 1, 1, 1, 0, 1, 1, 1],
+ [1, 1, 1, 0, 0, 1, 1, 1],
+ [1, 1, 0, 0, 1, 1, 1, 1],
+ [1, 1, 0, 1, 1, 1, 1, 1],
+ [1, 1, 0, 1, 1, 1, 1, 1],
+ [1, 1, 0, 1, 1, 1, 1, 1]]
+ >>> maze = [[1, 0, 0],
+ ... [0, 1, 1],
+ ... [1, 0, 1]]
+ >>> solve_maze(maze,0,1,len(maze)-1,len(maze)-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: No solution exists!
+
+ >>> maze = [[0, 0],
+ ... [1, 1]]
+ >>> solve_maze(maze,0,0,len(maze)-1,len(maze)-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: No solution exists!
>>> maze = [[0, 1],
... [1, 0]]
- >>> solve_maze(maze)
- No solution exists!
- False
+ >>> solve_maze(maze,2,0,len(maze)-1,len(maze)-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid source or destination coordinates
+
+ >>> maze = [[1, 0, 0],
+ ... [0, 1, 0],
+ ... [1, 0, 0]]
+ >>> solve_maze(maze,0,1,len(maze),len(maze)-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid source or destination coordinates
"""
size = len(maze)
+ # Check if source and destination coordinates are Invalid.
+ if not (0 <= source_row <= size - 1 and 0 <= source_column <= size - 1) or (
+ not (0 <= destination_row <= size - 1 and 0 <= destination_column <= size - 1)
+ ):
+ raise ValueError("Invalid source or destination coordinates")
# We need to create solution object to save path.
- solutions = [[0 for _ in range(size)] for _ in range(size)]
- solved = run_maze(maze, 0, 0, solutions)
+ solutions = [[1 for _ in range(size)] for _ in range(size)]
+ solved = run_maze(
+ maze, source_row, source_column, destination_row, destination_column, solutions
+ )
if solved:
- print("\n".join(str(row) for row in solutions))
+ return solutions
else:
- print("No solution exists!")
- return solved
+ raise ValueError("No solution exists!")
-def run_maze(maze, i, j, solutions):
+def run_maze(
+ maze: list[list[int]],
+ i: int,
+ j: int,
+ destination_row: int,
+ destination_column: int,
+ solutions: list[list[int]],
+) -> bool:
"""
This method is recursive starting from (i, j) and going in one of four directions:
up, down, left, right.
If a path is found to destination it returns True otherwise it returns False.
- Parameters:
- maze(2D matrix) : maze
+ Parameters
+ maze: A two dimensional matrix of zeros and ones.
i, j : coordinates of matrix
- solutions(2D matrix) : solutions
+ solutions: A two dimensional matrix of solutions.
Returns:
Boolean if path is found True, Otherwise False.
"""
size = len(maze)
# Final check point.
- if i == j == (size - 1):
- solutions[i][j] = 1
+ if i == destination_row and j == destination_column and maze[i][j] == 0:
+ solutions[i][j] = 0
return True
- lower_flag = (not (i < 0)) and (not (j < 0)) # Check lower bounds
+ lower_flag = (not i < 0) and (not j < 0) # Check lower bounds
upper_flag = (i < size) and (j < size) # Check upper bounds
if lower_flag and upper_flag:
# check for already visited and block points.
- block_flag = (not (solutions[i][j])) and (not (maze[i][j]))
+ block_flag = (solutions[i][j]) and (not maze[i][j])
if block_flag:
# check visited
- solutions[i][j] = 1
+ solutions[i][j] = 0
# check for directions
if (
- run_maze(maze, i + 1, j, solutions)
- or run_maze(maze, i, j + 1, solutions)
- or run_maze(maze, i - 1, j, solutions)
- or run_maze(maze, i, j - 1, solutions)
+ run_maze(maze, i + 1, j, destination_row, destination_column, solutions)
+ or run_maze(
+ maze, i, j + 1, destination_row, destination_column, solutions
+ )
+ or run_maze(
+ maze, i - 1, j, destination_row, destination_column, solutions
+ )
+ or run_maze(
+ maze, i, j - 1, destination_row, destination_column, solutions
+ )
):
return True
- solutions[i][j] = 0
+ solutions[i][j] = 1
return False
+ return False
if __name__ == "__main__":
import doctest
- doctest.testmod()
+ doctest.testmod(optionflags=doctest.NORMALIZE_WHITESPACE)
diff --git a/backtracking/sudoku.py b/backtracking/sudoku.py
index 614bdb8530ac..cabeebb90433 100644
--- a/backtracking/sudoku.py
+++ b/backtracking/sudoku.py
@@ -1,20 +1,21 @@
-from typing import List, Tuple, Union
-
-Matrix = List[List[int]]
-
"""
- Given a partially filled 9×9 2D array, the objective is to fill a 9×9
- square grid with digits numbered 1 to 9, so that every row, column, and
- and each of the nine 3×3 sub-grids contains all of the digits.
-
- This can be solved using Backtracking and is similar to n-queens.
- We check to see if a cell is safe or not and recursively call the
- function on the next column to see if it returns True. if yes, we
- have solved the puzzle. else, we backtrack and place another number
- in that cell and repeat this process.
+Given a partially filled 9x9 2D array, the objective is to fill a 9x9
+square grid with digits numbered 1 to 9, so that every row, column, and
+and each of the nine 3x3 sub-grids contains all of the digits.
+
+This can be solved using Backtracking and is similar to n-queens.
+We check to see if a cell is safe or not and recursively call the
+function on the next column to see if it returns True. if yes, we
+have solved the puzzle. else, we backtrack and place another number
+in that cell and repeat this process.
"""
+
+from __future__ import annotations
+
+Matrix = list[list[int]]
+
# assigning initial values to the grid
-initial_grid = [
+initial_grid: Matrix = [
[3, 0, 6, 5, 0, 8, 4, 0, 0],
[5, 2, 0, 0, 0, 0, 0, 0, 0],
[0, 8, 7, 0, 0, 0, 0, 3, 1],
@@ -27,7 +28,7 @@
]
# a grid with no solution
-no_solution = [
+no_solution: Matrix = [
[5, 0, 6, 5, 0, 8, 4, 0, 3],
[5, 2, 0, 0, 0, 0, 0, 0, 2],
[1, 8, 7, 0, 0, 0, 0, 3, 1],
@@ -48,7 +49,7 @@ def is_safe(grid: Matrix, row: int, column: int, n: int) -> bool:
is found) else returns True if it is 'safe'
"""
for i in range(9):
- if grid[row][i] == n or grid[i][column] == n:
+ if n in {grid[row][i], grid[i][column]}:
return False
for i in range(3):
@@ -59,28 +60,7 @@ def is_safe(grid: Matrix, row: int, column: int, n: int) -> bool:
return True
-def is_completed(grid: Matrix) -> bool:
- """
- This function checks if the puzzle is completed or not.
- it is completed when all the cells are assigned with a non-zero number.
-
- >>> is_completed([[0]])
- False
- >>> is_completed([[1]])
- True
- >>> is_completed([[1, 2], [0, 4]])
- False
- >>> is_completed([[1, 2], [3, 4]])
- True
- >>> is_completed(initial_grid)
- False
- >>> is_completed(no_solution)
- False
- """
- return all(all(cell != 0 for cell in row) for row in grid)
-
-
-def find_empty_location(grid: Matrix) -> Tuple[int, int]:
+def find_empty_location(grid: Matrix) -> tuple[int, int] | None:
"""
This function finds an empty location so that we can assign a number
for that particular row and column.
@@ -89,9 +69,10 @@ def find_empty_location(grid: Matrix) -> Tuple[int, int]:
for j in range(9):
if grid[i][j] == 0:
return i, j
+ return None
-def sudoku(grid: Matrix) -> Union[Matrix, bool]:
+def sudoku(grid: Matrix) -> Matrix | None:
"""
Takes a partially filled-in grid and attempts to assign values to
all unassigned locations in such a way to meet the requirements
@@ -107,25 +88,25 @@ def sudoku(grid: Matrix) -> Union[Matrix, bool]:
[1, 3, 8, 9, 4, 7, 2, 5, 6],
[6, 9, 2, 3, 5, 1, 8, 7, 4],
[7, 4, 5, 2, 8, 6, 3, 1, 9]]
- >>> sudoku(no_solution)
- False
+ >>> sudoku(no_solution) is None
+ True
"""
-
- if is_completed(grid):
+ if location := find_empty_location(grid):
+ row, column = location
+ else:
+ # If the location is ``None``, then the grid is solved.
return grid
- row, column = find_empty_location(grid)
-
for digit in range(1, 10):
if is_safe(grid, row, column, digit):
grid[row][column] = digit
- if sudoku(grid):
+ if sudoku(grid) is not None:
return grid
grid[row][column] = 0
- return False
+ return None
def print_solution(grid: Matrix) -> None:
@@ -141,11 +122,12 @@ def print_solution(grid: Matrix) -> None:
if __name__ == "__main__":
# make a copy of grid so that you can compare with the unmodified grid
- for grid in (initial_grid, no_solution):
- grid = list(map(list, grid))
- solution = sudoku(grid)
- if solution:
- print("grid after solving:")
+ for example_grid in (initial_grid, no_solution):
+ print("\nExample grid:\n" + "=" * 20)
+ print_solution(example_grid)
+ print("\nExample grid solution:")
+ solution = sudoku(example_grid)
+ if solution is not None:
print_solution(solution)
else:
print("Cannot find a solution.")
diff --git a/backtracking/sum_of_subsets.py b/backtracking/sum_of_subsets.py
index 425ddcff927e..f34d3ca34339 100644
--- a/backtracking/sum_of_subsets.py
+++ b/backtracking/sum_of_subsets.py
@@ -1,23 +1,32 @@
"""
- The sum-of-subsetsproblem states that a set of non-negative integers, and a
- value M, determine all possible subsets of the given set whose summation sum
- equal to given M.
+The sum-of-subsetsproblem states that a set of non-negative integers, and a
+value M, determine all possible subsets of the given set whose summation sum
+equal to given M.
- Summation of the chosen numbers must be equal to given number M and one number
- can be used only once.
+Summation of the chosen numbers must be equal to given number M and one number
+can be used only once.
"""
+from __future__ import annotations
-def generate_sum_of_subsets_soln(nums, max_sum):
- result = []
- path = []
+
+def generate_sum_of_subsets_soln(nums: list[int], max_sum: int) -> list[list[int]]:
+ result: list[list[int]] = []
+ path: list[int] = []
num_index = 0
remaining_nums_sum = sum(nums)
create_state_space_tree(nums, max_sum, num_index, path, result, remaining_nums_sum)
return result
-def create_state_space_tree(nums, max_sum, num_index, path, result, remaining_nums_sum):
+def create_state_space_tree(
+ nums: list[int],
+ max_sum: int,
+ num_index: int,
+ path: list[int],
+ result: list[list[int]],
+ remaining_nums_sum: int,
+) -> None:
"""
Creates a state space tree to iterate through each branch using DFS.
It terminates the branching of a node when any of the two conditions
@@ -31,14 +40,14 @@ def create_state_space_tree(nums, max_sum, num_index, path, result, remaining_nu
if sum(path) == max_sum:
result.append(path)
return
- for num_index in range(num_index, len(nums)):
+ for index in range(num_index, len(nums)):
create_state_space_tree(
nums,
max_sum,
- num_index + 1,
- path + [nums[num_index]],
+ index + 1,
+ [*path, nums[index]],
result,
- remaining_nums_sum - nums[num_index],
+ remaining_nums_sum - nums[index],
)
diff --git a/backtracking/word_break.py b/backtracking/word_break.py
new file mode 100644
index 000000000000..1f2ab073f499
--- /dev/null
+++ b/backtracking/word_break.py
@@ -0,0 +1,71 @@
+"""
+Word Break Problem is a well-known problem in computer science.
+Given a string and a dictionary of words, the task is to determine if
+the string can be segmented into a sequence of one or more dictionary words.
+
+Wikipedia: https://en.wikipedia.org/wiki/Word_break_problem
+"""
+
+
+def backtrack(input_string: str, word_dict: set[str], start: int) -> bool:
+ """
+ Helper function that uses backtracking to determine if a valid
+ word segmentation is possible starting from index 'start'.
+
+ Parameters:
+ input_string (str): The input string to be segmented.
+ word_dict (set[str]): A set of valid dictionary words.
+ start (int): The starting index of the substring to be checked.
+
+ Returns:
+ bool: True if a valid segmentation is possible, otherwise False.
+
+ Example:
+ >>> backtrack("leetcode", {"leet", "code"}, 0)
+ True
+
+ >>> backtrack("applepenapple", {"apple", "pen"}, 0)
+ True
+
+ >>> backtrack("catsandog", {"cats", "dog", "sand", "and", "cat"}, 0)
+ False
+ """
+
+ # Base case: if the starting index has reached the end of the string
+ if start == len(input_string):
+ return True
+
+ # Try every possible substring from 'start' to 'end'
+ for end in range(start + 1, len(input_string) + 1):
+ if input_string[start:end] in word_dict and backtrack(
+ input_string, word_dict, end
+ ):
+ return True
+
+ return False
+
+
+def word_break(input_string: str, word_dict: set[str]) -> bool:
+ """
+ Determines if the input string can be segmented into a sequence of
+ valid dictionary words using backtracking.
+
+ Parameters:
+ input_string (str): The input string to segment.
+ word_dict (set[str]): The set of valid words.
+
+ Returns:
+ bool: True if the string can be segmented into valid words, otherwise False.
+
+ Example:
+ >>> word_break("leetcode", {"leet", "code"})
+ True
+
+ >>> word_break("applepenapple", {"apple", "pen"})
+ True
+
+ >>> word_break("catsandog", {"cats", "dog", "sand", "and", "cat"})
+ False
+ """
+
+ return backtrack(input_string, word_dict, 0)
diff --git a/backtracking/word_ladder.py b/backtracking/word_ladder.py
new file mode 100644
index 000000000000..7d9fd00f6669
--- /dev/null
+++ b/backtracking/word_ladder.py
@@ -0,0 +1,100 @@
+"""
+Word Ladder is a classic problem in computer science.
+The problem is to transform a start word into an end word
+by changing one letter at a time.
+Each intermediate word must be a valid word from a given list of words.
+The goal is to find a transformation sequence
+from the start word to the end word.
+
+Wikipedia: https://en.wikipedia.org/wiki/Word_ladder
+"""
+
+import string
+
+
+def backtrack(
+ current_word: str, path: list[str], end_word: str, word_set: set[str]
+) -> list[str]:
+ """
+ Helper function to perform backtracking to find the transformation
+ from the current_word to the end_word.
+
+ Parameters:
+ current_word (str): The current word in the transformation sequence.
+ path (list[str]): The list of transformations from begin_word to current_word.
+ end_word (str): The target word for transformation.
+ word_set (set[str]): The set of valid words for transformation.
+
+ Returns:
+ list[str]: The list of transformations from begin_word to end_word.
+ Returns an empty list if there is no valid
+ transformation from current_word to end_word.
+
+ Example:
+ >>> backtrack("hit", ["hit"], "cog", {"hot", "dot", "dog", "lot", "log", "cog"})
+ ['hit', 'hot', 'dot', 'lot', 'log', 'cog']
+
+ >>> backtrack("hit", ["hit"], "cog", {"hot", "dot", "dog", "lot", "log"})
+ []
+
+ >>> backtrack("lead", ["lead"], "gold", {"load", "goad", "gold", "lead", "lord"})
+ ['lead', 'lead', 'load', 'goad', 'gold']
+
+ >>> backtrack("game", ["game"], "code", {"came", "cage", "code", "cade", "gave"})
+ ['game', 'came', 'cade', 'code']
+ """
+
+ # Base case: If the current word is the end word, return the path
+ if current_word == end_word:
+ return path
+
+ # Try all possible single-letter transformations
+ for i in range(len(current_word)):
+ for c in string.ascii_lowercase: # Try changing each letter
+ transformed_word = current_word[:i] + c + current_word[i + 1 :]
+ if transformed_word in word_set:
+ word_set.remove(transformed_word)
+ # Recur with the new word added to the path
+ result = backtrack(
+ transformed_word, [*path, transformed_word], end_word, word_set
+ )
+ if result: # valid transformation found
+ return result
+ word_set.add(transformed_word) # backtrack
+
+ return [] # No valid transformation found
+
+
+def word_ladder(begin_word: str, end_word: str, word_set: set[str]) -> list[str]:
+ """
+ Solve the Word Ladder problem using Backtracking and return
+ the list of transformations from begin_word to end_word.
+
+ Parameters:
+ begin_word (str): The word from which the transformation starts.
+ end_word (str): The target word for transformation.
+ word_list (list[str]): The list of valid words for transformation.
+
+ Returns:
+ list[str]: The list of transformations from begin_word to end_word.
+ Returns an empty list if there is no valid transformation.
+
+ Example:
+ >>> word_ladder("hit", "cog", ["hot", "dot", "dog", "lot", "log", "cog"])
+ ['hit', 'hot', 'dot', 'lot', 'log', 'cog']
+
+ >>> word_ladder("hit", "cog", ["hot", "dot", "dog", "lot", "log"])
+ []
+
+ >>> word_ladder("lead", "gold", ["load", "goad", "gold", "lead", "lord"])
+ ['lead', 'lead', 'load', 'goad', 'gold']
+
+ >>> word_ladder("game", "code", ["came", "cage", "code", "cade", "gave"])
+ ['game', 'came', 'cade', 'code']
+ """
+
+ if end_word not in word_set: # no valid transformation possible
+ return []
+
+ # Perform backtracking starting from the begin_word
+ return backtrack(begin_word, [begin_word], end_word, word_set)
diff --git a/backtracking/word_search.py b/backtracking/word_search.py
new file mode 100644
index 000000000000..8a9b2f1b5359
--- /dev/null
+++ b/backtracking/word_search.py
@@ -0,0 +1,162 @@
+"""
+Author : Alexander Pantyukhin
+Date : November 24, 2022
+
+Task:
+Given an m x n grid of characters board and a string word,
+return true if word exists in the grid.
+
+The word can be constructed from letters of sequentially adjacent cells,
+where adjacent cells are horizontally or vertically neighboring.
+The same letter cell may not be used more than once.
+
+Example:
+
+Matrix:
+---------
+|A|B|C|E|
+|S|F|C|S|
+|A|D|E|E|
+---------
+
+Word:
+"ABCCED"
+
+Result:
+True
+
+Implementation notes: Use backtracking approach.
+At each point, check all neighbors to try to find the next letter of the word.
+
+leetcode: https://leetcode.com/problems/word-search/
+
+"""
+
+
+def get_point_key(len_board: int, len_board_column: int, row: int, column: int) -> int:
+ """
+ Returns the hash key of matrix indexes.
+
+ >>> get_point_key(10, 20, 1, 0)
+ 200
+ """
+
+ return len_board * len_board_column * row + column
+
+
+def exits_word(
+ board: list[list[str]],
+ word: str,
+ row: int,
+ column: int,
+ word_index: int,
+ visited_points_set: set[int],
+) -> bool:
+ """
+ Return True if it's possible to search the word suffix
+ starting from the word_index.
+
+ >>> exits_word([["A"]], "B", 0, 0, 0, set())
+ False
+ """
+
+ if board[row][column] != word[word_index]:
+ return False
+
+ if word_index == len(word) - 1:
+ return True
+
+ traverts_directions = [(0, 1), (0, -1), (-1, 0), (1, 0)]
+ len_board = len(board)
+ len_board_column = len(board[0])
+ for direction in traverts_directions:
+ next_i = row + direction[0]
+ next_j = column + direction[1]
+ if not (0 <= next_i < len_board and 0 <= next_j < len_board_column):
+ continue
+
+ key = get_point_key(len_board, len_board_column, next_i, next_j)
+ if key in visited_points_set:
+ continue
+
+ visited_points_set.add(key)
+ if exits_word(board, word, next_i, next_j, word_index + 1, visited_points_set):
+ return True
+
+ visited_points_set.remove(key)
+
+ return False
+
+
+def word_exists(board: list[list[str]], word: str) -> bool:
+ """
+ >>> word_exists([["A","B","C","E"],["S","F","C","S"],["A","D","E","E"]], "ABCCED")
+ True
+ >>> word_exists([["A","B","C","E"],["S","F","C","S"],["A","D","E","E"]], "SEE")
+ True
+ >>> word_exists([["A","B","C","E"],["S","F","C","S"],["A","D","E","E"]], "ABCB")
+ False
+ >>> word_exists([["A"]], "A")
+ True
+ >>> word_exists([["B", "A", "A"], ["A", "A", "A"], ["A", "B", "A"]], "ABB")
+ False
+ >>> word_exists([["A"]], 123)
+ Traceback (most recent call last):
+ ...
+ ValueError: The word parameter should be a string of length greater than 0.
+ >>> word_exists([["A"]], "")
+ Traceback (most recent call last):
+ ...
+ ValueError: The word parameter should be a string of length greater than 0.
+ >>> word_exists([[]], "AB")
+ Traceback (most recent call last):
+ ...
+ ValueError: The board should be a non empty matrix of single chars strings.
+ >>> word_exists([], "AB")
+ Traceback (most recent call last):
+ ...
+ ValueError: The board should be a non empty matrix of single chars strings.
+ >>> word_exists([["A"], [21]], "AB")
+ Traceback (most recent call last):
+ ...
+ ValueError: The board should be a non empty matrix of single chars strings.
+ """
+
+ # Validate board
+ board_error_message = (
+ "The board should be a non empty matrix of single chars strings."
+ )
+
+ len_board = len(board)
+ if not isinstance(board, list) or len(board) == 0:
+ raise ValueError(board_error_message)
+
+ for row in board:
+ if not isinstance(row, list) or len(row) == 0:
+ raise ValueError(board_error_message)
+
+ for item in row:
+ if not isinstance(item, str) or len(item) != 1:
+ raise ValueError(board_error_message)
+
+ # Validate word
+ if not isinstance(word, str) or len(word) == 0:
+ raise ValueError(
+ "The word parameter should be a string of length greater than 0."
+ )
+
+ len_board_column = len(board[0])
+ for i in range(len_board):
+ for j in range(len_board_column):
+ if exits_word(
+ board, word, i, j, 0, {get_point_key(len_board, len_board_column, i, j)}
+ ):
+ return True
+
+ return False
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/README.md b/bit_manipulation/README.md
index 2ef1661524f2..3f5e028beb8e 100644
--- a/bit_manipulation/README.md
+++ b/bit_manipulation/README.md
@@ -1,7 +1,11 @@
-https://docs.python.org/3/reference/expressions.html#binary-bitwise-operations
-https://docs.python.org/3/reference/expressions.html#unary-arithmetic-and-bitwise-operations
-https://docs.python.org/3/library/stdtypes.html#bitwise-operations-on-integer-types
+# Bit manipulation
-https://wiki.python.org/moin/BitManipulation
-https://wiki.python.org/moin/BitwiseOperators
-https://www.tutorialspoint.com/python3/bitwise_operators_example.htm
+Bit manipulation is the act of manipulating bits to detect errors (hamming code), encrypts and decrypts messages (more on that in the 'ciphers' folder) or just do anything at the lowest level of your computer.
+
+*
+*
+*
+*
+*
+*
+*
diff --git a/bit_manipulation/binary_and_operator.py b/bit_manipulation/binary_and_operator.py
index f1b910f8cc9b..f33b8b1c0ab4 100644
--- a/bit_manipulation/binary_and_operator.py
+++ b/bit_manipulation/binary_and_operator.py
@@ -1,7 +1,7 @@
# https://www.tutorialspoint.com/python3/bitwise_operators_example.htm
-def binary_and(a: int, b: int):
+def binary_and(a: int, b: int) -> str:
"""
Take in 2 integers, convert them to binary,
return a binary number that is the
@@ -22,21 +22,21 @@ def binary_and(a: int, b: int):
>>> binary_and(0, -1)
Traceback (most recent call last):
...
- ValueError: the value of both input must be positive
+ ValueError: the value of both inputs must be positive
>>> binary_and(0, 1.1)
Traceback (most recent call last):
...
- TypeError: 'float' object cannot be interpreted as an integer
+ ValueError: Unknown format code 'b' for object of type 'float'
>>> binary_and("0", "1")
Traceback (most recent call last):
...
TypeError: '<' not supported between instances of 'str' and 'int'
"""
if a < 0 or b < 0:
- raise ValueError("the value of both input must be positive")
+ raise ValueError("the value of both inputs must be positive")
- a_binary = str(bin(a))[2:] # remove the leading "0b"
- b_binary = str(bin(b))[2:] # remove the leading "0b"
+ a_binary = format(a, "b")
+ b_binary = format(b, "b")
max_len = max(len(a_binary), len(b_binary))
diff --git a/bit_manipulation/binary_coded_decimal.py b/bit_manipulation/binary_coded_decimal.py
new file mode 100644
index 000000000000..676fd6d54fc5
--- /dev/null
+++ b/bit_manipulation/binary_coded_decimal.py
@@ -0,0 +1,29 @@
+def binary_coded_decimal(number: int) -> str:
+ """
+ Find binary coded decimal (bcd) of integer base 10.
+ Each digit of the number is represented by a 4-bit binary.
+ Example:
+ >>> binary_coded_decimal(-2)
+ '0b0000'
+ >>> binary_coded_decimal(-1)
+ '0b0000'
+ >>> binary_coded_decimal(0)
+ '0b0000'
+ >>> binary_coded_decimal(3)
+ '0b0011'
+ >>> binary_coded_decimal(2)
+ '0b0010'
+ >>> binary_coded_decimal(12)
+ '0b00010010'
+ >>> binary_coded_decimal(987)
+ '0b100110000111'
+ """
+ return "0b" + "".join(
+ str(bin(int(digit)))[2:].zfill(4) for digit in str(max(0, number))
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/binary_count_setbits.py b/bit_manipulation/binary_count_setbits.py
new file mode 100644
index 000000000000..3c92694533aa
--- /dev/null
+++ b/bit_manipulation/binary_count_setbits.py
@@ -0,0 +1,41 @@
+def binary_count_setbits(a: int) -> int:
+ """
+ Take in 1 integer, return a number that is
+ the number of 1's in binary representation of that number.
+
+ >>> binary_count_setbits(25)
+ 3
+ >>> binary_count_setbits(36)
+ 2
+ >>> binary_count_setbits(16)
+ 1
+ >>> binary_count_setbits(58)
+ 4
+ >>> binary_count_setbits(4294967295)
+ 32
+ >>> binary_count_setbits(0)
+ 0
+ >>> binary_count_setbits(-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value must be a positive integer
+ >>> binary_count_setbits(0.8)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be a 'int' type
+ >>> binary_count_setbits("0")
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'str' and 'int'
+ """
+ if a < 0:
+ raise ValueError("Input value must be a positive integer")
+ elif isinstance(a, float):
+ raise TypeError("Input value must be a 'int' type")
+ return bin(a).count("1")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/binary_count_trailing_zeros.py b/bit_manipulation/binary_count_trailing_zeros.py
new file mode 100644
index 000000000000..f401c4ab9266
--- /dev/null
+++ b/bit_manipulation/binary_count_trailing_zeros.py
@@ -0,0 +1,44 @@
+from math import log2
+
+
+def binary_count_trailing_zeros(a: int) -> int:
+ """
+ Take in 1 integer, return a number that is
+ the number of trailing zeros in binary representation of that number.
+
+ >>> binary_count_trailing_zeros(25)
+ 0
+ >>> binary_count_trailing_zeros(36)
+ 2
+ >>> binary_count_trailing_zeros(16)
+ 4
+ >>> binary_count_trailing_zeros(58)
+ 1
+ >>> binary_count_trailing_zeros(4294967296)
+ 32
+ >>> binary_count_trailing_zeros(0)
+ 0
+ >>> binary_count_trailing_zeros(-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value must be a positive integer
+ >>> binary_count_trailing_zeros(0.8)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be a 'int' type
+ >>> binary_count_trailing_zeros("0")
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'str' and 'int'
+ """
+ if a < 0:
+ raise ValueError("Input value must be a positive integer")
+ elif isinstance(a, float):
+ raise TypeError("Input value must be a 'int' type")
+ return 0 if (a == 0) else int(log2(a & -a))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/binary_or_operator.py b/bit_manipulation/binary_or_operator.py
index e83a86d6a8bc..95f61f1da64e 100644
--- a/bit_manipulation/binary_or_operator.py
+++ b/bit_manipulation/binary_or_operator.py
@@ -1,7 +1,7 @@
# https://www.tutorialspoint.com/python3/bitwise_operators_example.htm
-def binary_or(a: int, b: int):
+def binary_or(a: int, b: int) -> str:
"""
Take in 2 integers, convert them to binary, and return a binary number that is the
result of a binary or operation on the integers provided.
@@ -21,7 +21,7 @@ def binary_or(a: int, b: int):
>>> binary_or(0, -1)
Traceback (most recent call last):
...
- ValueError: the value of both input must be positive
+ ValueError: the value of both inputs must be positive
>>> binary_or(0, 1.1)
Traceback (most recent call last):
...
@@ -32,7 +32,7 @@ def binary_or(a: int, b: int):
TypeError: '<' not supported between instances of 'str' and 'int'
"""
if a < 0 or b < 0:
- raise ValueError("the value of both input must be positive")
+ raise ValueError("the value of both inputs must be positive")
a_binary = str(bin(a))[2:] # remove the leading "0b"
b_binary = str(bin(b))[2:]
max_len = max(len(a_binary), len(b_binary))
diff --git a/bit_manipulation/binary_shifts.py b/bit_manipulation/binary_shifts.py
new file mode 100644
index 000000000000..50dc27ffeef6
--- /dev/null
+++ b/bit_manipulation/binary_shifts.py
@@ -0,0 +1,109 @@
+# Information on binary shifts:
+# https://docs.python.org/3/library/stdtypes.html#bitwise-operations-on-integer-types
+# https://www.interviewcake.com/concept/java/bit-shift
+
+
+def logical_left_shift(number: int, shift_amount: int) -> str:
+ """
+ Take in 2 positive integers.
+ 'number' is the integer to be logically left shifted 'shift_amount' times.
+ i.e. (number << shift_amount)
+ Return the shifted binary representation.
+
+ >>> logical_left_shift(0, 1)
+ '0b00'
+ >>> logical_left_shift(1, 1)
+ '0b10'
+ >>> logical_left_shift(1, 5)
+ '0b100000'
+ >>> logical_left_shift(17, 2)
+ '0b1000100'
+ >>> logical_left_shift(1983, 4)
+ '0b111101111110000'
+ >>> logical_left_shift(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: both inputs must be positive integers
+ """
+ if number < 0 or shift_amount < 0:
+ raise ValueError("both inputs must be positive integers")
+
+ binary_number = str(bin(number))
+ binary_number += "0" * shift_amount
+ return binary_number
+
+
+def logical_right_shift(number: int, shift_amount: int) -> str:
+ """
+ Take in positive 2 integers.
+ 'number' is the integer to be logically right shifted 'shift_amount' times.
+ i.e. (number >>> shift_amount)
+ Return the shifted binary representation.
+
+ >>> logical_right_shift(0, 1)
+ '0b0'
+ >>> logical_right_shift(1, 1)
+ '0b0'
+ >>> logical_right_shift(1, 5)
+ '0b0'
+ >>> logical_right_shift(17, 2)
+ '0b100'
+ >>> logical_right_shift(1983, 4)
+ '0b1111011'
+ >>> logical_right_shift(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: both inputs must be positive integers
+ """
+ if number < 0 or shift_amount < 0:
+ raise ValueError("both inputs must be positive integers")
+
+ binary_number = str(bin(number))[2:]
+ if shift_amount >= len(binary_number):
+ return "0b0"
+ shifted_binary_number = binary_number[: len(binary_number) - shift_amount]
+ return "0b" + shifted_binary_number
+
+
+def arithmetic_right_shift(number: int, shift_amount: int) -> str:
+ """
+ Take in 2 integers.
+ 'number' is the integer to be arithmetically right shifted 'shift_amount' times.
+ i.e. (number >> shift_amount)
+ Return the shifted binary representation.
+
+ >>> arithmetic_right_shift(0, 1)
+ '0b00'
+ >>> arithmetic_right_shift(1, 1)
+ '0b00'
+ >>> arithmetic_right_shift(-1, 1)
+ '0b11'
+ >>> arithmetic_right_shift(17, 2)
+ '0b000100'
+ >>> arithmetic_right_shift(-17, 2)
+ '0b111011'
+ >>> arithmetic_right_shift(-1983, 4)
+ '0b111110000100'
+ """
+ if number >= 0: # Get binary representation of positive number
+ binary_number = "0" + str(bin(number)).strip("-")[2:]
+ else: # Get binary (2's complement) representation of negative number
+ binary_number_length = len(bin(number)[3:]) # Find 2's complement of number
+ binary_number = bin(abs(number) - (1 << binary_number_length))[3:]
+ binary_number = (
+ "1" + "0" * (binary_number_length - len(binary_number)) + binary_number
+ )
+
+ if shift_amount >= len(binary_number):
+ return "0b" + binary_number[0] * len(binary_number)
+ return (
+ "0b"
+ + binary_number[0] * shift_amount
+ + binary_number[: len(binary_number) - shift_amount]
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/binary_twos_complement.py b/bit_manipulation/binary_twos_complement.py
new file mode 100644
index 000000000000..2c064ec142d7
--- /dev/null
+++ b/bit_manipulation/binary_twos_complement.py
@@ -0,0 +1,43 @@
+# Information on 2's complement: https://en.wikipedia.org/wiki/Two%27s_complement
+
+
+def twos_complement(number: int) -> str:
+ """
+ Take in a negative integer 'number'.
+ Return the two's complement representation of 'number'.
+
+ >>> twos_complement(0)
+ '0b0'
+ >>> twos_complement(-1)
+ '0b11'
+ >>> twos_complement(-5)
+ '0b1011'
+ >>> twos_complement(-17)
+ '0b101111'
+ >>> twos_complement(-207)
+ '0b100110001'
+ >>> twos_complement(1)
+ Traceback (most recent call last):
+ ...
+ ValueError: input must be a negative integer
+ """
+ if number > 0:
+ raise ValueError("input must be a negative integer")
+ binary_number_length = len(bin(number)[3:])
+ twos_complement_number = bin(abs(number) - (1 << binary_number_length))[3:]
+ twos_complement_number = (
+ (
+ "1"
+ + "0" * (binary_number_length - len(twos_complement_number))
+ + twos_complement_number
+ )
+ if number < 0
+ else "0"
+ )
+ return "0b" + twos_complement_number
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/binary_xor_operator.py b/bit_manipulation/binary_xor_operator.py
index 0edf2ba6606d..6206c70a99f6 100644
--- a/bit_manipulation/binary_xor_operator.py
+++ b/bit_manipulation/binary_xor_operator.py
@@ -1,7 +1,7 @@
# https://www.tutorialspoint.com/python3/bitwise_operators_example.htm
-def binary_xor(a: int, b: int):
+def binary_xor(a: int, b: int) -> str:
"""
Take in 2 integers, convert them to binary,
return a binary number that is the
@@ -22,7 +22,7 @@ def binary_xor(a: int, b: int):
>>> binary_xor(0, -1)
Traceback (most recent call last):
...
- ValueError: the value of both input must be positive
+ ValueError: the value of both inputs must be positive
>>> binary_xor(0, 1.1)
Traceback (most recent call last):
...
@@ -33,7 +33,7 @@ def binary_xor(a: int, b: int):
TypeError: '<' not supported between instances of 'str' and 'int'
"""
if a < 0 or b < 0:
- raise ValueError("the value of both input must be positive")
+ raise ValueError("the value of both inputs must be positive")
a_binary = str(bin(a))[2:] # remove the leading "0b"
b_binary = str(bin(b))[2:] # remove the leading "0b"
diff --git a/bit_manipulation/bitwise_addition_recursive.py b/bit_manipulation/bitwise_addition_recursive.py
new file mode 100644
index 000000000000..70eaf6887b64
--- /dev/null
+++ b/bit_manipulation/bitwise_addition_recursive.py
@@ -0,0 +1,55 @@
+"""
+Calculates the sum of two non-negative integers using bitwise operators
+Wikipedia explanation: https://en.wikipedia.org/wiki/Binary_number
+"""
+
+
+def bitwise_addition_recursive(number: int, other_number: int) -> int:
+ """
+ >>> bitwise_addition_recursive(4, 5)
+ 9
+ >>> bitwise_addition_recursive(8, 9)
+ 17
+ >>> bitwise_addition_recursive(0, 4)
+ 4
+ >>> bitwise_addition_recursive(4.5, 9)
+ Traceback (most recent call last):
+ ...
+ TypeError: Both arguments MUST be integers!
+ >>> bitwise_addition_recursive('4', 9)
+ Traceback (most recent call last):
+ ...
+ TypeError: Both arguments MUST be integers!
+ >>> bitwise_addition_recursive('4.5', 9)
+ Traceback (most recent call last):
+ ...
+ TypeError: Both arguments MUST be integers!
+ >>> bitwise_addition_recursive(-1, 9)
+ Traceback (most recent call last):
+ ...
+ ValueError: Both arguments MUST be non-negative!
+ >>> bitwise_addition_recursive(1, -9)
+ Traceback (most recent call last):
+ ...
+ ValueError: Both arguments MUST be non-negative!
+ """
+
+ if not isinstance(number, int) or not isinstance(other_number, int):
+ raise TypeError("Both arguments MUST be integers!")
+
+ if number < 0 or other_number < 0:
+ raise ValueError("Both arguments MUST be non-negative!")
+
+ bitwise_sum = number ^ other_number
+ carry = number & other_number
+
+ if carry == 0:
+ return bitwise_sum
+
+ return bitwise_addition_recursive(bitwise_sum, carry << 1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/count_1s_brian_kernighan_method.py b/bit_manipulation/count_1s_brian_kernighan_method.py
new file mode 100644
index 000000000000..2ed81b09d675
--- /dev/null
+++ b/bit_manipulation/count_1s_brian_kernighan_method.py
@@ -0,0 +1,46 @@
+def get_1s_count(number: int) -> int:
+ """
+ Count the number of set bits in a 32 bit integer using Brian Kernighan's way.
+ Ref - https://graphics.stanford.edu/~seander/bithacks.html#CountBitsSetKernighan
+ >>> get_1s_count(25)
+ 3
+ >>> get_1s_count(37)
+ 3
+ >>> get_1s_count(21)
+ 3
+ >>> get_1s_count(58)
+ 4
+ >>> get_1s_count(0)
+ 0
+ >>> get_1s_count(256)
+ 1
+ >>> get_1s_count(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ >>> get_1s_count(0.8)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ >>> get_1s_count("25")
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ """
+ if not isinstance(number, int) or number < 0:
+ raise ValueError("Input must be a non-negative integer")
+
+ count = 0
+ while number:
+ # This way we arrive at next set bit (next 1) instead of looping
+ # through each bit and checking for 1s hence the
+ # loop won't run 32 times it will only run the number of `1` times
+ number &= number - 1
+ count += 1
+ return count
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/count_number_of_one_bits.py b/bit_manipulation/count_number_of_one_bits.py
new file mode 100644
index 000000000000..f0c9f927620a
--- /dev/null
+++ b/bit_manipulation/count_number_of_one_bits.py
@@ -0,0 +1,93 @@
+from timeit import timeit
+
+
+def get_set_bits_count_using_brian_kernighans_algorithm(number: int) -> int:
+ """
+ Count the number of set bits in a 32 bit integer
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(25)
+ 3
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(37)
+ 3
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(21)
+ 3
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(58)
+ 4
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(0)
+ 0
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(256)
+ 1
+ >>> get_set_bits_count_using_brian_kernighans_algorithm(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: the value of input must not be negative
+ """
+ if number < 0:
+ raise ValueError("the value of input must not be negative")
+ result = 0
+ while number:
+ number &= number - 1
+ result += 1
+ return result
+
+
+def get_set_bits_count_using_modulo_operator(number: int) -> int:
+ """
+ Count the number of set bits in a 32 bit integer
+ >>> get_set_bits_count_using_modulo_operator(25)
+ 3
+ >>> get_set_bits_count_using_modulo_operator(37)
+ 3
+ >>> get_set_bits_count_using_modulo_operator(21)
+ 3
+ >>> get_set_bits_count_using_modulo_operator(58)
+ 4
+ >>> get_set_bits_count_using_modulo_operator(0)
+ 0
+ >>> get_set_bits_count_using_modulo_operator(256)
+ 1
+ >>> get_set_bits_count_using_modulo_operator(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: the value of input must not be negative
+ """
+ if number < 0:
+ raise ValueError("the value of input must not be negative")
+ result = 0
+ while number:
+ if number % 2 == 1:
+ result += 1
+ number >>= 1
+ return result
+
+
+def benchmark() -> None:
+ """
+ Benchmark code for comparing 2 functions, with different length int values.
+ Brian Kernighan's algorithm is consistently faster than using modulo_operator.
+ """
+
+ def do_benchmark(number: int) -> None:
+ setup = "import __main__ as z"
+ print(f"Benchmark when {number = }:")
+ print(f"{get_set_bits_count_using_modulo_operator(number) = }")
+ timing = timeit(
+ f"z.get_set_bits_count_using_modulo_operator({number})", setup=setup
+ )
+ print(f"timeit() runs in {timing} seconds")
+ print(f"{get_set_bits_count_using_brian_kernighans_algorithm(number) = }")
+ timing = timeit(
+ f"z.get_set_bits_count_using_brian_kernighans_algorithm({number})",
+ setup=setup,
+ )
+ print(f"timeit() runs in {timing} seconds")
+
+ for number in (25, 37, 58, 0):
+ do_benchmark(number)
+ print()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ benchmark()
diff --git a/bit_manipulation/excess_3_code.py b/bit_manipulation/excess_3_code.py
new file mode 100644
index 000000000000..7beaabd90e8a
--- /dev/null
+++ b/bit_manipulation/excess_3_code.py
@@ -0,0 +1,27 @@
+def excess_3_code(number: int) -> str:
+ """
+ Find excess-3 code of integer base 10.
+ Add 3 to all digits in a decimal number then convert to a binary-coded decimal.
+ https://en.wikipedia.org/wiki/Excess-3
+
+ >>> excess_3_code(0)
+ '0b0011'
+ >>> excess_3_code(3)
+ '0b0110'
+ >>> excess_3_code(2)
+ '0b0101'
+ >>> excess_3_code(20)
+ '0b01010011'
+ >>> excess_3_code(120)
+ '0b010001010011'
+ """
+ num = ""
+ for digit in str(max(0, number)):
+ num += str(bin(int(digit) + 3))[2:].zfill(4)
+ return "0b" + num
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/find_previous_power_of_two.py b/bit_manipulation/find_previous_power_of_two.py
new file mode 100644
index 000000000000..8ac74ac98478
--- /dev/null
+++ b/bit_manipulation/find_previous_power_of_two.py
@@ -0,0 +1,30 @@
+def find_previous_power_of_two(number: int) -> int:
+ """
+ Find the largest power of two that is less than or equal to a given integer.
+ https://stackoverflow.com/questions/1322510
+
+ >>> [find_previous_power_of_two(i) for i in range(18)]
+ [0, 1, 2, 2, 4, 4, 4, 4, 8, 8, 8, 8, 8, 8, 8, 8, 16, 16]
+ >>> find_previous_power_of_two(-5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ >>> find_previous_power_of_two(10.5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ """
+ if not isinstance(number, int) or number < 0:
+ raise ValueError("Input must be a non-negative integer")
+ if number == 0:
+ return 0
+ power = 1
+ while power <= number:
+ power <<= 1 # Equivalent to multiplying by 2
+ return power >> 1 if number > 1 else 1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/gray_code_sequence.py b/bit_manipulation/gray_code_sequence.py
new file mode 100644
index 000000000000..636578d89754
--- /dev/null
+++ b/bit_manipulation/gray_code_sequence.py
@@ -0,0 +1,94 @@
+def gray_code(bit_count: int) -> list:
+ """
+ Takes in an integer n and returns a n-bit
+ gray code sequence
+ An n-bit gray code sequence is a sequence of 2^n
+ integers where:
+
+ a) Every integer is between [0,2^n -1] inclusive
+ b) The sequence begins with 0
+ c) An integer appears at most one times in the sequence
+ d)The binary representation of every pair of integers differ
+ by exactly one bit
+ e) The binary representation of first and last bit also
+ differ by exactly one bit
+
+ >>> gray_code(2)
+ [0, 1, 3, 2]
+
+ >>> gray_code(1)
+ [0, 1]
+
+ >>> gray_code(3)
+ [0, 1, 3, 2, 6, 7, 5, 4]
+
+ >>> gray_code(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The given input must be positive
+
+ >>> gray_code(10.6)
+ Traceback (most recent call last):
+ ...
+ TypeError: unsupported operand type(s) for <<: 'int' and 'float'
+ """
+
+ # bit count represents no. of bits in the gray code
+ if bit_count < 0:
+ raise ValueError("The given input must be positive")
+
+ # get the generated string sequence
+ sequence = gray_code_sequence_string(bit_count)
+ #
+ # convert them to integers
+ for i in range(len(sequence)):
+ sequence[i] = int(sequence[i], 2)
+
+ return sequence
+
+
+def gray_code_sequence_string(bit_count: int) -> list:
+ """
+ Will output the n-bit grey sequence as a
+ string of bits
+
+ >>> gray_code_sequence_string(2)
+ ['00', '01', '11', '10']
+
+ >>> gray_code_sequence_string(1)
+ ['0', '1']
+ """
+
+ # The approach is a recursive one
+ # Base case achieved when either n = 0 or n=1
+ if bit_count == 0:
+ return ["0"]
+
+ if bit_count == 1:
+ return ["0", "1"]
+
+ seq_len = 1 << bit_count # defines the length of the sequence
+ # 1<< n is equivalent to 2^n
+
+ # recursive answer will generate answer for n-1 bits
+ smaller_sequence = gray_code_sequence_string(bit_count - 1)
+
+ sequence = []
+
+ # append 0 to first half of the smaller sequence generated
+ for i in range(seq_len // 2):
+ generated_no = "0" + smaller_sequence[i]
+ sequence.append(generated_no)
+
+ # append 1 to second half ... start from the end of the list
+ for i in reversed(range(seq_len // 2)):
+ generated_no = "1" + smaller_sequence[i]
+ sequence.append(generated_no)
+
+ return sequence
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/highest_set_bit.py b/bit_manipulation/highest_set_bit.py
new file mode 100644
index 000000000000..21d92dcb9492
--- /dev/null
+++ b/bit_manipulation/highest_set_bit.py
@@ -0,0 +1,34 @@
+def get_highest_set_bit_position(number: int) -> int:
+ """
+ Returns position of the highest set bit of a number.
+ Ref - https://graphics.stanford.edu/~seander/bithacks.html#IntegerLogObvious
+ >>> get_highest_set_bit_position(25)
+ 5
+ >>> get_highest_set_bit_position(37)
+ 6
+ >>> get_highest_set_bit_position(1)
+ 1
+ >>> get_highest_set_bit_position(4)
+ 3
+ >>> get_highest_set_bit_position(0)
+ 0
+ >>> get_highest_set_bit_position(0.8)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be an 'int' type
+ """
+ if not isinstance(number, int):
+ raise TypeError("Input value must be an 'int' type")
+
+ position = 0
+ while number:
+ position += 1
+ number >>= 1
+
+ return position
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/index_of_rightmost_set_bit.py b/bit_manipulation/index_of_rightmost_set_bit.py
new file mode 100644
index 000000000000..c9c911660b08
--- /dev/null
+++ b/bit_manipulation/index_of_rightmost_set_bit.py
@@ -0,0 +1,51 @@
+# Reference: https://www.geeksforgeeks.org/position-of-rightmost-set-bit/
+
+
+def get_index_of_rightmost_set_bit(number: int) -> int:
+ """
+ Take in a positive integer 'number'.
+ Returns the zero-based index of first set bit in that 'number' from right.
+ Returns -1, If no set bit found.
+
+ >>> get_index_of_rightmost_set_bit(0)
+ -1
+ >>> get_index_of_rightmost_set_bit(5)
+ 0
+ >>> get_index_of_rightmost_set_bit(36)
+ 2
+ >>> get_index_of_rightmost_set_bit(8)
+ 3
+ >>> get_index_of_rightmost_set_bit(-18)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ >>> get_index_of_rightmost_set_bit('test')
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ >>> get_index_of_rightmost_set_bit(1.25)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a non-negative integer
+ """
+
+ if not isinstance(number, int) or number < 0:
+ raise ValueError("Input must be a non-negative integer")
+
+ intermediate = number & ~(number - 1)
+ index = 0
+ while intermediate:
+ intermediate >>= 1
+ index += 1
+ return index - 1
+
+
+if __name__ == "__main__":
+ """
+ Finding the index of rightmost set bit has some very peculiar use-cases,
+ especially in finding missing or/and repeating numbers in a list of
+ positive integers.
+ """
+ import doctest
+
+ doctest.testmod(verbose=True)
diff --git a/bit_manipulation/is_even.py b/bit_manipulation/is_even.py
new file mode 100644
index 000000000000..6f95a1160797
--- /dev/null
+++ b/bit_manipulation/is_even.py
@@ -0,0 +1,37 @@
+def is_even(number: int) -> bool:
+ """
+ return true if the input integer is even
+ Explanation: Lets take a look at the following decimal to binary conversions
+ 2 => 10
+ 14 => 1110
+ 100 => 1100100
+ 3 => 11
+ 13 => 1101
+ 101 => 1100101
+ from the above examples we can observe that
+ for all the odd integers there is always 1 set bit at the end
+ also, 1 in binary can be represented as 001, 00001, or 0000001
+ so for any odd integer n => n&1 is always equals 1 else the integer is even
+
+ >>> is_even(1)
+ False
+ >>> is_even(4)
+ True
+ >>> is_even(9)
+ False
+ >>> is_even(15)
+ False
+ >>> is_even(40)
+ True
+ >>> is_even(100)
+ True
+ >>> is_even(101)
+ False
+ """
+ return number & 1 == 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/is_power_of_two.py b/bit_manipulation/is_power_of_two.py
new file mode 100644
index 000000000000..023e979fe51c
--- /dev/null
+++ b/bit_manipulation/is_power_of_two.py
@@ -0,0 +1,57 @@
+"""
+Author : Alexander Pantyukhin
+Date : November 1, 2022
+
+Task:
+Given a positive int number. Return True if this number is power of 2
+or False otherwise.
+
+Implementation notes: Use bit manipulation.
+For example if the number is the power of two it's bits representation:
+n = 0..100..00
+n - 1 = 0..011..11
+
+n & (n - 1) - no intersections = 0
+"""
+
+
+def is_power_of_two(number: int) -> bool:
+ """
+ Return True if this number is power of 2 or False otherwise.
+
+ >>> is_power_of_two(0)
+ True
+ >>> is_power_of_two(1)
+ True
+ >>> is_power_of_two(2)
+ True
+ >>> is_power_of_two(4)
+ True
+ >>> is_power_of_two(6)
+ False
+ >>> is_power_of_two(8)
+ True
+ >>> is_power_of_two(17)
+ False
+ >>> is_power_of_two(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: number must not be negative
+ >>> is_power_of_two(1.2)
+ Traceback (most recent call last):
+ ...
+ TypeError: unsupported operand type(s) for &: 'float' and 'float'
+
+ # Test all powers of 2 from 0 to 10,000
+ >>> all(is_power_of_two(int(2 ** i)) for i in range(10000))
+ True
+ """
+ if number < 0:
+ raise ValueError("number must not be negative")
+ return number & (number - 1) == 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/largest_pow_of_two_le_num.py b/bit_manipulation/largest_pow_of_two_le_num.py
new file mode 100644
index 000000000000..6ef827312199
--- /dev/null
+++ b/bit_manipulation/largest_pow_of_two_le_num.py
@@ -0,0 +1,60 @@
+"""
+Author : Naman Sharma
+Date : October 2, 2023
+
+Task:
+To Find the largest power of 2 less than or equal to a given number.
+
+Implementation notes: Use bit manipulation.
+We start from 1 & left shift the set bit to check if (res<<1)<=number.
+Each left bit shift represents a pow of 2.
+
+For example:
+number: 15
+res: 1 0b1
+ 2 0b10
+ 4 0b100
+ 8 0b1000
+ 16 0b10000 (Exit)
+"""
+
+
+def largest_pow_of_two_le_num(number: int) -> int:
+ """
+ Return the largest power of two less than or equal to a number.
+
+ >>> largest_pow_of_two_le_num(0)
+ 0
+ >>> largest_pow_of_two_le_num(1)
+ 1
+ >>> largest_pow_of_two_le_num(-1)
+ 0
+ >>> largest_pow_of_two_le_num(3)
+ 2
+ >>> largest_pow_of_two_le_num(15)
+ 8
+ >>> largest_pow_of_two_le_num(99)
+ 64
+ >>> largest_pow_of_two_le_num(178)
+ 128
+ >>> largest_pow_of_two_le_num(999999)
+ 524288
+ >>> largest_pow_of_two_le_num(99.9)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be a 'int' type
+ """
+ if isinstance(number, float):
+ raise TypeError("Input value must be a 'int' type")
+ if number <= 0:
+ return 0
+ res = 1
+ while (res << 1) <= number:
+ res <<= 1
+ return res
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/missing_number.py b/bit_manipulation/missing_number.py
new file mode 100644
index 000000000000..554887b17562
--- /dev/null
+++ b/bit_manipulation/missing_number.py
@@ -0,0 +1,40 @@
+def find_missing_number(nums: list[int]) -> int:
+ """
+ Finds the missing number in a list of consecutive integers.
+
+ Args:
+ nums: A list of integers.
+
+ Returns:
+ The missing number.
+
+ Example:
+ >>> find_missing_number([0, 1, 3, 4])
+ 2
+ >>> find_missing_number([4, 3, 1, 0])
+ 2
+ >>> find_missing_number([-4, -3, -1, 0])
+ -2
+ >>> find_missing_number([-2, 2, 1, 3, 0])
+ -1
+ >>> find_missing_number([1, 3, 4, 5, 6])
+ 2
+ >>> find_missing_number([6, 5, 4, 2, 1])
+ 3
+ >>> find_missing_number([6, 1, 5, 3, 4])
+ 2
+ """
+ low = min(nums)
+ high = max(nums)
+ missing_number = high
+
+ for i in range(low, high):
+ missing_number ^= i ^ nums[i - low]
+
+ return missing_number
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/numbers_different_signs.py b/bit_manipulation/numbers_different_signs.py
new file mode 100644
index 000000000000..cf8b6d86f1eb
--- /dev/null
+++ b/bit_manipulation/numbers_different_signs.py
@@ -0,0 +1,39 @@
+"""
+Author : Alexander Pantyukhin
+Date : November 30, 2022
+
+Task:
+Given two int numbers. Return True these numbers have opposite signs
+or False otherwise.
+
+Implementation notes: Use bit manipulation.
+Use XOR for two numbers.
+"""
+
+
+def different_signs(num1: int, num2: int) -> bool:
+ """
+ Return True if numbers have opposite signs False otherwise.
+
+ >>> different_signs(1, -1)
+ True
+ >>> different_signs(1, 1)
+ False
+ >>> different_signs(1000000000000000000000000000, -1000000000000000000000000000)
+ True
+ >>> different_signs(-1000000000000000000000000000, 1000000000000000000000000000)
+ True
+ >>> different_signs(50, 278)
+ False
+ >>> different_signs(0, 2)
+ False
+ >>> different_signs(2, 0)
+ False
+ """
+ return num1 ^ num2 < 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/power_of_4.py b/bit_manipulation/power_of_4.py
new file mode 100644
index 000000000000..09e6e28621df
--- /dev/null
+++ b/bit_manipulation/power_of_4.py
@@ -0,0 +1,67 @@
+"""
+
+Task:
+Given a positive int number. Return True if this number is power of 4
+or False otherwise.
+
+Implementation notes: Use bit manipulation.
+For example if the number is the power of 2 it's bits representation:
+n = 0..100..00
+n - 1 = 0..011..11
+
+n & (n - 1) - no intersections = 0
+If the number is a power of 4 then it should be a power of 2
+and the set bit should be at an odd position.
+"""
+
+
+def power_of_4(number: int) -> bool:
+ """
+ Return True if this number is power of 4 or False otherwise.
+
+ >>> power_of_4(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be positive
+ >>> power_of_4(1)
+ True
+ >>> power_of_4(2)
+ False
+ >>> power_of_4(4)
+ True
+ >>> power_of_4(6)
+ False
+ >>> power_of_4(8)
+ False
+ >>> power_of_4(17)
+ False
+ >>> power_of_4(64)
+ True
+ >>> power_of_4(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be positive
+ >>> power_of_4(1.2)
+ Traceback (most recent call last):
+ ...
+ TypeError: number must be an integer
+
+ """
+ if not isinstance(number, int):
+ raise TypeError("number must be an integer")
+ if number <= 0:
+ raise ValueError("number must be positive")
+ if number & (number - 1) == 0:
+ c = 0
+ while number:
+ c += 1
+ number >>= 1
+ return c % 2 == 1
+ else:
+ return False
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/reverse_bits.py b/bit_manipulation/reverse_bits.py
new file mode 100644
index 000000000000..74b4f2563234
--- /dev/null
+++ b/bit_manipulation/reverse_bits.py
@@ -0,0 +1,86 @@
+def get_reverse_bit_string(number: int) -> str:
+ """
+ return the bit string of an integer
+
+ >>> get_reverse_bit_string(9)
+ '10010000000000000000000000000000'
+ >>> get_reverse_bit_string(43)
+ '11010100000000000000000000000000'
+ >>> get_reverse_bit_string(2873)
+ '10011100110100000000000000000000'
+ >>> get_reverse_bit_string("this is not a number")
+ Traceback (most recent call last):
+ ...
+ TypeError: operation can not be conducted on a object of type str
+ """
+ if not isinstance(number, int):
+ msg = (
+ "operation can not be conducted on a object of type "
+ f"{type(number).__name__}"
+ )
+ raise TypeError(msg)
+ bit_string = ""
+ for _ in range(32):
+ bit_string += str(number % 2)
+ number = number >> 1
+ return bit_string
+
+
+def reverse_bit(number: int) -> str:
+ """
+ Take in an 32 bit integer, reverse its bits,
+ return a string of reverse bits
+
+ result of a reverse_bit and operation on the integer provided.
+
+ >>> reverse_bit(25)
+ '00000000000000000000000000011001'
+ >>> reverse_bit(37)
+ '00000000000000000000000000100101'
+ >>> reverse_bit(21)
+ '00000000000000000000000000010101'
+ >>> reverse_bit(58)
+ '00000000000000000000000000111010'
+ >>> reverse_bit(0)
+ '00000000000000000000000000000000'
+ >>> reverse_bit(256)
+ '00000000000000000000000100000000'
+ >>> reverse_bit(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: the value of input must be positive
+
+ >>> reverse_bit(1.1)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be a 'int' type
+
+ >>> reverse_bit("0")
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'str' and 'int'
+ """
+ if number < 0:
+ raise ValueError("the value of input must be positive")
+ elif isinstance(number, float):
+ raise TypeError("Input value must be a 'int' type")
+ elif isinstance(number, str):
+ raise TypeError("'<' not supported between instances of 'str' and 'int'")
+ result = 0
+ # iterator over [1 to 32],since we are dealing with 32 bit integer
+ for _ in range(1, 33):
+ # left shift the bits by unity
+ result = result << 1
+ # get the end bit
+ end_bit = number % 2
+ # right shift the bits by unity
+ number = number >> 1
+ # add that bit to our ans
+ result = result | end_bit
+ return get_reverse_bit_string(result)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/bit_manipulation/single_bit_manipulation_operations.py b/bit_manipulation/single_bit_manipulation_operations.py
index 114eafe3235b..fcbf033ccb24 100644
--- a/bit_manipulation/single_bit_manipulation_operations.py
+++ b/bit_manipulation/single_bit_manipulation_operations.py
@@ -3,13 +3,13 @@
"""Provide the functionality to manipulate a single bit."""
-def set_bit(number: int, position: int):
+def set_bit(number: int, position: int) -> int:
"""
Set the bit at position to 1.
Details: perform bitwise or for given number and X.
- Where X is a number with all the bits – zeroes and bit on given
- position – one.
+ Where X is a number with all the bits - zeroes and bit on given
+ position - one.
>>> set_bit(0b1101, 1) # 0b1111
15
@@ -21,13 +21,13 @@ def set_bit(number: int, position: int):
return number | (1 << position)
-def clear_bit(number: int, position: int):
+def clear_bit(number: int, position: int) -> int:
"""
Set the bit at position to 0.
Details: perform bitwise and for given number and X.
- Where X is a number with all the bits – ones and bit on given
- position – zero.
+ Where X is a number with all the bits - ones and bit on given
+ position - zero.
>>> clear_bit(0b10010, 1) # 0b10000
16
@@ -37,13 +37,13 @@ def clear_bit(number: int, position: int):
return number & ~(1 << position)
-def flip_bit(number: int, position: int):
+def flip_bit(number: int, position: int) -> int:
"""
Flip the bit at position.
Details: perform bitwise xor for given number and X.
- Where X is a number with all the bits – zeroes and bit on given
- position – one.
+ Where X is a number with all the bits - zeroes and bit on given
+ position - one.
>>> flip_bit(0b101, 1) # 0b111
7
@@ -74,6 +74,26 @@ def is_bit_set(number: int, position: int) -> bool:
return ((number >> position) & 1) == 1
+def get_bit(number: int, position: int) -> int:
+ """
+ Get the bit at the given position
+
+ Details: perform bitwise and for the given number and X,
+ Where X is a number with all the bits - zeroes and bit on given position - one.
+ If the result is not equal to 0, then the bit on the given position is 1, else 0.
+
+ >>> get_bit(0b1010, 0)
+ 0
+ >>> get_bit(0b1010, 1)
+ 1
+ >>> get_bit(0b1010, 2)
+ 0
+ >>> get_bit(0b1010, 3)
+ 1
+ """
+ return int((number & (1 << position)) != 0)
+
+
if __name__ == "__main__":
import doctest
diff --git a/bit_manipulation/swap_all_odd_and_even_bits.py b/bit_manipulation/swap_all_odd_and_even_bits.py
new file mode 100644
index 000000000000..5ec84417bea6
--- /dev/null
+++ b/bit_manipulation/swap_all_odd_and_even_bits.py
@@ -0,0 +1,58 @@
+def show_bits(before: int, after: int) -> str:
+ """
+ >>> print(show_bits(0, 0xFFFF))
+ 0: 00000000
+ 65535: 1111111111111111
+ """
+ return f"{before:>5}: {before:08b}\n{after:>5}: {after:08b}"
+
+
+def swap_odd_even_bits(num: int) -> int:
+ """
+ 1. We use bitwise AND operations to separate the even bits (0, 2, 4, 6, etc.) and
+ odd bits (1, 3, 5, 7, etc.) in the input number.
+ 2. We then right-shift the even bits by 1 position and left-shift the odd bits by
+ 1 position to swap them.
+ 3. Finally, we combine the swapped even and odd bits using a bitwise OR operation
+ to obtain the final result.
+ >>> print(show_bits(0, swap_odd_even_bits(0)))
+ 0: 00000000
+ 0: 00000000
+ >>> print(show_bits(1, swap_odd_even_bits(1)))
+ 1: 00000001
+ 2: 00000010
+ >>> print(show_bits(2, swap_odd_even_bits(2)))
+ 2: 00000010
+ 1: 00000001
+ >>> print(show_bits(3, swap_odd_even_bits(3)))
+ 3: 00000011
+ 3: 00000011
+ >>> print(show_bits(4, swap_odd_even_bits(4)))
+ 4: 00000100
+ 8: 00001000
+ >>> print(show_bits(5, swap_odd_even_bits(5)))
+ 5: 00000101
+ 10: 00001010
+ >>> print(show_bits(6, swap_odd_even_bits(6)))
+ 6: 00000110
+ 9: 00001001
+ >>> print(show_bits(23, swap_odd_even_bits(23)))
+ 23: 00010111
+ 43: 00101011
+ """
+ # Get all even bits - 0xAAAAAAAA is a 32-bit number with all even bits set to 1
+ even_bits = num & 0xAAAAAAAA
+
+ # Get all odd bits - 0x55555555 is a 32-bit number with all odd bits set to 1
+ odd_bits = num & 0x55555555
+
+ # Right shift even bits and left shift odd bits and swap them
+ return even_bits >> 1 | odd_bits << 1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ for i in (-1, 0, 1, 2, 3, 4, 23, 24):
+ print(show_bits(i, swap_odd_even_bits(i)), "\n")
diff --git a/blockchain/README.md b/blockchain/README.md
new file mode 100644
index 000000000000..b5fab7b36eaa
--- /dev/null
+++ b/blockchain/README.md
@@ -0,0 +1,45 @@
+# Blockchain
+
+A Blockchain is a type of **distributed ledger** technology (DLT) that consists of growing list of records, called **blocks**, that are securely linked together using **cryptography**.
+
+Let's breakdown the terminologies in the above definition. We find below terminologies,
+
+- Digital Ledger Technology (DLT)
+- Blocks
+- Cryptography
+
+## Digital Ledger Technology
+
+ It is otherwise called as distributed ledger technology. It is simply the opposite of centralized database. Firstly, what is a **ledger**? A ledger is a book or collection of accounts that records account transactions.
+
+ *Why is Blockchain addressed as digital ledger if it can record more than account transactions? What other transaction details and information can it hold?*
+
+Digital Ledger Technology is just a ledger which is shared among multiple nodes. This way there exist no need for central authority to hold the info. Okay, how is it differentiated from central database and what are their benefits?
+
+There is an organization which has 4 branches whose data are stored in a centralized database. So even if one branch needs any data from ledger they need an approval from database in charge. And if one hacks the central database he gets to tamper and control all the data.
+
+Now lets assume every branch has a copy of the ledger and then once anything is added to the ledger by anyone branch it is gonna automatically reflect in all other ledgers available in other branch. This is done using Peer-to-peer network.
+
+So this means even if information is tampered in one branch we can find out. If one branch is hacked we can be alerted ,so we can safeguard other branches. Now, assume these branches as computers or nodes and the ledger is a transaction record or digital receipt. If one ledger is hacked in a node we can detect since there will be a mismatch in comparison with other node information. So this is the concept of Digital Ledger Technology.
+
+*Is it required for all nodes to have access to all information in other nodes? Wouldn't this require enormous storage space in each node?*
+
+## Blocks
+
+In short a block is nothing but collections of records with a labelled header. These are connected cryptographically. Once a new block is added to a chain, the previous block is connected, more precisely said as locked and hence, will remain unaltered. We can understand this concept once we get a clear understanding of working mechanism of blockchain.
+
+## Cryptography
+
+It is the practice and study of secure communication techniques in the midst of adversarial behavior. More broadly, cryptography is the creation and analysis of protocols that prevent third parties or the general public from accessing private messages.
+
+*Which cryptography technology is most widely used in blockchain and why?*
+
+So, in general, blockchain technology is a distributed record holder which records the information about ownership of an asset. To define precisely,
+> Blockchain is a distributed, immutable ledger that makes it easier to record transactions and track assets in a corporate network.
+An asset could be tangible (such as a house, car, cash, or land) or intangible (such as a business) (intellectual property, patents, copyrights, branding). A blockchain network can track and sell almost anything of value, lowering risk and costs for everyone involved.
+
+So this is all about introduction to blockchain technology. To learn more about the topic refer below links....
+*
+*
+*
+*
diff --git a/blockchain/diophantine_equation.py b/blockchain/diophantine_equation.py
index 751b0efb7227..ae6a145d2922 100644
--- a/blockchain/diophantine_equation.py
+++ b/blockchain/diophantine_equation.py
@@ -1,12 +1,16 @@
-# Diophantine Equation : Given integers a,b,c ( at least one of a and b != 0), the
-# diophantine equation a*x + b*y = c has a solution (where x and y are integers)
-# iff gcd(a,b) divides c.
+from __future__ import annotations
-# GCD ( Greatest Common Divisor ) or HCF ( Highest Common Factor )
+from maths.greatest_common_divisor import greatest_common_divisor
-def diophantine(a, b, c):
+def diophantine(a: int, b: int, c: int) -> tuple[float, float]:
"""
+ Diophantine Equation : Given integers a,b,c ( at least one of a and b != 0), the
+ diophantine equation a*x + b*y = c has a solution (where x and y are integers)
+ iff greatest_common_divisor(a,b) divides c.
+
+ GCD ( Greatest Common Divisor ) or HCF ( Highest Common Factor )
+
>>> diophantine(10,6,14)
(-7.0, 14.0)
@@ -20,25 +24,25 @@ def diophantine(a, b, c):
assert (
c % greatest_common_divisor(a, b) == 0
- ) # greatest_common_divisor(a,b) function implemented below
+ ) # greatest_common_divisor(a,b) is in maths directory
(d, x, y) = extended_gcd(a, b) # extended_gcd(a,b) function implemented below
r = c / d
return (r * x, r * y)
-# Lemma : if n|ab and gcd(a,n) = 1, then n|b.
-
-# Finding All solutions of Diophantine Equations:
+def diophantine_all_soln(a: int, b: int, c: int, n: int = 2) -> None:
+ """
+ Lemma : if n|ab and gcd(a,n) = 1, then n|b.
-# Theorem : Let gcd(a,b) = d, a = d*p, b = d*q. If (x0,y0) is a solution of Diophantine
-# Equation a*x + b*y = c. a*x0 + b*y0 = c, then all the solutions have the form
-# a(x0 + t*q) + b(y0 - t*p) = c, where t is an arbitrary integer.
+ Finding All solutions of Diophantine Equations:
-# n is the number of solution you want, n = 2 by default
+ Theorem : Let gcd(a,b) = d, a = d*p, b = d*q. If (x0,y0) is a solution of
+ Diophantine Equation a*x + b*y = c. a*x0 + b*y0 = c, then all the
+ solutions have the form a(x0 + t*q) + b(y0 - t*p) = c,
+ where t is an arbitrary integer.
+ n is the number of solution you want, n = 2 by default
-def diophantine_all_soln(a, b, c, n=2):
- """
>>> diophantine_all_soln(10, 6, 14)
-7.0 14.0
-4.0 9.0
@@ -67,39 +71,11 @@ def diophantine_all_soln(a, b, c, n=2):
print(x, y)
-# Euclid's Lemma : d divides a and b, if and only if d divides a-b and b
-
-# Euclid's Algorithm
-
-
-def greatest_common_divisor(a, b):
- """
- >>> greatest_common_divisor(7,5)
- 1
-
- Note : In number theory, two integers a and b are said to be relatively prime,
- mutually prime, or co-prime if the only positive integer (factor) that
- divides both of them is 1 i.e., gcd(a,b) = 1.
-
- >>> greatest_common_divisor(121, 11)
- 11
-
+def extended_gcd(a: int, b: int) -> tuple[int, int, int]:
"""
- if a < b:
- a, b = b, a
-
- while a % b != 0:
- a, b = b, a % b
-
- return b
+ Extended Euclid's Algorithm : If d divides a and b and d = a*x + b*y for integers
+ x and y, then d = gcd(a,b)
-
-# Extended Euclid's Algorithm : If d divides a and b and d = a*x + b*y for integers
-# x and y, then d = gcd(a,b)
-
-
-def extended_gcd(a, b):
- """
>>> extended_gcd(10, 6)
(2, -1, 2)
@@ -107,7 +83,8 @@ def extended_gcd(a, b):
(1, -2, 3)
"""
- assert a >= 0 and b >= 0
+ assert a >= 0
+ assert b >= 0
if b == 0:
d, x, y = a, 1, 0
@@ -116,7 +93,8 @@ def extended_gcd(a, b):
x = q
y = p - q * (a // b)
- assert a % d == 0 and b % d == 0
+ assert a % d == 0
+ assert b % d == 0
assert d == a * x + b * y
return (d, x, y)
diff --git a/boolean_algebra/README.md b/boolean_algebra/README.md
new file mode 100644
index 000000000000..45969c855f9c
--- /dev/null
+++ b/boolean_algebra/README.md
@@ -0,0 +1,7 @@
+# Boolean Algebra
+
+Boolean algebra is used to do arithmetic with bits of values True (1) or False (0).
+There are three basic operations: 'and', 'or' and 'not'.
+
+*
+*
diff --git a/boolean_algebra/and_gate.py b/boolean_algebra/and_gate.py
new file mode 100644
index 000000000000..6ae66b5b0a77
--- /dev/null
+++ b/boolean_algebra/and_gate.py
@@ -0,0 +1,38 @@
+"""
+An AND Gate is a logic gate in boolean algebra which results to 1 (True) if both the
+inputs are 1, and 0 (False) otherwise.
+
+Following is the truth table of an AND Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 0 |
+ | 0 | 1 | 0 |
+ | 1 | 0 | 0 |
+ | 1 | 1 | 1 |
+ ------------------------------
+
+Refer - https://www.geeksforgeeks.org/logic-gates-in-python/
+"""
+
+
+def and_gate(input_1: int, input_2: int) -> int:
+ """
+ Calculate AND of the input values
+
+ >>> and_gate(0, 0)
+ 0
+ >>> and_gate(0, 1)
+ 0
+ >>> and_gate(1, 0)
+ 0
+ >>> and_gate(1, 1)
+ 1
+ """
+ return int(input_1 and input_2)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/imply_gate.py b/boolean_algebra/imply_gate.py
new file mode 100644
index 000000000000..b64ebaceb306
--- /dev/null
+++ b/boolean_algebra/imply_gate.py
@@ -0,0 +1,39 @@
+"""
+An IMPLY Gate is a logic gate in boolean algebra which results to 1 if
+either input 1 is 0, or if input 1 is 1, then the output is 1 only if input 2 is 1.
+It is true if input 1 implies input 2.
+
+Following is the truth table of an IMPLY Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 1 |
+ | 0 | 1 | 1 |
+ | 1 | 0 | 0 |
+ | 1 | 1 | 1 |
+ ------------------------------
+
+Refer - https://en.wikipedia.org/wiki/IMPLY_gate
+"""
+
+
+def imply_gate(input_1: int, input_2: int) -> int:
+ """
+ Calculate IMPLY of the input values
+
+ >>> imply_gate(0, 0)
+ 1
+ >>> imply_gate(0, 1)
+ 1
+ >>> imply_gate(1, 0)
+ 0
+ >>> imply_gate(1, 1)
+ 1
+ """
+ return int(input_1 == 0 or input_2 == 1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/karnaugh_map_simplification.py b/boolean_algebra/karnaugh_map_simplification.py
new file mode 100644
index 000000000000..c7f2d4c6b897
--- /dev/null
+++ b/boolean_algebra/karnaugh_map_simplification.py
@@ -0,0 +1,55 @@
+"""
+https://en.wikipedia.org/wiki/Karnaugh_map
+https://www.allaboutcircuits.com/technical-articles/karnaugh-map-boolean-algebraic-simplification-technique
+"""
+
+
+def simplify_kmap(kmap: list[list[int]]) -> str:
+ """
+ Simplify the Karnaugh map.
+ >>> simplify_kmap(kmap=[[0, 1], [1, 1]])
+ "A'B + AB' + AB"
+ >>> simplify_kmap(kmap=[[0, 0], [0, 0]])
+ ''
+ >>> simplify_kmap(kmap=[[0, 1], [1, -1]])
+ "A'B + AB' + AB"
+ >>> simplify_kmap(kmap=[[0, 1], [1, 2]])
+ "A'B + AB' + AB"
+ >>> simplify_kmap(kmap=[[0, 1], [1, 1.1]])
+ "A'B + AB' + AB"
+ >>> simplify_kmap(kmap=[[0, 1], [1, 'a']])
+ "A'B + AB' + AB"
+ """
+ simplified_f = []
+ for a, row in enumerate(kmap):
+ for b, item in enumerate(row):
+ if item:
+ term = ("A" if a else "A'") + ("B" if b else "B'")
+ simplified_f.append(term)
+ return " + ".join(simplified_f)
+
+
+def main() -> None:
+ """
+ Main function to create and simplify a K-Map.
+
+ >>> main()
+ [0, 1]
+ [1, 1]
+ Simplified Expression:
+ A'B + AB' + AB
+ """
+ kmap = [[0, 1], [1, 1]]
+
+ # Manually generate the product of [0, 1] and [0, 1]
+
+ for row in kmap:
+ print(row)
+
+ print("Simplified Expression:")
+ print(simplify_kmap(kmap))
+
+
+if __name__ == "__main__":
+ main()
+ print(f"{simplify_kmap(kmap=[[0, 1], [1, 1]]) = }")
diff --git a/boolean_algebra/multiplexer.py b/boolean_algebra/multiplexer.py
new file mode 100644
index 000000000000..7e65c785c829
--- /dev/null
+++ b/boolean_algebra/multiplexer.py
@@ -0,0 +1,42 @@
+def mux(input0: int, input1: int, select: int) -> int:
+ """
+ Implement a 2-to-1 Multiplexer.
+
+ :param input0: The first input value (0 or 1).
+ :param input1: The second input value (0 or 1).
+ :param select: The select signal (0 or 1) to choose between input0 and input1.
+ :return: The output based on the select signal. input1 if select else input0.
+
+ https://www.electrically4u.com/solved-problems-on-multiplexer
+ https://en.wikipedia.org/wiki/Multiplexer
+
+ >>> mux(0, 1, 0)
+ 0
+ >>> mux(0, 1, 1)
+ 1
+ >>> mux(1, 0, 0)
+ 1
+ >>> mux(1, 0, 1)
+ 0
+ >>> mux(2, 1, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inputs and select signal must be 0 or 1
+ >>> mux(0, -1, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inputs and select signal must be 0 or 1
+ >>> mux(0, 1, 1.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inputs and select signal must be 0 or 1
+ """
+ if all(i in (0, 1) for i in (input0, input1, select)):
+ return input1 if select else input0
+ raise ValueError("Inputs and select signal must be 0 or 1")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/nand_gate.py b/boolean_algebra/nand_gate.py
new file mode 100644
index 000000000000..ea7a6815dcc9
--- /dev/null
+++ b/boolean_algebra/nand_gate.py
@@ -0,0 +1,36 @@
+"""
+A NAND Gate is a logic gate in boolean algebra which results to 0 (False) if both
+the inputs are 1, and 1 (True) otherwise. It's similar to adding
+a NOT gate along with an AND gate.
+Following is the truth table of a NAND Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 1 |
+ | 0 | 1 | 1 |
+ | 1 | 0 | 1 |
+ | 1 | 1 | 0 |
+ ------------------------------
+Refer - https://www.geeksforgeeks.org/logic-gates-in-python/
+"""
+
+
+def nand_gate(input_1: int, input_2: int) -> int:
+ """
+ Calculate NAND of the input values
+ >>> nand_gate(0, 0)
+ 1
+ >>> nand_gate(0, 1)
+ 1
+ >>> nand_gate(1, 0)
+ 1
+ >>> nand_gate(1, 1)
+ 0
+ """
+ return int(not (input_1 and input_2))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/nimply_gate.py b/boolean_algebra/nimply_gate.py
new file mode 100644
index 000000000000..68e82c8db8d9
--- /dev/null
+++ b/boolean_algebra/nimply_gate.py
@@ -0,0 +1,39 @@
+"""
+An NIMPLY Gate is a logic gate in boolean algebra which results to 0 if
+either input 1 is 0, or if input 1 is 1, then it is 0 only if input 2 is 1.
+It is false if input 1 implies input 2. It is the negated form of imply
+
+Following is the truth table of an NIMPLY Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 0 |
+ | 0 | 1 | 0 |
+ | 1 | 0 | 1 |
+ | 1 | 1 | 0 |
+ ------------------------------
+
+Refer - https://en.wikipedia.org/wiki/NIMPLY_gate
+"""
+
+
+def nimply_gate(input_1: int, input_2: int) -> int:
+ """
+ Calculate NIMPLY of the input values
+
+ >>> nimply_gate(0, 0)
+ 0
+ >>> nimply_gate(0, 1)
+ 0
+ >>> nimply_gate(1, 0)
+ 1
+ >>> nimply_gate(1, 1)
+ 0
+ """
+ return int(input_1 == 1 and input_2 == 0)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/nor_gate.py b/boolean_algebra/nor_gate.py
new file mode 100644
index 000000000000..d4d6f0da23ea
--- /dev/null
+++ b/boolean_algebra/nor_gate.py
@@ -0,0 +1,68 @@
+"""
+A NOR Gate is a logic gate in boolean algebra which results in false(0) if any of the
+inputs is 1, and True(1) if all inputs are 0.
+Following is the truth table of a NOR Gate:
+ Truth Table of NOR Gate:
+ | Input 1 | Input 2 | Output |
+ | 0 | 0 | 1 |
+ | 0 | 1 | 0 |
+ | 1 | 0 | 0 |
+ | 1 | 1 | 0 |
+
+ Code provided by Akshaj Vishwanathan
+https://www.geeksforgeeks.org/logic-gates-in-python
+"""
+
+from collections.abc import Callable
+
+
+def nor_gate(input_1: int, input_2: int) -> int:
+ """
+ >>> nor_gate(0, 0)
+ 1
+ >>> nor_gate(0, 1)
+ 0
+ >>> nor_gate(1, 0)
+ 0
+ >>> nor_gate(1, 1)
+ 0
+ >>> nor_gate(0.0, 0.0)
+ 1
+ >>> nor_gate(0, -7)
+ 0
+ """
+ return int(input_1 == input_2 == 0)
+
+
+def truth_table(func: Callable) -> str:
+ """
+ >>> print(truth_table(nor_gate))
+ Truth Table of NOR Gate:
+ | Input 1 | Input 2 | Output |
+ | 0 | 0 | 1 |
+ | 0 | 1 | 0 |
+ | 1 | 0 | 0 |
+ | 1 | 1 | 0 |
+ """
+
+ def make_table_row(items: list | tuple) -> str:
+ """
+ >>> make_table_row(("One", "Two", "Three"))
+ '| One | Two | Three |'
+ """
+ return f"| {' | '.join(f'{item:^8}' for item in items)} |"
+
+ return "\n".join(
+ (
+ "Truth Table of NOR Gate:",
+ make_table_row(("Input 1", "Input 2", "Output")),
+ *[make_table_row((i, j, func(i, j))) for i in (0, 1) for j in (0, 1)],
+ )
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(truth_table(nor_gate))
diff --git a/boolean_algebra/not_gate.py b/boolean_algebra/not_gate.py
new file mode 100644
index 000000000000..cfa74cf42204
--- /dev/null
+++ b/boolean_algebra/not_gate.py
@@ -0,0 +1,30 @@
+"""
+A NOT Gate is a logic gate in boolean algebra which results to 0 (False) if the
+input is high, and 1 (True) if the input is low.
+Following is the truth table of a XOR Gate:
+ ------------------------------
+ | Input | Output |
+ ------------------------------
+ | 0 | 1 |
+ | 1 | 0 |
+ ------------------------------
+Refer - https://www.geeksforgeeks.org/logic-gates-in-python/
+"""
+
+
+def not_gate(input_1: int) -> int:
+ """
+ Calculate NOT of the input values
+ >>> not_gate(0)
+ 1
+ >>> not_gate(1)
+ 0
+ """
+
+ return 1 if input_1 == 0 else 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/or_gate.py b/boolean_algebra/or_gate.py
new file mode 100644
index 000000000000..0fd4e5a5dc18
--- /dev/null
+++ b/boolean_algebra/or_gate.py
@@ -0,0 +1,35 @@
+"""
+An OR Gate is a logic gate in boolean algebra which results to 0 (False) if both the
+inputs are 0, and 1 (True) otherwise.
+Following is the truth table of an AND Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 0 |
+ | 0 | 1 | 1 |
+ | 1 | 0 | 1 |
+ | 1 | 1 | 1 |
+ ------------------------------
+Refer - https://www.geeksforgeeks.org/logic-gates-in-python/
+"""
+
+
+def or_gate(input_1: int, input_2: int) -> int:
+ """
+ Calculate OR of the input values
+ >>> or_gate(0, 0)
+ 0
+ >>> or_gate(0, 1)
+ 1
+ >>> or_gate(1, 0)
+ 1
+ >>> or_gate(1, 1)
+ 1
+ """
+ return int((input_1, input_2).count(1) != 0)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/quine_mc_cluskey.py b/boolean_algebra/quine_mc_cluskey.py
index 19bac336f6c5..8e22e66726d4 100644
--- a/boolean_algebra/quine_mc_cluskey.py
+++ b/boolean_algebra/quine_mc_cluskey.py
@@ -1,40 +1,46 @@
-def compare_string(string1: str, string2: str) -> str:
+from __future__ import annotations
+
+from collections.abc import Sequence
+from typing import Literal
+
+
+def compare_string(string1: str, string2: str) -> str | Literal[False]:
"""
>>> compare_string('0010','0110')
'0_10'
>>> compare_string('0110','1101')
- -1
+ False
"""
- l1 = list(string1)
- l2 = list(string2)
+ list1 = list(string1)
+ list2 = list(string2)
count = 0
- for i in range(len(l1)):
- if l1[i] != l2[i]:
+ for i in range(len(list1)):
+ if list1[i] != list2[i]:
count += 1
- l1[i] = "_"
+ list1[i] = "_"
if count > 1:
- return -1
+ return False
else:
- return "".join(l1)
+ return "".join(list1)
-def check(binary: [str]) -> [str]:
+def check(binary: list[str]) -> list[str]:
"""
>>> check(['0.00.01.5'])
['0.00.01.5']
"""
pi = []
- while 1:
+ while True:
check1 = ["$"] * len(binary)
temp = []
for i in range(len(binary)):
for j in range(i + 1, len(binary)):
k = compare_string(binary[i], binary[j])
- if k != -1:
+ if k is False:
check1[i] = "*"
check1[j] = "*"
- temp.append(k)
+ temp.append("X")
for i in range(len(binary)):
if check1[i] == "$":
pi.append(binary[i])
@@ -43,19 +49,18 @@ def check(binary: [str]) -> [str]:
binary = list(set(temp))
-def decimal_to_binary(no_of_variable: int, minterms: [float]) -> [str]:
+def decimal_to_binary(no_of_variable: int, minterms: Sequence[float]) -> list[str]:
"""
>>> decimal_to_binary(3,[1.5])
['0.00.01.5']
"""
temp = []
- s = ""
- for m in minterms:
- for i in range(no_of_variable):
- s = str(m % 2) + s
- m //= 2
- temp.append(s)
- s = ""
+ for minterm in minterms:
+ string = ""
+ for _ in range(no_of_variable):
+ string = str(minterm % 2) + string
+ minterm //= 2
+ temp.append(string)
return temp
@@ -67,19 +72,13 @@ def is_for_table(string1: str, string2: str, count: int) -> bool:
>>> is_for_table('01_','001',1)
False
"""
- l1 = list(string1)
- l2 = list(string2)
- count_n = 0
- for i in range(len(l1)):
- if l1[i] != l2[i]:
- count_n += 1
- if count_n == count:
- return True
- else:
- return False
+ list1 = list(string1)
+ list2 = list(string2)
+ count_n = sum(item1 != item2 for item1, item2 in zip(list1, list2))
+ return count_n == count
-def selection(chart: [[int]], prime_implicants: [str]) -> [str]:
+def selection(chart: list[list[int]], prime_implicants: list[str]) -> list[str]:
"""
>>> selection([[1]],['0.00.01.5'])
['0.00.01.5']
@@ -90,43 +89,39 @@ def selection(chart: [[int]], prime_implicants: [str]) -> [str]:
temp = []
select = [0] * len(chart)
for i in range(len(chart[0])):
- count = 0
- rem = -1
- for j in range(len(chart)):
- if chart[j][i] == 1:
- count += 1
- rem = j
+ count = sum(row[i] == 1 for row in chart)
if count == 1:
+ rem = max(j for j, row in enumerate(chart) if row[i] == 1)
select[rem] = 1
- for i in range(len(select)):
- if select[i] == 1:
- for j in range(len(chart[0])):
- if chart[i][j] == 1:
- for k in range(len(chart)):
- chart[k][j] = 0
- temp.append(prime_implicants[i])
- while 1:
- max_n = 0
- rem = -1
- count_n = 0
- for i in range(len(chart)):
- count_n = chart[i].count(1)
- if count_n > max_n:
- max_n = count_n
- rem = i
+ for i, item in enumerate(select):
+ if item != 1:
+ continue
+ for j in range(len(chart[0])):
+ if chart[i][j] != 1:
+ continue
+ for row in chart:
+ row[j] = 0
+ temp.append(prime_implicants[i])
+ while True:
+ counts = [chart[i].count(1) for i in range(len(chart))]
+ max_n = max(counts)
+ rem = counts.index(max_n)
if max_n == 0:
return temp
temp.append(prime_implicants[rem])
- for i in range(len(chart[0])):
- if chart[rem][i] == 1:
- for j in range(len(chart)):
- chart[j][i] = 0
+ for j in range(len(chart[0])):
+ if chart[rem][j] != 1:
+ continue
+ for i in range(len(chart)):
+ chart[i][j] = 0
-def prime_implicant_chart(prime_implicants: [str], binary: [str]) -> [[int]]:
+def prime_implicant_chart(
+ prime_implicants: list[str], binary: list[str]
+) -> list[list[int]]:
"""
>>> prime_implicant_chart(['0.00.01.5'],['0.00.01.5'])
[[1]]
@@ -141,10 +136,10 @@ def prime_implicant_chart(prime_implicants: [str], binary: [str]) -> [[int]]:
return chart
-def main():
+def main() -> None:
no_of_variable = int(input("Enter the no. of variables\n"))
minterms = [
- int(x)
+ float(x)
for x in input(
"Enter the decimal representation of Minterms 'Spaces Separated'\n"
).split()
diff --git a/boolean_algebra/xnor_gate.py b/boolean_algebra/xnor_gate.py
new file mode 100644
index 000000000000..05b756da2960
--- /dev/null
+++ b/boolean_algebra/xnor_gate.py
@@ -0,0 +1,37 @@
+"""
+A XNOR Gate is a logic gate in boolean algebra which results to 0 (False) if both the
+inputs are different, and 1 (True), if the inputs are same.
+It's similar to adding a NOT gate to an XOR gate
+
+Following is the truth table of a XNOR Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 1 |
+ | 0 | 1 | 0 |
+ | 1 | 0 | 0 |
+ | 1 | 1 | 1 |
+ ------------------------------
+Refer - https://www.geeksforgeeks.org/logic-gates-in-python/
+"""
+
+
+def xnor_gate(input_1: int, input_2: int) -> int:
+ """
+ Calculate XOR of the input values
+ >>> xnor_gate(0, 0)
+ 1
+ >>> xnor_gate(0, 1)
+ 0
+ >>> xnor_gate(1, 0)
+ 0
+ >>> xnor_gate(1, 1)
+ 1
+ """
+ return 1 if input_1 == input_2 else 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/boolean_algebra/xor_gate.py b/boolean_algebra/xor_gate.py
new file mode 100644
index 000000000000..f3922e426e3d
--- /dev/null
+++ b/boolean_algebra/xor_gate.py
@@ -0,0 +1,37 @@
+"""
+A XOR Gate is a logic gate in boolean algebra which results to 1 (True) if only one of
+the two inputs is 1, and 0 (False) if an even number of inputs are 1.
+Following is the truth table of a XOR Gate:
+ ------------------------------
+ | Input 1 | Input 2 | Output |
+ ------------------------------
+ | 0 | 0 | 0 |
+ | 0 | 1 | 1 |
+ | 1 | 0 | 1 |
+ | 1 | 1 | 0 |
+ ------------------------------
+
+Refer - https://www.geeksforgeeks.org/logic-gates-in-python/
+"""
+
+
+def xor_gate(input_1: int, input_2: int) -> int:
+ """
+ calculate xor of the input values
+
+ >>> xor_gate(0, 0)
+ 0
+ >>> xor_gate(0, 1)
+ 1
+ >>> xor_gate(1, 0)
+ 1
+ >>> xor_gate(1, 1)
+ 0
+ """
+ return (input_1, input_2).count(0) % 2
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/cellular_automata/README.md b/cellular_automata/README.md
index c3fa0516f5dd..c5681b33906c 100644
--- a/cellular_automata/README.md
+++ b/cellular_automata/README.md
@@ -1,4 +1,8 @@
# Cellular Automata
-* https://en.wikipedia.org/wiki/Cellular_automaton
-* https://mathworld.wolfram.com/ElementaryCellularAutomaton.html
+Cellular automata are a way to simulate the behavior of "life", no matter if it is a robot or cell.
+They usually follow simple rules but can lead to the creation of complex forms.
+The most popular cellular automaton is Conway's [Game of Life](https://en.wikipedia.org/wiki/Conway%27s_Game_of_Life).
+
+*
+*
diff --git a/cellular_automata/conways_game_of_life.py b/cellular_automata/conways_game_of_life.py
new file mode 100644
index 000000000000..485f0d47bd8b
--- /dev/null
+++ b/cellular_automata/conways_game_of_life.py
@@ -0,0 +1,96 @@
+"""
+Conway's Game of Life implemented in Python.
+https://en.wikipedia.org/wiki/Conway%27s_Game_of_Life
+"""
+
+from __future__ import annotations
+
+from PIL import Image
+
+# Define glider example
+GLIDER = [
+ [0, 1, 0, 0, 0, 0, 0, 0],
+ [0, 0, 1, 0, 0, 0, 0, 0],
+ [1, 1, 1, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0, 0],
+]
+
+# Define blinker example
+BLINKER = [[0, 1, 0], [0, 1, 0], [0, 1, 0]]
+
+
+def new_generation(cells: list[list[int]]) -> list[list[int]]:
+ """
+ Generates the next generation for a given state of Conway's Game of Life.
+ >>> new_generation(BLINKER)
+ [[0, 0, 0], [1, 1, 1], [0, 0, 0]]
+ """
+ next_generation = []
+ for i in range(len(cells)):
+ next_generation_row = []
+ for j in range(len(cells[i])):
+ # Get the number of live neighbours
+ neighbour_count = 0
+ if i > 0 and j > 0:
+ neighbour_count += cells[i - 1][j - 1]
+ if i > 0:
+ neighbour_count += cells[i - 1][j]
+ if i > 0 and j < len(cells[i]) - 1:
+ neighbour_count += cells[i - 1][j + 1]
+ if j > 0:
+ neighbour_count += cells[i][j - 1]
+ if j < len(cells[i]) - 1:
+ neighbour_count += cells[i][j + 1]
+ if i < len(cells) - 1 and j > 0:
+ neighbour_count += cells[i + 1][j - 1]
+ if i < len(cells) - 1:
+ neighbour_count += cells[i + 1][j]
+ if i < len(cells) - 1 and j < len(cells[i]) - 1:
+ neighbour_count += cells[i + 1][j + 1]
+
+ # Rules of the game of life (excerpt from Wikipedia):
+ # 1. Any live cell with two or three live neighbours survives.
+ # 2. Any dead cell with three live neighbours becomes a live cell.
+ # 3. All other live cells die in the next generation.
+ # Similarly, all other dead cells stay dead.
+ alive = cells[i][j] == 1
+ if (alive and 2 <= neighbour_count <= 3) or (
+ not alive and neighbour_count == 3
+ ):
+ next_generation_row.append(1)
+ else:
+ next_generation_row.append(0)
+
+ next_generation.append(next_generation_row)
+ return next_generation
+
+
+def generate_images(cells: list[list[int]], frames: int) -> list[Image.Image]:
+ """
+ Generates a list of images of subsequent Game of Life states.
+ """
+ images = []
+ for _ in range(frames):
+ # Create output image
+ img = Image.new("RGB", (len(cells[0]), len(cells)))
+ pixels = img.load()
+
+ # Save cells to image
+ for x in range(len(cells)):
+ for y in range(len(cells[0])):
+ colour = 255 - cells[y][x] * 255
+ pixels[x, y] = (colour, colour, colour)
+
+ # Save image
+ images.append(img)
+ cells = new_generation(cells)
+ return images
+
+
+if __name__ == "__main__":
+ images = generate_images(GLIDER, 16)
+ images[0].save("out.gif", save_all=True, append_images=images[1:])
diff --git a/other/game_of_life.py b/cellular_automata/game_of_life.py
similarity index 74%
rename from other/game_of_life.py
rename to cellular_automata/game_of_life.py
index 09863993dc3a..76276b272d65 100644
--- a/other/game_of_life.py
+++ b/cellular_automata/game_of_life.py
@@ -10,7 +10,7 @@
- 3.5
Usage:
- - $python3 game_o_life
+ - $python3 game_of_life
Game-Of-Life Rules:
@@ -26,7 +26,8 @@
4.
Any dead cell with exactly three live neighbours be-
comes a live cell, as if by reproduction.
- """
+"""
+
import random
import sys
@@ -34,25 +35,26 @@
from matplotlib import pyplot as plt
from matplotlib.colors import ListedColormap
-usage_doc = "Usage of script: script_nama "
+usage_doc = "Usage of script: script_name "
choice = [0] * 100 + [1] * 10
random.shuffle(choice)
-def create_canvas(size):
+def create_canvas(size: int) -> list[list[bool]]:
canvas = [[False for i in range(size)] for j in range(size)]
return canvas
-def seed(canvas):
+def seed(canvas: list[list[bool]]) -> None:
for i, row in enumerate(canvas):
for j, _ in enumerate(row):
canvas[i][j] = bool(random.getrandbits(1))
-def run(canvas):
- """This function runs the rules of game through all points, and changes their
+def run(canvas: list[list[bool]]) -> list[list[bool]]:
+ """
+ This function runs the rules of game through all points, and changes their
status accordingly.(in the same canvas)
@Args:
--
@@ -60,23 +62,20 @@ def run(canvas):
@returns:
--
- None
+ canvas of population after one step
"""
- canvas = np.array(canvas)
- next_gen_canvas = np.array(create_canvas(canvas.shape[0]))
- for r, row in enumerate(canvas):
+ current_canvas = np.array(canvas)
+ next_gen_canvas = np.array(create_canvas(current_canvas.shape[0]))
+ for r, row in enumerate(current_canvas):
for c, pt in enumerate(row):
- # print(r-1,r+2,c-1,c+2)
next_gen_canvas[r][c] = __judge_point(
- pt, canvas[r - 1 : r + 2, c - 1 : c + 2]
+ pt, current_canvas[r - 1 : r + 2, c - 1 : c + 2]
)
- canvas = next_gen_canvas
- del next_gen_canvas # cleaning memory as we move on.
- return canvas.tolist()
+ return next_gen_canvas.tolist()
-def __judge_point(pt, neighbours):
+def __judge_point(pt: bool, neighbours: list[list[bool]]) -> bool:
dead = 0
alive = 0
# finding dead or alive neighbours count.
@@ -98,13 +97,12 @@ def __judge_point(pt, neighbours):
if pt:
if alive < 2:
state = False
- elif alive == 2 or alive == 3:
+ elif alive in {2, 3}:
state = True
elif alive > 3:
state = False
- else:
- if alive == 3:
- state = True
+ elif alive == 3:
+ state = True
return state
diff --git a/cellular_automata/langtons_ant.py b/cellular_automata/langtons_ant.py
new file mode 100644
index 000000000000..9847c50a5c3e
--- /dev/null
+++ b/cellular_automata/langtons_ant.py
@@ -0,0 +1,106 @@
+"""
+Langton's ant
+
+@ https://en.wikipedia.org/wiki/Langton%27s_ant
+@ https://upload.wikimedia.org/wikipedia/commons/0/09/LangtonsAntAnimated.gif
+"""
+
+from functools import partial
+
+from matplotlib import pyplot as plt
+from matplotlib.animation import FuncAnimation
+
+WIDTH = 80
+HEIGHT = 80
+
+
+class LangtonsAnt:
+ """
+ Represents the main LangonsAnt algorithm.
+
+ >>> la = LangtonsAnt(2, 2)
+ >>> la.board
+ [[True, True], [True, True]]
+ >>> la.ant_position
+ (1, 1)
+ """
+
+ def __init__(self, width: int, height: int) -> None:
+ # Each square is either True or False where True is white and False is black
+ self.board = [[True] * width for _ in range(height)]
+ self.ant_position: tuple[int, int] = (width // 2, height // 2)
+
+ # Initially pointing left (similar to the wikipedia image)
+ # (0 = 0° | 1 = 90° | 2 = 180 ° | 3 = 270°)
+ self.ant_direction: int = 3
+
+ def move_ant(self, axes: plt.Axes | None, display: bool, _frame: int) -> None:
+ """
+ Performs three tasks:
+ 1. The ant turns either clockwise or anti-clockwise according to the colour
+ of the square that it is currently on. If the square is white, the ant
+ turns clockwise, and if the square is black the ant turns anti-clockwise
+ 2. The ant moves one square in the direction that it is currently facing
+ 3. The square the ant was previously on is inverted (White -> Black and
+ Black -> White)
+
+ If display is True, the board will also be displayed on the axes
+
+ >>> la = LangtonsAnt(2, 2)
+ >>> la.move_ant(None, True, 0)
+ >>> la.board
+ [[True, True], [True, False]]
+ >>> la.move_ant(None, True, 0)
+ >>> la.board
+ [[True, False], [True, False]]
+ """
+ directions = {
+ 0: (-1, 0), # 0°
+ 1: (0, 1), # 90°
+ 2: (1, 0), # 180°
+ 3: (0, -1), # 270°
+ }
+ x, y = self.ant_position
+
+ # Turn clockwise or anti-clockwise according to colour of square
+ if self.board[x][y] is True:
+ # The square is white so turn 90° clockwise
+ self.ant_direction = (self.ant_direction + 1) % 4
+ else:
+ # The square is black so turn 90° anti-clockwise
+ self.ant_direction = (self.ant_direction - 1) % 4
+
+ # Move ant
+ move_x, move_y = directions[self.ant_direction]
+ self.ant_position = (x + move_x, y + move_y)
+
+ # Flip colour of square
+ self.board[x][y] = not self.board[x][y]
+
+ if display and axes:
+ # Display the board on the axes
+ axes.get_xaxis().set_ticks([])
+ axes.get_yaxis().set_ticks([])
+ axes.imshow(self.board, cmap="gray", interpolation="nearest")
+
+ def display(self, frames: int = 100_000) -> None:
+ """
+ Displays the board without delay in a matplotlib plot
+ to visually understand and track the ant.
+
+ >>> _ = LangtonsAnt(WIDTH, HEIGHT)
+ """
+ fig, ax = plt.subplots()
+ # Assign animation to a variable to prevent it from getting garbage collected
+ self.animation = FuncAnimation(
+ fig, partial(self.move_ant, ax, True), frames=frames, interval=1
+ )
+ plt.show()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ LangtonsAnt(WIDTH, HEIGHT).display()
diff --git a/cellular_automata/nagel_schrekenberg.py b/cellular_automata/nagel_schrekenberg.py
new file mode 100644
index 000000000000..bcdca902afee
--- /dev/null
+++ b/cellular_automata/nagel_schrekenberg.py
@@ -0,0 +1,140 @@
+"""
+Simulate the evolution of a highway with only one road that is a loop.
+The highway is divided in cells, each cell can have at most one car in it.
+The highway is a loop so when a car comes to one end, it will come out on the other.
+Each car is represented by its speed (from 0 to 5).
+
+Some information about speed:
+ -1 means that the cell on the highway is empty
+ 0 to 5 are the speed of the cars with 0 being the lowest and 5 the highest
+
+highway: list[int] Where every position and speed of every car will be stored
+probability The probability that a driver will slow down
+initial_speed The speed of the cars a the start
+frequency How many cells there are between two cars at the start
+max_speed The maximum speed a car can go to
+number_of_cells How many cell are there in the highway
+number_of_update How many times will the position be updated
+
+More information here: https://en.wikipedia.org/wiki/Nagel%E2%80%93Schreckenberg_model
+
+Examples for doctest:
+>>> simulate(construct_highway(6, 3, 0), 2, 0, 2)
+[[0, -1, -1, 0, -1, -1], [-1, 1, -1, -1, 1, -1], [-1, -1, 1, -1, -1, 1]]
+>>> simulate(construct_highway(5, 2, -2), 3, 0, 2)
+[[0, -1, 0, -1, 0], [0, -1, 0, -1, -1], [0, -1, -1, 1, -1], [-1, 1, -1, 0, -1]]
+"""
+
+from random import randint, random
+
+
+def construct_highway(
+ number_of_cells: int,
+ frequency: int,
+ initial_speed: int,
+ random_frequency: bool = False,
+ random_speed: bool = False,
+ max_speed: int = 5,
+) -> list:
+ """
+ Build the highway following the parameters given
+ >>> construct_highway(10, 2, 6)
+ [[6, -1, 6, -1, 6, -1, 6, -1, 6, -1]]
+ >>> construct_highway(10, 10, 2)
+ [[2, -1, -1, -1, -1, -1, -1, -1, -1, -1]]
+ """
+
+ highway = [[-1] * number_of_cells] # Create a highway without any car
+ i = 0
+ initial_speed = max(initial_speed, 0)
+ while i < number_of_cells:
+ highway[0][i] = (
+ randint(0, max_speed) if random_speed else initial_speed
+ ) # Place the cars
+ i += (
+ randint(1, max_speed * 2) if random_frequency else frequency
+ ) # Arbitrary number, may need tuning
+ return highway
+
+
+def get_distance(highway_now: list, car_index: int) -> int:
+ """
+ Get the distance between a car (at index car_index) and the next car
+ >>> get_distance([6, -1, 6, -1, 6], 2)
+ 1
+ >>> get_distance([2, -1, -1, -1, 3, 1, 0, 1, 3, 2], 0)
+ 3
+ >>> get_distance([-1, -1, -1, -1, 2, -1, -1, -1, 3], -1)
+ 4
+ """
+
+ distance = 0
+ cells = highway_now[car_index + 1 :]
+ for cell in range(len(cells)): # May need a better name for this
+ if cells[cell] != -1: # If the cell is not empty then
+ return distance # we have the distance we wanted
+ distance += 1
+ # Here if the car is near the end of the highway
+ return distance + get_distance(highway_now, -1)
+
+
+def update(highway_now: list, probability: float, max_speed: int) -> list:
+ """
+ Update the speed of the cars
+ >>> update([-1, -1, -1, -1, -1, 2, -1, -1, -1, -1, 3], 0.0, 5)
+ [-1, -1, -1, -1, -1, 3, -1, -1, -1, -1, 4]
+ >>> update([-1, -1, 2, -1, -1, -1, -1, 3], 0.0, 5)
+ [-1, -1, 3, -1, -1, -1, -1, 1]
+ """
+
+ number_of_cells = len(highway_now)
+ # Beforce calculations, the highway is empty
+ next_highway = [-1] * number_of_cells
+
+ for car_index in range(number_of_cells):
+ if highway_now[car_index] != -1:
+ # Add 1 to the current speed of the car and cap the speed
+ next_highway[car_index] = min(highway_now[car_index] + 1, max_speed)
+ # Number of empty cell before the next car
+ dn = get_distance(highway_now, car_index) - 1
+ # We can't have the car causing an accident
+ next_highway[car_index] = min(next_highway[car_index], dn)
+ if random() < probability:
+ # Randomly, a driver will slow down
+ next_highway[car_index] = max(next_highway[car_index] - 1, 0)
+ return next_highway
+
+
+def simulate(
+ highway: list, number_of_update: int, probability: float, max_speed: int
+) -> list:
+ """
+ The main function, it will simulate the evolution of the highway
+ >>> simulate([[-1, 2, -1, -1, -1, 3]], 2, 0.0, 3)
+ [[-1, 2, -1, -1, -1, 3], [-1, -1, -1, 2, -1, 0], [1, -1, -1, 0, -1, -1]]
+ >>> simulate([[-1, 2, -1, 3]], 4, 0.0, 3)
+ [[-1, 2, -1, 3], [-1, 0, -1, 0], [-1, 0, -1, 0], [-1, 0, -1, 0], [-1, 0, -1, 0]]
+ """
+
+ number_of_cells = len(highway[0])
+
+ for i in range(number_of_update):
+ next_speeds_calculated = update(highway[i], probability, max_speed)
+ real_next_speeds = [-1] * number_of_cells
+
+ for car_index in range(number_of_cells):
+ speed = next_speeds_calculated[car_index]
+ if speed != -1:
+ # Change the position based on the speed (with % to create the loop)
+ index = (car_index + speed) % number_of_cells
+ # Commit the change of position
+ real_next_speeds[index] = speed
+ highway.append(real_next_speeds)
+
+ return highway
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/cellular_automata/wa_tor.py b/cellular_automata/wa_tor.py
new file mode 100644
index 000000000000..29f7ea510bfe
--- /dev/null
+++ b/cellular_automata/wa_tor.py
@@ -0,0 +1,548 @@
+"""
+Wa-Tor algorithm (1984)
+
+| @ https://en.wikipedia.org/wiki/Wa-Tor
+| @ https://beltoforion.de/en/wator/
+| @ https://beltoforion.de/en/wator/images/wator_medium.webm
+
+This solution aims to completely remove any systematic approach
+to the Wa-Tor planet, and utilise fully random methods.
+
+The constants are a working set that allows the Wa-Tor planet
+to result in one of the three possible results.
+"""
+
+from collections.abc import Callable
+from random import randint, shuffle
+from time import sleep
+from typing import Literal
+
+WIDTH = 50 # Width of the Wa-Tor planet
+HEIGHT = 50 # Height of the Wa-Tor planet
+
+PREY_INITIAL_COUNT = 30 # The initial number of prey entities
+PREY_REPRODUCTION_TIME = 5 # The chronons before reproducing
+
+PREDATOR_INITIAL_COUNT = 50 # The initial number of predator entities
+# The initial energy value of predator entities
+PREDATOR_INITIAL_ENERGY_VALUE = 15
+# The energy value provided when consuming prey
+PREDATOR_FOOD_VALUE = 5
+PREDATOR_REPRODUCTION_TIME = 20 # The chronons before reproducing
+
+MAX_ENTITIES = 500 # The max number of organisms on the board
+# The number of entities to delete from the unbalanced side
+DELETE_UNBALANCED_ENTITIES = 50
+
+
+class Entity:
+ """
+ Represents an entity (either prey or predator).
+
+ >>> e = Entity(True, coords=(0, 0))
+ >>> e.prey
+ True
+ >>> e.coords
+ (0, 0)
+ >>> e.alive
+ True
+ """
+
+ def __init__(self, prey: bool, coords: tuple[int, int]) -> None:
+ self.prey = prey
+ # The (row, col) pos of the entity
+ self.coords = coords
+
+ self.remaining_reproduction_time = (
+ PREY_REPRODUCTION_TIME if prey else PREDATOR_REPRODUCTION_TIME
+ )
+ self.energy_value = None if prey is True else PREDATOR_INITIAL_ENERGY_VALUE
+ self.alive = True
+
+ def reset_reproduction_time(self) -> None:
+ """
+ >>> e = Entity(True, coords=(0, 0))
+ >>> e.reset_reproduction_time()
+ >>> e.remaining_reproduction_time == PREY_REPRODUCTION_TIME
+ True
+ >>> e = Entity(False, coords=(0, 0))
+ >>> e.reset_reproduction_time()
+ >>> e.remaining_reproduction_time == PREDATOR_REPRODUCTION_TIME
+ True
+ """
+ self.remaining_reproduction_time = (
+ PREY_REPRODUCTION_TIME if self.prey is True else PREDATOR_REPRODUCTION_TIME
+ )
+
+ def __repr__(self) -> str:
+ """
+ >>> Entity(prey=True, coords=(1, 1))
+ Entity(prey=True, coords=(1, 1), remaining_reproduction_time=5)
+ >>> Entity(prey=False, coords=(2, 1)) # doctest: +NORMALIZE_WHITESPACE
+ Entity(prey=False, coords=(2, 1),
+ remaining_reproduction_time=20, energy_value=15)
+ """
+ repr_ = (
+ f"Entity(prey={self.prey}, coords={self.coords}, "
+ f"remaining_reproduction_time={self.remaining_reproduction_time}"
+ )
+ if self.energy_value is not None:
+ repr_ += f", energy_value={self.energy_value}"
+ return f"{repr_})"
+
+
+class WaTor:
+ """
+ Represents the main Wa-Tor algorithm.
+
+ :attr time_passed: A function that is called every time
+ time passes (a chronon) in order to visually display
+ the new Wa-Tor planet. The `time_passed` function can block
+ using ``time.sleep`` to slow the algorithm progression.
+
+ >>> wt = WaTor(10, 15)
+ >>> wt.width
+ 10
+ >>> wt.height
+ 15
+ >>> len(wt.planet)
+ 15
+ >>> len(wt.planet[0])
+ 10
+ >>> len(wt.get_entities()) == PREDATOR_INITIAL_COUNT + PREY_INITIAL_COUNT
+ True
+ """
+
+ time_passed: Callable[["WaTor", int], None] | None
+
+ def __init__(self, width: int, height: int) -> None:
+ self.width = width
+ self.height = height
+ self.time_passed = None
+
+ self.planet: list[list[Entity | None]] = [[None] * width for _ in range(height)]
+
+ # Populate planet with predators and prey randomly
+ for _ in range(PREY_INITIAL_COUNT):
+ self.add_entity(prey=True)
+ for _ in range(PREDATOR_INITIAL_COUNT):
+ self.add_entity(prey=False)
+ self.set_planet(self.planet)
+
+ def set_planet(self, planet: list[list[Entity | None]]) -> None:
+ """
+ Ease of access for testing
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> planet = [
+ ... [None, None, None],
+ ... [None, Entity(True, coords=(1, 1)), None]
+ ... ]
+ >>> wt.set_planet(planet)
+ >>> wt.planet == planet
+ True
+ >>> wt.width
+ 3
+ >>> wt.height
+ 2
+ """
+ self.planet = planet
+ self.width = len(planet[0])
+ self.height = len(planet)
+
+ def add_entity(self, prey: bool) -> None:
+ """
+ Adds an entity, making sure the entity does
+ not override another entity
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> wt.set_planet([[None, None], [None, None]])
+ >>> wt.add_entity(True)
+ >>> len(wt.get_entities())
+ 1
+ >>> wt.add_entity(False)
+ >>> len(wt.get_entities())
+ 2
+ """
+ while True:
+ row, col = randint(0, self.height - 1), randint(0, self.width - 1)
+ if self.planet[row][col] is None:
+ self.planet[row][col] = Entity(prey=prey, coords=(row, col))
+ return
+
+ def get_entities(self) -> list[Entity]:
+ """
+ Returns a list of all the entities within the planet.
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> len(wt.get_entities()) == PREDATOR_INITIAL_COUNT + PREY_INITIAL_COUNT
+ True
+ """
+ return [entity for column in self.planet for entity in column if entity]
+
+ def balance_predators_and_prey(self) -> None:
+ """
+ Balances predators and preys so that prey
+ can not dominate the predators, blocking up
+ space for them to reproduce.
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> for i in range(2000):
+ ... row, col = i // HEIGHT, i % WIDTH
+ ... wt.planet[row][col] = Entity(True, coords=(row, col))
+ >>> entities = len(wt.get_entities())
+ >>> wt.balance_predators_and_prey()
+ >>> len(wt.get_entities()) == entities
+ False
+ """
+ entities = self.get_entities()
+ shuffle(entities)
+
+ if len(entities) >= MAX_ENTITIES - MAX_ENTITIES / 10:
+ prey = [entity for entity in entities if entity.prey]
+ predators = [entity for entity in entities if not entity.prey]
+
+ prey_count, predator_count = len(prey), len(predators)
+
+ entities_to_purge = (
+ prey[:DELETE_UNBALANCED_ENTITIES]
+ if prey_count > predator_count
+ else predators[:DELETE_UNBALANCED_ENTITIES]
+ )
+ for entity in entities_to_purge:
+ self.planet[entity.coords[0]][entity.coords[1]] = None
+
+ def get_surrounding_prey(self, entity: Entity) -> list[Entity]:
+ """
+ Returns all the prey entities around (N, S, E, W) a predator entity.
+
+ Subtly different to the `move_and_reproduce`.
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> wt.set_planet([
+ ... [None, Entity(True, (0, 1)), None],
+ ... [None, Entity(False, (1, 1)), None],
+ ... [None, Entity(True, (2, 1)), None]])
+ >>> wt.get_surrounding_prey(
+ ... Entity(False, (1, 1))) # doctest: +NORMALIZE_WHITESPACE
+ [Entity(prey=True, coords=(0, 1), remaining_reproduction_time=5),
+ Entity(prey=True, coords=(2, 1), remaining_reproduction_time=5)]
+ >>> wt.set_planet([[Entity(False, (0, 0))]])
+ >>> wt.get_surrounding_prey(Entity(False, (0, 0)))
+ []
+ >>> wt.set_planet([
+ ... [Entity(True, (0, 0)), Entity(False, (1, 0)), Entity(False, (2, 0))],
+ ... [None, Entity(False, (1, 1)), Entity(True, (2, 1))],
+ ... [None, None, None]])
+ >>> wt.get_surrounding_prey(Entity(False, (1, 0)))
+ [Entity(prey=True, coords=(0, 0), remaining_reproduction_time=5)]
+ """
+ row, col = entity.coords
+ adjacent: list[tuple[int, int]] = [
+ (row - 1, col), # North
+ (row + 1, col), # South
+ (row, col - 1), # West
+ (row, col + 1), # East
+ ]
+
+ return [
+ ent
+ for r, c in adjacent
+ if 0 <= r < self.height
+ and 0 <= c < self.width
+ and (ent := self.planet[r][c]) is not None
+ and ent.prey
+ ]
+
+ def move_and_reproduce(
+ self, entity: Entity, direction_orders: list[Literal["N", "E", "S", "W"]]
+ ) -> None:
+ """
+ Attempts to move to an unoccupied neighbouring square
+ in either of the four directions (North, South, East, West).
+ If the move was successful and the `remaining_reproduction_time` is
+ equal to 0, then a new prey or predator can also be created
+ in the previous square.
+
+ :param direction_orders: Ordered list (like priority queue) depicting
+ order to attempt to move. Removes any systematic
+ approach of checking neighbouring squares.
+
+ >>> planet = [
+ ... [None, None, None],
+ ... [None, Entity(True, coords=(1, 1)), None],
+ ... [None, None, None]
+ ... ]
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> wt.set_planet(planet)
+ >>> wt.move_and_reproduce(Entity(True, coords=(1, 1)), direction_orders=["N"])
+ >>> wt.planet # doctest: +NORMALIZE_WHITESPACE
+ [[None, Entity(prey=True, coords=(0, 1), remaining_reproduction_time=4), None],
+ [None, None, None],
+ [None, None, None]]
+ >>> wt.planet[0][0] = Entity(True, coords=(0, 0))
+ >>> wt.move_and_reproduce(Entity(True, coords=(0, 1)),
+ ... direction_orders=["N", "W", "E", "S"])
+ >>> wt.planet # doctest: +NORMALIZE_WHITESPACE
+ [[Entity(prey=True, coords=(0, 0), remaining_reproduction_time=5), None,
+ Entity(prey=True, coords=(0, 2), remaining_reproduction_time=4)],
+ [None, None, None],
+ [None, None, None]]
+ >>> wt.planet[0][1] = wt.planet[0][2]
+ >>> wt.planet[0][2] = None
+ >>> wt.move_and_reproduce(Entity(True, coords=(0, 1)),
+ ... direction_orders=["N", "W", "S", "E"])
+ >>> wt.planet # doctest: +NORMALIZE_WHITESPACE
+ [[Entity(prey=True, coords=(0, 0), remaining_reproduction_time=5), None, None],
+ [None, Entity(prey=True, coords=(1, 1), remaining_reproduction_time=4), None],
+ [None, None, None]]
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> reproducable_entity = Entity(False, coords=(0, 1))
+ >>> reproducable_entity.remaining_reproduction_time = 0
+ >>> wt.planet = [[None, reproducable_entity]]
+ >>> wt.move_and_reproduce(reproducable_entity,
+ ... direction_orders=["N", "W", "S", "E"])
+ >>> wt.planet # doctest: +NORMALIZE_WHITESPACE
+ [[Entity(prey=False, coords=(0, 0),
+ remaining_reproduction_time=20, energy_value=15),
+ Entity(prey=False, coords=(0, 1), remaining_reproduction_time=20,
+ energy_value=15)]]
+ """
+ row, col = coords = entity.coords
+
+ adjacent_squares: dict[Literal["N", "E", "S", "W"], tuple[int, int]] = {
+ "N": (row - 1, col), # North
+ "S": (row + 1, col), # South
+ "W": (row, col - 1), # West
+ "E": (row, col + 1), # East
+ }
+ # Weight adjacent locations
+ adjacent: list[tuple[int, int]] = []
+ for order in direction_orders:
+ adjacent.append(adjacent_squares[order])
+
+ for r, c in adjacent:
+ if (
+ 0 <= r < self.height
+ and 0 <= c < self.width
+ and self.planet[r][c] is None
+ ):
+ # Move entity to empty adjacent square
+ self.planet[r][c] = entity
+ self.planet[row][col] = None
+ entity.coords = (r, c)
+ break
+
+ # (2.) See if it possible to reproduce in previous square
+ if coords != entity.coords and entity.remaining_reproduction_time <= 0:
+ # Check if the entities on the planet is less than the max limit
+ if len(self.get_entities()) < MAX_ENTITIES:
+ # Reproduce in previous square
+ self.planet[row][col] = Entity(prey=entity.prey, coords=coords)
+ entity.reset_reproduction_time()
+ else:
+ entity.remaining_reproduction_time -= 1
+
+ def perform_prey_actions(
+ self, entity: Entity, direction_orders: list[Literal["N", "E", "S", "W"]]
+ ) -> None:
+ """
+ Performs the actions for a prey entity
+
+ For prey the rules are:
+ 1. At each chronon, a prey moves randomly to one of the adjacent unoccupied
+ squares. If there are no free squares, no movement takes place.
+ 2. Once a prey has survived a certain number of chronons it may reproduce.
+ This is done as it moves to a neighbouring square,
+ leaving behind a new prey in its old position.
+ Its reproduction time is also reset to zero.
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> reproducable_entity = Entity(True, coords=(0, 1))
+ >>> reproducable_entity.remaining_reproduction_time = 0
+ >>> wt.planet = [[None, reproducable_entity]]
+ >>> wt.perform_prey_actions(reproducable_entity,
+ ... direction_orders=["N", "W", "S", "E"])
+ >>> wt.planet # doctest: +NORMALIZE_WHITESPACE
+ [[Entity(prey=True, coords=(0, 0), remaining_reproduction_time=5),
+ Entity(prey=True, coords=(0, 1), remaining_reproduction_time=5)]]
+ """
+ self.move_and_reproduce(entity, direction_orders)
+
+ def perform_predator_actions(
+ self,
+ entity: Entity,
+ occupied_by_prey_coords: tuple[int, int] | None,
+ direction_orders: list[Literal["N", "E", "S", "W"]],
+ ) -> None:
+ """
+ Performs the actions for a predator entity
+
+ :param occupied_by_prey_coords: Move to this location if there is prey there
+
+ For predators the rules are:
+ 1. At each chronon, a predator moves randomly to an adjacent square occupied
+ by a prey. If there is none, the predator moves to a random adjacent
+ unoccupied square. If there are no free squares, no movement takes place.
+ 2. At each chronon, each predator is deprived of a unit of energy.
+ 3. Upon reaching zero energy, a predator dies.
+ 4. If a predator moves to a square occupied by a prey,
+ it eats the prey and earns a certain amount of energy.
+ 5. Once a predator has survived a certain number of chronons
+ it may reproduce in exactly the same way as the prey.
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> wt.set_planet([[Entity(True, coords=(0, 0)), Entity(False, coords=(0, 1))]])
+ >>> wt.perform_predator_actions(Entity(False, coords=(0, 1)), (0, 0), [])
+ >>> wt.planet # doctest: +NORMALIZE_WHITESPACE
+ [[Entity(prey=False, coords=(0, 0),
+ remaining_reproduction_time=20, energy_value=19), None]]
+ """
+ assert entity.energy_value is not None # [type checking]
+
+ # (3.) If the entity has 0 energy, it will die
+ if entity.energy_value == 0:
+ self.planet[entity.coords[0]][entity.coords[1]] = None
+ return
+
+ # (1.) Move to entity if possible
+ if occupied_by_prey_coords is not None:
+ # Kill the prey
+ prey = self.planet[occupied_by_prey_coords[0]][occupied_by_prey_coords[1]]
+ assert prey is not None
+ prey.alive = False
+
+ # Move onto prey
+ self.planet[occupied_by_prey_coords[0]][occupied_by_prey_coords[1]] = entity
+ self.planet[entity.coords[0]][entity.coords[1]] = None
+
+ entity.coords = occupied_by_prey_coords
+ # (4.) Eats the prey and earns energy
+ entity.energy_value += PREDATOR_FOOD_VALUE
+ else:
+ # (5.) If it has survived the certain number of chronons it will also
+ # reproduce in this function
+ self.move_and_reproduce(entity, direction_orders)
+
+ # (2.) Each chronon, the predator is deprived of a unit of energy
+ entity.energy_value -= 1
+
+ def run(self, *, iteration_count: int) -> None:
+ """
+ Emulate time passing by looping `iteration_count` times
+
+ >>> wt = WaTor(WIDTH, HEIGHT)
+ >>> wt.run(iteration_count=PREDATOR_INITIAL_ENERGY_VALUE - 1)
+ >>> len(list(filter(lambda entity: entity.prey is False,
+ ... wt.get_entities()))) >= PREDATOR_INITIAL_COUNT
+ True
+ """
+ for iter_num in range(iteration_count):
+ # Generate list of all entities in order to randomly
+ # pop an entity at a time to simulate true randomness
+ # This removes the systematic approach of iterating
+ # through each entity width by height
+ all_entities = self.get_entities()
+
+ for __ in range(len(all_entities)):
+ entity = all_entities.pop(randint(0, len(all_entities) - 1))
+ if entity.alive is False:
+ continue
+
+ directions: list[Literal["N", "E", "S", "W"]] = ["N", "E", "S", "W"]
+ shuffle(directions) # Randomly shuffle directions
+
+ if entity.prey:
+ self.perform_prey_actions(entity, directions)
+ else:
+ # Create list of surrounding prey
+ surrounding_prey = self.get_surrounding_prey(entity)
+ surrounding_prey_coords = None
+
+ if surrounding_prey:
+ # Again, randomly shuffle directions
+ shuffle(surrounding_prey)
+ surrounding_prey_coords = surrounding_prey[0].coords
+
+ self.perform_predator_actions(
+ entity, surrounding_prey_coords, directions
+ )
+
+ # Balance out the predators and prey
+ self.balance_predators_and_prey()
+
+ if self.time_passed is not None:
+ # Call time_passed function for Wa-Tor planet
+ # visualisation in a terminal or a graph.
+ self.time_passed(self, iter_num)
+
+
+def visualise(wt: WaTor, iter_number: int, *, colour: bool = True) -> None:
+ """
+ Visually displays the Wa-Tor planet using
+ an ascii code in terminal to clear and re-print
+ the Wa-Tor planet at intervals.
+
+ Uses ascii colour codes to colourfully display the predators and prey:
+ * (0x60f197) Prey = ``#``
+ * (0xfffff) Predator = ``x``
+
+ >>> wt = WaTor(30, 30)
+ >>> wt.set_planet([
+ ... [Entity(True, coords=(0, 0)), Entity(False, coords=(0, 1)), None],
+ ... [Entity(False, coords=(1, 0)), None, Entity(False, coords=(1, 2))],
+ ... [None, Entity(True, coords=(2, 1)), None]
+ ... ])
+ >>> visualise(wt, 0, colour=False) # doctest: +NORMALIZE_WHITESPACE
+ # x .
+ x . x
+ . # .
+
+ Iteration: 0 | Prey count: 2 | Predator count: 3 |
+ """
+ if colour:
+ __import__("os").system("")
+ print("\x1b[0;0H\x1b[2J\x1b[?25l")
+
+ reprint = "\x1b[0;0H" if colour else ""
+ ansi_colour_end = "\x1b[0m " if colour else " "
+
+ planet = wt.planet
+ output = ""
+
+ # Iterate over every entity in the planet
+ for row in planet:
+ for entity in row:
+ if entity is None:
+ output += " . "
+ else:
+ if colour is True:
+ output += (
+ "\x1b[38;2;96;241;151m"
+ if entity.prey
+ else "\x1b[38;2;255;255;15m"
+ )
+ output += f" {'#' if entity.prey else 'x'}{ansi_colour_end}"
+
+ output += "\n"
+
+ entities = wt.get_entities()
+ prey_count = sum(entity.prey for entity in entities)
+
+ print(
+ f"{output}\n Iteration: {iter_number} | Prey count: {prey_count} | "
+ f"Predator count: {len(entities) - prey_count} | {reprint}"
+ )
+ # Block the thread to be able to visualise seeing the algorithm
+ sleep(0.05)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ wt = WaTor(WIDTH, HEIGHT)
+ wt.time_passed = visualise
+ wt.run(iteration_count=100_000)
diff --git a/ciphers/README.md b/ciphers/README.md
new file mode 100644
index 000000000000..fa09874f38e5
--- /dev/null
+++ b/ciphers/README.md
@@ -0,0 +1,7 @@
+# Ciphers
+
+Ciphers are used to protect data from people that are not allowed to have it. They are everywhere on the internet to protect your connections.
+
+*
+*
+*
diff --git a/ciphers/a1z26.py b/ciphers/a1z26.py
index 92710ec44b0e..a1377ea6d397 100644
--- a/ciphers/a1z26.py
+++ b/ciphers/a1z26.py
@@ -6,8 +6,10 @@
http://bestcodes.weebly.com/a1z26.html
"""
+from __future__ import annotations
-def encode(plain: str) -> list:
+
+def encode(plain: str) -> list[int]:
"""
>>> encode("myname")
[13, 25, 14, 1, 13, 5]
@@ -15,7 +17,7 @@ def encode(plain: str) -> list:
return [ord(elem) - 96 for elem in plain]
-def decode(encoded: list) -> str:
+def decode(encoded: list[int]) -> str:
"""
>>> decode([13, 25, 14, 1, 13, 5])
'myname'
@@ -23,8 +25,8 @@ def decode(encoded: list) -> str:
return "".join(chr(elem + 96) for elem in encoded)
-def main():
- encoded = encode(input("->").strip().lower())
+def main() -> None:
+ encoded = encode(input("-> ").strip().lower())
print("Encoded: ", encoded)
print("Decoded:", decode(encoded))
diff --git a/ciphers/affine_cipher.py b/ciphers/affine_cipher.py
index cf8c0d5f4c1d..10d16367cced 100644
--- a/ciphers/affine_cipher.py
+++ b/ciphers/affine_cipher.py
@@ -1,6 +1,8 @@
import random
import sys
+from maths.greatest_common_divisor import gcd_by_iterative
+
from . import cryptomath_module as cryptomath
SYMBOLS = (
@@ -9,46 +11,26 @@
)
-def main():
- """
- >>> key = get_random_key()
- >>> msg = "This is a test!"
- >>> decrypt_message(key, encrypt_message(key, msg)) == msg
- True
- """
- message = input("Enter message: ").strip()
- key = int(input("Enter key [2000 - 9000]: ").strip())
- mode = input("Encrypt/Decrypt [E/D]: ").strip().lower()
-
- if mode.startswith("e"):
- mode = "encrypt"
- translated = encrypt_message(key, message)
- elif mode.startswith("d"):
- mode = "decrypt"
- translated = decrypt_message(key, message)
- print(f"\n{mode.title()}ed text: \n{translated}")
-
-
-def check_keys(keyA: int, keyB: int, mode: str) -> None:
+def check_keys(key_a: int, key_b: int, mode: str) -> None:
if mode == "encrypt":
- if keyA == 1:
+ if key_a == 1:
sys.exit(
"The affine cipher becomes weak when key "
"A is set to 1. Choose different key"
)
- if keyB == 0:
+ if key_b == 0:
sys.exit(
"The affine cipher becomes weak when key "
"B is set to 0. Choose different key"
)
- if keyA < 0 or keyB < 0 or keyB > len(SYMBOLS) - 1:
+ if key_a < 0 or key_b < 0 or key_b > len(SYMBOLS) - 1:
sys.exit(
"Key A must be greater than 0 and key B must "
f"be between 0 and {len(SYMBOLS) - 1}."
)
- if cryptomath.gcd(keyA, len(SYMBOLS)) != 1:
+ if gcd_by_iterative(key_a, len(SYMBOLS)) != 1:
sys.exit(
- f"Key A {keyA} and the symbol set size {len(SYMBOLS)} "
+ f"Key A {key_a} and the symbol set size {len(SYMBOLS)} "
"are not relatively prime. Choose a different key."
)
@@ -59,16 +41,16 @@ def encrypt_message(key: int, message: str) -> str:
... 'substitution cipher.')
'VL}p MM{I}p~{HL}Gp{vp pFsH}pxMpyxIx JHL O}F{~pvuOvF{FuF{xIp~{HL}Gi'
"""
- keyA, keyB = divmod(key, len(SYMBOLS))
- check_keys(keyA, keyB, "encrypt")
- cipherText = ""
+ key_a, key_b = divmod(key, len(SYMBOLS))
+ check_keys(key_a, key_b, "encrypt")
+ cipher_text = ""
for symbol in message:
if symbol in SYMBOLS:
- symIndex = SYMBOLS.find(symbol)
- cipherText += SYMBOLS[(symIndex * keyA + keyB) % len(SYMBOLS)]
+ sym_index = SYMBOLS.find(symbol)
+ cipher_text += SYMBOLS[(sym_index * key_a + key_b) % len(SYMBOLS)]
else:
- cipherText += symbol
- return cipherText
+ cipher_text += symbol
+ return cipher_text
def decrypt_message(key: int, message: str) -> str:
@@ -77,25 +59,47 @@ def decrypt_message(key: int, message: str) -> str:
... '{xIp~{HL}Gi')
'The affine cipher is a type of monoalphabetic substitution cipher.'
"""
- keyA, keyB = divmod(key, len(SYMBOLS))
- check_keys(keyA, keyB, "decrypt")
- plainText = ""
- modInverseOfkeyA = cryptomath.findModInverse(keyA, len(SYMBOLS))
+ key_a, key_b = divmod(key, len(SYMBOLS))
+ check_keys(key_a, key_b, "decrypt")
+ plain_text = ""
+ mod_inverse_of_key_a = cryptomath.find_mod_inverse(key_a, len(SYMBOLS))
for symbol in message:
if symbol in SYMBOLS:
- symIndex = SYMBOLS.find(symbol)
- plainText += SYMBOLS[(symIndex - keyB) * modInverseOfkeyA % len(SYMBOLS)]
+ sym_index = SYMBOLS.find(symbol)
+ plain_text += SYMBOLS[
+ (sym_index - key_b) * mod_inverse_of_key_a % len(SYMBOLS)
+ ]
else:
- plainText += symbol
- return plainText
+ plain_text += symbol
+ return plain_text
def get_random_key() -> int:
while True:
- keyA = random.randint(2, len(SYMBOLS))
- keyB = random.randint(2, len(SYMBOLS))
- if cryptomath.gcd(keyA, len(SYMBOLS)) == 1 and keyB % len(SYMBOLS) != 0:
- return keyA * len(SYMBOLS) + keyB
+ key_b = random.randint(2, len(SYMBOLS))
+ key_b = random.randint(2, len(SYMBOLS))
+ if gcd_by_iterative(key_b, len(SYMBOLS)) == 1 and key_b % len(SYMBOLS) != 0:
+ return key_b * len(SYMBOLS) + key_b
+
+
+def main() -> None:
+ """
+ >>> key = get_random_key()
+ >>> msg = "This is a test!"
+ >>> decrypt_message(key, encrypt_message(key, msg)) == msg
+ True
+ """
+ message = input("Enter message: ").strip()
+ key = int(input("Enter key [2000 - 9000]: ").strip())
+ mode = input("Encrypt/Decrypt [E/D]: ").strip().lower()
+
+ if mode.startswith("e"):
+ mode = "encrypt"
+ translated = encrypt_message(key, message)
+ elif mode.startswith("d"):
+ mode = "decrypt"
+ translated = decrypt_message(key, message)
+ print(f"\n{mode.title()}ed text: \n{translated}")
if __name__ == "__main__":
diff --git a/ciphers/atbash.py b/ciphers/atbash.py
index c17d1e34f37a..4e8f663ed02d 100644
--- a/ciphers/atbash.py
+++ b/ciphers/atbash.py
@@ -1,4 +1,5 @@
-""" https://en.wikipedia.org/wiki/Atbash """
+"""https://en.wikipedia.org/wiki/Atbash"""
+
import string
@@ -38,29 +39,16 @@ def atbash(sequence: str) -> str:
def benchmark() -> None:
- """Let's benchmark them side-by-side..."""
+ """Let's benchmark our functions side-by-side..."""
from timeit import timeit
print("Running performance benchmarks...")
- print(
- "> atbash_slow()",
- timeit(
- "atbash_slow(printable)",
- setup="from string import printable ; from __main__ import atbash_slow",
- ),
- "seconds",
- )
- print(
- "> atbash()",
- timeit(
- "atbash(printable)",
- setup="from string import printable ; from __main__ import atbash",
- ),
- "seconds",
- )
+ setup = "from string import printable ; from __main__ import atbash, atbash_slow"
+ print(f"> atbash_slow(): {timeit('atbash_slow(printable)', setup=setup)} seconds")
+ print(f"> atbash(): {timeit('atbash(printable)', setup=setup)} seconds")
if __name__ == "__main__":
- for sequence in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"):
- print(f"{sequence} encrypted in atbash: {atbash(sequence)}")
+ for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"):
+ print(f"{example} encrypted in atbash: {atbash(example)}")
benchmark()
diff --git a/ciphers/autokey.py b/ciphers/autokey.py
new file mode 100644
index 000000000000..7751a32d7546
--- /dev/null
+++ b/ciphers/autokey.py
@@ -0,0 +1,150 @@
+"""
+https://en.wikipedia.org/wiki/Autokey_cipher
+
+An autokey cipher (also known as the autoclave cipher) is a cipher that
+incorporates the message (the plaintext) into the key.
+The key is generated from the message in some automated fashion,
+sometimes by selecting certain letters from the text or, more commonly,
+by adding a short primer key to the front of the message.
+"""
+
+
+def encrypt(plaintext: str, key: str) -> str:
+ """
+ Encrypt a given `plaintext` (string) and `key` (string), returning the
+ encrypted ciphertext.
+
+ >>> encrypt("hello world", "coffee")
+ 'jsqqs avvwo'
+ >>> encrypt("coffee is good as python", "TheAlgorithms")
+ 'vvjfpk wj ohvp su ddylsv'
+ >>> encrypt("coffee is good as python", 2)
+ Traceback (most recent call last):
+ ...
+ TypeError: key must be a string
+ >>> encrypt("", "TheAlgorithms")
+ Traceback (most recent call last):
+ ...
+ ValueError: plaintext is empty
+ >>> encrypt("coffee is good as python", "")
+ Traceback (most recent call last):
+ ...
+ ValueError: key is empty
+ >>> encrypt(527.26, "TheAlgorithms")
+ Traceback (most recent call last):
+ ...
+ TypeError: plaintext must be a string
+ """
+ if not isinstance(plaintext, str):
+ raise TypeError("plaintext must be a string")
+ if not isinstance(key, str):
+ raise TypeError("key must be a string")
+
+ if not plaintext:
+ raise ValueError("plaintext is empty")
+ if not key:
+ raise ValueError("key is empty")
+
+ key += plaintext
+ plaintext = plaintext.lower()
+ key = key.lower()
+ plaintext_iterator = 0
+ key_iterator = 0
+ ciphertext = ""
+ while plaintext_iterator < len(plaintext):
+ if (
+ ord(plaintext[plaintext_iterator]) < 97
+ or ord(plaintext[plaintext_iterator]) > 122
+ ):
+ ciphertext += plaintext[plaintext_iterator]
+ plaintext_iterator += 1
+ elif ord(key[key_iterator]) < 97 or ord(key[key_iterator]) > 122:
+ key_iterator += 1
+ else:
+ ciphertext += chr(
+ (
+ (ord(plaintext[plaintext_iterator]) - 97 + ord(key[key_iterator]))
+ - 97
+ )
+ % 26
+ + 97
+ )
+ key_iterator += 1
+ plaintext_iterator += 1
+ return ciphertext
+
+
+def decrypt(ciphertext: str, key: str) -> str:
+ """
+ Decrypt a given `ciphertext` (string) and `key` (string), returning the decrypted
+ ciphertext.
+
+ >>> decrypt("jsqqs avvwo", "coffee")
+ 'hello world'
+ >>> decrypt("vvjfpk wj ohvp su ddylsv", "TheAlgorithms")
+ 'coffee is good as python'
+ >>> decrypt("vvjfpk wj ohvp su ddylsv", "")
+ Traceback (most recent call last):
+ ...
+ ValueError: key is empty
+ >>> decrypt(527.26, "TheAlgorithms")
+ Traceback (most recent call last):
+ ...
+ TypeError: ciphertext must be a string
+ >>> decrypt("", "TheAlgorithms")
+ Traceback (most recent call last):
+ ...
+ ValueError: ciphertext is empty
+ >>> decrypt("vvjfpk wj ohvp su ddylsv", 2)
+ Traceback (most recent call last):
+ ...
+ TypeError: key must be a string
+ """
+ if not isinstance(ciphertext, str):
+ raise TypeError("ciphertext must be a string")
+ if not isinstance(key, str):
+ raise TypeError("key must be a string")
+
+ if not ciphertext:
+ raise ValueError("ciphertext is empty")
+ if not key:
+ raise ValueError("key is empty")
+
+ key = key.lower()
+ ciphertext_iterator = 0
+ key_iterator = 0
+ plaintext = ""
+ while ciphertext_iterator < len(ciphertext):
+ if (
+ ord(ciphertext[ciphertext_iterator]) < 97
+ or ord(ciphertext[ciphertext_iterator]) > 122
+ ):
+ plaintext += ciphertext[ciphertext_iterator]
+ else:
+ plaintext += chr(
+ (ord(ciphertext[ciphertext_iterator]) - ord(key[key_iterator])) % 26
+ + 97
+ )
+ key += chr(
+ (ord(ciphertext[ciphertext_iterator]) - ord(key[key_iterator])) % 26
+ + 97
+ )
+ key_iterator += 1
+ ciphertext_iterator += 1
+ return plaintext
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ operation = int(input("Type 1 to encrypt or 2 to decrypt:"))
+ if operation == 1:
+ plaintext = input("Typeplaintext to be encrypted:\n")
+ key = input("Type the key:\n")
+ print(encrypt(plaintext, key))
+ elif operation == 2:
+ ciphertext = input("Type the ciphertext to be decrypted:\n")
+ key = input("Type the key:\n")
+ print(decrypt(ciphertext, key))
+ decrypt("jsqqs avvwo", "coffee")
diff --git a/ciphers/baconian_cipher.py b/ciphers/baconian_cipher.py
new file mode 100644
index 000000000000..f146ba91b78f
--- /dev/null
+++ b/ciphers/baconian_cipher.py
@@ -0,0 +1,89 @@
+"""
+Program to encode and decode Baconian or Bacon's Cipher
+Wikipedia reference : https://en.wikipedia.org/wiki/Bacon%27s_cipher
+"""
+
+encode_dict = {
+ "a": "AAAAA",
+ "b": "AAAAB",
+ "c": "AAABA",
+ "d": "AAABB",
+ "e": "AABAA",
+ "f": "AABAB",
+ "g": "AABBA",
+ "h": "AABBB",
+ "i": "ABAAA",
+ "j": "BBBAA",
+ "k": "ABAAB",
+ "l": "ABABA",
+ "m": "ABABB",
+ "n": "ABBAA",
+ "o": "ABBAB",
+ "p": "ABBBA",
+ "q": "ABBBB",
+ "r": "BAAAA",
+ "s": "BAAAB",
+ "t": "BAABA",
+ "u": "BAABB",
+ "v": "BBBAB",
+ "w": "BABAA",
+ "x": "BABAB",
+ "y": "BABBA",
+ "z": "BABBB",
+ " ": " ",
+}
+
+
+decode_dict = {value: key for key, value in encode_dict.items()}
+
+
+def encode(word: str) -> str:
+ """
+ Encodes to Baconian cipher
+
+ >>> encode("hello")
+ 'AABBBAABAAABABAABABAABBAB'
+ >>> encode("hello world")
+ 'AABBBAABAAABABAABABAABBAB BABAAABBABBAAAAABABAAAABB'
+ >>> encode("hello world!")
+ Traceback (most recent call last):
+ ...
+ Exception: encode() accepts only letters of the alphabet and spaces
+ """
+ encoded = ""
+ for letter in word.lower():
+ if letter.isalpha() or letter == " ":
+ encoded += encode_dict[letter]
+ else:
+ raise Exception("encode() accepts only letters of the alphabet and spaces")
+ return encoded
+
+
+def decode(coded: str) -> str:
+ """
+ Decodes from Baconian cipher
+
+ >>> decode("AABBBAABAAABABAABABAABBAB BABAAABBABBAAAAABABAAAABB")
+ 'hello world'
+ >>> decode("AABBBAABAAABABAABABAABBAB")
+ 'hello'
+ >>> decode("AABBBAABAAABABAABABAABBAB BABAAABBABBAAAAABABAAAABB!")
+ Traceback (most recent call last):
+ ...
+ Exception: decode() accepts only 'A', 'B' and spaces
+ """
+ if set(coded) - {"A", "B", " "} != set():
+ raise Exception("decode() accepts only 'A', 'B' and spaces")
+ decoded = ""
+ for word in coded.split():
+ while len(word) != 0:
+ decoded += decode_dict[word[:5]]
+ word = word[5:]
+ decoded += " "
+ return decoded.strip()
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/ciphers/base16.py b/ciphers/base16.py
index 0210315d54e6..6cd62846fc87 100644
--- a/ciphers/base16.py
+++ b/ciphers/base16.py
@@ -1,13 +1,66 @@
-import base64
+def base16_encode(data: bytes) -> str:
+ """
+ Encodes the given bytes into base16.
+ >>> base16_encode(b'Hello World!')
+ '48656C6C6F20576F726C6421'
+ >>> base16_encode(b'HELLO WORLD!')
+ '48454C4C4F20574F524C4421'
+ >>> base16_encode(b'')
+ ''
+ """
+ # Turn the data into a list of integers (where each integer is a byte),
+ # Then turn each byte into its hexadecimal representation, make sure
+ # it is uppercase, and then join everything together and return it.
+ return "".join([hex(byte)[2:].zfill(2).upper() for byte in list(data)])
-def main():
- inp = input("->")
- encoded = inp.encode("utf-8") # encoded the input (we need a bytes like object)
- b16encoded = base64.b16encode(encoded) # b16encoded the encoded string
- print(b16encoded)
- print(base64.b16decode(b16encoded).decode("utf-8")) # decoded it
+
+def base16_decode(data: str) -> bytes:
+ """
+ Decodes the given base16 encoded data into bytes.
+
+ >>> base16_decode('48656C6C6F20576F726C6421')
+ b'Hello World!'
+ >>> base16_decode('48454C4C4F20574F524C4421')
+ b'HELLO WORLD!'
+ >>> base16_decode('')
+ b''
+ >>> base16_decode('486')
+ Traceback (most recent call last):
+ ...
+ ValueError: Base16 encoded data is invalid:
+ Data does not have an even number of hex digits.
+ >>> base16_decode('48656c6c6f20576f726c6421')
+ Traceback (most recent call last):
+ ...
+ ValueError: Base16 encoded data is invalid:
+ Data is not uppercase hex or it contains invalid characters.
+ >>> base16_decode('This is not base64 encoded data.')
+ Traceback (most recent call last):
+ ...
+ ValueError: Base16 encoded data is invalid:
+ Data is not uppercase hex or it contains invalid characters.
+ """
+ # Check data validity, following RFC3548
+ # https://www.ietf.org/rfc/rfc3548.txt
+ if (len(data) % 2) != 0:
+ raise ValueError(
+ """Base16 encoded data is invalid:
+Data does not have an even number of hex digits."""
+ )
+ # Check the character set - the standard base16 alphabet
+ # is uppercase according to RFC3548 section 6
+ if not set(data) <= set("0123456789ABCDEF"):
+ raise ValueError(
+ """Base16 encoded data is invalid:
+Data is not uppercase hex or it contains invalid characters."""
+ )
+ # For every two hexadecimal digits (= a byte), turn it into an integer.
+ # Then, string the result together into bytes, and return it.
+ return bytes(int(data[i] + data[i + 1], 16) for i in range(0, len(data), 2))
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod()
diff --git a/ciphers/base32.py b/ciphers/base32.py
index 5bba8c4dd685..911afa2452c0 100644
--- a/ciphers/base32.py
+++ b/ciphers/base32.py
@@ -1,13 +1,46 @@
-import base64
+"""
+Base32 encoding and decoding
+https://en.wikipedia.org/wiki/Base32
+"""
-def main():
- inp = input("->")
- encoded = inp.encode("utf-8") # encoded the input (we need a bytes like object)
- b32encoded = base64.b32encode(encoded) # b32encoded the encoded string
- print(b32encoded)
- print(base64.b32decode(b32encoded).decode("utf-8")) # decoded it
+B32_CHARSET = "ABCDEFGHIJKLMNOPQRSTUVWXYZ234567"
+
+
+def base32_encode(data: bytes) -> bytes:
+ """
+ >>> base32_encode(b"Hello World!")
+ b'JBSWY3DPEBLW64TMMQQQ===='
+ >>> base32_encode(b"123456")
+ b'GEZDGNBVGY======'
+ >>> base32_encode(b"some long complex string")
+ b'ONXW2ZJANRXW4ZZAMNXW24DMMV4CA43UOJUW4ZY='
+ """
+ binary_data = "".join(bin(ord(d))[2:].zfill(8) for d in data.decode("utf-8"))
+ binary_data = binary_data.ljust(5 * ((len(binary_data) // 5) + 1), "0")
+ b32_chunks = map("".join, zip(*[iter(binary_data)] * 5))
+ b32_result = "".join(B32_CHARSET[int(chunk, 2)] for chunk in b32_chunks)
+ return bytes(b32_result.ljust(8 * ((len(b32_result) // 8) + 1), "="), "utf-8")
+
+
+def base32_decode(data: bytes) -> bytes:
+ """
+ >>> base32_decode(b'JBSWY3DPEBLW64TMMQQQ====')
+ b'Hello World!'
+ >>> base32_decode(b'GEZDGNBVGY======')
+ b'123456'
+ >>> base32_decode(b'ONXW2ZJANRXW4ZZAMNXW24DMMV4CA43UOJUW4ZY=')
+ b'some long complex string'
+ """
+ binary_chunks = "".join(
+ bin(B32_CHARSET.index(_d))[2:].zfill(5)
+ for _d in data.decode("utf-8").strip("=")
+ )
+ binary_data = list(map("".join, zip(*[iter(binary_chunks)] * 8)))
+ return bytes("".join([chr(int(_d, 2)) for _d in binary_data]), "utf-8")
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod()
diff --git a/ciphers/base64_cipher.py b/ciphers/base64_cipher.py
index 1dbe74a20fe7..038d13963d95 100644
--- a/ciphers/base64_cipher.py
+++ b/ciphers/base64_cipher.py
@@ -1,89 +1,142 @@
-def encode_base64(text: str) -> str:
- r"""
- >>> encode_base64('WELCOME to base64 encoding 😁')
- 'V0VMQ09NRSB0byBiYXNlNjQgZW5jb2Rpbmcg8J+YgQ=='
- >>> encode_base64('AÅᐃ𐀏🤓')
- 'QcOF4ZCD8JCAj/CfpJM='
- >>> encode_base64('A'*60)
- 'QUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFB\r\nQUFB'
+B64_CHARSET = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/"
+
+
+def base64_encode(data: bytes) -> bytes:
+ """Encodes data according to RFC4648.
+
+ The data is first transformed to binary and appended with binary digits so that its
+ length becomes a multiple of 6, then each 6 binary digits will match a character in
+ the B64_CHARSET string. The number of appended binary digits would later determine
+ how many "=" signs should be added, the padding.
+ For every 2 binary digits added, a "=" sign is added in the output.
+ We can add any binary digits to make it a multiple of 6, for instance, consider the
+ following example:
+ "AA" -> 0010100100101001 -> 001010 010010 1001
+ As can be seen above, 2 more binary digits should be added, so there's 4
+ possibilities here: 00, 01, 10 or 11.
+ That being said, Base64 encoding can be used in Steganography to hide data in these
+ appended digits.
+
+ >>> from base64 import b64encode
+ >>> a = b"This pull request is part of Hacktoberfest20!"
+ >>> b = b"https://tools.ietf.org/html/rfc4648"
+ >>> c = b"A"
+ >>> base64_encode(a) == b64encode(a)
+ True
+ >>> base64_encode(b) == b64encode(b)
+ True
+ >>> base64_encode(c) == b64encode(c)
+ True
+ >>> base64_encode("abc")
+ Traceback (most recent call last):
+ ...
+ TypeError: a bytes-like object is required, not 'str'
"""
- base64_chars = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/"
+ # Make sure the supplied data is a bytes-like object
+ if not isinstance(data, bytes):
+ msg = f"a bytes-like object is required, not '{data.__class__.__name__}'"
+ raise TypeError(msg)
- byte_text = bytes(text, "utf-8") # put text in bytes for unicode support
- r = "" # the result
- c = -len(byte_text) % 3 # the length of padding
- p = "=" * c # the padding
- s = byte_text + b"\x00" * c # the text to encode
+ binary_stream = "".join(bin(byte)[2:].zfill(8) for byte in data)
- i = 0
- while i < len(s):
- if i > 0 and ((i / 3 * 4) % 76) == 0:
- r = r + "\r\n" # for unix newline, put "\n"
+ padding_needed = len(binary_stream) % 6 != 0
- n = (s[i] << 16) + (s[i + 1] << 8) + s[i + 2]
+ if padding_needed:
+ # The padding that will be added later
+ padding = b"=" * ((6 - len(binary_stream) % 6) // 2)
- n1 = (n >> 18) & 63
- n2 = (n >> 12) & 63
- n3 = (n >> 6) & 63
- n4 = n & 63
-
- r += base64_chars[n1] + base64_chars[n2] + base64_chars[n3] + base64_chars[n4]
- i += 3
-
- return r[0 : len(r) - len(p)] + p
-
-
-def decode_base64(text: str) -> str:
- r"""
- >>> decode_base64('V0VMQ09NRSB0byBiYXNlNjQgZW5jb2Rpbmcg8J+YgQ==')
- 'WELCOME to base64 encoding 😁'
- >>> decode_base64('QcOF4ZCD8JCAj/CfpJM=')
- 'AÅᐃ𐀏🤓'
- >>> decode_base64("QUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUF"
- ... "BQUFBQUFBQUFB\r\nQUFB")
- 'AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA'
+ # Append binary_stream with arbitrary binary digits (0's by default) to make its
+ # length a multiple of 6.
+ binary_stream += "0" * (6 - len(binary_stream) % 6)
+ else:
+ padding = b""
+
+ # Encode every 6 binary digits to their corresponding Base64 character
+ return (
+ "".join(
+ B64_CHARSET[int(binary_stream[index : index + 6], 2)]
+ for index in range(0, len(binary_stream), 6)
+ ).encode()
+ + padding
+ )
+
+
+def base64_decode(encoded_data: str) -> bytes:
+ """Decodes data according to RFC4648.
+
+ This does the reverse operation of base64_encode.
+ We first transform the encoded data back to a binary stream, take off the
+ previously appended binary digits according to the padding, at this point we
+ would have a binary stream whose length is multiple of 8, the last step is
+ to convert every 8 bits to a byte.
+
+ >>> from base64 import b64decode
+ >>> a = "VGhpcyBwdWxsIHJlcXVlc3QgaXMgcGFydCBvZiBIYWNrdG9iZXJmZXN0MjAh"
+ >>> b = "aHR0cHM6Ly90b29scy5pZXRmLm9yZy9odG1sL3JmYzQ2NDg="
+ >>> c = "QQ=="
+ >>> base64_decode(a) == b64decode(a)
+ True
+ >>> base64_decode(b) == b64decode(b)
+ True
+ >>> base64_decode(c) == b64decode(c)
+ True
+ >>> base64_decode("abc")
+ Traceback (most recent call last):
+ ...
+ AssertionError: Incorrect padding
"""
- base64_chars = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/"
- s = ""
-
- for i in text:
- if i in base64_chars:
- s += i
- c = ""
- else:
- if i == "=":
- c += "="
-
- p = ""
- if c == "=":
- p = "A"
+ # Make sure encoded_data is either a string or a bytes-like object
+ if not isinstance(encoded_data, bytes) and not isinstance(encoded_data, str):
+ msg = (
+ "argument should be a bytes-like object or ASCII string, "
+ f"not '{encoded_data.__class__.__name__}'"
+ )
+ raise TypeError(msg)
+
+ # In case encoded_data is a bytes-like object, make sure it contains only
+ # ASCII characters so we convert it to a string object
+ if isinstance(encoded_data, bytes):
+ try:
+ encoded_data = encoded_data.decode("utf-8")
+ except UnicodeDecodeError:
+ raise ValueError("base64 encoded data should only contain ASCII characters")
+
+ padding = encoded_data.count("=")
+
+ # Check if the encoded string contains non base64 characters
+ if padding:
+ assert all(char in B64_CHARSET for char in encoded_data[:-padding]), (
+ "Invalid base64 character(s) found."
+ )
else:
- if c == "==":
- p = "AA"
-
- r = b""
- s = s + p
-
- i = 0
- while i < len(s):
- n = (
- (base64_chars.index(s[i]) << 18)
- + (base64_chars.index(s[i + 1]) << 12)
- + (base64_chars.index(s[i + 2]) << 6)
- + base64_chars.index(s[i + 3])
+ assert all(char in B64_CHARSET for char in encoded_data), (
+ "Invalid base64 character(s) found."
)
- r += bytes([(n >> 16) & 255]) + bytes([(n >> 8) & 255]) + bytes([n & 255])
+ # Check the padding
+ assert len(encoded_data) % 4 == 0 and padding < 3, "Incorrect padding"
- i += 4
+ if padding:
+ # Remove padding if there is one
+ encoded_data = encoded_data[:-padding]
- return str(r[0 : len(r) - len(p)], "utf-8")
+ binary_stream = "".join(
+ bin(B64_CHARSET.index(char))[2:].zfill(6) for char in encoded_data
+ )[: -padding * 2]
+ else:
+ binary_stream = "".join(
+ bin(B64_CHARSET.index(char))[2:].zfill(6) for char in encoded_data
+ )
+ data = [
+ int(binary_stream[index : index + 8], 2)
+ for index in range(0, len(binary_stream), 8)
+ ]
-def main():
- print(encode_base64("WELCOME to base64 encoding 😁"))
- print(decode_base64(encode_base64("WELCOME to base64 encoding 😁")))
+ return bytes(data)
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod()
diff --git a/ciphers/base85.py b/ciphers/base85.py
index ebfd0480f794..f0228e5052dd 100644
--- a/ciphers/base85.py
+++ b/ciphers/base85.py
@@ -1,13 +1,58 @@
-import base64
+"""
+Base85 (Ascii85) encoding and decoding
+https://en.wikipedia.org/wiki/Ascii85
+"""
-def main():
- inp = input("->")
- encoded = inp.encode("utf-8") # encoded the input (we need a bytes like object)
- a85encoded = base64.a85encode(encoded) # a85encoded the encoded string
- print(a85encoded)
- print(base64.a85decode(a85encoded).decode("utf-8")) # decoded it
+
+def _base10_to_85(d: int) -> str:
+ return "".join(chr(d % 85 + 33)) + _base10_to_85(d // 85) if d > 0 else ""
+
+
+def _base85_to_10(digits: list) -> int:
+ return sum(char * 85**i for i, char in enumerate(reversed(digits)))
+
+
+def ascii85_encode(data: bytes) -> bytes:
+ """
+ >>> ascii85_encode(b"")
+ b''
+ >>> ascii85_encode(b"12345")
+ b'0etOA2#'
+ >>> ascii85_encode(b"base 85")
+ b'@UX=h+?24'
+ """
+ binary_data = "".join(bin(ord(d))[2:].zfill(8) for d in data.decode("utf-8"))
+ null_values = (32 * ((len(binary_data) // 32) + 1) - len(binary_data)) // 8
+ binary_data = binary_data.ljust(32 * ((len(binary_data) // 32) + 1), "0")
+ b85_chunks = [int(_s, 2) for _s in map("".join, zip(*[iter(binary_data)] * 32))]
+ result = "".join(_base10_to_85(chunk)[::-1] for chunk in b85_chunks)
+ return bytes(result[:-null_values] if null_values % 4 != 0 else result, "utf-8")
+
+
+def ascii85_decode(data: bytes) -> bytes:
+ """
+ >>> ascii85_decode(b"")
+ b''
+ >>> ascii85_decode(b"0etOA2#")
+ b'12345'
+ >>> ascii85_decode(b"@UX=h+?24")
+ b'base 85'
+ """
+ null_values = 5 * ((len(data) // 5) + 1) - len(data)
+ binary_data = data.decode("utf-8") + "u" * null_values
+ b85_chunks = map("".join, zip(*[iter(binary_data)] * 5))
+ b85_segments = [[ord(_s) - 33 for _s in chunk] for chunk in b85_chunks]
+ results = [bin(_base85_to_10(chunk))[2::].zfill(32) for chunk in b85_segments]
+ char_chunks = [
+ [chr(int(_s, 2)) for _s in map("".join, zip(*[iter(r)] * 8))] for r in results
+ ]
+ result = "".join("".join(char) for char in char_chunks)
+ offset = int(null_values % 5 == 0)
+ return bytes(result[: offset - null_values], "utf-8")
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod()
diff --git a/ciphers/beaufort_cipher.py b/ciphers/beaufort_cipher.py
index c885dec74001..788fc72b89c3 100644
--- a/ciphers/beaufort_cipher.py
+++ b/ciphers/beaufort_cipher.py
@@ -5,7 +5,7 @@
from string import ascii_uppercase
dict1 = {char: i for i, char in enumerate(ascii_uppercase)}
-dict2 = {i: char for i, char in enumerate(ascii_uppercase)}
+dict2 = dict(enumerate(ascii_uppercase))
# This function generates the key in
@@ -66,7 +66,7 @@ def original_text(cipher_text: str, key_new: str) -> str:
return or_txt
-def main():
+def main() -> None:
message = "THE GERMAN ATTACK"
key = "SECRET"
key_new = generate_key(message, key)
diff --git a/ciphers/bifid.py b/ciphers/bifid.py
new file mode 100644
index 000000000000..a15b381640aa
--- /dev/null
+++ b/ciphers/bifid.py
@@ -0,0 +1,111 @@
+#!/usr/bin/env python3
+
+"""
+The Bifid Cipher uses a Polybius Square to encipher a message in a way that
+makes it fairly difficult to decipher without knowing the secret.
+
+https://www.braingle.com/brainteasers/codes/bifid.php
+"""
+
+import numpy as np
+
+SQUARE = [
+ ["a", "b", "c", "d", "e"],
+ ["f", "g", "h", "i", "k"],
+ ["l", "m", "n", "o", "p"],
+ ["q", "r", "s", "t", "u"],
+ ["v", "w", "x", "y", "z"],
+]
+
+
+class BifidCipher:
+ def __init__(self) -> None:
+ self.SQUARE = np.array(SQUARE)
+
+ def letter_to_numbers(self, letter: str) -> np.ndarray:
+ """
+ Return the pair of numbers that represents the given letter in the
+ polybius square
+
+ >>> np.array_equal(BifidCipher().letter_to_numbers('a'), [1,1])
+ True
+
+ >>> np.array_equal(BifidCipher().letter_to_numbers('u'), [4,5])
+ True
+ """
+ index1, index2 = np.where(letter == self.SQUARE)
+ indexes = np.concatenate([index1 + 1, index2 + 1])
+ return indexes
+
+ def numbers_to_letter(self, index1: int, index2: int) -> str:
+ """
+ Return the letter corresponding to the position [index1, index2] in
+ the polybius square
+
+ >>> BifidCipher().numbers_to_letter(4, 5) == "u"
+ True
+
+ >>> BifidCipher().numbers_to_letter(1, 1) == "a"
+ True
+ """
+ letter = self.SQUARE[index1 - 1, index2 - 1]
+ return letter
+
+ def encode(self, message: str) -> str:
+ """
+ Return the encoded version of message according to the polybius cipher
+
+ >>> BifidCipher().encode('testmessage') == 'qtltbdxrxlk'
+ True
+
+ >>> BifidCipher().encode('Test Message') == 'qtltbdxrxlk'
+ True
+
+ >>> BifidCipher().encode('test j') == BifidCipher().encode('test i')
+ True
+ """
+ message = message.lower()
+ message = message.replace(" ", "")
+ message = message.replace("j", "i")
+
+ first_step = np.empty((2, len(message)))
+ for letter_index in range(len(message)):
+ numbers = self.letter_to_numbers(message[letter_index])
+
+ first_step[0, letter_index] = numbers[0]
+ first_step[1, letter_index] = numbers[1]
+
+ second_step = first_step.reshape(2 * len(message))
+ encoded_message = ""
+ for numbers_index in range(len(message)):
+ index1 = int(second_step[numbers_index * 2])
+ index2 = int(second_step[(numbers_index * 2) + 1])
+ letter = self.numbers_to_letter(index1, index2)
+ encoded_message = encoded_message + letter
+
+ return encoded_message
+
+ def decode(self, message: str) -> str:
+ """
+ Return the decoded version of message according to the polybius cipher
+
+ >>> BifidCipher().decode('qtltbdxrxlk') == 'testmessage'
+ True
+ """
+ message = message.lower()
+ message.replace(" ", "")
+ first_step = np.empty(2 * len(message))
+ for letter_index in range(len(message)):
+ numbers = self.letter_to_numbers(message[letter_index])
+ first_step[letter_index * 2] = numbers[0]
+ first_step[letter_index * 2 + 1] = numbers[1]
+
+ second_step = first_step.reshape((2, len(message)))
+ decoded_message = ""
+ for numbers_index in range(len(message)):
+ index1 = int(second_step[0, numbers_index])
+ index2 = int(second_step[1, numbers_index])
+ letter = self.numbers_to_letter(index1, index2)
+ decoded_message = decoded_message + letter
+
+ return decoded_message
diff --git a/ciphers/brute_force_caesar_cipher.py b/ciphers/brute_force_caesar_cipher.py
index 13a165245403..458d08db2628 100644
--- a/ciphers/brute_force_caesar_cipher.py
+++ b/ciphers/brute_force_caesar_cipher.py
@@ -1,3 +1,6 @@
+import string
+
+
def decrypt(message: str) -> None:
"""
>>> decrypt('TMDETUX PMDVU')
@@ -28,22 +31,21 @@ def decrypt(message: str) -> None:
Decryption using Key #24: VOFGVWZ ROFXW
Decryption using Key #25: UNEFUVY QNEWV
"""
- LETTERS = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
- for key in range(len(LETTERS)):
+ for key in range(len(string.ascii_uppercase)):
translated = ""
for symbol in message:
- if symbol in LETTERS:
- num = LETTERS.find(symbol)
+ if symbol in string.ascii_uppercase:
+ num = string.ascii_uppercase.find(symbol)
num = num - key
if num < 0:
- num = num + len(LETTERS)
- translated = translated + LETTERS[num]
+ num = num + len(string.ascii_uppercase)
+ translated = translated + string.ascii_uppercase[num]
else:
translated = translated + symbol
print(f"Decryption using Key #{key}: {translated}")
-def main():
+def main() -> None:
message = input("Encrypted message: ")
message = message.upper()
decrypt(message)
diff --git a/ciphers/caesar_cipher.py b/ciphers/caesar_cipher.py
index 4038919e5dde..1cf4d67cbaed 100644
--- a/ciphers/caesar_cipher.py
+++ b/ciphers/caesar_cipher.py
@@ -1,55 +1,64 @@
+from __future__ import annotations
+
from string import ascii_letters
-def encrypt(input_string: str, key: int, alphabet=None) -> str:
+def encrypt(input_string: str, key: int, alphabet: str | None = None) -> str:
"""
encrypt
=======
+
Encodes a given string with the caesar cipher and returns the encoded
message
Parameters:
-----------
- * input_string: the plain-text that needs to be encoded
- * key: the number of letters to shift the message by
+
+ * `input_string`: the plain-text that needs to be encoded
+ * `key`: the number of letters to shift the message by
Optional:
- * alphabet (None): the alphabet used to encode the cipher, if not
+
+ * `alphabet` (``None``): the alphabet used to encode the cipher, if not
specified, the standard english alphabet with upper and lowercase
letters is used
Returns:
+
* A string containing the encoded cipher-text
More on the caesar cipher
=========================
+
The caesar cipher is named after Julius Caesar who used it when sending
secret military messages to his troops. This is a simple substitution cipher
- where very character in the plain-text is shifted by a certain number known
+ where every character in the plain-text is shifted by a certain number known
as the "key" or "shift".
Example:
Say we have the following message:
- "Hello, captain"
+ ``Hello, captain``
And our alphabet is made up of lower and uppercase letters:
- "abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ"
+ ``abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ``
- And our shift is "2"
+ And our shift is ``2``
- We can then encode the message, one letter at a time. "H" would become "J",
- since "J" is two letters away, and so on. If the shift is ever two large, or
+ We can then encode the message, one letter at a time. ``H`` would become ``J``,
+ since ``J`` is two letters away, and so on. If the shift is ever two large, or
our letter is at the end of the alphabet, we just start at the beginning
- ("Z" would shift to "a" then "b" and so on).
+ (``Z`` would shift to ``a`` then ``b`` and so on).
- Our final message would be "Jgnnq, ecrvckp"
+ Our final message would be ``Jgnnq, ecrvckp``
Further reading
===============
+
* https://en.m.wikipedia.org/wiki/Caesar_cipher
Doctests
========
+
>>> encrypt('The quick brown fox jumps over the lazy dog', 8)
'bpm yCqks jzwEv nwF rCuxA wDmz Bpm tiHG lwo'
@@ -79,27 +88,32 @@ def encrypt(input_string: str, key: int, alphabet=None) -> str:
return result
-def decrypt(input_string: str, key: int, alphabet=None) -> str:
+def decrypt(input_string: str, key: int, alphabet: str | None = None) -> str:
"""
decrypt
=======
+
Decodes a given string of cipher-text and returns the decoded plain-text
Parameters:
-----------
- * input_string: the cipher-text that needs to be decoded
- * key: the number of letters to shift the message backwards by to decode
+
+ * `input_string`: the cipher-text that needs to be decoded
+ * `key`: the number of letters to shift the message backwards by to decode
Optional:
- * alphabet (None): the alphabet used to decode the cipher, if not
+
+ * `alphabet` (``None``): the alphabet used to decode the cipher, if not
specified, the standard english alphabet with upper and lowercase
letters is used
Returns:
+
* A string containing the decoded plain-text
More on the caesar cipher
=========================
+
The caesar cipher is named after Julius Caesar who used it when sending
secret military messages to his troops. This is a simple substitution cipher
where very character in the plain-text is shifted by a certain number known
@@ -108,27 +122,29 @@ def decrypt(input_string: str, key: int, alphabet=None) -> str:
Example:
Say we have the following cipher-text:
- "Jgnnq, ecrvckp"
+ ``Jgnnq, ecrvckp``
And our alphabet is made up of lower and uppercase letters:
- "abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ"
+ ``abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ``
- And our shift is "2"
+ And our shift is ``2``
To decode the message, we would do the same thing as encoding, but in
- reverse. The first letter, "J" would become "H" (remember: we are decoding)
- because "H" is two letters in reverse (to the left) of "J". We would
- continue doing this. A letter like "a" would shift back to the end of
- the alphabet, and would become "Z" or "Y" and so on.
+ reverse. The first letter, ``J`` would become ``H`` (remember: we are decoding)
+ because ``H`` is two letters in reverse (to the left) of ``J``. We would
+ continue doing this. A letter like ``a`` would shift back to the end of
+ the alphabet, and would become ``Z`` or ``Y`` and so on.
- Our final message would be "Hello, captain"
+ Our final message would be ``Hello, captain``
Further reading
===============
+
* https://en.m.wikipedia.org/wiki/Caesar_cipher
Doctests
========
+
>>> decrypt('bpm yCqks jzwEv nwF rCuxA wDmz Bpm tiHG lwo', 8)
'The quick brown fox jumps over the lazy dog'
@@ -144,45 +160,48 @@ def decrypt(input_string: str, key: int, alphabet=None) -> str:
return encrypt(input_string, key, alphabet)
-def brute_force(input_string: str, alphabet=None) -> dict:
+def brute_force(input_string: str, alphabet: str | None = None) -> dict[int, str]:
"""
brute_force
===========
+
Returns all the possible combinations of keys and the decoded strings in the
form of a dictionary
Parameters:
-----------
- * input_string: the cipher-text that needs to be used during brute-force
+
+ * `input_string`: the cipher-text that needs to be used during brute-force
Optional:
- * alphabet: (None): the alphabet used to decode the cipher, if not
+
+ * `alphabet` (``None``): the alphabet used to decode the cipher, if not
specified, the standard english alphabet with upper and lowercase
letters is used
More about brute force
======================
+
Brute force is when a person intercepts a message or password, not knowing
the key and tries every single combination. This is easy with the caesar
cipher since there are only all the letters in the alphabet. The more
complex the cipher, the larger amount of time it will take to do brute force
Ex:
- Say we have a 5 letter alphabet (abcde), for simplicity and we intercepted the
- following message:
-
- "dbc"
-
+ Say we have a ``5`` letter alphabet (``abcde``), for simplicity and we intercepted
+ the following message: ``dbc``,
we could then just write out every combination:
- ecd... and so on, until we reach a combination that makes sense:
- "cab"
+ ``ecd``... and so on, until we reach a combination that makes sense:
+ ``cab``
Further reading
===============
+
* https://en.wikipedia.org/wiki/Brute_force
Doctests
========
+
>>> brute_force("jFyuMy xIH'N vLONy zILwy Gy!")[20]
"Please don't brute force me!"
@@ -193,33 +212,20 @@ def brute_force(input_string: str, alphabet=None) -> dict:
# Set default alphabet to lower and upper case english chars
alpha = alphabet or ascii_letters
- # The key during testing (will increase)
- key = 1
-
- # The encoded result
- result = ""
-
# To store data on all the combinations
brute_force_data = {}
# Cycle through each combination
- while key <= len(alpha):
- # Decrypt the message
- result = decrypt(input_string, key, alpha)
-
- # Update the data
- brute_force_data[key] = result
-
- # Reset result and increase the key
- result = ""
- key += 1
+ for key in range(1, len(alpha) + 1):
+ # Decrypt the message and store the result in the data
+ brute_force_data[key] = decrypt(input_string, key, alpha)
return brute_force_data
-def main():
+if __name__ == "__main__":
while True:
- print(f'\n{"-" * 10}\n Menu\n{"-" * 10}')
+ print(f"\n{'-' * 10}\n Menu\n{'-' * 10}")
print(*["1.Encrypt", "2.Decrypt", "3.BruteForce", "4.Quit"], sep="\n")
# get user input
@@ -248,7 +254,3 @@ def main():
elif choice == "4":
print("Goodbye.")
break
-
-
-if __name__ == "__main__":
- main()
diff --git a/ciphers/cryptomath_module.py b/ciphers/cryptomath_module.py
index ffeac1617f64..02e94e4b9e92 100644
--- a/ciphers/cryptomath_module.py
+++ b/ciphers/cryptomath_module.py
@@ -1,12 +1,10 @@
-def gcd(a: int, b: int) -> int:
- while a != 0:
- a, b = b % a, a
- return b
+from maths.greatest_common_divisor import gcd_by_iterative
-def findModInverse(a: int, m: int) -> int:
- if gcd(a, m) != 1:
- return None
+def find_mod_inverse(a: int, m: int) -> int:
+ if gcd_by_iterative(a, m) != 1:
+ msg = f"mod inverse of {a!r} and {m!r} does not exist"
+ raise ValueError(msg)
u1, u2, u3 = 1, 0, a
v1, v2, v3 = 0, 1, m
while v3 != 0:
diff --git a/ciphers/decrypt_caesar_with_chi_squared.py b/ciphers/decrypt_caesar_with_chi_squared.py
index 41b4a12ba453..fb95c0f90628 100644
--- a/ciphers/decrypt_caesar_with_chi_squared.py
+++ b/ciphers/decrypt_caesar_with_chi_squared.py
@@ -1,44 +1,41 @@
#!/usr/bin/env python3
-
-from typing import Tuple
+from __future__ import annotations
def decrypt_caesar_with_chi_squared(
ciphertext: str,
- cipher_alphabet: str = None,
- frequencies_dict: str = None,
- case_sensetive: bool = False,
-) -> Tuple[int, float, str]:
+ cipher_alphabet: list[str] | None = None,
+ frequencies_dict: dict[str, float] | None = None,
+ case_sensitive: bool = False,
+) -> tuple[int, float, str]:
"""
Basic Usage
===========
+
Arguments:
- * ciphertext (str): the text to decode (encoded with the caesar cipher)
+ * `ciphertext` (str): the text to decode (encoded with the caesar cipher)
Optional Arguments:
- * cipher_alphabet (list): the alphabet used for the cipher (each letter is
- a string separated by commas)
- * frequencies_dict (dict): a dictionary of word frequencies where keys are
- the letters and values are a percentage representation of the frequency as
- a decimal/float
- * case_sensetive (bool): a boolean value: True if the case matters during
- decryption, False if it doesn't
+ * `cipher_alphabet` (list): the alphabet used for the cipher (each letter is
+ a string separated by commas)
+ * `frequencies_dict` (dict): a dictionary of word frequencies where keys are
+ the letters and values are a percentage representation of the frequency as
+ a decimal/float
+ * `case_sensitive` (bool): a boolean value: ``True`` if the case matters during
+ decryption, ``False`` if it doesn't
Returns:
- * A tuple in the form of:
- (
- most_likely_cipher,
- most_likely_cipher_chi_squared_value,
- decoded_most_likely_cipher
- )
+ * A tuple in the form of:
+ (`most_likely_cipher`, `most_likely_cipher_chi_squared_value`,
+ `decoded_most_likely_cipher`)
- where...
- - most_likely_cipher is an integer representing the shift of the smallest
- chi-squared statistic (most likely key)
- - most_likely_cipher_chi_squared_value is a float representing the
- chi-squared statistic of the most likely shift
- - decoded_most_likely_cipher is a string with the decoded cipher
- (decoded by the most_likely_cipher key)
+ where...
+ - `most_likely_cipher` is an integer representing the shift of the smallest
+ chi-squared statistic (most likely key)
+ - `most_likely_cipher_chi_squared_value` is a float representing the
+ chi-squared statistic of the most likely shift
+ - `decoded_most_likely_cipher` is a string with the decoded cipher
+ (decoded by the most_likely_cipher key)
The Chi-squared test
@@ -46,52 +43,57 @@ def decrypt_caesar_with_chi_squared(
The caesar cipher
-----------------
+
The caesar cipher is a very insecure encryption algorithm, however it has
been used since Julius Caesar. The cipher is a simple substitution cipher
where each character in the plain text is replaced by a character in the
alphabet a certain number of characters after the original character. The
number of characters away is called the shift or key. For example:
- Plain text: hello
- Key: 1
- Cipher text: ifmmp
- (each letter in hello has been shifted one to the right in the eng. alphabet)
+ | Plain text: ``hello``
+ | Key: ``1``
+ | Cipher text: ``ifmmp``
+ | (each letter in ``hello`` has been shifted one to the right in the eng. alphabet)
As you can imagine, this doesn't provide lots of security. In fact
decrypting ciphertext by brute-force is extremely easy even by hand. However
- one way to do that is the chi-squared test.
+ one way to do that is the chi-squared test.
The chi-squared test
- -------------------
+ --------------------
+
Each letter in the english alphabet has a frequency, or the amount of times
it shows up compared to other letters (usually expressed as a decimal
representing the percentage likelihood). The most common letter in the
- english language is "e" with a frequency of 0.11162 or 11.162%. The test is
- completed in the following fashion.
+ english language is ``e`` with a frequency of ``0.11162`` or ``11.162%``.
+ The test is completed in the following fashion.
1. The ciphertext is decoded in a brute force way (every combination of the
- 26 possible combinations)
+ ``26`` possible combinations)
2. For every combination, for each letter in the combination, the average
amount of times the letter should appear the message is calculated by
- multiplying the total number of characters by the frequency of the letter
+ multiplying the total number of characters by the frequency of the letter.
- For example:
- In a message of 100 characters, e should appear around 11.162 times.
+ | For example:
+ | In a message of ``100`` characters, ``e`` should appear around ``11.162``
+ times.
- 3. Then, to calculate the margin of error (the amount of times the letter
- SHOULD appear with the amount of times the letter DOES appear), we use
- the chi-squared test. The following formula is used:
+ 3. Then, to calculate the margin of error (the amount of times the letter
+ SHOULD appear with the amount of times the letter DOES appear), we use
+ the chi-squared test. The following formula is used:
- Let:
- - n be the number of times the letter actually appears
- - p be the predicted value of the number of times the letter should
- appear (see #2)
- - let v be the chi-squared test result (referred to here as chi-squared
- value/statistic)
+ Let:
+ - n be the number of times the letter actually appears
+ - p be the predicted value of the number of times the letter should
+ appear (see item ``2``)
+ - let v be the chi-squared test result (referred to here as chi-squared
+ value/statistic)
- (n - p)^2
- --------- = v
- p
+ ::
+
+ (n - p)^2
+ --------- = v
+ p
4. Each chi squared value for each letter is then added up to the total.
The total is the chi-squared statistic for that encryption key.
@@ -99,16 +101,16 @@ def decrypt_caesar_with_chi_squared(
to be the decoded answer.
Further Reading
- ================
+ ===============
- * http://practicalcryptography.com/cryptanalysis/text-characterisation/chi-squared-
- statistic/
+ * http://practicalcryptography.com/cryptanalysis/text-characterisation/chi-squared-statistic/
* https://en.wikipedia.org/wiki/Letter_frequency
* https://en.wikipedia.org/wiki/Chi-squared_test
* https://en.m.wikipedia.org/wiki/Caesar_cipher
Doctests
========
+
>>> decrypt_caesar_with_chi_squared(
... 'dof pz aol jhlzhy jpwoly zv wvwbshy? pa pz avv lhzf av jyhjr!'
... ) # doctest: +NORMALIZE_WHITESPACE
@@ -118,14 +120,17 @@ def decrypt_caesar_with_chi_squared(
>>> decrypt_caesar_with_chi_squared('crybd cdbsxq')
(10, 233.35343938980898, 'short string')
+ >>> decrypt_caesar_with_chi_squared('Crybd Cdbsxq', case_sensitive=True)
+ (10, 233.35343938980898, 'Short String')
+
>>> decrypt_caesar_with_chi_squared(12)
Traceback (most recent call last):
AttributeError: 'int' object has no attribute 'lower'
"""
alphabet_letters = cipher_alphabet or [chr(i) for i in range(97, 123)]
- frequencies_dict = frequencies_dict or {}
- if frequencies_dict == {}:
+ # If the argument is None or the user provided an empty dictionary
+ if not frequencies_dict:
# Frequencies of letters in the english language (how much they show up)
frequencies = {
"a": 0.08497,
@@ -159,11 +164,11 @@ def decrypt_caesar_with_chi_squared(
# Custom frequencies dictionary
frequencies = frequencies_dict
- if not case_sensetive:
+ if not case_sensitive:
ciphertext = ciphertext.lower()
# Chi squared statistic values
- chi_squared_statistic_values = {}
+ chi_squared_statistic_values: dict[int, tuple[float, str]] = {}
# cycle through all of the shifts
for shift in range(len(alphabet_letters)):
@@ -173,10 +178,14 @@ def decrypt_caesar_with_chi_squared(
for letter in ciphertext:
try:
# Try to index the letter in the alphabet
- new_key = (alphabet_letters.index(letter) - shift) % len(
+ new_key = (alphabet_letters.index(letter.lower()) - shift) % len(
alphabet_letters
)
- decrypted_with_shift += alphabet_letters[new_key]
+ decrypted_with_shift += (
+ alphabet_letters[new_key].upper()
+ if case_sensitive and letter.isupper()
+ else alphabet_letters[new_key]
+ )
except ValueError:
# Append the character if it isn't in the alphabet
decrypted_with_shift += letter
@@ -185,10 +194,11 @@ def decrypt_caesar_with_chi_squared(
# Loop through each letter in the decoded message with the shift
for letter in decrypted_with_shift:
- if case_sensetive:
+ if case_sensitive:
+ letter = letter.lower()
if letter in frequencies:
# Get the amount of times the letter occurs in the message
- occurrences = decrypted_with_shift.count(letter)
+ occurrences = decrypted_with_shift.lower().count(letter)
# Get the excepcted amount of times the letter should appear based
# on letter frequencies
@@ -199,38 +209,41 @@ def decrypt_caesar_with_chi_squared(
# Add the margin of error to the total chi squared statistic
chi_squared_statistic += chi_letter_value
- else:
- if letter.lower() in frequencies:
- # Get the amount of times the letter occurs in the message
- occurrences = decrypted_with_shift.count(letter)
+ elif letter.lower() in frequencies:
+ # Get the amount of times the letter occurs in the message
+ occurrences = decrypted_with_shift.count(letter)
- # Get the excepcted amount of times the letter should appear based
- # on letter frequencies
- expected = frequencies[letter] * occurrences
+ # Get the excepcted amount of times the letter should appear based
+ # on letter frequencies
+ expected = frequencies[letter] * occurrences
- # Complete the chi squared statistic formula
- chi_letter_value = ((occurrences - expected) ** 2) / expected
+ # Complete the chi squared statistic formula
+ chi_letter_value = ((occurrences - expected) ** 2) / expected
- # Add the margin of error to the total chi squared statistic
- chi_squared_statistic += chi_letter_value
+ # Add the margin of error to the total chi squared statistic
+ chi_squared_statistic += chi_letter_value
# Add the data to the chi_squared_statistic_values dictionary
- chi_squared_statistic_values[shift] = [
+ chi_squared_statistic_values[shift] = (
chi_squared_statistic,
decrypted_with_shift,
- ]
+ )
# Get the most likely cipher by finding the cipher with the smallest chi squared
# statistic
- most_likely_cipher = min(
- chi_squared_statistic_values, key=chi_squared_statistic_values.get
+ def chi_squared_statistic_values_sorting_key(key: int) -> tuple[float, str]:
+ return chi_squared_statistic_values[key]
+
+ most_likely_cipher: int = min(
+ chi_squared_statistic_values,
+ key=chi_squared_statistic_values_sorting_key,
)
# Get all the data from the most likely cipher (key, decoded message)
- most_likely_cipher_chi_squared_value = chi_squared_statistic_values[
- most_likely_cipher
- ][0]
- decoded_most_likely_cipher = chi_squared_statistic_values[most_likely_cipher][1]
+ (
+ most_likely_cipher_chi_squared_value,
+ decoded_most_likely_cipher,
+ ) = chi_squared_statistic_values[most_likely_cipher]
# Return the data on the most likely shift
return (
diff --git a/ciphers/deterministic_miller_rabin.py b/ciphers/deterministic_miller_rabin.py
index d7fcb67e936c..2191caf630a7 100644
--- a/ciphers/deterministic_miller_rabin.py
+++ b/ciphers/deterministic_miller_rabin.py
@@ -73,7 +73,7 @@ def miller_rabin(n: int, allow_probable: bool = False) -> bool:
for prime in plist:
pr = False
for r in range(s):
- m = pow(prime, d * 2 ** r, n)
+ m = pow(prime, d * 2**r, n)
# see article for analysis explanation for m
if (r == 0 and m == 1) or ((m + 1) % n == 0):
pr = True
diff --git a/ciphers/diffie.py b/ciphers/diffie.py
index 44b12bf9d103..1e1e868999b6 100644
--- a/ciphers/diffie.py
+++ b/ciphers/diffie.py
@@ -1,25 +1,53 @@
-def find_primitive(n: int) -> int:
- for r in range(1, n):
+from __future__ import annotations
+
+
+def find_primitive(modulus: int) -> int | None:
+ """
+ Find a primitive root modulo modulus, if one exists.
+
+ Args:
+ modulus : The modulus for which to find a primitive root.
+
+ Returns:
+ The primitive root if one exists, or None if there is none.
+
+ Examples:
+ >>> find_primitive(7) # Modulo 7 has primitive root 3
+ 3
+ >>> find_primitive(11) # Modulo 11 has primitive root 2
+ 2
+ >>> find_primitive(8) == None # Modulo 8 has no primitive root
+ True
+ """
+ for r in range(1, modulus):
li = []
- for x in range(n - 1):
- val = pow(r, x, n)
+ for x in range(modulus - 1):
+ val = pow(r, x, modulus)
if val in li:
break
li.append(val)
else:
return r
+ return None
if __name__ == "__main__":
- q = int(input("Enter a prime number q: "))
- a = find_primitive(q)
- a_private = int(input("Enter private key of A: "))
- a_public = pow(a, a_private, q)
- b_private = int(input("Enter private key of B: "))
- b_public = pow(a, b_private, q)
-
- a_secret = pow(b_public, a_private, q)
- b_secret = pow(a_public, b_private, q)
-
- print("The key value generated by A is: ", a_secret)
- print("The key value generated by B is: ", b_secret)
+ import doctest
+
+ doctest.testmod()
+
+ prime = int(input("Enter a prime number q: "))
+ primitive_root = find_primitive(prime)
+ if primitive_root is None:
+ print(f"Cannot find the primitive for the value: {primitive_root!r}")
+ else:
+ a_private = int(input("Enter private key of A: "))
+ a_public = pow(primitive_root, a_private, prime)
+ b_private = int(input("Enter private key of B: "))
+ b_public = pow(primitive_root, b_private, prime)
+
+ a_secret = pow(b_public, a_private, prime)
+ b_secret = pow(a_public, b_private, prime)
+
+ print("The key value generated by A is: ", a_secret)
+ print("The key value generated by B is: ", b_secret)
diff --git a/ciphers/diffie_hellman.py b/ciphers/diffie_hellman.py
new file mode 100644
index 000000000000..aec7fb3eaf17
--- /dev/null
+++ b/ciphers/diffie_hellman.py
@@ -0,0 +1,267 @@
+from binascii import hexlify
+from hashlib import sha256
+from os import urandom
+
+# RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for
+# Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526
+
+primes = {
+ # 1536-bit
+ 5: {
+ "prime": int(
+ "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1"
+ "29024E088A67CC74020BBEA63B139B22514A08798E3404DD"
+ "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245"
+ "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED"
+ "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D"
+ "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F"
+ "83655D23DCA3AD961C62F356208552BB9ED529077096966D"
+ "670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF",
+ base=16,
+ ),
+ "generator": 2,
+ },
+ # 2048-bit
+ 14: {
+ "prime": int(
+ "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1"
+ "29024E088A67CC74020BBEA63B139B22514A08798E3404DD"
+ "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245"
+ "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED"
+ "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D"
+ "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F"
+ "83655D23DCA3AD961C62F356208552BB9ED529077096966D"
+ "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B"
+ "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9"
+ "DE2BCBF6955817183995497CEA956AE515D2261898FA0510"
+ "15728E5A8AACAA68FFFFFFFFFFFFFFFF",
+ base=16,
+ ),
+ "generator": 2,
+ },
+ # 3072-bit
+ 15: {
+ "prime": int(
+ "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1"
+ "29024E088A67CC74020BBEA63B139B22514A08798E3404DD"
+ "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245"
+ "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED"
+ "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D"
+ "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F"
+ "83655D23DCA3AD961C62F356208552BB9ED529077096966D"
+ "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B"
+ "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9"
+ "DE2BCBF6955817183995497CEA956AE515D2261898FA0510"
+ "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64"
+ "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7"
+ "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B"
+ "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C"
+ "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31"
+ "43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF",
+ base=16,
+ ),
+ "generator": 2,
+ },
+ # 4096-bit
+ 16: {
+ "prime": int(
+ "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1"
+ "29024E088A67CC74020BBEA63B139B22514A08798E3404DD"
+ "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245"
+ "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED"
+ "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D"
+ "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F"
+ "83655D23DCA3AD961C62F356208552BB9ED529077096966D"
+ "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B"
+ "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9"
+ "DE2BCBF6955817183995497CEA956AE515D2261898FA0510"
+ "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64"
+ "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7"
+ "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B"
+ "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C"
+ "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31"
+ "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7"
+ "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA"
+ "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6"
+ "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED"
+ "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9"
+ "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199"
+ "FFFFFFFFFFFFFFFF",
+ base=16,
+ ),
+ "generator": 2,
+ },
+ # 6144-bit
+ 17: {
+ "prime": int(
+ "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08"
+ "8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B"
+ "302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9"
+ "A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6"
+ "49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8"
+ "FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D"
+ "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C"
+ "180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718"
+ "3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D"
+ "04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D"
+ "B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226"
+ "1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C"
+ "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC"
+ "E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26"
+ "99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB"
+ "04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2"
+ "233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127"
+ "D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492"
+ "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406"
+ "AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918"
+ "DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151"
+ "2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03"
+ "F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F"
+ "BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA"
+ "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B"
+ "B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632"
+ "387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E"
+ "6DCC4024FFFFFFFFFFFFFFFF",
+ base=16,
+ ),
+ "generator": 2,
+ },
+ # 8192-bit
+ 18: {
+ "prime": int(
+ "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1"
+ "29024E088A67CC74020BBEA63B139B22514A08798E3404DD"
+ "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245"
+ "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED"
+ "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D"
+ "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F"
+ "83655D23DCA3AD961C62F356208552BB9ED529077096966D"
+ "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B"
+ "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9"
+ "DE2BCBF6955817183995497CEA956AE515D2261898FA0510"
+ "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64"
+ "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7"
+ "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B"
+ "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C"
+ "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31"
+ "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7"
+ "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA"
+ "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6"
+ "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED"
+ "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9"
+ "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492"
+ "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD"
+ "F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831"
+ "179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B"
+ "DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF"
+ "5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6"
+ "D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3"
+ "23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA"
+ "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328"
+ "06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C"
+ "DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE"
+ "12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4"
+ "38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300"
+ "741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568"
+ "3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9"
+ "22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B"
+ "4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A"
+ "062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36"
+ "4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1"
+ "B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92"
+ "4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47"
+ "9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71"
+ "60C980DD98EDD3DFFFFFFFFFFFFFFFFF",
+ base=16,
+ ),
+ "generator": 2,
+ },
+}
+
+
+class DiffieHellman:
+ """
+ Class to represent the Diffie-Hellman key exchange protocol
+
+
+ >>> alice = DiffieHellman()
+ >>> bob = DiffieHellman()
+
+ >>> alice_private = alice.get_private_key()
+ >>> alice_public = alice.generate_public_key()
+
+ >>> bob_private = bob.get_private_key()
+ >>> bob_public = bob.generate_public_key()
+
+ >>> # generating shared key using the DH object
+ >>> alice_shared = alice.generate_shared_key(bob_public)
+ >>> bob_shared = bob.generate_shared_key(alice_public)
+
+ >>> assert alice_shared == bob_shared
+
+ >>> # generating shared key using static methods
+ >>> alice_shared = DiffieHellman.generate_shared_key_static(
+ ... alice_private, bob_public
+ ... )
+ >>> bob_shared = DiffieHellman.generate_shared_key_static(
+ ... bob_private, alice_public
+ ... )
+
+ >>> assert alice_shared == bob_shared
+ """
+
+ # Current minimum recommendation is 2048 bit (group 14)
+ def __init__(self, group: int = 14) -> None:
+ if group not in primes:
+ raise ValueError("Unsupported Group")
+ self.prime = primes[group]["prime"]
+ self.generator = primes[group]["generator"]
+
+ self.__private_key = int(hexlify(urandom(32)), base=16)
+
+ def get_private_key(self) -> str:
+ return hex(self.__private_key)[2:]
+
+ def generate_public_key(self) -> str:
+ public_key = pow(self.generator, self.__private_key, self.prime)
+ return hex(public_key)[2:]
+
+ def is_valid_public_key(self, key: int) -> bool:
+ # check if the other public key is valid based on NIST SP800-56
+ return (
+ 2 <= key <= self.prime - 2
+ and pow(key, (self.prime - 1) // 2, self.prime) == 1
+ )
+
+ def generate_shared_key(self, other_key_str: str) -> str:
+ other_key = int(other_key_str, base=16)
+ if not self.is_valid_public_key(other_key):
+ raise ValueError("Invalid public key")
+ shared_key = pow(other_key, self.__private_key, self.prime)
+ return sha256(str(shared_key).encode()).hexdigest()
+
+ @staticmethod
+ def is_valid_public_key_static(remote_public_key_str: int, prime: int) -> bool:
+ # check if the other public key is valid based on NIST SP800-56
+ return (
+ 2 <= remote_public_key_str <= prime - 2
+ and pow(remote_public_key_str, (prime - 1) // 2, prime) == 1
+ )
+
+ @staticmethod
+ def generate_shared_key_static(
+ local_private_key_str: str, remote_public_key_str: str, group: int = 14
+ ) -> str:
+ local_private_key = int(local_private_key_str, base=16)
+ remote_public_key = int(remote_public_key_str, base=16)
+ prime = primes[group]["prime"]
+ if not DiffieHellman.is_valid_public_key_static(remote_public_key, prime):
+ raise ValueError("Invalid public key")
+ shared_key = pow(remote_public_key, local_private_key, prime)
+ return sha256(str(shared_key).encode()).hexdigest()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/ciphers/elgamal_key_generator.py b/ciphers/elgamal_key_generator.py
index 52cf69074187..17ba55c0d013 100644
--- a/ciphers/elgamal_key_generator.py
+++ b/ciphers/elgamal_key_generator.py
@@ -2,24 +2,18 @@
import random
import sys
-from . import cryptomath_module as cryptoMath
-from . import rabin_miller as rabinMiller
+from . import cryptomath_module as cryptomath
+from . import rabin_miller
min_primitive_root = 3
-def main():
- print("Making key files...")
- makeKeyFiles("elgamal", 2048)
- print("Key files generation successful")
-
-
# I have written my code naively same as definition of primitive root
# however every time I run this program, memory exceeded...
# so I used 4.80 Algorithm in
# Handbook of Applied Cryptography(CRC Press, ISBN : 0-8493-8523-7, October 1996)
# and it seems to run nicely!
-def primitiveRoot(p_val: int) -> int:
+def primitive_root(p_val: int) -> int:
print("Generating primitive root of p")
while True:
g = random.randrange(3, p_val)
@@ -30,41 +24,42 @@ def primitiveRoot(p_val: int) -> int:
return g
-def generateKey(keySize: int) -> ((int, int, int, int), (int, int)):
+def generate_key(key_size: int) -> tuple[tuple[int, int, int, int], tuple[int, int]]:
print("Generating prime p...")
- p = rabinMiller.generateLargePrime(keySize) # select large prime number.
- e_1 = primitiveRoot(p) # one primitive root on modulo p.
+ p = rabin_miller.generate_large_prime(key_size) # select large prime number.
+ e_1 = primitive_root(p) # one primitive root on modulo p.
d = random.randrange(3, p) # private_key -> have to be greater than 2 for safety.
- e_2 = cryptoMath.findModInverse(pow(e_1, d, p), p)
+ e_2 = cryptomath.find_mod_inverse(pow(e_1, d, p), p)
- publicKey = (keySize, e_1, e_2, p)
- privateKey = (keySize, d)
+ public_key = (key_size, e_1, e_2, p)
+ private_key = (key_size, d)
- return publicKey, privateKey
+ return public_key, private_key
-def makeKeyFiles(name: str, keySize: int):
- if os.path.exists("%s_pubkey.txt" % name) or os.path.exists(
- "%s_privkey.txt" % name
- ):
+def make_key_files(name: str, key_size: int) -> None:
+ if os.path.exists(f"{name}_pubkey.txt") or os.path.exists(f"{name}_privkey.txt"):
print("\nWARNING:")
print(
- '"%s_pubkey.txt" or "%s_privkey.txt" already exists. \n'
+ f'"{name}_pubkey.txt" or "{name}_privkey.txt" already exists. \n'
"Use a different name or delete these files and re-run this program."
- % (name, name)
)
sys.exit()
- publicKey, privateKey = generateKey(keySize)
- print("\nWriting public key to file %s_pubkey.txt..." % name)
- with open("%s_pubkey.txt" % name, "w") as fo:
- fo.write(
- "%d,%d,%d,%d" % (publicKey[0], publicKey[1], publicKey[2], publicKey[3])
- )
+ public_key, private_key = generate_key(key_size)
+ print(f"\nWriting public key to file {name}_pubkey.txt...")
+ with open(f"{name}_pubkey.txt", "w") as fo:
+ fo.write(f"{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}")
- print("Writing private key to file %s_privkey.txt..." % name)
- with open("%s_privkey.txt" % name, "w") as fo:
- fo.write("%d,%d" % (privateKey[0], privateKey[1]))
+ print(f"Writing private key to file {name}_privkey.txt...")
+ with open(f"{name}_privkey.txt", "w") as fo:
+ fo.write(f"{private_key[0]},{private_key[1]}")
+
+
+def main() -> None:
+ print("Making key files...")
+ make_key_files("elgamal", 2048)
+ print("Key files generation successful")
if __name__ == "__main__":
diff --git a/ciphers/enigma_machine2.py b/ciphers/enigma_machine2.py
index 4344db0056fd..e42fdd82ed41 100644
--- a/ciphers/enigma_machine2.py
+++ b/ciphers/enigma_machine2.py
@@ -1,20 +1,28 @@
"""
-Wikipedia: https://en.wikipedia.org/wiki/Enigma_machine
-Video explanation: https://youtu.be/QwQVMqfoB2E
-Also check out Numberphile's and Computerphile's videos on this topic
+| Wikipedia: https://en.wikipedia.org/wiki/Enigma_machine
+| Video explanation: https://youtu.be/QwQVMqfoB2E
+| Also check out Numberphile's and Computerphile's videos on this topic
-This module contains function 'enigma' which emulates
+This module contains function ``enigma`` which emulates
the famous Enigma machine from WWII.
+
Module includes:
-- enigma function
+
+- ``enigma`` function
- showcase of function usage
-- 9 randnomly generated rotors
+- ``9`` randomly generated rotors
- reflector (aka static rotor)
- original alphabet
Created by TrapinchO
"""
+from __future__ import annotations
+
+RotorPositionT = tuple[int, int, int]
+RotorSelectionT = tuple[str, str, str]
+
+
# used alphabet --------------------------
# from string.ascii_uppercase
abc = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
@@ -63,9 +71,11 @@
rotor9 = "KOAEGVDHXPQZMLFTYWJNBRCIUS"
-def _validator(rotpos: tuple, rotsel: tuple, pb: str) -> tuple:
+def _validator(
+ rotpos: RotorPositionT, rotsel: RotorSelectionT, pb: str
+) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]:
"""
- Checks if the values can be used for the 'enigma' function
+ Checks if the values can be used for the ``enigma`` function
>>> _validator((1,1,1), (rotor1, rotor2, rotor3), 'POLAND')
((1, 1, 1), ('EGZWVONAHDCLFQMSIPJBYUKXTR', 'FOBHMDKEXQNRAULPGSJVTYICZW', \
@@ -75,36 +85,33 @@ def _validator(rotpos: tuple, rotsel: tuple, pb: str) -> tuple:
:param rotpos: rotor_positon
:param rotsel: rotor_selection
:param pb: plugb -> validated and transformed
- :return: (rotpos, rotsel, pb)
+ :return: (`rotpos`, `rotsel`, `pb`)
"""
# Checks if there are 3 unique rotors
- unique_rotsel = len(set(rotsel))
- if unique_rotsel < 3:
- raise Exception(f"Please use 3 unique rotors (not {unique_rotsel})")
+ if (unique_rotsel := len(set(rotsel))) < 3:
+ msg = f"Please use 3 unique rotors (not {unique_rotsel})"
+ raise Exception(msg)
# Checks if rotor positions are valid
rotorpos1, rotorpos2, rotorpos3 = rotpos
if not 0 < rotorpos1 <= len(abc):
- raise ValueError(
- f"First rotor position is not within range of 1..26 (" f"{rotorpos1}"
- )
+ msg = f"First rotor position is not within range of 1..26 ({rotorpos1}"
+ raise ValueError(msg)
if not 0 < rotorpos2 <= len(abc):
- raise ValueError(
- f"Second rotor position is not within range of 1..26 (" f"{rotorpos2})"
- )
+ msg = f"Second rotor position is not within range of 1..26 ({rotorpos2})"
+ raise ValueError(msg)
if not 0 < rotorpos3 <= len(abc):
- raise ValueError(
- f"Third rotor position is not within range of 1..26 (" f"{rotorpos3})"
- )
+ msg = f"Third rotor position is not within range of 1..26 ({rotorpos3})"
+ raise ValueError(msg)
# Validates string and returns dict
- pb = _plugboard(pb)
+ pbdict = _plugboard(pb)
- return rotpos, rotsel, pb
+ return rotpos, rotsel, pbdict
-def _plugboard(pbstring: str) -> dict:
+def _plugboard(pbstring: str) -> dict[str, str]:
"""
https://en.wikipedia.org/wiki/Enigma_machine#Plugboard
@@ -113,9 +120,10 @@ def _plugboard(pbstring: str) -> dict:
>>> _plugboard('POLAND')
{'P': 'O', 'O': 'P', 'L': 'A', 'A': 'L', 'N': 'D', 'D': 'N'}
- In the code, 'pb' stands for 'plugboard'
+ In the code, ``pb`` stands for ``plugboard``
Pairs can be separated by spaces
+
:param pbstring: string containing plugboard setting for the Enigma machine
:return: dictionary containing converted pairs
"""
@@ -124,9 +132,11 @@ def _plugboard(pbstring: str) -> dict:
# a) is type string
# b) has even length (so pairs can be made)
if not isinstance(pbstring, str):
- raise TypeError(f"Plugboard setting isn't type string ({type(pbstring)})")
+ msg = f"Plugboard setting isn't type string ({type(pbstring)})"
+ raise TypeError(msg)
elif len(pbstring) % 2 != 0:
- raise Exception(f"Odd number of symbols ({len(pbstring)})")
+ msg = f"Odd number of symbols ({len(pbstring)})"
+ raise Exception(msg)
elif pbstring == "":
return {}
@@ -136,54 +146,59 @@ def _plugboard(pbstring: str) -> dict:
tmppbl = set()
for i in pbstring:
if i not in abc:
- raise Exception(f"'{i}' not in list of symbols")
+ msg = f"'{i}' not in list of symbols"
+ raise Exception(msg)
elif i in tmppbl:
- raise Exception(f"Duplicate symbol ({i})")
+ msg = f"Duplicate symbol ({i})"
+ raise Exception(msg)
else:
tmppbl.add(i)
del tmppbl
# Created the dictionary
pb = {}
- for i in range(0, len(pbstring) - 1, 2):
- pb[pbstring[i]] = pbstring[i + 1]
- pb[pbstring[i + 1]] = pbstring[i]
+ for j in range(0, len(pbstring) - 1, 2):
+ pb[pbstring[j]] = pbstring[j + 1]
+ pb[pbstring[j + 1]] = pbstring[j]
return pb
def enigma(
text: str,
- rotor_position: tuple,
- rotor_selection: tuple = (rotor1, rotor2, rotor3),
+ rotor_position: RotorPositionT,
+ rotor_selection: RotorSelectionT = (rotor1, rotor2, rotor3),
plugb: str = "",
) -> str:
"""
- The only difference with real-world enigma is that I allowed string input.
+ The only difference with real-world enigma is that ``I`` allowed string input.
All characters are converted to uppercase. (non-letter symbol are ignored)
- How it works:
- (for every letter in the message)
+
+ | How it works:
+ | (for every letter in the message)
- Input letter goes into the plugboard.
- If it is connected to another one, switch it.
+ If it is connected to another one, switch it.
+
+ - Letter goes through ``3`` rotors.
+ Each rotor can be represented as ``2`` sets of symbol, where one is shuffled.
+ Each symbol from the first set has corresponding symbol in
+ the second set and vice versa.
- - Letter goes through 3 rotors.
- Each rotor can be represented as 2 sets of symbol, where one is shuffled.
- Each symbol from the first set has corresponding symbol in
- the second set and vice versa.
+ example::
- example:
- | ABCDEFGHIJKLMNOPQRSTUVWXYZ | e.g. F=D and D=F
- | VKLEPDBGRNWTFCJOHQAMUZYIXS |
+ | ABCDEFGHIJKLMNOPQRSTUVWXYZ | e.g. F=D and D=F
+ | VKLEPDBGRNWTFCJOHQAMUZYIXS |
- Symbol then goes through reflector (static rotor).
- There it is switched with paired symbol
- The reflector can be represented as2 sets, each with half of the alphanet.
- There are usually 10 pairs of letters.
+ There it is switched with paired symbol.
+ The reflector can be represented as ``2`` sets, each with half of the alphanet.
+ There are usually ``10`` pairs of letters.
- Example:
- | ABCDEFGHIJKLM | e.g. E is paired to X
- | ZYXWVUTSRQPON | so when E goes in X goes out and vice versa
+ Example::
+
+ | ABCDEFGHIJKLM | e.g. E is paired to X
+ | ZYXWVUTSRQPON | so when E goes in X goes out and vice versa
- Letter then goes through the rotors again
@@ -202,9 +217,9 @@ def enigma(
:param text: input message
- :param rotor_position: tuple with 3 values in range 1..26
- :param rotor_selection: tuple with 3 rotors ()
- :param plugb: string containing plugboard configuration (default '')
+ :param rotor_position: tuple with ``3`` values in range ``1``.. ``26``
+ :param rotor_selection: tuple with ``3`` rotors
+ :param plugb: string containing plugboard configuration (default ``''``)
:return: en/decrypted string
"""
@@ -224,7 +239,6 @@ def enigma(
# encryption/decryption process --------------------------
for symbol in text:
if symbol in abc:
-
# 1st plugboard --------------------------
if symbol in plugboard:
symbol = plugboard[symbol]
diff --git a/ciphers/fractionated_morse_cipher.py b/ciphers/fractionated_morse_cipher.py
new file mode 100644
index 000000000000..6c4c415abac1
--- /dev/null
+++ b/ciphers/fractionated_morse_cipher.py
@@ -0,0 +1,168 @@
+"""
+Python program for the Fractionated Morse Cipher.
+
+The Fractionated Morse cipher first converts the plaintext to Morse code,
+then enciphers fixed-size blocks of Morse code back to letters.
+This procedure means plaintext letters are mixed into the ciphertext letters,
+making it more secure than substitution ciphers.
+
+http://practicalcryptography.com/ciphers/fractionated-morse-cipher/
+"""
+
+import string
+
+MORSE_CODE_DICT = {
+ "A": ".-",
+ "B": "-...",
+ "C": "-.-.",
+ "D": "-..",
+ "E": ".",
+ "F": "..-.",
+ "G": "--.",
+ "H": "....",
+ "I": "..",
+ "J": ".---",
+ "K": "-.-",
+ "L": ".-..",
+ "M": "--",
+ "N": "-.",
+ "O": "---",
+ "P": ".--.",
+ "Q": "--.-",
+ "R": ".-.",
+ "S": "...",
+ "T": "-",
+ "U": "..-",
+ "V": "...-",
+ "W": ".--",
+ "X": "-..-",
+ "Y": "-.--",
+ "Z": "--..",
+ " ": "",
+}
+
+# Define possible trigrams of Morse code
+MORSE_COMBINATIONS = [
+ "...",
+ "..-",
+ "..x",
+ ".-.",
+ ".--",
+ ".-x",
+ ".x.",
+ ".x-",
+ ".xx",
+ "-..",
+ "-.-",
+ "-.x",
+ "--.",
+ "---",
+ "--x",
+ "-x.",
+ "-x-",
+ "-xx",
+ "x..",
+ "x.-",
+ "x.x",
+ "x-.",
+ "x--",
+ "x-x",
+ "xx.",
+ "xx-",
+ "xxx",
+]
+
+# Create a reverse dictionary for Morse code
+REVERSE_DICT = {value: key for key, value in MORSE_CODE_DICT.items()}
+
+
+def encode_to_morse(plaintext: str) -> str:
+ """Encode a plaintext message into Morse code.
+
+ Args:
+ plaintext: The plaintext message to encode.
+
+ Returns:
+ The Morse code representation of the plaintext message.
+
+ Example:
+ >>> encode_to_morse("defend the east")
+ '-..x.x..-.x.x-.x-..xx-x....x.xx.x.-x...x-'
+ """
+ return "x".join([MORSE_CODE_DICT.get(letter.upper(), "") for letter in plaintext])
+
+
+def encrypt_fractionated_morse(plaintext: str, key: str) -> str:
+ """Encrypt a plaintext message using Fractionated Morse Cipher.
+
+ Args:
+ plaintext: The plaintext message to encrypt.
+ key: The encryption key.
+
+ Returns:
+ The encrypted ciphertext.
+
+ Example:
+ >>> encrypt_fractionated_morse("defend the east","Roundtable")
+ 'ESOAVVLJRSSTRX'
+
+ """
+ morse_code = encode_to_morse(plaintext)
+ key = key.upper() + string.ascii_uppercase
+ key = "".join(sorted(set(key), key=key.find))
+
+ # Ensure morse_code length is a multiple of 3
+ padding_length = 3 - (len(morse_code) % 3)
+ morse_code += "x" * padding_length
+
+ fractionated_morse_dict = {v: k for k, v in zip(key, MORSE_COMBINATIONS)}
+ fractionated_morse_dict["xxx"] = ""
+ encrypted_text = "".join(
+ [
+ fractionated_morse_dict[morse_code[i : i + 3]]
+ for i in range(0, len(morse_code), 3)
+ ]
+ )
+ return encrypted_text
+
+
+def decrypt_fractionated_morse(ciphertext: str, key: str) -> str:
+ """Decrypt a ciphertext message encrypted with Fractionated Morse Cipher.
+
+ Args:
+ ciphertext: The ciphertext message to decrypt.
+ key: The decryption key.
+
+ Returns:
+ The decrypted plaintext message.
+
+ Example:
+ >>> decrypt_fractionated_morse("ESOAVVLJRSSTRX","Roundtable")
+ 'DEFEND THE EAST'
+ """
+ key = key.upper() + string.ascii_uppercase
+ key = "".join(sorted(set(key), key=key.find))
+
+ inverse_fractionated_morse_dict = dict(zip(key, MORSE_COMBINATIONS))
+ morse_code = "".join(
+ [inverse_fractionated_morse_dict.get(letter, "") for letter in ciphertext]
+ )
+ decrypted_text = "".join(
+ [REVERSE_DICT[code] for code in morse_code.split("x")]
+ ).strip()
+ return decrypted_text
+
+
+if __name__ == "__main__":
+ """
+ Example usage of Fractionated Morse Cipher.
+ """
+ plaintext = "defend the east"
+ print("Plain Text:", plaintext)
+ key = "ROUNDTABLE"
+
+ ciphertext = encrypt_fractionated_morse(plaintext, key)
+ print("Encrypted:", ciphertext)
+
+ decrypted_text = decrypt_fractionated_morse(ciphertext, key)
+ print("Decrypted:", decrypted_text)
diff --git a/ciphers/gronsfeld_cipher.py b/ciphers/gronsfeld_cipher.py
new file mode 100644
index 000000000000..8fbeab4307fc
--- /dev/null
+++ b/ciphers/gronsfeld_cipher.py
@@ -0,0 +1,45 @@
+from string import ascii_uppercase
+
+
+def gronsfeld(text: str, key: str) -> str:
+ """
+ Encrypt plaintext with the Gronsfeld cipher
+
+ >>> gronsfeld('hello', '412')
+ 'LFNPP'
+ >>> gronsfeld('hello', '123')
+ 'IGOMQ'
+ >>> gronsfeld('', '123')
+ ''
+ >>> gronsfeld('yes, ¥€$ - _!@#%?', '0')
+ 'YES, ¥€$ - _!@#%?'
+ >>> gronsfeld('yes, ¥€$ - _!@#%?', '01')
+ 'YFS, ¥€$ - _!@#%?'
+ >>> gronsfeld('yes, ¥€$ - _!@#%?', '012')
+ 'YFU, ¥€$ - _!@#%?'
+ >>> gronsfeld('yes, ¥€$ - _!@#%?', '')
+ Traceback (most recent call last):
+ ...
+ ZeroDivisionError: integer modulo by zero
+ """
+ ascii_len = len(ascii_uppercase)
+ key_len = len(key)
+ encrypted_text = ""
+ keys = [int(char) for char in key]
+ upper_case_text = text.upper()
+
+ for i, char in enumerate(upper_case_text):
+ if char in ascii_uppercase:
+ new_position = (ascii_uppercase.index(char) + keys[i % key_len]) % ascii_len
+ shifted_letter = ascii_uppercase[new_position]
+ encrypted_text += shifted_letter
+ else:
+ encrypted_text += char
+
+ return encrypted_text
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/ciphers/hill_cipher.py b/ciphers/hill_cipher.py
index 8237abf6aa5d..33b2529f017b 100644
--- a/ciphers/hill_cipher.py
+++ b/ciphers/hill_cipher.py
@@ -35,23 +35,12 @@
https://www.youtube.com/watch?v=4RhLNDqcjpA
"""
-import string
-import numpy
+import string
+import numpy as np
-def greatest_common_divisor(a: int, b: int) -> int:
- """
- >>> greatest_common_divisor(4, 8)
- 4
- >>> greatest_common_divisor(8, 4)
- 4
- >>> greatest_common_divisor(4, 7)
- 1
- >>> greatest_common_divisor(0, 10)
- 10
- """
- return b if a == 0 else greatest_common_divisor(b % a, a)
+from maths.greatest_common_divisor import greatest_common_divisor
class HillCipher:
@@ -60,22 +49,21 @@ class HillCipher:
# i.e. a total of 36 characters
# take x and return x % len(key_string)
- modulus = numpy.vectorize(lambda x: x % 36)
+ modulus = np.vectorize(lambda x: x % 36)
- to_int = numpy.vectorize(lambda x: round(x))
+ to_int = np.vectorize(round)
- def __init__(self, encrypt_key: int):
+ def __init__(self, encrypt_key: np.ndarray) -> None:
"""
encrypt_key is an NxN numpy array
"""
self.encrypt_key = self.modulus(encrypt_key) # mod36 calc's on the encrypt key
self.check_determinant() # validate the determinant of the encryption key
- self.decrypt_key = None
self.break_key = encrypt_key.shape[0]
def replace_letters(self, letter: str) -> int:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.replace_letters('T')
19
>>> hill_cipher.replace_letters('0')
@@ -85,7 +73,7 @@ def replace_letters(self, letter: str) -> int:
def replace_digits(self, num: int) -> str:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.replace_digits(19)
'T'
>>> hill_cipher.replace_digits(26)
@@ -95,24 +83,25 @@ def replace_digits(self, num: int) -> str:
def check_determinant(self) -> None:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.check_determinant()
"""
- det = round(numpy.linalg.det(self.encrypt_key))
+ det = round(np.linalg.det(self.encrypt_key))
if det < 0:
det = det % len(self.key_string)
req_l = len(self.key_string)
if greatest_common_divisor(det, len(self.key_string)) != 1:
- raise ValueError(
- f"determinant modular {req_l} of encryption key({det}) is not co prime "
- f"w.r.t {req_l}.\nTry another key."
+ msg = (
+ f"determinant modular {req_l} of encryption key({det}) "
+ f"is not co prime w.r.t {req_l}.\nTry another key."
)
+ raise ValueError(msg)
def process_text(self, text: str) -> str:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.process_text('Testing Hill Cipher')
'TESTINGHILLCIPHERR'
>>> hill_cipher.process_text('hello')
@@ -128,7 +117,7 @@ def process_text(self, text: str) -> str:
def encrypt(self, text: str) -> str:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.encrypt('testing hill cipher')
'WHXYJOLM9C6XT085LL'
>>> hill_cipher.encrypt('hello')
@@ -139,8 +128,8 @@ def encrypt(self, text: str) -> str:
for i in range(0, len(text) - self.break_key + 1, self.break_key):
batch = text[i : i + self.break_key]
- batch_vec = [self.replace_letters(char) for char in batch]
- batch_vec = numpy.array([batch_vec]).T
+ vec = [self.replace_letters(char) for char in batch]
+ batch_vec = np.array([vec]).T
batch_encrypted = self.modulus(self.encrypt_key.dot(batch_vec)).T.tolist()[
0
]
@@ -151,14 +140,14 @@ def encrypt(self, text: str) -> str:
return encrypted
- def make_decrypt_key(self):
+ def make_decrypt_key(self) -> np.ndarray:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.make_decrypt_key()
array([[ 6, 25],
[ 5, 26]])
"""
- det = round(numpy.linalg.det(self.encrypt_key))
+ det = round(np.linalg.det(self.encrypt_key))
if det < 0:
det = det % len(self.key_string)
@@ -169,32 +158,28 @@ def make_decrypt_key(self):
break
inv_key = (
- det_inv
- * numpy.linalg.det(self.encrypt_key)
- * numpy.linalg.inv(self.encrypt_key)
+ det_inv * np.linalg.det(self.encrypt_key) * np.linalg.inv(self.encrypt_key)
)
return self.to_int(self.modulus(inv_key))
def decrypt(self, text: str) -> str:
"""
- >>> hill_cipher = HillCipher(numpy.array([[2, 5], [1, 6]]))
+ >>> hill_cipher = HillCipher(np.array([[2, 5], [1, 6]]))
>>> hill_cipher.decrypt('WHXYJOLM9C6XT085LL')
'TESTINGHILLCIPHERR'
>>> hill_cipher.decrypt('85FF00')
'HELLOO'
"""
- self.decrypt_key = self.make_decrypt_key()
+ decrypt_key = self.make_decrypt_key()
text = self.process_text(text.upper())
decrypted = ""
for i in range(0, len(text) - self.break_key + 1, self.break_key):
batch = text[i : i + self.break_key]
- batch_vec = [self.replace_letters(char) for char in batch]
- batch_vec = numpy.array([batch_vec]).T
- batch_decrypted = self.modulus(self.decrypt_key.dot(batch_vec)).T.tolist()[
- 0
- ]
+ vec = [self.replace_letters(char) for char in batch]
+ batch_vec = np.array([vec]).T
+ batch_decrypted = self.modulus(decrypt_key.dot(batch_vec)).T.tolist()[0]
decrypted_batch = "".join(
self.replace_digits(num) for num in batch_decrypted
)
@@ -203,16 +188,16 @@ def decrypt(self, text: str) -> str:
return decrypted
-def main():
- N = int(input("Enter the order of the encryption key: "))
+def main() -> None:
+ n = int(input("Enter the order of the encryption key: "))
hill_matrix = []
print("Enter each row of the encryption key with space separated integers")
- for i in range(N):
+ for _ in range(n):
row = [int(x) for x in input().split()]
hill_matrix.append(row)
- hc = HillCipher(numpy.array(hill_matrix))
+ hc = HillCipher(np.array(hill_matrix))
print("Would you like to encrypt or decrypt some text? (1 or 2)")
option = input("\n1. Encrypt\n2. Decrypt\n")
diff --git a/ciphers/mixed_keyword_cypher.py b/ciphers/mixed_keyword_cypher.py
index 59298d310ce0..1b186108a73e 100644
--- a/ciphers/mixed_keyword_cypher.py
+++ b/ciphers/mixed_keyword_cypher.py
@@ -1,7 +1,11 @@
-def mixed_keyword(key: str = "college", pt: str = "UNIVERSITY") -> str:
- """
+from string import ascii_uppercase
+
- For key:hello
+def mixed_keyword(
+ keyword: str, plaintext: str, verbose: bool = False, alphabet: str = ascii_uppercase
+) -> str:
+ """
+ For keyword: hello
H E L O
A B C D
@@ -12,57 +16,60 @@ def mixed_keyword(key: str = "college", pt: str = "UNIVERSITY") -> str:
Y Z
and map vertically
- >>> mixed_keyword("college", "UNIVERSITY") # doctest: +NORMALIZE_WHITESPACE
+ >>> mixed_keyword("college", "UNIVERSITY", True) # doctest: +NORMALIZE_WHITESPACE
{'A': 'C', 'B': 'A', 'C': 'I', 'D': 'P', 'E': 'U', 'F': 'Z', 'G': 'O', 'H': 'B',
'I': 'J', 'J': 'Q', 'K': 'V', 'L': 'L', 'M': 'D', 'N': 'K', 'O': 'R', 'P': 'W',
'Q': 'E', 'R': 'F', 'S': 'M', 'T': 'S', 'U': 'X', 'V': 'G', 'W': 'H', 'X': 'N',
'Y': 'T', 'Z': 'Y'}
'XKJGUFMJST'
+
+ >>> mixed_keyword("college", "UNIVERSITY", False) # doctest: +NORMALIZE_WHITESPACE
+ 'XKJGUFMJST'
"""
- key = key.upper()
- pt = pt.upper()
- temp = []
- for i in key:
- if i not in temp:
- temp.append(i)
- len_temp = len(temp)
- # print(temp)
- alpha = []
- modalpha = []
- for i in range(65, 91):
- t = chr(i)
- alpha.append(t)
- if t not in temp:
- temp.append(t)
- # print(temp)
- r = int(26 / 4)
- # print(r)
- k = 0
- for i in range(r):
- t = []
- for j in range(len_temp):
- t.append(temp[k])
- if not (k < 25):
- break
- k += 1
- modalpha.append(t)
- # print(modalpha)
- d = {}
- j = 0
- k = 0
- for j in range(len_temp):
- for i in modalpha:
- if not (len(i) - 1 >= j):
- break
- d[alpha[k]] = i[j]
- if not k < 25:
+ keyword = keyword.upper()
+ plaintext = plaintext.upper()
+ alphabet_set = set(alphabet)
+
+ # create a list of unique characters in the keyword - their order matters
+ # it determines how we will map plaintext characters to the ciphertext
+ unique_chars = []
+ for char in keyword:
+ if char in alphabet_set and char not in unique_chars:
+ unique_chars.append(char)
+ # the number of those unique characters will determine the number of rows
+ num_unique_chars_in_keyword = len(unique_chars)
+
+ # create a shifted version of the alphabet
+ shifted_alphabet = unique_chars + [
+ char for char in alphabet if char not in unique_chars
+ ]
+
+ # create a modified alphabet by splitting the shifted alphabet into rows
+ modified_alphabet = [
+ shifted_alphabet[k : k + num_unique_chars_in_keyword]
+ for k in range(0, 26, num_unique_chars_in_keyword)
+ ]
+
+ # map the alphabet characters to the modified alphabet characters
+ # going 'vertically' through the modified alphabet - consider columns first
+ mapping = {}
+ letter_index = 0
+ for column in range(num_unique_chars_in_keyword):
+ for row in modified_alphabet:
+ # if current row (the last one) is too short, break out of loop
+ if len(row) <= column:
break
- k += 1
- print(d)
- cypher = ""
- for i in pt:
- cypher += d[i]
- return cypher
+
+ # map current letter to letter in modified alphabet
+ mapping[alphabet[letter_index]] = row[column]
+ letter_index += 1
+
+ if verbose:
+ print(mapping)
+ # create the encrypted text by mapping the plaintext to the modified alphabet
+ return "".join(mapping.get(char, char) for char in plaintext)
-print(mixed_keyword("college", "UNIVERSITY"))
+if __name__ == "__main__":
+ # example use
+ print(mixed_keyword("college", "UNIVERSITY"))
diff --git a/ciphers/mono_alphabetic_ciphers.py b/ciphers/mono_alphabetic_ciphers.py
index 0a29d6442896..46013f4936bc 100644
--- a/ciphers/mono_alphabetic_ciphers.py
+++ b/ciphers/mono_alphabetic_ciphers.py
@@ -1,7 +1,11 @@
+from typing import Literal
+
LETTERS = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
-def translate_message(key, message, mode):
+def translate_message(
+ key: str, message: str, mode: Literal["encrypt", "decrypt"]
+) -> str:
"""
>>> translate_message("QWERTYUIOPASDFGHJKLZXCVBNM","Hello World","encrypt")
'Pcssi Bidsm'
@@ -40,7 +44,7 @@ def decrypt_message(key: str, message: str) -> str:
return translate_message(key, message, "decrypt")
-def main():
+def main() -> None:
message = "Hello World"
key = "QWERTYUIOPASDFGHJKLZXCVBNM"
mode = "decrypt" # set to 'encrypt' or 'decrypt'
diff --git a/ciphers/morse_code.py b/ciphers/morse_code.py
new file mode 100644
index 000000000000..0370c26fe4a6
--- /dev/null
+++ b/ciphers/morse_code.py
@@ -0,0 +1,58 @@
+#!/usr/bin/env python3
+
+"""
+Python program to translate to and from Morse code.
+
+https://en.wikipedia.org/wiki/Morse_code
+"""
+
+# fmt: off
+MORSE_CODE_DICT = {
+ "A": ".-", "B": "-...", "C": "-.-.", "D": "-..", "E": ".", "F": "..-.", "G": "--.",
+ "H": "....", "I": "..", "J": ".---", "K": "-.-", "L": ".-..", "M": "--", "N": "-.",
+ "O": "---", "P": ".--.", "Q": "--.-", "R": ".-.", "S": "...", "T": "-", "U": "..-",
+ "V": "...-", "W": ".--", "X": "-..-", "Y": "-.--", "Z": "--..", "1": ".----",
+ "2": "..---", "3": "...--", "4": "....-", "5": ".....", "6": "-....", "7": "--...",
+ "8": "---..", "9": "----.", "0": "-----", "&": ".-...", "@": ".--.-.",
+ ":": "---...", ",": "--..--", ".": ".-.-.-", "'": ".----.", '"': ".-..-.",
+ "?": "..--..", "/": "-..-.", "=": "-...-", "+": ".-.-.", "-": "-....-",
+ "(": "-.--.", ")": "-.--.-", "!": "-.-.--", " ": "/"
+} # Exclamation mark is not in ITU-R recommendation
+# fmt: on
+REVERSE_DICT = {value: key for key, value in MORSE_CODE_DICT.items()}
+
+
+def encrypt(message: str) -> str:
+ """
+ >>> encrypt("Sos!")
+ '... --- ... -.-.--'
+ >>> encrypt("SOS!") == encrypt("sos!")
+ True
+ """
+ return " ".join(MORSE_CODE_DICT[char] for char in message.upper())
+
+
+def decrypt(message: str) -> str:
+ """
+ >>> decrypt('... --- ... -.-.--')
+ 'SOS!'
+ """
+ return "".join(REVERSE_DICT[char] for char in message.split())
+
+
+def main() -> None:
+ """
+ >>> s = "".join(MORSE_CODE_DICT)
+ >>> decrypt(encrypt(s)) == s
+ True
+ """
+ message = "Morse code here!"
+ print(message)
+ message = encrypt(message)
+ print(message)
+ message = decrypt(message)
+ print(message)
+
+
+if __name__ == "__main__":
+ main()
diff --git a/ciphers/morse_code_implementation.py b/ciphers/morse_code_implementation.py
deleted file mode 100644
index 1cce2ef8b386..000000000000
--- a/ciphers/morse_code_implementation.py
+++ /dev/null
@@ -1,97 +0,0 @@
-# Python program to implement Morse Code Translator
-
-# Dictionary representing the morse code chart
-MORSE_CODE_DICT = {
- "A": ".-",
- "B": "-...",
- "C": "-.-.",
- "D": "-..",
- "E": ".",
- "F": "..-.",
- "G": "--.",
- "H": "....",
- "I": "..",
- "J": ".---",
- "K": "-.-",
- "L": ".-..",
- "M": "--",
- "N": "-.",
- "O": "---",
- "P": ".--.",
- "Q": "--.-",
- "R": ".-.",
- "S": "...",
- "T": "-",
- "U": "..-",
- "V": "...-",
- "W": ".--",
- "X": "-..-",
- "Y": "-.--",
- "Z": "--..",
- "1": ".----",
- "2": "..---",
- "3": "...--",
- "4": "....-",
- "5": ".....",
- "6": "-....",
- "7": "--...",
- "8": "---..",
- "9": "----.",
- "0": "-----",
- "&": ".-...",
- "@": ".--.-.",
- ":": "---...",
- ",": "--..--",
- ".": ".-.-.-",
- "'": ".----.",
- '"': ".-..-.",
- "?": "..--..",
- "/": "-..-.",
- "=": "-...-",
- "+": ".-.-.",
- "-": "-....-",
- "(": "-.--.",
- ")": "-.--.-",
- # Exclamation mark is not in ITU-R recommendation
- "!": "-.-.--",
-}
-
-
-def encrypt(message: str) -> str:
- cipher = ""
- for letter in message:
- if letter != " ":
- cipher += MORSE_CODE_DICT[letter] + " "
- else:
- cipher += "/ "
-
- # Remove trailing space added on line 64
- return cipher[:-1]
-
-
-def decrypt(message: str) -> str:
- decipher = ""
- letters = message.split(" ")
- for letter in letters:
- if letter != "/":
- decipher += list(MORSE_CODE_DICT.keys())[
- list(MORSE_CODE_DICT.values()).index(letter)
- ]
- else:
- decipher += " "
-
- return decipher
-
-
-def main():
- message = "Morse code here"
- result = encrypt(message.upper())
- print(result)
-
- message = result
- result = decrypt(message)
- print(result)
-
-
-if __name__ == "__main__":
- main()
diff --git a/ciphers/onepad_cipher.py b/ciphers/onepad_cipher.py
index a91f2b4d31c5..c4fb22e14a06 100644
--- a/ciphers/onepad_cipher.py
+++ b/ciphers/onepad_cipher.py
@@ -2,8 +2,29 @@
class Onepad:
- def encrypt(self, text: str) -> ([str], [int]):
- """Function to encrypt text using pseudo-random numbers"""
+ @staticmethod
+ def encrypt(text: str) -> tuple[list[int], list[int]]:
+ """
+ Function to encrypt text using pseudo-random numbers
+ >>> Onepad().encrypt("")
+ ([], [])
+ >>> Onepad().encrypt([])
+ ([], [])
+ >>> random.seed(1)
+ >>> Onepad().encrypt(" ")
+ ([6969], [69])
+ >>> random.seed(1)
+ >>> Onepad().encrypt("Hello")
+ ([9729, 114756, 4653, 31309, 10492], [69, 292, 33, 131, 61])
+ >>> Onepad().encrypt(1)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'int' object is not iterable
+ >>> Onepad().encrypt(1.1)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'float' object is not iterable
+ """
plain = [ord(i) for i in text]
key = []
cipher = []
@@ -14,14 +35,27 @@ def encrypt(self, text: str) -> ([str], [int]):
key.append(k)
return cipher, key
- def decrypt(self, cipher: [str], key: [int]) -> str:
- """Function to decrypt text using pseudo-random numbers."""
+ @staticmethod
+ def decrypt(cipher: list[int], key: list[int]) -> str:
+ """
+ Function to decrypt text using pseudo-random numbers.
+ >>> Onepad().decrypt([], [])
+ ''
+ >>> Onepad().decrypt([35], [])
+ ''
+ >>> Onepad().decrypt([], [35])
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> random.seed(1)
+ >>> Onepad().decrypt([9729, 114756, 4653, 31309, 10492], [69, 292, 33, 131, 61])
+ 'Hello'
+ """
plain = []
for i in range(len(key)):
p = int((cipher[i] - (key[i]) ** 2) / key[i])
plain.append(chr(p))
- plain = "".join([i for i in plain])
- return plain
+ return "".join(plain)
if __name__ == "__main__":
diff --git a/ciphers/permutation_cipher.py b/ciphers/permutation_cipher.py
new file mode 100644
index 000000000000..9e1c64a7b4ea
--- /dev/null
+++ b/ciphers/permutation_cipher.py
@@ -0,0 +1,143 @@
+"""
+The permutation cipher, also called the transposition cipher, is a simple encryption
+technique that rearranges the characters in a message based on a secret key. It
+divides the message into blocks and applies a permutation to the characters within
+each block according to the key. The key is a sequence of unique integers that
+determine the order of character rearrangement.
+
+For more info: https://www.nku.edu/~christensen/1402%20permutation%20ciphers.pdf
+"""
+
+import random
+
+
+def generate_valid_block_size(message_length: int) -> int:
+ """
+ Generate a valid block size that is a factor of the message length.
+
+ Args:
+ message_length (int): The length of the message.
+
+ Returns:
+ int: A valid block size.
+
+ Example:
+ >>> random.seed(1)
+ >>> generate_valid_block_size(12)
+ 3
+ """
+ block_sizes = [
+ block_size
+ for block_size in range(2, message_length + 1)
+ if message_length % block_size == 0
+ ]
+ return random.choice(block_sizes)
+
+
+def generate_permutation_key(block_size: int) -> list[int]:
+ """
+ Generate a random permutation key of a specified block size.
+
+ Args:
+ block_size (int): The size of each permutation block.
+
+ Returns:
+ list[int]: A list containing a random permutation of digits.
+
+ Example:
+ >>> random.seed(0)
+ >>> generate_permutation_key(4)
+ [2, 0, 1, 3]
+ """
+ digits = list(range(block_size))
+ random.shuffle(digits)
+ return digits
+
+
+def encrypt(
+ message: str, key: list[int] | None = None, block_size: int | None = None
+) -> tuple[str, list[int]]:
+ """
+ Encrypt a message using a permutation cipher with block rearrangement using a key.
+
+ Args:
+ message (str): The plaintext message to be encrypted.
+ key (list[int]): The permutation key for decryption.
+ block_size (int): The size of each permutation block.
+
+ Returns:
+ tuple: A tuple containing the encrypted message and the encryption key.
+
+ Example:
+ >>> encrypted_message, key = encrypt("HELLO WORLD")
+ >>> decrypted_message = decrypt(encrypted_message, key)
+ >>> decrypted_message
+ 'HELLO WORLD'
+ """
+ message = message.upper()
+ message_length = len(message)
+
+ if key is None or block_size is None:
+ block_size = generate_valid_block_size(message_length)
+ key = generate_permutation_key(block_size)
+
+ encrypted_message = ""
+
+ for i in range(0, message_length, block_size):
+ block = message[i : i + block_size]
+ rearranged_block = [block[digit] for digit in key]
+ encrypted_message += "".join(rearranged_block)
+
+ return encrypted_message, key
+
+
+def decrypt(encrypted_message: str, key: list[int]) -> str:
+ """
+ Decrypt an encrypted message using a permutation cipher with block rearrangement.
+
+ Args:
+ encrypted_message (str): The encrypted message.
+ key (list[int]): The permutation key for decryption.
+
+ Returns:
+ str: The decrypted plaintext message.
+
+ Example:
+ >>> encrypted_message, key = encrypt("HELLO WORLD")
+ >>> decrypted_message = decrypt(encrypted_message, key)
+ >>> decrypted_message
+ 'HELLO WORLD'
+ """
+ key_length = len(key)
+ decrypted_message = ""
+
+ for i in range(0, len(encrypted_message), key_length):
+ block = encrypted_message[i : i + key_length]
+ original_block = [""] * key_length
+ for j, digit in enumerate(key):
+ original_block[digit] = block[j]
+ decrypted_message += "".join(original_block)
+
+ return decrypted_message
+
+
+def main() -> None:
+ """
+ Driver function to pass message to get encrypted, then decrypted.
+
+ Example:
+ >>> main()
+ Decrypted message: HELLO WORLD
+ """
+ message = "HELLO WORLD"
+ encrypted_message, key = encrypt(message)
+
+ decrypted_message = decrypt(encrypted_message, key)
+ print(f"Decrypted message: {decrypted_message}")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ main()
diff --git a/ciphers/playfair_cipher.py b/ciphers/playfair_cipher.py
index 219437448e53..d48f113f02e0 100644
--- a/ciphers/playfair_cipher.py
+++ b/ciphers/playfair_cipher.py
@@ -1,8 +1,30 @@
+"""
+https://en.wikipedia.org/wiki/Playfair_cipher#Description
+
+The Playfair cipher was developed by Charles Wheatstone in 1854
+It's use was heavily promotedby Lord Playfair, hence its name
+
+Some features of the Playfair cipher are:
+
+1) It was the first literal diagram substitution cipher
+2) It is a manual symmetric encryption technique
+3) It is a multiple letter encryption cipher
+
+The implementation in the code below encodes alphabets only.
+It removes spaces, special characters and numbers from the
+code.
+
+Playfair is no longer used by military forces because of known
+insecurities and of the advent of automated encryption devices.
+This cipher is regarded as insecure since before World War I.
+"""
+
import itertools
import string
+from collections.abc import Generator, Iterable
-def chunker(seq, size):
+def chunker(seq: Iterable[str], size: int) -> Generator[tuple[str, ...]]:
it = iter(seq)
while True:
chunk = tuple(itertools.islice(it, size))
@@ -37,8 +59,7 @@ def prepare_input(dirty: str) -> str:
return clean
-def generate_table(key: str) -> [str]:
-
+def generate_table(key: str) -> list[str]:
# I and J are used interchangeably to allow
# us to use a 5x5 table (25 letters)
alphabet = "ABCDEFGHIKLMNOPQRSTUVWXYZ"
@@ -60,11 +81,26 @@ def generate_table(key: str) -> [str]:
def encode(plaintext: str, key: str) -> str:
+ """
+ Encode the given plaintext using the Playfair cipher.
+ Takes the plaintext and the key as input and returns the encoded string.
+
+ >>> encode("Hello", "MONARCHY")
+ 'CFSUPM'
+ >>> encode("attack on the left flank", "EMERGENCY")
+ 'DQZSBYFSDZFMFNLOHFDRSG'
+ >>> encode("Sorry!", "SPECIAL")
+ 'AVXETX'
+ >>> encode("Number 1", "NUMBER")
+ 'UMBENF'
+ >>> encode("Photosynthesis!", "THE SUN")
+ 'OEMHQHVCHESUKE'
+ """
+
table = generate_table(key)
plaintext = prepare_input(plaintext)
ciphertext = ""
- # https://en.wikipedia.org/wiki/Playfair_cipher#Description
for char1, char2 in chunker(plaintext, 2):
row1, col1 = divmod(table.index(char1), 5)
row2, col2 = divmod(table.index(char2), 5)
@@ -83,10 +119,20 @@ def encode(plaintext: str, key: str) -> str:
def decode(ciphertext: str, key: str) -> str:
+ """
+ Decode the input string using the provided key.
+
+ >>> decode("BMZFAZRZDH", "HAZARD")
+ 'FIREHAZARD'
+ >>> decode("HNBWBPQT", "AUTOMOBILE")
+ 'DRIVINGX'
+ >>> decode("SLYSSAQS", "CASTLE")
+ 'ATXTACKX'
+ """
+
table = generate_table(key)
plaintext = ""
- # https://en.wikipedia.org/wiki/Playfair_cipher#Description
for char1, char2 in chunker(ciphertext, 2):
row1, col1 = divmod(table.index(char1), 5)
row2, col2 = divmod(table.index(char2), 5)
@@ -102,3 +148,12 @@ def decode(ciphertext: str, key: str) -> str:
plaintext += table[row2 * 5 + col1]
return plaintext
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ print("Encoded:", encode("BYE AND THANKS", "GREETING"))
+ print("Decoded:", decode("CXRBANRLBALQ", "GREETING"))
diff --git a/ciphers/polybius.py b/ciphers/polybius.py
new file mode 100644
index 000000000000..d83badf4ac0a
--- /dev/null
+++ b/ciphers/polybius.py
@@ -0,0 +1,96 @@
+#!/usr/bin/env python3
+
+"""
+A Polybius Square is a table that allows someone to translate letters into numbers.
+
+https://www.braingle.com/brainteasers/codes/polybius.php
+"""
+
+import numpy as np
+
+SQUARE = [
+ ["a", "b", "c", "d", "e"],
+ ["f", "g", "h", "i", "k"],
+ ["l", "m", "n", "o", "p"],
+ ["q", "r", "s", "t", "u"],
+ ["v", "w", "x", "y", "z"],
+]
+
+
+class PolybiusCipher:
+ def __init__(self) -> None:
+ self.SQUARE = np.array(SQUARE)
+
+ def letter_to_numbers(self, letter: str) -> np.ndarray:
+ """
+ Return the pair of numbers that represents the given letter in the
+ polybius square
+ >>> np.array_equal(PolybiusCipher().letter_to_numbers('a'), [1,1])
+ True
+
+ >>> np.array_equal(PolybiusCipher().letter_to_numbers('u'), [4,5])
+ True
+ """
+ index1, index2 = np.where(letter == self.SQUARE)
+ indexes = np.concatenate([index1 + 1, index2 + 1])
+ return indexes
+
+ def numbers_to_letter(self, index1: int, index2: int) -> str:
+ """
+ Return the letter corresponding to the position [index1, index2] in
+ the polybius square
+
+ >>> PolybiusCipher().numbers_to_letter(4, 5) == "u"
+ True
+
+ >>> PolybiusCipher().numbers_to_letter(1, 1) == "a"
+ True
+ """
+ return self.SQUARE[index1 - 1, index2 - 1]
+
+ def encode(self, message: str) -> str:
+ """
+ Return the encoded version of message according to the polybius cipher
+
+ >>> PolybiusCipher().encode("test message") == "44154344 32154343112215"
+ True
+
+ >>> PolybiusCipher().encode("Test Message") == "44154344 32154343112215"
+ True
+ """
+ message = message.lower()
+ message = message.replace("j", "i")
+
+ encoded_message = ""
+ for letter_index in range(len(message)):
+ if message[letter_index] != " ":
+ numbers = self.letter_to_numbers(message[letter_index])
+ encoded_message = encoded_message + str(numbers[0]) + str(numbers[1])
+ elif message[letter_index] == " ":
+ encoded_message = encoded_message + " "
+
+ return encoded_message
+
+ def decode(self, message: str) -> str:
+ """
+ Return the decoded version of message according to the polybius cipher
+
+ >>> PolybiusCipher().decode("44154344 32154343112215") == "test message"
+ True
+
+ >>> PolybiusCipher().decode("4415434432154343112215") == "testmessage"
+ True
+ """
+ message = message.replace(" ", " ")
+ decoded_message = ""
+ for numbers_index in range(int(len(message) / 2)):
+ if message[numbers_index * 2] != " ":
+ index1 = message[numbers_index * 2]
+ index2 = message[numbers_index * 2 + 1]
+
+ letter = self.numbers_to_letter(int(index1), int(index2))
+ decoded_message = decoded_message + letter
+ elif message[numbers_index * 2] == " ":
+ decoded_message = decoded_message + " "
+
+ return decoded_message
diff --git a/ciphers/porta_cipher.py b/ciphers/porta_cipher.py
index 29043c4c9fac..498ae294041e 100644
--- a/ciphers/porta_cipher.py
+++ b/ciphers/porta_cipher.py
@@ -28,7 +28,7 @@
}
-def generate_table(key: str) -> [(str, str)]:
+def generate_table(key: str) -> list[tuple[str, str]]:
"""
>>> generate_table('marvin') # doctest: +NORMALIZE_WHITESPACE
[('ABCDEFGHIJKLM', 'UVWXYZNOPQRST'), ('ABCDEFGHIJKLM', 'NOPQRSTUVWXYZ'),
@@ -60,30 +60,21 @@ def decrypt(key: str, words: str) -> str:
return encrypt(key, words)
-def get_position(table: [(str, str)], char: str) -> (int, int) or (None, None):
+def get_position(table: tuple[str, str], char: str) -> tuple[int, int]:
"""
- >>> table = [
- ... ('ABCDEFGHIJKLM', 'UVWXYZNOPQRST'), ('ABCDEFGHIJKLM', 'NOPQRSTUVWXYZ'),
- ... ('ABCDEFGHIJKLM', 'STUVWXYZNOPQR'), ('ABCDEFGHIJKLM', 'QRSTUVWXYZNOP'),
- ... ('ABCDEFGHIJKLM', 'WXYZNOPQRSTUV'), ('ABCDEFGHIJKLM', 'UVWXYZNOPQRST')]
- >>> get_position(table, 'A')
- (None, None)
+ >>> get_position(generate_table('marvin')[0], 'M')
+ (0, 12)
"""
- if char in table[0]:
- row = 0
- else:
- row = 1 if char in table[1] else -1
- return (None, None) if row == -1 else (row, table[row].index(char))
+ # `char` is either in the 0th row or the 1st row
+ row = 0 if char in table[0] else 1
+ col = table[row].index(char)
+ return row, col
-def get_opponent(table: [(str, str)], char: str) -> str:
+def get_opponent(table: tuple[str, str], char: str) -> str:
"""
- >>> table = [
- ... ('ABCDEFGHIJKLM', 'UVWXYZNOPQRST'), ('ABCDEFGHIJKLM', 'NOPQRSTUVWXYZ'),
- ... ('ABCDEFGHIJKLM', 'STUVWXYZNOPQR'), ('ABCDEFGHIJKLM', 'QRSTUVWXYZNOP'),
- ... ('ABCDEFGHIJKLM', 'WXYZNOPQRSTUV'), ('ABCDEFGHIJKLM', 'UVWXYZNOPQRST')]
- >>> get_opponent(table, 'A')
- 'A'
+ >>> get_opponent(generate_table('marvin')[0], 'M')
+ 'T'
"""
row, col = get_position(table, char.upper())
if row == 1:
@@ -97,14 +88,16 @@ def get_opponent(table: [(str, str)], char: str) -> str:
doctest.testmod() # Fist ensure that all our tests are passing...
"""
- ENTER KEY: marvin
- ENTER TEXT TO ENCRYPT: jessica
- ENCRYPTED: QRACRWU
- DECRYPTED WITH KEY: JESSICA
+ Demo:
+
+ Enter key: marvin
+ Enter text to encrypt: jessica
+ Encrypted: QRACRWU
+ Decrypted with key: JESSICA
"""
- key = input("ENTER KEY: ").strip()
- text = input("ENTER TEXT TO ENCRYPT: ").strip()
+ key = input("Enter key: ").strip()
+ text = input("Enter text to encrypt: ").strip()
cipher_text = encrypt(key, text)
- print(f"ENCRYPTED: {cipher_text}")
- print(f"DECRYPTED WITH KEY: {decrypt(key, cipher_text)}")
+ print(f"Encrypted: {cipher_text}")
+ print(f"Decrypted with key: {decrypt(key, cipher_text)}")
diff --git a/ciphers/prehistoric_men.txt b/ciphers/prehistoric_men.txt
index a58e533a8405..8d1b2bd8c8d1 100644
--- a/ciphers/prehistoric_men.txt
+++ b/ciphers/prehistoric_men.txt
@@ -40,8 +40,8 @@ Transcriber's note:
version referred to above. One example of this might
occur in the second paragraph under "Choppers and
Adze-like Tools", page 46, which contains the phrase
- an adze cutting edge is ? shaped. The symbol before
- shaped looks like a sharply-italicized sans-serif L.
+ �an adze cutting edge is ? shaped�. The symbol before
+ �shaped� looks like a sharply-italicized sans-serif �L�.
Devices that cannot display that symbol may substitute
a question mark, a square, or other symbol.
@@ -98,7 +98,7 @@ forced or pedantic; at least I have done my very best to tell the story
simply and clearly.
Many friends have aided in the preparation of the book. The whimsical
-charm of Miss Susan Richerts illustrations add enormously to the
+charm of Miss Susan Richert�s illustrations add enormously to the
spirit I wanted. She gave freely of her own time on the drawings and
in planning the book with me. My colleagues at the University of
Chicago, especially Professor Wilton M. Krogman (now of the University
@@ -108,7 +108,7 @@ the Department of Anthropology, gave me counsel in matters bearing on
their special fields, and the Department of Anthropology bore some of
the expense of the illustrations. From Mrs. Irma Hunter and Mr. Arnold
Maremont, who are not archeologists at all and have only an intelligent
-laymans notion of archeology, I had sound advice on how best to tell
+layman�s notion of archeology, I had sound advice on how best to tell
the story. I am deeply indebted to all these friends.
While I was preparing the second edition, I had the great fortune
@@ -117,13 +117,13 @@ Washburn, now of the Department of Anthropology of the University of
California, and the fourth, fifth, and sixth chapters with Professor
Hallum L. Movius, Jr., of the Peabody Museum, Harvard University. The
book has gained greatly in accuracy thereby. In matters of dating,
-Professor Movius and the indications of Professor W. F. Libbys Carbon
+Professor Movius and the indications of Professor W. F. Libby�s Carbon
14 chronology project have both encouraged me to choose the lowest
dates now current for the events of the Pleistocene Ice Age. There is
still no certain way of fixing a direct chronology for most of the
-Pleistocene, but Professor Libbys method appears very promising for
+Pleistocene, but Professor Libby�s method appears very promising for
its end range and for proto-historic dates. In any case, this book
-names periods, and new dates may be written in against mine, if new
+names �periods,� and new dates may be written in against mine, if new
and better dating systems appear.
I wish to thank Dr. Clifford C. Gregg, Director of Chicago Natural
@@ -150,7 +150,7 @@ Clark Howell of the Department of Anthropology of the University of
Chicago in reworking the earlier chapters, and he was very patient in
the matter, which I sincerely appreciate.
-All of Mrs. Susan Richert Allens original drawings appear, but a few
+All of Mrs. Susan Richert Allen�s original drawings appear, but a few
necessary corrections have been made in some of the charts and some new
drawings have been added by Mr. John Pfiffner, Staff Artist, Chicago
Natural History Museum.
@@ -200,7 +200,7 @@ HOW WE LEARN about Prehistoric Men
Prehistory means the time before written history began. Actually, more
-than 99 per cent of mans story is prehistory. Man is at least half a
+than 99 per cent of man�s story is prehistory. Man is at least half a
million years old, but he did not begin to write history (or to write
anything) until about 5,000 years ago.
@@ -216,7 +216,7 @@ The scientists who study the bones and teeth and any other parts
they find of the bodies of prehistoric men, are called _physical
anthropologists_. Physical anthropologists are trained, much like
doctors, to know all about the human body. They study living people,
-too; they know more about the biological facts of human races than
+too; they know more about the biological facts of human �races� than
anybody else. If the police find a badly decayed body in a trunk,
they ask a physical anthropologist to tell them what the person
originally looked like. The physical anthropologists who specialize in
@@ -228,14 +228,14 @@ ARCHEOLOGISTS
There is a kind of scientist who studies the things that prehistoric
men made and did. Such a scientist is called an _archeologist_. It is
-the archeologists business to look for the stone and metal tools, the
+the archeologist�s business to look for the stone and metal tools, the
pottery, the graves, and the caves or huts of the men who lived before
history began.
But there is more to archeology than just looking for things. In
-Professor V. Gordon Childes words, archeology furnishes a sort of
+Professor V. Gordon Childe�s words, archeology �furnishes a sort of
history of human activity, provided always that the actions have
-produced concrete results and left recognizable material traces. You
+produced concrete results and left recognizable material traces.� You
will see that there are at least three points in what Childe says:
1. The archeologists have to find the traces of things left behind by
@@ -245,7 +245,7 @@ will see that there are at least three points in what Childe says:
too soft or too breakable to last through the years. However,
3. The archeologist must use whatever he can find to tell a story--to
- make a sort of history--from the objects and living-places and
+ make a �sort of history�--from the objects and living-places and
graves that have escaped destruction.
What I mean is this: Let us say you are walking through a dump yard,
@@ -253,8 +253,8 @@ and you find a rusty old spark plug. If you want to think about what
the spark plug means, you quickly remember that it is a part of an
automobile motor. This tells you something about the man who threw
the spark plug on the dump. He either had an automobile, or he knew
-or lived near someone who did. He cant have lived so very long ago,
-youll remember, because spark plugs and automobiles are only about
+or lived near someone who did. He can�t have lived so very long ago,
+you�ll remember, because spark plugs and automobiles are only about
sixty years old.
When you think about the old spark plug in this way you have
@@ -264,8 +264,8 @@ It is the same way with the man-made things we archeologists find
and put in museums. Usually, only a few of these objects are pretty
to look at; but each of them has some sort of story to tell. Making
the interpretation of his finds is the most important part of the
-archeologists job. It is the way he gets at the sort of history of
-human activity which is expected of archeology.
+archeologist�s job. It is the way he gets at the �sort of history of
+human activity� which is expected of archeology.
SOME OTHER SCIENTISTS
@@ -274,7 +274,7 @@ There are many other scientists who help the archeologist and the
physical anthropologist find out about prehistoric men. The geologists
help us tell the age of the rocks or caves or gravel beds in which
human bones or man-made objects are found. There are other scientists
-with names which all begin with paleo (the Greek word for old). The
+with names which all begin with �paleo� (the Greek word for �old�). The
_paleontologists_ study fossil animals. There are also, for example,
such scientists as _paleobotanists_ and _paleoclimatologists_, who
study ancient plants and climates. These scientists help us to know
@@ -306,20 +306,20 @@ systems.
The rate of disappearance of radioactivity as time passes.[1]]
[1] It is important that the limitations of the radioactive carbon
- dating system be held in mind. As the statistics involved in
+ �dating� system be held in mind. As the statistics involved in
the system are used, there are two chances in three that the
- date of the sample falls within the range given as plus or
- minus an added number of years. For example, the date for the
- Jarmo village (see chart), given as 6750 200 B.C., really
+ �date� of the sample falls within the range given as plus or
+ minus an added number of years. For example, the �date� for the
+ Jarmo village (see chart), given as 6750 � 200 B.C., really
means that there are only two chances in three that the real
date of the charcoal sampled fell between 6950 and 6550 B.C.
We have also begun to suspect that there are ways in which the
- samples themselves may have become contaminated, either on
+ samples themselves may have become �contaminated,� either on
the early or on the late side. We now tend to be suspicious of
single radioactive carbon determinations, or of determinations
from one site alone. But as a fabric of consistent
determinations for several or more sites of one archeological
- period, we gain confidence in the dates.
+ period, we gain confidence in the dates.
HOW THE SCIENTISTS FIND OUT
@@ -330,9 +330,9 @@ about prehistoric men. We also need a word about _how_ they find out.
All our finds came by accident until about a hundred years ago. Men
digging wells, or digging in caves for fertilizer, often turned up
ancient swords or pots or stone arrowheads. People also found some odd
-pieces of stone that didnt look like natural forms, but they also
-didnt look like any known tool. As a result, the people who found them
-gave them queer names; for example, thunderbolts. The people thought
+pieces of stone that didn�t look like natural forms, but they also
+didn�t look like any known tool. As a result, the people who found them
+gave them queer names; for example, �thunderbolts.� The people thought
the strange stones came to earth as bolts of lightning. We know now
that these strange stones were prehistoric stone tools.
@@ -349,7 +349,7 @@ story of cave men on Mount Carmel, in Palestine, began to be known.
Planned archeological digging is only about a century old. Even before
this, however, a few men realized the significance of objects they dug
from the ground; one of these early archeologists was our own Thomas
-Jefferson. The first real mound-digger was a German grocers clerk,
+Jefferson. The first real mound-digger was a German grocer�s clerk,
Heinrich Schliemann. Schliemann made a fortune as a merchant, first
in Europe and then in the California gold-rush of 1849. He became an
American citizen. Then he retired and had both money and time to test
@@ -389,16 +389,16 @@ used had been a soft, unbaked mud-brick, and most of the debris
consisted of fallen or rain-melted mud from these mud-bricks.
This idea of _stratification_, like the cake layers, was already a
-familiar one to the geologists by Schliemanns time. They could show
+familiar one to the geologists by Schliemann�s time. They could show
that their lowest layer of rock was oldest or earliest, and that the
-overlying layers became more recent as one moved upward. Schliemanns
+overlying layers became more recent as one moved upward. Schliemann�s
digging proved the same thing at Troy. His first (lowest and earliest)
city had at least nine layers above it; he thought that the second
-layer contained the remains of Homers Troy. We now know that Homeric
+layer contained the remains of Homer�s Troy. We now know that Homeric
Troy was layer VIIa from the bottom; also, we count eleven layers or
sub-layers in total.
-Schliemanns work marks the beginnings of modern archeology. Scholars
+Schliemann�s work marks the beginnings of modern archeology. Scholars
soon set out to dig on ancient sites, from Egypt to Central America.
@@ -410,21 +410,21 @@ Archeologists began to get ideas as to the kinds of objects that
belonged together. If you compared a mail-order catalogue of 1890 with
one of today, you would see a lot of differences. If you really studied
the two catalogues hard, you would also begin to see that certain
-objects go together. Horseshoes and metal buggy tires and pieces of
+objects �go together.� Horseshoes and metal buggy tires and pieces of
harness would begin to fit into a picture with certain kinds of coal
stoves and furniture and china dishes and kerosene lamps. Our friend
the spark plug, and radios and electric refrigerators and light bulbs
would fit into a picture with different kinds of furniture and dishes
-and tools. You wont be old enough to remember the kind of hats that
-women wore in 1890, but youve probably seen pictures of them, and you
-know very well they couldnt be worn with the fashions of today.
+and tools. You won�t be old enough to remember the kind of hats that
+women wore in 1890, but you�ve probably seen pictures of them, and you
+know very well they couldn�t be worn with the fashions of today.
This is one of the ways that archeologists study their materials.
The various tools and weapons and jewelry, the pottery, the kinds
of houses, and even the ways of burying the dead tend to fit into
pictures. Some archeologists call all of the things that go together to
make such a picture an _assemblage_. The assemblage of the first layer
-of Schliemanns Troy was as different from that of the seventh layer as
+of Schliemann�s Troy was as different from that of the seventh layer as
our 1900 mail-order catalogue is from the one of today.
The archeologists who came after Schliemann began to notice other
@@ -433,23 +433,23 @@ idea that people will buy better mousetraps goes back into very
ancient times. Today, if we make good automobiles or radios, we can
sell some of them in Turkey or even in Timbuktu. This means that a
few present-day types of American automobiles and radios form part
-of present-day assemblages in both Turkey and Timbuktu. The total
-present-day assemblage of Turkey is quite different from that of
+of present-day �assemblages� in both Turkey and Timbuktu. The total
+present-day �assemblage� of Turkey is quite different from that of
Timbuktu or that of America, but they have at least some automobiles
and some radios in common.
Now these automobiles and radios will eventually wear out. Let us
suppose we could go to some remote part of Turkey or to Timbuktu in a
-dream. We dont know what the date is, in our dream, but we see all
+dream. We don�t know what the date is, in our dream, but we see all
sorts of strange things and ways of living in both places. Nobody
tells us what the date is. But suddenly we see a 1936 Ford; so we
know that in our dream it has to be at least the year 1936, and only
as many years after that as we could reasonably expect a Ford to keep
-in running order. The Ford would probably break down in twenty years
-time, so the Turkish or Timbuktu assemblage were seeing in our dream
+in running order. The Ford would probably break down in twenty years�
+time, so the Turkish or Timbuktu �assemblage� we�re seeing in our dream
has to date at about A.D. 1936-56.
-Archeologists not only date their ancient materials in this way; they
+Archeologists not only �date� their ancient materials in this way; they
also see over what distances and between which peoples trading was
done. It turns out that there was a good deal of trading in ancient
times, probably all on a barter and exchange basis.
@@ -480,13 +480,13 @@ site. They find the remains of everything that would last through
time, in several different layers. They know that the assemblage in
the bottom layer was laid down earlier than the assemblage in the next
layer above, and so on up to the topmost layer, which is the latest.
-They look at the results of other digs and find that some other
+They look at the results of other �digs� and find that some other
archeologist 900 miles away has found ax-heads in his lowest layer,
exactly like the ax-heads of their fifth layer. This means that their
fifth layer must have been lived in at about the same time as was the
first layer in the site 200 miles away. It also may mean that the
people who lived in the two layers knew and traded with each other. Or
-it could mean that they didnt necessarily know each other, but simply
+it could mean that they didn�t necessarily know each other, but simply
that both traded with a third group at about the same time.
You can see that the more we dig and find, the more clearly the main
@@ -501,8 +501,8 @@ those of domesticated animals, for instance, sheep or cattle, and
therefore the people must have kept herds.
More important than anything else--as our structure grows more
-complicated and our materials increase--is the fact that a sort
-of history of human activity does begin to appear. The habits or
+complicated and our materials increase--is the fact that �a sort
+of history of human activity� does begin to appear. The habits or
traditions that men formed in the making of their tools and in the
ways they did things, begin to stand out for us. How characteristic
were these habits and traditions? What areas did they spread over?
@@ -519,7 +519,7 @@ method--chemical tests of the bones--that will enable them to discover
what the blood-type may have been. One thing is sure. We have never
found a group of skeletons so absolutely similar among themselves--so
cast from a single mould, so to speak--that we could claim to have a
-pure race. I am sure we never shall.
+�pure� race. I am sure we never shall.
We become particularly interested in any signs of change--when new
materials and tool types and ways of doing things replace old ones. We
@@ -527,7 +527,7 @@ watch for signs of social change and progress in one way or another.
We must do all this without one word of written history to aid us.
Everything we are concerned with goes back to the time _before_ men
-learned to write. That is the prehistorians job--to find out what
+learned to write. That is the prehistorian�s job--to find out what
happened before history began.
@@ -538,9 +538,9 @@ THE CHANGING WORLD in which Prehistoric Men Lived
[Illustration]
-Mankind, well say, is at least a half million years old. It is very
+Mankind, we�ll say, is at least a half million years old. It is very
hard to understand how long a time half a million years really is.
-If we were to compare this whole length of time to one day, wed get
+If we were to compare this whole length of time to one day, we�d get
something like this: The present time is midnight, and Jesus was
born just five minutes and thirty-six seconds ago. Earliest history
began less than fifteen minutes ago. Everything before 11:45 was in
@@ -569,7 +569,7 @@ book; it would mainly affect the dates earlier than 25,000 years ago.
CHANGES IN ENVIRONMENT
-The earth probably hasnt changed much in the last 5,000 years (250
+The earth probably hasn�t changed much in the last 5,000 years (250
generations). Men have built things on its surface and dug into it and
drawn boundaries on maps of it, but the places where rivers, lakes,
seas, and mountains now stand have changed very little.
@@ -605,7 +605,7 @@ the glaciers covered most of Canada and the northern United States and
reached down to southern England and France in Europe. Smaller ice
sheets sat like caps on the Rockies, the Alps, and the Himalayas. The
continental glaciation only happened north of the equator, however, so
-remember that Ice Age is only half true.
+remember that �Ice Age� is only half true.
As you know, the amount of water on and about the earth does not vary.
These large glaciers contained millions of tons of water frozen into
@@ -677,9 +677,9 @@ their dead.
At about the time when the last great glacier was finally melting away,
men in the Near East made the first basic change in human economy.
They began to plant grain, and they learned to raise and herd certain
-animals. This meant that they could store food in granaries and on the
-hoof against the bad times of the year. This first really basic change
-in mans way of living has been called the food-producing revolution.
+animals. This meant that they could store food in granaries and �on the
+hoof� against the bad times of the year. This first really basic change
+in man�s way of living has been called the �food-producing revolution.�
By the time it happened, a modern kind of climate was beginning. Men
had already grown to look as they do now. Know-how in ways of living
had developed and progressed, slowly but surely, up to a point. It was
@@ -698,25 +698,25 @@ Prehistoric Men THEMSELVES
DO WE KNOW WHERE MAN ORIGINATED?
-For a long time some scientists thought the cradle of mankind was in
+For a long time some scientists thought the �cradle of mankind� was in
central Asia. Other scientists insisted it was in Africa, and still
-others said it might have been in Europe. Actually, we dont know
-where it was. We dont even know that there was only _one_ cradle.
-If we had to choose a cradle at this moment, we would probably say
+others said it might have been in Europe. Actually, we don�t know
+where it was. We don�t even know that there was only _one_ �cradle.�
+If we had to choose a �cradle� at this moment, we would probably say
Africa. But the southern portions of Asia and Europe may also have been
included in the general area. The scene of the early development of
-mankind was certainly the Old World. It is pretty certain men didnt
+mankind was certainly the Old World. It is pretty certain men didn�t
reach North or South America until almost the end of the Ice Age--had
they done so earlier we would certainly have found some trace of them
by now.
The earliest tools we have yet found come from central and south
-Africa. By the dating system Im using, these tools must be over
+Africa. By the dating system I�m using, these tools must be over
500,000 years old. There are now reports that a few such early tools
have been found--at the Sterkfontein cave in South Africa--along with
-the bones of small fossil men called australopithecines.
+the bones of small fossil men called �australopithecines.�
-Not all scientists would agree that the australopithecines were men,
+Not all scientists would agree that the australopithecines were �men,�
or would agree that the tools were made by the australopithecines
themselves. For these sticklers, the earliest bones of men come from
the island of Java. The date would be about 450,000 years ago. So far,
@@ -727,12 +727,12 @@ Let me say it another way. How old are the earliest traces of men we
now have? Over half a million years. This was a time when the first
alpine glaciation was happening in the north. What has been found so
far? The tools which the men of those times made, in different parts
-of Africa. It is now fairly generally agreed that the men who made
-the tools were the australopithecines. There is also a more man-like
+of Africa. It is now fairly generally agreed that the �men� who made
+the tools were the australopithecines. There is also a more �man-like�
jawbone at Kanam in Kenya, but its find-spot has been questioned. The
next earliest bones we have were found in Java, and they may be almost
a hundred thousand years younger than the earliest African finds. We
-havent yet found the tools of these early Javanese. Our knowledge of
+haven�t yet found the tools of these early Javanese. Our knowledge of
tool-using in Africa spreads quickly as time goes on: soon after the
appearance of tools in the south we shall have them from as far north
as Algeria.
@@ -758,30 +758,30 @@ prove it.
MEN AND APES
Many people used to get extremely upset at the ill-formed notion
-that man descended from the apes. Such words were much more likely
-to start fights or monkey trials than the correct notion that all
+that �man descended from the apes.� Such words were much more likely
+to start fights or �monkey trials� than the correct notion that all
living animals, including man, ascended or evolved from a single-celled
organism which lived in the primeval seas hundreds of millions of years
-ago. Men are mammals, of the order called Primates, and mans living
-relatives are the great apes. Men didnt descend from the apes or
+ago. Men are mammals, of the order called Primates, and man�s living
+relatives are the great apes. Men didn�t �descend� from the apes or
apes from men, and mankind must have had much closer relatives who have
since become extinct.
Men stand erect. They also walk and run on their two feet. Apes are
happiest in trees, swinging with their arms from branch to branch.
Few branches of trees will hold the mighty gorilla, although he still
-manages to sleep in trees. Apes cant stand really erect in our sense,
+manages to sleep in trees. Apes can�t stand really erect in our sense,
and when they have to run on the ground, they use the knuckles of their
hands as well as their feet.
A key group of fossil bones here are the south African
australopithecines. These are called the _Australopithecinae_ or
-man-apes or sometimes even ape-men. We do not _know_ that they were
+�man-apes� or sometimes even �ape-men.� We do not _know_ that they were
directly ancestral to men but they can hardly have been so to apes.
-Presently Ill describe them a bit more. The reason I mention them
+Presently I�ll describe them a bit more. The reason I mention them
here is that while they had brains no larger than those of apes, their
hipbones were enough like ours so that they must have stood erect.
-There is no good reason to think they couldnt have walked as we do.
+There is no good reason to think they couldn�t have walked as we do.
BRAINS, HANDS, AND TOOLS
@@ -801,12 +801,12 @@ Nobody knows which of these three is most important, or which came
first. Most probably the growth of all three things was very much
blended together. If you think about each of the things, you will see
what I mean. Unless your hand is more flexible than a paw, and your
-thumb will work against (or oppose) your fingers, you cant hold a tool
-very well. But you wouldnt get the idea of using a tool unless you had
+thumb will work against (or oppose) your fingers, you can�t hold a tool
+very well. But you wouldn�t get the idea of using a tool unless you had
enough brain to help you see cause and effect. And it is rather hard to
see how your hand and brain would develop unless they had something to
-practice on--like using tools. In Professor Krogmans words, the hand
-must become the obedient servant of the eye and the brain. It is the
+practice on--like using tools. In Professor Krogman�s words, �the hand
+must become the obedient servant of the eye and the brain.� It is the
_co-ordination_ of these things that counts.
Many other things must have been happening to the bodies of the
@@ -820,17 +820,17 @@ little by little, all together. Men became men very slowly.
WHEN SHALL WE CALL MEN MEN?
-What do I mean when I say men? People who looked pretty much as we
+What do I mean when I say �men�? People who looked pretty much as we
do, and who used different tools to do different things, are men to me.
-Well probably never know whether the earliest ones talked or not. They
+We�ll probably never know whether the earliest ones talked or not. They
probably had vocal cords, so they could make sounds, but did they know
how to make sounds work as symbols to carry meanings? But if the fossil
-bones look like our skeletons, and if we find tools which well agree
-couldnt have been made by nature or by animals, then Id say we had
+bones look like our skeletons, and if we find tools which we�ll agree
+couldn�t have been made by nature or by animals, then I�d say we had
traces of _men_.
The australopithecine finds of the Transvaal and Bechuanaland, in
-south Africa, are bound to come into the discussion here. Ive already
+south Africa, are bound to come into the discussion here. I�ve already
told you that the australopithecines could have stood upright and
walked on their two hind legs. They come from the very base of the
Pleistocene or Ice Age, and a few coarse stone tools have been found
@@ -848,17 +848,17 @@ bones. The doubt as to whether the australopithecines used the tools
themselves goes like this--just suppose some man-like creature (whose
bones we have not yet found) made the tools and used them to kill
and butcher australopithecines. Hence a few experts tend to let
-australopithecines still hang in limbo as man-apes.
+australopithecines still hang in limbo as �man-apes.�
THE EARLIEST MEN WE KNOW
-Ill postpone talking about the tools of early men until the next
+I�ll postpone talking about the tools of early men until the next
chapter. The men whose bones were the earliest of the Java lot have
been given the name _Meganthropus_. The bones are very fragmentary. We
would not understand them very well unless we had the somewhat later
-Javanese lot--the more commonly known _Pithecanthropus_ or Java
-man--against which to refer them for study. One of the less well-known
+Javanese lot--the more commonly known _Pithecanthropus_ or �Java
+man�--against which to refer them for study. One of the less well-known
and earliest fragments, a piece of lower jaw and some teeth, rather
strongly resembles the lower jaws and teeth of the australopithecine
type. Was _Meganthropus_ a sort of half-way point between the
@@ -872,7 +872,7 @@ finds of Java man were made in 1891-92 by Dr. Eugene Dubois, a Dutch
doctor in the colonial service. Finds have continued to be made. There
are now bones enough to account for four skulls. There are also four
jaws and some odd teeth and thigh bones. Java man, generally speaking,
-was about five feet six inches tall, and didnt hold his head very
+was about five feet six inches tall, and didn�t hold his head very
erect. His skull was very thick and heavy and had room for little more
than two-thirds as large a brain as we have. He had big teeth and a big
jaw and enormous eyebrow ridges.
@@ -885,22 +885,22 @@ belonged to his near descendants.
Remember that there are several varieties of men in the whole early
Java lot, at least two of which are earlier than the _Pithecanthropus_,
-Java man. Some of the earlier ones seem to have gone in for
+�Java man.� Some of the earlier ones seem to have gone in for
bigness, in tooth-size at least. _Meganthropus_ is one of these
earlier varieties. As we said, he _may_ turn out to be a link to
the australopithecines, who _may_ or _may not_ be ancestral to men.
_Meganthropus_ is best understandable in terms of _Pithecanthropus_,
who appeared later in the same general area. _Pithecanthropus_ is
pretty well understandable from the bones he left us, and also because
-of his strong resemblance to the fully tool-using cave-dwelling Peking
-man, _Sinanthropus_, about whom we shall talk next. But you can see
+of his strong resemblance to the fully tool-using cave-dwelling �Peking
+man,� _Sinanthropus_, about whom we shall talk next. But you can see
that the physical anthropologists and prehistoric archeologists still
have a lot of work to do on the problem of earliest men.
PEKING MEN AND SOME EARLY WESTERNERS
-The earliest known Chinese are called _Sinanthropus_, or Peking man,
+The earliest known Chinese are called _Sinanthropus_, or �Peking man,�
because the finds were made near that city. In World War II, the United
States Marine guard at our Embassy in Peking tried to help get the
bones out of the city before the Japanese attack. Nobody knows where
@@ -913,9 +913,9 @@ casts of the bones.
Peking man lived in a cave in a limestone hill, made tools, cracked
animal bones to get the marrow out, and used fire. Incidentally, the
bones of Peking man were found because Chinese dig for what they call
-dragon bones and dragon teeth. Uneducated Chinese buy these things
+�dragon bones� and �dragon teeth.� Uneducated Chinese buy these things
in their drug stores and grind them into powder for medicine. The
-dragon teeth and bones are really fossils of ancient animals, and
+�dragon teeth� and �bones� are really fossils of ancient animals, and
sometimes of men. The people who supply the drug stores have learned
where to dig for strange bones and teeth. Paleontologists who get to
China go to the drug stores to buy fossils. In a roundabout way, this
@@ -924,7 +924,7 @@ is how the fallen-in cave of Peking man at Choukoutien was discovered.
Peking man was not quite as tall as Java man but he probably stood
straighter. His skull looked very much like that of the Java skull
except that it had room for a slightly larger brain. His face was less
-brutish than was Java mans face, but this isnt saying much.
+brutish than was Java man�s face, but this isn�t saying much.
Peking man dates from early in the interglacial period following the
second alpine glaciation. He probably lived close to 350,000 years
@@ -946,9 +946,9 @@ big ridges over the eyes. The more fragmentary skull from Swanscombe in
England (p. 11) has been much more carefully studied. Only the top and
back of that skull have been found. Since the skull rounds up nicely,
it has been assumed that the face and forehead must have been quite
-modern. Careful comparison with Steinheim shows that this was not
+�modern.� Careful comparison with Steinheim shows that this was not
necessarily so. This is important because it bears on the question of
-how early truly modern man appeared.
+how early truly �modern� man appeared.
Recently two fragmentary jaws were found at Ternafine in Algeria,
northwest Africa. They look like the jaws of Peking man. Tools were
@@ -971,22 +971,22 @@ modern Australian natives. During parts of the Ice Age there was a land
bridge all the way from Java to Australia.
-TWO ENGLISHMEN WHO WERENT OLD
+TWO ENGLISHMEN WHO WEREN�T OLD
The older textbooks contain descriptions of two English finds which
were thought to be very old. These were called Piltdown (_Eoanthropus
dawsoni_) and Galley Hill. The skulls were very modern in appearance.
In 1948-49, British scientists began making chemical tests which proved
that neither of these finds is very old. It is now known that both
-Piltdown man and the tools which were said to have been found with
+�Piltdown man� and the tools which were said to have been found with
him were part of an elaborate fake!
-TYPICAL CAVE MEN
+TYPICAL �CAVE MEN�
The next men we have to talk about are all members of a related group.
-These are the Neanderthal group. Neanderthal man himself was found in
-the Neander Valley, near Dsseldorf, Germany, in 1856. He was the first
+These are the Neanderthal group. �Neanderthal man� himself was found in
+the Neander Valley, near D�sseldorf, Germany, in 1856. He was the first
human fossil to be recognized as such.
[Illustration: PRINCIPAL KNOWN TYPES OF FOSSIL MEN
@@ -999,7 +999,7 @@ human fossil to be recognized as such.
PITHECANTHROPUS]
Some of us think that the neanderthaloids proper are only those people
-of western Europe who didnt get out before the beginning of the last
+of western Europe who didn�t get out before the beginning of the last
great glaciation, and who found themselves hemmed in by the glaciers
in the Alps and northern Europe. Being hemmed in, they intermarried
a bit too much and developed into a special type. Professor F. Clark
@@ -1010,7 +1010,7 @@ pre-neanderthaloids. There are traces of these pre-neanderthaloids
pretty much throughout Europe during the third interglacial period--say
100,000 years ago. The pre-neanderthaloids are represented by such
finds as the ones at Ehringsdorf in Germany and Saccopastore in Italy.
-I wont describe them for you, since they are simply less extreme than
+I won�t describe them for you, since they are simply less extreme than
the neanderthaloids proper--about half way between Steinheim and the
classic Neanderthal people.
@@ -1019,24 +1019,24 @@ get caught in the pocket of the southwest corner of Europe at the onset
of the last great glaciation became the classic Neanderthalers. Out in
the Near East, Howell thinks, it is possible to see traces of people
evolving from the pre-neanderthaloid type toward that of fully modern
-man. Certainly, we dont see such extreme cases of neanderthaloidism
+man. Certainly, we don�t see such extreme cases of �neanderthaloidism�
outside of western Europe.
There are at least a dozen good examples in the main or classic
Neanderthal group in Europe. They date to just before and in the
earlier part of the last great glaciation (85,000 to 40,000 years ago).
-Many of the finds have been made in caves. The cave men the movies
+Many of the finds have been made in caves. The �cave men� the movies
and the cartoonists show you are probably meant to be Neanderthalers.
-Im not at all sure they dragged their women by the hair; the women
+I�m not at all sure they dragged their women by the hair; the women
were probably pretty tough, too!
Neanderthal men had large bony heads, but plenty of room for brains.
Some had brain cases even larger than the average for modern man. Their
faces were heavy, and they had eyebrow ridges of bone, but the ridges
were not as big as those of Java man. Their foreheads were very low,
-and they didnt have much chin. They were about five feet three inches
-tall, but were heavy and barrel-chested. But the Neanderthalers didnt
-slouch as much as theyve been blamed for, either.
+and they didn�t have much chin. They were about five feet three inches
+tall, but were heavy and barrel-chested. But the Neanderthalers didn�t
+slouch as much as they�ve been blamed for, either.
One important thing about the Neanderthal group is that there is a fair
number of them to study. Just as important is the fact that we know
@@ -1059,10 +1059,10 @@ different-looking people.
EARLY MODERN MEN
-How early is modern man (_Homo sapiens_), the wise man? Some people
+How early is modern man (_Homo sapiens_), the �wise man�? Some people
have thought that he was very early, a few still think so. Piltdown
and Galley Hill, which were quite modern in anatomical appearance and
-_supposedly_ very early in date, were the best evidence for very
+_supposedly_ very early in date, were the best �evidence� for very
early modern men. Now that Piltdown has been liquidated and Galley Hill
is known to be very late, what is left of the idea?
@@ -1073,13 +1073,13 @@ the Ternafine jaws, you might come to the conclusion that the crown of
the Swanscombe head was that of a modern-like man.
Two more skulls, again without faces, are available from a French
-cave site, Fontchevade. They come from the time of the last great
+cave site, Font�chevade. They come from the time of the last great
interglacial, as did the pre-neanderthaloids. The crowns of the
-Fontchevade skulls also look quite modern. There is a bit of the
+Font�chevade skulls also look quite modern. There is a bit of the
forehead preserved on one of these skulls and the brow-ridge is not
heavy. Nevertheless, there is a suggestion that the bones belonged to
an immature individual. In this case, his (or even more so, if _her_)
-brow-ridges would have been weak anyway. The case for the Fontchevade
+brow-ridges would have been weak anyway. The case for the Font�chevade
fossils, as modern type men, is little stronger than that for
Swanscombe, although Professor Vallois believes it a good case.
@@ -1101,8 +1101,8 @@ of the onset of colder weather, when the last glaciation was beginning
in the north--say 75,000 years ago.
The 70 per cent modern group came from only one cave, Mugharet es-Skhul
-(cave of the kids). The other group, from several caves, had bones of
-men of the type weve been calling pre-neanderthaloid which we noted
+(�cave of the kids�). The other group, from several caves, had bones of
+men of the type we�ve been calling pre-neanderthaloid which we noted
were widespread in Europe and beyond. The tools which came with each
of these finds were generally similar, and McCown and Keith, and other
scholars since their study, have tended to assume that both the Skhul
@@ -1131,26 +1131,26 @@ important fossil men of later Europe are shown in the chart on page
DIFFERENCES IN THE EARLY MODERNS
The main early European moderns have been divided into two groups, the
-Cro-Magnon group and the Combe Capelle-Brnn group. Cro-Magnon people
+Cro-Magnon group and the Combe Capelle-Br�nn group. Cro-Magnon people
were tall and big-boned, with large, long, and rugged heads. They
must have been built like many present-day Scandinavians. The Combe
-Capelle-Brnn people were shorter; they had narrow heads and faces, and
-big eyebrow-ridges. Of course we dont find the skin or hair of these
-people. But there is little doubt they were Caucasoids (Whites).
+Capelle-Br�nn people were shorter; they had narrow heads and faces, and
+big eyebrow-ridges. Of course we don�t find the skin or hair of these
+people. But there is little doubt they were Caucasoids (�Whites�).
Another important find came in the Italian Riviera, near Monte Carlo.
Here, in a cave near Grimaldi, there was a grave containing a woman
and a young boy, buried together. The two skeletons were first called
-Negroid because some features of their bones were thought to resemble
+�Negroid� because some features of their bones were thought to resemble
certain features of modern African Negro bones. But more recently,
Professor E. A. Hooton and other experts questioned the use of the word
-Negroid in describing the Grimaldi skeletons. It is true that nothing
+�Negroid� in describing the Grimaldi skeletons. It is true that nothing
is known of the skin color, hair form, or any other fleshy feature of
-the Grimaldi people, so that the word Negroid in its usual meaning is
+the Grimaldi people, so that the word �Negroid� in its usual meaning is
not proper here. It is also not clear whether the features of the bones
-claimed to be Negroid are really so at all.
+claimed to be �Negroid� are really so at all.
-From a place called Wadjak, in Java, we have proto-Australoid skulls
+From a place called Wadjak, in Java, we have �proto-Australoid� skulls
which closely resemble those of modern Australian natives. Some of
the skulls found in South Africa, especially the Boskop skull, look
like those of modern Bushmen, but are much bigger. The ancestors of
@@ -1159,12 +1159,12 @@ Desert. True African Negroes were forest people who apparently expanded
out of the west central African area only in the last several thousand
years. Although dark in skin color, neither the Australians nor the
Bushmen are Negroes; neither the Wadjak nor the Boskop skulls are
-Negroid.
+�Negroid.�
-As weve already mentioned, Professor Weidenreich believed that Peking
+As we�ve already mentioned, Professor Weidenreich believed that Peking
man was already on the way to becoming a Mongoloid. Anyway, the
-Mongoloids would seem to have been present by the time of the Upper
-Cave at Choukoutien, the _Sinanthropus_ find-spot.
+Mongoloids would seem to have been present by the time of the �Upper
+Cave� at Choukoutien, the _Sinanthropus_ find-spot.
WHAT THE DIFFERENCES MEAN
@@ -1175,14 +1175,14 @@ From area to area, men tended to look somewhat different, just as
they do today. This is all quite natural. People _tended_ to mate
near home; in the anthropological jargon, they made up geographically
localized breeding populations. The simple continental division of
-stocks--black = Africa, yellow = Asia, white = Europe--is too simple
+�stocks�--black = Africa, yellow = Asia, white = Europe--is too simple
a picture to fit the facts. People became accustomed to life in some
-particular area within a continent (we might call it a natural area).
+particular area within a continent (we might call it a �natural area�).
As they went on living there, they evolved towards some particular
physical variety. It would, of course, have been difficult to draw
a clear boundary between two adjacent areas. There must always have
been some mating across the boundaries in every case. One thing human
-beings dont do, and never have done, is to mate for purity. It is
+beings don�t do, and never have done, is to mate for �purity.� It is
self-righteous nonsense when we try to kid ourselves into thinking that
they do.
@@ -1195,28 +1195,28 @@ and they must do the writing about races. I shall, however, give two
modern definitions of race, and then make one comment.
Dr. William G. Boyd, professor of Immunochemistry, School of
- Medicine, Boston University: We may define a human race as a
+ Medicine, Boston University: �We may define a human race as a
population which differs significantly from other human populations
in regard to the frequency of one or more of the genes it
- possesses.
+ possesses.�
Professor Sherwood L. Washburn, professor of Physical Anthropology,
- Department of Anthropology, the University of California: A race
+ Department of Anthropology, the University of California: �A �race�
is a group of genetically similar populations, and races intergrade
- because there are always intermediate populations.
+ because there are always intermediate populations.�
My comment is that the ideas involved here are all biological: they
concern groups, _not_ individuals. Boyd and Washburn may differ a bit
-on what they want to consider a population, but a population is a
+on what they want to consider a �population,� but a population is a
group nevertheless, and genetics is biology to the hilt. Now a lot of
people still think of race in terms of how people dress or fix their
food or of other habits or customs they have. The next step is to talk
-about racial purity. None of this has anything whatever to do with
+about racial �purity.� None of this has anything whatever to do with
race proper, which is a matter of the biology of groups.
-Incidentally, Im told that if man very carefully _controls_
+Incidentally, I�m told that if man very carefully _controls_
the breeding of certain animals over generations--dogs, cattle,
-chickens--he might achieve a pure race of animals. But he doesnt do
+chickens--he might achieve a �pure� race of animals. But he doesn�t do
it. Some unfortunate genetic trait soon turns up, so this has just as
carefully to be bred out again, and so on.
@@ -1240,20 +1240,20 @@ date to the second great interglacial period, about 350,000 years ago.
Piltdown and Galley Hill are out, and with them, much of the starch
in the old idea that there were two distinct lines of development
-in human evolution: (1) a line of paleoanthropic development from
+in human evolution: (1) a line of �paleoanthropic� development from
Heidelberg to the Neanderthalers where it became extinct, and (2) a
-very early modern line, through Piltdown, Galley Hill, Swanscombe, to
+very early �modern� line, through Piltdown, Galley Hill, Swanscombe, to
us. Swanscombe, Steinheim, and Ternafine are just as easily cases of
very early pre-neanderthaloids.
The pre-neanderthaloids were very widespread during the third
interglacial: Ehringsdorf, Saccopastore, some of the Mount Carmel
-people, and probably Fontchevade are cases in point. A variety of
+people, and probably Font�chevade are cases in point. A variety of
their descendants can be seen, from Java (Solo), Africa (Rhodesian
man), and about the Mediterranean and in western Europe. As the acute
cold of the last glaciation set in, the western Europeans found
themselves surrounded by water, ice, or bitter cold tundra. To vastly
-over-simplify it, they bred in and became classic neanderthaloids.
+over-simplify it, they �bred in� and became classic neanderthaloids.
But on Mount Carmel, the Skhul cave-find with its 70 per cent modern
features shows what could happen elsewhere at the same time.
@@ -1263,12 +1263,12 @@ modern skeletons of men. The modern skeletons differ from place to
place, just as different groups of men living in different places still
look different.
-What became of the Neanderthalers? Nobody can tell me for sure. Ive a
-hunch they were simply bred out again when the cold weather was over.
+What became of the Neanderthalers? Nobody can tell me for sure. I�ve a
+hunch they were simply �bred out� again when the cold weather was over.
Many Americans, as the years go by, are no longer ashamed to claim they
-have Indian blood in their veins. Give us a few more generations
+have �Indian blood in their veins.� Give us a few more generations
and there will not be very many other Americans left to whom we can
-brag about it. It certainly isnt inconceivable to me to imagine a
+brag about it. It certainly isn�t inconceivable to me to imagine a
little Cro-Magnon boy bragging to his friends about his tough, strong,
Neanderthaler great-great-great-great-grandfather!
@@ -1281,15 +1281,15 @@ Cultural BEGINNINGS
Men, unlike the lower animals, are made up of much more than flesh and
-blood and bones; for men have culture.
+blood and bones; for men have �culture.�
WHAT IS CULTURE?
-Culture is a word with many meanings. The doctors speak of making a
-culture of a certain kind of bacteria, and ants are said to have a
-culture. Then there is the Emily Post kind of culture--you say a
-person is cultured, or that he isnt, depending on such things as
+�Culture� is a word with many meanings. The doctors speak of making a
+�culture� of a certain kind of bacteria, and ants are said to have a
+�culture.� Then there is the Emily Post kind of �culture�--you say a
+person is �cultured,� or that he isn�t, depending on such things as
whether or not he eats peas with his knife.
The anthropologists use the word too, and argue heatedly over its finer
@@ -1300,7 +1300,7 @@ men from another. In this sense, a CULTURE means the way the members
of a group of people think and believe and live, the tools they make,
and the way they do things. Professor Robert Redfield says a culture
is an organized or formalized body of conventional understandings.
-Conventional understandings means the whole set of rules, beliefs,
+�Conventional understandings� means the whole set of rules, beliefs,
and standards which a group of people lives by. These understandings
show themselves in art, and in the other things a people may make and
do. The understandings continue to last, through tradition, from one
@@ -1325,12 +1325,12 @@ Egyptians. I mean their beliefs as to why grain grew, as well as their
ability to make tools with which to reap the grain. I mean their
beliefs about life after death. What I am thinking about as culture is
a thing which lasted in time. If any one Egyptian, even the Pharaoh,
-died, it didnt affect the Egyptian culture of that particular moment.
+died, it didn�t affect the Egyptian culture of that particular moment.
PREHISTORIC CULTURES
-For that long period of mans history that is all prehistory, we have
+For that long period of man�s history that is all prehistory, we have
no written descriptions of cultures. We find only the tools men made,
the places where they lived, the graves in which they buried their
dead. Fortunately for us, these tools and living places and graves all
@@ -1345,15 +1345,15 @@ of the classic European Neanderthal group of men, we have found few
cave-dwelling places of very early prehistoric men. First, there is the
fallen-in cave where Peking man was found, near Peking. Then there are
two or three other _early_, but not _very early_, possibilities. The
-finds at the base of the French cave of Fontchevade, those in one of
+finds at the base of the French cave of Font�chevade, those in one of
the Makapan caves in South Africa, and several open sites such as Dr.
-L. S. B. Leakeys Olorgesailie in Kenya doubtless all lie earlier than
+L. S. B. Leakey�s Olorgesailie in Kenya doubtless all lie earlier than
the time of the main European Neanderthal group, but none are so early
as the Peking finds.
You can see that we know very little about the home life of earlier
prehistoric men. We find different kinds of early stone tools, but we
-cant even be really sure which tools may have been used together.
+can�t even be really sure which tools may have been used together.
WHY LITTLE HAS LASTED FROM EARLY TIMES
@@ -1380,11 +1380,11 @@ there first! The front of this enormous sheet of ice moved down over
the country, crushing and breaking and plowing up everything, like a
gigantic bulldozer. You can see what happened to our camp site.
-Everything the glacier couldnt break, it pushed along in front of it
+Everything the glacier couldn�t break, it pushed along in front of it
or plowed beneath it. Rocks were ground to gravel, and soil was caught
into the ice, which afterwards melted and ran off as muddy water. Hard
-tools of flint sometimes remained whole. Human bones werent so hard;
-its a wonder _any_ of them lasted. Gushing streams of melt water
+tools of flint sometimes remained whole. Human bones weren�t so hard;
+it�s a wonder _any_ of them lasted. Gushing streams of melt water
flushed out the debris from underneath the glacier, and water flowed
off the surface and through great crevasses. The hard materials these
waters carried were even more rolled and ground up. Finally, such
@@ -1407,26 +1407,26 @@ all up, and so we cannot say which particular sets of tools belonged
together in the first place.
-EOLITHS
+�EOLITHS�
But what sort of tools do we find earliest? For almost a century,
people have been picking up odd bits of flint and other stone in the
oldest Ice Age gravels in England and France. It is now thought these
-odd bits of stone werent actually worked by prehistoric men. The
-stones were given a name, _eoliths_, or dawn stones. You can see them
+odd bits of stone weren�t actually worked by prehistoric men. The
+stones were given a name, _eoliths_, or �dawn stones.� You can see them
in many museums; but you can be pretty sure that very few of them were
actually fashioned by men.
-It is impossible to pick out eoliths that seem to be made in any
-one _tradition_. By tradition I mean a set of habits for making one
-kind of tool for some particular job. No two eoliths look very much
+It is impossible to pick out �eoliths� that seem to be made in any
+one _tradition_. By �tradition� I mean a set of habits for making one
+kind of tool for some particular job. No two �eoliths� look very much
alike: tools made as part of some one tradition all look much alike.
-Now its easy to suppose that the very earliest prehistoric men picked
-up and used almost any sort of stone. This wouldnt be surprising; you
-and I do it when we go camping. In other words, some of these eoliths
+Now it�s easy to suppose that the very earliest prehistoric men picked
+up and used almost any sort of stone. This wouldn�t be surprising; you
+and I do it when we go camping. In other words, some of these �eoliths�
may actually have been used by prehistoric men. They must have used
anything that might be handy when they needed it. We could have figured
-that out without the eoliths.
+that out without the �eoliths.�
THE ROAD TO STANDARDIZATION
@@ -1434,7 +1434,7 @@ THE ROAD TO STANDARDIZATION
Reasoning from what we know or can easily imagine, there should have
been three major steps in the prehistory of tool-making. The first step
would have been simple _utilization_ of what was at hand. This is the
-step into which the eoliths would fall. The second step would have
+step into which the �eoliths� would fall. The second step would have
been _fashioning_--the haphazard preparation of a tool when there was a
need for it. Probably many of the earlier pebble tools, which I shall
describe next, fall into this group. The third step would have been
@@ -1447,7 +1447,7 @@ tradition appears.
PEBBLE TOOLS
-At the beginning of the last chapter, youll remember that I said there
+At the beginning of the last chapter, you�ll remember that I said there
were tools from very early geological beds. The earliest bones of men
have not yet been found in such early beds although the Sterkfontein
australopithecine cave approaches this early date. The earliest tools
@@ -1467,7 +1467,7 @@ Old World besides Africa; in fact, some prehistorians already claim
to have identified a few. Since the forms and the distinct ways of
making the earlier pebble tools had not yet sufficiently jelled into
a set tradition, they are difficult for us to recognize. It is not
-so difficult, however, if there are great numbers of possibles
+so difficult, however, if there are great numbers of �possibles�
available. A little later in time the tradition becomes more clearly
set, and pebble tools are easier to recognize. So far, really large
collections of pebble tools have only been found and examined in Africa.
@@ -1475,9 +1475,9 @@ collections of pebble tools have only been found and examined in Africa.
CORE-BIFACE TOOLS
-The next tradition well look at is the _core_ or biface one. The tools
+The next tradition we�ll look at is the _core_ or biface one. The tools
are large pear-shaped pieces of stone trimmed flat on the two opposite
-sides or faces. Hence biface has been used to describe these tools.
+sides or �faces.� Hence �biface� has been used to describe these tools.
The front view is like that of a pear with a rather pointed top, and
the back view looks almost exactly the same. Look at them side on, and
you can see that the front and back faces are the same and have been
@@ -1488,7 +1488,7 @@ illustration.
[Illustration: ABBEVILLIAN BIFACE]
We have very little idea of the way in which these core-bifaces were
-used. They have been called hand axes, but this probably gives the
+used. They have been called �hand axes,� but this probably gives the
wrong idea, for an ax, to us, is not a pointed tool. All of these early
tools must have been used for a number of jobs--chopping, scraping,
cutting, hitting, picking, and prying. Since the core-bifaces tend to
@@ -1505,7 +1505,7 @@ a big block of stone. You had to break off the flake in such a way that
it was broad and thin, and also had a good sharp cutting edge. Once you
really got on to the trick of doing it, this was probably a simpler way
to make a good cutting tool than preparing a biface. You have to know
-how, though; Ive tried it and have mashed my fingers more than once.
+how, though; I�ve tried it and have mashed my fingers more than once.
The flake tools look as if they were meant mainly for chopping,
scraping, and cutting jobs. When one made a flake tool, the idea seems
@@ -1535,9 +1535,9 @@ tradition. It probably has its earliest roots in the pebble tool
tradition of African type. There are several kinds of tools in this
tradition, but all differ from the western core-bifaces and flakes.
There are broad, heavy scrapers or cleavers, and tools with an
-adze-like cutting edge. These last-named tools are called hand adzes,
-just as the core-bifaces of the west have often been called hand
-axes. The section of an adze cutting edge is ? shaped; the section of
+adze-like cutting edge. These last-named tools are called �hand adzes,�
+just as the core-bifaces of the west have often been called �hand
+axes.� The section of an adze cutting edge is ? shaped; the section of
an ax is < shaped.
[Illustration: ANYATHIAN ADZE-LIKE TOOL]
@@ -1581,17 +1581,17 @@ stratification.[3]
Soan (India)
Flake:
- Typical Mousterian
+ �Typical Mousterian�
Levalloiso-Mousterian
Levalloisian
Tayacian
Clactonian (localized in England)
Core-biface:
- Some blended elements in Mousterian
+ Some blended elements in �Mousterian�
Micoquian (= Acheulean 6 and 7)
Acheulean
- Abbevillian (once called Chellean)
+ Abbevillian (once called �Chellean�)
Pebble tool:
Oldowan
@@ -1608,8 +1608,8 @@ out of glacial gravels the easiest thing to do first is to isolate
individual types of tools into groups. First you put a bushel-basketful
of tools on a table and begin matching up types. Then you give names to
the groups of each type. The groups and the types are really matters of
-the archeologists choice; in real life, they were probably less exact
-than the archeologists lists of them. We now know pretty well in which
+the archeologists� choice; in real life, they were probably less exact
+than the archeologists� lists of them. We now know pretty well in which
of the early traditions the various early groups belong.
@@ -1635,9 +1635,9 @@ production must have been passed on from one generation to another.
I could even guess that the notions of the ideal type of one or the
other of these tools stood out in the minds of men of those times
-somewhat like a symbol of perfect tool for good job. If this were
-so--remember its only a wild guess of mine--then men were already
-symbol users. Now lets go on a further step to the fact that the words
+somewhat like a symbol of �perfect tool for good job.� If this were
+so--remember it�s only a wild guess of mine--then men were already
+symbol users. Now let�s go on a further step to the fact that the words
men speak are simply sounds, each different sound being a symbol for a
different meaning. If standardized tool-making suggests symbol-making,
is it also possible that crude word-symbols were also being made? I
@@ -1650,7 +1650,7 @@ of our second step is more suggestive, although we may not yet feel
sure that many of the earlier pebble tools were man-made products. But
with the step to standardization and the appearance of the traditions,
I believe we must surely be dealing with the traces of culture-bearing
-_men_. The conventional understandings which Professor Redfields
+_men_. The �conventional understandings� which Professor Redfield�s
definition of culture suggests are now evidenced for us in the
persistent habits for the preparation of stone tools. Were we able to
see the other things these prehistoric men must have made--in materials
@@ -1666,19 +1666,19 @@ In the last chapter, I told you that many of the older archeologists
and human paleontologists used to think that modern man was very old.
The supposed ages of Piltdown and Galley Hill were given as evidence
of the great age of anatomically modern man, and some interpretations
-of the Swanscombe and Fontchevade fossils were taken to support
+of the Swanscombe and Font�chevade fossils were taken to support
this view. The conclusion was that there were two parallel lines or
-phyla of men already present well back in the Pleistocene. The
-first of these, the more primitive or paleoanthropic line, was
+�phyla� of men already present well back in the Pleistocene. The
+first of these, the more primitive or �paleoanthropic� line, was
said to include Heidelberg, the proto-neanderthaloids and classic
-Neanderthal. The more anatomically modern or neanthropic line was
+Neanderthal. The more anatomically modern or �neanthropic� line was
thought to consist of Piltdown and the others mentioned above. The
Neanderthaler or paleoanthropic line was thought to have become extinct
after the first phase of the last great glaciation. Of course, the
modern or neanthropic line was believed to have persisted into the
-present, as the basis for the worlds population today. But with
+present, as the basis for the world�s population today. But with
Piltdown liquidated, Galley Hill known to be very late, and Swanscombe
-and Fontchevade otherwise interpreted, there is little left of the
+and Font�chevade otherwise interpreted, there is little left of the
so-called parallel phyla theory.
While the theory was in vogue, however, and as long as the European
@@ -1695,9 +1695,9 @@ where they had actually been dropped by the men who made and used
them. The tools came, rather, from the secondary hodge-podge of the
glacial gravels. I tried to give you a picture of the bulldozing action
of glaciers (p. 40) and of the erosion and weathering that were
-side-effects of a glacially conditioned climate on the earths surface.
+side-effects of a glacially conditioned climate on the earth�s surface.
As we said above, if one simply plucks tools out of the redeposited
-gravels, his natural tendency is to type the tools by groups, and to
+gravels, his natural tendency is to �type� the tools by groups, and to
think that the groups stand for something _on their own_.
In 1906, M. Victor Commont actually made a rare find of what seems
@@ -1705,15 +1705,15 @@ to have been a kind of workshop site, on a terrace above the Somme
river in France. Here, Commont realized, flake tools appeared clearly
in direct association with core-biface tools. Few prehistorians paid
attention to Commont or his site, however. It was easier to believe
-that flake tools represented a distinct culture and that this
-culture was that of the Neanderthaler or paleoanthropic line, and
-that the core-bifaces stood for another culture which was that of the
+that flake tools represented a distinct �culture� and that this
+�culture� was that of the Neanderthaler or paleoanthropic line, and
+that the core-bifaces stood for another �culture� which was that of the
supposed early modern or neanthropic line. Of course, I am obviously
skipping many details here. Some later sites with Neanderthal fossils
do seem to have only flake tools, but other such sites have both types
of tools. The flake tools which appeared _with_ the core-bifaces
in the Swanscombe gravels were never made much of, although it
-was embarrassing for the parallel phyla people that Fontchevade
+was embarrassing for the parallel phyla people that Font�chevade
ran heavily to flake tools. All in all, the parallel phyla theory
flourished because it seemed so neat and easy to understand.
@@ -1722,20 +1722,20 @@ TRADITIONS ARE TOOL-MAKING HABITS, NOT CULTURES
In case you think I simply enjoy beating a dead horse, look in any
standard book on prehistory written twenty (or even ten) years ago, or
-in most encyclopedias. Youll find that each of the individual tool
-types, of the West, at least, was supposed to represent a culture.
-The cultures were believed to correspond to parallel lines of human
+in most encyclopedias. You�ll find that each of the individual tool
+types, of the West, at least, was supposed to represent a �culture.�
+The �cultures� were believed to correspond to parallel lines of human
evolution.
In 1937, Mr. Harper Kelley strongly re-emphasized the importance
-of Commonts workshop site and the presence of flake tools with
-core-bifaces. Next followed Dr. Movius clear delineation of the
+of Commont�s workshop site and the presence of flake tools with
+core-bifaces. Next followed Dr. Movius� clear delineation of the
chopper-chopping tool tradition of the Far East. This spoiled the nice
symmetry of the flake-tool = paleoanthropic, core-biface = neanthropic
equations. Then came increasing understanding of the importance of
the pebble tools in Africa, and the location of several more workshop
sites there, especially at Olorgesailie in Kenya. Finally came the
-liquidation of Piltdown and the deflation of Galley Hills date. So it
+liquidation of Piltdown and the deflation of Galley Hill�s date. So it
is at last possible to picture an individual prehistoric man making a
flake tool to do one job and a core-biface tool to do another. Commont
showed us this picture in 1906, but few believed him.
@@ -1751,7 +1751,7 @@ that of the cave on Mount Carmel in Palestine, where the blended
pre-neanderthaloid, 70 per cent modern-type skulls were found. Here, in
the same level with the skulls, were 9,784 flint tools. Of these, only
three--doubtless strays--were core-bifaces; all the rest were flake
-tools or flake chips. We noted above how the Fontchevade cave ran to
+tools or flake chips. We noted above how the Font�chevade cave ran to
flake tools. The only conclusion I would draw from this is that times
and circumstances did exist in which prehistoric men needed only flake
tools. So they only made flake tools for those particular times and
@@ -1773,13 +1773,13 @@ piece of bone. From the gravels which yield the Clactonian flakes of
England comes the fire-hardened point of a wooden spear. There are
also the chance finds of the fossil human bones themselves, of which
we spoke in the last chapter. Aside from the cave of Peking man, none
-of the earliest tools have been found in caves. Open air or workshop
+of the earliest tools have been found in caves. Open air or �workshop�
sites which do not seem to have been disturbed later by some geological
agency are very rare.
The chart on page 65 shows graphically what the situation in
west-central Europe seems to have been. It is not yet certain whether
-there were pebble tools there or not. The Fontchevade cave comes
+there were pebble tools there or not. The Font�chevade cave comes
into the picture about 100,000 years ago or more. But for the earlier
hundreds of thousands of years--below the red-dotted line on the
chart--the tools we find come almost entirely from the haphazard
@@ -1790,13 +1790,13 @@ kinds of all-purpose tools. Almost any one of them could be used for
hacking, chopping, cutting, and scraping; so the men who used them must
have been living in a rough and ready sort of way. They found or hunted
their food wherever they could. In the anthropological jargon, they
-were food-gatherers, pure and simple.
+were �food-gatherers,� pure and simple.
Because of the mixture in the gravels and in the materials they
-carried, we cant be sure which animals these men hunted. Bones of
+carried, we can�t be sure which animals these men hunted. Bones of
the larger animals turn up in the gravels, but they could just as
well belong to the animals who hunted the men, rather than the other
-way about. We dont know. This is why camp sites like Commonts and
+way about. We don�t know. This is why camp sites like Commont�s and
Olorgesailie in Kenya are so important when we do find them. The animal
bones at Olorgesailie belonged to various mammals of extremely large
size. Probably they were taken in pit-traps, but there are a number of
@@ -1809,18 +1809,18 @@ animal.
Professor F. Clark Howell recently returned from excavating another
important open air site at Isimila in Tanganyika. The site yielded
the bones of many fossil animals and also thousands of core-bifaces,
-flakes, and choppers. But Howells reconstruction of the food-getting
-habits of the Isimila people certainly suggests that the word hunting
-is too dignified for what they did; scavenging would be much nearer
+flakes, and choppers. But Howell�s reconstruction of the food-getting
+habits of the Isimila people certainly suggests that the word �hunting�
+is too dignified for what they did; �scavenging� would be much nearer
the mark.
During a great part of this time the climate was warm and pleasant. The
second interglacial period (the time between the second and third great
alpine glaciations) lasted a long time, and during much of this time
-the climate may have been even better than ours is now. We dont know
+the climate may have been even better than ours is now. We don�t know
that earlier prehistoric men in Europe or Africa lived in caves. They
may not have needed to; much of the weather may have been so nice that
-they lived in the open. Perhaps they didnt wear clothes, either.
+they lived in the open. Perhaps they didn�t wear clothes, either.
WHAT THE PEKING CAVE-FINDS TELL US
@@ -1832,7 +1832,7 @@ were bones of dangerous animals, members of the wolf, bear, and cat
families. Some of the cat bones belonged to beasts larger than tigers.
There were also bones of other wild animals: buffalo, camel, deer,
elephants, horses, sheep, and even ostriches. Seventy per cent of the
-animals Peking man killed were fallow deer. Its much too cold and dry
+animals Peking man killed were fallow deer. It�s much too cold and dry
in north China for all these animals to live there today. So this list
helps us know that the weather was reasonably warm, and that there was
enough rain to grow grass for the grazing animals. The list also helps
@@ -1840,7 +1840,7 @@ the paleontologists to date the find.
Peking man also seems to have eaten plant food, for there are hackberry
seeds in the debris of the cave. His tools were made of sandstone and
-quartz and sometimes of a rather bad flint. As weve already seen, they
+quartz and sometimes of a rather bad flint. As we�ve already seen, they
belong in the chopper-tool tradition. It seems fairly clear that some
of the edges were chipped by right-handed people. There are also many
split pieces of heavy bone. Peking man probably split them so he could
@@ -1850,10 +1850,10 @@ Many of these split bones were the bones of Peking men. Each one of the
skulls had already had the base broken out of it. In no case were any
of the bones resting together in their natural relation to one another.
There is nothing like a burial; all of the bones are scattered. Now
-its true that animals could have scattered bodies that were not cared
+it�s true that animals could have scattered bodies that were not cared
for or buried. But splitting bones lengthwise and carefully removing
the base of a skull call for both the tools and the people to use them.
-Its pretty clear who the people were. Peking man was a cannibal.
+It�s pretty clear who the people were. Peking man was a cannibal.
* * * * *
@@ -1862,8 +1862,8 @@ prehistoric men. In those days life was rough. You evidently had to
watch out not only for dangerous animals but also for your fellow men.
You ate whatever you could catch or find growing. But you had sense
enough to build fires, and you had already formed certain habits for
-making the kinds of stone tools you needed. Thats about all we know.
-But I think well have to admit that cultural beginnings had been made,
+making the kinds of stone tools you needed. That�s about all we know.
+But I think we�ll have to admit that cultural beginnings had been made,
and that these early people were really _men_.
@@ -1876,16 +1876,16 @@ MORE EVIDENCE of Culture
While the dating is not yet sure, the material that we get from caves
in Europe must go back to about 100,000 years ago; the time of the
-classic Neanderthal group followed soon afterwards. We dont know why
+classic Neanderthal group followed soon afterwards. We don�t know why
there is no earlier material in the caves; apparently they were not
used before the last interglacial phase (the period just before the
last great glaciation). We know that men of the classic Neanderthal
group were living in caves from about 75,000 to 45,000 years ago.
New radioactive carbon dates even suggest that some of the traces of
-culture well describe in this chapter may have lasted to about 35,000
+culture we�ll describe in this chapter may have lasted to about 35,000
years ago. Probably some of the pre-neanderthaloid types of men had
also lived in caves. But we have so far found their bones in caves only
-in Palestine and at Fontchevade.
+in Palestine and at Font�chevade.
THE CAVE LAYERS
@@ -1893,7 +1893,7 @@ THE CAVE LAYERS
In parts of France, some peasants still live in caves. In prehistoric
time, many generations of people lived in them. As a result, many
caves have deep layers of debris. The first people moved in and lived
-on the rock floor. They threw on the floor whatever they didnt want,
+on the rock floor. They threw on the floor whatever they didn�t want,
and they tracked in mud; nobody bothered to clean house in those days.
Their debris--junk and mud and garbage and what not--became packed
into a layer. As time went on, and generations passed, the layer grew
@@ -1910,20 +1910,20 @@ earliest to latest. This is the _stratification_ we talked about (p.
[Illustration: SECTION OF SHELTER ON LOWER TERRACE, LE MOUSTIER]
-While we may find a mix-up in caves, its not nearly as bad as the
+While we may find a mix-up in caves, it�s not nearly as bad as the
mixing up that was done by glaciers. The animal bones and shells, the
fireplaces, the bones of men, and the tools the men made all belong
-together, if they come from one layer. Thats the reason why the cave
+together, if they come from one layer. That�s the reason why the cave
of Peking man is so important. It is also the reason why the caves in
Europe and the Near East are so important. We can get an idea of which
things belong together and which lot came earliest and which latest.
In most cases, prehistoric men lived only in the mouths of caves.
-They didnt like the dark inner chambers as places to live in. They
+They didn�t like the dark inner chambers as places to live in. They
preferred rock-shelters, at the bases of overhanging cliffs, if there
was enough overhang to give shelter. When the weather was good, they no
-doubt lived in the open air as well. Ill go on using the term cave
-since its more familiar, but remember that I really mean rock-shelter,
+doubt lived in the open air as well. I�ll go on using the term �cave�
+since it�s more familiar, but remember that I really mean rock-shelter,
as a place in which people actually lived.
The most important European cave sites are in Spain, France, and
@@ -1933,29 +1933,29 @@ found when the out-of-the-way parts of Europe, Africa, and Asia are
studied.
-AN INDUSTRY DEFINED
+AN �INDUSTRY� DEFINED
We have already seen that the earliest European cave materials are
-those from the cave of Fontchevade. Movius feels certain that the
+those from the cave of Font�chevade. Movius feels certain that the
lowest materials here date back well into the third interglacial stage,
-that which lay between the Riss (next to the last) and the Wrm I
+that which lay between the Riss (next to the last) and the W�rm I
(first stage of the last) alpine glaciations. This material consists
of an _industry_ of stone tools, apparently all made in the flake
-tradition. This is the first time we have used the word industry.
+tradition. This is the first time we have used the word �industry.�
It is useful to call all of the different tools found together in one
layer and made of _one kind of material_ an industry; that is, the
tools must be found together as men left them. Tools taken from the
glacial gravels (or from windswept desert surfaces or river gravels
-or any geological deposit) are not together in this sense. We might
-say the latter have only geological, not archeological context.
+or any geological deposit) are not �together� in this sense. We might
+say the latter have only �geological,� not �archeological� context.
Archeological context means finding things just as men left them. We
-can tell what tools go together in an industrial sense only if we
+can tell what tools go together in an �industrial� sense only if we
have archeological context.
-Up to now, the only things we could have called industries were the
+Up to now, the only things we could have called �industries� were the
worked stone industry and perhaps the worked (?) bone industry of the
Peking cave. We could add some of the very clear cases of open air
-sites, like Olorgesailie. We couldnt use the term for the stone tools
+sites, like Olorgesailie. We couldn�t use the term for the stone tools
from the glacial gravels, because we do not know which tools belonged
together. But when the cave materials begin to appear in Europe, we can
begin to speak of industries. Most of the European caves of this time
@@ -1964,16 +1964,16 @@ contain industries of flint tools alone.
THE EARLIEST EUROPEAN CAVE LAYERS
-Weve just mentioned the industry from what is said to be the oldest
+We�ve just mentioned the industry from what is said to be the oldest
inhabited cave in Europe; that is, the industry from the deepest layer
-of the site at Fontchevade. Apparently it doesnt amount to much. The
+of the site at Font�chevade. Apparently it doesn�t amount to much. The
tools are made of stone, in the flake tradition, and are very poorly
worked. This industry is called _Tayacian_. Its type tool seems to be
a smallish flake tool, but there are also larger flakes which seem to
have been fashioned for hacking. In fact, the type tool seems to be
simply a smaller edition of the Clactonian tool (pictured on p. 45).
-None of the Fontchevade tools are really good. There are scrapers,
+None of the Font�chevade tools are really good. There are scrapers,
and more or less pointed tools, and tools that may have been used
for hacking and chopping. Many of the tools from the earlier glacial
gravels are better made than those of this first industry we see in
@@ -2005,7 +2005,7 @@ core-biface and the flake traditions. The core-biface tools usually
make up less than half of all the tools in the industry. However,
the name of the biface type of tool is generally given to the whole
industry. It is called the _Acheulean_, actually a late form of it, as
-Acheulean is also used for earlier core-biface tools taken from the
+�Acheulean� is also used for earlier core-biface tools taken from the
glacial gravels. In western Europe, the name used is _Upper Acheulean_
or _Micoquian_. The same terms have been borrowed to name layers E and
F in the Tabun cave, on Mount Carmel in Palestine.
@@ -2029,7 +2029,7 @@ those used for at least one of the flake industries we shall mention
presently.
There is very little else in these early cave layers. We do not have
-a proper industry of bone tools. There are traces of fire, and of
+a proper �industry� of bone tools. There are traces of fire, and of
animal bones, and a few shells. In Palestine, there are many more
bones of deer than of gazelle in these layers; the deer lives in a
wetter climate than does the gazelle. In the European cave layers, the
@@ -2043,18 +2043,18 @@ bones of fossil men definitely in place with this industry.
FLAKE INDUSTRIES FROM THE CAVES
Two more stone industries--the _Levalloisian_ and the
-_Mousterian_--turn up at approximately the same time in the European
+�_Mousterian_�--turn up at approximately the same time in the European
cave layers. Their tools seem to be mainly in the flake tradition,
but according to some of the authorities their preparation also shows
some combination with the habits by which the core-biface tools were
prepared.
-Now notice that I dont tell you the Levalloisian and the Mousterian
+Now notice that I don�t tell you the Levalloisian and the �Mousterian�
layers are both above the late Acheulean layers. Look at the cave
-section (p. 57) and youll find that some Mousterian of Acheulean
-tradition appears above some typical Mousterian. This means that
+section (p. 57) and you�ll find that some �Mousterian of Acheulean
+tradition� appears above some �typical Mousterian.� This means that
there may be some kinds of Acheulean industries that are later than
-some kinds of Mousterian. The same is true of the Levalloisian.
+some kinds of �Mousterian.� The same is true of the Levalloisian.
There were now several different kinds of habits that men used in
making stone tools. These habits were based on either one or the other
@@ -2072,7 +2072,7 @@ were no patent laws in those days.
The extremely complicated interrelationships of the different habits
used by the tool-makers of this range of time are at last being
-systematically studied. M. Franois Bordes has developed a statistical
+systematically studied. M. Fran�ois Bordes has developed a statistical
method of great importance for understanding these tool preparation
habits.
@@ -2081,22 +2081,22 @@ THE LEVALLOISIAN AND MOUSTERIAN
The easiest Levalloisian tool to spot is a big flake tool. The trick
in making it was to fashion carefully a big chunk of stone (called
-the Levalloisian tortoise core, because it resembles the shape of
+the Levalloisian �tortoise core,� because it resembles the shape of
a turtle-shell) and then to whack this in such a way that a large
flake flew off. This large thin flake, with sharp cutting edges, is
the finished Levalloisian tool. There were various other tools in a
Levalloisian industry, but this is the characteristic _Levalloisian_
tool.
-There are several typical Mousterian stone tools. Different from
-the tools of the Levalloisian type, these were made from disc-like
-cores. There are medium-sized flake side scrapers. There are also
-some small pointed tools and some small hand axes. The last of these
+There are several �typical Mousterian� stone tools. Different from
+the tools of the Levalloisian type, these were made from �disc-like
+cores.� There are medium-sized flake �side scrapers.� There are also
+some small pointed tools and some small �hand axes.� The last of these
tool types is often a flake worked on both of the flat sides (that
is, bifacially). There are also pieces of flint worked into the form
of crude balls. The pointed tools may have been fixed on shafts to
make short jabbing spears; the round flint balls may have been used as
-bolas. Actually, we dont _know_ what either tool was used for. The
+bolas. Actually, we don�t _know_ what either tool was used for. The
points and side scrapers are illustrated (pp. 64 and 66).
[Illustration: LEVALLOIS FLAKE]
@@ -2108,9 +2108,9 @@ Nowadays the archeologists are less and less sure of the importance
of any one specific tool type and name. Twenty years ago, they used
to speak simply of Acheulean or Levalloisian or Mousterian tools.
Now, more and more, _all_ of the tools from some one layer in a
-cave are called an industry, which is given a mixed name. Thus we
-have Levalloiso-Mousterian, and Acheuleo-Levalloisian, and even
-Acheuleo-Mousterian (or Mousterian of Acheulean tradition). Bordes
+cave are called an �industry,� which is given a mixed name. Thus we
+have �Levalloiso-Mousterian,� and �Acheuleo-Levalloisian,� and even
+�Acheuleo-Mousterian� (or �Mousterian of Acheulean tradition�). Bordes�
systematic work is beginning to clear up some of our confusion.
The time of these late Acheuleo-Levalloiso-Mousterioid industries
@@ -2120,16 +2120,16 @@ phase of the last great glaciation. It was also the time that the
classic group of Neanderthal men was living in Europe. A number of
the Neanderthal fossil finds come from these cave layers. Before the
different habits of tool preparation were understood it used to be
-popular to say Neanderthal man was Mousterian man. I think this is
-wrong. What used to be called Mousterian is now known to be a variety
+popular to say Neanderthal man was �Mousterian man.� I think this is
+wrong. What used to be called �Mousterian� is now known to be a variety
of industries with tools of both core-biface and flake habits, and
-so mixed that the word Mousterian used alone really doesnt mean
+so mixed that the word �Mousterian� used alone really doesn�t mean
anything. The Neanderthalers doubtless understood the tool preparation
habits by means of which Acheulean, Levalloisian and Mousterian type
tools were produced. We also have the more modern-like Mount Carmel
people, found in a cave layer of Palestine with tools almost entirely
-in the flake tradition, called Levalloiso-Mousterian, and the
-Fontchevade-Tayacian (p. 59).
+in the flake tradition, called �Levalloiso-Mousterian,� and the
+Font�chevade-Tayacian (p. 59).
[Illustration: MOUSTERIAN POINT]
@@ -2165,7 +2165,7 @@ which seem to have served as anvils or chopping blocks, are fairly
common.
Bits of mineral, used as coloring matter, have also been found. We
-dont know what the color was used for.
+don�t know what the color was used for.
[Illustration: MOUSTERIAN SIDE SCRAPER]
@@ -2230,7 +2230,7 @@ might suggest some notion of hoarding up the spirits or the strength of
bears killed in the hunt. Probably the people lived in small groups,
as hunting and food-gathering seldom provide enough food for large
groups of people. These groups probably had some kind of leader or
-chief. Very likely the rude beginnings of rules for community life
+�chief.� Very likely the rude beginnings of rules for community life
and politics, and even law, were being made. But what these were, we
do not know. We can only guess about such things, as we can only guess
about many others; for example, how the idea of a family must have been
@@ -2246,8 +2246,8 @@ small. The mixtures and blendings of the habits used in making stone
tools must mean that there were also mixtures and blends in many of
the other ideas and beliefs of these small groups. And what this
probably means is that there was no one _culture_ of the time. It is
-certainly unlikely that there were simply three cultures, Acheulean,
-Levalloisian, and Mousterian, as has been thought in the past.
+certainly unlikely that there were simply three cultures, �Acheulean,�
+�Levalloisian,� and �Mousterian,� as has been thought in the past.
Rather there must have been a great variety of loosely related cultures
at about the same stage of advancement. We could say, too, that here
we really begin to see, for the first time, that remarkable ability
@@ -2272,7 +2272,7 @@ related habits for the making of tools. But the men who made them must
have looked much like the men of the West. Their tools were different,
but just as useful.
-As to what the men of the West looked like, Ive already hinted at all
+As to what the men of the West looked like, I�ve already hinted at all
we know so far (pp. 29 ff.). The Neanderthalers were present at
the time. Some more modern-like men must have been about, too, since
fossils of them have turned up at Mount Carmel in Palestine, and at
@@ -2306,7 +2306,7 @@ A NEW TRADITION APPEARS
Something new was probably beginning to happen in the
European-Mediterranean area about 40,000 years ago, though all the
rest of the Old World seems to have been going on as it had been. I
-cant be sure of this because the information we are using as a basis
+can�t be sure of this because the information we are using as a basis
for dates is very inaccurate for the areas outside of Europe and the
Mediterranean.
@@ -2325,7 +2325,7 @@ drawing shows. It has sharp cutting edges, and makes a very useful
knife. The real trick is to be able to make one. It is almost
impossible to make a blade out of any stone but flint or a natural
volcanic glass called obsidian. And even if you have flint or obsidian,
-you first have to work up a special cone-shaped blade-core, from
+you first have to work up a special cone-shaped �blade-core,� from
which to whack off blades.
[Illustration: PLAIN BLADE]
@@ -2351,8 +2351,8 @@ found in equally early cave levels in Syria; their popularity there
seems to fluctuate a bit. Some more or less parallel-sided flakes are
known in the Levalloisian industry in France, but they are probably
no earlier than Tabun E. The Tabun blades are part of a local late
-Acheulean industry, which is characterized by core-biface hand
-axes, but which has many flake tools as well. Professor F. E.
+�Acheulean� industry, which is characterized by core-biface �hand
+axes,� but which has many flake tools as well. Professor F. E.
Zeuner believes that this industry may be more than 120,000 years old;
actually its date has not yet been fixed, but it is very old--older
than the fossil finds of modern-like men in the same caves.
@@ -2371,7 +2371,7 @@ We are not sure just where the earliest _persisting_ habits for the
production of blade tools developed. Impressed by the very early
momentary appearance of blades at Tabun on Mount Carmel, Professor
Dorothy A. Garrod first favored the Near East as a center of origin.
-She spoke of some as yet unidentified Asiatic centre, which she
+She spoke of �some as yet unidentified Asiatic centre,� which she
thought might be in the highlands of Iran or just beyond. But more
recent work has been done in this area, especially by Professor Coon,
and the blade tools do not seem to have an early appearance there. When
@@ -2395,21 +2395,21 @@ core (and the striking of the Levalloisian flake from it) might have
followed through to the conical core and punch technique for the
production of blades. Professor Garrod is much impressed with the speed
of change during the later phases of the last glaciation, and its
-probable consequences. She speaks of the greater number of industries
+probable consequences. She speaks of �the greater number of industries
having enough individual character to be classified as distinct ...
-since evolution now starts to outstrip diffusion. Her evolution here
+since evolution now starts to outstrip diffusion.� Her �evolution� here
is of course an industrial evolution rather than a biological one.
Certainly the people of Europe had begun to make blade tools during
the warm spell after the first phase of the last glaciation. By about
40,000 years ago blades were well established. The bones of the blade
-tool makers weve found so far indicate that anatomically modern men
+tool makers we�ve found so far indicate that anatomically modern men
had now certainly appeared. Unfortunately, only a few fossil men have
so far been found from the very beginning of the blade tool range in
Europe (or elsewhere). What I certainly shall _not_ tell you is that
conquering bands of fine, strong, anatomically modern men, armed with
superior blade tools, came sweeping out of the East to exterminate the
-lowly Neanderthalers. Even if we dont know exactly what happened, Id
-lay a good bet it wasnt that simple.
+lowly Neanderthalers. Even if we don�t know exactly what happened, I�d
+lay a good bet it wasn�t that simple.
We do know a good deal about different blade industries in Europe.
Almost all of them come from cave layers. There is a great deal of
@@ -2418,7 +2418,7 @@ this complication; in fact, it doubtless simplifies it too much. But
it may suggest all the complication of industries which is going
on at this time. You will note that the upper portion of my much
simpler chart (p. 65) covers the same material (in the section
-marked Various Blade-Tool Industries). That chart is certainly too
+marked �Various Blade-Tool Industries�). That chart is certainly too
simplified.
You will realize that all this complication comes not only from
@@ -2429,7 +2429,7 @@ a good deal of climatic change at this time. The plants and animals
that men used for food were changing, too. The great variety of tools
and industries we now find reflect these changes and the ability of men
to keep up with the times. Now, for example, is the first time we are
-sure that there are tools to _make_ other tools. They also show mens
+sure that there are tools to _make_ other tools. They also show men�s
increasing ability to adapt themselves.
@@ -2437,15 +2437,15 @@ SPECIAL TYPES OF BLADE TOOLS
The most useful tools that appear at this time were made from blades.
- 1. The backed blade. This is a knife made of a flint blade, with
- one edge purposely blunted, probably to save the users fingers
+ 1. The �backed� blade. This is a knife made of a flint blade, with
+ one edge purposely blunted, probably to save the user�s fingers
from being cut. There are several shapes of backed blades (p.
73).
[Illustration: TWO BURINS]
- 2. The _burin_ or graver. The burin was the original chisel. Its
- cutting edge is _transverse_, like a chisels. Some burins are
+ 2. The _burin_ or �graver.� The burin was the original chisel. Its
+ cutting edge is _transverse_, like a chisel�s. Some burins are
made like a screw-driver, save that burins are sharp. Others have
edges more like the blade of a chisel or a push plane, with
only one bevel. Burins were probably used to make slots in wood
@@ -2456,29 +2456,29 @@ The most useful tools that appear at this time were made from blades.
[Illustration: TANGED POINT]
- 3. The tanged point. These stone points were used to tip arrows or
+ 3. The �tanged� point. These stone points were used to tip arrows or
light spears. They were made from blades, and they had a long tang
at the bottom where they were fixed to the shaft. At the place
where the tang met the main body of the stone point, there was
- a marked shoulder, the beginnings of a barb. Such points had
+ a marked �shoulder,� the beginnings of a barb. Such points had
either one or two shoulders.
[Illustration: NOTCHED BLADE]
- 4. The notched or strangulated blade. Along with the points for
+ 4. The �notched� or �strangulated� blade. Along with the points for
arrows or light spears must go a tool to prepare the arrow or
- spear shaft. Today, such a tool would be called a draw-knife or
- a spoke-shave, and this is what the notched blades probably are.
+ spear shaft. Today, such a tool would be called a �draw-knife� or
+ a �spoke-shave,� and this is what the notched blades probably are.
Our spoke-shaves have sharp straight cutting blades and really
- shave. Notched blades of flint probably scraped rather than cut.
+ �shave.� Notched blades of flint probably scraped rather than cut.
- 5. The awl, drill, or borer. These blade tools are worked out
+ 5. The �awl,� �drill,� or �borer.� These blade tools are worked out
to a spike-like point. They must have been used for making holes
in wood, bone, shell, skin, or other things.
[Illustration: DRILL OR AWL]
- 6. The end-scraper on a blade is a tool with one or both ends
+ 6. The �end-scraper on a blade� is a tool with one or both ends
worked so as to give a good scraping edge. It could have been used
to hollow out wood or bone, scrape hides, remove bark from trees,
and a number of other things (p. 78).
@@ -2489,11 +2489,11 @@ usually made of blades, but the best examples are so carefully worked
on both sides (bifacially) that it is impossible to see the original
blade. This tool is
- 7. The laurel leaf point. Some of these tools were long and
+ 7. The �laurel leaf� point. Some of these tools were long and
dagger-like, and must have been used as knives or daggers. Others
- were small, called willow leaf, and must have been mounted on
+ were small, called �willow leaf,� and must have been mounted on
spear or arrow shafts. Another typical Solutrean tool is the
- shouldered point. Both the laurel leaf and shouldered point
+ �shouldered� point. Both the �laurel leaf� and �shouldered� point
types are illustrated (see above and p. 79).
[Illustration: END-SCRAPER ON A BLADE]
@@ -2507,17 +2507,17 @@ second is a core tool.
[Illustration: SHOULDERED POINT]
- 8. The keel-shaped round scraper is usually small and quite round,
+ 8. The �keel-shaped round scraper� is usually small and quite round,
and has had chips removed up to a peak in the center. It is called
- keel-shaped because it is supposed to look (when upside down)
+ �keel-shaped� because it is supposed to look (when upside down)
like a section through a boat. Actually, it looks more like a tent
or an umbrella. Its outer edges are sharp all the way around, and
it was probably a general purpose scraping tool (see illustration,
p. 81).
- 9. The keel-shaped nosed scraper is a much larger and heavier tool
+ 9. The �keel-shaped nosed scraper� is a much larger and heavier tool
than the round scraper. It was made on a core with a flat bottom,
- and has one nicely worked end or nose. Such tools are usually
+ and has one nicely worked end or �nose.� Such tools are usually
large enough to be easily grasped, and probably were used like
push planes (see illustration, p. 81).
@@ -2530,7 +2530,7 @@ the most easily recognized blade tools, although they show differences
in detail at different times. There are also many other kinds. Not
all of these tools appear in any one industry at one time. Thus the
different industries shown in the chart (p. 72) each have only some
-of the blade tools weve just listed, and also a few flake tools. Some
+of the blade tools we�ve just listed, and also a few flake tools. Some
industries even have a few core tools. The particular types of blade
tools appearing in one cave layer or another, and the frequency of
appearance of the different types, tell which industry we have in each
@@ -2545,15 +2545,15 @@ to appear. There are knives, pins, needles with eyes, and little
double-pointed straight bars of bone that were probably fish-hooks. The
fish-line would have been fastened in the center of the bar; when the
fish swallowed the bait, the bar would have caught cross-wise in the
-fishs mouth.
+fish�s mouth.
One quite special kind of bone tool is a long flat point for a light
spear. It has a deep notch cut up into the breadth of its base, and is
-called a split-based bone point (p. 82). We know examples of bone
+called a �split-based bone point� (p. 82). We know examples of bone
beads from these times, and of bone handles for flint tools. Pierced
teeth of some animals were worn as beads or pendants, but I am not sure
-that elks teeth were worn this early. There are even spool-shaped
-buttons or toggles.
+that elks� teeth were worn this early. There are even spool-shaped
+�buttons� or toggles.
[Illustration: SPLIT-BASED BONE POINT]
@@ -2595,12 +2595,12 @@ almost to have served as sketch blocks. The surfaces of these various
objects may show animals, or rather abstract floral designs, or
geometric designs.
-[Illustration: VENUS FIGURINE FROM WILLENDORF]
+[Illustration: �VENUS� FIGURINE FROM WILLENDORF]
Some of the movable art is not done on tools. The most remarkable
examples of this class are little figures of women. These women seem to
be pregnant, and their most female characteristics are much emphasized.
-It is thought that these Venus or Mother-goddess figurines may be
+It is thought that these �Venus� or �Mother-goddess� figurines may be
meant to show the great forces of nature--fertility and the birth of
life.
@@ -2616,21 +2616,21 @@ are different styles in the cave art. The really great cave art is
pretty well restricted to southern France and Cantabrian (northwestern)
Spain.
-There are several interesting things about the Franco-Cantabrian cave
+There are several interesting things about the �Franco-Cantabrian� cave
art. It was done deep down in the darkest and most dangerous parts of
the caves, although the men lived only in the openings of caves. If you
think what they must have had for lights--crude lamps of hollowed stone
have been found, which must have burned some kind of oil or grease,
with a matted hair or fiber wick--and of the animals that may have
-lurked in the caves, youll understand the part about danger. Then,
-too, were sure the pictures these people painted were not simply to be
+lurked in the caves, you�ll understand the part about danger. Then,
+too, we�re sure the pictures these people painted were not simply to be
looked at and admired, for they painted one picture right over other
pictures which had been done earlier. Clearly, it was the _act_ of
_painting_ that counted. The painter had to go way down into the most
mysterious depths of the earth and create an animal in paint. Possibly
he believed that by doing this he gained some sort of magic power over
the same kind of animal when he hunted it in the open air. It certainly
-doesnt look as if he cared very much about the picture he painted--as
+doesn�t look as if he cared very much about the picture he painted--as
a finished product to be admired--for he or somebody else soon went
down and painted another animal right over the one he had done.
@@ -2683,10 +2683,10 @@ it.
Their art is another example of the direction the human mind was
taking. And when I say human, I mean it in the fullest sense, for this
is the time in which fully modern man has appeared. On page 34, we
-spoke of the Cro-Magnon group and of the Combe Capelle-Brnn group of
-Caucasoids and of the Grimaldi Negroids, who are no longer believed
+spoke of the Cro-Magnon group and of the Combe Capelle-Br�nn group of
+Caucasoids and of the Grimaldi �Negroids,� who are no longer believed
to be Negroid. I doubt that any one of these groups produced most of
-the achievements of the times. Its not yet absolutely sure which
+the achievements of the times. It�s not yet absolutely sure which
particular group produced the great cave art. The artists were almost
certainly a blend of several (no doubt already mixed) groups. The pair
of Grimaldians were buried in a grave with a sprinkling of red ochre,
@@ -2705,9 +2705,9 @@ also found about the shore of the Mediterranean basin, and it moved
into northern Europe as the last glaciation pulled northward. People
began making blade tools of very small size. They learned how to chip
very slender and tiny blades from a prepared core. Then they made these
-little blades into tiny triangles, half-moons (lunates), trapezoids,
+little blades into tiny triangles, half-moons (�lunates�), trapezoids,
and several other geometric forms. These little tools are called
-microliths. They are so small that most of them must have been fixed
+�microliths.� They are so small that most of them must have been fixed
in handles or shafts.
[Illustration: MICROLITHS
@@ -2726,7 +2726,7 @@ One corner of each little triangle stuck out, and the whole thing
made a fine barbed harpoon. In historic times in Egypt, geometric
trapezoidal microliths were still in use as arrowheads. They were
fastened--broad end out--on the end of an arrow shaft. It seems queer
-to give an arrow a point shaped like a T. Actually, the little points
+to give an arrow a point shaped like a �T.� Actually, the little points
were very sharp, and must have pierced the hides of animals very
easily. We also think that the broader cutting edge of the point may
have caused more bleeding than a pointed arrowhead would. In hunting
@@ -2739,7 +2739,7 @@ is some evidence that they appear early in the Near East. Their use
was very common in northwest Africa but this came later. The microlith
makers who reached south Russia and central Europe possibly moved up
out of the Near East. Or it may have been the other way around; we
-simply dont yet know.
+simply don�t yet know.
Remember that the microliths we are talking about here were made from
carefully prepared little blades, and are often geometric in outline.
@@ -2749,7 +2749,7 @@ even some flake scrapers, in most microlithic industries. I emphasize
this bladelet and the geometric character of the microlithic industries
of the western Old World, since there has sometimes been confusion in
the matter. Sometimes small flake chips, utilized as minute pointed
-tools, have been called microliths. They may be _microlithic_ in size
+tools, have been called �microliths.� They may be _microlithic_ in size
in terms of the general meaning of the word, but they do not seem to
belong to the sub-tradition of the blade tool preparation habits which
we have been discussing here.
@@ -2763,10 +2763,10 @@ in western Asia too, and early, although Professor Garrod is no longer
sure that the whole tradition originated in the Near East. If you look
again at my chart (p. 72) you will note that in western Asia I list
some of the names of the western European industries, but with the
-qualification -like (for example, Gravettian-like). The western
+qualification �-like� (for example, �Gravettian-like�). The western
Asiatic blade-tool industries do vaguely recall some aspects of those
of western Europe, but we would probably be better off if we used
-completely local names for them. The Emiran of my chart is such an
+completely local names for them. The �Emiran� of my chart is such an
example; its industry includes a long spike-like blade point which has
no western European counterpart.
@@ -2774,13 +2774,13 @@ When we last spoke of Africa (p. 66), I told you that stone tools
there were continuing in the Levalloisian flake tradition, and were
becoming smaller. At some time during this process, two new tool
types appeared in northern Africa: one was the Aterian point with
-a tang (p. 67), and the other was a sort of laurel leaf point,
-called the Sbaikian. These two tool types were both produced from
+a tang (p. 67), and the other was a sort of �laurel leaf� point,
+called the �Sbaikian.� These two tool types were both produced from
flakes. The Sbaikian points, especially, are roughly similar to some
of the Solutrean points of Europe. It has been suggested that both the
Sbaikian and Aterian points may be seen on their way to France through
their appearance in the Spanish cave deposits of Parpallo, but there is
-also a rival pre-Solutrean in central Europe. We still do not know
+also a rival �pre-Solutrean� in central Europe. We still do not know
whether there was any contact between the makers of these north African
tools and the Solutrean tool-makers. What does seem clear is that the
blade-tool tradition itself arrived late in northern Africa.
@@ -2788,11 +2788,11 @@ blade-tool tradition itself arrived late in northern Africa.
NETHER AFRICA
-Blade tools and laurel leaf points and some other probably late
+Blade tools and �laurel leaf� points and some other probably late
stone tool types also appear in central and southern Africa. There
are geometric microliths on bladelets and even some coarse pottery in
east Africa. There is as yet no good way of telling just where these
-items belong in time; in broad geological terms they are late.
+items belong in time; in broad geological terms they are �late.�
Some people have guessed that they are as early as similar European
and Near Eastern examples, but I doubt it. The makers of small-sized
Levalloisian flake tools occupied much of Africa until very late in
@@ -2823,18 +2823,18 @@ ancestors of the American Indians came from Asia.
The stone-tool traditions of Europe, Africa, the Near and Middle East,
and central Siberia, did _not_ move into the New World. With only a
very few special or late exceptions, there are _no_ core-bifaces,
-flakes, or blade tools of the Old World. Such things just havent been
+flakes, or blade tools of the Old World. Such things just haven�t been
found here.
-This is why I say its a shame we dont know more of the end of the
+This is why I say it�s a shame we don�t know more of the end of the
chopper-tool tradition in the Far East. According to Weidenreich,
the Mongoloids were in the Far East long before the end of the last
glaciation. If the genetics of the blood group types do demand a
non-Mongoloid ancestry for the American Indians, who else may have been
in the Far East 25,000 years ago? We know a little about the habits
for making stone tools which these first people brought with them,
-and these habits dont conform with those of the western Old World.
-Wed better keep our eyes open for whatever happened to the end of
+and these habits don�t conform with those of the western Old World.
+We�d better keep our eyes open for whatever happened to the end of
the chopper-tool tradition in northern China; already there are hints
that it lasted late there. Also we should watch future excavations
in eastern Siberia. Perhaps we shall find the chopper-tool tradition
@@ -2846,13 +2846,13 @@ THE NEW ERA
Perhaps it comes in part from the way I read the evidence and perhaps
in part it is only intuition, but I feel that the materials of this
chapter suggest a new era in the ways of life. Before about 40,000
-years ago, people simply gathered their food, wandering over large
+years ago, people simply �gathered� their food, wandering over large
areas to scavenge or to hunt in a simple sort of way. But here we
-have seen them settling-in more, perhaps restricting themselves in
+have seen them �settling-in� more, perhaps restricting themselves in
their wanderings and adapting themselves to a given locality in more
intensive ways. This intensification might be suggested by the word
-collecting. The ways of life we described in the earlier chapters
-were food-gathering ways, but now an era of food-collecting has
+�collecting.� The ways of life we described in the earlier chapters
+were �food-gathering� ways, but now an era of �food-collecting� has
begun. We shall see further intensifications of it in the next chapter.
@@ -2883,8 +2883,8 @@ The last great glaciation of the Ice Age was a two-part affair, with a
sub-phase at the end of the second part. In Europe the last sub-phase
of this glaciation commenced somewhere around 15,000 years ago. Then
the glaciers began to melt back, for the last time. Remember that
-Professor Antevs (p. 19) isnt sure the Ice Age is over yet! This
-melting sometimes went by fits and starts, and the weather wasnt
+Professor Antevs (p. 19) isn�t sure the Ice Age is over yet! This
+melting sometimes went by fits and starts, and the weather wasn�t
always changing for the better; but there was at least one time when
European weather was even better than it is now.
@@ -2927,16 +2927,16 @@ Sweden. Much of this north European material comes from bogs and swamps
where it had become water-logged and has kept very well. Thus we have
much more complete _assemblages_[4] than for any time earlier.
- [4] Assemblage is a useful word when there are different kinds of
+ [4] �Assemblage� is a useful word when there are different kinds of
archeological materials belonging together, from one area and of
- one time. An assemblage is made up of a number of industries
+ one time. An assemblage is made up of a number of �industries�
(that is, all the tools in chipped stone, all the tools in
bone, all the tools in wood, the traces of houses, etc.) and
everything else that manages to survive, such as the art, the
burials, the bones of the animals used as food, and the traces
of plant foods; in fact, everything that has been left to us
and can be used to help reconstruct the lives of the people to
- whom it once belonged. Our own present-day assemblage would be
+ whom it once belonged. Our own present-day �assemblage� would be
the sum total of all the objects in our mail-order catalogues,
department stores and supply houses of every sort, our churches,
our art galleries and other buildings, together with our roads,
@@ -2976,7 +2976,7 @@ found.
It seems likely that the Maglemosian bog finds are remains of summer
camps, and that in winter the people moved to higher and drier regions.
-Childe calls them the Forest folk; they probably lived much the
+Childe calls them the �Forest folk�; they probably lived much the
same sort of life as did our pre-agricultural Indians of the north
central states. They hunted small game or deer; they did a great deal
of fishing; they collected what plant food they could find. In fact,
@@ -3010,7 +3010,7 @@ South of the north European belt the hunting-food-collecting peoples
were living on as best they could during this time. One interesting
group, which seems to have kept to the regions of sandy soil and scrub
forest, made great quantities of geometric microliths. These are the
-materials called _Tardenoisian_. The materials of the Forest folk of
+materials called _Tardenoisian_. The materials of the �Forest folk� of
France and central Europe generally are called _Azilian_; Dr. Movius
believes the term might best be restricted to the area south of the
Loire River.
@@ -3032,24 +3032,24 @@ to it than this.
Professor Mathiassen of Copenhagen, who knows the archeological remains
of this time very well, poses a question. He speaks of the material
-as being neither rich nor progressive, in fact rather stagnant, but
-he goes on to add that the people had a certain receptiveness and
+as being neither rich nor progressive, in fact �rather stagnant,� but
+he goes on to add that the people had a certain �receptiveness� and
were able to adapt themselves quickly when the next change did come.
-My own understanding of the situation is that the Forest folk made
+My own understanding of the situation is that the �Forest folk� made
nothing as spectacular as had the producers of the earlier Magdalenian
assemblage and the Franco-Cantabrian art. On the other hand, they
_seem_ to have been making many more different kinds of tools for many
more different kinds of tasks than had their Ice Age forerunners. I
-emphasize seem because the preservation in the Maglemosian bogs
+emphasize �seem� because the preservation in the Maglemosian bogs
is very complete; certainly we cannot list anywhere near as many
different things for earlier times as we did for the Maglemosians
(p. 94). I believe this experimentation with all kinds of new tools
and gadgets, this intensification of adaptiveness (p. 91), this
-receptiveness, even if it is still only pointed toward hunting,
+�receptiveness,� even if it is still only pointed toward hunting,
fishing, and food-collecting, is an important thing.
Remember that the only marker we have handy for the _beginning_ of
-this tendency toward receptiveness and experimentation is the
+this tendency toward �receptiveness� and experimentation is the
little microlithic blade tools of various geometric forms. These, we
saw, began before the last ice had melted away, and they lasted on
in use for a very long time. I wish there were a better marker than
@@ -3063,7 +3063,7 @@ CHANGES IN OTHER AREAS?
All this last section was about Europe. How about the rest of the world
when the last glaciers were melting away?
-We simply dont know much about this particular time in other parts
+We simply don�t know much about this particular time in other parts
of the world except in Europe, the Mediterranean basin and the Middle
East. People were certainly continuing to move into the New World by
way of Siberia and the Bering Strait about this time. But for the
@@ -3075,10 +3075,10 @@ clear information.
REAL CHANGE AND PRELUDE IN THE NEAR EAST
The appearance of the microliths and the developments made by the
-Forest folk of northwestern Europe also mark an end. They show us
+�Forest folk� of northwestern Europe also mark an end. They show us
the terminal phase of the old food-collecting way of life. It grows
increasingly clear that at about the same time that the Maglemosian and
-other Forest folk were adapting themselves to hunting, fishing, and
+other �Forest folk� were adapting themselves to hunting, fishing, and
collecting in new ways to fit the post-glacial environment, something
completely new was being made ready in western Asia.
@@ -3098,7 +3098,7 @@ simply gathering or collecting it. When their food-production
became reasonably effective, people could and did settle down in
village-farming communities. With the appearance of the little farming
villages, a new way of life was actually under way. Professor Childe
-has good reason to speak of the food-producing revolution, for it was
+has good reason to speak of the �food-producing revolution,� for it was
indeed a revolution.
@@ -3117,8 +3117,8 @@ before the _how_ and _why_ answers begin to appear. Anthropologically
trained archeologists are fascinated with the cultures of men in times
of great change. About ten or twelve thousand years ago, the general
level of culture in many parts of the world seems to have been ready
-for change. In northwestern Europe, we saw that cultures changed
-just enough so that they would not have to change. We linked this to
+for change. In northwestern Europe, we saw that cultures �changed
+just enough so that they would not have to change.� We linked this to
environmental changes with the coming of post-glacial times.
In western Asia, we archeologists can prove that the food-producing
@@ -3155,7 +3155,7 @@ living as the Maglemosians did? These are the questions we still have
to face.
-CULTURAL RECEPTIVENESS AND PROMISING ENVIRONMENTS
+CULTURAL �RECEPTIVENESS� AND PROMISING ENVIRONMENTS
Until the archeologists and the natural scientists--botanists,
geologists, zoologists, and general ecologists--have spent many more
@@ -3163,15 +3163,15 @@ years on the problem, we shall not have full _how_ and _why_ answers. I
do think, however, that we are beginning to understand what to look for.
We shall have to learn much more of what makes the cultures of men
-receptive and experimental. Did change in the environment alone
-force it? Was it simply a case of Professor Toynbees challenge and
-response? I cannot believe the answer is quite that simple. Were it
-so simple, we should want to know why the change hadnt come earlier,
+�receptive� and experimental. Did change in the environment alone
+force it? Was it simply a case of Professor Toynbee�s �challenge and
+response?� I cannot believe the answer is quite that simple. Were it
+so simple, we should want to know why the change hadn�t come earlier,
along with earlier environmental changes. We shall not know the answer,
however, until we have excavated the traces of many more cultures of
the time in question. We shall doubtless also have to learn more about,
and think imaginatively about, the simpler cultures still left today.
-The mechanics of culture in general will be bound to interest us.
+The �mechanics� of culture in general will be bound to interest us.
It will also be necessary to learn much more of the environments of
10,000 to 12,000 years ago. In which regions of the world were the
@@ -3228,7 +3228,7 @@ THE OLD THEORY TOO SIMPLE FOR THE FACTS
This theory was set up before we really knew anything in detail about
the later prehistory of the Near and Middle East. We now know that
-the facts which have been found dont fit the old theory at all well.
+the facts which have been found don�t fit the old theory at all well.
Also, I have yet to find an American meteorologist who feels that we
know enough about the changes in the weather pattern to say that it can
have been so simple and direct. And, of course, the glacial ice which
@@ -3238,7 +3238,7 @@ of great alpine glaciers, and long periods of warm weather in between.
If the rain belt moved north as the glaciers melted for the last time,
it must have moved in the same direction in earlier times. Thus, the
forced neighborliness of men, plants, and animals in river valleys and
-oases must also have happened earlier. Why didnt domestication happen
+oases must also have happened earlier. Why didn�t domestication happen
earlier, then?
Furthermore, it does not seem to be in the oases and river valleys
@@ -3275,20 +3275,20 @@ archeologists, probably through habit, favor an old scheme of Grecized
names for the subdivisions: paleolithic, mesolithic, neolithic. I
refuse to use these words myself. They have meant too many different
things to too many different people and have tended to hide some pretty
-fuzzy thinking. Probably you havent even noticed my own scheme of
-subdivision up to now, but Id better tell you in general what it is.
+fuzzy thinking. Probably you haven�t even noticed my own scheme of
+subdivision up to now, but I�d better tell you in general what it is.
I think of the earliest great group of archeological materials, from
which we can deduce only a food-gathering way of culture, as the
-_food-gathering stage_. I say stage rather than age, because it
+_food-gathering stage_. I say �stage� rather than �age,� because it
is not quite over yet; there are still a few primitive people in
out-of-the-way parts of the world who remain in the _food-gathering
stage_. In fact, Professor Julian Steward would probably prefer to call
it a food-gathering _level_ of existence, rather than a stage. This
would be perfectly acceptable to me. I also tend to find myself using
_collecting_, rather than _gathering_, for the more recent aspects or
-era of the stage, as the word collecting appears to have more sense
-of purposefulness and specialization than does gathering (see p.
+era of the stage, as the word �collecting� appears to have more sense
+of purposefulness and specialization than does �gathering� (see p.
91).
Now, while I think we could make several possible subdivisions of the
@@ -3297,22 +3297,22 @@ believe the only one which means much to us here is the last or
_terminal sub-era of food-collecting_ of the whole food-gathering
stage. The microliths seem to mark its approach in the northwestern
part of the Old World. It is really shown best in the Old World by
-the materials of the Forest folk, the cultural adaptation to the
+the materials of the �Forest folk,� the cultural adaptation to the
post-glacial environment in northwestern Europe. We talked about
-the Forest folk at the beginning of this chapter, and I used the
+the �Forest folk� at the beginning of this chapter, and I used the
Maglemosian assemblage of Denmark as an example.
[5] It is difficult to find words which have a sequence or gradation
of meaning with respect to both development and a range of time
in the past, or with a range of time from somewhere in the past
which is perhaps not yet ended. One standard Webster definition
- of _stage_ is: One of the steps into which the material
- development of man ... is divided. I cannot find any dictionary
+ of _stage_ is: �One of the steps into which the material
+ development of man ... is divided.� I cannot find any dictionary
definition that suggests which of the words, _stage_ or _era_,
has the meaning of a longer span of time. Therefore, I have
chosen to let my eras be shorter, and to subdivide my stages
- into eras. Webster gives _era_ as: A signal stage of history,
- an epoch. When I want to subdivide my eras, I find myself using
+ into eras. Webster gives _era_ as: �A signal stage of history,
+ an epoch.� When I want to subdivide my eras, I find myself using
_sub-eras_. Thus I speak of the _eras_ within a _stage_ and of
the _sub-eras_ within an _era_; that is, I do so when I feel
that I really have to, and when the evidence is clear enough to
@@ -3328,9 +3328,9 @@ realms of culture. It is rather that for most of prehistoric time the
materials left to the archeologists tend to limit our deductions to
technology and economics.
-Im so soon out of my competence, as conventional ancient history
+I�m so soon out of my competence, as conventional ancient history
begins, that I shall only suggest the earlier eras of the
-food-producing stage to you. This book is about prehistory, and Im not
+food-producing stage to you. This book is about prehistory, and I�m not
a universal historian.
@@ -3339,28 +3339,28 @@ THE TWO EARLIEST ERAS OF THE FOOD-PRODUCING STAGE
The food-producing stage seems to appear in western Asia with really
revolutionary suddenness. It is seen by the relative speed with which
the traces of new crafts appear in the earliest village-farming
-community sites weve dug. It is seen by the spread and multiplication
+community sites we�ve dug. It is seen by the spread and multiplication
of these sites themselves, and the remarkable growth in human
-population we deduce from this increase in sites. Well look at some
+population we deduce from this increase in sites. We�ll look at some
of these sites and the archeological traces they yield in the next
chapter. When such village sites begin to appear, I believe we are in
the _era of the primary village-farming community_. I also believe this
is the second era of the food-producing stage.
The first era of the food-producing stage, I believe, was an _era of
-incipient cultivation and animal domestication_. I keep saying I
-believe because the actual evidence for this earlier era is so slight
+incipient cultivation and animal domestication_. I keep saying �I
+believe� because the actual evidence for this earlier era is so slight
that one has to set it up mainly by playing a hunch for it. The reason
for playing the hunch goes about as follows.
One thing we seem to be able to see, in the food-collecting era in
general, is a tendency for people to begin to settle down. This
settling down seemed to become further intensified in the terminal
-era. How this is connected with Professor Mathiassens receptiveness
+era. How this is connected with Professor Mathiassen�s �receptiveness�
and the tendency to be experimental, we do not exactly know. The
evidence from the New World comes into play here as well as that from
the Old World. With this settling down in one place, the people of the
-terminal era--especially the Forest folk whom we know best--began
+terminal era--especially the �Forest folk� whom we know best--began
making a great variety of new things. I remarked about this earlier in
the chapter. Dr. Robert M. Adams is of the opinion that this atmosphere
of experimentation with new tools--with new ways of collecting food--is
@@ -3368,9 +3368,9 @@ the kind of atmosphere in which one might expect trials at planting
and at animal domestication to have been made. We first begin to find
traces of more permanent life in outdoor camp sites, although caves
were still inhabited at the beginning of the terminal era. It is not
-surprising at all that the Forest folk had already domesticated the
+surprising at all that the �Forest folk� had already domesticated the
dog. In this sense, the whole era of food-collecting was becoming ready
-and almost incipient for cultivation and animal domestication.
+and almost �incipient� for cultivation and animal domestication.
Northwestern Europe was not the place for really effective beginnings
in agriculture and animal domestication. These would have had to take
@@ -3425,13 +3425,13 @@ zone which surrounds the drainage basin of the Tigris and Euphrates
Rivers at elevations of from approximately 2,000 to 5,000 feet. The
lower alluvial land of the Tigris-Euphrates basin itself has very
little rainfall. Some years ago Professor James Henry Breasted called
-the alluvial lands of the Tigris-Euphrates a part of the fertile
-crescent. These alluvial lands are very fertile if irrigated. Breasted
+the alluvial lands of the Tigris-Euphrates a part of the �fertile
+crescent.� These alluvial lands are very fertile if irrigated. Breasted
was most interested in the oriental civilizations of conventional
ancient history, and irrigation had been discovered before they
appeared.
-The country of hilly flanks above Breasteds crescent receives from
+The country of hilly flanks above Breasted�s crescent receives from
10 to 20 or more inches of winter rainfall each year, which is about
what Kansas has. Above the hilly-flanks zone tower the peaks and ridges
of the Lebanon-Amanus chain bordering the coast-line from Palestine
@@ -3440,7 +3440,7 @@ range of the Iraq-Iran borderland. This rugged mountain frame for our
hilly-flanks zone rises to some magnificent alpine scenery, with peaks
of from ten to fifteen thousand feet in elevation. There are several
gaps in the Mediterranean coastal portion of the frame, through which
-the winters rain-bearing winds from the sea may break so as to carry
+the winter�s rain-bearing winds from the sea may break so as to carry
rain to the foothills of the Taurus and the Zagros.
The picture I hope you will have from this description is that of an
@@ -3482,7 +3482,7 @@ hilly-flanks zone in their wild state.
With a single exception--that of the dog--the earliest positive
evidence of domestication includes the two forms of wheat, the barley,
and the goat. The evidence comes from within the hilly-flanks zone.
-However, it comes from a settled village proper, Jarmo (which Ill
+However, it comes from a settled village proper, Jarmo (which I�ll
describe in the next chapter), and is thus from the era of the primary
village-farming community. We are still without positive evidence of
domesticated grain and animals in the first era of the food-producing
@@ -3534,9 +3534,9 @@ and the spread of ideas of people who had passed on into one of the
more developed eras. In many cases, the terminal era of food-collecting
was ended by the incoming of the food-producing peoples themselves.
For example, the practices of food-production were carried into Europe
-by the actual movement of some numbers of peoples (we dont know how
+by the actual movement of some numbers of peoples (we don�t know how
many) who had reached at least the level of the primary village-farming
-community. The Forest folk learned food-production from them. There
+community. The �Forest folk� learned food-production from them. There
was never an era of incipient cultivation and domestication proper in
Europe, if my hunch is right.
@@ -3547,16 +3547,16 @@ The way I see it, two things were required in order that an era of
incipient cultivation and domestication could begin. First, there had
to be the natural environment of a nuclear area, with its whole group
of plants and animals capable of domestication. This is the aspect of
-the matter which weve said is directly given by nature. But it is
+the matter which we�ve said is directly given by nature. But it is
quite possible that such an environment with such a group of plants
and animals in it may have existed well before ten thousand years ago
in the Near East. It is also quite possible that the same promising
condition may have existed in regions which never developed into
nuclear areas proper. Here, again, we come back to the cultural factor.
-I think it was that atmosphere of experimentation weve talked about
-once or twice before. I cant define it for you, other than to say that
+I think it was that �atmosphere of experimentation� we�ve talked about
+once or twice before. I can�t define it for you, other than to say that
by the end of the Ice Age, the general level of many cultures was ready
-for change. Ask me how and why this was so, and Ill tell you we dont
+for change. Ask me how and why this was so, and I�ll tell you we don�t
know yet, and that if we did understand this kind of question, there
would be no need for me to go on being a prehistorian!
@@ -3590,7 +3590,7 @@ such collections for the modern wild forms of animals and plants from
some of our nuclear areas. In the nuclear area in the Near East, some
of the wild animals, at least, have already become extinct. There are
no longer wild cattle or wild horses in western Asia. We know they were
-there from the finds weve made in caves of late Ice Age times, and
+there from the finds we�ve made in caves of late Ice Age times, and
from some slightly later sites.
@@ -3601,7 +3601,7 @@ incipient era of cultivation and animal domestication. I am closing
this chapter with descriptions of two of the best Near Eastern examples
I know of. You may not be satisfied that what I am able to describe
makes a full-bodied era of development at all. Remember, however, that
-Ive told you Im largely playing a kind of a hunch, and also that the
+I�ve told you I�m largely playing a kind of a hunch, and also that the
archeological materials of this era will always be extremely difficult
to interpret. At the beginning of any new way of life, there will be a
great tendency for people to make-do, at first, with tools and habits
@@ -3613,7 +3613,7 @@ THE NATUFIAN, AN ASSEMBLAGE OF THE INCIPIENT ERA
The assemblage called the Natufian comes from the upper layers of a
number of caves in Palestine. Traces of its flint industry have also
-turned up in Syria and Lebanon. We dont know just how old it is. I
+turned up in Syria and Lebanon. We don�t know just how old it is. I
guess that it probably falls within five hundred years either way of
about 5000 B.C.
@@ -3662,7 +3662,7 @@ pendants. There were also beads and pendants of pierced teeth and shell.
A number of Natufian burials have been found in the caves; some burials
were grouped together in one grave. The people who were buried within
the Mount Carmel cave were laid on their backs in an extended position,
-while those on the terrace seem to have been flexed (placed in their
+while those on the terrace seem to have been �flexed� (placed in their
graves in a curled-up position). This may mean no more than that it was
easier to dig a long hole in cave dirt than in the hard-packed dirt of
the terrace. The people often had some kind of object buried with them,
@@ -3679,7 +3679,7 @@ beads.
GROUND STONE
BONE]
-The animal bones of the Natufian layers show beasts of a modern type,
+The animal bones of the Natufian layers show beasts of a �modern� type,
but with some differences from those of present-day Palestine. The
bones of the gazelle far outnumber those of the deer; since gazelles
like a much drier climate than deer, Palestine must then have had much
@@ -3692,9 +3692,9 @@ Maglemosian of northern Europe. More recently, it has been reported
that a domesticated goat is also part of the Natufian finds.
The study of the human bones from the Natufian burials is not yet
-complete. Until Professor McCowns study becomes available, we may note
-Professor Coons assessment that these people were of a basically
-Mediterranean type.
+complete. Until Professor McCown�s study becomes available, we may note
+Professor Coon�s assessment that these people were of a �basically
+Mediterranean type.�
THE KARIM SHAHIR ASSEMBLAGE
@@ -3704,11 +3704,11 @@ of a temporary open site or encampment. It lies on the top of a bluff
in the Kurdish hill-country of northeastern Iraq. It was dug by Dr.
Bruce Howe of the expedition I directed in 1950-51 for the Oriental
Institute and the American Schools of Oriental Research. In 1954-55,
-our expedition located another site, Mlefaat, with general resemblance
+our expedition located another site, M�lefaat, with general resemblance
to Karim Shahir, but about a hundred miles north of it. In 1956, Dr.
Ralph Solecki located still another Karim Shahir type of site called
Zawi Chemi Shanidar. The Zawi Chemi site has a radiocarbon date of 8900
- 300 B.C.
+� 300 B.C.
Karim Shahir has evidence of only one very shallow level of occupation.
It was probably not lived on very long, although the people who lived
@@ -3717,7 +3717,7 @@ layer yielded great numbers of fist-sized cracked pieces of limestone,
which had been carried up from the bed of a stream at the bottom of the
bluff. We think these cracked stones had something to do with a kind of
architecture, but we were unable to find positive traces of hut plans.
-At Mlefaat and Zawi Chemi, there were traces of rounded hut plans.
+At M�lefaat and Zawi Chemi, there were traces of rounded hut plans.
As in the Natufian, the great bulk of small objects of the Karim Shahir
assemblage was in chipped flint. A large proportion of the flint tools
@@ -3737,7 +3737,7 @@ clay figurines which seemed to be of animal form.
UNBAKED CLAY
SHELL
BONE
- ARCHITECTURE]
+ �ARCHITECTURE�]
Karim Shahir did not yield direct evidence of the kind of vegetable
food its people ate. The animal bones showed a considerable
@@ -3746,7 +3746,7 @@ domestication--sheep, goat, cattle, horse, dog--as compared with animal
bones from the earlier cave sites of the area, which have a high
proportion of bones of wild forms like deer and gazelle. But we do not
know that any of the Karim Shahir animals were actually domesticated.
-Some of them may have been, in an incipient way, but we have no means
+Some of them may have been, in an �incipient� way, but we have no means
at the moment that will tell us from the bones alone.
@@ -3761,7 +3761,7 @@ goat, and the general animal situation at Karim Shahir to hint at an
incipient approach to food-production. At Karim Shahir, there was the
tendency to settle down out in the open; this is echoed by the new
reports of open air Natufian sites. The large number of cracked stones
-certainly indicates that it was worth the peoples while to have some
+certainly indicates that it was worth the peoples� while to have some
kind of structure, even if the site as a whole was short-lived.
It is a part of my hunch that these things all point toward
@@ -3771,13 +3771,13 @@ which we shall look at next, are fully food-producing, the Natufian
and Karim Shahir folk had not yet arrived. I think they were part of
a general build-up to full scale food-production. They were possibly
controlling a few animals of several kinds and perhaps one or two
-plants, without realizing the full possibilities of this control as a
+plants, without realizing the full possibilities of this �control� as a
new way of life.
This is why I think of the Karim Shahir and Natufian folk as being at
a level, or in an era, of incipient cultivation and domestication. But
we shall have to do a great deal more excavation in this range of time
-before well get the kind of positive information we need.
+before we�ll get the kind of positive information we need.
SUMMARY
@@ -3798,7 +3798,7 @@ history.
We know the earliest village-farming communities appeared in western
Asia, in a nuclear area. We do not yet know why the Near Eastern
-experiment came first, or why it didnt happen earlier in some other
+experiment came first, or why it didn�t happen earlier in some other
nuclear area. Apparently, the level of culture and the promise of the
natural environment were ready first in western Asia. The next sites
we look at will show a simple but effective food-production already
@@ -3835,7 +3835,7 @@ contrast between food-collecting and food-producing as ways of life.
THE DIFFERENCE BETWEEN FOOD-COLLECTORS AND FOOD-PRODUCERS
-Childe used the word revolution because of the radical change that
+Childe used the word �revolution� because of the radical change that
took place in the habits and customs of man. Food-collectors--that is,
hunters, fishers, berry- and nut-gatherers--had to live in small groups
or bands, for they had to be ready to move wherever their food supply
@@ -3851,7 +3851,7 @@ for clothing beyond the tools that were probably used to dress the
skins of animals; no time to think of much of anything but food and
protection and disposal of the dead when death did come: an existence
which takes nature as it finds it, which does little or nothing to
-modify nature--all in all, a savages existence, and a very tough one.
+modify nature--all in all, a savage�s existence, and a very tough one.
A man who spends his whole life following animals just to kill them to
eat, or moving from one berry patch to another, is really living just
like an animal himself.
@@ -3859,10 +3859,10 @@ like an animal himself.
THE FOOD-PRODUCING ECONOMY
-Against this picture let me try to draw another--that of mans life
-after food-production had begun. His meat was stored on the hoof,
+Against this picture let me try to draw another--that of man�s life
+after food-production had begun. His meat was stored �on the hoof,�
his grain in silos or great pottery jars. He lived in a house: it was
-worth his while to build one, because he couldnt move far from his
+worth his while to build one, because he couldn�t move far from his
fields and flocks. In his neighborhood enough food could be grown
and enough animals bred so that many people were kept busy. They all
lived close to their flocks and fields, in a village. The village was
@@ -3872,7 +3872,7 @@ Children and old men could shepherd the animals by day or help with
the lighter work in the fields. After the crops had been harvested the
younger men might go hunting and some of them would fish, but the food
they brought in was only an addition to the food in the village; the
-villagers wouldnt starve, even if the hunters and fishermen came home
+villagers wouldn�t starve, even if the hunters and fishermen came home
empty-handed.
There was more time to do different things, too. They began to modify
@@ -3885,23 +3885,23 @@ people in the village who were becoming full-time craftsmen.
Other things were changing, too. The villagers must have had
to agree on new rules for living together. The head man of the
village had problems different from those of the chief of the small
-food-collectors band. If somebodys flock of sheep spoiled a wheat
+food-collectors� band. If somebody�s flock of sheep spoiled a wheat
field, the owner wanted payment for the grain he lost. The chief of
the hunters was never bothered with such questions. Even the gods
had changed. The spirits and the magic that had been used by hunters
-werent of any use to the villagers. They needed gods who would watch
+weren�t of any use to the villagers. They needed gods who would watch
over the fields and the flocks, and they eventually began to erect
buildings where their gods might dwell, and where the men who knew most
about the gods might live.
-WAS FOOD-PRODUCTION A REVOLUTION?
+WAS FOOD-PRODUCTION A �REVOLUTION�?
If you can see the difference between these two pictures--between
life in the food-collecting stage and life after food-production
-had begun--youll see why Professor Childe speaks of a revolution.
-By revolution, he doesnt mean that it happened over night or that
-it happened only once. We dont know exactly how long it took. Some
+had begun--you�ll see why Professor Childe speaks of a revolution.
+By revolution, he doesn�t mean that it happened over night or that
+it happened only once. We don�t know exactly how long it took. Some
people think that all these changes may have occurred in less than
500 years, but I doubt that. The incipient era was probably an affair
of some duration. Once the level of the village-farming community had
@@ -3915,7 +3915,7 @@ been achieved with truly revolutionary suddenness.
GAPS IN OUR KNOWLEDGE OF THE NEAR EAST
-If youll look again at the chart (p. 111) youll see that I have
+If you�ll look again at the chart (p. 111) you�ll see that I have
very few sites and assemblages to name in the incipient era of
cultivation and domestication, and not many in the earlier part of
the primary village-farming level either. Thanks in no small part
@@ -3926,20 +3926,20 @@ yard-stick here. But I am far from being able to show you a series of
Sears Roebuck catalogues, even century by century, for any part of
the nuclear area. There is still a great deal of earth to move, and a
great mass of material to recover and interpret before we even begin to
-understand how and why.
+understand �how� and �why.�
Perhaps here, because this kind of archeology is really my specialty,
-youll excuse it if I become personal for a moment. I very much look
+you�ll excuse it if I become personal for a moment. I very much look
forward to having further part in closing some of the gaps in knowledge
-of the Near East. This is not, as Ive told you, the spectacular
+of the Near East. This is not, as I�ve told you, the spectacular
range of Near Eastern archeology. There are no royal tombs, no gold,
no great buildings or sculpture, no writing, in fact nothing to
excite the normal museum at all. Nevertheless it is a range which,
idea-wise, gives the archeologist tremendous satisfaction. The country
of the hilly flanks is an exciting combination of green grasslands
and mountainous ridges. The Kurds, who inhabit the part of the area
-in which Ive worked most recently, are an extremely interesting and
-hospitable people. Archeologists dont become rich, but Ill forego
+in which I�ve worked most recently, are an extremely interesting and
+hospitable people. Archeologists don�t become rich, but I�ll forego
the Cadillac for any bright spring morning in the Kurdish hills, on a
good site with a happy crew of workmen and an interested and efficient
staff. It is probably impossible to convey the full feeling which life
@@ -3965,15 +3965,15 @@ like the use of pottery borrowed from the more developed era of the
same time in the nuclear area. The same general explanation doubtless
holds true for certain materials in Egypt, along the upper Nile and in
the Kharga oasis: these materials, called Sebilian III, the Khartoum
-neolithic, and the Khargan microlithic, are from surface sites,
+�neolithic,� and the Khargan microlithic, are from surface sites,
not from caves. The chart (p. 111) shows where I would place these
materials in era and time.
[Illustration: THE HILLY FLANKS OF THE CRESCENT AND EARLY SITES OF THE
NEAR EAST]
-Both Mlefaat and Dr. Soleckis Zawi Chemi Shanidar site appear to have
-been slightly more settled in than was Karim Shahir itself. But I do
+Both M�lefaat and Dr. Solecki�s Zawi Chemi Shanidar site appear to have
+been slightly more �settled in� than was Karim Shahir itself. But I do
not think they belong to the era of farming-villages proper. The first
site of this era, in the hills of Iraqi Kurdistan, is Jarmo, on which
we have spent three seasons of work. Following Jarmo comes a variety of
@@ -3989,9 +3989,9 @@ times when their various cultures flourished, there must have been
many little villages which shared the same general assemblage. We are
only now beginning to locate them again. Thus, if I speak of Jarmo,
or Jericho, or Sialk as single examples of their particular kinds of
-assemblages, I dont mean that they were unique at all. I think I could
+assemblages, I don�t mean that they were unique at all. I think I could
take you to the sites of at least three more Jarmos, within twenty
-miles of the original one. They are there, but they simply havent yet
+miles of the original one. They are there, but they simply haven�t yet
been excavated. In 1956, a Danish expedition discovered material of
Jarmo type at Shimshara, only two dozen miles northeast of Jarmo, and
below an assemblage of Hassunan type (which I shall describe presently).
@@ -4000,15 +4000,15 @@ below an assemblage of Hassunan type (which I shall describe presently).
THE GAP BETWEEN KARIM SHAHIR AND JARMO
As we see the matter now, there is probably still a gap in the
-available archeological record between the Karim Shahir-Mlefaat-Zawi
+available archeological record between the Karim Shahir-M�lefaat-Zawi
Chemi group (of the incipient era) and that of Jarmo (of the
village-farming era). Although some items of the Jarmo type materials
do reflect the beginnings of traditions set in the Karim Shahir group
(see p. 120), there is not a clear continuity. Moreover--to the
degree that we may trust a few radiocarbon dates--there would appear
to be around two thousand years of difference in time. The single
-available Zawi Chemi date is 8900 300 B.C.; the most reasonable
-group of dates from Jarmo average to about 6750 200 B.C. I am
+available Zawi Chemi �date� is 8900 � 300 B.C.; the most reasonable
+group of �dates� from Jarmo average to about 6750 � 200 B.C. I am
uncertain about this two thousand years--I do not think it can have
been so long.
@@ -4021,7 +4021,7 @@ JARMO, IN THE KURDISH HILLS, IRAQ
The site of Jarmo has a depth of deposit of about twenty-seven feet,
and approximately a dozen layers of architectural renovation and
-change. Nevertheless it is a one period site: its assemblage remains
+change. Nevertheless it is a �one period� site: its assemblage remains
essentially the same throughout, although one or two new items are
added in later levels. It covers about four acres of the top of a
bluff, below which runs a small stream. Jarmo lies in the hill country
@@ -4078,7 +4078,7 @@ human beings in clay; one type of human figurine they favored was that
of a markedly pregnant woman, probably the expression of some sort of
fertility spirit. They provided their house floors with baked-in-place
depressions, either as basins or hearths, and later with domed ovens of
-clay. As weve noted, the houses themselves were of clay or mud; one
+clay. As we�ve noted, the houses themselves were of clay or mud; one
could almost say they were built up like a house-sized pot. Then,
finally, the idea of making portable pottery itself appeared, although
I very much doubt that the people of the Jarmo village discovered the
@@ -4095,11 +4095,11 @@ over three hundred miles to the north. Already a bulk carrying trade
had been established--the forerunner of commerce--and the routes were
set by which, in later times, the metal trade was to move.
-There are now twelve radioactive carbon dates from Jarmo. The most
-reasonable cluster of determinations averages to about 6750 200
-B.C., although there is a completely unreasonable range of dates
+There are now twelve radioactive carbon �dates� from Jarmo. The most
+reasonable cluster of determinations averages to about 6750 � 200
+B.C., although there is a completely unreasonable range of �dates�
running from 3250 to 9250 B.C.! _If_ I am right in what I take to be
-reasonable, the first flush of the food-producing revolution had been
+�reasonable,� the first flush of the food-producing revolution had been
achieved almost nine thousand years ago.
@@ -4117,7 +4117,7 @@ it, but the Hassunan sites seem to cluster at slightly lower elevations
than those we have been talking about so far.
The catalogue of the Hassuna assemblage is of course more full and
-elaborate than that of Jarmo. The Iraqi governments archeologists
+elaborate than that of Jarmo. The Iraqi government�s archeologists
who dug Hassuna itself, exposed evidence of increasing architectural
know-how. The walls of houses were still formed of puddled mud;
sun-dried bricks appear only in later periods. There were now several
@@ -4130,16 +4130,16 @@ largely disappeared by Hassunan times. The flint work of the Hassunan
catalogue is, by and large, a wretched affair. We might guess that the
kinaesthetic concentration of the Hassuna craftsmen now went into other
categories; that is, they suddenly discovered they might have more fun
-working with the newer materials. Its a shame, for example, that none
+working with the newer materials. It�s a shame, for example, that none
of their weaving is preserved for us.
The two available radiocarbon determinations from Hassunan contexts
-stand at about 5100 and 5600 B.C. 250 years.
+stand at about 5100 and 5600 B.C. � 250 years.
OTHER EARLY VILLAGE SITES IN THE NUCLEAR AREA
-Ill now name and very briefly describe a few of the other early
+I�ll now name and very briefly describe a few of the other early
village assemblages either in or adjacent to the hilly flanks of the
crescent. Unfortunately, we do not have radioactive carbon dates for
many of these materials. We may guess that some particular assemblage,
@@ -4177,7 +4177,7 @@ ecological niche, some seven hundred feet below sea level; it is
geographically within the hilly-flanks zone but environmentally not
part of it.
-Several radiocarbon dates for Jericho fall within the range of those
+Several radiocarbon �dates� for Jericho fall within the range of those
I find reasonable for Jarmo, and their internal statistical consistency
is far better than that for the Jarmo determinations. It is not yet
clear exactly what this means.
@@ -4226,7 +4226,7 @@ how things were made are different; the Sialk assemblage represents
still another cultural pattern. I suspect it appeared a bit later
in time than did that of Hassuna. There is an important new item in
the Sialk catalogue. The Sialk people made small drills or pins of
-hammered copper. Thus the metallurgists specialized craft had made its
+hammered copper. Thus the metallurgist�s specialized craft had made its
appearance.
There is at least one very early Iranian site on the inward slopes
@@ -4246,7 +4246,7 @@ shore of the Fayum lake. The Fayum materials come mainly from grain
bins or silos. Another site, Merimde, in the western part of the Nile
delta, shows the remains of a true village, but it may be slightly
later than the settlement of the Fayum. There are radioactive carbon
-dates for the Fayum materials at about 4275 B.C. 320 years, which
+�dates� for the Fayum materials at about 4275 B.C. � 320 years, which
is almost fifteen hundred years later than the determinations suggested
for the Hassunan or Syro-Cilician assemblages. I suspect that this
is a somewhat over-extended indication of the time it took for the
@@ -4260,13 +4260,13 @@ the mound called Shaheinab. The Shaheinab catalogue roughly corresponds
to that of the Fayum; the distance between the two places, as the Nile
flows, is roughly 1,500 miles. Thus it took almost a thousand years for
the new way of life to be carried as far south into Africa as Khartoum;
-the two Shaheinab dates average about 3300 B.C. 400 years.
+the two Shaheinab �dates� average about 3300 B.C. � 400 years.
If the movement was up the Nile (southward), as these dates suggest,
then I suspect that the earliest available village material of middle
Egypt, the so-called Tasian, is also later than that of the Fayum. The
Tasian materials come from a few graves near a village called Deir
-Tasa, and I have an uncomfortable feeling that the Tasian assemblage
+Tasa, and I have an uncomfortable feeling that the Tasian �assemblage�
may be mainly an artificial selection of poor examples of objects which
belong in the following range of time.
@@ -4280,7 +4280,7 @@ spread outward in space from the nuclear area, as time went on. There
is good archeological evidence that both these processes took place.
For the hill country of northeastern Iraq, in the nuclear area, we
have already noticed how the succession (still with gaps) from Karim
-Shahir, through Mlefaat and Jarmo, to Hassuna can be charted (see
+Shahir, through M�lefaat and Jarmo, to Hassuna can be charted (see
chart, p. 111). In the next chapter, we shall continue this charting
and description of what happened in Iraq upward through time. We also
watched traces of the new way of life move through space up the Nile
@@ -4299,7 +4299,7 @@ appearance of the village-farming community there--is still an open
one. In the last chapter, we noted the probability of an independent
nuclear area in southeastern Asia. Professor Carl Sauer strongly
champions the great importance of this area as _the_ original center
-of agricultural pursuits, as a kind of cradle of all incipient eras
+of agricultural pursuits, as a kind of �cradle� of all incipient eras
of the Old World at least. While there is certainly not the slightest
archeological evidence to allow us to go that far, we may easily expect
that an early southeast Asian development would have been felt in
@@ -4311,13 +4311,13 @@ way of life moved well beyond Khartoum in Africa.
THE SPREAD OF THE VILLAGE-FARMING COMMUNITY WAY OF LIFE INTO EUROPE
-How about Europe? I wont give you many details. You can easily imagine
+How about Europe? I won�t give you many details. You can easily imagine
that the late prehistoric prelude to European history is a complicated
affair. We all know very well how complicated an area Europe is now,
with its welter of different languages and cultures. Remember, however,
that a great deal of archeology has been done on the late prehistory of
Europe, and very little on that of further Asia and Africa. If we knew
-as much about these areas as we do of Europe, I expect wed find them
+as much about these areas as we do of Europe, I expect we�d find them
just as complicated.
This much is clear for Europe, as far as the spread of the
@@ -4329,21 +4329,21 @@ in western Asia. I do not, of course, mean that there were traveling
salesmen who carried these ideas and things to Europe with a commercial
gleam in their eyes. The process took time, and the ideas and things
must have been passed on from one group of people to the next. There
-was also some actual movement of peoples, but we dont know the size of
+was also some actual movement of peoples, but we don�t know the size of
the groups that moved.
-The story of the colonization of Europe by the first farmers is
+The story of the �colonization� of Europe by the first farmers is
thus one of (1) the movement from the eastern Mediterranean lands
of some people who were farmers; (2) the spread of ideas and things
beyond the Near East itself and beyond the paths along which the
-colonists moved; and (3) the adaptations of the ideas and things
-by the indigenous Forest folk, about whose receptiveness Professor
+�colonists� moved; and (3) the adaptations of the ideas and things
+by the indigenous �Forest folk�, about whose �receptiveness� Professor
Mathiassen speaks (p. 97). It is important to note that the resulting
cultures in the new European environment were European, not Near
-Eastern. The late Professor Childe remarked that the peoples of the
+Eastern. The late Professor Childe remarked that �the peoples of the
West were not slavish imitators; they adapted the gifts from the East
... into a new and organic whole capable of developing on its own
-original lines.
+original lines.�
THE WAYS TO EUROPE
@@ -4389,19 +4389,19 @@ Hill, the earliest known trace of village-farming communities in
England, is about 2500 B.C. I would expect about 5500 B.C. to be a
safe date to give for the well-developed early village communities of
Syro-Cilicia. We suspect that the spread throughout Europe did not
-proceed at an even rate. Professor Piggott writes that at a date
+proceed at an even rate. Professor Piggott writes that �at a date
probably about 2600 B.C., simple agricultural communities were being
established in Spain and southern France, and from the latter region a
spread northwards can be traced ... from points on the French seaboard
of the [English] Channel ... there were emigrations of a certain number
of these tribes by boat, across to the chalk lands of Wessex and Sussex
[in England], probably not more than three or four generations later
-than the formation of the south French colonies.
+than the formation of the south French colonies.�
New radiocarbon determinations are becoming available all the
time--already several suggest that the food-producing way of life
had reached the lower Rhine and Holland by 4000 B.C. But not all
-prehistorians accept these dates, so I do not show them on my map
+prehistorians accept these �dates,� so I do not show them on my map
(p. 139).
@@ -4427,7 +4427,7 @@ concentric sets of banks and ditches. Traces of oblong timber houses
have been found, but not within the enclosures. The second type of
structure is mine-shafts, dug down into the chalk beds where good
flint for the making of axes or hoes could be found. The third type
-of structure is long simple mounds or unchambered barrows, in one
+of structure is long simple mounds or �unchambered barrows,� in one
end of which burials were made. It has been commonly believed that the
Windmill Hill assemblage belonged entirely to the cultural tradition
which moved up through France to the Channel. Professor Piggott is now
@@ -4443,12 +4443,12 @@ consists mainly of tombs and the contents of tombs, with only very
rare settlement sites. The tombs were of some size and received the
bodies of many people. The tombs themselves were built of stone, heaped
over with earth; the stones enclosed a passage to a central chamber
-(passage graves), or to a simple long gallery, along the sides of
-which the bodies were laid (gallery graves). The general type of
-construction is called megalithic (= great stone), and the whole
+(�passage graves�), or to a simple long gallery, along the sides of
+which the bodies were laid (�gallery graves�). The general type of
+construction is called �megalithic� (= great stone), and the whole
earth-mounded structure is often called a _barrow_. Since many have
-proper chambers, in one sense or another, we used the term unchambered
-barrow above to distinguish those of the Windmill Hill type from these
+proper chambers, in one sense or another, we used the term �unchambered
+barrow� above to distinguish those of the Windmill Hill type from these
megalithic structures. There is some evidence for sacrifice, libations,
and ceremonial fires, and it is clear that some form of community
ritual was focused on the megalithic tombs.
@@ -4466,7 +4466,7 @@ The third early British group of antiquities of this general time
It is not so certain that the people who made this assemblage, called
Peterborough, were actually farmers. While they may on occasion have
practiced a simple agriculture, many items of their assemblage link
-them closely with that of the Forest folk of earlier times in
+them closely with that of the �Forest folk� of earlier times in
England and in the Baltic countries. Their pottery is decorated with
impressions of cords and is quite different from that of Windmill Hill
and the megalithic builders. In addition, the distribution of their
@@ -4479,7 +4479,7 @@ to acquire the raw material for stone axes.
A probably slightly later culture, whose traces are best known from
Skara Brae on Orkney, also had its roots in those cultures of the
-Baltic area which fused out of the meeting of the Forest folk and
+Baltic area which fused out of the meeting of the �Forest folk� and
the peoples who took the eastern way into Europe. Skara Brae is very
well preserved, having been built of thin stone slabs about which
dune-sand drifted after the village died. The individual houses, the
@@ -4498,14 +4498,14 @@ details which I have omitted in order to shorten the story.
I believe some of the difficulty we have in understanding the
establishment of the first farming communities in Europe is with
-the word colonization. We have a natural tendency to think of
-colonization as it has happened within the last few centuries. In the
+the word �colonization.� We have a natural tendency to think of
+�colonization� as it has happened within the last few centuries. In the
case of the colonization of the Americas, for example, the colonists
came relatively quickly, and in increasingly vast numbers. They had
vastly superior technical, political, and war-making skills, compared
with those of the Indians. There was not much mixing with the Indians.
The case in Europe five or six thousand years ago must have been very
-different. I wonder if it is even proper to call people colonists
+different. I wonder if it is even proper to call people �colonists�
who move some miles to a new region, settle down and farm it for some
years, then move on again, generation after generation? The ideas and
the things which these new people carried were only _potentially_
@@ -4521,12 +4521,12 @@ migrants were moving by boat, long distances may have been covered in
a short time. Remember, however, we seem to have about three thousand
years between the early Syro-Cilician villages and Windmill Hill.
-Let me repeat Professor Childe again. The peoples of the West were
+Let me repeat Professor Childe again. �The peoples of the West were
not slavish imitators: they adapted the gifts from the East ... into
a new and organic whole capable of developing on its own original
-lines. Childe is of course completely conscious of the fact that his
-peoples of the West were in part the descendants of migrants who came
-originally from the East, bringing their gifts with them. This
+lines.� Childe is of course completely conscious of the fact that his
+�peoples of the West� were in part the descendants of migrants who came
+originally from the �East,� bringing their �gifts� with them. This
was the late prehistoric achievement of Europe--to take new ideas and
things and some migrant peoples and, by mixing them with the old in its
own environments, to forge a new and unique series of cultures.
@@ -4553,14 +4553,14 @@ things first happened there and also because I know it best.
There is another interesting thing, too. We have seen that the first
experiment in village-farming took place in the Near East. So did
-the first experiment in civilization. Both experiments took. The
+the first experiment in civilization. Both experiments �took.� The
traditions we live by today are based, ultimately, on those ancient
beginnings in food-production and civilization in the Near East.
-WHAT CIVILIZATION MEANS
+WHAT �CIVILIZATION� MEANS
-I shall not try to define civilization for you; rather, I shall
+I shall not try to define �civilization� for you; rather, I shall
tell you what the word brings to my mind. To me civilization means
urbanization: the fact that there are cities. It means a formal
political set-up--that there are kings or governing bodies that the
@@ -4606,7 +4606,7 @@ of Mexico, the Mayas of Yucatan and Guatemala, and the Incas of the
Andes were civilized.
-WHY DIDNT CIVILIZATION COME TO ALL FOOD-PRODUCERS?
+WHY DIDN�T CIVILIZATION COME TO ALL FOOD-PRODUCERS?
Once you have food-production, even at the well-advanced level of
the village-farming community, what else has to happen before you
@@ -4625,13 +4625,13 @@ early civilization, is still an open and very interesting question.
WHERE CIVILIZATION FIRST APPEARED IN THE NEAR EAST
You remember that our earliest village-farming communities lay along
-the hilly flanks of a great crescent. (See map on p. 125.)
-Professor Breasteds fertile crescent emphasized the rich river
+the hilly flanks of a great �crescent.� (See map on p. 125.)
+Professor Breasted�s �fertile crescent� emphasized the rich river
valleys of the Nile and the Tigris-Euphrates Rivers. Our hilly-flanks
area of the crescent zone arches up from Egypt through Palestine and
Syria, along southern Turkey into northern Iraq, and down along the
southwestern fringe of Iran. The earliest food-producing villages we
-know already existed in this area by about 6750 B.C. ( 200 years).
+know already existed in this area by about 6750 B.C. (� 200 years).
Now notice that this hilly-flanks zone does not include southern
Mesopotamia, the alluvial land of the lower Tigris and Euphrates in
@@ -4639,7 +4639,7 @@ Iraq, or the Nile Valley proper. The earliest known villages of classic
Mesopotamia and Egypt seem to appear fifteen hundred or more years
after those of the hilly-flanks zone. For example, the early Fayum
village which lies near a lake west of the Nile Valley proper (see p.
-135) has a radiocarbon date of 4275 B.C. 320 years. It was in the
+135) has a radiocarbon date of 4275 B.C. � 320 years. It was in the
river lands, however, that the immediate beginnings of civilization
were made.
@@ -4657,8 +4657,8 @@ THE HILLY-FLANKS ZONE VERSUS THE RIVER LANDS
Why did these two civilizations spring up in these two river
lands which apparently were not even part of the area where the
-village-farming community began? Why didnt we have the first
-civilizations in Palestine, Syria, north Iraq, or Iran, where were
+village-farming community began? Why didn�t we have the first
+civilizations in Palestine, Syria, north Iraq, or Iran, where we�re
sure food-production had had a long time to develop? I think the
probable answer gives a clue to the ways in which civilization began in
Egypt and Mesopotamia.
@@ -4669,7 +4669,7 @@ and Syria. There are pleasant mountain slopes, streams running out to
the sea, and rain, at least in the winter months. The rain belt and the
foothills of the Turkish mountains also extend to northern Iraq and on
to the Iranian plateau. The Iranian plateau has its mountain valleys,
-streams, and some rain. These hilly flanks of the crescent, through
+streams, and some rain. These hilly flanks of the �crescent,� through
most of its arc, are almost made-to-order for beginning farmers. The
grassy slopes of the higher hills would be pasture for their herds
and flocks. As soon as the earliest experiments with agriculture and
@@ -4720,10 +4720,10 @@ Obviously, we can no longer find the first dikes or reservoirs of
the Nile Valley, or the first canals or ditches of Mesopotamia. The
same land has been lived on far too long for any traces of the first
attempts to be left; or, especially in Egypt, it has been covered by
-the yearly deposits of silt, dropped by the river floods. But were
+the yearly deposits of silt, dropped by the river floods. But we�re
pretty sure the first food-producers of Egypt and southern Mesopotamia
must have made such dikes, canals, and ditches. In the first place,
-there cant have been enough rain for them to grow things otherwise.
+there can�t have been enough rain for them to grow things otherwise.
In the second place, the patterns for such projects seem to have been
pretty well set by historic times.
@@ -4733,10 +4733,10 @@ CONTROL OF THE RIVERS THE BUSINESS OF EVERYONE
Here, then, is a _part_ of the reason why civilization grew in Egypt
and Mesopotamia first--not in Palestine, Syria, or Iran. In the latter
areas, people could manage to produce their food as individuals. It
-wasnt too hard; there were rain and some streams, and good pasturage
+wasn�t too hard; there were rain and some streams, and good pasturage
for the animals even if a crop or two went wrong. In Egypt and
Mesopotamia, people had to put in a much greater amount of work, and
-this work couldnt be individual work. Whole villages or groups of
+this work couldn�t be individual work. Whole villages or groups of
people had to turn out to fix dikes or dig ditches. The dikes had to be
repaired and the ditches carefully cleared of silt each year, or they
would become useless.
@@ -4745,7 +4745,7 @@ There also had to be hard and fast rules. The person who lived nearest
the ditch or the reservoir must not be allowed to take all the water
and leave none for his neighbors. It was not only a business of
learning to control the rivers and of making their waters do the
-farmers work. It also meant controlling men. But once these men had
+farmer�s work. It also meant controlling men. But once these men had
managed both kinds of controls, what a wonderful yield they had! The
soil was already fertile, and the silt which came in the floods and
ditches kept adding fertile soil.
@@ -4756,7 +4756,7 @@ THE GERM OF CIVILIZATION IN EGYPT AND MESOPOTAMIA
This learning to work together for the common good was the real germ of
the Egyptian and the Mesopotamian civilizations. The bare elements of
civilization were already there: the need for a governing hand and for
-laws to see that the communities work was done and that the water was
+laws to see that the communities� work was done and that the water was
justly shared. You may object that there is a sort of chicken and egg
paradox in this idea. How could the people set up the rules until they
had managed to get a way to live, and how could they manage to get a
@@ -4781,12 +4781,12 @@ My explanation has been pointed particularly at Egypt and Mesopotamia.
I have already told you that the irrigation and water-control part of
it does not apply to the development of the Aztecs or the Mayas, or
perhaps anybody else. But I think that a fair part of the story of
-Egypt and Mesopotamia must be as Ive just told you.
+Egypt and Mesopotamia must be as I�ve just told you.
I am particularly anxious that you do _not_ understand me to mean that
irrigation _caused_ civilization. I am sure it was not that simple at
all. For, in fact, a complex and highly engineered irrigation system
-proper did not come until later times. Lets say rather that the simple
+proper did not come until later times. Let�s say rather that the simple
beginnings of irrigation allowed and in fact encouraged a great number
of things in the technological, political, social, and moral realms of
culture. We do not yet understand what all these things were or how
@@ -4842,7 +4842,7 @@ the mound which later became the holy Sumerian city of Eridu, Iraqi
archeologists uncovered a handsome painted pottery. Pottery of the same
type had been noticed earlier by German archeologists on the surface
of a small mound, awash in the spring floods, near the remains of the
-Biblical city of Erich (Sumerian = Uruk; Arabic = Warka). This Eridu
+Biblical city of Erich (Sumerian = Uruk; Arabic = Warka). This �Eridu�
pottery, which is about all we have of the assemblage of the people who
once produced it, may be seen as a blend of the Samarran and Halafian
painted pottery styles. This may over-simplify the case, but as yet we
@@ -4864,7 +4864,7 @@ seems to move into place before the Halaf manifestation is finished,
and to blend with it. The Ubaidian assemblage in the south is by far
the more spectacular. The development of the temple has been traced
at Eridu from a simple little structure to a monumental building some
-62 feet long, with a pilaster-decorated faade and an altar in its
+62 feet long, with a pilaster-decorated fa�ade and an altar in its
central chamber. There is painted Ubaidian pottery, but the style is
hurried and somewhat careless and gives the _impression_ of having been
a cheap mass-production means of decoration when compared with the
@@ -4879,7 +4879,7 @@ turtle-like faces are another item in the southern Ubaidian assemblage.
There is a large Ubaid cemetery at Eridu, much of it still awaiting
excavation. The few skeletons so far tentatively studied reveal a
-completely modern type of Mediterraneanoid; the individuals whom the
+completely modern type of �Mediterraneanoid�; the individuals whom the
skeletons represent would undoubtedly blend perfectly into the modern
population of southern Iraq. What the Ubaidian assemblage says to us is
that these people had already adapted themselves and their culture to
@@ -4925,7 +4925,7 @@ woven stuffs must have been the mediums of exchange. Over what area did
the trading net-work of Ubaid extend? We start with the idea that the
Ubaidian assemblage is most richly developed in the south. We assume, I
think, correctly, that it represents a cultural flowering of the south.
-On the basis of the pottery of the still elusive Eridu immigrants
+On the basis of the pottery of the still elusive �Eridu� immigrants
who had first followed the rivers into alluvial Mesopotamia, we get
the notion that the characteristic painted pottery style of Ubaid
was developed in the southland. If this reconstruction is correct
@@ -4935,7 +4935,7 @@ assemblage of (and from the southern point of view, _fairly_ pure)
Ubaidian material in northern Iraq. The pottery appears all along the
Iranian flanks, even well east of the head of the Persian Gulf, and
ends in a later and spectacular flourish in an extremely handsome
-painted style called the Susa style. Ubaidian pottery has been noted
+painted style called the �Susa� style. Ubaidian pottery has been noted
up the valleys of both of the great rivers, well north of the Iraqi
and Syrian borders on the southern flanks of the Anatolian plateau.
It reaches the Mediterranean Sea and the valley of the Orontes in
@@ -4965,10 +4965,10 @@ Mesopotamia.
Next, much to our annoyance, we have what is almost a temporary
black-out. According to the system of terminology I favor, our next
-assemblage after that of Ubaid is called the _Warka_ phase, from
+�assemblage� after that of Ubaid is called the _Warka_ phase, from
the Arabic name for the site of Uruk or Erich. We know it only from
six or seven levels in a narrow test-pit at Warka, and from an even
-smaller hole at another site. This assemblage, so far, is known only
+smaller hole at another site. This �assemblage,� so far, is known only
by its pottery, some of which still bears Ubaidian style painting. The
characteristic Warkan pottery is unpainted, with smoothed red or gray
surfaces and peculiar shapes. Unquestionably, there must be a great
@@ -4979,7 +4979,7 @@ have to excavate it!
THE DAWN OF CIVILIZATION
After our exasperation with the almost unknown Warka interlude,
-following the brilliant false dawn of Ubaid, we move next to an
+following the brilliant �false dawn� of Ubaid, we move next to an
assemblage which yields traces of a preponderance of those elements
which we noted (p. 144) as meaning civilization. This assemblage
is that called _Proto-Literate_; it already contains writing. On
@@ -4988,8 +4988,8 @@ history--and no longer prehistory--the assemblage is named for the
historical implications of its content, and no longer after the name of
the site where it was first found. Since some of the older books used
site-names for this assemblage, I will tell you that the Proto-Literate
-includes the latter half of what used to be called the Uruk period
-_plus_ all of what used to be called the Jemdet Nasr period. It shows
+includes the latter half of what used to be called the �Uruk period�
+_plus_ all of what used to be called the �Jemdet Nasr period.� It shows
a consistent development from beginning to end.
I shall, in fact, leave much of the description and the historic
@@ -5033,18 +5033,18 @@ mental block seems to have been removed.
Clay tablets bearing pictographic signs are the Proto-Literate
forerunners of cuneiform writing. The earliest examples are not well
-understood but they seem to be devices for making accounts and
-for remembering accounts. Different from the later case in Egypt,
+understood but they seem to be �devices for making accounts and
+for remembering accounts.� Different from the later case in Egypt,
where writing appears fully formed in the earliest examples, the
development from simple pictographic signs to proper cuneiform writing
may be traced, step by step, in Mesopotamia. It is most probable
that the development of writing was connected with the temple and
-the need for keeping account of the temples possessions. Professor
+the need for keeping account of the temple�s possessions. Professor
Jacobsen sees writing as a means for overcoming space, time, and the
-increasing complications of human affairs: Literacy, which began
+increasing complications of human affairs: �Literacy, which began
with ... civilization, enhanced mightily those very tendencies in its
development which characterize it as a civilization and mark it off as
-such from other types of culture.
+such from other types of culture.�
[Illustration: RELIEF ON A PROTO-LITERATE STONE VASE, WARKA
@@ -5098,7 +5098,7 @@ civilized way of life.
I suppose you could say that the difference in the approach is that as
a prehistorian I have been looking forward or upward in time, while the
-historians look backward to glimpse what Ive been describing here. My
+historians look backward to glimpse what I�ve been describing here. My
base-line was half a million years ago with a being who had little more
than the capacity to make tools and fire to distinguish him from the
animals about him. Thus my point of view and that of the conventional
@@ -5114,17 +5114,17 @@ End of PREHISTORY
[Illustration]
-Youll doubtless easily recall your general course in ancient history:
+You�ll doubtless easily recall your general course in ancient history:
how the Sumerian dynasties of Mesopotamia were supplanted by those of
Babylonia, how the Hittite kingdom appeared in Anatolian Turkey, and
about the three great phases of Egyptian history. The literate kingdom
of Crete arose, and by 1500 B.C. there were splendid fortified Mycenean
towns on the mainland of Greece. This was the time--about the whole
eastern end of the Mediterranean--of what Professor Breasted called the
-first great internationalism, with flourishing trade, international
+�first great internationalism,� with flourishing trade, international
treaties, and royal marriages between Egyptians, Babylonians, and
-Hittites. By 1200 B.C., the whole thing had fragmented: the peoples of
-the sea were restless in their isles, and the great ancient centers in
+Hittites. By 1200 B.C., the whole thing had fragmented: �the peoples of
+the sea were restless in their isles,� and the great ancient centers in
Egypt, Mesopotamia, and Anatolia were eclipsed. Numerous smaller states
arose--Assyria, Phoenicia, Israel--and the Trojan war was fought.
Finally Assyria became the paramount power of all the Near East,
@@ -5135,7 +5135,7 @@ but casting them with its own tradition into a new mould, arose in
mainland Greece.
I once shocked my Classical colleagues to the core by referring to
-Greece as a second degree derived civilization, but there is much
+Greece as �a second degree derived civilization,� but there is much
truth in this. The principles of bronze- and then of iron-working, of
the alphabet, and of many other elements in Greek culture were borrowed
from western Asia. Our debt to the Greeks is too well known for me even
@@ -5146,7 +5146,7 @@ Greece fell in its turn to Rome, and in 55 B.C. Caesar invaded Britain.
I last spoke of Britain on page 142; I had chosen it as my single
example for telling you something of how the earliest farming
communities were established in Europe. Now I will continue with
-Britains later prehistory, so you may sense something of the end of
+Britain�s later prehistory, so you may sense something of the end of
prehistory itself. Remember that Britain is simply a single example
we select; the same thing could be done for all the other countries
of Europe, and will be possible also, some day, for further Asia and
@@ -5186,20 +5186,20 @@ few Battle-axe folk elements, including, in fact, stone battle-axes,
reached England with the earliest Beaker folk,[6] coming from the
Rhineland.
- [6] The British authors use the term Beaker folk to mean both
+ [6] The British authors use the term �Beaker folk� to mean both
archeological assemblage and human physical type. They speak
- of a ... tall, heavy-boned, rugged, and round-headed strain
+ of a �... tall, heavy-boned, rugged, and round-headed� strain
which they take to have developed, apparently in the Rhineland,
by a mixture of the original (Spanish?) beaker-makers and
the northeast European battle-axe makers. However, since the
science of physical anthropology is very much in flux at the
moment, and since I am not able to assess the evidence for these
- physical types, I _do not_ use the term folk in this book with
+ physical types, I _do not_ use the term �folk� in this book with
its usual meaning of standardized physical type. When I use
- folk here, I mean simply _the makers of a given archeological
+ �folk� here, I mean simply _the makers of a given archeological
assemblage_. The difficulty only comes when assemblages are
named for some item in them; it is too clumsy to make an
- adjective of the item and refer to a beakerian assemblage.
+ adjective of the item and refer to a �beakerian� assemblage.
The Beaker folk settled earliest in the agriculturally fertile south
and east. There seem to have been several phases of Beaker folk
@@ -5211,7 +5211,7 @@ folk are known. They buried their dead singly, sometimes in conspicuous
individual barrows with the dead warrior in his full trappings. The
spectacular element in the assemblage of the Beaker folk is a group
of large circular monuments with ditches and with uprights of wood or
-stone. These henges became truly monumental several hundred years
+stone. These �henges� became truly monumental several hundred years
later; while they were occasionally dedicated with a burial, they were
not primarily tombs. The effect of the invasion of the Beaker folk
seems to cut across the whole fabric of life in Britain.
@@ -5221,7 +5221,7 @@ seems to cut across the whole fabric of life in Britain.
There was, however, a second major element in British life at this
time. It shows itself in the less well understood traces of a group
again called after one of the items in their catalogue, the Food-vessel
-folk. There are many burials in these food-vessel pots in northern
+folk. There are many burials in these �food-vessel� pots in northern
England, Scotland, and Ireland, and the pottery itself seems to
link back to that of the Peterborough assemblage. Like the earlier
Peterborough people in the highland zone before them, the makers of
@@ -5238,8 +5238,8 @@ MORE INVASIONS
About 1500 B.C., the situation became further complicated by the
arrival of new people in the region of southern England anciently
called Wessex. The traces suggest the Brittany coast of France as a
-source, and the people seem at first to have been a small but heroic
-group of aristocrats. Their heroes are buried with wealth and
+source, and the people seem at first to have been a small but �heroic�
+group of aristocrats. Their �heroes� are buried with wealth and
ceremony, surrounded by their axes and daggers of bronze, their gold
ornaments, and amber and jet beads. These rich finds show that the
trade-linkage these warriors patronized spread from the Baltic sources
@@ -5265,10 +5265,10 @@ which must have been necessary before such a great monument could have
been built.
-THIS ENGLAND
+�THIS ENGLAND�
The range from 1900 to about 1400 B.C. includes the time of development
-of the archeological features usually called the Early Bronze Age
+of the archeological features usually called the �Early Bronze Age�
in Britain. In fact, traces of the Wessex warriors persisted down to
about 1200 B.C. The main regions of the island were populated, and the
adjustments to the highland and lowland zones were distinct and well
@@ -5279,7 +5279,7 @@ trading role, separated from the European continent but conveniently
adjacent to it. The tin of Cornwall--so important in the production
of good bronze--as well as the copper of the west and of Ireland,
taken with the gold of Ireland and the general excellence of Irish
-metal work, assured Britain a traders place in the then known world.
+metal work, assured Britain a trader�s place in the then known world.
Contacts with the eastern Mediterranean may have been by sea, with
Cornish tin as the attraction, or may have been made by the Food-vessel
middlemen on their trips to the Baltic coast. There they would have
@@ -5292,9 +5292,9 @@ relative isolation gave some peace and also gave time for a leveling
and further fusion of culture. The separate cultural traditions began
to have more in common. The growing of barley, the herding of sheep and
cattle, and the production of woolen garments were already features
-common to all Britains inhabitants save a few in the remote highlands,
+common to all Britain�s inhabitants save a few in the remote highlands,
the far north, and the distant islands not yet fully touched by
-food-production. The personality of Britain was being formed.
+food-production. The �personality of Britain� was being formed.
CREMATION BURIALS BEGIN
@@ -5325,9 +5325,9 @@ which we shall mention below.
The British cremation-burial-in-urns folk survived a long time in the
highland zone. In the general British scheme, they make up what is
-called the Middle Bronze Age, but in the highland zone they last
+called the �Middle Bronze Age,� but in the highland zone they last
until after 900 B.C. and are considered to be a specialized highland
-Late Bronze Age. In the highland zone, these later cremation-burial
+�Late Bronze Age.� In the highland zone, these later cremation-burial
folk seem to have continued the older Food-vessel tradition of being
middlemen in the metal market.
@@ -5379,12 +5379,12 @@ to get a picture of estate or tribal boundaries which included village
communities; we find a variety of tools in bronze, and even whetstones
which show that iron has been honed on them (although the scarce iron
has not been found). Let me give you the picture in Professor S.
-Piggotts words: The ... Late Bronze Age of southern England was but
+Piggott�s words: �The ... Late Bronze Age of southern England was but
the forerunner of the earliest Iron Age in the same region, not only in
the techniques of agriculture, but almost certainly in terms of ethnic
kinship ... we can with some assurance talk of the Celts ... the great
early Celtic expansion of the Continent is recognized to be that of the
-Urnfield people.
+Urnfield people.�
Thus, certainly by 500 B.C., there were people in Britain, some of
whose descendants we may recognize today in name or language in remote
@@ -5399,11 +5399,11 @@ efficient set of tools than does bronze. Iron tools seem first to
have been made in quantity in Hittite Anatolia about 1500 B.C. In
continental Europe, the earliest, so-called Hallstatt, iron-using
cultures appeared in Germany soon after 750 B.C. Somewhat later,
-Greek and especially Etruscan exports of _objets dart_--which moved
+Greek and especially Etruscan exports of _objets d�art_--which moved
with a flourishing trans-Alpine wine trade--influenced the Hallstatt
iron-working tradition. Still later new classical motifs, together with
older Hallstatt, oriental, and northern nomad motifs, gave rise to a
-new style in metal decoration which characterizes the so-called La Tne
+new style in metal decoration which characterizes the so-called La T�ne
phase.
A few iron users reached Britain a little before 400 B.C. Not long
@@ -5422,7 +5422,7 @@ HILL-FORTS AND FARMS
The earliest iron-users seem to have entrenched themselves temporarily
within hill-top forts, mainly in the south. Gradually, they moved
inland, establishing _individual_ farm sites with extensive systems
-of rectangular fields. We recognize these fields by the lynchets or
+of rectangular fields. We recognize these fields by the �lynchets� or
lines of soil-creep which plowing left on the slopes of hills. New
crops appeared; there were now bread wheat, oats, and rye, as well as
barley.
@@ -5434,7 +5434,7 @@ various outbuildings and pits for the storage of grain. Weaving was
done on the farm, but not blacksmithing, which must have been a
specialized trade. Save for the lack of firearms, the place might
almost be taken for a farmstead on the American frontier in the early
-1800s.
+1800�s.
Toward 250 B.C. there seems to have been a hasty attempt to repair the
hill-forts and to build new ones, evidently in response to signs of
@@ -5446,9 +5446,9 @@ THE SECOND PHASE
Perhaps the hill-forts were not entirely effective or perhaps a
compromise was reached. In any case, the newcomers from the Marne
district did establish themselves, first in the southeast and then to
-the north and west. They brought iron with decoration of the La Tne
+the north and west. They brought iron with decoration of the La T�ne
type and also the two-wheeled chariot. Like the Wessex warriors of
-over a thousand years earlier, they made heroes graves, with their
+over a thousand years earlier, they made �heroes�� graves, with their
warriors buried in the war-chariots and dressed in full trappings.
[Illustration: CELTIC BUCKLE]
@@ -5457,7 +5457,7 @@ The metal work of these Marnian newcomers is excellent. The peculiar
Celtic art style, based originally on the classic tendril motif,
is colorful and virile, and fits with Greek and Roman descriptions
of Celtic love of color in dress. There is a strong trace of these
-newcomers northward in Yorkshire, linked by Ptolemys description to
+newcomers northward in Yorkshire, linked by Ptolemy�s description to
the Parisii, doubtless part of the Celtic tribe which originally gave
its name to Paris on the Seine. Near Glastonbury, in Somerset, two
villages in swamps have been excavated. They seem to date toward the
@@ -5469,7 +5469,7 @@ villagers.
In Scotland, which yields its first iron tools at a date of about 100
B.C., and in northern Ireland even slightly earlier, the effects of the
-two phases of newcomers tend especially to blend. Hill-forts, brochs
+two phases of newcomers tend especially to blend. Hill-forts, �brochs�
(stone-built round towers) and a variety of other strange structures
seem to appear as the new ideas develop in the comparative isolation of
northern Britain.
@@ -5493,27 +5493,27 @@ at last, we can even begin to speak of dynasties and individuals.
Some time before 55 B.C., the Catuvellauni, originally from the Marne
district in France, had possessed themselves of a large part of
southeastern England. They evidently sailed up the Thames and built a
-town of over a hundred acres in area. Here ruled Cassivellaunus, the
-first man in England whose name we know, and whose town Caesar sacked.
+town of over a hundred acres in area. Here ruled Cassivellaunus, �the
+first man in England whose name we know,� and whose town Caesar sacked.
The town sprang up elsewhere again, however.
THE END OF PREHISTORY
Prehistory, strictly speaking, is now over in southern Britain.
-Claudius effective invasion took place in 43 A.D.; by 83 A.D., a raid
+Claudius� effective invasion took place in 43 A.D.; by 83 A.D., a raid
had been made as far north as Aberdeen in Scotland. But by 127 A.D.,
Hadrian had completed his wall from the Solway to the Tyne, and the
Romans settled behind it. In Scotland, Romanization can have affected
-the countryside very little. Professor Piggott adds that ... it is
+the countryside very little. Professor Piggott adds that �... it is
when the pressure of Romanization is relaxed by the break-up of the
Dark Ages that we see again the Celtic metal-smiths handling their
material with the same consummate skill as they had before the Roman
Conquest, and with traditional styles that had not even then forgotten
-their Marnian and Belgic heritage.
+their Marnian and Belgic heritage.�
In fact, many centuries go by, in Britain as well as in the rest of
-Europe, before the archeologists task is complete and the historian on
+Europe, before the archeologist�s task is complete and the historian on
his own is able to describe the ways of men in the past.
@@ -5524,7 +5524,7 @@ you will have noticed how often I had to refer to the European
continent itself. Britain, beyond the English Channel for all of her
later prehistory, had a much simpler course of events than did most of
the rest of Europe in later prehistoric times. This holds, in spite
-of all the invasions and reverberations from the continent. Most
+of all the �invasions� and �reverberations� from the continent. Most
of Europe was the scene of an even more complicated ebb and flow of
cultural change, save in some of its more remote mountain valleys and
peninsulas.
@@ -5536,7 +5536,7 @@ accounts and some good general accounts of part of the range from about
3000 B.C. to A.D. 1. I suspect that the difficulty of making a good
book that covers all of its later prehistory is another aspect of what
makes Europe so very complicated a continent today. The prehistoric
-foundations for Europes very complicated set of civilizations,
+foundations for Europe�s very complicated set of civilizations,
cultures, and sub-cultures--which begin to appear as history
proceeds--were in themselves very complicated.
@@ -5552,8 +5552,8 @@ of their journeys. But by the same token, they had had time en route to
take on their characteristic European aspects.
Some time ago, Sir Cyril Fox wrote a famous book called _The
-Personality of Britain_, sub-titled Its Influence on Inhabitant and
-Invader in Prehistoric and Early Historic Times. We have not gone
+Personality of Britain_, sub-titled �Its Influence on Inhabitant and
+Invader in Prehistoric and Early Historic Times.� We have not gone
into the post-Roman early historic period here; there are still the
Anglo-Saxons and Normans to account for as well as the effects of
the Romans. But what I have tried to do was to begin the story of
@@ -5570,7 +5570,7 @@ Summary
In the pages you have read so far, you have been brought through the
-earliest 99 per cent of the story of mans life on this planet. I have
+earliest 99 per cent of the story of man�s life on this planet. I have
left only 1 per cent of the story for the historians to tell.
@@ -5601,7 +5601,7 @@ But I think there may have been a few. Certainly the pace of the
first act accelerated with the swing from simple gathering to more
intensified collecting. The great cave art of France and Spain was
probably an expression of a climax. Even the ideas of burying the dead
-and of the Venus figurines must also point to levels of human thought
+and of the �Venus� figurines must also point to levels of human thought
and activity that were over and above pure food-getting.
@@ -5629,7 +5629,7 @@ five thousand years after the second act began. But it could never have
happened in the first act at all.
There is another curious thing about the first act. Many of the players
-didnt know it was over and they kept on with their roles long after
+didn�t know it was over and they kept on with their roles long after
the second act had begun. On the edges of the stage there are today
some players who are still going on with the first act. The Eskimos,
and the native Australians, and certain tribes in the Amazon jungle are
@@ -5680,20 +5680,20 @@ act may have lessons for us and give depth to our thinking. I know
there are at least _some_ lessons, even in the present incomplete
state of our knowledge. The players who began the second act--that of
food-production--separately, in different parts of the world, were not
-all of one pure race nor did they have pure cultural traditions.
+all of one �pure race� nor did they have �pure� cultural traditions.
Some apparently quite mixed Mediterraneans got off to the first start
on the second act and brought it to its first two climaxes as well.
Peoples of quite different physical type achieved the first climaxes in
China and in the New World.
In our British example of how the late prehistory of Europe worked, we
-listed a continuous series of invasions and reverberations. After
+listed a continuous series of �invasions� and �reverberations.� After
each of these came fusion. Even though the Channel protected Britain
from some of the extreme complications of the mixture and fusion of
continental Europe, you can see how silly it would be to refer to a
-pure British race or a pure British culture. We speak of the United
-States as a melting pot. But this is nothing new. Actually, Britain
-and all the rest of the world have been melting pots at one time or
+�pure� British race or a �pure� British culture. We speak of the United
+States as a �melting pot.� But this is nothing new. Actually, Britain
+and all the rest of the world have been �melting pots� at one time or
another.
By the time the written records of Mesopotamia and Egypt begin to turn
@@ -5703,12 +5703,12 @@ itself, we are thrown back on prehistoric archeology. And this is as
true for China, India, Middle America, and the Andes, as it is for the
Near East.
-There are lessons to be learned from all of mans past, not simply
+There are lessons to be learned from all of man�s past, not simply
lessons of how to fight battles or win peace conferences, but of how
human society evolves from one stage to another. Many of these lessons
can only be looked for in the prehistoric past. So far, we have only
made a beginning. There is much still to do, and many gaps in the story
-are yet to be filled. The prehistorians job is to find the evidence,
+are yet to be filled. The prehistorian�s job is to find the evidence,
to fill the gaps, and to discover the lessons men have learned in the
past. As I see it, this is not only an exciting but a very practical
goal for which to strive.
@@ -5745,7 +5745,7 @@ paperbound books.)
GEOCHRONOLOGY AND THE ICE AGE
-(Two general books. Some Pleistocene geologists disagree with Zeuners
+(Two general books. Some Pleistocene geologists disagree with Zeuner�s
interpretation of the dating evidence, but their points of view appear
in professional journals, in articles too cumbersome to list here.)
@@ -5815,7 +5815,7 @@ GENERAL PREHISTORY
Press.
Movius, Hallam L., Jr.
- Old World Prehistory: Paleolithic in _Anthropology Today_.
+ �Old World Prehistory: Paleolithic� in _Anthropology Today_.
Kroeber, A. L., ed. 1953. University of Chicago Press.
Oakley, Kenneth P.
@@ -5826,7 +5826,7 @@ GENERAL PREHISTORY
_British Prehistory._ 1949. Oxford University Press.
Pittioni, Richard
- _Die Urgeschichtlichen Grundlagen der Europischen Kultur._
+ _Die Urgeschichtlichen Grundlagen der Europ�ischen Kultur._
1949. Deuticke. (A single book which does attempt to cover the
whole range of European prehistory to ca. 1 A.D.)
@@ -5834,7 +5834,7 @@ GENERAL PREHISTORY
THE NEAR EAST
Adams, Robert M.
- Developmental Stages in Ancient Mesopotamia, _in_ Steward,
+ �Developmental Stages in Ancient Mesopotamia,� _in_ Steward,
Julian, _et al_, _Irrigation Civilizations: A Comparative
Study_. 1955. Pan American Union.
@@ -6000,7 +6000,7 @@ Index
Bolas, 54
- Bordes, Franois, 62
+ Bordes, Fran�ois, 62
Borer, 77
@@ -6028,7 +6028,7 @@ Index
killed by stampede, 86
Burials, 66, 86;
- in henges, 164;
+ in �henges,� 164;
in urns, 168
Burins, 75
@@ -6085,7 +6085,7 @@ Index
Combe Capelle, 30
- Combe Capelle-Brnn group, 34
+ Combe Capelle-Br�nn group, 34
Commont, Victor, 51
@@ -6097,7 +6097,7 @@ Index
Corrals for cattle, 140
- Cradle of mankind, 136
+ �Cradle of mankind,� 136
Cremation, 167
@@ -6123,7 +6123,7 @@ Index
Domestication, of animals, 100, 105, 107;
of plants, 100
- Dragon teeth fossils in China, 28
+ �Dragon teeth� fossils in China, 28
Drill, 77
@@ -6176,9 +6176,9 @@ Index
Fayum, 135;
radiocarbon date, 146
- Fertile Crescent, 107, 146
+ �Fertile Crescent,� 107, 146
- Figurines, Venus, 84;
+ Figurines, �Venus,� 84;
at Jarmo, 128;
at Ubaid, 153
@@ -6197,7 +6197,7 @@ Index
Flint industry, 127
- Fontchevade, 32, 56, 58
+ Font�chevade, 32, 56, 58
Food-collecting, 104, 121;
end of, 104
@@ -6223,7 +6223,7 @@ Index
Food-vessel folk, 164
- Forest folk, 97, 98, 104, 110
+ �Forest folk,� 97, 98, 104, 110
Fox, Sir Cyril, 174
@@ -6379,7 +6379,7 @@ Index
Land bridges in Mediterranean, 19
- La Tne phase, 170
+ La T�ne phase, 170
Laurel leaf point, 78, 89
@@ -6404,7 +6404,7 @@ Index
Mammoth, 93;
in cave art, 85
- Man-apes, 26
+ �Man-apes,� 26
Mango, 107
@@ -6435,7 +6435,7 @@ Index
Microliths, 87;
at Jarmo, 130;
- lunates, 87;
+ �lunates,� 87;
trapezoids, 87;
triangles, 87
@@ -6443,7 +6443,7 @@ Index
Mine-shafts, 140
- Mlefaat, 126, 127
+ M�lefaat, 126, 127
Mongoloids, 29, 90
@@ -6453,9 +6453,9 @@ Index
Mount Carmel, 11, 33, 52, 59, 64, 69, 113, 114
- Mousterian man, 64
+ �Mousterian man,� 64
- Mousterian tools, 61, 62;
+ �Mousterian� tools, 61, 62;
of Acheulean tradition, 62
Movius, H. L., 47
@@ -6471,7 +6471,7 @@ Index
Near East, beginnings of civilization in, 20, 144;
cave sites, 58;
climate in Ice Age, 99;
- Fertile Crescent, 107, 146;
+ �Fertile Crescent,� 107, 146;
food-production in, 99;
Natufian assemblage in, 113-115;
stone tools, 114
@@ -6539,7 +6539,7 @@ Index
Pig, wild, 108
- Piltdown man, 29
+ �Piltdown man,� 29
Pins, 80
@@ -6578,7 +6578,7 @@ Index
Race, 35;
biological, 36;
- pure, 16
+ �pure,� 16
Radioactivity, 9, 10
@@ -6795,7 +6795,7 @@ Index
Writing, 158;
cuneiform, 158
- Wrm I glaciation, 58
+ W�rm I glaciation, 58
Zebu cattle, domestication of, 107
@@ -6810,7 +6810,7 @@ Index
-Transcribers note:
+Transcriber�s note:
Punctuation, hyphenation, and spelling were made consistent when a
predominant preference was found in this book; otherwise they were not
diff --git a/ciphers/rabin_miller.py b/ciphers/rabin_miller.py
index 65c162984ece..410d559d4315 100644
--- a/ciphers/rabin_miller.py
+++ b/ciphers/rabin_miller.py
@@ -3,7 +3,7 @@
import random
-def rabinMiller(num: int) -> bool:
+def rabin_miller(num: int) -> bool:
s = num - 1
t = 0
@@ -11,7 +11,7 @@ def rabinMiller(num: int) -> bool:
s = s // 2
t += 1
- for trials in range(5):
+ for _ in range(5):
a = random.randrange(2, num - 1)
v = pow(a, s, num)
if v != 1:
@@ -21,15 +21,15 @@ def rabinMiller(num: int) -> bool:
return False
else:
i = i + 1
- v = (v ** 2) % num
+ v = (v**2) % num
return True
-def isPrime(num: int) -> bool:
+def is_prime_low_num(num: int) -> bool:
if num < 2:
return False
- lowPrimes = [
+ low_primes = [
2,
3,
5,
@@ -200,24 +200,24 @@ def isPrime(num: int) -> bool:
997,
]
- if num in lowPrimes:
+ if num in low_primes:
return True
- for prime in lowPrimes:
+ for prime in low_primes:
if (num % prime) == 0:
return False
- return rabinMiller(num)
+ return rabin_miller(num)
-def generateLargePrime(keysize: int = 1024) -> int:
+def generate_large_prime(keysize: int = 1024) -> int:
while True:
num = random.randrange(2 ** (keysize - 1), 2 ** (keysize))
- if isPrime(num):
+ if is_prime_low_num(num):
return num
if __name__ == "__main__":
- num = generateLargePrime()
+ num = generate_large_prime()
print(("Prime number:", num))
- print(("isPrime:", isPrime(num)))
+ print(("is_prime_low_num:", is_prime_low_num(num)))
diff --git a/ciphers/rail_fence_cipher.py b/ciphers/rail_fence_cipher.py
index 2596415207ae..5b2311a115e4 100644
--- a/ciphers/rail_fence_cipher.py
+++ b/ciphers/rail_fence_cipher.py
@@ -1,4 +1,4 @@
-""" https://en.wikipedia.org/wiki/Rail_fence_cipher """
+"""https://en.wikipedia.org/wiki/Rail_fence_cipher"""
def encrypt(input_string: str, key: int) -> str:
@@ -20,7 +20,7 @@ def encrypt(input_string: str, key: int) -> str:
...
TypeError: sequence item 0: expected str instance, int found
"""
- grid = [[] for _ in range(key)]
+ temp_grid: list[list[str]] = [[] for _ in range(key)]
lowest = key - 1
if key <= 0:
@@ -31,8 +31,8 @@ def encrypt(input_string: str, key: int) -> str:
for position, character in enumerate(input_string):
num = position % (lowest * 2) # puts it in bounds
num = min(num, lowest * 2 - num) # creates zigzag pattern
- grid[num].append(character)
- grid = ["".join(row) for row in grid]
+ temp_grid[num].append(character)
+ grid = ["".join(row) for row in temp_grid]
output_string = "".join(grid)
return output_string
@@ -63,7 +63,7 @@ def decrypt(input_string: str, key: int) -> str:
if key == 1:
return input_string
- temp_grid = [[] for _ in range(key)] # generates template
+ temp_grid: list[list[str]] = [[] for _ in range(key)] # generates template
for position in range(len(input_string)):
num = position % (lowest * 2) # puts it in bounds
num = min(num, lowest * 2 - num) # creates zigzag pattern
@@ -72,7 +72,7 @@ def decrypt(input_string: str, key: int) -> str:
counter = 0
for row in temp_grid: # fills in the characters
splice = input_string[counter : counter + len(row)]
- grid.append([character for character in splice])
+ grid.append(list(splice))
counter += len(row)
output_string = "" # reads as zigzag
@@ -84,7 +84,7 @@ def decrypt(input_string: str, key: int) -> str:
return output_string
-def bruteforce(input_string: str) -> dict:
+def bruteforce(input_string: str) -> dict[int, str]:
"""Uses decrypt function by guessing every key
>>> bruteforce("HWe olordll")[4]
diff --git a/ciphers/rot13.py b/ciphers/rot13.py
index 21dbda98eecc..b367c3215127 100644
--- a/ciphers/rot13.py
+++ b/ciphers/rot13.py
@@ -20,7 +20,7 @@ def dencrypt(s: str, n: int = 13) -> str:
return out
-def main():
+def main() -> None:
s0 = input("Enter message: ")
s1 = dencrypt(s0, 13)
diff --git a/ciphers/rsa_cipher.py b/ciphers/rsa_cipher.py
index 57c916a44d4b..ac9782a49fff 100644
--- a/ciphers/rsa_cipher.py
+++ b/ciphers/rsa_cipher.py
@@ -7,144 +7,142 @@
BYTE_SIZE = 256
-def main():
- filename = "encrypted_file.txt"
- response = input(r"Encrypt\Decrypt [e\d]: ")
-
- if response.lower().startswith("e"):
- mode = "encrypt"
- elif response.lower().startswith("d"):
- mode = "decrypt"
-
- if mode == "encrypt":
- if not os.path.exists("rsa_pubkey.txt"):
- rkg.makeKeyFiles("rsa", 1024)
-
- message = input("\nEnter message: ")
- pubKeyFilename = "rsa_pubkey.txt"
- print("Encrypting and writing to %s..." % (filename))
- encryptedText = encryptAndWriteToFile(filename, pubKeyFilename, message)
-
- print("\nEncrypted text:")
- print(encryptedText)
-
- elif mode == "decrypt":
- privKeyFilename = "rsa_privkey.txt"
- print("Reading from %s and decrypting..." % (filename))
- decryptedText = readFromFileAndDecrypt(filename, privKeyFilename)
- print("writing decryption to rsa_decryption.txt...")
- with open("rsa_decryption.txt", "w") as dec:
- dec.write(decryptedText)
-
- print("\nDecryption:")
- print(decryptedText)
-
-
-def getBlocksFromText(message: int, blockSize: int = DEFAULT_BLOCK_SIZE) -> [int]:
- messageBytes = message.encode("ascii")
-
- blockInts = []
- for blockStart in range(0, len(messageBytes), blockSize):
- blockInt = 0
- for i in range(blockStart, min(blockStart + blockSize, len(messageBytes))):
- blockInt += messageBytes[i] * (BYTE_SIZE ** (i % blockSize))
- blockInts.append(blockInt)
- return blockInts
-
-
-def getTextFromBlocks(
- blockInts: [int], messageLength: int, blockSize: int = DEFAULT_BLOCK_SIZE
+def get_blocks_from_text(
+ message: str, block_size: int = DEFAULT_BLOCK_SIZE
+) -> list[int]:
+ message_bytes = message.encode("ascii")
+
+ block_ints = []
+ for block_start in range(0, len(message_bytes), block_size):
+ block_int = 0
+ for i in range(block_start, min(block_start + block_size, len(message_bytes))):
+ block_int += message_bytes[i] * (BYTE_SIZE ** (i % block_size))
+ block_ints.append(block_int)
+ return block_ints
+
+
+def get_text_from_blocks(
+ block_ints: list[int], message_length: int, block_size: int = DEFAULT_BLOCK_SIZE
) -> str:
- message = []
- for blockInt in blockInts:
- blockMessage = []
- for i in range(blockSize - 1, -1, -1):
- if len(message) + i < messageLength:
- asciiNumber = blockInt // (BYTE_SIZE ** i)
- blockInt = blockInt % (BYTE_SIZE ** i)
- blockMessage.insert(0, chr(asciiNumber))
- message.extend(blockMessage)
+ message: list[str] = []
+ for block_int in block_ints:
+ block_message: list[str] = []
+ for i in range(block_size - 1, -1, -1):
+ if len(message) + i < message_length:
+ ascii_number = block_int // (BYTE_SIZE**i)
+ block_int = block_int % (BYTE_SIZE**i)
+ block_message.insert(0, chr(ascii_number))
+ message.extend(block_message)
return "".join(message)
-def encryptMessage(
- message: str, key: (int, int), blockSize: int = DEFAULT_BLOCK_SIZE
-) -> [int]:
- encryptedBlocks = []
+def encrypt_message(
+ message: str, key: tuple[int, int], block_size: int = DEFAULT_BLOCK_SIZE
+) -> list[int]:
+ encrypted_blocks = []
n, e = key
- for block in getBlocksFromText(message, blockSize):
- encryptedBlocks.append(pow(block, e, n))
- return encryptedBlocks
+ for block in get_blocks_from_text(message, block_size):
+ encrypted_blocks.append(pow(block, e, n))
+ return encrypted_blocks
-def decryptMessage(
- encryptedBlocks: [int],
- messageLength: int,
- key: (int, int),
- blockSize: int = DEFAULT_BLOCK_SIZE,
+def decrypt_message(
+ encrypted_blocks: list[int],
+ message_length: int,
+ key: tuple[int, int],
+ block_size: int = DEFAULT_BLOCK_SIZE,
) -> str:
- decryptedBlocks = []
+ decrypted_blocks = []
n, d = key
- for block in encryptedBlocks:
- decryptedBlocks.append(pow(block, d, n))
- return getTextFromBlocks(decryptedBlocks, messageLength, blockSize)
+ for block in encrypted_blocks:
+ decrypted_blocks.append(pow(block, d, n))
+ return get_text_from_blocks(decrypted_blocks, message_length, block_size)
-def readKeyFile(keyFilename: str) -> (int, int, int):
- with open(keyFilename) as fo:
+def read_key_file(key_filename: str) -> tuple[int, int, int]:
+ with open(key_filename) as fo:
content = fo.read()
- keySize, n, EorD = content.split(",")
- return (int(keySize), int(n), int(EorD))
+ key_size, n, eor_d = content.split(",")
+ return (int(key_size), int(n), int(eor_d))
-def encryptAndWriteToFile(
- messageFilename: str,
- keyFilename: str,
+def encrypt_and_write_to_file(
+ message_filename: str,
+ key_filename: str,
message: str,
- blockSize: int = DEFAULT_BLOCK_SIZE,
+ block_size: int = DEFAULT_BLOCK_SIZE,
) -> str:
- keySize, n, e = readKeyFile(keyFilename)
- if keySize < blockSize * 8:
+ key_size, n, e = read_key_file(key_filename)
+ if key_size < block_size * 8:
sys.exit(
- "ERROR: Block size is %s bits and key size is %s bits. The RSA cipher "
- "requires the block size to be equal to or greater than the key size. "
- "Either decrease the block size or use different keys."
- % (blockSize * 8, keySize)
+ f"ERROR: Block size is {block_size * 8} bits and key size is {key_size} "
+ "bits. The RSA cipher requires the block size to be equal to or greater "
+ "than the key size. Either decrease the block size or use different keys."
)
- encryptedBlocks = encryptMessage(message, (n, e), blockSize)
+ encrypted_blocks = [str(i) for i in encrypt_message(message, (n, e), block_size)]
- for i in range(len(encryptedBlocks)):
- encryptedBlocks[i] = str(encryptedBlocks[i])
- encryptedContent = ",".join(encryptedBlocks)
- encryptedContent = "{}_{}_{}".format(len(message), blockSize, encryptedContent)
- with open(messageFilename, "w") as fo:
- fo.write(encryptedContent)
- return encryptedContent
+ encrypted_content = ",".join(encrypted_blocks)
+ encrypted_content = f"{len(message)}_{block_size}_{encrypted_content}"
+ with open(message_filename, "w") as fo:
+ fo.write(encrypted_content)
+ return encrypted_content
-def readFromFileAndDecrypt(messageFilename: str, keyFilename: str) -> str:
- keySize, n, d = readKeyFile(keyFilename)
- with open(messageFilename) as fo:
+def read_from_file_and_decrypt(message_filename: str, key_filename: str) -> str:
+ key_size, n, d = read_key_file(key_filename)
+ with open(message_filename) as fo:
content = fo.read()
- messageLength, blockSize, encryptedMessage = content.split("_")
- messageLength = int(messageLength)
- blockSize = int(blockSize)
+ message_length_str, block_size_str, encrypted_message = content.split("_")
+ message_length = int(message_length_str)
+ block_size = int(block_size_str)
- if keySize < blockSize * 8:
+ if key_size < block_size * 8:
sys.exit(
- "ERROR: Block size is %s bits and key size is %s bits. The RSA cipher "
- "requires the block size to be equal to or greater than the key size. "
- "Did you specify the correct key file and encrypted file?"
- % (blockSize * 8, keySize)
+ f"ERROR: Block size is {block_size * 8} bits and key size is {key_size} "
+ "bits. The RSA cipher requires the block size to be equal to or greater "
+ "than the key size. Were the correct key file and encrypted file specified?"
)
- encryptedBlocks = []
- for block in encryptedMessage.split(","):
- encryptedBlocks.append(int(block))
+ encrypted_blocks = []
+ for block in encrypted_message.split(","):
+ encrypted_blocks.append(int(block))
+
+ return decrypt_message(encrypted_blocks, message_length, (n, d), block_size)
+
+
+def main() -> None:
+ filename = "encrypted_file.txt"
+ response = input(r"Encrypt\Decrypt [e\d]: ")
+
+ if response.lower().startswith("e"):
+ mode = "encrypt"
+ elif response.lower().startswith("d"):
+ mode = "decrypt"
- return decryptMessage(encryptedBlocks, messageLength, (n, d), blockSize)
+ if mode == "encrypt":
+ if not os.path.exists("rsa_pubkey.txt"):
+ rkg.make_key_files("rsa", 1024)
+
+ message = input("\nEnter message: ")
+ pubkey_filename = "rsa_pubkey.txt"
+ print(f"Encrypting and writing to {filename}...")
+ encrypted_text = encrypt_and_write_to_file(filename, pubkey_filename, message)
+
+ print("\nEncrypted text:")
+ print(encrypted_text)
+
+ elif mode == "decrypt":
+ privkey_filename = "rsa_privkey.txt"
+ print(f"Reading from {filename} and decrypting...")
+ decrypted_text = read_from_file_and_decrypt(filename, privkey_filename)
+ print("writing decryption to rsa_decryption.txt...")
+ with open("rsa_decryption.txt", "w") as dec:
+ dec.write(decrypted_text)
+
+ print("\nDecryption:")
+ print(decrypted_text)
if __name__ == "__main__":
diff --git a/ciphers/rsa_factorization.py b/ciphers/rsa_factorization.py
index b18aab609e2d..585b21fac856 100644
--- a/ciphers/rsa_factorization.py
+++ b/ciphers/rsa_factorization.py
@@ -3,26 +3,30 @@
The program can efficiently factor RSA prime number given the private key d and
public key e.
-Source: on page 3 of https://crypto.stanford.edu/~dabo/papers/RSA-survey.pdf
-More readable source: https://www.di-mgt.com.au/rsa_factorize_n.html
+
+| Source: on page ``3`` of https://crypto.stanford.edu/~dabo/papers/RSA-survey.pdf
+| More readable source: https://www.di-mgt.com.au/rsa_factorize_n.html
+
large number can take minutes to factor, therefore are not included in doctest.
"""
+
from __future__ import annotations
import math
import random
-def rsafactor(d: int, e: int, N: int) -> [int]:
+def rsafactor(d: int, e: int, n: int) -> list[int]:
"""
This function returns the factors of N, where p*q=N
- Return: [p, q]
+
+ Return: [p, q]
We call N the RSA modulus, e the encryption exponent, and d the decryption exponent.
The pair (N, e) is the public key. As its name suggests, it is public and is used to
- encrypt messages.
+ encrypt messages.
The pair (N, d) is the secret key or private key and is known only to the recipient
- of encrypted messages.
+ of encrypted messages.
>>> rsafactor(3, 16971, 25777)
[149, 173]
@@ -35,16 +39,16 @@ def rsafactor(d: int, e: int, N: int) -> [int]:
p = 0
q = 0
while p == 0:
- g = random.randint(2, N - 1)
+ g = random.randint(2, n - 1)
t = k
while True:
if t % 2 == 0:
t = t // 2
- x = (g ** t) % N
- y = math.gcd(x - 1, N)
+ x = (g**t) % n
+ y = math.gcd(x - 1, n)
if x > 1 and y > 1:
p = y
- q = N // y
+ q = n // y
break # find the correct factors
else:
break # t is not divisible by 2, break and choose another g
diff --git a/ciphers/rsa_key_generator.py b/ciphers/rsa_key_generator.py
index 5693aa637ee9..44970e8cbc15 100644
--- a/ciphers/rsa_key_generator.py
+++ b/ciphers/rsa_key_generator.py
@@ -1,59 +1,62 @@
import os
import random
import sys
-from typing import Tuple
-from . import cryptomath_module as cryptoMath
-from . import rabin_miller as rabinMiller
+from maths.greatest_common_divisor import gcd_by_iterative
+from . import cryptomath_module, rabin_miller
-def main():
+
+def main() -> None:
print("Making key files...")
- makeKeyFiles("rsa", 1024)
+ make_key_files("rsa", 1024)
print("Key files generation successful.")
-def generateKey(keySize: int) -> Tuple[Tuple[int, int], Tuple[int, int]]:
- print("Generating prime p...")
- p = rabinMiller.generateLargePrime(keySize)
- print("Generating prime q...")
- q = rabinMiller.generateLargePrime(keySize)
+def generate_key(key_size: int) -> tuple[tuple[int, int], tuple[int, int]]:
+ """
+ >>> random.seed(0) # for repeatability
+ >>> public_key, private_key = generate_key(8)
+ >>> public_key
+ (26569, 239)
+ >>> private_key
+ (26569, 2855)
+ """
+ p = rabin_miller.generate_large_prime(key_size)
+ q = rabin_miller.generate_large_prime(key_size)
n = p * q
- print("Generating e that is relatively prime to (p - 1) * (q - 1)...")
+ # Generate e that is relatively prime to (p - 1) * (q - 1)
while True:
- e = random.randrange(2 ** (keySize - 1), 2 ** (keySize))
- if cryptoMath.gcd(e, (p - 1) * (q - 1)) == 1:
+ e = random.randrange(2 ** (key_size - 1), 2 ** (key_size))
+ if gcd_by_iterative(e, (p - 1) * (q - 1)) == 1:
break
- print("Calculating d that is mod inverse of e...")
- d = cryptoMath.findModInverse(e, (p - 1) * (q - 1))
+ # Calculate d that is mod inverse of e
+ d = cryptomath_module.find_mod_inverse(e, (p - 1) * (q - 1))
- publicKey = (n, e)
- privateKey = (n, d)
- return (publicKey, privateKey)
+ public_key = (n, e)
+ private_key = (n, d)
+ return (public_key, private_key)
-def makeKeyFiles(name: int, keySize: int) -> None:
- if os.path.exists("%s_pubkey.txt" % (name)) or os.path.exists(
- "%s_privkey.txt" % (name)
- ):
+def make_key_files(name: str, key_size: int) -> None:
+ if os.path.exists(f"{name}_pubkey.txt") or os.path.exists(f"{name}_privkey.txt"):
print("\nWARNING:")
print(
- '"%s_pubkey.txt" or "%s_privkey.txt" already exists. \n'
+ f'"{name}_pubkey.txt" or "{name}_privkey.txt" already exists. \n'
"Use a different name or delete these files and re-run this program."
- % (name, name)
)
sys.exit()
- publicKey, privateKey = generateKey(keySize)
- print("\nWriting public key to file %s_pubkey.txt..." % name)
- with open("%s_pubkey.txt" % name, "w") as out_file:
- out_file.write("{},{},{}".format(keySize, publicKey[0], publicKey[1]))
+ public_key, private_key = generate_key(key_size)
+ print(f"\nWriting public key to file {name}_pubkey.txt...")
+ with open(f"{name}_pubkey.txt", "w") as out_file:
+ out_file.write(f"{key_size},{public_key[0]},{public_key[1]}")
- print("Writing private key to file %s_privkey.txt..." % name)
- with open("%s_privkey.txt" % name, "w") as out_file:
- out_file.write("{},{},{}".format(keySize, privateKey[0], privateKey[1]))
+ print(f"Writing private key to file {name}_privkey.txt...")
+ with open(f"{name}_privkey.txt", "w") as out_file:
+ out_file.write(f"{key_size},{private_key[0]},{private_key[1]}")
if __name__ == "__main__":
diff --git a/ciphers/running_key_cipher.py b/ciphers/running_key_cipher.py
new file mode 100644
index 000000000000..6bda417be898
--- /dev/null
+++ b/ciphers/running_key_cipher.py
@@ -0,0 +1,75 @@
+"""
+https://en.wikipedia.org/wiki/Running_key_cipher
+"""
+
+
+def running_key_encrypt(key: str, plaintext: str) -> str:
+ """
+ Encrypts the plaintext using the Running Key Cipher.
+
+ :param key: The running key (long piece of text).
+ :param plaintext: The plaintext to be encrypted.
+ :return: The ciphertext.
+ """
+ plaintext = plaintext.replace(" ", "").upper()
+ key = key.replace(" ", "").upper()
+ key_length = len(key)
+ ciphertext = []
+ ord_a = ord("A")
+
+ for i, char in enumerate(plaintext):
+ p = ord(char) - ord_a
+ k = ord(key[i % key_length]) - ord_a
+ c = (p + k) % 26
+ ciphertext.append(chr(c + ord_a))
+
+ return "".join(ciphertext)
+
+
+def running_key_decrypt(key: str, ciphertext: str) -> str:
+ """
+ Decrypts the ciphertext using the Running Key Cipher.
+
+ :param key: The running key (long piece of text).
+ :param ciphertext: The ciphertext to be decrypted.
+ :return: The plaintext.
+ """
+ ciphertext = ciphertext.replace(" ", "").upper()
+ key = key.replace(" ", "").upper()
+ key_length = len(key)
+ plaintext = []
+ ord_a = ord("A")
+
+ for i, char in enumerate(ciphertext):
+ c = ord(char) - ord_a
+ k = ord(key[i % key_length]) - ord_a
+ p = (c - k) % 26
+ plaintext.append(chr(p + ord_a))
+
+ return "".join(plaintext)
+
+
+def test_running_key_encrypt() -> None:
+ """
+ >>> key = "How does the duck know that? said Victor"
+ >>> ciphertext = running_key_encrypt(key, "DEFEND THIS")
+ >>> running_key_decrypt(key, ciphertext) == "DEFENDTHIS"
+ True
+ """
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ test_running_key_encrypt()
+
+ plaintext = input("Enter the plaintext: ").upper()
+ print(f"\n{plaintext = }")
+
+ key = "How does the duck know that? said Victor"
+ encrypted_text = running_key_encrypt(key, plaintext)
+ print(f"{encrypted_text = }")
+
+ decrypted_text = running_key_decrypt(key, encrypted_text)
+ print(f"{decrypted_text = }")
diff --git a/ciphers/shuffled_shift_cipher.py b/ciphers/shuffled_shift_cipher.py
index 22628f3c9d9e..08b2cab97c69 100644
--- a/ciphers/shuffled_shift_cipher.py
+++ b/ciphers/shuffled_shift_cipher.py
@@ -1,3 +1,5 @@
+from __future__ import annotations
+
import random
import string
@@ -7,7 +9,7 @@ class ShuffledShiftCipher:
This algorithm uses the Caesar Cipher algorithm but removes the option to
use brute force to decrypt the message.
- The passcode is a a random password from the selection buffer of
+ The passcode is a random password from the selection buffer of
1. uppercase letters of the English alphabet
2. lowercase letters of the English alphabet
3. digits from 0 to 9
@@ -26,7 +28,7 @@ class ShuffledShiftCipher:
cip2 = ShuffledShiftCipher()
"""
- def __init__(self, passcode: str = None):
+ def __init__(self, passcode: str | None = None) -> None:
"""
Initializes a cipher object with a passcode as it's entity
Note: No new passcode is generated if user provides a passcode
@@ -36,13 +38,13 @@ def __init__(self, passcode: str = None):
self.__key_list = self.__make_key_list()
self.__shift_key = self.__make_shift_key()
- def __str__(self):
+ def __str__(self) -> str:
"""
:return: passcode of the cipher object
"""
- return "Passcode is: " + "".join(self.__passcode)
+ return "".join(self.__passcode)
- def __neg_pos(self, iterlist: list) -> list:
+ def __neg_pos(self, iterlist: list[int]) -> list[int]:
"""
Mutates the list by changing the sign of each alternate element
@@ -54,7 +56,7 @@ def __neg_pos(self, iterlist: list) -> list:
iterlist[i] *= -1
return iterlist
- def __passcode_creator(self) -> list:
+ def __passcode_creator(self) -> list[str]:
"""
Creates a random password from the selection buffer of
1. uppercase letters of the English alphabet
@@ -65,10 +67,10 @@ def __passcode_creator(self) -> list:
:return: a password of a random length between 10 to 20
"""
choices = string.ascii_letters + string.digits
- password = [random.choice(choices) for i in range(random.randint(10, 20))]
+ password = [random.choice(choices) for _ in range(random.randint(10, 20))]
return password
- def __make_key_list(self) -> list:
+ def __make_key_list(self) -> list[str]:
"""
Shuffles the ordered character choices by pivoting at breakpoints
Breakpoints are the set of characters in the passcode
@@ -99,7 +101,7 @@ def __make_key_list(self) -> list:
# creates points known as breakpoints to break the key_list_options at those
# points and pivot each substring
breakpoints = sorted(set(self.__passcode))
- temp_list = []
+ temp_list: list[str] = []
# algorithm for creating a new shuffled list, keys_l, out of key_list_options
for i in key_list_options:
@@ -109,7 +111,7 @@ def __make_key_list(self) -> list:
# keys_l
if i in breakpoints or i == key_list_options[-1]:
keys_l.extend(temp_list[::-1])
- temp_list = []
+ temp_list.clear()
# returning a shuffled keys_l to prevent brute force guessing of shift key
return keys_l
@@ -167,7 +169,7 @@ def encrypt(self, plaintext: str) -> str:
return encoded_message
-def test_end_to_end(msg: str = "Hello, this is a modified Caesar cipher"):
+def test_end_to_end(msg: str = "Hello, this is a modified Caesar cipher") -> str:
"""
>>> test_end_to_end()
'Hello, this is a modified Caesar cipher'
diff --git a/ciphers/simple_keyword_cypher.py b/ciphers/simple_keyword_cypher.py
index 71c3083e9dfc..bde137d826c3 100644
--- a/ciphers/simple_keyword_cypher.py
+++ b/ciphers/simple_keyword_cypher.py
@@ -1,27 +1,30 @@
def remove_duplicates(key: str) -> str:
"""
Removes duplicate alphabetic characters in a keyword (letter is ignored after its
- first appearance).
+ first appearance).
+
:param key: Keyword to use
:return: String with duplicates removed
+
>>> remove_duplicates('Hello World!!')
'Helo Wrd'
"""
key_no_dups = ""
for ch in key:
- if ch == " " or ch not in key_no_dups and ch.isalpha():
+ if ch == " " or (ch not in key_no_dups and ch.isalpha()):
key_no_dups += ch
return key_no_dups
-def create_cipher_map(key: str) -> dict:
+def create_cipher_map(key: str) -> dict[str, str]:
"""
Returns a cipher map given a keyword.
+
:param key: keyword to use
:return: dictionary cipher map
"""
- # Create alphabet list
+ # Create a list of the letters in the alphabet
alphabet = [chr(i + 65) for i in range(26)]
# Remove duplicate characters from key
key = remove_duplicates(key.upper())
@@ -40,24 +43,28 @@ def create_cipher_map(key: str) -> dict:
return cipher_alphabet
-def encipher(message: str, cipher_map: dict) -> str:
+def encipher(message: str, cipher_map: dict[str, str]) -> str:
"""
Enciphers a message given a cipher map.
+
:param message: Message to encipher
:param cipher_map: Cipher map
:return: enciphered string
+
>>> encipher('Hello World!!', create_cipher_map('Goodbye!!'))
'CYJJM VMQJB!!'
"""
return "".join(cipher_map.get(ch, ch) for ch in message.upper())
-def decipher(message: str, cipher_map: dict) -> str:
+def decipher(message: str, cipher_map: dict[str, str]) -> str:
"""
Deciphers a message given a cipher map
+
:param message: Message to decipher
:param cipher_map: Dictionary mapping to use
:return: Deciphered string
+
>>> cipher_map = create_cipher_map('Goodbye!!')
>>> decipher(encipher('Hello World!!', cipher_map), cipher_map)
'HELLO WORLD!!'
@@ -67,9 +74,10 @@ def decipher(message: str, cipher_map: dict) -> str:
return "".join(rev_cipher_map.get(ch, ch) for ch in message.upper())
-def main():
+def main() -> None:
"""
Handles I/O
+
:return: void
"""
message = input("Enter message to encode or decode: ").strip()
diff --git a/ciphers/simple_substitution_cipher.py b/ciphers/simple_substitution_cipher.py
index 646ea449fc06..291a9bccd771 100644
--- a/ciphers/simple_substitution_cipher.py
+++ b/ciphers/simple_substitution_cipher.py
@@ -4,71 +4,71 @@
LETTERS = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
-def main():
+def main() -> None:
message = input("Enter message: ")
key = "LFWOAYUISVKMNXPBDCRJTQEGHZ"
resp = input("Encrypt/Decrypt [e/d]: ")
- checkValidKey(key)
+ check_valid_key(key)
if resp.lower().startswith("e"):
mode = "encrypt"
- translated = encryptMessage(key, message)
+ translated = encrypt_message(key, message)
elif resp.lower().startswith("d"):
mode = "decrypt"
- translated = decryptMessage(key, message)
+ translated = decrypt_message(key, message)
print(f"\n{mode.title()}ion: \n{translated}")
-def checkValidKey(key: str) -> None:
- keyList = list(key)
- lettersList = list(LETTERS)
- keyList.sort()
- lettersList.sort()
+def check_valid_key(key: str) -> None:
+ key_list = list(key)
+ letters_list = list(LETTERS)
+ key_list.sort()
+ letters_list.sort()
- if keyList != lettersList:
+ if key_list != letters_list:
sys.exit("Error in the key or symbol set.")
-def encryptMessage(key: str, message: str) -> str:
+def encrypt_message(key: str, message: str) -> str:
"""
- >>> encryptMessage('LFWOAYUISVKMNXPBDCRJTQEGHZ', 'Harshil Darji')
+ >>> encrypt_message('LFWOAYUISVKMNXPBDCRJTQEGHZ', 'Harshil Darji')
'Ilcrism Olcvs'
"""
- return translateMessage(key, message, "encrypt")
+ return translate_message(key, message, "encrypt")
-def decryptMessage(key: str, message: str) -> str:
+def decrypt_message(key: str, message: str) -> str:
"""
- >>> decryptMessage('LFWOAYUISVKMNXPBDCRJTQEGHZ', 'Ilcrism Olcvs')
+ >>> decrypt_message('LFWOAYUISVKMNXPBDCRJTQEGHZ', 'Ilcrism Olcvs')
'Harshil Darji'
"""
- return translateMessage(key, message, "decrypt")
+ return translate_message(key, message, "decrypt")
-def translateMessage(key: str, message: str, mode: str) -> str:
+def translate_message(key: str, message: str, mode: str) -> str:
translated = ""
- charsA = LETTERS
- charsB = key
+ chars_a = LETTERS
+ chars_b = key
if mode == "decrypt":
- charsA, charsB = charsB, charsA
+ chars_a, chars_b = chars_b, chars_a
for symbol in message:
- if symbol.upper() in charsA:
- symIndex = charsA.find(symbol.upper())
+ if symbol.upper() in chars_a:
+ sym_index = chars_a.find(symbol.upper())
if symbol.isupper():
- translated += charsB[symIndex].upper()
+ translated += chars_b[sym_index].upper()
else:
- translated += charsB[symIndex].lower()
+ translated += chars_b[sym_index].lower()
else:
translated += symbol
return translated
-def getRandomKey():
+def get_random_key() -> str:
key = list(LETTERS)
random.shuffle(key)
return "".join(key)
diff --git a/ciphers/trafid_cipher.py b/ciphers/trafid_cipher.py
deleted file mode 100644
index 328814f97744..000000000000
--- a/ciphers/trafid_cipher.py
+++ /dev/null
@@ -1,124 +0,0 @@
-# https://en.wikipedia.org/wiki/Trifid_cipher
-
-
-def __encryptPart(messagePart: str, character2Number: dict) -> str:
- one, two, three = "", "", ""
- tmp = []
-
- for character in messagePart:
- tmp.append(character2Number[character])
-
- for each in tmp:
- one += each[0]
- two += each[1]
- three += each[2]
-
- return one + two + three
-
-
-def __decryptPart(messagePart: str, character2Number: dict) -> (str, str, str):
- tmp, thisPart = "", ""
- result = []
-
- for character in messagePart:
- thisPart += character2Number[character]
-
- for digit in thisPart:
- tmp += digit
- if len(tmp) == len(messagePart):
- result.append(tmp)
- tmp = ""
-
- return result[0], result[1], result[2]
-
-
-def __prepare(message: str, alphabet: str) -> (str, str, dict, dict):
- # Validate message and alphabet, set to upper and remove spaces
- alphabet = alphabet.replace(" ", "").upper()
- message = message.replace(" ", "").upper()
-
- # Check length and characters
- if len(alphabet) != 27:
- raise KeyError("Length of alphabet has to be 27.")
- for each in message:
- if each not in alphabet:
- raise ValueError("Each message character has to be included in alphabet!")
-
- # Generate dictionares
- numbers = (
- "111",
- "112",
- "113",
- "121",
- "122",
- "123",
- "131",
- "132",
- "133",
- "211",
- "212",
- "213",
- "221",
- "222",
- "223",
- "231",
- "232",
- "233",
- "311",
- "312",
- "313",
- "321",
- "322",
- "323",
- "331",
- "332",
- "333",
- )
- character2Number = {}
- number2Character = {}
- for letter, number in zip(alphabet, numbers):
- character2Number[letter] = number
- number2Character[number] = letter
-
- return message, alphabet, character2Number, number2Character
-
-
-def encryptMessage(
- message: str, alphabet: str = "ABCDEFGHIJKLMNOPQRSTUVWXYZ.", period: int = 5
-) -> str:
- message, alphabet, character2Number, number2Character = __prepare(message, alphabet)
- encrypted, encrypted_numeric = "", ""
-
- for i in range(0, len(message) + 1, period):
- encrypted_numeric += __encryptPart(message[i : i + period], character2Number)
-
- for i in range(0, len(encrypted_numeric), 3):
- encrypted += number2Character[encrypted_numeric[i : i + 3]]
-
- return encrypted
-
-
-def decryptMessage(
- message: str, alphabet: str = "ABCDEFGHIJKLMNOPQRSTUVWXYZ.", period: int = 5
-) -> str:
- message, alphabet, character2Number, number2Character = __prepare(message, alphabet)
- decrypted_numeric = []
- decrypted = ""
-
- for i in range(0, len(message) + 1, period):
- a, b, c = __decryptPart(message[i : i + period], character2Number)
-
- for j in range(0, len(a)):
- decrypted_numeric.append(a[j] + b[j] + c[j])
-
- for each in decrypted_numeric:
- decrypted += number2Character[each]
-
- return decrypted
-
-
-if __name__ == "__main__":
- msg = "DEFEND THE EAST WALL OF THE CASTLE."
- encrypted = encryptMessage(msg, "EPSDUCVWYM.ZLKXNBTFGORIJHAQ")
- decrypted = decryptMessage(encrypted, "EPSDUCVWYM.ZLKXNBTFGORIJHAQ")
- print(f"Encrypted: {encrypted}\nDecrypted: {decrypted}")
diff --git a/ciphers/transposition_cipher.py b/ciphers/transposition_cipher.py
index 6a0a22d3e31d..76178cb6a1bc 100644
--- a/ciphers/transposition_cipher.py
+++ b/ciphers/transposition_cipher.py
@@ -8,59 +8,57 @@
"""
-def main():
+def main() -> None:
message = input("Enter message: ")
- key = int(input("Enter key [2-%s]: " % (len(message) - 1)))
+ key = int(input(f"Enter key [2-{len(message) - 1}]: "))
mode = input("Encryption/Decryption [e/d]: ")
if mode.lower().startswith("e"):
- text = encryptMessage(key, message)
+ text = encrypt_message(key, message)
elif mode.lower().startswith("d"):
- text = decryptMessage(key, message)
+ text = decrypt_message(key, message)
# Append pipe symbol (vertical bar) to identify spaces at the end.
- print("Output:\n%s" % (text + "|"))
+ print(f"Output:\n{text + '|'}")
-def encryptMessage(key: int, message: str) -> str:
+def encrypt_message(key: int, message: str) -> str:
"""
- >>> encryptMessage(6, 'Harshil Darji')
+ >>> encrypt_message(6, 'Harshil Darji')
'Hlia rDsahrij'
"""
- cipherText = [""] * key
+ cipher_text = [""] * key
for col in range(key):
pointer = col
while pointer < len(message):
- cipherText[col] += message[pointer]
+ cipher_text[col] += message[pointer]
pointer += key
- return "".join(cipherText)
+ return "".join(cipher_text)
-def decryptMessage(key: int, message: str) -> str:
+def decrypt_message(key: int, message: str) -> str:
"""
- >>> decryptMessage(6, 'Hlia rDsahrij')
+ >>> decrypt_message(6, 'Hlia rDsahrij')
'Harshil Darji'
"""
- numCols = math.ceil(len(message) / key)
- numRows = key
- numShadedBoxes = (numCols * numRows) - len(message)
- plainText = [""] * numCols
+ num_cols = math.ceil(len(message) / key)
+ num_rows = key
+ num_shaded_boxes = (num_cols * num_rows) - len(message)
+ plain_text = [""] * num_cols
col = 0
row = 0
for symbol in message:
- plainText[col] += symbol
+ plain_text[col] += symbol
col += 1
- if (
- (col == numCols)
- or (col == numCols - 1)
- and (row >= numRows - numShadedBoxes)
+ if (col == num_cols) or (
+ (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes)
):
col = 0
row += 1
- return "".join(plainText)
+ return "".join(plain_text)
if __name__ == "__main__":
diff --git a/ciphers/transposition_cipher_encrypt_decrypt_file.py b/ciphers/transposition_cipher_encrypt_decrypt_file.py
index 45aab056109a..b9630243d7f3 100644
--- a/ciphers/transposition_cipher_encrypt_decrypt_file.py
+++ b/ciphers/transposition_cipher_encrypt_decrypt_file.py
@@ -2,39 +2,39 @@
import sys
import time
-from . import transposition_cipher as transCipher
+from . import transposition_cipher as trans_cipher
-def main():
- inputFile = "Prehistoric Men.txt"
- outputFile = "Output.txt"
+def main() -> None:
+ input_file = "./prehistoric_men.txt"
+ output_file = "./Output.txt"
key = int(input("Enter key: "))
mode = input("Encrypt/Decrypt [e/d]: ")
- if not os.path.exists(inputFile):
- print("File %s does not exist. Quitting..." % inputFile)
+ if not os.path.exists(input_file):
+ print(f"File {input_file} does not exist. Quitting...")
sys.exit()
- if os.path.exists(outputFile):
- print("Overwrite %s? [y/n]" % outputFile)
+ if os.path.exists(output_file):
+ print(f"Overwrite {output_file}? [y/n]")
response = input("> ")
if not response.lower().startswith("y"):
sys.exit()
- startTime = time.time()
+ start_time = time.time()
if mode.lower().startswith("e"):
- with open(inputFile) as f:
+ with open(input_file) as f:
content = f.read()
- translated = transCipher.encryptMessage(key, content)
+ translated = trans_cipher.encrypt_message(key, content)
elif mode.lower().startswith("d"):
- with open(outputFile) as f:
+ with open(output_file) as f:
content = f.read()
- translated = transCipher.decryptMessage(key, content)
+ translated = trans_cipher.decrypt_message(key, content)
- with open(outputFile, "w") as outputObj:
- outputObj.write(translated)
+ with open(output_file, "w") as output_obj:
+ output_obj.write(translated)
- totalTime = round(time.time() - startTime, 2)
- print(("Done (", totalTime, "seconds )"))
+ total_time = round(time.time() - start_time, 2)
+ print(("Done (", total_time, "seconds )"))
if __name__ == "__main__":
diff --git a/ciphers/trifid_cipher.py b/ciphers/trifid_cipher.py
new file mode 100644
index 000000000000..13a47e9dd03b
--- /dev/null
+++ b/ciphers/trifid_cipher.py
@@ -0,0 +1,215 @@
+"""
+The trifid cipher uses a table to fractionate each plaintext letter into a trigram,
+mixes the constituents of the trigrams, and then applies the table in reverse to turn
+these mixed trigrams into ciphertext letters.
+
+https://en.wikipedia.org/wiki/Trifid_cipher
+"""
+
+from __future__ import annotations
+
+# fmt: off
+TEST_CHARACTER_TO_NUMBER = {
+ "A": "111", "B": "112", "C": "113", "D": "121", "E": "122", "F": "123", "G": "131",
+ "H": "132", "I": "133", "J": "211", "K": "212", "L": "213", "M": "221", "N": "222",
+ "O": "223", "P": "231", "Q": "232", "R": "233", "S": "311", "T": "312", "U": "313",
+ "V": "321", "W": "322", "X": "323", "Y": "331", "Z": "332", "+": "333",
+}
+# fmt: off
+
+TEST_NUMBER_TO_CHARACTER = {val: key for key, val in TEST_CHARACTER_TO_NUMBER.items()}
+
+
+def __encrypt_part(message_part: str, character_to_number: dict[str, str]) -> str:
+ """
+ Arrange the triagram value of each letter of `message_part` vertically and join
+ them horizontally.
+
+ >>> __encrypt_part('ASK', TEST_CHARACTER_TO_NUMBER)
+ '132111112'
+ """
+ one, two, three = "", "", ""
+ for each in (character_to_number[character] for character in message_part):
+ one += each[0]
+ two += each[1]
+ three += each[2]
+
+ return one + two + three
+
+
+def __decrypt_part(
+ message_part: str, character_to_number: dict[str, str]
+) -> tuple[str, str, str]:
+ """
+ Convert each letter of the input string into their respective trigram values, join
+ them and split them into three equal groups of strings which are returned.
+
+ >>> __decrypt_part('ABCDE', TEST_CHARACTER_TO_NUMBER)
+ ('11111', '21131', '21122')
+ """
+ this_part = "".join(character_to_number[character] for character in message_part)
+ result = []
+ tmp = ""
+ for digit in this_part:
+ tmp += digit
+ if len(tmp) == len(message_part):
+ result.append(tmp)
+ tmp = ""
+
+ return result[0], result[1], result[2]
+
+
+def __prepare(
+ message: str, alphabet: str
+) -> tuple[str, str, dict[str, str], dict[str, str]]:
+ """
+ A helper function that generates the triagrams and assigns each letter of the
+ alphabet to its corresponding triagram and stores this in a dictionary
+ (`character_to_number` and `number_to_character`) after confirming if the
+ alphabet's length is ``27``.
+
+ >>> test = __prepare('I aM a BOy','abCdeFghijkLmnopqrStuVwxYZ+')
+ >>> expected = ('IAMABOY','ABCDEFGHIJKLMNOPQRSTUVWXYZ+',
+ ... TEST_CHARACTER_TO_NUMBER, TEST_NUMBER_TO_CHARACTER)
+ >>> test == expected
+ True
+
+ Testing with incomplete alphabet
+
+ >>> __prepare('I aM a BOy','abCdeFghijkLmnopqrStuVw')
+ Traceback (most recent call last):
+ ...
+ KeyError: 'Length of alphabet has to be 27.'
+
+ Testing with extra long alphabets
+
+ >>> __prepare('I aM a BOy','abCdeFghijkLmnopqrStuVwxyzzwwtyyujjgfd')
+ Traceback (most recent call last):
+ ...
+ KeyError: 'Length of alphabet has to be 27.'
+
+ Testing with punctuation not in the given alphabet
+
+ >>> __prepare('am i a boy?','abCdeFghijkLmnopqrStuVwxYZ+')
+ Traceback (most recent call last):
+ ...
+ ValueError: Each message character has to be included in alphabet!
+
+ Testing with numbers
+
+ >>> __prepare(500,'abCdeFghijkLmnopqrStuVwxYZ+')
+ Traceback (most recent call last):
+ ...
+ AttributeError: 'int' object has no attribute 'replace'
+ """
+ # Validate message and alphabet, set to upper and remove spaces
+ alphabet = alphabet.replace(" ", "").upper()
+ message = message.replace(" ", "").upper()
+
+ # Check length and characters
+ if len(alphabet) != 27:
+ raise KeyError("Length of alphabet has to be 27.")
+ if any(char not in alphabet for char in message):
+ raise ValueError("Each message character has to be included in alphabet!")
+
+ # Generate dictionares
+ character_to_number = dict(zip(alphabet, TEST_CHARACTER_TO_NUMBER.values()))
+ number_to_character = {
+ number: letter for letter, number in character_to_number.items()
+ }
+
+ return message, alphabet, character_to_number, number_to_character
+
+
+def encrypt_message(
+ message: str, alphabet: str = "ABCDEFGHIJKLMNOPQRSTUVWXYZ.", period: int = 5
+) -> str:
+ """
+ encrypt_message
+ ===============
+
+ Encrypts a message using the trifid_cipher. Any punctuatuion chars that
+ would be used should be added to the alphabet.
+
+ PARAMETERS
+ ----------
+
+ * `message`: The message you want to encrypt.
+ * `alphabet` (optional): The characters to be used for the cipher .
+ * `period` (optional): The number of characters you want in a group whilst
+ encrypting.
+
+ >>> encrypt_message('I am a boy')
+ 'BCDGBQY'
+
+ >>> encrypt_message(' ')
+ ''
+
+ >>> encrypt_message(' aide toi le c iel ta id era ',
+ ... 'FELIXMARDSTBCGHJKNOPQUVWYZ+',5)
+ 'FMJFVOISSUFTFPUFEQQC'
+
+ """
+ message, alphabet, character_to_number, number_to_character = __prepare(
+ message, alphabet
+ )
+
+ encrypted_numeric = ""
+ for i in range(0, len(message) + 1, period):
+ encrypted_numeric += __encrypt_part(
+ message[i : i + period], character_to_number
+ )
+
+ encrypted = ""
+ for i in range(0, len(encrypted_numeric), 3):
+ encrypted += number_to_character[encrypted_numeric[i : i + 3]]
+ return encrypted
+
+
+def decrypt_message(
+ message: str, alphabet: str = "ABCDEFGHIJKLMNOPQRSTUVWXYZ.", period: int = 5
+) -> str:
+ """
+ decrypt_message
+ ===============
+
+ Decrypts a trifid_cipher encrypted message.
+
+ PARAMETERS
+ ----------
+
+ * `message`: The message you want to decrypt.
+ * `alphabet` (optional): The characters used for the cipher.
+ * `period` (optional): The number of characters used in grouping when it
+ was encrypted.
+
+ >>> decrypt_message('BCDGBQY')
+ 'IAMABOY'
+
+ Decrypting with your own alphabet and period
+
+ >>> decrypt_message('FMJFVOISSUFTFPUFEQQC','FELIXMARDSTBCGHJKNOPQUVWYZ+',5)
+ 'AIDETOILECIELTAIDERA'
+ """
+ message, alphabet, character_to_number, number_to_character = __prepare(
+ message, alphabet
+ )
+
+ decrypted_numeric = []
+ for i in range(0, len(message), period):
+ a, b, c = __decrypt_part(message[i : i + period], character_to_number)
+
+ for j in range(len(a)):
+ decrypted_numeric.append(a[j] + b[j] + c[j])
+
+ return "".join(number_to_character[each] for each in decrypted_numeric)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ msg = "DEFEND THE EAST WALL OF THE CASTLE."
+ encrypted = encrypt_message(msg, "EPSDUCVWYM.ZLKXNBTFGORIJHAQ")
+ decrypted = decrypt_message(encrypted, "EPSDUCVWYM.ZLKXNBTFGORIJHAQ")
+ print(f"Encrypted: {encrypted}\nDecrypted: {decrypted}")
diff --git a/ciphers/vernam_cipher.py b/ciphers/vernam_cipher.py
new file mode 100644
index 000000000000..197f28635a1c
--- /dev/null
+++ b/ciphers/vernam_cipher.py
@@ -0,0 +1,42 @@
+def vernam_encrypt(plaintext: str, key: str) -> str:
+ """
+ >>> vernam_encrypt("HELLO","KEY")
+ 'RIJVS'
+ """
+ ciphertext = ""
+ for i in range(len(plaintext)):
+ ct = ord(key[i % len(key)]) - 65 + ord(plaintext[i]) - 65
+ while ct > 25:
+ ct = ct - 26
+ ciphertext += chr(65 + ct)
+ return ciphertext
+
+
+def vernam_decrypt(ciphertext: str, key: str) -> str:
+ """
+ >>> vernam_decrypt("RIJVS","KEY")
+ 'HELLO'
+ """
+ decrypted_text = ""
+ for i in range(len(ciphertext)):
+ ct = ord(ciphertext[i]) - ord(key[i % len(key)])
+ while ct < 0:
+ ct = 26 + ct
+ decrypted_text += chr(65 + ct)
+ return decrypted_text
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+
+ # Example usage
+ plaintext = "HELLO"
+ key = "KEY"
+ encrypted_text = vernam_encrypt(plaintext, key)
+ decrypted_text = vernam_decrypt(encrypted_text, key)
+ print("\n\n")
+ print("Plaintext:", plaintext)
+ print("Encrypted:", encrypted_text)
+ print("Decrypted:", decrypted_text)
diff --git a/ciphers/vigenere_cipher.py b/ciphers/vigenere_cipher.py
index eb523d078005..e76161351fb1 100644
--- a/ciphers/vigenere_cipher.py
+++ b/ciphers/vigenere_cipher.py
@@ -1,50 +1,50 @@
LETTERS = "ABCDEFGHIJKLMNOPQRSTUVWXYZ"
-def main():
+def main() -> None:
message = input("Enter message: ")
key = input("Enter key [alphanumeric]: ")
mode = input("Encrypt/Decrypt [e/d]: ")
if mode.lower().startswith("e"):
mode = "encrypt"
- translated = encryptMessage(key, message)
+ translated = encrypt_message(key, message)
elif mode.lower().startswith("d"):
mode = "decrypt"
- translated = decryptMessage(key, message)
+ translated = decrypt_message(key, message)
- print("\n%sed message:" % mode.title())
+ print(f"\n{mode.title()}ed message:")
print(translated)
-def encryptMessage(key: str, message: str) -> str:
+def encrypt_message(key: str, message: str) -> str:
"""
- >>> encryptMessage('HDarji', 'This is Harshil Darji from Dharmaj.')
+ >>> encrypt_message('HDarji', 'This is Harshil Darji from Dharmaj.')
'Akij ra Odrjqqs Gaisq muod Mphumrs.'
"""
- return translateMessage(key, message, "encrypt")
+ return translate_message(key, message, "encrypt")
-def decryptMessage(key: str, message: str) -> str:
+def decrypt_message(key: str, message: str) -> str:
"""
- >>> decryptMessage('HDarji', 'Akij ra Odrjqqs Gaisq muod Mphumrs.')
+ >>> decrypt_message('HDarji', 'Akij ra Odrjqqs Gaisq muod Mphumrs.')
'This is Harshil Darji from Dharmaj.'
"""
- return translateMessage(key, message, "decrypt")
+ return translate_message(key, message, "decrypt")
-def translateMessage(key: str, message: str, mode: str) -> str:
+def translate_message(key: str, message: str, mode: str) -> str:
translated = []
- keyIndex = 0
+ key_index = 0
key = key.upper()
for symbol in message:
num = LETTERS.find(symbol.upper())
if num != -1:
if mode == "encrypt":
- num += LETTERS.find(key[keyIndex])
+ num += LETTERS.find(key[key_index])
elif mode == "decrypt":
- num -= LETTERS.find(key[keyIndex])
+ num -= LETTERS.find(key[key_index])
num %= len(LETTERS)
@@ -53,9 +53,9 @@ def translateMessage(key: str, message: str, mode: str) -> str:
elif symbol.islower():
translated.append(LETTERS[num].lower())
- keyIndex += 1
- if keyIndex == len(key):
- keyIndex = 0
+ key_index += 1
+ if key_index == len(key):
+ key_index = 0
else:
translated.append(symbol)
return "".join(translated)
diff --git a/ciphers/xor_cipher.py b/ciphers/xor_cipher.py
index 32a350d4e61c..24d88a0fd588 100644
--- a/ciphers/xor_cipher.py
+++ b/ciphers/xor_cipher.py
@@ -1,22 +1,24 @@
"""
- author: Christian Bender
- date: 21.12.2017
- class: XORCipher
-
- This class implements the XOR-cipher algorithm and provides
- some useful methods for encrypting and decrypting strings and
- files.
-
- Overview about methods
-
- - encrypt : list of char
- - decrypt : list of char
- - encrypt_string : str
- - decrypt_string : str
- - encrypt_file : boolean
- - decrypt_file : boolean
+author: Christian Bender
+date: 21.12.2017
+class: XORCipher
+
+This class implements the XOR-cipher algorithm and provides
+some useful methods for encrypting and decrypting strings and
+files.
+
+Overview about methods
+
+- encrypt : list of char
+- decrypt : list of char
+- encrypt_string : str
+- decrypt_string : str
+- encrypt_file : boolean
+- decrypt_file : boolean
"""
+from __future__ import annotations
+
class XORCipher:
def __init__(self, key: int = 0):
@@ -28,55 +30,75 @@ def __init__(self, key: int = 0):
# private field
self.__key = key
- def encrypt(self, content: str, key: int) -> [str]:
+ def encrypt(self, content: str, key: int) -> list[str]:
"""
input: 'content' of type string and 'key' of type int
output: encrypted string 'content' as a list of chars
if key not passed the method uses the key by the constructor.
otherwise key = 1
+
+ Empty list
+ >>> XORCipher().encrypt("", 5)
+ []
+
+ One key
+ >>> XORCipher().encrypt("hallo welt", 1)
+ ['i', '`', 'm', 'm', 'n', '!', 'v', 'd', 'm', 'u']
+
+ Normal key
+ >>> XORCipher().encrypt("HALLO WELT", 32)
+ ['h', 'a', 'l', 'l', 'o', '\\x00', 'w', 'e', 'l', 't']
+
+ Key greater than 255
+ >>> XORCipher().encrypt("hallo welt", 256)
+ ['h', 'a', 'l', 'l', 'o', ' ', 'w', 'e', 'l', 't']
"""
# precondition
- assert isinstance(key, int) and isinstance(content, str)
+ assert isinstance(key, int)
+ assert isinstance(content, str)
key = key or self.__key or 1
- # make sure key can be any size
- while key > 255:
- key -= 255
+ # make sure key is an appropriate size
+ key %= 256
- # This will be returned
- ans = []
-
- for ch in content:
- ans.append(chr(ord(ch) ^ key))
+ return [chr(ord(ch) ^ key) for ch in content]
- return ans
-
- def decrypt(self, content: str, key: int) -> [str]:
+ def decrypt(self, content: str, key: int) -> list[str]:
"""
input: 'content' of type list and 'key' of type int
output: decrypted string 'content' as a list of chars
if key not passed the method uses the key by the constructor.
otherwise key = 1
+
+ Empty list
+ >>> XORCipher().decrypt("", 5)
+ []
+
+ One key
+ >>> XORCipher().decrypt("hallo welt", 1)
+ ['i', '`', 'm', 'm', 'n', '!', 'v', 'd', 'm', 'u']
+
+ Normal key
+ >>> XORCipher().decrypt("HALLO WELT", 32)
+ ['h', 'a', 'l', 'l', 'o', '\\x00', 'w', 'e', 'l', 't']
+
+ Key greater than 255
+ >>> XORCipher().decrypt("hallo welt", 256)
+ ['h', 'a', 'l', 'l', 'o', ' ', 'w', 'e', 'l', 't']
"""
# precondition
- assert isinstance(key, int) and isinstance(content, list)
+ assert isinstance(key, int)
+ assert isinstance(content, str)
key = key or self.__key or 1
- # make sure key can be any size
- while key > 255:
- key -= 255
-
- # This will be returned
- ans = []
+ # make sure key is an appropriate size
+ key %= 256
- for ch in content:
- ans.append(chr(ord(ch) ^ key))
-
- return ans
+ return [chr(ord(ch) ^ key) for ch in content]
def encrypt_string(self, content: str, key: int = 0) -> str:
"""
@@ -84,16 +106,32 @@ def encrypt_string(self, content: str, key: int = 0) -> str:
output: encrypted string 'content'
if key not passed the method uses the key by the constructor.
otherwise key = 1
+
+ Empty list
+ >>> XORCipher().encrypt_string("", 5)
+ ''
+
+ One key
+ >>> XORCipher().encrypt_string("hallo welt", 1)
+ 'i`mmn!vdmu'
+
+ Normal key
+ >>> XORCipher().encrypt_string("HALLO WELT", 32)
+ 'hallo\\x00welt'
+
+ Key greater than 255
+ >>> XORCipher().encrypt_string("hallo welt", 256)
+ 'hallo welt'
"""
# precondition
- assert isinstance(key, int) and isinstance(content, str)
+ assert isinstance(key, int)
+ assert isinstance(content, str)
key = key or self.__key or 1
- # make sure key can be any size
- while key > 255:
- key -= 255
+ # make sure key is an appropriate size
+ key %= 256
# This will be returned
ans = ""
@@ -109,16 +147,32 @@ def decrypt_string(self, content: str, key: int = 0) -> str:
output: decrypted string 'content'
if key not passed the method uses the key by the constructor.
otherwise key = 1
+
+ Empty list
+ >>> XORCipher().decrypt_string("", 5)
+ ''
+
+ One key
+ >>> XORCipher().decrypt_string("hallo welt", 1)
+ 'i`mmn!vdmu'
+
+ Normal key
+ >>> XORCipher().decrypt_string("HALLO WELT", 32)
+ 'hallo\\x00welt'
+
+ Key greater than 255
+ >>> XORCipher().decrypt_string("hallo welt", 256)
+ 'hallo welt'
"""
# precondition
- assert isinstance(key, int) and isinstance(content, str)
+ assert isinstance(key, int)
+ assert isinstance(content, str)
key = key or self.__key or 1
- # make sure key can be any size
- while key > 255:
- key -= 255
+ # make sure key is an appropriate size
+ key %= 256
# This will be returned
ans = ""
@@ -138,15 +192,17 @@ def encrypt_file(self, file: str, key: int = 0) -> bool:
"""
# precondition
- assert isinstance(file, str) and isinstance(key, int)
+ assert isinstance(file, str)
+ assert isinstance(key, int)
- try:
- with open(file) as fin:
- with open("encrypt.out", "w+") as fout:
+ # make sure key is an appropriate size
+ key %= 256
- # actual encrypt-process
- for line in fin:
- fout.write(self.encrypt_string(line, key))
+ try:
+ with open(file) as fin, open("encrypt.out", "w+") as fout:
+ # actual encrypt-process
+ for line in fin:
+ fout.write(self.encrypt_string(line, key))
except OSError:
return False
@@ -163,15 +219,17 @@ def decrypt_file(self, file: str, key: int) -> bool:
"""
# precondition
- assert isinstance(file, str) and isinstance(key, int)
+ assert isinstance(file, str)
+ assert isinstance(key, int)
- try:
- with open(file) as fin:
- with open("decrypt.out", "w+") as fout:
+ # make sure key is an appropriate size
+ key %= 256
- # actual encrypt-process
- for line in fin:
- fout.write(self.decrypt_string(line, key))
+ try:
+ with open(file) as fin, open("decrypt.out", "w+") as fout:
+ # actual encrypt-process
+ for line in fin:
+ fout.write(self.decrypt_string(line, key))
except OSError:
return False
@@ -179,6 +237,11 @@ def decrypt_file(self, file: str, key: int) -> bool:
return True
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+
# Tests
# crypt = XORCipher()
# key = 67
diff --git a/compression/README.md b/compression/README.md
new file mode 100644
index 000000000000..bad7ae1a2f76
--- /dev/null
+++ b/compression/README.md
@@ -0,0 +1,10 @@
+# Compression
+
+Data compression is everywhere, you need it to store data without taking too much space.
+Either the compression loses some data (then we talk about lossy compression, such as .jpg) or it does not (and then it is lossless compression, such as .png)
+
+Lossless compression is mainly used for archive purpose as it allows storing data without losing information about the file archived. On the other hand, lossy compression is used for transfer of file where quality isn't necessarily what is required (i.e: images on Twitter).
+
+*
+*
+*
diff --git a/compression/burrows_wheeler.py b/compression/burrows_wheeler.py
index 1a6610915e65..857d677c904e 100644
--- a/compression/burrows_wheeler.py
+++ b/compression/burrows_wheeler.py
@@ -1,7 +1,7 @@
"""
https://en.wikipedia.org/wiki/Burrows%E2%80%93Wheeler_transform
-The Burrows–Wheeler transform (BWT, also called block-sorting compression)
+The Burrows-Wheeler transform (BWT, also called block-sorting compression)
rearranges a character string into runs of similar characters. This is useful
for compression, since it tends to be easy to compress a string that has runs
of repeated characters by techniques such as move-to-front transform and
@@ -10,8 +10,16 @@
original character. The BWT is thus a "free" method of improving the efficiency
of text compression algorithms, costing only some extra computation.
"""
+
from __future__ import annotations
+from typing import TypedDict
+
+
+class BWTTransformDict(TypedDict):
+ bwt_string: str
+ idx_original_string: int
+
def all_rotations(s: str) -> list[str]:
"""
@@ -43,7 +51,7 @@ def all_rotations(s: str) -> list[str]:
return [s[i:] + s[:i] for i in range(len(s))]
-def bwt_transform(s: str) -> dict:
+def bwt_transform(s: str) -> BWTTransformDict:
"""
:param s: The string that will be used at bwt algorithm
:return: the string composed of the last char of each row of the ordered
@@ -75,10 +83,11 @@ def bwt_transform(s: str) -> dict:
rotations = all_rotations(s)
rotations.sort() # sort the list of rotations in alphabetically order
# make a string composed of the last char of each rotation
- return {
+ response: BWTTransformDict = {
"bwt_string": "".join([word[-1] for word in rotations]),
"idx_original_string": rotations.index(s),
}
+ return response
def reverse_bwt(bwt_string: str, idx_original_string: int) -> str:
@@ -142,11 +151,11 @@ def reverse_bwt(bwt_string: str, idx_original_string: int) -> str:
raise ValueError("The parameter idx_original_string must not be lower than 0.")
if idx_original_string >= len(bwt_string):
raise ValueError(
- "The parameter idx_original_string must be lower than" " len(bwt_string)."
+ "The parameter idx_original_string must be lower than len(bwt_string)."
)
ordered_rotations = [""] * len(bwt_string)
- for x in range(len(bwt_string)):
+ for _ in range(len(bwt_string)):
for i in range(len(bwt_string)):
ordered_rotations[i] = bwt_string[i] + ordered_rotations[i]
ordered_rotations.sort()
@@ -157,11 +166,12 @@ def reverse_bwt(bwt_string: str, idx_original_string: int) -> str:
entry_msg = "Provide a string that I will generate its BWT transform: "
s = input(entry_msg).strip()
result = bwt_transform(s)
- bwt_output_msg = "Burrows Wheeler transform for string '{}' results in '{}'"
- print(bwt_output_msg.format(s, result["bwt_string"]))
+ print(
+ f"Burrows Wheeler transform for string '{s}' results "
+ f"in '{result['bwt_string']}'"
+ )
original_string = reverse_bwt(result["bwt_string"], result["idx_original_string"])
- fmt = (
- "Reversing Burrows Wheeler transform for entry '{}' we get original"
- " string '{}'"
+ print(
+ f"Reversing Burrows Wheeler transform for entry '{result['bwt_string']}' "
+ f"we get original string '{original_string}'"
)
- print(fmt.format(result["bwt_string"], original_string))
diff --git a/compression/huffman.py b/compression/huffman.py
index 3a3cbfa4b0c6..44eda6c03180 100644
--- a/compression/huffman.py
+++ b/compression/huffman.py
@@ -1,84 +1,89 @@
+from __future__ import annotations
+
import sys
class Letter:
- def __init__(self, letter, freq):
- self.letter = letter
- self.freq = freq
- self.bitstring = ""
+ def __init__(self, letter: str, freq: int):
+ self.letter: str = letter
+ self.freq: int = freq
+ self.bitstring: dict[str, str] = {}
- def __repr__(self):
+ def __repr__(self) -> str:
return f"{self.letter}:{self.freq}"
class TreeNode:
- def __init__(self, freq, left, right):
- self.freq = freq
- self.left = left
- self.right = right
+ def __init__(self, freq: int, left: Letter | TreeNode, right: Letter | TreeNode):
+ self.freq: int = freq
+ self.left: Letter | TreeNode = left
+ self.right: Letter | TreeNode = right
-def parse_file(file_path):
+def parse_file(file_path: str) -> list[Letter]:
"""
Read the file and build a dict of all letters and their
frequencies, then convert the dict into a list of Letters.
"""
- chars = {}
+ chars: dict[str, int] = {}
with open(file_path) as f:
while True:
c = f.read(1)
if not c:
break
- chars[c] = chars[c] + 1 if c in chars.keys() else 1
- return sorted([Letter(c, f) for c, f in chars.items()], key=lambda l: l.freq)
+ chars[c] = chars[c] + 1 if c in chars else 1
+ return sorted((Letter(c, f) for c, f in chars.items()), key=lambda x: x.freq)
-def build_tree(letters):
+def build_tree(letters: list[Letter]) -> Letter | TreeNode:
"""
Run through the list of Letters and build the min heap
for the Huffman Tree.
"""
- while len(letters) > 1:
- left = letters.pop(0)
- right = letters.pop(0)
+ response: list[Letter | TreeNode] = list(letters)
+ while len(response) > 1:
+ left = response.pop(0)
+ right = response.pop(0)
total_freq = left.freq + right.freq
node = TreeNode(total_freq, left, right)
- letters.append(node)
- letters.sort(key=lambda l: l.freq)
- return letters[0]
+ response.append(node)
+ response.sort(key=lambda x: x.freq)
+ return response[0]
-def traverse_tree(root, bitstring):
+def traverse_tree(root: Letter | TreeNode, bitstring: str) -> list[Letter]:
"""
Recursively traverse the Huffman Tree to set each
- Letter's bitstring, and return the list of Letters
+ Letter's bitstring dictionary, and return the list of Letters
"""
- if type(root) is Letter:
- root.bitstring = bitstring
+ if isinstance(root, Letter):
+ root.bitstring[root.letter] = bitstring
return [root]
+ treenode: TreeNode = root
letters = []
- letters += traverse_tree(root.left, bitstring + "0")
- letters += traverse_tree(root.right, bitstring + "1")
+ letters += traverse_tree(treenode.left, bitstring + "0")
+ letters += traverse_tree(treenode.right, bitstring + "1")
return letters
-def huffman(file_path):
+def huffman(file_path: str) -> None:
"""
Parse the file, build the tree, then run through the file
- again, using the list of Letters to find and print out the
+ again, using the letters dictionary to find and print out the
bitstring for each letter.
"""
letters_list = parse_file(file_path)
root = build_tree(letters_list)
- letters = traverse_tree(root, "")
- print(f"Huffman Coding of {file_path}: ")
+ letters = {
+ k: v for letter in traverse_tree(root, "") for k, v in letter.bitstring.items()
+ }
+ print(f"Huffman Coding of {file_path}: ")
with open(file_path) as f:
while True:
c = f.read(1)
if not c:
break
- le = list(filter(lambda l: l.letter == c, letters))[0]
- print(le.bitstring, end=" ")
+ print(letters[c], end=" ")
print()
diff --git a/compression/lempel_ziv.py b/compression/lempel_ziv.py
index 2d0601b27b34..648b029471bd 100644
--- a/compression/lempel_ziv.py
+++ b/compression/lempel_ziv.py
@@ -1,6 +1,6 @@
"""
- One of the several implementations of Lempel–Ziv–Welch compression algorithm
- https://en.wikipedia.org/wiki/Lempel%E2%80%93Ziv%E2%80%93Welch
+One of the several implementations of Lempel-Ziv-Welch compression algorithm
+https://en.wikipedia.org/wiki/Lempel%E2%80%93Ziv%E2%80%93Welch
"""
import math
@@ -26,7 +26,7 @@ def read_file_binary(file_path: str) -> str:
def add_key_to_lexicon(
- lexicon: dict, curr_string: str, index: int, last_match_id: int
+ lexicon: dict[str, str], curr_string: str, index: int, last_match_id: str
) -> None:
"""
Adds new strings (curr_string + "0", curr_string + "1") to the lexicon
@@ -35,15 +35,15 @@ def add_key_to_lexicon(
lexicon[curr_string + "0"] = last_match_id
if math.log2(index).is_integer():
- for curr_key in lexicon:
- lexicon[curr_key] = "0" + lexicon[curr_key]
+ for curr_key, value in lexicon.items():
+ lexicon[curr_key] = f"0{value}"
lexicon[curr_string + "1"] = bin(index)[2:]
def compress_data(data_bits: str) -> str:
"""
- Compresses given data_bits using Lempel–Ziv–Welch compression algorithm
+ Compresses given data_bits using Lempel-Ziv-Welch compression algorithm
and returns the result as a string
"""
lexicon = {"0": "0", "1": "1"}
@@ -110,7 +110,7 @@ def write_file_binary(file_path: str, to_write: str) -> None:
sys.exit()
-def compress(source_path, destination_path: str) -> None:
+def compress(source_path: str, destination_path: str) -> None:
"""
Reads source file, compresses it and writes the compressed result in destination
file
diff --git a/compression/lempel_ziv_decompress.py b/compression/lempel_ziv_decompress.py
index 4d3c2c0d2cf3..225e96236c2c 100644
--- a/compression/lempel_ziv_decompress.py
+++ b/compression/lempel_ziv_decompress.py
@@ -1,6 +1,6 @@
"""
- One of the several implementations of Lempel–Ziv–Welch decompression algorithm
- https://en.wikipedia.org/wiki/Lempel%E2%80%93Ziv%E2%80%93Welch
+One of the several implementations of Lempel-Ziv-Welch decompression algorithm
+https://en.wikipedia.org/wiki/Lempel%E2%80%93Ziv%E2%80%93Welch
"""
import math
@@ -26,7 +26,7 @@ def read_file_binary(file_path: str) -> str:
def decompress_data(data_bits: str) -> str:
"""
- Decompresses given data_bits using Lempel–Ziv–Welch compression algorithm
+ Decompresses given data_bits using Lempel-Ziv-Welch compression algorithm
and returns the result as a string
"""
lexicon = {"0": "0", "1": "1"}
@@ -43,10 +43,10 @@ def decompress_data(data_bits: str) -> str:
lexicon[curr_string] = last_match_id + "0"
if math.log2(index).is_integer():
- newLex = {}
+ new_lex = {}
for curr_key in list(lexicon):
- newLex["0" + curr_key] = lexicon.pop(curr_key)
- lexicon = newLex
+ new_lex["0" + curr_key] = lexicon.pop(curr_key)
+ lexicon = new_lex
lexicon[bin(index)[2:]] = last_match_id + "1"
index += 1
diff --git a/compression/lz77.py b/compression/lz77.py
new file mode 100644
index 000000000000..09b8b021e9d5
--- /dev/null
+++ b/compression/lz77.py
@@ -0,0 +1,225 @@
+"""
+LZ77 compression algorithm
+- lossless data compression published in papers by Abraham Lempel and Jacob Ziv in 1977
+- also known as LZ1 or sliding-window compression
+- form the basis for many variations including LZW, LZSS, LZMA and others
+
+It uses a “sliding window” method. Within the sliding window we have:
+ - search buffer
+ - look ahead buffer
+len(sliding_window) = len(search_buffer) + len(look_ahead_buffer)
+
+LZ77 manages a dictionary that uses triples composed of:
+ - Offset into search buffer, it's the distance between the start of a phrase and
+ the beginning of a file.
+ - Length of the match, it's the number of characters that make up a phrase.
+ - The indicator is represented by a character that is going to be encoded next.
+
+As a file is parsed, the dictionary is dynamically updated to reflect the compressed
+data contents and size.
+
+Examples:
+"cabracadabrarrarrad" <-> [(0, 0, 'c'), (0, 0, 'a'), (0, 0, 'b'), (0, 0, 'r'),
+ (3, 1, 'c'), (2, 1, 'd'), (7, 4, 'r'), (3, 5, 'd')]
+"ababcbababaa" <-> [(0, 0, 'a'), (0, 0, 'b'), (2, 2, 'c'), (4, 3, 'a'), (2, 2, 'a')]
+"aacaacabcabaaac" <-> [(0, 0, 'a'), (1, 1, 'c'), (3, 4, 'b'), (3, 3, 'a'), (1, 2, 'c')]
+
+Sources:
+en.wikipedia.org/wiki/LZ77_and_LZ78
+"""
+
+from dataclasses import dataclass
+
+__version__ = "0.1"
+__author__ = "Lucia Harcekova"
+
+
+@dataclass
+class Token:
+ """
+ Dataclass representing triplet called token consisting of length, offset
+ and indicator. This triplet is used during LZ77 compression.
+ """
+
+ offset: int
+ length: int
+ indicator: str
+
+ def __repr__(self) -> str:
+ """
+ >>> token = Token(1, 2, "c")
+ >>> repr(token)
+ '(1, 2, c)'
+ >>> str(token)
+ '(1, 2, c)'
+ """
+ return f"({self.offset}, {self.length}, {self.indicator})"
+
+
+class LZ77Compressor:
+ """
+ Class containing compress and decompress methods using LZ77 compression algorithm.
+ """
+
+ def __init__(self, window_size: int = 13, lookahead_buffer_size: int = 6) -> None:
+ self.window_size = window_size
+ self.lookahead_buffer_size = lookahead_buffer_size
+ self.search_buffer_size = self.window_size - self.lookahead_buffer_size
+
+ def compress(self, text: str) -> list[Token]:
+ """
+ Compress the given string text using LZ77 compression algorithm.
+
+ Args:
+ text: string to be compressed
+
+ Returns:
+ output: the compressed text as a list of Tokens
+
+ >>> lz77_compressor = LZ77Compressor()
+ >>> str(lz77_compressor.compress("ababcbababaa"))
+ '[(0, 0, a), (0, 0, b), (2, 2, c), (4, 3, a), (2, 2, a)]'
+ >>> str(lz77_compressor.compress("aacaacabcabaaac"))
+ '[(0, 0, a), (1, 1, c), (3, 4, b), (3, 3, a), (1, 2, c)]'
+ """
+
+ output = []
+ search_buffer = ""
+
+ # while there are still characters in text to compress
+ while text:
+ # find the next encoding phrase
+ # - triplet with offset, length, indicator (the next encoding character)
+ token = self._find_encoding_token(text, search_buffer)
+
+ # update the search buffer:
+ # - add new characters from text into it
+ # - check if size exceed the max search buffer size, if so, drop the
+ # oldest elements
+ search_buffer += text[: token.length + 1]
+ if len(search_buffer) > self.search_buffer_size:
+ search_buffer = search_buffer[-self.search_buffer_size :]
+
+ # update the text
+ text = text[token.length + 1 :]
+
+ # append the token to output
+ output.append(token)
+
+ return output
+
+ def decompress(self, tokens: list[Token]) -> str:
+ """
+ Convert the list of tokens into an output string.
+
+ Args:
+ tokens: list containing triplets (offset, length, char)
+
+ Returns:
+ output: decompressed text
+
+ Tests:
+ >>> lz77_compressor = LZ77Compressor()
+ >>> lz77_compressor.decompress([Token(0, 0, 'c'), Token(0, 0, 'a'),
+ ... Token(0, 0, 'b'), Token(0, 0, 'r'), Token(3, 1, 'c'),
+ ... Token(2, 1, 'd'), Token(7, 4, 'r'), Token(3, 5, 'd')])
+ 'cabracadabrarrarrad'
+ >>> lz77_compressor.decompress([Token(0, 0, 'a'), Token(0, 0, 'b'),
+ ... Token(2, 2, 'c'), Token(4, 3, 'a'), Token(2, 2, 'a')])
+ 'ababcbababaa'
+ >>> lz77_compressor.decompress([Token(0, 0, 'a'), Token(1, 1, 'c'),
+ ... Token(3, 4, 'b'), Token(3, 3, 'a'), Token(1, 2, 'c')])
+ 'aacaacabcabaaac'
+ """
+
+ output = ""
+
+ for token in tokens:
+ for _ in range(token.length):
+ output += output[-token.offset]
+ output += token.indicator
+
+ return output
+
+ def _find_encoding_token(self, text: str, search_buffer: str) -> Token:
+ """Finds the encoding token for the first character in the text.
+
+ Tests:
+ >>> lz77_compressor = LZ77Compressor()
+ >>> lz77_compressor._find_encoding_token("abrarrarrad", "abracad").offset
+ 7
+ >>> lz77_compressor._find_encoding_token("adabrarrarrad", "cabrac").length
+ 1
+ >>> lz77_compressor._find_encoding_token("abc", "xyz").offset
+ 0
+ >>> lz77_compressor._find_encoding_token("", "xyz").offset
+ Traceback (most recent call last):
+ ...
+ ValueError: We need some text to work with.
+ >>> lz77_compressor._find_encoding_token("abc", "").offset
+ 0
+ """
+
+ if not text:
+ raise ValueError("We need some text to work with.")
+
+ # Initialise result parameters to default values
+ length, offset = 0, 0
+
+ if not search_buffer:
+ return Token(offset, length, text[length])
+
+ for i, character in enumerate(search_buffer):
+ found_offset = len(search_buffer) - i
+ if character == text[0]:
+ found_length = self._match_length_from_index(text, search_buffer, 0, i)
+ # if the found length is bigger than the current or if it's equal,
+ # which means it's offset is smaller: update offset and length
+ if found_length >= length:
+ offset, length = found_offset, found_length
+
+ return Token(offset, length, text[length])
+
+ def _match_length_from_index(
+ self, text: str, window: str, text_index: int, window_index: int
+ ) -> int:
+ """Calculate the longest possible match of text and window characters from
+ text_index in text and window_index in window.
+
+ Args:
+ text: _description_
+ window: sliding window
+ text_index: index of character in text
+ window_index: index of character in sliding window
+
+ Returns:
+ The maximum match between text and window, from given indexes.
+
+ Tests:
+ >>> lz77_compressor = LZ77Compressor(13, 6)
+ >>> lz77_compressor._match_length_from_index("rarrad", "adabrar", 0, 4)
+ 5
+ >>> lz77_compressor._match_length_from_index("adabrarrarrad",
+ ... "cabrac", 0, 1)
+ 1
+ """
+ if not text or text[text_index] != window[window_index]:
+ return 0
+ return 1 + self._match_length_from_index(
+ text, window + text[text_index], text_index + 1, window_index + 1
+ )
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ # Initialize compressor class
+ lz77_compressor = LZ77Compressor(window_size=13, lookahead_buffer_size=6)
+
+ # Example
+ TEXT = "cabracadabrarrarrad"
+ compressed_text = lz77_compressor.compress(TEXT)
+ print(lz77_compressor.compress("ababcbababaa"))
+ decompressed_text = lz77_compressor.decompress(compressed_text)
+ assert decompressed_text == TEXT, "The LZ77 algorithm returned the invalid result."
diff --git a/compression/peak_signal_to_noise_ratio.py b/compression/peak_signal_to_noise_ratio.py
index 6c6c4c38a12a..284f2904a21d 100644
--- a/compression/peak_signal_to_noise_ratio.py
+++ b/compression/peak_signal_to_noise_ratio.py
@@ -11,17 +11,18 @@
import cv2
import numpy as np
+PIXEL_MAX = 255.0
-def psnr(original, contrast):
+
+def peak_signal_to_noise_ratio(original: float, contrast: float) -> float:
mse = np.mean((original - contrast) ** 2)
if mse == 0:
return 100
- PIXEL_MAX = 255.0
- PSNR = 20 * math.log10(PIXEL_MAX / math.sqrt(mse))
- return PSNR
+
+ return 20 * math.log10(PIXEL_MAX / math.sqrt(mse))
-def main():
+def main() -> None:
dir_path = os.path.dirname(os.path.realpath(__file__))
# Loading images (original image and compressed image)
original = cv2.imread(os.path.join(dir_path, "image_data/original_image.png"))
@@ -34,11 +35,11 @@ def main():
# Value expected: 29.73dB
print("-- First Test --")
- print(f"PSNR value is {psnr(original, contrast)} dB")
+ print(f"PSNR value is {peak_signal_to_noise_ratio(original, contrast)} dB")
# # Value expected: 31.53dB (Wikipedia Example)
print("\n-- Second Test --")
- print(f"PSNR value is {psnr(original2, contrast2)} dB")
+ print(f"PSNR value is {peak_signal_to_noise_ratio(original2, contrast2)} dB")
if __name__ == "__main__":
diff --git a/compression/run_length_encoding.py b/compression/run_length_encoding.py
new file mode 100644
index 000000000000..691e19095dc6
--- /dev/null
+++ b/compression/run_length_encoding.py
@@ -0,0 +1,48 @@
+# https://en.wikipedia.org/wiki/Run-length_encoding
+
+
+def run_length_encode(text: str) -> list:
+ """
+ Performs Run Length Encoding
+ >>> run_length_encode("AAAABBBCCDAA")
+ [('A', 4), ('B', 3), ('C', 2), ('D', 1), ('A', 2)]
+ >>> run_length_encode("A")
+ [('A', 1)]
+ >>> run_length_encode("AA")
+ [('A', 2)]
+ >>> run_length_encode("AAADDDDDDFFFCCCAAVVVV")
+ [('A', 3), ('D', 6), ('F', 3), ('C', 3), ('A', 2), ('V', 4)]
+ """
+ encoded = []
+ count = 1
+
+ for i in range(len(text)):
+ if i + 1 < len(text) and text[i] == text[i + 1]:
+ count += 1
+ else:
+ encoded.append((text[i], count))
+ count = 1
+
+ return encoded
+
+
+def run_length_decode(encoded: list) -> str:
+ """
+ Performs Run Length Decoding
+ >>> run_length_decode([('A', 4), ('B', 3), ('C', 2), ('D', 1), ('A', 2)])
+ 'AAAABBBCCDAA'
+ >>> run_length_decode([('A', 1)])
+ 'A'
+ >>> run_length_decode([('A', 2)])
+ 'AA'
+ >>> run_length_decode([('A', 3), ('D', 6), ('F', 3), ('C', 3), ('A', 2), ('V', 4)])
+ 'AAADDDDDDFFFCCCAAVVVV'
+ """
+ return "".join(char * length for char, length in encoded)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod(name="run_length_encode", verbose=True)
+ testmod(name="run_length_decode", verbose=True)
diff --git a/computer_vision/README.md b/computer_vision/README.md
index 94ee493086cc..61462567b662 100644
--- a/computer_vision/README.md
+++ b/computer_vision/README.md
@@ -1,7 +1,10 @@
-### Computer Vision
+# Computer Vision
+
+Computer vision is a field of computer science that works on enabling computers to see, identify and process images in the same way that human does, and provide appropriate output.
-Computer vision is a field of computer science that works on enabling computers to see,
-identify and process images in the same way that human vision does, and then provide appropriate output.
It is like imparting human intelligence and instincts to a computer.
Image processing and computer vision are a little different from each other. Image processing means applying some algorithms for transforming image from one form to the other like smoothing, contrasting, stretching, etc.
+
While computer vision comes from modelling image processing using the techniques of machine learning, computer vision applies machine learning to recognize patterns for interpretation of images (much like the process of visual reasoning of human vision).
+
+*
diff --git a/computer_vision/cnn_classification.py b/computer_vision/cnn_classification.py
new file mode 100644
index 000000000000..115333eba0d1
--- /dev/null
+++ b/computer_vision/cnn_classification.py
@@ -0,0 +1,100 @@
+"""
+Convolutional Neural Network
+
+Objective : To train a CNN model detect if TB is present in Lung X-ray or not.
+
+Resources CNN Theory :
+ https://en.wikipedia.org/wiki/Convolutional_neural_network
+Resources Tensorflow : https://www.tensorflow.org/tutorials/images/cnn
+
+Download dataset from :
+https://lhncbc.nlm.nih.gov/LHC-publications/pubs/TuberculosisChestXrayImageDataSets.html
+
+1. Download the dataset folder and create two folder training set and test set
+in the parent dataset folder
+2. Move 30-40 image from both TB positive and TB Negative folder
+in the test set folder
+3. The labels of the images will be extracted from the folder name
+the image is present in.
+
+"""
+
+# Part 1 - Building the CNN
+
+import numpy as np
+
+# Importing the Keras libraries and packages
+import tensorflow as tf
+from keras import layers, models
+
+if __name__ == "__main__":
+ # Initialising the CNN
+ # (Sequential- Building the model layer by layer)
+ classifier = models.Sequential()
+
+ # Step 1 - Convolution
+ # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel
+ # (3,3) is the kernel size (filter matrix)
+ classifier.add(
+ layers.Conv2D(32, (3, 3), input_shape=(64, 64, 3), activation="relu")
+ )
+
+ # Step 2 - Pooling
+ classifier.add(layers.MaxPooling2D(pool_size=(2, 2)))
+
+ # Adding a second convolutional layer
+ classifier.add(layers.Conv2D(32, (3, 3), activation="relu"))
+ classifier.add(layers.MaxPooling2D(pool_size=(2, 2)))
+
+ # Step 3 - Flattening
+ classifier.add(layers.Flatten())
+
+ # Step 4 - Full connection
+ classifier.add(layers.Dense(units=128, activation="relu"))
+ classifier.add(layers.Dense(units=1, activation="sigmoid"))
+
+ # Compiling the CNN
+ classifier.compile(
+ optimizer="adam", loss="binary_crossentropy", metrics=["accuracy"]
+ )
+
+ # Part 2 - Fitting the CNN to the images
+
+ # Load Trained model weights
+
+ # from keras.models import load_model
+ # regressor=load_model('cnn.h5')
+
+ train_datagen = tf.keras.preprocessing.image.ImageDataGenerator(
+ rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True
+ )
+
+ test_datagen = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255)
+
+ training_set = train_datagen.flow_from_directory(
+ "dataset/training_set", target_size=(64, 64), batch_size=32, class_mode="binary"
+ )
+
+ test_set = test_datagen.flow_from_directory(
+ "dataset/test_set", target_size=(64, 64), batch_size=32, class_mode="binary"
+ )
+
+ classifier.fit_generator(
+ training_set, steps_per_epoch=5, epochs=30, validation_data=test_set
+ )
+
+ classifier.save("cnn.h5")
+
+ # Part 3 - Making new predictions
+
+ test_image = tf.keras.preprocessing.image.load_img(
+ "dataset/single_prediction/image.png", target_size=(64, 64)
+ )
+ test_image = tf.keras.preprocessing.image.img_to_array(test_image)
+ test_image = np.expand_dims(test_image, axis=0)
+ result = classifier.predict(test_image)
+ # training_set.class_indices
+ if result[0][0] == 0:
+ prediction = "Normal"
+ if result[0][0] == 1:
+ prediction = "Abnormality detected"
diff --git a/computer_vision/flip_augmentation.py b/computer_vision/flip_augmentation.py
new file mode 100644
index 000000000000..7301424824df
--- /dev/null
+++ b/computer_vision/flip_augmentation.py
@@ -0,0 +1,128 @@
+import glob
+import os
+import random
+from string import ascii_lowercase, digits
+
+import cv2
+
+"""
+Flip image and bounding box for computer vision task
+https://paperswithcode.com/method/randomhorizontalflip
+"""
+
+# Params
+LABEL_DIR = ""
+IMAGE_DIR = ""
+OUTPUT_DIR = ""
+FLIP_TYPE = 1 # (0 is vertical, 1 is horizontal)
+
+
+def main() -> None:
+ """
+ Get images list and annotations list from input dir.
+ Update new images and annotations.
+ Save images and annotations in output dir.
+ """
+ img_paths, annos = get_dataset(LABEL_DIR, IMAGE_DIR)
+ print("Processing...")
+ new_images, new_annos, paths = update_image_and_anno(img_paths, annos, FLIP_TYPE)
+
+ for index, image in enumerate(new_images):
+ # Get random string code: '7b7ad245cdff75241935e4dd860f3bad'
+ letter_code = random_chars(32)
+ file_name = paths[index].split(os.sep)[-1].rsplit(".", 1)[0]
+ file_root = f"{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}"
+ cv2.imwrite(f"{file_root}.jpg", image, [cv2.IMWRITE_JPEG_QUALITY, 85])
+ print(f"Success {index + 1}/{len(new_images)} with {file_name}")
+ annos_list = []
+ for anno in new_annos[index]:
+ obj = f"{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}"
+ annos_list.append(obj)
+ with open(f"{file_root}.txt", "w") as outfile:
+ outfile.write("\n".join(line for line in annos_list))
+
+
+def get_dataset(label_dir: str, img_dir: str) -> tuple[list, list]:
+ """
+ - label_dir : Path to label include annotation of images
+ - img_dir : Path to folder contain images
+ Return : List of images path and labels
+ """
+ img_paths = []
+ labels = []
+ for label_file in glob.glob(os.path.join(label_dir, "*.txt")):
+ label_name = label_file.split(os.sep)[-1].rsplit(".", 1)[0]
+ with open(label_file) as in_file:
+ obj_lists = in_file.readlines()
+ img_path = os.path.join(img_dir, f"{label_name}.jpg")
+
+ boxes = []
+ for obj_list in obj_lists:
+ obj = obj_list.rstrip("\n").split(" ")
+ boxes.append(
+ [
+ int(obj[0]),
+ float(obj[1]),
+ float(obj[2]),
+ float(obj[3]),
+ float(obj[4]),
+ ]
+ )
+ if not boxes:
+ continue
+ img_paths.append(img_path)
+ labels.append(boxes)
+ return img_paths, labels
+
+
+def update_image_and_anno(
+ img_list: list, anno_list: list, flip_type: int = 1
+) -> tuple[list, list, list]:
+ """
+ - img_list : list of all images
+ - anno_list : list of all annotations of specific image
+ - flip_type : 0 is vertical, 1 is horizontal
+ Return:
+ - new_imgs_list : image after resize
+ - new_annos_lists : list of new annotation after scale
+ - path_list : list the name of image file
+ """
+ new_annos_lists = []
+ path_list = []
+ new_imgs_list = []
+ for idx in range(len(img_list)):
+ new_annos = []
+ path = img_list[idx]
+ path_list.append(path)
+ img_annos = anno_list[idx]
+ img = cv2.imread(path)
+ if flip_type == 1:
+ new_img = cv2.flip(img, flip_type)
+ for bbox in img_annos:
+ x_center_new = 1 - bbox[1]
+ new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]])
+ elif flip_type == 0:
+ new_img = cv2.flip(img, flip_type)
+ for bbox in img_annos:
+ y_center_new = 1 - bbox[2]
+ new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]])
+ new_annos_lists.append(new_annos)
+ new_imgs_list.append(new_img)
+ return new_imgs_list, new_annos_lists, path_list
+
+
+def random_chars(number_char: int = 32) -> str:
+ """
+ Automatic generate random 32 characters.
+ Get random string code: '7b7ad245cdff75241935e4dd860f3bad'
+ >>> len(random_chars(32))
+ 32
+ """
+ assert number_char > 1, "The number of character should greater than 1"
+ letter_code = ascii_lowercase + digits
+ return "".join(random.choice(letter_code) for _ in range(number_char))
+
+
+if __name__ == "__main__":
+ main()
+ print("DONE ✅")
diff --git a/computer_vision/haralick_descriptors.py b/computer_vision/haralick_descriptors.py
new file mode 100644
index 000000000000..54632160dcf2
--- /dev/null
+++ b/computer_vision/haralick_descriptors.py
@@ -0,0 +1,434 @@
+"""
+https://en.wikipedia.org/wiki/Image_texture
+https://en.wikipedia.org/wiki/Co-occurrence_matrix#Application_to_image_analysis
+"""
+
+import imageio.v2 as imageio
+import numpy as np
+
+
+def root_mean_square_error(original: np.ndarray, reference: np.ndarray) -> float:
+ """Simple implementation of Root Mean Squared Error
+ for two N dimensional numpy arrays.
+
+ Examples:
+ >>> root_mean_square_error(np.array([1, 2, 3]), np.array([1, 2, 3]))
+ 0.0
+ >>> root_mean_square_error(np.array([1, 2, 3]), np.array([2, 2, 2]))
+ 0.816496580927726
+ >>> root_mean_square_error(np.array([1, 2, 3]), np.array([6, 4, 2]))
+ 3.1622776601683795
+ """
+ return float(np.sqrt(((original - reference) ** 2).mean()))
+
+
+def normalize_image(
+ image: np.ndarray, cap: float = 255.0, data_type: np.dtype = np.uint8
+) -> np.ndarray:
+ """
+ Normalizes image in Numpy 2D array format, between ranges 0-cap,
+ as to fit uint8 type.
+
+ Args:
+ image: 2D numpy array representing image as matrix, with values in any range
+ cap: Maximum cap amount for normalization
+ data_type: numpy data type to set output variable to
+ Returns:
+ return 2D numpy array of type uint8, corresponding to limited range matrix
+
+ Examples:
+ >>> normalize_image(np.array([[1, 2, 3], [4, 5, 10]]),
+ ... cap=1.0, data_type=np.float64)
+ array([[0. , 0.11111111, 0.22222222],
+ [0.33333333, 0.44444444, 1. ]])
+ >>> normalize_image(np.array([[4, 4, 3], [1, 7, 2]]))
+ array([[127, 127, 85],
+ [ 0, 255, 42]], dtype=uint8)
+ """
+ normalized = (image - np.min(image)) / (np.max(image) - np.min(image)) * cap
+ return normalized.astype(data_type)
+
+
+def normalize_array(array: np.ndarray, cap: float = 1) -> np.ndarray:
+ """Normalizes a 1D array, between ranges 0-cap.
+
+ Args:
+ array: List containing values to be normalized between cap range.
+ cap: Maximum cap amount for normalization.
+ Returns:
+ return 1D numpy array, corresponding to limited range array
+
+ Examples:
+ >>> normalize_array(np.array([2, 3, 5, 7]))
+ array([0. , 0.2, 0.6, 1. ])
+ >>> normalize_array(np.array([[5], [7], [11], [13]]))
+ array([[0. ],
+ [0.25],
+ [0.75],
+ [1. ]])
+ """
+ diff = np.max(array) - np.min(array)
+ return (array - np.min(array)) / (1 if diff == 0 else diff) * cap
+
+
+def grayscale(image: np.ndarray) -> np.ndarray:
+ """
+ Uses luminance weights to transform RGB channel to greyscale, by
+ taking the dot product between the channel and the weights.
+
+ Example:
+ >>> grayscale(np.array([[[108, 201, 72], [255, 11, 127]],
+ ... [[56, 56, 56], [128, 255, 107]]]))
+ array([[158, 97],
+ [ 56, 200]], dtype=uint8)
+ """
+ return np.dot(image[:, :, 0:3], [0.299, 0.587, 0.114]).astype(np.uint8)
+
+
+def binarize(image: np.ndarray, threshold: float = 127.0) -> np.ndarray:
+ """
+ Binarizes a grayscale image based on a given threshold value,
+ setting values to 1 or 0 accordingly.
+
+ Examples:
+ >>> binarize(np.array([[128, 255], [101, 156]]))
+ array([[1, 1],
+ [0, 1]])
+ >>> binarize(np.array([[0.07, 1], [0.51, 0.3]]), threshold=0.5)
+ array([[0, 1],
+ [1, 0]])
+ """
+ return np.where(image > threshold, 1, 0)
+
+
+def transform(
+ image: np.ndarray, kind: str, kernel: np.ndarray | None = None
+) -> np.ndarray:
+ """
+ Simple image transformation using one of two available filter functions:
+ Erosion and Dilation.
+
+ Args:
+ image: binarized input image, onto which to apply transformation
+ kind: Can be either 'erosion', in which case the :func:np.max
+ function is called, or 'dilation', when :func:np.min is used instead.
+ kernel: n x n kernel with shape < :attr:image.shape,
+ to be used when applying convolution to original image
+
+ Returns:
+ returns a numpy array with same shape as input image,
+ corresponding to applied binary transformation.
+
+ Examples:
+ >>> img = np.array([[1, 0.5], [0.2, 0.7]])
+ >>> img = binarize(img, threshold=0.5)
+ >>> transform(img, 'erosion')
+ array([[1, 1],
+ [1, 1]], dtype=uint8)
+ >>> transform(img, 'dilation')
+ array([[0, 0],
+ [0, 0]], dtype=uint8)
+ """
+ if kernel is None:
+ kernel = np.ones((3, 3))
+
+ if kind == "erosion":
+ constant = 1
+ apply = np.max
+ else:
+ constant = 0
+ apply = np.min
+
+ center_x, center_y = (x // 2 for x in kernel.shape)
+
+ # Use padded image when applying convolution
+ # to not go out of bounds of the original the image
+ transformed = np.zeros(image.shape, dtype=np.uint8)
+ padded = np.pad(image, 1, "constant", constant_values=constant)
+
+ for x in range(center_x, padded.shape[0] - center_x):
+ for y in range(center_y, padded.shape[1] - center_y):
+ center = padded[
+ x - center_x : x + center_x + 1, y - center_y : y + center_y + 1
+ ]
+ # Apply transformation method to the centered section of the image
+ transformed[x - center_x, y - center_y] = apply(center[kernel == 1])
+
+ return transformed
+
+
+def opening_filter(image: np.ndarray, kernel: np.ndarray | None = None) -> np.ndarray:
+ """
+ Opening filter, defined as the sequence of
+ erosion and then a dilation filter on the same image.
+
+ Examples:
+ >>> img = np.array([[1, 0.5], [0.2, 0.7]])
+ >>> img = binarize(img, threshold=0.5)
+ >>> opening_filter(img)
+ array([[1, 1],
+ [1, 1]], dtype=uint8)
+ """
+ if kernel is None:
+ np.ones((3, 3))
+
+ return transform(transform(image, "dilation", kernel), "erosion", kernel)
+
+
+def closing_filter(image: np.ndarray, kernel: np.ndarray | None = None) -> np.ndarray:
+ """
+ Opening filter, defined as the sequence of
+ dilation and then erosion filter on the same image.
+
+ Examples:
+ >>> img = np.array([[1, 0.5], [0.2, 0.7]])
+ >>> img = binarize(img, threshold=0.5)
+ >>> closing_filter(img)
+ array([[0, 0],
+ [0, 0]], dtype=uint8)
+ """
+ if kernel is None:
+ kernel = np.ones((3, 3))
+ return transform(transform(image, "erosion", kernel), "dilation", kernel)
+
+
+def binary_mask(
+ image_gray: np.ndarray, image_map: np.ndarray
+) -> tuple[np.ndarray, np.ndarray]:
+ """
+ Apply binary mask, or thresholding based
+ on bit mask value (mapping mask is binary).
+
+ Returns the mapped true value mask and its complementary false value mask.
+
+ Example:
+ >>> img = np.array([[[108, 201, 72], [255, 11, 127]],
+ ... [[56, 56, 56], [128, 255, 107]]])
+ >>> gray = grayscale(img)
+ >>> binary = binarize(gray)
+ >>> morphological = opening_filter(binary)
+ >>> binary_mask(gray, morphological)
+ (array([[1, 1],
+ [1, 1]], dtype=uint8), array([[158, 97],
+ [ 56, 200]], dtype=uint8))
+ """
+ true_mask, false_mask = image_gray.copy(), image_gray.copy()
+ true_mask[image_map == 1] = 1
+ false_mask[image_map == 0] = 0
+
+ return true_mask, false_mask
+
+
+def matrix_concurrency(image: np.ndarray, coordinate: tuple[int, int]) -> np.ndarray:
+ """
+ Calculate sample co-occurrence matrix based on input image
+ as well as selected coordinates on image.
+
+ Implementation is made using basic iteration,
+ as function to be performed (np.max) is non-linear and therefore
+ not callable on the frequency domain.
+
+ Example:
+ >>> img = np.array([[[108, 201, 72], [255, 11, 127]],
+ ... [[56, 56, 56], [128, 255, 107]]])
+ >>> gray = grayscale(img)
+ >>> binary = binarize(gray)
+ >>> morphological = opening_filter(binary)
+ >>> mask_1 = binary_mask(gray, morphological)[0]
+ >>> matrix_concurrency(mask_1, (0, 1))
+ array([[0., 0.],
+ [0., 0.]])
+ """
+ matrix = np.zeros([np.max(image) + 1, np.max(image) + 1])
+
+ offset_x, offset_y = coordinate
+
+ for x in range(1, image.shape[0] - 1):
+ for y in range(1, image.shape[1] - 1):
+ base_pixel = image[x, y]
+ offset_pixel = image[x + offset_x, y + offset_y]
+
+ matrix[base_pixel, offset_pixel] += 1
+ matrix_sum = np.sum(matrix)
+ return matrix / (1 if matrix_sum == 0 else matrix_sum)
+
+
+def haralick_descriptors(matrix: np.ndarray) -> list[float]:
+ """Calculates all 8 Haralick descriptors based on co-occurrence input matrix.
+ All descriptors are as follows:
+ Maximum probability, Inverse Difference, Homogeneity, Entropy,
+ Energy, Dissimilarity, Contrast and Correlation
+
+ Args:
+ matrix: Co-occurrence matrix to use as base for calculating descriptors.
+
+ Returns:
+ Reverse ordered list of resulting descriptors
+
+ Example:
+ >>> img = np.array([[[108, 201, 72], [255, 11, 127]],
+ ... [[56, 56, 56], [128, 255, 107]]])
+ >>> gray = grayscale(img)
+ >>> binary = binarize(gray)
+ >>> morphological = opening_filter(binary)
+ >>> mask_1 = binary_mask(gray, morphological)[0]
+ >>> concurrency = matrix_concurrency(mask_1, (0, 1))
+ >>> [float(f) for f in haralick_descriptors(concurrency)]
+ [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ """
+ # Function np.indices could be used for bigger input types,
+ # but np.ogrid works just fine
+ i, j = np.ogrid[0 : matrix.shape[0], 0 : matrix.shape[1]] # np.indices()
+
+ # Pre-calculate frequent multiplication and subtraction
+ prod = np.multiply(i, j)
+ sub = np.subtract(i, j)
+
+ # Calculate numerical value of Maximum Probability
+ maximum_prob = np.max(matrix)
+ # Using the definition for each descriptor individually to calculate its matrix
+ correlation = prod * matrix
+ energy = np.power(matrix, 2)
+ contrast = matrix * np.power(sub, 2)
+
+ dissimilarity = matrix * np.abs(sub)
+ inverse_difference = matrix / (1 + np.abs(sub))
+ homogeneity = matrix / (1 + np.power(sub, 2))
+ entropy = -(matrix[matrix > 0] * np.log(matrix[matrix > 0]))
+
+ # Sum values for descriptors ranging from the first one to the last,
+ # as all are their respective origin matrix and not the resulting value yet.
+ return [
+ maximum_prob,
+ correlation.sum(),
+ energy.sum(),
+ contrast.sum(),
+ dissimilarity.sum(),
+ inverse_difference.sum(),
+ homogeneity.sum(),
+ entropy.sum(),
+ ]
+
+
+def get_descriptors(
+ masks: tuple[np.ndarray, np.ndarray], coordinate: tuple[int, int]
+) -> np.ndarray:
+ """
+ Calculate all Haralick descriptors for a sequence of
+ different co-occurrence matrices, given input masks and coordinates.
+
+ Example:
+ >>> img = np.array([[[108, 201, 72], [255, 11, 127]],
+ ... [[56, 56, 56], [128, 255, 107]]])
+ >>> gray = grayscale(img)
+ >>> binary = binarize(gray)
+ >>> morphological = opening_filter(binary)
+ >>> get_descriptors(binary_mask(gray, morphological), (0, 1))
+ array([0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0.])
+ """
+ descriptors = np.array(
+ [haralick_descriptors(matrix_concurrency(mask, coordinate)) for mask in masks]
+ )
+
+ # Concatenate each individual descriptor into
+ # one single list containing sequence of descriptors
+ return np.concatenate(descriptors, axis=None)
+
+
+def euclidean(point_1: np.ndarray, point_2: np.ndarray) -> float:
+ """
+ Simple method for calculating the euclidean distance between two points,
+ with type np.ndarray.
+
+ Example:
+ >>> a = np.array([1, 0, -2])
+ >>> b = np.array([2, -1, 1])
+ >>> euclidean(a, b)
+ 3.3166247903554
+ """
+ return float(np.sqrt(np.sum(np.square(point_1 - point_2))))
+
+
+def get_distances(descriptors: np.ndarray, base: int) -> list[tuple[int, float]]:
+ """
+ Calculate all Euclidean distances between a selected base descriptor
+ and all other Haralick descriptors
+ The resulting comparison is return in decreasing order,
+ showing which descriptor is the most similar to the selected base.
+
+ Args:
+ descriptors: Haralick descriptors to compare with base index
+ base: Haralick descriptor index to use as base when calculating respective
+ euclidean distance to other descriptors.
+
+ Returns:
+ Ordered distances between descriptors
+
+ Example:
+ >>> index = 1
+ >>> img = np.array([[[108, 201, 72], [255, 11, 127]],
+ ... [[56, 56, 56], [128, 255, 107]]])
+ >>> gray = grayscale(img)
+ >>> binary = binarize(gray)
+ >>> morphological = opening_filter(binary)
+ >>> get_distances(get_descriptors(
+ ... binary_mask(gray, morphological), (0, 1)),
+ ... index)
+ [(0, 0.0), (1, 0.0), (2, 0.0), (3, 0.0), (4, 0.0), (5, 0.0), \
+(6, 0.0), (7, 0.0), (8, 0.0), (9, 0.0), (10, 0.0), (11, 0.0), (12, 0.0), \
+(13, 0.0), (14, 0.0), (15, 0.0)]
+ """
+ distances = np.array(
+ [euclidean(descriptor, descriptors[base]) for descriptor in descriptors]
+ )
+ # Normalize distances between range [0, 1]
+ normalized_distances: list[float] = normalize_array(distances, 1).tolist()
+ enum_distances = list(enumerate(normalized_distances))
+ enum_distances.sort(key=lambda tup: tup[1], reverse=True)
+ return enum_distances
+
+
+if __name__ == "__main__":
+ # Index to compare haralick descriptors to
+ index = int(input())
+ q_value_list = [int(value) for value in input().split()]
+ q_value = (q_value_list[0], q_value_list[1])
+
+ # Format is the respective filter to apply,
+ # can be either 1 for the opening filter or else for the closing
+ parameters = {"format": int(input()), "threshold": int(input())}
+
+ # Number of images to perform methods on
+ b_number = int(input())
+
+ files, descriptors = [], []
+
+ for _ in range(b_number):
+ file = input().rstrip()
+ files.append(file)
+
+ # Open given image and calculate morphological filter,
+ # respective masks and correspondent Harralick Descriptors.
+ image = imageio.imread(file).astype(np.float32)
+ gray = grayscale(image)
+ threshold = binarize(gray, parameters["threshold"])
+
+ morphological = (
+ opening_filter(threshold)
+ if parameters["format"] == 1
+ else closing_filter(threshold)
+ )
+ masks = binary_mask(gray, morphological)
+ descriptors.append(get_descriptors(masks, q_value))
+
+ # Transform ordered distances array into a sequence of indexes
+ # corresponding to original file position
+ distances = get_distances(np.array(descriptors), index)
+ indexed_distances = np.array(distances).astype(np.uint8)[:, 0]
+
+ # Finally, print distances considering the Haralick descriptions from the base
+ # file to all other images using the morphology method of choice.
+ print(f"Query: {files[index]}")
+ print("Ranking:")
+ for idx, file_idx in enumerate(indexed_distances):
+ print(f"({idx}) {files[file_idx]}", end="\n")
diff --git a/computer_vision/harriscorner.py b/computer_vision/harris_corner.py
similarity index 83%
rename from computer_vision/harriscorner.py
rename to computer_vision/harris_corner.py
index fb7f560f7873..0cc7522bc3af 100644
--- a/computer_vision/harriscorner.py
+++ b/computer_vision/harris_corner.py
@@ -7,9 +7,8 @@
"""
-class Harris_Corner:
+class HarrisCorner:
def __init__(self, k: float, window_size: int):
-
"""
k : is an empirically determined constant in [0.04,0.06]
window_size : neighbourhoods considered
@@ -21,12 +20,10 @@ def __init__(self, k: float, window_size: int):
else:
raise ValueError("invalid k value")
- def __str__(self):
-
- return f"Harris Corner detection with k : {self.k}"
-
- def detect(self, img_path: str):
+ def __str__(self) -> str:
+ return str(self.k)
+ def detect(self, img_path: str) -> tuple[cv2.Mat, list[list[int]]]:
"""
Returns the image with corners identified
img_path : path of the image
@@ -35,12 +32,12 @@ def detect(self, img_path: str):
img = cv2.imread(img_path, 0)
h, w = img.shape
- corner_list = []
+ corner_list: list[list[int]] = []
color_img = img.copy()
color_img = cv2.cvtColor(color_img, cv2.COLOR_GRAY2RGB)
dy, dx = np.gradient(img)
- ixx = dx ** 2
- iyy = dy ** 2
+ ixx = dx**2
+ iyy = dy**2
ixy = dx * dy
k = 0.04
offset = self.window_size // 2
@@ -56,9 +53,9 @@ def detect(self, img_path: str):
y - offset : y + offset + 1, x - offset : x + offset + 1
].sum()
- det = (wxx * wyy) - (wxy ** 2)
+ det = (wxx * wyy) - (wxy**2)
trace = wxx + wyy
- r = det - k * (trace ** 2)
+ r = det - k * (trace**2)
# Can change the value
if r > 0.5:
corner_list.append([x, y, r])
@@ -69,7 +66,6 @@ def detect(self, img_path: str):
if __name__ == "__main__":
-
- edge_detect = Harris_Corner(0.04, 3)
+ edge_detect = HarrisCorner(0.04, 3)
color_img, _ = edge_detect.detect("path_to_image")
cv2.imwrite("detect.png", color_img)
diff --git a/computer_vision/horn_schunck.py b/computer_vision/horn_schunck.py
new file mode 100644
index 000000000000..f33b5b1c794b
--- /dev/null
+++ b/computer_vision/horn_schunck.py
@@ -0,0 +1,131 @@
+"""
+The Horn-Schunck method estimates the optical flow for every single pixel of
+a sequence of images.
+It works by assuming brightness constancy between two consecutive frames
+and smoothness in the optical flow.
+
+Useful resources:
+Wikipedia: https://en.wikipedia.org/wiki/Horn%E2%80%93Schunck_method
+Paper: http://image.diku.dk/imagecanon/material/HornSchunckOptical_Flow.pdf
+"""
+
+from typing import SupportsIndex
+
+import numpy as np
+from scipy.ndimage import convolve
+
+
+def warp(
+ image: np.ndarray, horizontal_flow: np.ndarray, vertical_flow: np.ndarray
+) -> np.ndarray:
+ """
+ Warps the pixels of an image into a new image using the horizontal and vertical
+ flows.
+ Pixels that are warped from an invalid location are set to 0.
+
+ Parameters:
+ image: Grayscale image
+ horizontal_flow: Horizontal flow
+ vertical_flow: Vertical flow
+
+ Returns: Warped image
+
+ >>> warp(np.array([[0, 1, 2], [0, 3, 0], [2, 2, 2]]), \
+ np.array([[0, 1, -1], [-1, 0, 0], [1, 1, 1]]), \
+ np.array([[0, 0, 0], [0, 1, 0], [0, 0, 1]]))
+ array([[0, 0, 0],
+ [3, 1, 0],
+ [0, 2, 3]])
+ """
+ flow = np.stack((horizontal_flow, vertical_flow), 2)
+
+ # Create a grid of all pixel coordinates and subtract the flow to get the
+ # target pixels coordinates
+ grid = np.stack(
+ np.meshgrid(np.arange(0, image.shape[1]), np.arange(0, image.shape[0])), 2
+ )
+ grid = np.round(grid - flow).astype(np.int32)
+
+ # Find the locations outside of the original image
+ invalid = (grid < 0) | (grid >= np.array([image.shape[1], image.shape[0]]))
+ grid[invalid] = 0
+
+ warped = image[grid[:, :, 1], grid[:, :, 0]]
+
+ # Set pixels at invalid locations to 0
+ warped[invalid[:, :, 0] | invalid[:, :, 1]] = 0
+
+ return warped
+
+
+def horn_schunck(
+ image0: np.ndarray,
+ image1: np.ndarray,
+ num_iter: SupportsIndex,
+ alpha: float | None = None,
+) -> tuple[np.ndarray, np.ndarray]:
+ """
+ This function performs the Horn-Schunck algorithm and returns the estimated
+ optical flow. It is assumed that the input images are grayscale and
+ normalized to be in [0, 1].
+
+ Parameters:
+ image0: First image of the sequence
+ image1: Second image of the sequence
+ alpha: Regularization constant
+ num_iter: Number of iterations performed
+
+ Returns: estimated horizontal & vertical flow
+
+ >>> np.round(horn_schunck(np.array([[0, 0, 2], [0, 0, 2]]), \
+ np.array([[0, 2, 0], [0, 2, 0]]), alpha=0.1, num_iter=110)).\
+ astype(np.int32)
+ array([[[ 0, -1, -1],
+ [ 0, -1, -1]],
+
+ [[ 0, 0, 0],
+ [ 0, 0, 0]]], dtype=int32)
+ """
+ if alpha is None:
+ alpha = 0.1
+
+ # Initialize flow
+ horizontal_flow = np.zeros_like(image0)
+ vertical_flow = np.zeros_like(image0)
+
+ # Prepare kernels for the calculation of the derivatives and the average velocity
+ kernel_x = np.array([[-1, 1], [-1, 1]]) * 0.25
+ kernel_y = np.array([[-1, -1], [1, 1]]) * 0.25
+ kernel_t = np.array([[1, 1], [1, 1]]) * 0.25
+ kernel_laplacian = np.array(
+ [[1 / 12, 1 / 6, 1 / 12], [1 / 6, 0, 1 / 6], [1 / 12, 1 / 6, 1 / 12]]
+ )
+
+ # Iteratively refine the flow
+ for _ in range(num_iter):
+ warped_image = warp(image0, horizontal_flow, vertical_flow)
+ derivative_x = convolve(warped_image, kernel_x) + convolve(image1, kernel_x)
+ derivative_y = convolve(warped_image, kernel_y) + convolve(image1, kernel_y)
+ derivative_t = convolve(warped_image, kernel_t) + convolve(image1, -kernel_t)
+
+ avg_horizontal_velocity = convolve(horizontal_flow, kernel_laplacian)
+ avg_vertical_velocity = convolve(vertical_flow, kernel_laplacian)
+
+ # This updates the flow as proposed in the paper (Step 12)
+ update = (
+ derivative_x * avg_horizontal_velocity
+ + derivative_y * avg_vertical_velocity
+ + derivative_t
+ )
+ update = update / (alpha**2 + derivative_x**2 + derivative_y**2)
+
+ horizontal_flow = avg_horizontal_velocity - derivative_x * update
+ vertical_flow = avg_vertical_velocity - derivative_y * update
+
+ return horizontal_flow, vertical_flow
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/computer_vision/intensity_based_segmentation.py b/computer_vision/intensity_based_segmentation.py
new file mode 100644
index 000000000000..7f2b1141acc4
--- /dev/null
+++ b/computer_vision/intensity_based_segmentation.py
@@ -0,0 +1,62 @@
+# Source: "https://www.ijcse.com/docs/IJCSE11-02-03-117.pdf"
+
+# Importing necessary libraries
+import matplotlib.pyplot as plt
+import numpy as np
+from PIL import Image
+
+
+def segment_image(image: np.ndarray, thresholds: list[int]) -> np.ndarray:
+ """
+ Performs image segmentation based on intensity thresholds.
+
+ Args:
+ image: Input grayscale image as a 2D array.
+ thresholds: Intensity thresholds to define segments.
+
+ Returns:
+ A labeled 2D array where each region corresponds to a threshold range.
+
+ Example:
+ >>> img = np.array([[80, 120, 180], [40, 90, 150], [20, 60, 100]])
+ >>> segment_image(img, [50, 100, 150])
+ array([[1, 2, 3],
+ [0, 1, 2],
+ [0, 1, 1]], dtype=int32)
+ """
+ # Initialize segmented array with zeros
+ segmented = np.zeros_like(image, dtype=np.int32)
+
+ # Assign labels based on thresholds
+ for i, threshold in enumerate(thresholds):
+ segmented[image > threshold] = i + 1
+
+ return segmented
+
+
+if __name__ == "__main__":
+ # Load the image
+ image_path = "path_to_image" # Replace with your image path
+ original_image = Image.open(image_path).convert("L")
+ image_array = np.array(original_image)
+
+ # Define thresholds
+ thresholds = [50, 100, 150, 200]
+
+ # Perform segmentation
+ segmented_image = segment_image(image_array, thresholds)
+
+ # Display the results
+ plt.figure(figsize=(10, 5))
+
+ plt.subplot(1, 2, 1)
+ plt.title("Original Image")
+ plt.imshow(image_array, cmap="gray")
+ plt.axis("off")
+
+ plt.subplot(1, 2, 2)
+ plt.title("Segmented Image")
+ plt.imshow(segmented_image, cmap="tab20")
+ plt.axis("off")
+
+ plt.show()
diff --git a/computer_vision/meanthreshold.py b/computer_vision/mean_threshold.py
similarity index 100%
rename from computer_vision/meanthreshold.py
rename to computer_vision/mean_threshold.py
diff --git a/computer_vision/mosaic_augmentation.py b/computer_vision/mosaic_augmentation.py
new file mode 100644
index 000000000000..d881347121ea
--- /dev/null
+++ b/computer_vision/mosaic_augmentation.py
@@ -0,0 +1,186 @@
+"""Source: https://github.com/jason9075/opencv-mosaic-data-aug"""
+
+import glob
+import os
+import random
+from string import ascii_lowercase, digits
+
+import cv2
+import numpy as np
+
+# Parameters
+OUTPUT_SIZE = (720, 1280) # Height, Width
+SCALE_RANGE = (0.4, 0.6) # if height or width lower than this scale, drop it.
+FILTER_TINY_SCALE = 1 / 100
+LABEL_DIR = ""
+IMG_DIR = ""
+OUTPUT_DIR = ""
+NUMBER_IMAGES = 250
+
+
+def main() -> None:
+ """
+ Get images list and annotations list from input dir.
+ Update new images and annotations.
+ Save images and annotations in output dir.
+ """
+ img_paths, annos = get_dataset(LABEL_DIR, IMG_DIR)
+ for index in range(NUMBER_IMAGES):
+ idxs = random.sample(range(len(annos)), 4)
+ new_image, new_annos, path = update_image_and_anno(
+ img_paths,
+ annos,
+ idxs,
+ OUTPUT_SIZE,
+ SCALE_RANGE,
+ filter_scale=FILTER_TINY_SCALE,
+ )
+
+ # Get random string code: '7b7ad245cdff75241935e4dd860f3bad'
+ letter_code = random_chars(32)
+ file_name = path.split(os.sep)[-1].rsplit(".", 1)[0]
+ file_root = f"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}"
+ cv2.imwrite(f"{file_root}.jpg", new_image, [cv2.IMWRITE_JPEG_QUALITY, 85])
+ print(f"Succeeded {index + 1}/{NUMBER_IMAGES} with {file_name}")
+ annos_list = []
+ for anno in new_annos:
+ width = anno[3] - anno[1]
+ height = anno[4] - anno[2]
+ x_center = anno[1] + width / 2
+ y_center = anno[2] + height / 2
+ obj = f"{anno[0]} {x_center} {y_center} {width} {height}"
+ annos_list.append(obj)
+ with open(f"{file_root}.txt", "w") as outfile:
+ outfile.write("\n".join(line for line in annos_list))
+
+
+def get_dataset(label_dir: str, img_dir: str) -> tuple[list, list]:
+ """
+ - label_dir : Path to label include annotation of images
+ - img_dir : Path to folder contain images
+ Return : List of images path and labels
+ """
+ img_paths = []
+ labels = []
+ for label_file in glob.glob(os.path.join(label_dir, "*.txt")):
+ label_name = label_file.split(os.sep)[-1].rsplit(".", 1)[0]
+ with open(label_file) as in_file:
+ obj_lists = in_file.readlines()
+ img_path = os.path.join(img_dir, f"{label_name}.jpg")
+
+ boxes = []
+ for obj_list in obj_lists:
+ obj = obj_list.rstrip("\n").split(" ")
+ xmin = float(obj[1]) - float(obj[3]) / 2
+ ymin = float(obj[2]) - float(obj[4]) / 2
+ xmax = float(obj[1]) + float(obj[3]) / 2
+ ymax = float(obj[2]) + float(obj[4]) / 2
+
+ boxes.append([int(obj[0]), xmin, ymin, xmax, ymax])
+ if not boxes:
+ continue
+ img_paths.append(img_path)
+ labels.append(boxes)
+ return img_paths, labels
+
+
+def update_image_and_anno(
+ all_img_list: list,
+ all_annos: list,
+ idxs: list[int],
+ output_size: tuple[int, int],
+ scale_range: tuple[float, float],
+ filter_scale: float = 0.0,
+) -> tuple[list, list, str]:
+ """
+ - all_img_list : list of all images
+ - all_annos : list of all annotations of specific image
+ - idxs : index of image in list
+ - output_size : size of output image (Height, Width)
+ - scale_range : range of scale image
+ - filter_scale : the condition of downscale image and bounding box
+ Return:
+ - output_img : image after resize
+ - new_anno : list of new annotation after scale
+ - path[0] : get the name of image file
+ """
+ output_img = np.zeros([output_size[0], output_size[1], 3], dtype=np.uint8)
+ scale_x = scale_range[0] + random.random() * (scale_range[1] - scale_range[0])
+ scale_y = scale_range[0] + random.random() * (scale_range[1] - scale_range[0])
+ divid_point_x = int(scale_x * output_size[1])
+ divid_point_y = int(scale_y * output_size[0])
+
+ new_anno = []
+ path_list = []
+ for i, index in enumerate(idxs):
+ path = all_img_list[index]
+ path_list.append(path)
+ img_annos = all_annos[index]
+ img = cv2.imread(path)
+ if i == 0: # top-left
+ img = cv2.resize(img, (divid_point_x, divid_point_y))
+ output_img[:divid_point_y, :divid_point_x, :] = img
+ for bbox in img_annos:
+ xmin = bbox[1] * scale_x
+ ymin = bbox[2] * scale_y
+ xmax = bbox[3] * scale_x
+ ymax = bbox[4] * scale_y
+ new_anno.append([bbox[0], xmin, ymin, xmax, ymax])
+ elif i == 1: # top-right
+ img = cv2.resize(img, (output_size[1] - divid_point_x, divid_point_y))
+ output_img[:divid_point_y, divid_point_x : output_size[1], :] = img
+ for bbox in img_annos:
+ xmin = scale_x + bbox[1] * (1 - scale_x)
+ ymin = bbox[2] * scale_y
+ xmax = scale_x + bbox[3] * (1 - scale_x)
+ ymax = bbox[4] * scale_y
+ new_anno.append([bbox[0], xmin, ymin, xmax, ymax])
+ elif i == 2: # bottom-left
+ img = cv2.resize(img, (divid_point_x, output_size[0] - divid_point_y))
+ output_img[divid_point_y : output_size[0], :divid_point_x, :] = img
+ for bbox in img_annos:
+ xmin = bbox[1] * scale_x
+ ymin = scale_y + bbox[2] * (1 - scale_y)
+ xmax = bbox[3] * scale_x
+ ymax = scale_y + bbox[4] * (1 - scale_y)
+ new_anno.append([bbox[0], xmin, ymin, xmax, ymax])
+ else: # bottom-right
+ img = cv2.resize(
+ img, (output_size[1] - divid_point_x, output_size[0] - divid_point_y)
+ )
+ output_img[
+ divid_point_y : output_size[0], divid_point_x : output_size[1], :
+ ] = img
+ for bbox in img_annos:
+ xmin = scale_x + bbox[1] * (1 - scale_x)
+ ymin = scale_y + bbox[2] * (1 - scale_y)
+ xmax = scale_x + bbox[3] * (1 - scale_x)
+ ymax = scale_y + bbox[4] * (1 - scale_y)
+ new_anno.append([bbox[0], xmin, ymin, xmax, ymax])
+
+ # Remove bounding box small than scale of filter
+ if filter_scale > 0:
+ new_anno = [
+ anno
+ for anno in new_anno
+ if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2])
+ ]
+
+ return output_img, new_anno, path_list[0]
+
+
+def random_chars(number_char: int) -> str:
+ """
+ Automatic generate random 32 characters.
+ Get random string code: '7b7ad245cdff75241935e4dd860f3bad'
+ >>> len(random_chars(32))
+ 32
+ """
+ assert number_char > 1, "The number of character should greater than 1"
+ letter_code = ascii_lowercase + digits
+ return "".join(random.choice(letter_code) for _ in range(number_char))
+
+
+if __name__ == "__main__":
+ main()
+ print("DONE ✅")
diff --git a/computer_vision/pooling_functions.py b/computer_vision/pooling_functions.py
new file mode 100644
index 000000000000..09beabcba82d
--- /dev/null
+++ b/computer_vision/pooling_functions.py
@@ -0,0 +1,135 @@
+# Source : https://computersciencewiki.org/index.php/Max-pooling_/_Pooling
+# Importing the libraries
+import numpy as np
+from PIL import Image
+
+
+# Maxpooling Function
+def maxpooling(arr: np.ndarray, size: int, stride: int) -> np.ndarray:
+ """
+ This function is used to perform maxpooling on the input array of 2D matrix(image)
+ Args:
+ arr: numpy array
+ size: size of pooling matrix
+ stride: the number of pixels shifts over the input matrix
+ Returns:
+ numpy array of maxpooled matrix
+ Sample Input Output:
+ >>> maxpooling([[1,2,3,4],[5,6,7,8],[9,10,11,12],[13,14,15,16]], 2, 2)
+ array([[ 6., 8.],
+ [14., 16.]])
+ >>> maxpooling([[147, 180, 122],[241, 76, 32],[126, 13, 157]], 2, 1)
+ array([[241., 180.],
+ [241., 157.]])
+ """
+ arr = np.array(arr)
+ if arr.shape[0] != arr.shape[1]:
+ raise ValueError("The input array is not a square matrix")
+ i = 0
+ j = 0
+ mat_i = 0
+ mat_j = 0
+
+ # compute the shape of the output matrix
+ maxpool_shape = (arr.shape[0] - size) // stride + 1
+ # initialize the output matrix with zeros of shape maxpool_shape
+ updated_arr = np.zeros((maxpool_shape, maxpool_shape))
+
+ while i < arr.shape[0]:
+ if i + size > arr.shape[0]:
+ # if the end of the matrix is reached, break
+ break
+ while j < arr.shape[1]:
+ # if the end of the matrix is reached, break
+ if j + size > arr.shape[1]:
+ break
+ # compute the maximum of the pooling matrix
+ updated_arr[mat_i][mat_j] = np.max(arr[i : i + size, j : j + size])
+ # shift the pooling matrix by stride of column pixels
+ j += stride
+ mat_j += 1
+
+ # shift the pooling matrix by stride of row pixels
+ i += stride
+ mat_i += 1
+
+ # reset the column index to 0
+ j = 0
+ mat_j = 0
+
+ return updated_arr
+
+
+# Averagepooling Function
+def avgpooling(arr: np.ndarray, size: int, stride: int) -> np.ndarray:
+ """
+ This function is used to perform avgpooling on the input array of 2D matrix(image)
+ Args:
+ arr: numpy array
+ size: size of pooling matrix
+ stride: the number of pixels shifts over the input matrix
+ Returns:
+ numpy array of avgpooled matrix
+ Sample Input Output:
+ >>> avgpooling([[1,2,3,4],[5,6,7,8],[9,10,11,12],[13,14,15,16]], 2, 2)
+ array([[ 3., 5.],
+ [11., 13.]])
+ >>> avgpooling([[147, 180, 122],[241, 76, 32],[126, 13, 157]], 2, 1)
+ array([[161., 102.],
+ [114., 69.]])
+ """
+ arr = np.array(arr)
+ if arr.shape[0] != arr.shape[1]:
+ raise ValueError("The input array is not a square matrix")
+ i = 0
+ j = 0
+ mat_i = 0
+ mat_j = 0
+
+ # compute the shape of the output matrix
+ avgpool_shape = (arr.shape[0] - size) // stride + 1
+ # initialize the output matrix with zeros of shape avgpool_shape
+ updated_arr = np.zeros((avgpool_shape, avgpool_shape))
+
+ while i < arr.shape[0]:
+ # if the end of the matrix is reached, break
+ if i + size > arr.shape[0]:
+ break
+ while j < arr.shape[1]:
+ # if the end of the matrix is reached, break
+ if j + size > arr.shape[1]:
+ break
+ # compute the average of the pooling matrix
+ updated_arr[mat_i][mat_j] = int(np.average(arr[i : i + size, j : j + size]))
+ # shift the pooling matrix by stride of column pixels
+ j += stride
+ mat_j += 1
+
+ # shift the pooling matrix by stride of row pixels
+ i += stride
+ mat_i += 1
+ # reset the column index to 0
+ j = 0
+ mat_j = 0
+
+ return updated_arr
+
+
+# Main Function
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod(name="avgpooling", verbose=True)
+
+ # Loading the image
+ image = Image.open("path_to_image")
+
+ # Converting the image to numpy array and maxpooling, displaying the result
+ # Ensure that the image is a square matrix
+
+ Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show()
+
+ # Converting the image to numpy array and averagepooling, displaying the result
+ # Ensure that the image is a square matrix
+
+ Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
diff --git a/conversions/README.md b/conversions/README.md
new file mode 100644
index 000000000000..ec3d931fd828
--- /dev/null
+++ b/conversions/README.md
@@ -0,0 +1,6 @@
+# Conversion
+
+Conversion programs convert a type of data, a number from a numerical base or unit into one of another type, base or unit, e.g. binary to decimal, integer to string or foot to meters.
+
+*
+*
diff --git a/conversions/astronomical_length_scale_conversion.py b/conversions/astronomical_length_scale_conversion.py
new file mode 100644
index 000000000000..0f413644906d
--- /dev/null
+++ b/conversions/astronomical_length_scale_conversion.py
@@ -0,0 +1,106 @@
+"""
+Conversion of length units.
+Available Units:
+Metre, Kilometre, Megametre, Gigametre,
+Terametre, Petametre, Exametre, Zettametre, Yottametre
+
+USAGE :
+-> Import this file into their respective project.
+-> Use the function length_conversion() for conversion of length units.
+-> Parameters :
+ -> value : The number of from units you want to convert
+ -> from_type : From which type you want to convert
+ -> to_type : To which type you want to convert
+
+REFERENCES :
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Meter
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Kilometer
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Orders_of_magnitude_(length)
+"""
+
+UNIT_SYMBOL = {
+ "meter": "m",
+ "kilometer": "km",
+ "megametre": "Mm",
+ "gigametre": "Gm",
+ "terametre": "Tm",
+ "petametre": "Pm",
+ "exametre": "Em",
+ "zettametre": "Zm",
+ "yottametre": "Ym",
+}
+# Exponent of the factor(meter)
+METRIC_CONVERSION = {
+ "m": 0,
+ "km": 3,
+ "Mm": 6,
+ "Gm": 9,
+ "Tm": 12,
+ "Pm": 15,
+ "Em": 18,
+ "Zm": 21,
+ "Ym": 24,
+}
+
+
+def length_conversion(value: float, from_type: str, to_type: str) -> float:
+ """
+ Conversion between astronomical length units.
+
+ >>> length_conversion(1, "meter", "kilometer")
+ 0.001
+ >>> length_conversion(1, "meter", "megametre")
+ 1e-06
+ >>> length_conversion(1, "gigametre", "meter")
+ 1000000000
+ >>> length_conversion(1, "gigametre", "terametre")
+ 0.001
+ >>> length_conversion(1, "petametre", "terametre")
+ 1000
+ >>> length_conversion(1, "petametre", "exametre")
+ 0.001
+ >>> length_conversion(1, "terametre", "zettametre")
+ 1e-09
+ >>> length_conversion(1, "yottametre", "zettametre")
+ 1000
+ >>> length_conversion(4, "wrongUnit", "inch")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid 'from_type' value: 'wrongUnit'.
+ Conversion abbreviations are: m, km, Mm, Gm, Tm, Pm, Em, Zm, Ym
+ """
+
+ from_sanitized = from_type.lower().strip("s")
+ to_sanitized = to_type.lower().strip("s")
+
+ from_sanitized = UNIT_SYMBOL.get(from_sanitized, from_sanitized)
+ to_sanitized = UNIT_SYMBOL.get(to_sanitized, to_sanitized)
+
+ if from_sanitized not in METRIC_CONVERSION:
+ msg = (
+ f"Invalid 'from_type' value: {from_type!r}.\n"
+ f"Conversion abbreviations are: {', '.join(METRIC_CONVERSION)}"
+ )
+ raise ValueError(msg)
+ if to_sanitized not in METRIC_CONVERSION:
+ msg = (
+ f"Invalid 'to_type' value: {to_type!r}.\n"
+ f"Conversion abbreviations are: {', '.join(METRIC_CONVERSION)}"
+ )
+ raise ValueError(msg)
+ from_exponent = METRIC_CONVERSION[from_sanitized]
+ to_exponent = METRIC_CONVERSION[to_sanitized]
+ exponent = 1
+
+ if from_exponent > to_exponent:
+ exponent = from_exponent - to_exponent
+ else:
+ exponent = -(to_exponent - from_exponent)
+
+ return value * pow(10, exponent)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/conversions/binary_to_decimal.py b/conversions/binary_to_decimal.py
index a7625e475bdc..914a9318c225 100644
--- a/conversions/binary_to_decimal.py
+++ b/conversions/binary_to_decimal.py
@@ -12,15 +12,15 @@ def bin_to_decimal(bin_string: str) -> int:
0
>>> bin_to_decimal("a")
Traceback (most recent call last):
- ...
+ ...
ValueError: Non-binary value was passed to the function
>>> bin_to_decimal("")
Traceback (most recent call last):
- ...
+ ...
ValueError: Empty string was passed to the function
>>> bin_to_decimal("39")
Traceback (most recent call last):
- ...
+ ...
ValueError: Non-binary value was passed to the function
"""
bin_string = str(bin_string).strip()
diff --git a/conversions/binary_to_hexadecimal.py b/conversions/binary_to_hexadecimal.py
new file mode 100644
index 000000000000..a3855bb70b52
--- /dev/null
+++ b/conversions/binary_to_hexadecimal.py
@@ -0,0 +1,66 @@
+BITS_TO_HEX = {
+ "0000": "0",
+ "0001": "1",
+ "0010": "2",
+ "0011": "3",
+ "0100": "4",
+ "0101": "5",
+ "0110": "6",
+ "0111": "7",
+ "1000": "8",
+ "1001": "9",
+ "1010": "a",
+ "1011": "b",
+ "1100": "c",
+ "1101": "d",
+ "1110": "e",
+ "1111": "f",
+}
+
+
+def bin_to_hexadecimal(binary_str: str) -> str:
+ """
+ Converting a binary string into hexadecimal using Grouping Method
+
+ >>> bin_to_hexadecimal('101011111')
+ '0x15f'
+ >>> bin_to_hexadecimal(' 1010 ')
+ '0x0a'
+ >>> bin_to_hexadecimal('-11101')
+ '-0x1d'
+ >>> bin_to_hexadecimal('a')
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-binary value was passed to the function
+ >>> bin_to_hexadecimal('')
+ Traceback (most recent call last):
+ ...
+ ValueError: Empty string was passed to the function
+ """
+ # Sanitising parameter
+ binary_str = str(binary_str).strip()
+
+ # Exceptions
+ if not binary_str:
+ raise ValueError("Empty string was passed to the function")
+ is_negative = binary_str[0] == "-"
+ binary_str = binary_str[1:] if is_negative else binary_str
+ if not all(char in "01" for char in binary_str):
+ raise ValueError("Non-binary value was passed to the function")
+
+ binary_str = (
+ "0" * (4 * (divmod(len(binary_str), 4)[0] + 1) - len(binary_str)) + binary_str
+ )
+
+ hexadecimal = []
+ for x in range(0, len(binary_str), 4):
+ hexadecimal.append(BITS_TO_HEX[binary_str[x : x + 4]])
+ hexadecimal_str = "0x" + "".join(hexadecimal)
+
+ return "-" + hexadecimal_str if is_negative else hexadecimal_str
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/conversions/binary_to_octal.py b/conversions/binary_to_octal.py
index 8b594887867e..82f81e06234a 100644
--- a/conversions/binary_to_octal.py
+++ b/conversions/binary_to_octal.py
@@ -9,11 +9,11 @@
>>> bin_to_octal("")
Traceback (most recent call last):
-...
+ ...
ValueError: Empty string was passed to the function
>>> bin_to_octal("a-1")
Traceback (most recent call last):
-...
+ ...
ValueError: Non-binary value was passed to the function
"""
@@ -28,7 +28,7 @@ def bin_to_octal(bin_string: str) -> str:
bin_string = "0" + bin_string
bin_string_in_3_list = [
bin_string[index : index + 3]
- for index, value in enumerate(bin_string)
+ for index in range(len(bin_string))
if index % 3 == 0
]
for bin_group in bin_string_in_3_list:
diff --git a/conversions/convert_number_to_words.py b/conversions/convert_number_to_words.py
new file mode 100644
index 000000000000..6aa43738b9fe
--- /dev/null
+++ b/conversions/convert_number_to_words.py
@@ -0,0 +1,205 @@
+from enum import Enum
+from typing import Literal
+
+
+class NumberingSystem(Enum):
+ SHORT = (
+ (15, "quadrillion"),
+ (12, "trillion"),
+ (9, "billion"),
+ (6, "million"),
+ (3, "thousand"),
+ (2, "hundred"),
+ )
+
+ LONG = (
+ (15, "billiard"),
+ (9, "milliard"),
+ (6, "million"),
+ (3, "thousand"),
+ (2, "hundred"),
+ )
+
+ INDIAN = (
+ (14, "crore crore"),
+ (12, "lakh crore"),
+ (7, "crore"),
+ (5, "lakh"),
+ (3, "thousand"),
+ (2, "hundred"),
+ )
+
+ @classmethod
+ def max_value(cls, system: str) -> int:
+ """
+ Gets the max value supported by the given number system.
+
+ >>> NumberingSystem.max_value("short") == 10**18 - 1
+ True
+ >>> NumberingSystem.max_value("long") == 10**21 - 1
+ True
+ >>> NumberingSystem.max_value("indian") == 10**19 - 1
+ True
+ """
+ match system_enum := cls[system.upper()]:
+ case cls.SHORT:
+ max_exp = system_enum.value[0][0] + 3
+ case cls.LONG:
+ max_exp = system_enum.value[0][0] + 6
+ case cls.INDIAN:
+ max_exp = 19
+ case _:
+ raise ValueError("Invalid numbering system")
+ return 10**max_exp - 1
+
+
+class NumberWords(Enum):
+ ONES = { # noqa: RUF012
+ 0: "",
+ 1: "one",
+ 2: "two",
+ 3: "three",
+ 4: "four",
+ 5: "five",
+ 6: "six",
+ 7: "seven",
+ 8: "eight",
+ 9: "nine",
+ }
+
+ TEENS = { # noqa: RUF012
+ 0: "ten",
+ 1: "eleven",
+ 2: "twelve",
+ 3: "thirteen",
+ 4: "fourteen",
+ 5: "fifteen",
+ 6: "sixteen",
+ 7: "seventeen",
+ 8: "eighteen",
+ 9: "nineteen",
+ }
+
+ TENS = { # noqa: RUF012
+ 2: "twenty",
+ 3: "thirty",
+ 4: "forty",
+ 5: "fifty",
+ 6: "sixty",
+ 7: "seventy",
+ 8: "eighty",
+ 9: "ninety",
+ }
+
+
+def convert_small_number(num: int) -> str:
+ """
+ Converts small, non-negative integers with irregular constructions in English (i.e.,
+ numbers under 100) into words.
+
+ >>> convert_small_number(0)
+ 'zero'
+ >>> convert_small_number(5)
+ 'five'
+ >>> convert_small_number(10)
+ 'ten'
+ >>> convert_small_number(15)
+ 'fifteen'
+ >>> convert_small_number(20)
+ 'twenty'
+ >>> convert_small_number(25)
+ 'twenty-five'
+ >>> convert_small_number(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: This function only accepts non-negative integers
+ >>> convert_small_number(123)
+ Traceback (most recent call last):
+ ...
+ ValueError: This function only converts numbers less than 100
+ """
+ if num < 0:
+ raise ValueError("This function only accepts non-negative integers")
+ if num >= 100:
+ raise ValueError("This function only converts numbers less than 100")
+ tens, ones = divmod(num, 10)
+ if tens == 0:
+ return NumberWords.ONES.value[ones] or "zero"
+ if tens == 1:
+ return NumberWords.TEENS.value[ones]
+ return (
+ NumberWords.TENS.value[tens]
+ + ("-" if NumberWords.ONES.value[ones] else "")
+ + NumberWords.ONES.value[ones]
+ )
+
+
+def convert_number(
+ num: int, system: Literal["short", "long", "indian"] = "short"
+) -> str:
+ """
+ Converts an integer to English words.
+
+ :param num: The integer to be converted
+ :param system: The numbering system (short, long, or Indian)
+
+ >>> convert_number(0)
+ 'zero'
+ >>> convert_number(1)
+ 'one'
+ >>> convert_number(100)
+ 'one hundred'
+ >>> convert_number(-100)
+ 'negative one hundred'
+ >>> convert_number(123_456_789_012_345) # doctest: +NORMALIZE_WHITESPACE
+ 'one hundred twenty-three trillion four hundred fifty-six billion
+ seven hundred eighty-nine million twelve thousand three hundred forty-five'
+ >>> convert_number(123_456_789_012_345, "long") # doctest: +NORMALIZE_WHITESPACE
+ 'one hundred twenty-three thousand four hundred fifty-six milliard
+ seven hundred eighty-nine million twelve thousand three hundred forty-five'
+ >>> convert_number(12_34_56_78_90_12_345, "indian") # doctest: +NORMALIZE_WHITESPACE
+ 'one crore crore twenty-three lakh crore
+ forty-five thousand six hundred seventy-eight crore
+ ninety lakh twelve thousand three hundred forty-five'
+ >>> convert_number(10**18)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input number is too large
+ >>> convert_number(10**21, "long")
+ Traceback (most recent call last):
+ ...
+ ValueError: Input number is too large
+ >>> convert_number(10**19, "indian")
+ Traceback (most recent call last):
+ ...
+ ValueError: Input number is too large
+ """
+ word_groups = []
+
+ if num < 0:
+ word_groups.append("negative")
+ num *= -1
+
+ if num > NumberingSystem.max_value(system):
+ raise ValueError("Input number is too large")
+
+ for power, unit in NumberingSystem[system.upper()].value:
+ digit_group, num = divmod(num, 10**power)
+ if digit_group > 0:
+ word_group = (
+ convert_number(digit_group, system)
+ if digit_group >= 100
+ else convert_small_number(digit_group)
+ )
+ word_groups.append(f"{word_group} {unit}")
+ if num > 0 or not word_groups: # word_groups is only empty if input num was 0
+ word_groups.append(convert_small_number(num))
+ return " ".join(word_groups)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ print(f"{convert_number(123456789) = }")
diff --git a/conversions/decimal_to_any.py b/conversions/decimal_to_any.py
index 3c72a7732ac6..c9c2e9a5fb71 100644
--- a/conversions/decimal_to_any.py
+++ b/conversions/decimal_to_any.py
@@ -1,5 +1,9 @@
"""Convert a positive Decimal Number to Any Other Representation"""
+from string import ascii_uppercase
+
+ALPHABET_VALUES = {str(ord(c) - 55): c for c in ascii_uppercase}
+
def decimal_to_any(num: int, base: int) -> str:
"""
@@ -25,32 +29,32 @@ def decimal_to_any(num: int, base: int) -> str:
>>> # negatives will error
>>> decimal_to_any(-45, 8) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
ValueError: parameter must be positive int
>>> # floats will error
>>> decimal_to_any(34.4, 6) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
TypeError: int() can't convert non-string with explicit base
>>> # a float base will error
>>> decimal_to_any(5, 2.5) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
TypeError: 'float' object cannot be interpreted as an integer
>>> # a str base will error
>>> decimal_to_any(10, '16') # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
TypeError: 'str' object cannot be interpreted as an integer
>>> # a base less than 2 will error
>>> decimal_to_any(7, 0) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
ValueError: base must be >= 2
>>> # a base greater than 36 will error
>>> decimal_to_any(34, 37) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
ValueError: base must be <= 36
"""
if isinstance(num, float):
@@ -65,13 +69,6 @@ def decimal_to_any(num: int, base: int) -> str:
raise ValueError("base must be >= 2")
if base > 36:
raise ValueError("base must be <= 36")
- # fmt: off
- ALPHABET_VALUES = {'10': 'A', '11': 'B', '12': 'C', '13': 'D', '14': 'E', '15': 'F',
- '16': 'G', '17': 'H', '18': 'I', '19': 'J', '20': 'K', '21': 'L',
- '22': 'M', '23': 'N', '24': 'O', '25': 'P', '26': 'Q', '27': 'R',
- '28': 'S', '29': 'T', '30': 'U', '31': 'V', '32': 'W', '33': 'X',
- '34': 'Y', '35': 'Z'}
- # fmt: on
new_value = ""
mod = 0
div = 0
@@ -79,8 +76,9 @@ def decimal_to_any(num: int, base: int) -> str:
div, mod = divmod(num, base)
if base >= 11 and 9 < mod < 36:
actual_value = ALPHABET_VALUES[str(mod)]
- mod = actual_value
- new_value += str(mod)
+ else:
+ actual_value = str(mod)
+ new_value += actual_value
div = num // base
num = div
if div == 0:
diff --git a/conversions/decimal_to_binary.py b/conversions/decimal_to_binary.py
index 7e83aee4f7a5..cf2b6040ec2a 100644
--- a/conversions/decimal_to_binary.py
+++ b/conversions/decimal_to_binary.py
@@ -1,36 +1,35 @@
"""Convert a Decimal Number to a Binary Number."""
-def decimal_to_binary(num: int) -> str:
-
+def decimal_to_binary_iterative(num: int) -> str:
"""
Convert an Integer Decimal Number to a Binary Number as str.
- >>> decimal_to_binary(0)
+ >>> decimal_to_binary_iterative(0)
'0b0'
- >>> decimal_to_binary(2)
+ >>> decimal_to_binary_iterative(2)
'0b10'
- >>> decimal_to_binary(7)
+ >>> decimal_to_binary_iterative(7)
'0b111'
- >>> decimal_to_binary(35)
+ >>> decimal_to_binary_iterative(35)
'0b100011'
>>> # negatives work too
- >>> decimal_to_binary(-2)
+ >>> decimal_to_binary_iterative(-2)
'-0b10'
>>> # other floats will error
- >>> decimal_to_binary(16.16) # doctest: +ELLIPSIS
+ >>> decimal_to_binary_iterative(16.16) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
TypeError: 'float' object cannot be interpreted as an integer
>>> # strings will error as well
- >>> decimal_to_binary('0xfffff') # doctest: +ELLIPSIS
+ >>> decimal_to_binary_iterative('0xfffff') # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
TypeError: 'str' object cannot be interpreted as an integer
"""
- if type(num) == float:
+ if isinstance(num, float):
raise TypeError("'float' object cannot be interpreted as an integer")
- if type(num) == str:
+ if isinstance(num, str):
raise TypeError("'str' object cannot be interpreted as an integer")
if num == 0:
@@ -42,7 +41,7 @@ def decimal_to_binary(num: int) -> str:
negative = True
num = -num
- binary = []
+ binary: list[int] = []
while num > 0:
binary.insert(0, num % 2)
num >>= 1
@@ -53,7 +52,58 @@ def decimal_to_binary(num: int) -> str:
return "0b" + "".join(str(e) for e in binary)
+def decimal_to_binary_recursive_helper(decimal: int) -> str:
+ """
+ Take a positive integer value and return its binary equivalent.
+ >>> decimal_to_binary_recursive_helper(1000)
+ '1111101000'
+ >>> decimal_to_binary_recursive_helper("72")
+ '1001000'
+ >>> decimal_to_binary_recursive_helper("number")
+ Traceback (most recent call last):
+ ...
+ ValueError: invalid literal for int() with base 10: 'number'
+ """
+ decimal = int(decimal)
+ if decimal in (0, 1): # Exit cases for the recursion
+ return str(decimal)
+ div, mod = divmod(decimal, 2)
+ return decimal_to_binary_recursive_helper(div) + str(mod)
+
+
+def decimal_to_binary_recursive(number: str) -> str:
+ """
+ Take an integer value and raise ValueError for wrong inputs,
+ call the function above and return the output with prefix "0b" & "-0b"
+ for positive and negative integers respectively.
+ >>> decimal_to_binary_recursive(0)
+ '0b0'
+ >>> decimal_to_binary_recursive(40)
+ '0b101000'
+ >>> decimal_to_binary_recursive(-40)
+ '-0b101000'
+ >>> decimal_to_binary_recursive(40.8)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value is not an integer
+ >>> decimal_to_binary_recursive("forty")
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value is not an integer
+ """
+ number = str(number).strip()
+ if not number:
+ raise ValueError("No input value was provided")
+ negative = "-" if number.startswith("-") else ""
+ number = number.lstrip("-")
+ if not number.isnumeric():
+ raise ValueError("Input value is not an integer")
+ return f"{negative}0b{decimal_to_binary_recursive_helper(int(number))}"
+
+
if __name__ == "__main__":
import doctest
doctest.testmod()
+
+ print(decimal_to_binary_recursive(input("Input a decimal number: ")))
diff --git a/conversions/decimal_to_binary_recursion.py b/conversions/decimal_to_binary_recursion.py
deleted file mode 100644
index c149ea86592f..000000000000
--- a/conversions/decimal_to_binary_recursion.py
+++ /dev/null
@@ -1,53 +0,0 @@
-def binary_recursive(decimal: int) -> str:
- """
- Take a positive integer value and return its binary equivalent.
- >>> binary_recursive(1000)
- '1111101000'
- >>> binary_recursive("72")
- '1001000'
- >>> binary_recursive("number")
- Traceback (most recent call last):
- ...
- ValueError: invalid literal for int() with base 10: 'number'
- """
- decimal = int(decimal)
- if decimal in (0, 1): # Exit cases for the recursion
- return str(decimal)
- div, mod = divmod(decimal, 2)
- return binary_recursive(div) + str(mod)
-
-
-def main(number: str) -> str:
- """
- Take an integer value and raise ValueError for wrong inputs,
- call the function above and return the output with prefix "0b" & "-0b"
- for positive and negative integers respectively.
- >>> main(0)
- '0b0'
- >>> main(40)
- '0b101000'
- >>> main(-40)
- '-0b101000'
- >>> main(40.8)
- Traceback (most recent call last):
- ...
- ValueError: Input value is not an integer
- >>> main("forty")
- Traceback (most recent call last):
- ...
- ValueError: Input value is not an integer
- """
- number = str(number).strip()
- if not number:
- raise ValueError("No input value was provided")
- negative = "-" if number.startswith("-") else ""
- number = number.lstrip("-")
- if not number.isnumeric():
- raise ValueError("Input value is not an integer")
- return f"{negative}0b{binary_recursive(int(number))}"
-
-
-if __name__ == "__main__":
- from doctest import testmod
-
- testmod()
diff --git a/conversions/decimal_to_hexadecimal.py b/conversions/decimal_to_hexadecimal.py
index 433f78dfecb7..ee79592de5ca 100644
--- a/conversions/decimal_to_hexadecimal.py
+++ b/conversions/decimal_to_hexadecimal.py
@@ -1,4 +1,4 @@
-""" Convert Base 10 (Decimal) Values to Hexadecimal Representations """
+"""Convert Base 10 (Decimal) Values to Hexadecimal Representations"""
# set decimal value for each hexadecimal digit
values = {
@@ -21,7 +21,7 @@
}
-def decimal_to_hexadecimal(decimal):
+def decimal_to_hexadecimal(decimal: float) -> str:
"""
take integer decimal value, return hexadecimal representation as str beginning
with 0x
@@ -46,18 +46,20 @@ def decimal_to_hexadecimal(decimal):
>>> # other floats will error
>>> decimal_to_hexadecimal(16.16) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
AssertionError
>>> # strings will error as well
>>> decimal_to_hexadecimal('0xfffff') # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
AssertionError
>>> # results are the same when compared to Python's default hex function
>>> decimal_to_hexadecimal(-256) == hex(-256)
True
"""
- assert type(decimal) in (int, float) and decimal == int(decimal)
+ assert isinstance(decimal, (int, float))
+ assert decimal == int(decimal)
+ decimal = int(decimal)
hexadecimal = ""
negative = False
if decimal < 0:
diff --git a/conversions/decimal_to_octal.py b/conversions/decimal_to_octal.py
index 8dc04830ad87..4c313bddf64c 100644
--- a/conversions/decimal_to_octal.py
+++ b/conversions/decimal_to_octal.py
@@ -17,14 +17,14 @@ def decimal_to_octal(num: int) -> str:
counter = 0
while num > 0:
remainder = num % 8
- octal = octal + (remainder * math.pow(10, counter))
+ octal = octal + (remainder * math.floor(math.pow(10, counter)))
counter += 1
num = math.floor(num / 8) # basically /= 8 without remainder if any
# This formatting removes trailing '.0' from `octal`.
return f"0o{int(octal)}"
-def main():
+def main() -> None:
"""Print octal equivalents of decimal numbers."""
print("\n2 in octal is:")
print(decimal_to_octal(2)) # = 2
diff --git a/conversions/energy_conversions.py b/conversions/energy_conversions.py
new file mode 100644
index 000000000000..51de6b313928
--- /dev/null
+++ b/conversions/energy_conversions.py
@@ -0,0 +1,114 @@
+"""
+Conversion of energy units.
+
+Available units: joule, kilojoule, megajoule, gigajoule,\
+ wattsecond, watthour, kilowatthour, newtonmeter, calorie_nutr,\
+ kilocalorie_nutr, electronvolt, britishthermalunit_it, footpound
+
+USAGE :
+-> Import this file into their respective project.
+-> Use the function energy_conversion() for conversion of energy units.
+-> Parameters :
+ -> from_type : From which type you want to convert
+ -> to_type : To which type you want to convert
+ -> value : the value which you want to convert
+
+REFERENCES :
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Units_of_energy
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Joule
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Kilowatt-hour
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Newton-metre
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Calorie
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Electronvolt
+-> Wikipedia reference: https://en.wikipedia.org/wiki/British_thermal_unit
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Foot-pound_(energy)
+-> Unit converter reference: https://www.unitconverters.net/energy-converter.html
+"""
+
+ENERGY_CONVERSION: dict[str, float] = {
+ "joule": 1.0,
+ "kilojoule": 1_000,
+ "megajoule": 1_000_000,
+ "gigajoule": 1_000_000_000,
+ "wattsecond": 1.0,
+ "watthour": 3_600,
+ "kilowatthour": 3_600_000,
+ "newtonmeter": 1.0,
+ "calorie_nutr": 4_186.8,
+ "kilocalorie_nutr": 4_186_800.00,
+ "electronvolt": 1.602_176_634e-19,
+ "britishthermalunit_it": 1_055.055_85,
+ "footpound": 1.355_818,
+}
+
+
+def energy_conversion(from_type: str, to_type: str, value: float) -> float:
+ """
+ Conversion of energy units.
+ >>> energy_conversion("joule", "joule", 1)
+ 1.0
+ >>> energy_conversion("joule", "kilojoule", 1)
+ 0.001
+ >>> energy_conversion("joule", "megajoule", 1)
+ 1e-06
+ >>> energy_conversion("joule", "gigajoule", 1)
+ 1e-09
+ >>> energy_conversion("joule", "wattsecond", 1)
+ 1.0
+ >>> energy_conversion("joule", "watthour", 1)
+ 0.0002777777777777778
+ >>> energy_conversion("joule", "kilowatthour", 1)
+ 2.7777777777777776e-07
+ >>> energy_conversion("joule", "newtonmeter", 1)
+ 1.0
+ >>> energy_conversion("joule", "calorie_nutr", 1)
+ 0.00023884589662749592
+ >>> energy_conversion("joule", "kilocalorie_nutr", 1)
+ 2.388458966274959e-07
+ >>> energy_conversion("joule", "electronvolt", 1)
+ 6.241509074460763e+18
+ >>> energy_conversion("joule", "britishthermalunit_it", 1)
+ 0.0009478171226670134
+ >>> energy_conversion("joule", "footpound", 1)
+ 0.7375621211696556
+ >>> energy_conversion("joule", "megajoule", 1000)
+ 0.001
+ >>> energy_conversion("calorie_nutr", "kilocalorie_nutr", 1000)
+ 1.0
+ >>> energy_conversion("kilowatthour", "joule", 10)
+ 36000000.0
+ >>> energy_conversion("britishthermalunit_it", "footpound", 1)
+ 778.1692306784539
+ >>> energy_conversion("watthour", "joule", "a") # doctest: +ELLIPSIS
+ Traceback (most recent call last):
+ ...
+ TypeError: unsupported operand type(s) for /: 'str' and 'float'
+ >>> energy_conversion("wrongunit", "joule", 1) # doctest: +ELLIPSIS
+ Traceback (most recent call last):
+ ...
+ ValueError: Incorrect 'from_type' or 'to_type' value: 'wrongunit', 'joule'
+ Valid values are: joule, ... footpound
+ >>> energy_conversion("joule", "wrongunit", 1) # doctest: +ELLIPSIS
+ Traceback (most recent call last):
+ ...
+ ValueError: Incorrect 'from_type' or 'to_type' value: 'joule', 'wrongunit'
+ Valid values are: joule, ... footpound
+ >>> energy_conversion("123", "abc", 1) # doctest: +ELLIPSIS
+ Traceback (most recent call last):
+ ...
+ ValueError: Incorrect 'from_type' or 'to_type' value: '123', 'abc'
+ Valid values are: joule, ... footpound
+ """
+ if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION:
+ msg = (
+ f"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n"
+ f"Valid values are: {', '.join(ENERGY_CONVERSION)}"
+ )
+ raise ValueError(msg)
+ return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/excel_title_to_column.py b/conversions/excel_title_to_column.py
new file mode 100644
index 000000000000..d77031ec26f2
--- /dev/null
+++ b/conversions/excel_title_to_column.py
@@ -0,0 +1,33 @@
+def excel_title_to_column(column_title: str) -> int:
+ """
+ Given a string column_title that represents
+ the column title in an Excel sheet, return
+ its corresponding column number.
+
+ >>> excel_title_to_column("A")
+ 1
+ >>> excel_title_to_column("B")
+ 2
+ >>> excel_title_to_column("AB")
+ 28
+ >>> excel_title_to_column("Z")
+ 26
+ """
+ assert column_title.isupper()
+ answer = 0
+ index = len(column_title) - 1
+ power = 0
+
+ while index >= 0:
+ value = (ord(column_title[index]) - 64) * pow(26, power)
+ answer += value
+ power += 1
+ index -= 1
+
+ return answer
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/conversions/hex_to_bin.py b/conversions/hex_to_bin.py
new file mode 100644
index 000000000000..b872ab5cbce6
--- /dev/null
+++ b/conversions/hex_to_bin.py
@@ -0,0 +1,56 @@
+def hex_to_bin(hex_num: str) -> int:
+ """
+ Convert a hexadecimal value to its binary equivalent
+ #https://stackoverflow.com/questions/1425493/convert-hex-to-binary
+ Here, we have used the bitwise right shift operator: >>
+ Shifts the bits of the number to the right and fills 0 on voids left as a result.
+ Similar effect as of dividing the number with some power of two.
+ Example:
+ a = 10
+ a >> 1 = 5
+
+ >>> hex_to_bin("AC")
+ 10101100
+ >>> hex_to_bin("9A4")
+ 100110100100
+ >>> hex_to_bin(" 12f ")
+ 100101111
+ >>> hex_to_bin("FfFf")
+ 1111111111111111
+ >>> hex_to_bin("-fFfF")
+ -1111111111111111
+ >>> hex_to_bin("F-f")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid value was passed to the function
+ >>> hex_to_bin("")
+ Traceback (most recent call last):
+ ...
+ ValueError: No value was passed to the function
+ """
+
+ hex_num = hex_num.strip()
+ if not hex_num:
+ raise ValueError("No value was passed to the function")
+
+ is_negative = hex_num[0] == "-"
+ if is_negative:
+ hex_num = hex_num[1:]
+
+ try:
+ int_num = int(hex_num, 16)
+ except ValueError:
+ raise ValueError("Invalid value was passed to the function")
+
+ bin_str = ""
+ while int_num > 0:
+ bin_str = str(int_num % 2) + bin_str
+ int_num >>= 1
+
+ return int(("-" + bin_str) if is_negative else bin_str)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/hexadecimal_to_decimal.py b/conversions/hexadecimal_to_decimal.py
index beb1c2c3ded6..209e4aebb368 100644
--- a/conversions/hexadecimal_to_decimal.py
+++ b/conversions/hexadecimal_to_decimal.py
@@ -18,15 +18,15 @@ def hex_to_decimal(hex_string: str) -> int:
-255
>>> hex_to_decimal("F-f")
Traceback (most recent call last):
- ...
+ ...
ValueError: Non-hexadecimal value was passed to the function
>>> hex_to_decimal("")
Traceback (most recent call last):
- ...
+ ...
ValueError: Empty string was passed to the function
>>> hex_to_decimal("12m")
Traceback (most recent call last):
- ...
+ ...
ValueError: Non-hexadecimal value was passed to the function
"""
hex_string = hex_string.strip().lower()
diff --git a/conversions/ipv4_conversion.py b/conversions/ipv4_conversion.py
new file mode 100644
index 000000000000..862309b7251e
--- /dev/null
+++ b/conversions/ipv4_conversion.py
@@ -0,0 +1,85 @@
+# https://www.geeksforgeeks.org/convert-ip-address-to-integer-and-vice-versa/
+
+
+def ipv4_to_decimal(ipv4_address: str) -> int:
+ """
+ Convert an IPv4 address to its decimal representation.
+
+ Args:
+ ip_address: A string representing an IPv4 address (e.g., "192.168.0.1").
+
+ Returns:
+ int: The decimal representation of the IP address.
+
+ >>> ipv4_to_decimal("192.168.0.1")
+ 3232235521
+ >>> ipv4_to_decimal("10.0.0.255")
+ 167772415
+ >>> ipv4_to_decimal("10.0.255")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid IPv4 address format
+ >>> ipv4_to_decimal("10.0.0.256")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid IPv4 octet 256
+ """
+
+ octets = [int(octet) for octet in ipv4_address.split(".")]
+ if len(octets) != 4:
+ raise ValueError("Invalid IPv4 address format")
+
+ decimal_ipv4 = 0
+ for octet in octets:
+ if not 0 <= octet <= 255:
+ raise ValueError(f"Invalid IPv4 octet {octet}") # noqa: EM102
+ decimal_ipv4 = (decimal_ipv4 << 8) + int(octet)
+
+ return decimal_ipv4
+
+
+def alt_ipv4_to_decimal(ipv4_address: str) -> int:
+ """
+ >>> alt_ipv4_to_decimal("192.168.0.1")
+ 3232235521
+ >>> alt_ipv4_to_decimal("10.0.0.255")
+ 167772415
+ """
+ return int("0x" + "".join(f"{int(i):02x}" for i in ipv4_address.split(".")), 16)
+
+
+def decimal_to_ipv4(decimal_ipv4: int) -> str:
+ """
+ Convert a decimal representation of an IP address to its IPv4 format.
+
+ Args:
+ decimal_ipv4: An integer representing the decimal IP address.
+
+ Returns:
+ The IPv4 representation of the decimal IP address.
+
+ >>> decimal_to_ipv4(3232235521)
+ '192.168.0.1'
+ >>> decimal_to_ipv4(167772415)
+ '10.0.0.255'
+ >>> decimal_to_ipv4(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid decimal IPv4 address
+ """
+
+ if not (0 <= decimal_ipv4 <= 4294967295):
+ raise ValueError("Invalid decimal IPv4 address")
+
+ ip_parts = []
+ for _ in range(4):
+ ip_parts.append(str(decimal_ipv4 & 255))
+ decimal_ipv4 >>= 8
+
+ return ".".join(reversed(ip_parts))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/length_conversion.py b/conversions/length_conversion.py
new file mode 100644
index 000000000000..07fa93a198c7
--- /dev/null
+++ b/conversions/length_conversion.py
@@ -0,0 +1,132 @@
+"""
+Conversion of length units.
+Available Units:- Metre,Kilometre,Feet,Inch,Centimeter,Yard,Foot,Mile,Millimeter
+
+USAGE :
+-> Import this file into their respective project.
+-> Use the function length_conversion() for conversion of length units.
+-> Parameters :
+ -> value : The number of from units you want to convert
+ -> from_type : From which type you want to convert
+ -> to_type : To which type you want to convert
+
+REFERENCES :
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Meter
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Kilometer
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Feet
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Inch
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Centimeter
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Yard
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Foot
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Mile
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Millimeter
+"""
+
+from typing import NamedTuple
+
+
+class FromTo(NamedTuple):
+ from_factor: float
+ to_factor: float
+
+
+TYPE_CONVERSION = {
+ "millimeter": "mm",
+ "centimeter": "cm",
+ "meter": "m",
+ "kilometer": "km",
+ "inch": "in",
+ "inche": "in", # Trailing 's' has been stripped off
+ "feet": "ft",
+ "foot": "ft",
+ "yard": "yd",
+ "mile": "mi",
+}
+
+METRIC_CONVERSION = {
+ "mm": FromTo(0.001, 1000),
+ "cm": FromTo(0.01, 100),
+ "m": FromTo(1, 1),
+ "km": FromTo(1000, 0.001),
+ "in": FromTo(0.0254, 39.3701),
+ "ft": FromTo(0.3048, 3.28084),
+ "yd": FromTo(0.9144, 1.09361),
+ "mi": FromTo(1609.34, 0.000621371),
+}
+
+
+def length_conversion(value: float, from_type: str, to_type: str) -> float:
+ """
+ Conversion between length units.
+
+ >>> length_conversion(4, "METER", "FEET")
+ 13.12336
+ >>> length_conversion(4, "M", "FT")
+ 13.12336
+ >>> length_conversion(1, "meter", "kilometer")
+ 0.001
+ >>> length_conversion(1, "kilometer", "inch")
+ 39370.1
+ >>> length_conversion(3, "kilometer", "mile")
+ 1.8641130000000001
+ >>> length_conversion(2, "feet", "meter")
+ 0.6096
+ >>> length_conversion(4, "feet", "yard")
+ 1.333329312
+ >>> length_conversion(1, "inch", "meter")
+ 0.0254
+ >>> length_conversion(2, "inch", "mile")
+ 3.15656468e-05
+ >>> length_conversion(2, "centimeter", "millimeter")
+ 20.0
+ >>> length_conversion(2, "centimeter", "yard")
+ 0.0218722
+ >>> length_conversion(4, "yard", "meter")
+ 3.6576
+ >>> length_conversion(4, "yard", "kilometer")
+ 0.0036576
+ >>> length_conversion(3, "foot", "meter")
+ 0.9144000000000001
+ >>> length_conversion(3, "foot", "inch")
+ 36.00001944
+ >>> length_conversion(4, "mile", "kilometer")
+ 6.43736
+ >>> length_conversion(2, "miles", "InChEs")
+ 126719.753468
+ >>> length_conversion(3, "millimeter", "centimeter")
+ 0.3
+ >>> length_conversion(3, "mm", "in")
+ 0.1181103
+ >>> length_conversion(4, "wrongUnit", "inch")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid 'from_type' value: 'wrongUnit'.
+ Conversion abbreviations are: mm, cm, m, km, in, ft, yd, mi
+ """
+ new_from = from_type.lower().rstrip("s")
+ new_from = TYPE_CONVERSION.get(new_from, new_from)
+ new_to = to_type.lower().rstrip("s")
+ new_to = TYPE_CONVERSION.get(new_to, new_to)
+ if new_from not in METRIC_CONVERSION:
+ msg = (
+ f"Invalid 'from_type' value: {from_type!r}.\n"
+ f"Conversion abbreviations are: {', '.join(METRIC_CONVERSION)}"
+ )
+ raise ValueError(msg)
+ if new_to not in METRIC_CONVERSION:
+ msg = (
+ f"Invalid 'to_type' value: {to_type!r}.\n"
+ f"Conversion abbreviations are: {', '.join(METRIC_CONVERSION)}"
+ )
+ raise ValueError(msg)
+ return (
+ value
+ * METRIC_CONVERSION[new_from].from_factor
+ * METRIC_CONVERSION[new_to].to_factor
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/molecular_chemistry.py b/conversions/molecular_chemistry.py
index 8c68459965b0..51ffe534dd0d 100644
--- a/conversions/molecular_chemistry.py
+++ b/conversions/molecular_chemistry.py
@@ -20,7 +20,7 @@ def molarity_to_normality(nfactor: int, moles: float, volume: float) -> float:
>>> molarity_to_normality(4, 11.4, 5.7)
8
"""
- return round((float(moles / volume) * nfactor))
+ return round(float(moles / volume) * nfactor)
def moles_to_pressure(volume: float, moles: float, temperature: float) -> float:
@@ -86,7 +86,6 @@ def pressure_and_volume_to_temperature(
if __name__ == "__main__":
-
import doctest
doctest.testmod()
diff --git a/conversions/octal_to_binary.py b/conversions/octal_to_binary.py
new file mode 100644
index 000000000000..84e1e85f33ca
--- /dev/null
+++ b/conversions/octal_to_binary.py
@@ -0,0 +1,54 @@
+"""
+* Author: Bama Charan Chhandogi (https://github.com/BamaCharanChhandogi)
+* Description: Convert a Octal number to Binary.
+
+References for better understanding:
+https://en.wikipedia.org/wiki/Binary_number
+https://en.wikipedia.org/wiki/Octal
+"""
+
+
+def octal_to_binary(octal_number: str) -> str:
+ """
+ Convert an Octal number to Binary.
+
+ >>> octal_to_binary("17")
+ '001111'
+ >>> octal_to_binary("7")
+ '111'
+ >>> octal_to_binary("Av")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> octal_to_binary("@#")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> octal_to_binary("")
+ Traceback (most recent call last):
+ ...
+ ValueError: Empty string was passed to the function
+ """
+ if not octal_number:
+ raise ValueError("Empty string was passed to the function")
+
+ binary_number = ""
+ octal_digits = "01234567"
+ for digit in octal_number:
+ if digit not in octal_digits:
+ raise ValueError("Non-octal value was passed to the function")
+
+ binary_digit = ""
+ value = int(digit)
+ for _ in range(3):
+ binary_digit = str(value % 2) + binary_digit
+ value //= 2
+ binary_number += binary_digit
+
+ return binary_number
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/octal_to_decimal b/conversions/octal_to_decimal
deleted file mode 100644
index a5b027e3ae8d..000000000000
--- a/conversions/octal_to_decimal
+++ /dev/null
@@ -1,37 +0,0 @@
-def oct_to_decimal(oct_string: str) -> int:
- """
- Convert a octal value to its decimal equivalent
-
- >>> oct_to_decimal("12")
- 10
- >>> oct_to_decimal(" 12 ")
- 10
- >>> oct_to_decimal("-45")
- -37
- >>> oct_to_decimal("2-0Fm")
- ValueError: Non-octal value was passed to the function
- >>> oct_to_decimal("")
- ValueError: Empty string value was passed to the function
- >>> oct_to_decimal("19")
- ValueError: Non-octal value was passed to the function
- """
- oct_string = str(oct_string).strip()
- if not oct_string:
- raise ValueError("Empty string was passed to the function")
- is_negative = oct_string[0] == "-"
- if is_negative:
- oct_string = oct_string[1:]
- if not all(0 <= int(char) <= 7 for char in oct_string):
- raise ValueError("Non-octal value was passed to the function")
- decimal_number = 0
- for char in oct_string:
- decimal_number = 8 * decimal_number + int(char)
- if is_negative:
- decimal_number = -decimal_number
- return decimal_number
-
-
-if __name__ == "__main__":
- from doctest import testmod
-
- testmod()
diff --git a/conversions/octal_to_decimal.py b/conversions/octal_to_decimal.py
new file mode 100644
index 000000000000..7f006f20e0c8
--- /dev/null
+++ b/conversions/octal_to_decimal.py
@@ -0,0 +1,79 @@
+def oct_to_decimal(oct_string: str) -> int:
+ """
+ Convert a octal value to its decimal equivalent
+
+ >>> oct_to_decimal("")
+ Traceback (most recent call last):
+ ...
+ ValueError: Empty string was passed to the function
+ >>> oct_to_decimal("-")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("e")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("8")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("-e")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("-8")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("1")
+ 1
+ >>> oct_to_decimal("-1")
+ -1
+ >>> oct_to_decimal("12")
+ 10
+ >>> oct_to_decimal(" 12 ")
+ 10
+ >>> oct_to_decimal("-45")
+ -37
+ >>> oct_to_decimal("-")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("0")
+ 0
+ >>> oct_to_decimal("-4055")
+ -2093
+ >>> oct_to_decimal("2-0Fm")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ >>> oct_to_decimal("")
+ Traceback (most recent call last):
+ ...
+ ValueError: Empty string was passed to the function
+ >>> oct_to_decimal("19")
+ Traceback (most recent call last):
+ ...
+ ValueError: Non-octal value was passed to the function
+ """
+ oct_string = str(oct_string).strip()
+ if not oct_string:
+ raise ValueError("Empty string was passed to the function")
+ is_negative = oct_string[0] == "-"
+ if is_negative:
+ oct_string = oct_string[1:]
+ if not oct_string.isdigit() or not all(0 <= int(char) <= 7 for char in oct_string):
+ raise ValueError("Non-octal value was passed to the function")
+ decimal_number = 0
+ for char in oct_string:
+ decimal_number = 8 * decimal_number + int(char)
+ if is_negative:
+ decimal_number = -decimal_number
+ return decimal_number
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/conversions/octal_to_hexadecimal.py b/conversions/octal_to_hexadecimal.py
new file mode 100644
index 000000000000..0615d79b5c53
--- /dev/null
+++ b/conversions/octal_to_hexadecimal.py
@@ -0,0 +1,65 @@
+def octal_to_hex(octal: str) -> str:
+ """
+ Convert an Octal number to Hexadecimal number.
+ For more information: https://en.wikipedia.org/wiki/Octal
+
+ >>> octal_to_hex("100")
+ '0x40'
+ >>> octal_to_hex("235")
+ '0x9D'
+ >>> octal_to_hex(17)
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a string as input
+ >>> octal_to_hex("Av")
+ Traceback (most recent call last):
+ ...
+ ValueError: Not a Valid Octal Number
+ >>> octal_to_hex("")
+ Traceback (most recent call last):
+ ...
+ ValueError: Empty string was passed to the function
+ """
+
+ if not isinstance(octal, str):
+ raise TypeError("Expected a string as input")
+ if octal.startswith("0o"):
+ octal = octal[2:]
+ if octal == "":
+ raise ValueError("Empty string was passed to the function")
+ if any(char not in "01234567" for char in octal):
+ raise ValueError("Not a Valid Octal Number")
+
+ decimal = 0
+ for char in octal:
+ decimal <<= 3
+ decimal |= int(char)
+
+ hex_char = "0123456789ABCDEF"
+
+ revhex = ""
+ while decimal:
+ revhex += hex_char[decimal & 15]
+ decimal >>= 4
+
+ return "0x" + revhex[::-1]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ nums = ["030", "100", "247", "235", "007"]
+
+ ## Main Tests
+
+ for num in nums:
+ hexadecimal = octal_to_hex(num)
+ expected = "0x" + hex(int(num, 8))[2:].upper()
+
+ assert hexadecimal == expected
+
+ print(f"Hex of '0o{num}' is: {hexadecimal}")
+ print(f"Expected was: {expected}")
+ print("---")
diff --git a/conversions/prefix_conversions.py b/conversions/prefix_conversions.py
index c2440d1cf886..714677f3b242 100644
--- a/conversions/prefix_conversions.py
+++ b/conversions/prefix_conversions.py
@@ -1,11 +1,13 @@
"""
Convert International System of Units (SI) and Binary prefixes
"""
+
+from __future__ import annotations
+
from enum import Enum
-from typing import Union
-class SI_Unit(Enum):
+class SIUnit(Enum):
yotta = 24
zetta = 21
exa = 18
@@ -28,7 +30,7 @@ class SI_Unit(Enum):
yocto = -24
-class Binary_Unit(Enum):
+class BinaryUnit(Enum):
yotta = 8
zetta = 7
exa = 6
@@ -41,17 +43,17 @@ class Binary_Unit(Enum):
def convert_si_prefix(
known_amount: float,
- known_prefix: Union[str, SI_Unit],
- unknown_prefix: Union[str, SI_Unit],
+ known_prefix: str | SIUnit,
+ unknown_prefix: str | SIUnit,
) -> float:
"""
Wikipedia reference: https://en.wikipedia.org/wiki/Binary_prefix
Wikipedia reference: https://en.wikipedia.org/wiki/International_System_of_Units
- >>> convert_si_prefix(1, SI_Unit.giga, SI_Unit.mega)
+ >>> convert_si_prefix(1, SIUnit.giga, SIUnit.mega)
1000
- >>> convert_si_prefix(1, SI_Unit.mega, SI_Unit.giga)
+ >>> convert_si_prefix(1, SIUnit.mega, SIUnit.giga)
0.001
- >>> convert_si_prefix(1, SI_Unit.kilo, SI_Unit.kilo)
+ >>> convert_si_prefix(1, SIUnit.kilo, SIUnit.kilo)
1
>>> convert_si_prefix(1, 'giga', 'mega')
1000
@@ -59,25 +61,27 @@ def convert_si_prefix(
1000
"""
if isinstance(known_prefix, str):
- known_prefix: SI_Unit = SI_Unit[known_prefix.lower()]
+ known_prefix = SIUnit[known_prefix.lower()]
if isinstance(unknown_prefix, str):
- unknown_prefix: SI_Unit = SI_Unit[unknown_prefix.lower()]
- unknown_amount = known_amount * (10 ** (known_prefix.value - unknown_prefix.value))
+ unknown_prefix = SIUnit[unknown_prefix.lower()]
+ unknown_amount: float = known_amount * (
+ 10 ** (known_prefix.value - unknown_prefix.value)
+ )
return unknown_amount
def convert_binary_prefix(
known_amount: float,
- known_prefix: Union[str, Binary_Unit],
- unknown_prefix: Union[str, Binary_Unit],
+ known_prefix: str | BinaryUnit,
+ unknown_prefix: str | BinaryUnit,
) -> float:
"""
Wikipedia reference: https://en.wikipedia.org/wiki/Metric_prefix
- >>> convert_binary_prefix(1, Binary_Unit.giga, Binary_Unit.mega)
+ >>> convert_binary_prefix(1, BinaryUnit.giga, BinaryUnit.mega)
1024
- >>> convert_binary_prefix(1, Binary_Unit.mega, Binary_Unit.giga)
+ >>> convert_binary_prefix(1, BinaryUnit.mega, BinaryUnit.giga)
0.0009765625
- >>> convert_binary_prefix(1, Binary_Unit.kilo, Binary_Unit.kilo)
+ >>> convert_binary_prefix(1, BinaryUnit.kilo, BinaryUnit.kilo)
1
>>> convert_binary_prefix(1, 'giga', 'mega')
1024
@@ -85,10 +89,10 @@ def convert_binary_prefix(
1024
"""
if isinstance(known_prefix, str):
- known_prefix: Binary_Unit = Binary_Unit[known_prefix.lower()]
+ known_prefix = BinaryUnit[known_prefix.lower()]
if isinstance(unknown_prefix, str):
- unknown_prefix: Binary_Unit = Binary_Unit[unknown_prefix.lower()]
- unknown_amount = known_amount * (
+ unknown_prefix = BinaryUnit[unknown_prefix.lower()]
+ unknown_amount: float = known_amount * (
2 ** ((known_prefix.value - unknown_prefix.value) * 10)
)
return unknown_amount
diff --git a/conversions/prefix_conversions_string.py b/conversions/prefix_conversions_string.py
new file mode 100644
index 000000000000..c5fef49874ca
--- /dev/null
+++ b/conversions/prefix_conversions_string.py
@@ -0,0 +1,121 @@
+"""
+* Author: Manuel Di Lullo (https://github.com/manueldilullo)
+* Description: Convert a number to use the correct SI or Binary unit prefix.
+
+Inspired by prefix_conversion.py file in this repository by lance-pyles
+
+URL: https://en.wikipedia.org/wiki/Metric_prefix#List_of_SI_prefixes
+URL: https://en.wikipedia.org/wiki/Binary_prefix
+"""
+
+from __future__ import annotations
+
+from enum import Enum, unique
+from typing import TypeVar
+
+# Create a generic variable that can be 'Enum', or any subclass.
+T = TypeVar("T", bound="Enum")
+
+
+@unique
+class BinaryUnit(Enum):
+ yotta = 80
+ zetta = 70
+ exa = 60
+ peta = 50
+ tera = 40
+ giga = 30
+ mega = 20
+ kilo = 10
+
+
+@unique
+class SIUnit(Enum):
+ yotta = 24
+ zetta = 21
+ exa = 18
+ peta = 15
+ tera = 12
+ giga = 9
+ mega = 6
+ kilo = 3
+ hecto = 2
+ deca = 1
+ deci = -1
+ centi = -2
+ milli = -3
+ micro = -6
+ nano = -9
+ pico = -12
+ femto = -15
+ atto = -18
+ zepto = -21
+ yocto = -24
+
+ @classmethod
+ def get_positive(cls) -> dict:
+ """
+ Returns a dictionary with only the elements of this enum
+ that has a positive value
+ >>> from itertools import islice
+ >>> positive = SIUnit.get_positive()
+ >>> inc = iter(positive.items())
+ >>> dict(islice(inc, len(positive) // 2))
+ {'yotta': 24, 'zetta': 21, 'exa': 18, 'peta': 15, 'tera': 12}
+ >>> dict(inc)
+ {'giga': 9, 'mega': 6, 'kilo': 3, 'hecto': 2, 'deca': 1}
+ """
+ return {unit.name: unit.value for unit in cls if unit.value > 0}
+
+ @classmethod
+ def get_negative(cls) -> dict:
+ """
+ Returns a dictionary with only the elements of this enum
+ that has a negative value
+ @example
+ >>> from itertools import islice
+ >>> negative = SIUnit.get_negative()
+ >>> inc = iter(negative.items())
+ >>> dict(islice(inc, len(negative) // 2))
+ {'deci': -1, 'centi': -2, 'milli': -3, 'micro': -6, 'nano': -9}
+ >>> dict(inc)
+ {'pico': -12, 'femto': -15, 'atto': -18, 'zepto': -21, 'yocto': -24}
+ """
+ return {unit.name: unit.value for unit in cls if unit.value < 0}
+
+
+def add_si_prefix(value: float) -> str:
+ """
+ Function that converts a number to his version with SI prefix
+ @input value (an integer)
+ @example:
+ >>> add_si_prefix(10000)
+ '10.0 kilo'
+ """
+ prefixes = SIUnit.get_positive() if value > 0 else SIUnit.get_negative()
+ for name_prefix, value_prefix in prefixes.items():
+ numerical_part = value / (10**value_prefix)
+ if numerical_part > 1:
+ return f"{numerical_part!s} {name_prefix}"
+ return str(value)
+
+
+def add_binary_prefix(value: float) -> str:
+ """
+ Function that converts a number to his version with Binary prefix
+ @input value (an integer)
+ @example:
+ >>> add_binary_prefix(65536)
+ '64.0 kilo'
+ """
+ for prefix in BinaryUnit:
+ numerical_part = value / (2**prefix.value)
+ if numerical_part > 1:
+ return f"{numerical_part!s} {prefix.name}"
+ return str(value)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/pressure_conversions.py b/conversions/pressure_conversions.py
new file mode 100644
index 000000000000..fe78b1382677
--- /dev/null
+++ b/conversions/pressure_conversions.py
@@ -0,0 +1,87 @@
+"""
+Conversion of pressure units.
+Available Units:- Pascal,Bar,Kilopascal,Megapascal,psi(pound per square inch),
+inHg(in mercury column),torr,atm
+USAGE :
+-> Import this file into their respective project.
+-> Use the function pressure_conversion() for conversion of pressure units.
+-> Parameters :
+ -> value : The number of from units you want to convert
+ -> from_type : From which type you want to convert
+ -> to_type : To which type you want to convert
+REFERENCES :
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Pascal_(unit)
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Pound_per_square_inch
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Inch_of_mercury
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Torr
+-> https://en.wikipedia.org/wiki/Standard_atmosphere_(unit)
+-> https://msestudent.com/what-are-the-units-of-pressure/
+-> https://www.unitconverters.net/pressure-converter.html
+"""
+
+from typing import NamedTuple
+
+
+class FromTo(NamedTuple):
+ from_factor: float
+ to_factor: float
+
+
+PRESSURE_CONVERSION = {
+ "atm": FromTo(1, 1),
+ "pascal": FromTo(0.0000098, 101325),
+ "bar": FromTo(0.986923, 1.01325),
+ "kilopascal": FromTo(0.00986923, 101.325),
+ "megapascal": FromTo(9.86923, 0.101325),
+ "psi": FromTo(0.068046, 14.6959),
+ "inHg": FromTo(0.0334211, 29.9213),
+ "torr": FromTo(0.00131579, 760),
+}
+
+
+def pressure_conversion(value: float, from_type: str, to_type: str) -> float:
+ """
+ Conversion between pressure units.
+ >>> pressure_conversion(4, "atm", "pascal")
+ 405300
+ >>> pressure_conversion(1, "pascal", "psi")
+ 0.00014401981999999998
+ >>> pressure_conversion(1, "bar", "atm")
+ 0.986923
+ >>> pressure_conversion(3, "kilopascal", "bar")
+ 0.029999991892499998
+ >>> pressure_conversion(2, "megapascal", "psi")
+ 290.074434314
+ >>> pressure_conversion(4, "psi", "torr")
+ 206.85984
+ >>> pressure_conversion(1, "inHg", "atm")
+ 0.0334211
+ >>> pressure_conversion(1, "torr", "psi")
+ 0.019336718261000002
+ >>> pressure_conversion(4, "wrongUnit", "atm")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid 'from_type' value: 'wrongUnit' Supported values are:
+ atm, pascal, bar, kilopascal, megapascal, psi, inHg, torr
+ """
+ if from_type not in PRESSURE_CONVERSION:
+ raise ValueError(
+ f"Invalid 'from_type' value: {from_type!r} Supported values are:\n"
+ + ", ".join(PRESSURE_CONVERSION)
+ )
+ if to_type not in PRESSURE_CONVERSION:
+ raise ValueError(
+ f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n"
+ + ", ".join(PRESSURE_CONVERSION)
+ )
+ return (
+ value
+ * PRESSURE_CONVERSION[from_type].from_factor
+ * PRESSURE_CONVERSION[to_type].to_factor
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/rectangular_to_polar.py b/conversions/rectangular_to_polar.py
new file mode 100644
index 000000000000..bed97d7410ec
--- /dev/null
+++ b/conversions/rectangular_to_polar.py
@@ -0,0 +1,32 @@
+import math
+
+
+def rectangular_to_polar(real: float, img: float) -> tuple[float, float]:
+ """
+ https://en.wikipedia.org/wiki/Polar_coordinate_system
+
+ >>> rectangular_to_polar(5,-5)
+ (7.07, -45.0)
+ >>> rectangular_to_polar(-1,1)
+ (1.41, 135.0)
+ >>> rectangular_to_polar(-1,-1)
+ (1.41, -135.0)
+ >>> rectangular_to_polar(1e-10,1e-10)
+ (0.0, 45.0)
+ >>> rectangular_to_polar(-1e-10,1e-10)
+ (0.0, 135.0)
+ >>> rectangular_to_polar(9.75,5.93)
+ (11.41, 31.31)
+ >>> rectangular_to_polar(10000,99999)
+ (100497.76, 84.29)
+ """
+
+ mod = round(math.sqrt((real**2) + (img**2)), 2)
+ ang = round(math.degrees(math.atan2(img, real)), 2)
+ return (mod, ang)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/rgb_cmyk_conversion.py b/conversions/rgb_cmyk_conversion.py
new file mode 100644
index 000000000000..07d65b704c44
--- /dev/null
+++ b/conversions/rgb_cmyk_conversion.py
@@ -0,0 +1,71 @@
+def rgb_to_cmyk(r_input: int, g_input: int, b_input: int) -> tuple[int, int, int, int]:
+ """
+ Simple RGB to CMYK conversion. Returns percentages of CMYK paint.
+ https://www.programmingalgorithms.com/algorithm/rgb-to-cmyk/
+
+ Note: this is a very popular algorithm that converts colors linearly and gives
+ only approximate results. Actual preparation for printing requires advanced color
+ conversion considering the color profiles and parameters of the target device.
+
+ >>> rgb_to_cmyk(255, 200, "a")
+ Traceback (most recent call last):
+ ...
+ ValueError: Expected int, found (, , )
+
+ >>> rgb_to_cmyk(255, 255, 999)
+ Traceback (most recent call last):
+ ...
+ ValueError: Expected int of the range 0..255
+
+ >>> rgb_to_cmyk(255, 255, 255) # white
+ (0, 0, 0, 0)
+
+ >>> rgb_to_cmyk(128, 128, 128) # gray
+ (0, 0, 0, 50)
+
+ >>> rgb_to_cmyk(0, 0, 0) # black
+ (0, 0, 0, 100)
+
+ >>> rgb_to_cmyk(255, 0, 0) # red
+ (0, 100, 100, 0)
+
+ >>> rgb_to_cmyk(0, 255, 0) # green
+ (100, 0, 100, 0)
+
+ >>> rgb_to_cmyk(0, 0, 255) # blue
+ (100, 100, 0, 0)
+ """
+
+ if (
+ not isinstance(r_input, int)
+ or not isinstance(g_input, int)
+ or not isinstance(b_input, int)
+ ):
+ msg = f"Expected int, found {type(r_input), type(g_input), type(b_input)}"
+ raise ValueError(msg)
+
+ if not 0 <= r_input < 256 or not 0 <= g_input < 256 or not 0 <= b_input < 256:
+ raise ValueError("Expected int of the range 0..255")
+
+ # changing range from 0..255 to 0..1
+ r = r_input / 255
+ g = g_input / 255
+ b = b_input / 255
+
+ k = 1 - max(r, g, b)
+
+ if k == 1: # pure black
+ return 0, 0, 0, 100
+
+ c = round(100 * (1 - r - k) / (1 - k))
+ m = round(100 * (1 - g - k) / (1 - k))
+ y = round(100 * (1 - b - k) / (1 - k))
+ k = round(100 * k)
+
+ return c, m, y, k
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/conversions/rgb_hsv_conversion.py b/conversions/rgb_hsv_conversion.py
new file mode 100644
index 000000000000..74b3d33e49e7
--- /dev/null
+++ b/conversions/rgb_hsv_conversion.py
@@ -0,0 +1,159 @@
+"""
+The RGB color model is an additive color model in which red, green, and blue light
+are added together in various ways to reproduce a broad array of colors. The name
+of the model comes from the initials of the three additive primary colors, red,
+green, and blue. Meanwhile, the HSV representation models how colors appear under
+light. In it, colors are represented using three components: hue, saturation and
+(brightness-)value. This file provides functions for converting colors from one
+representation to the other.
+
+(description adapted from https://en.wikipedia.org/wiki/RGB_color_model and
+https://en.wikipedia.org/wiki/HSL_and_HSV).
+"""
+
+
+def hsv_to_rgb(hue: float, saturation: float, value: float) -> list[int]:
+ """
+ Conversion from the HSV-representation to the RGB-representation.
+ Expected RGB-values taken from
+ https://www.rapidtables.com/convert/color/hsv-to-rgb.html
+
+ >>> hsv_to_rgb(0, 0, 0)
+ [0, 0, 0]
+ >>> hsv_to_rgb(0, 0, 1)
+ [255, 255, 255]
+ >>> hsv_to_rgb(0, 1, 1)
+ [255, 0, 0]
+ >>> hsv_to_rgb(60, 1, 1)
+ [255, 255, 0]
+ >>> hsv_to_rgb(120, 1, 1)
+ [0, 255, 0]
+ >>> hsv_to_rgb(240, 1, 1)
+ [0, 0, 255]
+ >>> hsv_to_rgb(300, 1, 1)
+ [255, 0, 255]
+ >>> hsv_to_rgb(180, 0.5, 0.5)
+ [64, 128, 128]
+ >>> hsv_to_rgb(234, 0.14, 0.88)
+ [193, 196, 224]
+ >>> hsv_to_rgb(330, 0.75, 0.5)
+ [128, 32, 80]
+ """
+ if hue < 0 or hue > 360:
+ raise Exception("hue should be between 0 and 360")
+
+ if saturation < 0 or saturation > 1:
+ raise Exception("saturation should be between 0 and 1")
+
+ if value < 0 or value > 1:
+ raise Exception("value should be between 0 and 1")
+
+ chroma = value * saturation
+ hue_section = hue / 60
+ second_largest_component = chroma * (1 - abs(hue_section % 2 - 1))
+ match_value = value - chroma
+
+ if hue_section >= 0 and hue_section <= 1:
+ red = round(255 * (chroma + match_value))
+ green = round(255 * (second_largest_component + match_value))
+ blue = round(255 * (match_value))
+ elif hue_section > 1 and hue_section <= 2:
+ red = round(255 * (second_largest_component + match_value))
+ green = round(255 * (chroma + match_value))
+ blue = round(255 * (match_value))
+ elif hue_section > 2 and hue_section <= 3:
+ red = round(255 * (match_value))
+ green = round(255 * (chroma + match_value))
+ blue = round(255 * (second_largest_component + match_value))
+ elif hue_section > 3 and hue_section <= 4:
+ red = round(255 * (match_value))
+ green = round(255 * (second_largest_component + match_value))
+ blue = round(255 * (chroma + match_value))
+ elif hue_section > 4 and hue_section <= 5:
+ red = round(255 * (second_largest_component + match_value))
+ green = round(255 * (match_value))
+ blue = round(255 * (chroma + match_value))
+ else:
+ red = round(255 * (chroma + match_value))
+ green = round(255 * (match_value))
+ blue = round(255 * (second_largest_component + match_value))
+
+ return [red, green, blue]
+
+
+def rgb_to_hsv(red: int, green: int, blue: int) -> list[float]:
+ """
+ Conversion from the RGB-representation to the HSV-representation.
+ The tested values are the reverse values from the hsv_to_rgb-doctests.
+ Function "approximately_equal_hsv" is needed because of small deviations due to
+ rounding for the RGB-values.
+
+ >>> approximately_equal_hsv(rgb_to_hsv(0, 0, 0), [0, 0, 0])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(255, 255, 255), [0, 0, 1])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(255, 0, 0), [0, 1, 1])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(255, 255, 0), [60, 1, 1])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(0, 255, 0), [120, 1, 1])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(0, 0, 255), [240, 1, 1])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(255, 0, 255), [300, 1, 1])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(64, 128, 128), [180, 0.5, 0.5])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(193, 196, 224), [234, 0.14, 0.88])
+ True
+ >>> approximately_equal_hsv(rgb_to_hsv(128, 32, 80), [330, 0.75, 0.5])
+ True
+ """
+ if red < 0 or red > 255:
+ raise Exception("red should be between 0 and 255")
+
+ if green < 0 or green > 255:
+ raise Exception("green should be between 0 and 255")
+
+ if blue < 0 or blue > 255:
+ raise Exception("blue should be between 0 and 255")
+
+ float_red = red / 255
+ float_green = green / 255
+ float_blue = blue / 255
+ value = max(float_red, float_green, float_blue)
+ chroma = value - min(float_red, float_green, float_blue)
+ saturation = 0 if value == 0 else chroma / value
+
+ if chroma == 0:
+ hue = 0.0
+ elif value == float_red:
+ hue = 60 * (0 + (float_green - float_blue) / chroma)
+ elif value == float_green:
+ hue = 60 * (2 + (float_blue - float_red) / chroma)
+ else:
+ hue = 60 * (4 + (float_red - float_green) / chroma)
+
+ hue = (hue + 360) % 360
+
+ return [hue, saturation, value]
+
+
+def approximately_equal_hsv(hsv_1: list[float], hsv_2: list[float]) -> bool:
+ """
+ Utility-function to check that two hsv-colors are approximately equal
+
+ >>> approximately_equal_hsv([0, 0, 0], [0, 0, 0])
+ True
+ >>> approximately_equal_hsv([180, 0.5, 0.3], [179.9999, 0.500001, 0.30001])
+ True
+ >>> approximately_equal_hsv([0, 0, 0], [1, 0, 0])
+ False
+ >>> approximately_equal_hsv([180, 0.5, 0.3], [179.9999, 0.6, 0.30001])
+ False
+ """
+ check_hue = abs(hsv_1[0] - hsv_2[0]) < 0.2
+ check_saturation = abs(hsv_1[1] - hsv_2[1]) < 0.002
+ check_value = abs(hsv_1[2] - hsv_2[2]) < 0.002
+
+ return check_hue and check_saturation and check_value
diff --git a/conversions/roman_to_integer.py b/conversions/roman_numerals.py
similarity index 52%
rename from conversions/roman_to_integer.py
rename to conversions/roman_numerals.py
index ce52b6fb7cbb..75af2ac72882 100644
--- a/conversions/roman_to_integer.py
+++ b/conversions/roman_numerals.py
@@ -1,3 +1,20 @@
+ROMAN = [
+ (1000, "M"),
+ (900, "CM"),
+ (500, "D"),
+ (400, "CD"),
+ (100, "C"),
+ (90, "XC"),
+ (50, "L"),
+ (40, "XL"),
+ (10, "X"),
+ (9, "IX"),
+ (5, "V"),
+ (4, "IV"),
+ (1, "I"),
+]
+
+
def roman_to_int(roman: str) -> int:
"""
LeetCode No. 13 Roman to Integer
@@ -21,6 +38,23 @@ def roman_to_int(roman: str) -> int:
return total
+def int_to_roman(number: int) -> str:
+ """
+ Given a integer, convert it to an roman numeral.
+ https://en.wikipedia.org/wiki/Roman_numerals
+ >>> tests = {"III": 3, "CLIV": 154, "MIX": 1009, "MMD": 2500, "MMMCMXCIX": 3999}
+ >>> all(int_to_roman(value) == key for key, value in tests.items())
+ True
+ """
+ result = []
+ for arabic, roman in ROMAN:
+ (factor, number) = divmod(number, arabic)
+ result.append(roman * factor)
+ if number == 0:
+ break
+ return "".join(result)
+
+
if __name__ == "__main__":
import doctest
diff --git a/conversions/speed_conversions.py b/conversions/speed_conversions.py
new file mode 100644
index 000000000000..ba497119d3f5
--- /dev/null
+++ b/conversions/speed_conversions.py
@@ -0,0 +1,71 @@
+"""
+Convert speed units
+
+https://en.wikipedia.org/wiki/Kilometres_per_hour
+https://en.wikipedia.org/wiki/Miles_per_hour
+https://en.wikipedia.org/wiki/Knot_(unit)
+https://en.wikipedia.org/wiki/Metre_per_second
+"""
+
+speed_chart: dict[str, float] = {
+ "km/h": 1.0,
+ "m/s": 3.6,
+ "mph": 1.609344,
+ "knot": 1.852,
+}
+
+speed_chart_inverse: dict[str, float] = {
+ "km/h": 1.0,
+ "m/s": 0.277777778,
+ "mph": 0.621371192,
+ "knot": 0.539956803,
+}
+
+
+def convert_speed(speed: float, unit_from: str, unit_to: str) -> float:
+ """
+ Convert speed from one unit to another using the speed_chart above.
+
+ "km/h": 1.0,
+ "m/s": 3.6,
+ "mph": 1.609344,
+ "knot": 1.852,
+
+ >>> convert_speed(100, "km/h", "m/s")
+ 27.778
+ >>> convert_speed(100, "km/h", "mph")
+ 62.137
+ >>> convert_speed(100, "km/h", "knot")
+ 53.996
+ >>> convert_speed(100, "m/s", "km/h")
+ 360.0
+ >>> convert_speed(100, "m/s", "mph")
+ 223.694
+ >>> convert_speed(100, "m/s", "knot")
+ 194.384
+ >>> convert_speed(100, "mph", "km/h")
+ 160.934
+ >>> convert_speed(100, "mph", "m/s")
+ 44.704
+ >>> convert_speed(100, "mph", "knot")
+ 86.898
+ >>> convert_speed(100, "knot", "km/h")
+ 185.2
+ >>> convert_speed(100, "knot", "m/s")
+ 51.444
+ >>> convert_speed(100, "knot", "mph")
+ 115.078
+ """
+ if unit_to not in speed_chart or unit_from not in speed_chart_inverse:
+ msg = (
+ f"Incorrect 'from_type' or 'to_type' value: {unit_from!r}, {unit_to!r}\n"
+ f"Valid values are: {', '.join(speed_chart_inverse)}"
+ )
+ raise ValueError(msg)
+ return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to], 3)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/temperature_conversions.py b/conversions/temperature_conversions.py
index 167c9dc64727..dde1d2f0f166 100644
--- a/conversions/temperature_conversions.py
+++ b/conversions/temperature_conversions.py
@@ -1,4 +1,4 @@
-""" Convert between different units of temperature """
+"""Convert between different units of temperature"""
def celsius_to_fahrenheit(celsius: float, ndigits: int = 2) -> float:
@@ -23,7 +23,7 @@ def celsius_to_fahrenheit(celsius: float, ndigits: int = 2) -> float:
104.0
>>> celsius_to_fahrenheit("celsius")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'celsius'
"""
return round((float(celsius) * 9 / 5) + 32, ndigits)
@@ -47,7 +47,7 @@ def celsius_to_kelvin(celsius: float, ndigits: int = 2) -> float:
313.15
>>> celsius_to_kelvin("celsius")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'celsius'
"""
return round(float(celsius) + 273.15, ndigits)
@@ -71,7 +71,7 @@ def celsius_to_rankine(celsius: float, ndigits: int = 2) -> float:
563.67
>>> celsius_to_rankine("celsius")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'celsius'
"""
return round((float(celsius) * 9 / 5) + 491.67, ndigits)
@@ -101,7 +101,7 @@ def fahrenheit_to_celsius(fahrenheit: float, ndigits: int = 2) -> float:
37.78
>>> fahrenheit_to_celsius("fahrenheit")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'fahrenheit'
"""
return round((float(fahrenheit) - 32) * 5 / 9, ndigits)
@@ -131,7 +131,7 @@ def fahrenheit_to_kelvin(fahrenheit: float, ndigits: int = 2) -> float:
310.93
>>> fahrenheit_to_kelvin("fahrenheit")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'fahrenheit'
"""
return round(((float(fahrenheit) - 32) * 5 / 9) + 273.15, ndigits)
@@ -161,7 +161,7 @@ def fahrenheit_to_rankine(fahrenheit: float, ndigits: int = 2) -> float:
559.67
>>> fahrenheit_to_rankine("fahrenheit")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'fahrenheit'
"""
return round(float(fahrenheit) + 459.67, ndigits)
@@ -185,7 +185,7 @@ def kelvin_to_celsius(kelvin: float, ndigits: int = 2) -> float:
42.35
>>> kelvin_to_celsius("kelvin")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'kelvin'
"""
return round(float(kelvin) - 273.15, ndigits)
@@ -209,7 +209,7 @@ def kelvin_to_fahrenheit(kelvin: float, ndigits: int = 2) -> float:
108.23
>>> kelvin_to_fahrenheit("kelvin")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'kelvin'
"""
return round(((float(kelvin) - 273.15) * 9 / 5) + 32, ndigits)
@@ -233,7 +233,7 @@ def kelvin_to_rankine(kelvin: float, ndigits: int = 2) -> float:
72.0
>>> kelvin_to_rankine("kelvin")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'kelvin'
"""
return round((float(kelvin) * 9 / 5), ndigits)
@@ -257,7 +257,7 @@ def rankine_to_celsius(rankine: float, ndigits: int = 2) -> float:
-97.87
>>> rankine_to_celsius("rankine")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'rankine'
"""
return round((float(rankine) - 491.67) * 5 / 9, ndigits)
@@ -277,7 +277,7 @@ def rankine_to_fahrenheit(rankine: float, ndigits: int = 2) -> float:
-144.17
>>> rankine_to_fahrenheit("rankine")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'rankine'
"""
return round(float(rankine) - 459.67, ndigits)
@@ -297,7 +297,7 @@ def rankine_to_kelvin(rankine: float, ndigits: int = 2) -> float:
22.22
>>> rankine_to_kelvin("rankine")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'rankine'
"""
return round((float(rankine) * 5 / 9), ndigits)
@@ -316,7 +316,7 @@ def reaumur_to_kelvin(reaumur: float, ndigits: int = 2) -> float:
323.15
>>> reaumur_to_kelvin("reaumur")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'reaumur'
"""
return round((float(reaumur) * 1.25 + 273.15), ndigits)
@@ -335,7 +335,7 @@ def reaumur_to_fahrenheit(reaumur: float, ndigits: int = 2) -> float:
122.0
>>> reaumur_to_fahrenheit("reaumur")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'reaumur'
"""
return round((float(reaumur) * 2.25 + 32), ndigits)
@@ -354,7 +354,7 @@ def reaumur_to_celsius(reaumur: float, ndigits: int = 2) -> float:
50.0
>>> reaumur_to_celsius("reaumur")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'reaumur'
"""
return round((float(reaumur) * 1.25), ndigits)
@@ -373,14 +373,13 @@ def reaumur_to_rankine(reaumur: float, ndigits: int = 2) -> float:
581.67
>>> reaumur_to_rankine("reaumur")
Traceback (most recent call last):
- ...
+ ...
ValueError: could not convert string to float: 'reaumur'
"""
return round((float(reaumur) * 2.25 + 32 + 459.67), ndigits)
if __name__ == "__main__":
-
import doctest
doctest.testmod()
diff --git a/conversions/time_conversions.py b/conversions/time_conversions.py
new file mode 100644
index 000000000000..8c30f5bc4a45
--- /dev/null
+++ b/conversions/time_conversions.py
@@ -0,0 +1,86 @@
+"""
+A unit of time is any particular time interval, used as a standard way of measuring or
+expressing duration. The base unit of time in the International System of Units (SI),
+and by extension most of the Western world, is the second, defined as about 9 billion
+oscillations of the caesium atom.
+
+https://en.wikipedia.org/wiki/Unit_of_time
+"""
+
+time_chart: dict[str, float] = {
+ "seconds": 1.0,
+ "minutes": 60.0, # 1 minute = 60 sec
+ "hours": 3600.0, # 1 hour = 60 minutes = 3600 seconds
+ "days": 86400.0, # 1 day = 24 hours = 1440 min = 86400 sec
+ "weeks": 604800.0, # 1 week=7d=168hr=10080min = 604800 sec
+ "months": 2629800.0, # Approximate value for a month in seconds
+ "years": 31557600.0, # Approximate value for a year in seconds
+}
+
+time_chart_inverse: dict[str, float] = {
+ key: 1 / value for key, value in time_chart.items()
+}
+
+
+def convert_time(time_value: float, unit_from: str, unit_to: str) -> float:
+ """
+ Convert time from one unit to another using the time_chart above.
+
+ >>> convert_time(3600, "seconds", "hours")
+ 1.0
+ >>> convert_time(3500, "Seconds", "Hours")
+ 0.972
+ >>> convert_time(1, "DaYs", "hours")
+ 24.0
+ >>> convert_time(120, "minutes", "SeCoNdS")
+ 7200.0
+ >>> convert_time(2, "WEEKS", "days")
+ 14.0
+ >>> convert_time(0.5, "hours", "MINUTES")
+ 30.0
+ >>> convert_time(-3600, "seconds", "hours")
+ Traceback (most recent call last):
+ ...
+ ValueError: 'time_value' must be a non-negative number.
+ >>> convert_time("Hello", "hours", "minutes")
+ Traceback (most recent call last):
+ ...
+ ValueError: 'time_value' must be a non-negative number.
+ >>> convert_time([0, 1, 2], "weeks", "days")
+ Traceback (most recent call last):
+ ...
+ ValueError: 'time_value' must be a non-negative number.
+ >>> convert_time(1, "cool", "century") # doctest: +ELLIPSIS
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid unit cool is not in seconds, minutes, hours, days, weeks, ...
+ >>> convert_time(1, "seconds", "hot") # doctest: +ELLIPSIS
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid unit hot is not in seconds, minutes, hours, days, weeks, ...
+ """
+ if not isinstance(time_value, (int, float)) or time_value < 0:
+ msg = "'time_value' must be a non-negative number."
+ raise ValueError(msg)
+
+ unit_from = unit_from.lower()
+ unit_to = unit_to.lower()
+ if unit_from not in time_chart or unit_to not in time_chart:
+ invalid_unit = unit_from if unit_from not in time_chart else unit_to
+ msg = f"Invalid unit {invalid_unit} is not in {', '.join(time_chart)}."
+ raise ValueError(msg)
+
+ return round(
+ time_value * time_chart[unit_from] * time_chart_inverse[unit_to],
+ 3,
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(f"{convert_time(3600,'seconds', 'hours') = :,}")
+ print(f"{convert_time(360, 'days', 'months') = :,}")
+ print(f"{convert_time(360, 'months', 'years') = :,}")
+ print(f"{convert_time(1, 'years', 'seconds') = :,}")
diff --git a/conversions/volume_conversions.py b/conversions/volume_conversions.py
new file mode 100644
index 000000000000..cb240380534b
--- /dev/null
+++ b/conversions/volume_conversions.py
@@ -0,0 +1,83 @@
+"""
+Conversion of volume units.
+Available Units:- Cubic metre,Litre,KiloLitre,Gallon,Cubic yard,Cubic foot,cup
+USAGE :
+-> Import this file into their respective project.
+-> Use the function length_conversion() for conversion of volume units.
+-> Parameters :
+ -> value : The number of from units you want to convert
+ -> from_type : From which type you want to convert
+ -> to_type : To which type you want to convert
+REFERENCES :
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Cubic_metre
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Litre
+-> Wikipedia reference: https://en.wiktionary.org/wiki/kilolitre
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Gallon
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Cubic_yard
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Cubic_foot
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Cup_(unit)
+"""
+
+from typing import NamedTuple
+
+
+class FromTo(NamedTuple):
+ from_factor: float
+ to_factor: float
+
+
+METRIC_CONVERSION = {
+ "cubic meter": FromTo(1, 1),
+ "litre": FromTo(0.001, 1000),
+ "kilolitre": FromTo(1, 1),
+ "gallon": FromTo(0.00454, 264.172),
+ "cubic yard": FromTo(0.76455, 1.30795),
+ "cubic foot": FromTo(0.028, 35.3147),
+ "cup": FromTo(0.000236588, 4226.75),
+}
+
+
+def volume_conversion(value: float, from_type: str, to_type: str) -> float:
+ """
+ Conversion between volume units.
+ >>> volume_conversion(4, "cubic meter", "litre")
+ 4000
+ >>> volume_conversion(1, "litre", "gallon")
+ 0.264172
+ >>> volume_conversion(1, "kilolitre", "cubic meter")
+ 1
+ >>> volume_conversion(3, "gallon", "cubic yard")
+ 0.017814279
+ >>> volume_conversion(2, "cubic yard", "litre")
+ 1529.1
+ >>> volume_conversion(4, "cubic foot", "cup")
+ 473.396
+ >>> volume_conversion(1, "cup", "kilolitre")
+ 0.000236588
+ >>> volume_conversion(4, "wrongUnit", "litre")
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid 'from_type' value: 'wrongUnit' Supported values are:
+ cubic meter, litre, kilolitre, gallon, cubic yard, cubic foot, cup
+ """
+ if from_type not in METRIC_CONVERSION:
+ raise ValueError(
+ f"Invalid 'from_type' value: {from_type!r} Supported values are:\n"
+ + ", ".join(METRIC_CONVERSION)
+ )
+ if to_type not in METRIC_CONVERSION:
+ raise ValueError(
+ f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n"
+ + ", ".join(METRIC_CONVERSION)
+ )
+ return (
+ value
+ * METRIC_CONVERSION[from_type].from_factor
+ * METRIC_CONVERSION[to_type].to_factor
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/conversions/weight_conversion.py b/conversions/weight_conversion.py
new file mode 100644
index 000000000000..0777aead9f02
--- /dev/null
+++ b/conversions/weight_conversion.py
@@ -0,0 +1,319 @@
+"""
+Conversion of weight units.
+
+__author__ = "Anubhav Solanki"
+__license__ = "MIT"
+__version__ = "1.1.0"
+__maintainer__ = "Anubhav Solanki"
+__email__ = "anubhavsolanki0@gmail.com"
+
+USAGE :
+-> Import this file into their respective project.
+-> Use the function weight_conversion() for conversion of weight units.
+-> Parameters :
+ -> from_type : From which type you want to convert
+ -> to_type : To which type you want to convert
+ -> value : the value which you want to convert
+
+REFERENCES :
+
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Kilogram
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Gram
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Millimetre
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Tonne
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Long_ton
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Short_ton
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Pound
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Ounce
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Fineness#Karat
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Dalton_(unit)
+-> Wikipedia reference: https://en.wikipedia.org/wiki/Stone_(unit)
+"""
+
+KILOGRAM_CHART: dict[str, float] = {
+ "kilogram": 1,
+ "gram": pow(10, 3),
+ "milligram": pow(10, 6),
+ "metric-ton": pow(10, -3),
+ "long-ton": 0.0009842073,
+ "short-ton": 0.0011023122,
+ "pound": 2.2046244202,
+ "stone": 0.1574731728,
+ "ounce": 35.273990723,
+ "carrat": 5000,
+ "atomic-mass-unit": 6.022136652e26,
+}
+
+WEIGHT_TYPE_CHART: dict[str, float] = {
+ "kilogram": 1,
+ "gram": pow(10, -3),
+ "milligram": pow(10, -6),
+ "metric-ton": pow(10, 3),
+ "long-ton": 1016.04608,
+ "short-ton": 907.184,
+ "pound": 0.453592,
+ "stone": 6.35029,
+ "ounce": 0.0283495,
+ "carrat": 0.0002,
+ "atomic-mass-unit": 1.660540199e-27,
+}
+
+
+def weight_conversion(from_type: str, to_type: str, value: float) -> float:
+ """
+ Conversion of weight unit with the help of KILOGRAM_CHART
+
+ "kilogram" : 1,
+ "gram" : pow(10, 3),
+ "milligram" : pow(10, 6),
+ "metric-ton" : pow(10, -3),
+ "long-ton" : 0.0009842073,
+ "short-ton" : 0.0011023122,
+ "pound" : 2.2046244202,
+ "stone": 0.1574731728,
+ "ounce" : 35.273990723,
+ "carrat" : 5000,
+ "atomic-mass-unit" : 6.022136652E+26
+
+ >>> weight_conversion("kilogram","kilogram",4)
+ 4
+ >>> weight_conversion("kilogram","gram",1)
+ 1000
+ >>> weight_conversion("kilogram","milligram",4)
+ 4000000
+ >>> weight_conversion("kilogram","metric-ton",4)
+ 0.004
+ >>> weight_conversion("kilogram","long-ton",3)
+ 0.0029526219
+ >>> weight_conversion("kilogram","short-ton",1)
+ 0.0011023122
+ >>> weight_conversion("kilogram","pound",4)
+ 8.8184976808
+ >>> weight_conversion("kilogram","stone",5)
+ 0.7873658640000001
+ >>> weight_conversion("kilogram","ounce",4)
+ 141.095962892
+ >>> weight_conversion("kilogram","carrat",3)
+ 15000
+ >>> weight_conversion("kilogram","atomic-mass-unit",1)
+ 6.022136652e+26
+ >>> weight_conversion("gram","kilogram",1)
+ 0.001
+ >>> weight_conversion("gram","gram",3)
+ 3.0
+ >>> weight_conversion("gram","milligram",2)
+ 2000.0
+ >>> weight_conversion("gram","metric-ton",4)
+ 4e-06
+ >>> weight_conversion("gram","long-ton",3)
+ 2.9526219e-06
+ >>> weight_conversion("gram","short-ton",3)
+ 3.3069366000000003e-06
+ >>> weight_conversion("gram","pound",3)
+ 0.0066138732606
+ >>> weight_conversion("gram","stone",4)
+ 0.0006298926912000001
+ >>> weight_conversion("gram","ounce",1)
+ 0.035273990723
+ >>> weight_conversion("gram","carrat",2)
+ 10.0
+ >>> weight_conversion("gram","atomic-mass-unit",1)
+ 6.022136652e+23
+ >>> weight_conversion("milligram","kilogram",1)
+ 1e-06
+ >>> weight_conversion("milligram","gram",2)
+ 0.002
+ >>> weight_conversion("milligram","milligram",3)
+ 3.0
+ >>> weight_conversion("milligram","metric-ton",3)
+ 3e-09
+ >>> weight_conversion("milligram","long-ton",3)
+ 2.9526219e-09
+ >>> weight_conversion("milligram","short-ton",1)
+ 1.1023122e-09
+ >>> weight_conversion("milligram","pound",3)
+ 6.6138732605999995e-06
+ >>> weight_conversion("milligram","ounce",2)
+ 7.054798144599999e-05
+ >>> weight_conversion("milligram","carrat",1)
+ 0.005
+ >>> weight_conversion("milligram","atomic-mass-unit",1)
+ 6.022136652e+20
+ >>> weight_conversion("metric-ton","kilogram",2)
+ 2000
+ >>> weight_conversion("metric-ton","gram",2)
+ 2000000
+ >>> weight_conversion("metric-ton","milligram",3)
+ 3000000000
+ >>> weight_conversion("metric-ton","metric-ton",2)
+ 2.0
+ >>> weight_conversion("metric-ton","long-ton",3)
+ 2.9526219
+ >>> weight_conversion("metric-ton","short-ton",2)
+ 2.2046244
+ >>> weight_conversion("metric-ton","pound",3)
+ 6613.8732606
+ >>> weight_conversion("metric-ton","ounce",4)
+ 141095.96289199998
+ >>> weight_conversion("metric-ton","carrat",4)
+ 20000000
+ >>> weight_conversion("metric-ton","atomic-mass-unit",1)
+ 6.022136652e+29
+ >>> weight_conversion("long-ton","kilogram",4)
+ 4064.18432
+ >>> weight_conversion("long-ton","gram",4)
+ 4064184.32
+ >>> weight_conversion("long-ton","milligram",3)
+ 3048138240.0
+ >>> weight_conversion("long-ton","metric-ton",4)
+ 4.06418432
+ >>> weight_conversion("long-ton","long-ton",3)
+ 2.999999907217152
+ >>> weight_conversion("long-ton","short-ton",1)
+ 1.119999989746176
+ >>> weight_conversion("long-ton","pound",3)
+ 6720.000000049448
+ >>> weight_conversion("long-ton","ounce",1)
+ 35840.000000060514
+ >>> weight_conversion("long-ton","carrat",4)
+ 20320921.599999998
+ >>> weight_conversion("long-ton","atomic-mass-unit",4)
+ 2.4475073353955697e+30
+ >>> weight_conversion("short-ton","kilogram",3)
+ 2721.5519999999997
+ >>> weight_conversion("short-ton","gram",3)
+ 2721552.0
+ >>> weight_conversion("short-ton","milligram",1)
+ 907184000.0
+ >>> weight_conversion("short-ton","metric-ton",4)
+ 3.628736
+ >>> weight_conversion("short-ton","long-ton",3)
+ 2.6785713457296
+ >>> weight_conversion("short-ton","short-ton",3)
+ 2.9999999725344
+ >>> weight_conversion("short-ton","pound",2)
+ 4000.0000000294335
+ >>> weight_conversion("short-ton","ounce",4)
+ 128000.00000021611
+ >>> weight_conversion("short-ton","carrat",4)
+ 18143680.0
+ >>> weight_conversion("short-ton","atomic-mass-unit",1)
+ 5.463186016507968e+29
+ >>> weight_conversion("pound","kilogram",4)
+ 1.814368
+ >>> weight_conversion("pound","gram",2)
+ 907.184
+ >>> weight_conversion("pound","milligram",3)
+ 1360776.0
+ >>> weight_conversion("pound","metric-ton",3)
+ 0.001360776
+ >>> weight_conversion("pound","long-ton",2)
+ 0.0008928571152432
+ >>> weight_conversion("pound","short-ton",1)
+ 0.0004999999954224
+ >>> weight_conversion("pound","pound",3)
+ 3.0000000000220752
+ >>> weight_conversion("pound","ounce",1)
+ 16.000000000027015
+ >>> weight_conversion("pound","carrat",1)
+ 2267.96
+ >>> weight_conversion("pound","atomic-mass-unit",4)
+ 1.0926372033015936e+27
+ >>> weight_conversion("stone","kilogram",5)
+ 31.751450000000002
+ >>> weight_conversion("stone","gram",2)
+ 12700.58
+ >>> weight_conversion("stone","milligram",3)
+ 19050870.0
+ >>> weight_conversion("stone","metric-ton",3)
+ 0.01905087
+ >>> weight_conversion("stone","long-ton",3)
+ 0.018750005325351003
+ >>> weight_conversion("stone","short-ton",3)
+ 0.021000006421614002
+ >>> weight_conversion("stone","pound",2)
+ 28.00000881870372
+ >>> weight_conversion("stone","ounce",1)
+ 224.00007054835967
+ >>> weight_conversion("stone","carrat",2)
+ 63502.9
+ >>> weight_conversion("ounce","kilogram",3)
+ 0.0850485
+ >>> weight_conversion("ounce","gram",3)
+ 85.0485
+ >>> weight_conversion("ounce","milligram",4)
+ 113398.0
+ >>> weight_conversion("ounce","metric-ton",4)
+ 0.000113398
+ >>> weight_conversion("ounce","long-ton",4)
+ 0.0001116071394054
+ >>> weight_conversion("ounce","short-ton",4)
+ 0.0001249999988556
+ >>> weight_conversion("ounce","pound",1)
+ 0.0625000000004599
+ >>> weight_conversion("ounce","ounce",2)
+ 2.000000000003377
+ >>> weight_conversion("ounce","carrat",1)
+ 141.7475
+ >>> weight_conversion("ounce","atomic-mass-unit",1)
+ 1.70724563015874e+25
+ >>> weight_conversion("carrat","kilogram",1)
+ 0.0002
+ >>> weight_conversion("carrat","gram",4)
+ 0.8
+ >>> weight_conversion("carrat","milligram",2)
+ 400.0
+ >>> weight_conversion("carrat","metric-ton",2)
+ 4.0000000000000003e-07
+ >>> weight_conversion("carrat","long-ton",3)
+ 5.9052438e-07
+ >>> weight_conversion("carrat","short-ton",4)
+ 8.818497600000002e-07
+ >>> weight_conversion("carrat","pound",1)
+ 0.00044092488404000004
+ >>> weight_conversion("carrat","ounce",2)
+ 0.0141095962892
+ >>> weight_conversion("carrat","carrat",4)
+ 4.0
+ >>> weight_conversion("carrat","atomic-mass-unit",4)
+ 4.8177093216e+23
+ >>> weight_conversion("atomic-mass-unit","kilogram",4)
+ 6.642160796e-27
+ >>> weight_conversion("atomic-mass-unit","gram",2)
+ 3.321080398e-24
+ >>> weight_conversion("atomic-mass-unit","milligram",2)
+ 3.3210803980000002e-21
+ >>> weight_conversion("atomic-mass-unit","metric-ton",3)
+ 4.9816205970000004e-30
+ >>> weight_conversion("atomic-mass-unit","long-ton",3)
+ 4.9029473573977584e-30
+ >>> weight_conversion("atomic-mass-unit","short-ton",1)
+ 1.830433719948128e-30
+ >>> weight_conversion("atomic-mass-unit","pound",3)
+ 1.0982602420317504e-26
+ >>> weight_conversion("atomic-mass-unit","ounce",2)
+ 1.1714775914938915e-25
+ >>> weight_conversion("atomic-mass-unit","carrat",2)
+ 1.660540199e-23
+ >>> weight_conversion("atomic-mass-unit","atomic-mass-unit",2)
+ 1.999999998903455
+ >>> weight_conversion("slug", "kilogram", 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid 'from_type' or 'to_type' value: 'slug', 'kilogram'
+ Supported values are: kilogram, gram, milligram, metric-ton, long-ton, short-ton, \
+pound, stone, ounce, carrat, atomic-mass-unit
+ """
+ if to_type not in KILOGRAM_CHART or from_type not in WEIGHT_TYPE_CHART:
+ msg = (
+ f"Invalid 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n"
+ f"Supported values are: {', '.join(WEIGHT_TYPE_CHART)}"
+ )
+ raise ValueError(msg)
+ return value * KILOGRAM_CHART[to_type] * WEIGHT_TYPE_CHART[from_type]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/arithmetic_analysis/image_data/__init__.py b/data_structures/arrays/__init__.py
similarity index 100%
rename from arithmetic_analysis/image_data/__init__.py
rename to data_structures/arrays/__init__.py
diff --git a/data_structures/arrays/equilibrium_index_in_array.py b/data_structures/arrays/equilibrium_index_in_array.py
new file mode 100644
index 000000000000..0717a45d9f4b
--- /dev/null
+++ b/data_structures/arrays/equilibrium_index_in_array.py
@@ -0,0 +1,58 @@
+"""
+Find the Equilibrium Index of an Array.
+Reference: https://www.geeksforgeeks.org/equilibrium-index-of-an-array/
+
+Python doctest can be run with the following command:
+python -m doctest -v equilibrium_index_in_array.py
+
+Given a sequence arr[] of size n, this function returns
+an equilibrium index (if any) or -1 if no equilibrium index exists.
+
+The equilibrium index of an array is an index such that the sum of
+elements at lower indexes is equal to the sum of elements at higher indexes.
+
+
+
+Example Input:
+arr = [-7, 1, 5, 2, -4, 3, 0]
+Output: 3
+
+"""
+
+
+def equilibrium_index(arr: list[int]) -> int:
+ """
+ Find the equilibrium index of an array.
+
+ Args:
+ arr (list[int]): The input array of integers.
+
+ Returns:
+ int: The equilibrium index or -1 if no equilibrium index exists.
+
+ Examples:
+ >>> equilibrium_index([-7, 1, 5, 2, -4, 3, 0])
+ 3
+ >>> equilibrium_index([1, 2, 3, 4, 5])
+ -1
+ >>> equilibrium_index([1, 1, 1, 1, 1])
+ 2
+ >>> equilibrium_index([2, 4, 6, 8, 10, 3])
+ -1
+ """
+ total_sum = sum(arr)
+ left_sum = 0
+
+ for i, value in enumerate(arr):
+ total_sum -= value
+ if left_sum == total_sum:
+ return i
+ left_sum += value
+
+ return -1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/find_triplets_with_0_sum.py b/data_structures/arrays/find_triplets_with_0_sum.py
new file mode 100644
index 000000000000..52e521906873
--- /dev/null
+++ b/data_structures/arrays/find_triplets_with_0_sum.py
@@ -0,0 +1,87 @@
+from itertools import combinations
+
+
+def find_triplets_with_0_sum(nums: list[int]) -> list[list[int]]:
+ """
+ Given a list of integers, return elements a, b, c such that a + b + c = 0.
+ Args:
+ nums: list of integers
+ Returns:
+ list of lists of integers where sum(each_list) == 0
+ Examples:
+ >>> find_triplets_with_0_sum([-1, 0, 1, 2, -1, -4])
+ [[-1, -1, 2], [-1, 0, 1]]
+ >>> find_triplets_with_0_sum([])
+ []
+ >>> find_triplets_with_0_sum([0, 0, 0])
+ [[0, 0, 0]]
+ >>> find_triplets_with_0_sum([1, 2, 3, 0, -1, -2, -3])
+ [[-3, 0, 3], [-3, 1, 2], [-2, -1, 3], [-2, 0, 2], [-1, 0, 1]]
+ """
+ return [
+ list(x)
+ for x in sorted({abc for abc in combinations(sorted(nums), 3) if not sum(abc)})
+ ]
+
+
+def find_triplets_with_0_sum_hashing(arr: list[int]) -> list[list[int]]:
+ """
+ Function for finding the triplets with a given sum in the array using hashing.
+
+ Given a list of integers, return elements a, b, c such that a + b + c = 0.
+
+ Args:
+ nums: list of integers
+ Returns:
+ list of lists of integers where sum(each_list) == 0
+ Examples:
+ >>> find_triplets_with_0_sum_hashing([-1, 0, 1, 2, -1, -4])
+ [[-1, 0, 1], [-1, -1, 2]]
+ >>> find_triplets_with_0_sum_hashing([])
+ []
+ >>> find_triplets_with_0_sum_hashing([0, 0, 0])
+ [[0, 0, 0]]
+ >>> find_triplets_with_0_sum_hashing([1, 2, 3, 0, -1, -2, -3])
+ [[-1, 0, 1], [-3, 1, 2], [-2, 0, 2], [-2, -1, 3], [-3, 0, 3]]
+
+ Time complexity: O(N^2)
+ Auxiliary Space: O(N)
+
+ """
+ target_sum = 0
+
+ # Initialize the final output array with blank.
+ output_arr = []
+
+ # Set the initial element as arr[i].
+ for index, item in enumerate(arr[:-2]):
+ # to store second elements that can complement the final sum.
+ set_initialize = set()
+
+ # current sum needed for reaching the target sum
+ current_sum = target_sum - item
+
+ # Traverse the subarray arr[i+1:].
+ for other_item in arr[index + 1 :]:
+ # required value for the second element
+ required_value = current_sum - other_item
+
+ # Verify if the desired value exists in the set.
+ if required_value in set_initialize:
+ # finding triplet elements combination.
+ combination_array = sorted([item, other_item, required_value])
+ if combination_array not in output_arr:
+ output_arr.append(combination_array)
+
+ # Include the current element in the set
+ # for subsequent complement verification.
+ set_initialize.add(other_item)
+
+ # Return all the triplet combinations.
+ return output_arr
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/data_structures/arrays/index_2d_array_in_1d.py b/data_structures/arrays/index_2d_array_in_1d.py
new file mode 100644
index 000000000000..27a9fa5f9121
--- /dev/null
+++ b/data_structures/arrays/index_2d_array_in_1d.py
@@ -0,0 +1,105 @@
+"""
+Retrieves the value of an 0-indexed 1D index from a 2D array.
+There are two ways to retrieve value(s):
+
+1. Index2DArrayIterator(matrix) -> Iterator[int]
+This iterator allows you to iterate through a 2D array by passing in the matrix and
+calling next(your_iterator). You can also use the iterator in a loop.
+Examples:
+list(Index2DArrayIterator(matrix))
+set(Index2DArrayIterator(matrix))
+tuple(Index2DArrayIterator(matrix))
+sum(Index2DArrayIterator(matrix))
+-5 in Index2DArrayIterator(matrix)
+
+2. index_2d_array_in_1d(array: list[int], index: int) -> int
+This function allows you to provide a 2D array and a 0-indexed 1D integer index,
+and retrieves the integer value at that index.
+
+Python doctests can be run using this command:
+python3 -m doctest -v index_2d_array_in_1d.py
+"""
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class Index2DArrayIterator:
+ matrix: list[list[int]]
+
+ def __iter__(self) -> Iterator[int]:
+ """
+ >>> tuple(Index2DArrayIterator([[5], [-523], [-1], [34], [0]]))
+ (5, -523, -1, 34, 0)
+ >>> tuple(Index2DArrayIterator([[5, -523, -1], [34, 0]]))
+ (5, -523, -1, 34, 0)
+ >>> tuple(Index2DArrayIterator([[5, -523, -1, 34, 0]]))
+ (5, -523, -1, 34, 0)
+ >>> t = Index2DArrayIterator([[5, 2, 25], [23, 14, 5], [324, -1, 0]])
+ >>> tuple(t)
+ (5, 2, 25, 23, 14, 5, 324, -1, 0)
+ >>> list(t)
+ [5, 2, 25, 23, 14, 5, 324, -1, 0]
+ >>> sorted(t)
+ [-1, 0, 2, 5, 5, 14, 23, 25, 324]
+ >>> tuple(t)[3]
+ 23
+ >>> sum(t)
+ 397
+ >>> -1 in t
+ True
+ >>> t = iter(Index2DArrayIterator([[5], [-523], [-1], [34], [0]]))
+ >>> next(t)
+ 5
+ >>> next(t)
+ -523
+ """
+ for row in self.matrix:
+ yield from row
+
+
+def index_2d_array_in_1d(array: list[list[int]], index: int) -> int:
+ """
+ Retrieves the value of the one-dimensional index from a two-dimensional array.
+
+ Args:
+ array: A 2D array of integers where all rows are the same size and all
+ columns are the same size.
+ index: A 1D index.
+
+ Returns:
+ int: The 0-indexed value of the 1D index in the array.
+
+ Examples:
+ >>> index_2d_array_in_1d([[0, 1, 2, 3], [4, 5, 6, 7], [8, 9, 10, 11]], 5)
+ 5
+ >>> index_2d_array_in_1d([[0, 1, 2, 3], [4, 5, 6, 7], [8, 9, 10, 11]], -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: index out of range
+ >>> index_2d_array_in_1d([[0, 1, 2, 3], [4, 5, 6, 7], [8, 9, 10, 11]], 12)
+ Traceback (most recent call last):
+ ...
+ ValueError: index out of range
+ >>> index_2d_array_in_1d([[]], 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: no items in array
+ """
+ rows = len(array)
+ cols = len(array[0])
+
+ if rows == 0 or cols == 0:
+ raise ValueError("no items in array")
+
+ if index < 0 or index >= rows * cols:
+ raise ValueError("index out of range")
+
+ return array[index // cols][index % cols]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/kth_largest_element.py b/data_structures/arrays/kth_largest_element.py
new file mode 100644
index 000000000000..f25cc68e9b72
--- /dev/null
+++ b/data_structures/arrays/kth_largest_element.py
@@ -0,0 +1,117 @@
+"""
+Given an array of integers and an integer k, find the kth largest element in the array.
+
+https://stackoverflow.com/questions/251781
+"""
+
+
+def partition(arr: list[int], low: int, high: int) -> int:
+ """
+ Partitions list based on the pivot element.
+
+ This function rearranges the elements in the input list 'elements' such that
+ all elements greater than or equal to the chosen pivot are on the right side
+ of the pivot, and all elements smaller than the pivot are on the left side.
+
+ Args:
+ arr: The list to be partitioned
+ low: The lower index of the list
+ high: The higher index of the list
+
+ Returns:
+ int: The index of pivot element after partitioning
+
+ Examples:
+ >>> partition([3, 1, 4, 5, 9, 2, 6, 5, 3, 5], 0, 9)
+ 4
+ >>> partition([7, 1, 4, 5, 9, 2, 6, 5, 8], 0, 8)
+ 1
+ >>> partition(['apple', 'cherry', 'date', 'banana'], 0, 3)
+ 2
+ >>> partition([3.1, 1.2, 5.6, 4.7], 0, 3)
+ 1
+ """
+ pivot = arr[high]
+ i = low - 1
+ for j in range(low, high):
+ if arr[j] >= pivot:
+ i += 1
+ arr[i], arr[j] = arr[j], arr[i]
+ arr[i + 1], arr[high] = arr[high], arr[i + 1]
+ return i + 1
+
+
+def kth_largest_element(arr: list[int], position: int) -> int:
+ """
+ Finds the kth largest element in a list.
+ Should deliver similar results to:
+ ```python
+ def kth_largest_element(arr, position):
+ return sorted(arr)[-position]
+ ```
+
+ Args:
+ nums: The list of numbers.
+ k: The position of the desired kth largest element.
+
+ Returns:
+ int: The kth largest element.
+
+ Examples:
+ >>> kth_largest_element([3, 1, 4, 1, 5, 9, 2, 6, 5, 3, 5], 3)
+ 5
+ >>> kth_largest_element([2, 5, 6, 1, 9, 3, 8, 4, 7, 3, 5], 1)
+ 9
+ >>> kth_largest_element([2, 5, 6, 1, 9, 3, 8, 4, 7, 3, 5], -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid value of 'position'
+ >>> kth_largest_element([9, 1, 3, 6, 7, 9, 8, 4, 2, 4, 9], 110)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid value of 'position'
+ >>> kth_largest_element([1, 2, 4, 3, 5, 9, 7, 6, 5, 9, 3], 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid value of 'position'
+ >>> kth_largest_element(['apple', 'cherry', 'date', 'banana'], 2)
+ 'cherry'
+ >>> kth_largest_element([3.1, 1.2, 5.6, 4.7,7.9,5,0], 2)
+ 5.6
+ >>> kth_largest_element([-2, -5, -4, -1], 1)
+ -1
+ >>> kth_largest_element([], 1)
+ -1
+ >>> kth_largest_element([3.1, 1.2, 5.6, 4.7, 7.9, 5, 0], 1.5)
+ Traceback (most recent call last):
+ ...
+ ValueError: The position should be an integer
+ >>> kth_largest_element((4, 6, 1, 2), 4)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'tuple' object does not support item assignment
+ """
+ if not arr:
+ return -1
+ if not isinstance(position, int):
+ raise ValueError("The position should be an integer")
+ if not 1 <= position <= len(arr):
+ raise ValueError("Invalid value of 'position'")
+ low, high = 0, len(arr) - 1
+ while low <= high:
+ if low > len(arr) - 1 or high < 0:
+ return -1
+ pivot_index = partition(arr, low, high)
+ if pivot_index == position - 1:
+ return arr[pivot_index]
+ elif pivot_index > position - 1:
+ high = pivot_index - 1
+ else:
+ low = pivot_index + 1
+ return -1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/median_two_array.py b/data_structures/arrays/median_two_array.py
new file mode 100644
index 000000000000..972b0ee44201
--- /dev/null
+++ b/data_structures/arrays/median_two_array.py
@@ -0,0 +1,61 @@
+"""
+https://www.enjoyalgorithms.com/blog/median-of-two-sorted-arrays
+"""
+
+
+def find_median_sorted_arrays(nums1: list[int], nums2: list[int]) -> float:
+ """
+ Find the median of two arrays.
+
+ Args:
+ nums1: The first array.
+ nums2: The second array.
+
+ Returns:
+ The median of the two arrays.
+
+ Examples:
+ >>> find_median_sorted_arrays([1, 3], [2])
+ 2.0
+
+ >>> find_median_sorted_arrays([1, 2], [3, 4])
+ 2.5
+
+ >>> find_median_sorted_arrays([0, 0], [0, 0])
+ 0.0
+
+ >>> find_median_sorted_arrays([], [])
+ Traceback (most recent call last):
+ ...
+ ValueError: Both input arrays are empty.
+
+ >>> find_median_sorted_arrays([], [1])
+ 1.0
+
+ >>> find_median_sorted_arrays([-1000], [1000])
+ 0.0
+
+ >>> find_median_sorted_arrays([-1.1, -2.2], [-3.3, -4.4])
+ -2.75
+ """
+ if not nums1 and not nums2:
+ raise ValueError("Both input arrays are empty.")
+
+ # Merge the arrays into a single sorted array.
+ merged = sorted(nums1 + nums2)
+ total = len(merged)
+
+ if total % 2 == 1: # If the total number of elements is odd
+ return float(merged[total // 2]) # then return the middle element
+
+ # If the total number of elements is even, calculate
+ # the average of the two middle elements as the median.
+ middle1 = merged[total // 2 - 1]
+ middle2 = merged[total // 2]
+ return (float(middle1) + float(middle2)) / 2.0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/monotonic_array.py b/data_structures/arrays/monotonic_array.py
new file mode 100644
index 000000000000..342d443a9cfc
--- /dev/null
+++ b/data_structures/arrays/monotonic_array.py
@@ -0,0 +1,37 @@
+# https://leetcode.com/problems/monotonic-array/
+def is_monotonic(nums: list[int]) -> bool:
+ """
+ Check if a list is monotonic.
+
+ >>> is_monotonic([1, 2, 2, 3])
+ True
+ >>> is_monotonic([6, 5, 4, 4])
+ True
+ >>> is_monotonic([1, 3, 2])
+ False
+ >>> is_monotonic([1,2,3,4,5,6,5])
+ False
+ >>> is_monotonic([-3,-2,-1])
+ True
+ >>> is_monotonic([-5,-6,-7])
+ True
+ >>> is_monotonic([0,0,0])
+ True
+ >>> is_monotonic([-100,0,100])
+ True
+ """
+ return all(nums[i] <= nums[i + 1] for i in range(len(nums) - 1)) or all(
+ nums[i] >= nums[i + 1] for i in range(len(nums) - 1)
+ )
+
+
+# Test the function with your examples
+if __name__ == "__main__":
+ # Test the function with your examples
+ print(is_monotonic([1, 2, 2, 3])) # Output: True
+ print(is_monotonic([6, 5, 4, 4])) # Output: True
+ print(is_monotonic([1, 3, 2])) # Output: False
+
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/pairs_with_given_sum.py b/data_structures/arrays/pairs_with_given_sum.py
new file mode 100644
index 000000000000..b27bd78e1e0f
--- /dev/null
+++ b/data_structures/arrays/pairs_with_given_sum.py
@@ -0,0 +1,29 @@
+#!/usr/bin/env python3
+
+"""
+Given an array of integers and an integer req_sum, find the number of pairs of array
+elements whose sum is equal to req_sum.
+
+https://practice.geeksforgeeks.org/problems/count-pairs-with-given-sum5022/0
+"""
+
+from itertools import combinations
+
+
+def pairs_with_sum(arr: list, req_sum: int) -> int:
+ """
+ Return the no. of pairs with sum "sum"
+ >>> pairs_with_sum([1, 5, 7, 1], 6)
+ 2
+ >>> pairs_with_sum([1, 1, 1, 1, 1, 1, 1, 1], 2)
+ 28
+ >>> pairs_with_sum([1, 7, 6, 2, 5, 4, 3, 1, 9, 8], 7)
+ 4
+ """
+ return len([1 for a, b in combinations(arr, 2) if a + b == req_sum])
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/data_structures/arrays/permutations.py b/data_structures/arrays/permutations.py
new file mode 100644
index 000000000000..4906dd5c2ae1
--- /dev/null
+++ b/data_structures/arrays/permutations.py
@@ -0,0 +1,48 @@
+def permute_recursive(nums: list[int]) -> list[list[int]]:
+ """
+ Return all permutations.
+
+ >>> permute_recursive([1, 2, 3])
+ [[3, 2, 1], [2, 3, 1], [1, 3, 2], [3, 1, 2], [2, 1, 3], [1, 2, 3]]
+ """
+ result: list[list[int]] = []
+ if len(nums) == 0:
+ return [[]]
+ for _ in range(len(nums)):
+ n = nums.pop(0)
+ permutations = permute_recursive(nums.copy())
+ for perm in permutations:
+ perm.append(n)
+ result.extend(permutations)
+ nums.append(n)
+ return result
+
+
+def permute_backtrack(nums: list[int]) -> list[list[int]]:
+ """
+ Return all permutations of the given list.
+
+ >>> permute_backtrack([1, 2, 3])
+ [[1, 2, 3], [1, 3, 2], [2, 1, 3], [2, 3, 1], [3, 2, 1], [3, 1, 2]]
+ """
+
+ def backtrack(start: int) -> None:
+ if start == len(nums) - 1:
+ output.append(nums[:])
+ else:
+ for i in range(start, len(nums)):
+ nums[start], nums[i] = nums[i], nums[start]
+ backtrack(start + 1)
+ nums[start], nums[i] = nums[i], nums[start] # backtrack
+
+ output: list[list[int]] = []
+ backtrack(0)
+ return output
+
+
+if __name__ == "__main__":
+ import doctest
+
+ result = permute_backtrack([1, 2, 3])
+ print(result)
+ doctest.testmod()
diff --git a/data_structures/arrays/prefix_sum.py b/data_structures/arrays/prefix_sum.py
new file mode 100644
index 000000000000..717b5f9d7e7e
--- /dev/null
+++ b/data_structures/arrays/prefix_sum.py
@@ -0,0 +1,96 @@
+"""
+Author : Alexander Pantyukhin
+Date : November 3, 2022
+
+Implement the class of prefix sum with useful functions based on it.
+
+"""
+
+
+class PrefixSum:
+ def __init__(self, array: list[int]) -> None:
+ len_array = len(array)
+ self.prefix_sum = [0] * len_array
+
+ if len_array > 0:
+ self.prefix_sum[0] = array[0]
+
+ for i in range(1, len_array):
+ self.prefix_sum[i] = self.prefix_sum[i - 1] + array[i]
+
+ def get_sum(self, start: int, end: int) -> int:
+ """
+ The function returns the sum of array from the start to the end indexes.
+ Runtime : O(1)
+ Space: O(1)
+
+ >>> PrefixSum([1,2,3]).get_sum(0, 2)
+ 6
+ >>> PrefixSum([1,2,3]).get_sum(1, 2)
+ 5
+ >>> PrefixSum([1,2,3]).get_sum(2, 2)
+ 3
+ >>> PrefixSum([]).get_sum(0, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: The array is empty.
+ >>> PrefixSum([1,2,3]).get_sum(-1, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid range specified.
+ >>> PrefixSum([1,2,3]).get_sum(2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid range specified.
+ >>> PrefixSum([1,2,3]).get_sum(2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid range specified.
+ """
+ if not self.prefix_sum:
+ raise ValueError("The array is empty.")
+
+ if start < 0 or end >= len(self.prefix_sum) or start > end:
+ raise ValueError("Invalid range specified.")
+
+ if start == 0:
+ return self.prefix_sum[end]
+
+ return self.prefix_sum[end] - self.prefix_sum[start - 1]
+
+ def contains_sum(self, target_sum: int) -> bool:
+ """
+ The function returns True if array contains the target_sum,
+ False otherwise.
+
+ Runtime : O(n)
+ Space: O(n)
+
+ >>> PrefixSum([1,2,3]).contains_sum(6)
+ True
+ >>> PrefixSum([1,2,3]).contains_sum(5)
+ True
+ >>> PrefixSum([1,2,3]).contains_sum(3)
+ True
+ >>> PrefixSum([1,2,3]).contains_sum(4)
+ False
+ >>> PrefixSum([1,2,3]).contains_sum(7)
+ False
+ >>> PrefixSum([1,-2,3]).contains_sum(2)
+ True
+ """
+
+ sums = {0}
+ for sum_item in self.prefix_sum:
+ if sum_item - target_sum in sums:
+ return True
+
+ sums.add(sum_item)
+
+ return False
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/product_sum.py b/data_structures/arrays/product_sum.py
new file mode 100644
index 000000000000..4fb906f369ab
--- /dev/null
+++ b/data_structures/arrays/product_sum.py
@@ -0,0 +1,98 @@
+"""
+Calculate the Product Sum from a Special Array.
+reference: https://dev.to/sfrasica/algorithms-product-sum-from-an-array-dc6
+
+Python doctests can be run with the following command:
+python -m doctest -v product_sum.py
+
+Calculate the product sum of a "special" array which can contain integers or nested
+arrays. The product sum is obtained by adding all elements and multiplying by their
+respective depths.
+
+For example, in the array [x, y], the product sum is (x + y). In the array [x, [y, z]],
+the product sum is x + 2 * (y + z). In the array [x, [y, [z]]],
+the product sum is x + 2 * (y + 3z).
+
+Example Input:
+[5, 2, [-7, 1], 3, [6, [-13, 8], 4]]
+Output: 12
+
+"""
+
+
+def product_sum(arr: list[int | list], depth: int) -> int:
+ """
+ Recursively calculates the product sum of an array.
+
+ The product sum of an array is defined as the sum of its elements multiplied by
+ their respective depths. If an element is a list, its product sum is calculated
+ recursively by multiplying the sum of its elements with its depth plus one.
+
+ Args:
+ arr: The array of integers and nested lists.
+ depth: The current depth level.
+
+ Returns:
+ int: The product sum of the array.
+
+ Examples:
+ >>> product_sum([1, 2, 3], 1)
+ 6
+ >>> product_sum([-1, 2, [-3, 4]], 2)
+ 8
+ >>> product_sum([1, 2, 3], -1)
+ -6
+ >>> product_sum([1, 2, 3], 0)
+ 0
+ >>> product_sum([1, 2, 3], 7)
+ 42
+ >>> product_sum((1, 2, 3), 7)
+ 42
+ >>> product_sum({1, 2, 3}, 7)
+ 42
+ >>> product_sum([1, -1], 1)
+ 0
+ >>> product_sum([1, -2], 1)
+ -1
+ >>> product_sum([-3.5, [1, [0.5]]], 1)
+ 1.5
+
+ """
+ total_sum = 0
+ for ele in arr:
+ total_sum += product_sum(ele, depth + 1) if isinstance(ele, list) else ele
+ return total_sum * depth
+
+
+def product_sum_array(array: list[int | list]) -> int:
+ """
+ Calculates the product sum of an array.
+
+ Args:
+ array (List[Union[int, List]]): The array of integers and nested lists.
+
+ Returns:
+ int: The product sum of the array.
+
+ Examples:
+ >>> product_sum_array([1, 2, 3])
+ 6
+ >>> product_sum_array([1, [2, 3]])
+ 11
+ >>> product_sum_array([1, [2, [3, 4]]])
+ 47
+ >>> product_sum_array([0])
+ 0
+ >>> product_sum_array([-3.5, [1, [0.5]]])
+ 1.5
+ >>> product_sum_array([1, -2])
+ -1
+
+ """
+ return product_sum(array, 1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/arrays/sparse_table.py b/data_structures/arrays/sparse_table.py
new file mode 100644
index 000000000000..4606fe908607
--- /dev/null
+++ b/data_structures/arrays/sparse_table.py
@@ -0,0 +1,95 @@
+"""
+Sparse table is a data structure that allows answering range queries on
+a static number list, i.e. the elements do not change throughout all the queries.
+
+The implementation below will solve the problem of Range Minimum Query:
+Finding the minimum value of a subset [L..R] of a static number list.
+
+Overall time complexity: O(nlogn)
+Overall space complexity: O(nlogn)
+
+Wikipedia link: https://en.wikipedia.org/wiki/Range_minimum_query
+"""
+
+from math import log2
+
+
+def build_sparse_table(number_list: list[int]) -> list[list[int]]:
+ """
+ Precompute range minimum queries with power of two length and store the precomputed
+ values in a table.
+
+ >>> build_sparse_table([8, 1, 0, 3, 4, 9, 3])
+ [[8, 1, 0, 3, 4, 9, 3], [1, 0, 0, 3, 4, 3, 0], [0, 0, 0, 3, 0, 0, 0]]
+ >>> build_sparse_table([3, 1, 9])
+ [[3, 1, 9], [1, 1, 0]]
+ >>> build_sparse_table([])
+ Traceback (most recent call last):
+ ...
+ ValueError: empty number list not allowed
+ """
+ if not number_list:
+ raise ValueError("empty number list not allowed")
+
+ length = len(number_list)
+ # Initialise sparse_table -- sparse_table[j][i] represents the minimum value of the
+ # subset of length (2 ** j) of number_list, starting from index i.
+
+ # smallest power of 2 subset length that fully covers number_list
+ row = int(log2(length)) + 1
+ sparse_table = [[0 for i in range(length)] for j in range(row)]
+
+ # minimum of subset of length 1 is that value itself
+ for i, value in enumerate(number_list):
+ sparse_table[0][i] = value
+ j = 1
+
+ # compute the minimum value for all intervals with size (2 ** j)
+ while (1 << j) <= length:
+ i = 0
+ # while subset starting from i still have at least (2 ** j) elements
+ while (i + (1 << j) - 1) < length:
+ # split range [i, i + 2 ** j] and find minimum of 2 halves
+ sparse_table[j][i] = min(
+ sparse_table[j - 1][i + (1 << (j - 1))], sparse_table[j - 1][i]
+ )
+ i += 1
+ j += 1
+ return sparse_table
+
+
+def query(sparse_table: list[list[int]], left_bound: int, right_bound: int) -> int:
+ """
+ >>> query(build_sparse_table([8, 1, 0, 3, 4, 9, 3]), 0, 4)
+ 0
+ >>> query(build_sparse_table([8, 1, 0, 3, 4, 9, 3]), 4, 6)
+ 3
+ >>> query(build_sparse_table([3, 1, 9]), 2, 2)
+ 9
+ >>> query(build_sparse_table([3, 1, 9]), 0, 1)
+ 1
+ >>> query(build_sparse_table([8, 1, 0, 3, 4, 9, 3]), 0, 11)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> query(build_sparse_table([]), 0, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: empty number list not allowed
+ """
+ if left_bound < 0 or right_bound >= len(sparse_table[0]):
+ raise IndexError("list index out of range")
+
+ # highest subset length of power of 2 that is within range [left_bound, right_bound]
+ j = int(log2(right_bound - left_bound + 1))
+
+ # minimum of 2 overlapping smaller subsets:
+ # [left_bound, left_bound + 2 ** j - 1] and [right_bound - 2 ** j + 1, right_bound]
+ return min(sparse_table[j][right_bound - (1 << j) + 1], sparse_table[j][left_bound])
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ print(f"{query(build_sparse_table([3, 1, 9]), 2, 2) = }")
diff --git a/data_structures/arrays/sudoku_solver.py b/data_structures/arrays/sudoku_solver.py
new file mode 100644
index 000000000000..4c722f12fd6e
--- /dev/null
+++ b/data_structures/arrays/sudoku_solver.py
@@ -0,0 +1,246 @@
+"""
+Please do not modify this file! It is published at https://norvig.com/sudoku.html with
+only minimal changes to work with modern versions of Python. If you have improvements,
+please make them in a separate file.
+"""
+
+import random
+import time
+
+
+def cross(items_a, items_b):
+ """
+ Cross product of elements in A and elements in B.
+ """
+ return [a + b for a in items_a for b in items_b]
+
+
+digits = "123456789"
+rows = "ABCDEFGHI"
+cols = digits
+squares = cross(rows, cols)
+unitlist = (
+ [cross(rows, c) for c in cols]
+ + [cross(r, cols) for r in rows]
+ + [cross(rs, cs) for rs in ("ABC", "DEF", "GHI") for cs in ("123", "456", "789")]
+)
+units = {s: [u for u in unitlist if s in u] for s in squares}
+peers = {s: {x for u in units[s] for x in u} - {s} for s in squares}
+
+
+def test():
+ """A set of unit tests."""
+ assert len(squares) == 81
+ assert len(unitlist) == 27
+ assert all(len(units[s]) == 3 for s in squares)
+ assert all(len(peers[s]) == 20 for s in squares)
+ assert units["C2"] == [
+ ["A2", "B2", "C2", "D2", "E2", "F2", "G2", "H2", "I2"],
+ ["C1", "C2", "C3", "C4", "C5", "C6", "C7", "C8", "C9"],
+ ["A1", "A2", "A3", "B1", "B2", "B3", "C1", "C2", "C3"],
+ ]
+ # fmt: off
+ assert peers["C2"] == {
+ "A2", "B2", "D2", "E2", "F2", "G2", "H2", "I2", "C1", "C3",
+ "C4", "C5", "C6", "C7", "C8", "C9", "A1", "A3", "B1", "B3"
+ }
+ # fmt: on
+ print("All tests pass.")
+
+
+def parse_grid(grid):
+ """
+ Convert grid to a dict of possible values, {square: digits}, or
+ return False if a contradiction is detected.
+ """
+ ## To start, every square can be any digit; then assign values from the grid.
+ values = dict.fromkeys(squares, digits)
+ for s, d in grid_values(grid).items():
+ if d in digits and not assign(values, s, d):
+ return False ## (Fail if we can't assign d to square s.)
+ return values
+
+
+def grid_values(grid):
+ """
+ Convert grid into a dict of {square: char} with '0' or '.' for empties.
+ """
+ chars = [c for c in grid if c in digits or c in "0."]
+ assert len(chars) == 81
+ return dict(zip(squares, chars))
+
+
+def assign(values, s, d):
+ """
+ Eliminate all the other values (except d) from values[s] and propagate.
+ Return values, except return False if a contradiction is detected.
+ """
+ other_values = values[s].replace(d, "")
+ if all(eliminate(values, s, d2) for d2 in other_values):
+ return values
+ else:
+ return False
+
+
+def eliminate(values, s, d):
+ """
+ Eliminate d from values[s]; propagate when values or places <= 2.
+ Return values, except return False if a contradiction is detected.
+ """
+ if d not in values[s]:
+ return values ## Already eliminated
+ values[s] = values[s].replace(d, "")
+ ## (1) If a square s is reduced to one value d2, then eliminate d2 from the peers.
+ if len(values[s]) == 0:
+ return False ## Contradiction: removed last value
+ elif len(values[s]) == 1:
+ d2 = values[s]
+ if not all(eliminate(values, s2, d2) for s2 in peers[s]):
+ return False
+ ## (2) If a unit u is reduced to only one place for a value d, then put it there.
+ for u in units[s]:
+ dplaces = [s for s in u if d in values[s]]
+ if len(dplaces) == 0:
+ return False ## Contradiction: no place for this value
+ # d can only be in one place in unit; assign it there
+ elif len(dplaces) == 1 and not assign(values, dplaces[0], d):
+ return False
+ return values
+
+
+def display(values):
+ """
+ Display these values as a 2-D grid.
+ """
+ width = 1 + max(len(values[s]) for s in squares)
+ line = "+".join(["-" * (width * 3)] * 3)
+ for r in rows:
+ print(
+ "".join(
+ values[r + c].center(width) + ("|" if c in "36" else "") for c in cols
+ )
+ )
+ if r in "CF":
+ print(line)
+ print()
+
+
+def solve(grid):
+ """
+ Solve the grid.
+ """
+ return search(parse_grid(grid))
+
+
+def some(seq):
+ """Return some element of seq that is true."""
+ for e in seq:
+ if e:
+ return e
+ return False
+
+
+def search(values):
+ """
+ Using depth-first search and propagation, try all possible values.
+ """
+ if values is False:
+ return False ## Failed earlier
+ if all(len(values[s]) == 1 for s in squares):
+ return values ## Solved!
+ ## Chose the unfilled square s with the fewest possibilities
+ n, s = min((len(values[s]), s) for s in squares if len(values[s]) > 1)
+ return some(search(assign(values.copy(), s, d)) for d in values[s])
+
+
+def solve_all(grids, name="", showif=0.0):
+ """
+ Attempt to solve a sequence of grids. Report results.
+ When showif is a number of seconds, display puzzles that take longer.
+ When showif is None, don't display any puzzles.
+ """
+
+ def time_solve(grid):
+ start = time.monotonic()
+ values = solve(grid)
+ t = time.monotonic() - start
+ ## Display puzzles that take long enough
+ if showif is not None and t > showif:
+ display(grid_values(grid))
+ if values:
+ display(values)
+ print(f"({t:.5f} seconds)\n")
+ return (t, solved(values))
+
+ times, results = zip(*[time_solve(grid) for grid in grids])
+ if (n := len(grids)) > 1:
+ print(
+ "Solved %d of %d %s puzzles (avg %.2f secs (%d Hz), max %.2f secs)." # noqa: UP031
+ % (sum(results), n, name, sum(times) / n, n / sum(times), max(times))
+ )
+
+
+def solved(values):
+ """
+ A puzzle is solved if each unit is a permutation of the digits 1 to 9.
+ """
+
+ def unitsolved(unit):
+ return {values[s] for s in unit} == set(digits)
+
+ return values is not False and all(unitsolved(unit) for unit in unitlist)
+
+
+def from_file(filename, sep="\n"):
+ "Parse a file into a list of strings, separated by sep."
+ with open(filename) as file:
+ return file.read().strip().split(sep)
+
+
+def random_puzzle(assignments=17):
+ """
+ Make a random puzzle with N or more assignments. Restart on contradictions.
+ Note the resulting puzzle is not guaranteed to be solvable, but empirically
+ about 99.8% of them are solvable. Some have multiple solutions.
+ """
+ values = dict.fromkeys(squares, digits)
+ for s in shuffled(squares):
+ if not assign(values, s, random.choice(values[s])):
+ break
+ ds = [values[s] for s in squares if len(values[s]) == 1]
+ if len(ds) >= assignments and len(set(ds)) >= 8:
+ return "".join(values[s] if len(values[s]) == 1 else "." for s in squares)
+ return random_puzzle(assignments) ## Give up and make a new puzzle
+
+
+def shuffled(seq):
+ """
+ Return a randomly shuffled copy of the input sequence.
+ """
+ seq = list(seq)
+ random.shuffle(seq)
+ return seq
+
+
+grid1 = (
+ "003020600900305001001806400008102900700000008006708200002609500800203009005010300"
+)
+grid2 = (
+ "4.....8.5.3..........7......2.....6.....8.4......1.......6.3.7.5..2.....1.4......"
+)
+hard1 = (
+ ".....6....59.....82....8....45........3........6..3.54...325..6.................."
+)
+
+if __name__ == "__main__":
+ test()
+ # solve_all(from_file("easy50.txt", '========'), "easy", None)
+ # solve_all(from_file("top95.txt"), "hard", None)
+ # solve_all(from_file("hardest.txt"), "hardest", None)
+ solve_all([random_puzzle() for _ in range(99)], "random", 100.0)
+ for puzzle in (grid1, grid2): # , hard1): # Takes 22 sec to solve on my M1 Mac.
+ display(parse_grid(puzzle))
+ start = time.monotonic()
+ solve(puzzle)
+ t = time.monotonic() - start
+ print(f"Solved: {t:.5f} sec")
diff --git a/data_structures/binary_tree/README.md b/data_structures/binary_tree/README.md
new file mode 100644
index 000000000000..ebe727b6589d
--- /dev/null
+++ b/data_structures/binary_tree/README.md
@@ -0,0 +1,111 @@
+# Binary Tree Traversal
+
+## Overview
+
+The combination of binary trees being data structures and traversal being an algorithm relates to classic problems, either directly or indirectly.
+
+> If you can grasp the traversal of binary trees, the traversal of other complicated trees will be easy for you.
+
+The following are some common ways to traverse trees.
+
+- Depth First Traversals (DFS): In-order, Pre-order, Post-order
+
+- Level Order Traversal or Breadth First or Traversal (BFS)
+
+There are applications for both DFS and BFS.
+
+Stack can be used to simplify the process of DFS traversal. Besides, since tree is a recursive data structure, recursion and stack are two key points for DFS.
+
+Graph for DFS:
+
+
+
+The key point of BFS is how to determine whether the traversal of each level has been completed. The answer is to use a variable as a flag to represent the end of the traversal of current level.
+
+## Pre-order Traversal
+
+The traversal order of pre-order traversal is `root-left-right`.
+
+Algorithm Pre-order
+
+1. Visit the root node and push it into a stack.
+
+2. Pop a node from the stack, and push its right and left child node into the stack respectively.
+
+3. Repeat step 2.
+
+Conclusion: This problem involves the classic recursive data structure (i.e. a binary tree), and the algorithm above demonstrates how a simplified solution can be reached by using a stack.
+
+If you look at the bigger picture, you'll find that the process of traversal is as followed. `Visit the left subtrees respectively from top to bottom, and visit the right subtrees respectively from bottom to top`. If we are to implement it from this perspective, things will be somewhat different. For the `top to bottom` part we can simply use recursion, and for the `bottom to top` part we can turn to stack.
+
+## In-order Traversal
+
+The traversal order of in-order traversal is `left-root-right`.
+
+So the root node is not printed first. Things are getting a bit complicated here.
+
+Algorithm In-order
+
+1. Visit the root and push it into a stack.
+
+2. If there is a left child node, push it into the stack. Repeat this process until a leaf node reached.
+
+ > At this point the root node and all the left nodes are in the stack.
+
+3. Start popping nodes from the stack. If a node has a right child node, push the child node into the stack. Repeat step 2.
+
+It's worth pointing out that the in-order traversal of a binary search tree (BST) is a sorted array, which is helpful for coming up simplified solutions for some problems.
+
+## Post-order Traversal
+
+The traversal order of post-order traversal is `left-right-root`.
+
+This one is a bit of a challenge. It deserves the `hard` tag of LeetCode.
+
+In this case, the root node is printed not as the first but the last one. A cunning way to do it is to:
+
+Record whether the current node has been visited. If 1) it's a leaf node or 2) both its left and right subtrees have been traversed, then it can be popped from the stack.
+
+As for `1) it's a leaf node`, you can easily tell whether a node is a leaf if both its left and right are `null`.
+
+As for `2) both its left and right subtrees have been traversed`, we only need a variable to record whether a node has been visited or not. In the worst case, we need to record the status for every single node and the space complexity is `O(n)`. But if you come to think about it, as we are using a stack and start printing the result from the leaf nodes, it makes sense that we only record the status for the current node popping from the stack, reducing the space complexity to `O(1)`.
+
+## Level Order Traversal
+
+The key point of level order traversal is how do we know whether the traversal of each level is done. The answer is that we use a variable as a flag representing the end of the traversal of the current level.
+
+
+
+Algorithm Level-order
+
+1. Visit the root node, put it in a FIFO queue, put in the queue a special flag (we are using `null` here).
+
+2. Dequeue a node.
+
+3. If the node equals `null`, it means that all nodes of the current level have been visited. If the queue is empty, we do nothing. Or else we put in another `null`.
+
+4. If the node is not `null`, meaning the traversal of current level has not finished yet, we enqueue its left subtree and right subtree respectively.
+
+## Bi-color marking
+
+We know that there is a tri-color marking in garbage collection algorithm, which works as described below.
+
+- The white color represents "not visited".
+
+- The gray color represents "not all child nodes visited".
+
+- The black color represents "all child nodes visited".
+
+Enlightened by tri-color marking, a bi-color marking method can be invented to solve all three traversal problems with one solution.
+
+The core idea is as follow.
+
+- Use a color to mark whether a node has been visited or not. Nodes yet to be visited are marked as white and visited nodes are marked as gray.
+
+- If we are visiting a white node, turn it into gray, and push its right child node, itself, and it's left child node into the stack respectively.
+
+- If we are visiting a gray node, print it.
+
+Implementation of pre-order and post-order traversal algorithms can be easily done by changing the order of pushing the child nodes into the stack.
+
+Reference: [LeetCode](https://github.com/azl397985856/leetcode/blob/master/thinkings/binary-tree-traversal.en.md)
diff --git a/data_structures/binary_tree/avl_tree.py b/data_structures/binary_tree/avl_tree.py
index 3362610b9303..8558305eefe4 100644
--- a/data_structures/binary_tree/avl_tree.py
+++ b/data_structures/binary_tree/avl_tree.py
@@ -6,86 +6,85 @@
python avl_tree.py
"""
+from __future__ import annotations
+
import math
import random
+from typing import Any
-class my_queue:
- def __init__(self):
- self.data = []
- self.head = 0
- self.tail = 0
+class MyQueue:
+ def __init__(self) -> None:
+ self.data: list[Any] = []
+ self.head: int = 0
+ self.tail: int = 0
- def is_empty(self):
+ def is_empty(self) -> bool:
return self.head == self.tail
- def push(self, data):
+ def push(self, data: Any) -> None:
self.data.append(data)
self.tail = self.tail + 1
- def pop(self):
+ def pop(self) -> Any:
ret = self.data[self.head]
self.head = self.head + 1
return ret
- def count(self):
+ def count(self) -> int:
return self.tail - self.head
- def print(self):
+ def print_queue(self) -> None:
print(self.data)
print("**************")
print(self.data[self.head : self.tail])
-class my_node:
- def __init__(self, data):
+class MyNode:
+ def __init__(self, data: Any) -> None:
self.data = data
- self.left = None
- self.right = None
- self.height = 1
+ self.left: MyNode | None = None
+ self.right: MyNode | None = None
+ self.height: int = 1
- def get_data(self):
+ def get_data(self) -> Any:
return self.data
- def get_left(self):
+ def get_left(self) -> MyNode | None:
return self.left
- def get_right(self):
+ def get_right(self) -> MyNode | None:
return self.right
- def get_height(self):
+ def get_height(self) -> int:
return self.height
- def set_data(self, data):
+ def set_data(self, data: Any) -> None:
self.data = data
- return
- def set_left(self, node):
+ def set_left(self, node: MyNode | None) -> None:
self.left = node
- return
- def set_right(self, node):
+ def set_right(self, node: MyNode | None) -> None:
self.right = node
- return
- def set_height(self, height):
+ def set_height(self, height: int) -> None:
self.height = height
- return
-def get_height(node):
+def get_height(node: MyNode | None) -> int:
if node is None:
return 0
return node.get_height()
-def my_max(a, b):
+def my_max(a: int, b: int) -> int:
if a > b:
return a
return b
-def right_rotation(node):
+def right_rotation(node: MyNode) -> MyNode:
r"""
A B
/ \ / \
@@ -98,6 +97,7 @@ def right_rotation(node):
"""
print("left rotation node:", node.get_data())
ret = node.get_left()
+ assert ret is not None
node.set_left(ret.get_right())
ret.set_right(node)
h1 = my_max(get_height(node.get_right()), get_height(node.get_left())) + 1
@@ -107,12 +107,13 @@ def right_rotation(node):
return ret
-def left_rotation(node):
+def left_rotation(node: MyNode) -> MyNode:
"""
a mirror symmetry rotation of the left_rotation
"""
print("right rotation node:", node.get_data())
ret = node.get_right()
+ assert ret is not None
node.set_right(ret.get_left())
ret.set_left(node)
h1 = my_max(get_height(node.get_right()), get_height(node.get_left())) + 1
@@ -122,7 +123,7 @@ def left_rotation(node):
return ret
-def lr_rotation(node):
+def lr_rotation(node: MyNode) -> MyNode:
r"""
A A Br
/ \ / \ / \
@@ -133,25 +134,31 @@ def lr_rotation(node):
UB Bl
RR = right_rotation LR = left_rotation
"""
- node.set_left(left_rotation(node.get_left()))
+ left_child = node.get_left()
+ assert left_child is not None
+ node.set_left(left_rotation(left_child))
return right_rotation(node)
-def rl_rotation(node):
- node.set_right(right_rotation(node.get_right()))
+def rl_rotation(node: MyNode) -> MyNode:
+ right_child = node.get_right()
+ assert right_child is not None
+ node.set_right(right_rotation(right_child))
return left_rotation(node)
-def insert_node(node, data):
+def insert_node(node: MyNode | None, data: Any) -> MyNode | None:
if node is None:
- return my_node(data)
+ return MyNode(data)
if data < node.get_data():
node.set_left(insert_node(node.get_left(), data))
if (
get_height(node.get_left()) - get_height(node.get_right()) == 2
): # an unbalance detected
+ left_child = node.get_left()
+ assert left_child is not None
if (
- data < node.get_left().get_data()
+ data < left_child.get_data()
): # new node is the left child of the left child
node = right_rotation(node)
else:
@@ -159,7 +166,9 @@ def insert_node(node, data):
else:
node.set_right(insert_node(node.get_right(), data))
if get_height(node.get_right()) - get_height(node.get_left()) == 2:
- if data < node.get_right().get_data():
+ right_child = node.get_right()
+ assert right_child is not None
+ if data < right_child.get_data():
node = rl_rotation(node)
else:
node = left_rotation(node)
@@ -168,52 +177,63 @@ def insert_node(node, data):
return node
-def get_rightMost(root):
- while root.get_right() is not None:
- root = root.get_right()
+def get_right_most(root: MyNode) -> Any:
+ while True:
+ right_child = root.get_right()
+ if right_child is None:
+ break
+ root = right_child
return root.get_data()
-def get_leftMost(root):
- while root.get_left() is not None:
- root = root.get_left()
+def get_left_most(root: MyNode) -> Any:
+ while True:
+ left_child = root.get_left()
+ if left_child is None:
+ break
+ root = left_child
return root.get_data()
-def del_node(root, data):
+def del_node(root: MyNode, data: Any) -> MyNode | None:
+ left_child = root.get_left()
+ right_child = root.get_right()
if root.get_data() == data:
- if root.get_left() is not None and root.get_right() is not None:
- temp_data = get_leftMost(root.get_right())
+ if left_child is not None and right_child is not None:
+ temp_data = get_left_most(right_child)
root.set_data(temp_data)
- root.set_right(del_node(root.get_right(), temp_data))
- elif root.get_left() is not None:
- root = root.get_left()
+ root.set_right(del_node(right_child, temp_data))
+ elif left_child is not None:
+ root = left_child
+ elif right_child is not None:
+ root = right_child
else:
- root = root.get_right()
+ return None
elif root.get_data() > data:
- if root.get_left() is None:
+ if left_child is None:
print("No such data")
return root
else:
- root.set_left(del_node(root.get_left(), data))
- elif root.get_data() < data:
- if root.get_right() is None:
- return root
- else:
- root.set_right(del_node(root.get_right(), data))
- if root is None:
+ root.set_left(del_node(left_child, data))
+ # root.get_data() < data
+ elif right_child is None:
return root
- if get_height(root.get_right()) - get_height(root.get_left()) == 2:
- if get_height(root.get_right().get_right()) > get_height(
- root.get_right().get_left()
- ):
+ else:
+ root.set_right(del_node(right_child, data))
+
+ # Re-fetch left_child and right_child references
+ left_child = root.get_left()
+ right_child = root.get_right()
+
+ if get_height(right_child) - get_height(left_child) == 2:
+ assert right_child is not None
+ if get_height(right_child.get_right()) > get_height(right_child.get_left()):
root = left_rotation(root)
else:
root = rl_rotation(root)
- elif get_height(root.get_right()) - get_height(root.get_left()) == -2:
- if get_height(root.get_left().get_left()) > get_height(
- root.get_left().get_right()
- ):
+ elif get_height(right_child) - get_height(left_child) == -2:
+ assert left_child is not None
+ if get_height(left_child.get_left()) > get_height(left_child.get_right()):
root = right_rotation(root)
else:
root = lr_rotation(root)
@@ -256,27 +276,28 @@ class AVLtree:
*************************************
"""
- def __init__(self):
- self.root = None
+ def __init__(self) -> None:
+ self.root: MyNode | None = None
- def get_height(self):
- # print("yyy")
+ def get_height(self) -> int:
return get_height(self.root)
- def insert(self, data):
+ def insert(self, data: Any) -> None:
print("insert:" + str(data))
self.root = insert_node(self.root, data)
- def del_node(self, data):
+ def del_node(self, data: Any) -> None:
print("delete:" + str(data))
if self.root is None:
print("Tree is empty!")
return
self.root = del_node(self.root, data)
- def __str__(self): # a level traversale, gives a more intuitive look on the tree
+ def __str__(
+ self,
+ ) -> str: # a level traversale, gives a more intuitive look on the tree
output = ""
- q = my_queue()
+ q = MyQueue()
q.push(self.root)
layer = self.get_height()
if layer == 0:
@@ -308,7 +329,7 @@ def __str__(self): # a level traversale, gives a more intuitive look on the tre
return output
-def _test():
+def _test() -> None:
import doctest
doctest.testmod()
diff --git a/data_structures/binary_tree/basic_binary_tree.py b/data_structures/binary_tree/basic_binary_tree.py
index 575b157ee78a..9d4c1bdbb57a 100644
--- a/data_structures/binary_tree/basic_binary_tree.py
+++ b/data_structures/binary_tree/basic_binary_tree.py
@@ -1,101 +1,110 @@
-from typing import Optional
+from __future__ import annotations
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+@dataclass
class Node:
- """
- A Node has data variable and pointers to Nodes to its left and right.
- """
-
- def __init__(self, data: int) -> None:
- self.data = data
- self.left: Optional[Node] = None
- self.right: Optional[Node] = None
-
-
-def display(tree: Optional[Node]) -> None: # In Order traversal of the tree
- """
- >>> root = Node(1)
- >>> root.left = Node(0)
- >>> root.right = Node(2)
- >>> display(root)
- 0
- 1
- 2
- >>> display(root.right)
- 2
- """
- if tree:
- display(tree.left)
- print(tree.data)
- display(tree.right)
-
-
-def depth_of_tree(tree: Optional[Node]) -> int:
- """
- Recursive function that returns the depth of a binary tree.
-
- >>> root = Node(0)
- >>> depth_of_tree(root)
- 1
- >>> root.left = Node(0)
- >>> depth_of_tree(root)
- 2
- >>> root.right = Node(0)
- >>> depth_of_tree(root)
- 2
- >>> root.left.right = Node(0)
- >>> depth_of_tree(root)
- 3
- >>> depth_of_tree(root.left)
- 2
- """
- return 1 + max(depth_of_tree(tree.left), depth_of_tree(tree.right)) if tree else 0
-
-
-def is_full_binary_tree(tree: Node) -> bool:
- """
- Returns True if this is a full binary tree
-
- >>> root = Node(0)
- >>> is_full_binary_tree(root)
- True
- >>> root.left = Node(0)
- >>> is_full_binary_tree(root)
- False
- >>> root.right = Node(0)
- >>> is_full_binary_tree(root)
- True
- >>> root.left.left = Node(0)
- >>> is_full_binary_tree(root)
- False
- >>> root.right.right = Node(0)
- >>> is_full_binary_tree(root)
- False
- """
- if not tree:
- return True
- if tree.left and tree.right:
- return is_full_binary_tree(tree.left) and is_full_binary_tree(tree.right)
- else:
- return not tree.left and not tree.right
-
-
-def main() -> None: # Main function for testing.
- tree = Node(1)
- tree.left = Node(2)
- tree.right = Node(3)
- tree.left.left = Node(4)
- tree.left.right = Node(5)
- tree.left.right.left = Node(6)
- tree.right.left = Node(7)
- tree.right.left.left = Node(8)
- tree.right.left.left.right = Node(9)
-
- print(is_full_binary_tree(tree))
- print(depth_of_tree(tree))
- print("Tree is: ")
- display(tree)
+ data: int
+ left: Node | None = None
+ right: Node | None = None
+
+ def __iter__(self) -> Iterator[int]:
+ if self.left:
+ yield from self.left
+ yield self.data
+ if self.right:
+ yield from self.right
+
+ def __len__(self) -> int:
+ return sum(1 for _ in self)
+
+ def is_full(self) -> bool:
+ if not self or (not self.left and not self.right):
+ return True
+ if self.left and self.right:
+ return self.left.is_full() and self.right.is_full()
+ return False
+
+
+@dataclass
+class BinaryTree:
+ root: Node
+
+ def __iter__(self) -> Iterator[int]:
+ return iter(self.root)
+
+ def __len__(self) -> int:
+ return len(self.root)
+
+ @classmethod
+ def small_tree(cls) -> BinaryTree:
+ """
+ Return a small binary tree with 3 nodes.
+ >>> binary_tree = BinaryTree.small_tree()
+ >>> len(binary_tree)
+ 3
+ >>> list(binary_tree)
+ [1, 2, 3]
+ """
+ binary_tree = BinaryTree(Node(2))
+ binary_tree.root.left = Node(1)
+ binary_tree.root.right = Node(3)
+ return binary_tree
+
+ @classmethod
+ def medium_tree(cls) -> BinaryTree:
+ """
+ Return a medium binary tree with 3 nodes.
+ >>> binary_tree = BinaryTree.medium_tree()
+ >>> len(binary_tree)
+ 7
+ >>> list(binary_tree)
+ [1, 2, 3, 4, 5, 6, 7]
+ """
+ binary_tree = BinaryTree(Node(4))
+ binary_tree.root.left = two = Node(2)
+ two.left = Node(1)
+ two.right = Node(3)
+ binary_tree.root.right = five = Node(5)
+ five.right = six = Node(6)
+ six.right = Node(7)
+ return binary_tree
+
+ def depth(self) -> int:
+ """
+ Returns the depth of the tree
+
+ >>> BinaryTree(Node(1)).depth()
+ 1
+ >>> BinaryTree.small_tree().depth()
+ 2
+ >>> BinaryTree.medium_tree().depth()
+ 4
+ """
+ return self._depth(self.root)
+
+ def _depth(self, node: Node | None) -> int:
+ if not node:
+ return 0
+ return 1 + max(self._depth(node.left), self._depth(node.right))
+
+ def is_full(self) -> bool:
+ """
+ Returns True if the tree is full
+
+ >>> BinaryTree(Node(1)).is_full()
+ True
+ >>> BinaryTree.small_tree().is_full()
+ True
+ >>> BinaryTree.medium_tree().is_full()
+ False
+ """
+ return self.root.is_full()
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/binary_search_tree.py b/data_structures/binary_tree/binary_search_tree.py
index 45c3933fe899..3f214d0113a4 100644
--- a/data_structures/binary_tree/binary_search_tree.py
+++ b/data_structures/binary_tree/binary_search_tree.py
@@ -1,59 +1,183 @@
-"""
+r"""
A binary search Tree
+
+Example
+ 8
+ / \
+ 3 10
+ / \ \
+ 1 6 14
+ / \ /
+ 4 7 13
+
+>>> t = BinarySearchTree().insert(8, 3, 6, 1, 10, 14, 13, 4, 7)
+>>> print(" ".join(repr(i.value) for i in t.traversal_tree()))
+8 3 1 6 4 7 10 14 13
+
+>>> tuple(i.value for i in t.traversal_tree(inorder))
+(1, 3, 4, 6, 7, 8, 10, 13, 14)
+>>> tuple(t)
+(1, 3, 4, 6, 7, 8, 10, 13, 14)
+>>> t.find_kth_smallest(3, t.root)
+4
+>>> tuple(t)[3-1]
+4
+
+>>> print(" ".join(repr(i.value) for i in t.traversal_tree(postorder)))
+1 4 7 6 3 13 14 10 8
+>>> t.remove(20)
+Traceback (most recent call last):
+ ...
+ValueError: Value 20 not found
+>>> BinarySearchTree().search(6)
+Traceback (most recent call last):
+ ...
+IndexError: Warning: Tree is empty! please use another.
+
+Other example:
+
+>>> testlist = (8, 3, 6, 1, 10, 14, 13, 4, 7)
+>>> t = BinarySearchTree()
+>>> for i in testlist:
+... t.insert(i) # doctest: +ELLIPSIS
+BinarySearchTree(root=8)
+BinarySearchTree(root={'8': (3, None)})
+BinarySearchTree(root={'8': ({'3': (None, 6)}, None)})
+BinarySearchTree(root={'8': ({'3': (1, 6)}, None)})
+BinarySearchTree(root={'8': ({'3': (1, 6)}, 10)})
+BinarySearchTree(root={'8': ({'3': (1, 6)}, {'10': (None, 14)})})
+BinarySearchTree(root={'8': ({'3': (1, 6)}, {'10': (None, {'14': (13, None)})})})
+BinarySearchTree(root={'8': ({'3': (1, {'6': (4, None)})}, {'10': (None, {'14': ...
+BinarySearchTree(root={'8': ({'3': (1, {'6': (4, 7)})}, {'10': (None, {'14': (13, ...
+
+Prints all the elements of the list in order traversal
+>>> print(t)
+{'8': ({'3': (1, {'6': (4, 7)})}, {'10': (None, {'14': (13, None)})})}
+
+Test existence
+>>> t.search(6) is not None
+True
+>>> 6 in t
+True
+>>> t.search(-1) is not None
+False
+>>> -1 in t
+False
+
+>>> t.search(6).is_right
+True
+>>> t.search(1).is_right
+False
+
+>>> t.get_max().value
+14
+>>> max(t)
+14
+>>> t.get_min().value
+1
+>>> min(t)
+1
+>>> t.empty()
+False
+>>> not t
+False
+>>> for i in testlist:
+... t.remove(i)
+>>> t.empty()
+True
+>>> not t
+True
"""
+from __future__ import annotations
+
+from collections.abc import Iterable, Iterator
+from dataclasses import dataclass
+from typing import Any, Self
+
+@dataclass
class Node:
- def __init__(self, value, parent):
- self.value = value
- self.parent = parent # Added in order to delete a node easier
- self.left = None
- self.right = None
+ value: int
+ left: Node | None = None
+ right: Node | None = None
+ parent: Node | None = None # Added in order to delete a node easier
- def __repr__(self):
+ def __iter__(self) -> Iterator[int]:
+ """
+ >>> list(Node(0))
+ [0]
+ >>> list(Node(0, Node(-1), Node(1), None))
+ [-1, 0, 1]
+ """
+ yield from self.left or []
+ yield self.value
+ yield from self.right or []
+
+ def __repr__(self) -> str:
from pprint import pformat
if self.left is None and self.right is None:
return str(self.value)
- return pformat({"%s" % (self.value): (self.left, self.right)}, indent=1)
+ return pformat({f"{self.value}": (self.left, self.right)}, indent=1)
+
+ @property
+ def is_right(self) -> bool:
+ return bool(self.parent and self is self.parent.right)
+@dataclass
class BinarySearchTree:
- def __init__(self, root=None):
- self.root = root
+ root: Node | None = None
- def __str__(self):
+ def __bool__(self) -> bool:
+ return bool(self.root)
+
+ def __iter__(self) -> Iterator[int]:
+ yield from self.root or []
+
+ def __str__(self) -> str:
"""
Return a string of all the Nodes using in order traversal
"""
return str(self.root)
- def __reassign_nodes(self, node, new_children):
+ def __reassign_nodes(self, node: Node, new_children: Node | None) -> None:
if new_children is not None: # reset its kids
new_children.parent = node.parent
if node.parent is not None: # reset its parent
- if self.is_right(node): # If it is the right children
+ if node.is_right: # If it is the right child
node.parent.right = new_children
else:
node.parent.left = new_children
else:
self.root = new_children
- def is_right(self, node):
- return node == node.parent.right
-
- def empty(self):
- return self.root is None
+ def empty(self) -> bool:
+ """
+ Returns True if the tree does not have any element(s).
+ False if the tree has element(s).
+
+ >>> BinarySearchTree().empty()
+ True
+ >>> BinarySearchTree().insert(1).empty()
+ False
+ >>> BinarySearchTree().insert(8, 3, 6, 1, 10, 14, 13, 4, 7).empty()
+ False
+ """
+ return not self.root
- def __insert(self, value):
+ def __insert(self, value) -> None:
"""
Insert a new node in Binary Search Tree with value label
"""
- new_node = Node(value, None) # create a new Node
+ new_node = Node(value) # create a new Node
if self.empty(): # if Tree is empty
self.root = new_node # set its root
else: # Tree is not empty
parent_node = self.root # from root
+ if parent_node is None:
+ return
while True: # While we don't get to a leaf
if value < parent_node.value: # We go left
if parent_node.left is None:
@@ -61,20 +185,43 @@ def __insert(self, value):
break
else:
parent_node = parent_node.left
+ elif parent_node.right is None:
+ parent_node.right = new_node
+ break
else:
- if parent_node.right is None:
- parent_node.right = new_node
- break
- else:
- parent_node = parent_node.right
+ parent_node = parent_node.right
new_node.parent = parent_node
- def insert(self, *values):
+ def insert(self, *values) -> Self:
for value in values:
self.__insert(value)
return self
- def search(self, value):
+ def search(self, value) -> Node | None:
+ """
+ >>> tree = BinarySearchTree().insert(10, 20, 30, 40, 50)
+ >>> tree.search(10)
+ {'10': (None, {'20': (None, {'30': (None, {'40': (None, 50)})})})}
+ >>> tree.search(20)
+ {'20': (None, {'30': (None, {'40': (None, 50)})})}
+ >>> tree.search(30)
+ {'30': (None, {'40': (None, 50)})}
+ >>> tree.search(40)
+ {'40': (None, 50)}
+ >>> tree.search(50)
+ 50
+ >>> tree.search(5) is None # element not present
+ True
+ >>> tree.search(0) is None # element not present
+ True
+ >>> tree.search(-5) is None # element not present
+ True
+ >>> BinarySearchTree().search(10)
+ Traceback (most recent call last):
+ ...
+ IndexError: Warning: Tree is empty! please use another.
+ """
+
if self.empty():
raise IndexError("Warning: Tree is empty! please use another.")
else:
@@ -84,54 +231,81 @@ def search(self, value):
node = node.left if value < node.value else node.right
return node
- def get_max(self, node=None):
+ def get_max(self, node: Node | None = None) -> Node | None:
"""
We go deep on the right branch
+
+ >>> BinarySearchTree().insert(10, 20, 30, 40, 50).get_max()
+ 50
+ >>> BinarySearchTree().insert(-5, -1, 0.1, -0.3, -4.5).get_max()
+ {'0.1': (-0.3, None)}
+ >>> BinarySearchTree().insert(1, 78.3, 30, 74.0, 1).get_max()
+ {'78.3': ({'30': (1, 74.0)}, None)}
+ >>> BinarySearchTree().insert(1, 783, 30, 740, 1).get_max()
+ {'783': ({'30': (1, 740)}, None)}
"""
if node is None:
+ if self.root is None:
+ return None
node = self.root
+
if not self.empty():
while node.right is not None:
node = node.right
return node
- def get_min(self, node=None):
+ def get_min(self, node: Node | None = None) -> Node | None:
"""
We go deep on the left branch
+
+ >>> BinarySearchTree().insert(10, 20, 30, 40, 50).get_min()
+ {'10': (None, {'20': (None, {'30': (None, {'40': (None, 50)})})})}
+ >>> BinarySearchTree().insert(-5, -1, 0, -0.3, -4.5).get_min()
+ {'-5': (None, {'-1': (-4.5, {'0': (-0.3, None)})})}
+ >>> BinarySearchTree().insert(1, 78.3, 30, 74.0, 1).get_min()
+ {'1': (None, {'78.3': ({'30': (1, 74.0)}, None)})}
+ >>> BinarySearchTree().insert(1, 783, 30, 740, 1).get_min()
+ {'1': (None, {'783': ({'30': (1, 740)}, None)})}
"""
if node is None:
node = self.root
+ if self.root is None:
+ return None
if not self.empty():
node = self.root
while node.left is not None:
node = node.left
return node
- def remove(self, value):
- node = self.search(value) # Look for the node with that label
- if node is not None:
- if node.left is None and node.right is None: # If it has no children
- self.__reassign_nodes(node, None)
- elif node.left is None: # Has only right children
- self.__reassign_nodes(node, node.right)
- elif node.right is None: # Has only left children
- self.__reassign_nodes(node, node.left)
- else:
- tmp_node = self.get_max(
- node.left
- ) # Gets the max value of the left branch
- self.remove(tmp_node.value)
- node.value = (
- tmp_node.value
- ) # Assigns the value to the node to delete and keep tree structure
-
- def preorder_traverse(self, node):
+ def remove(self, value: int) -> None:
+ # Look for the node with that label
+ node = self.search(value)
+ if node is None:
+ msg = f"Value {value} not found"
+ raise ValueError(msg)
+
+ if node.left is None and node.right is None: # If it has no children
+ self.__reassign_nodes(node, None)
+ elif node.left is None: # Has only right children
+ self.__reassign_nodes(node, node.right)
+ elif node.right is None: # Has only left children
+ self.__reassign_nodes(node, node.left)
+ else:
+ predecessor = self.get_max(
+ node.left
+ ) # Gets the max value of the left branch
+ self.remove(predecessor.value) # type: ignore[union-attr]
+ node.value = (
+ predecessor.value # type: ignore[union-attr]
+ ) # Assigns the value to the node to delete and keep tree structure
+
+ def preorder_traverse(self, node: Node | None) -> Iterable:
if node is not None:
yield node # Preorder Traversal
yield from self.preorder_traverse(node.left)
yield from self.preorder_traverse(node.right)
- def traversal_tree(self, traversal_function=None):
+ def traversal_tree(self, traversal_function=None) -> Any:
"""
This function traversal the tree.
You can pass a function to traversal the tree as needed by client code
@@ -141,7 +315,7 @@ def traversal_tree(self, traversal_function=None):
else:
return traversal_function(self.root)
- def inorder(self, arr: list, node: Node):
+ def inorder(self, arr: list, node: Node | None) -> None:
"""Perform an inorder traversal and append values of the nodes to
a list named arr"""
if node:
@@ -150,72 +324,33 @@ def inorder(self, arr: list, node: Node):
self.inorder(arr, node.right)
def find_kth_smallest(self, k: int, node: Node) -> int:
- """Return the kth smallest element in a binary search tree """
- arr = []
+ """Return the kth smallest element in a binary search tree"""
+ arr: list[int] = []
self.inorder(arr, node) # append all values to list using inorder traversal
return arr[k - 1]
-def postorder(curr_node):
+def inorder(curr_node: Node | None) -> list[Node]:
"""
- postOrder (left, right, self)
+ inorder (left, self, right)
"""
- node_list = list()
+ node_list = []
if curr_node is not None:
- node_list = postorder(curr_node.left) + postorder(curr_node.right) + [curr_node]
+ node_list = [*inorder(curr_node.left), curr_node, *inorder(curr_node.right)]
return node_list
-def binary_search_tree():
- r"""
- Example
- 8
- / \
- 3 10
- / \ \
- 1 6 14
- / \ /
- 4 7 13
-
- >>> t = BinarySearchTree().insert(8, 3, 6, 1, 10, 14, 13, 4, 7)
- >>> print(" ".join(repr(i.value) for i in t.traversal_tree()))
- 8 3 1 6 4 7 10 14 13
- >>> print(" ".join(repr(i.value) for i in t.traversal_tree(postorder)))
- 1 4 7 6 3 13 14 10 8
- >>> BinarySearchTree().search(6)
- Traceback (most recent call last):
- ...
- IndexError: Warning: Tree is empty! please use another.
+def postorder(curr_node: Node | None) -> list[Node]:
"""
- testlist = (8, 3, 6, 1, 10, 14, 13, 4, 7)
- t = BinarySearchTree()
- for i in testlist:
- t.insert(i)
-
- # Prints all the elements of the list in order traversal
- print(t)
-
- if t.search(6) is not None:
- print("The value 6 exists")
- else:
- print("The value 6 doesn't exist")
-
- if t.search(-1) is not None:
- print("The value -1 exists")
- else:
- print("The value -1 doesn't exist")
-
- if not t.empty():
- print("Max Value: ", t.get_max().value)
- print("Min Value: ", t.get_min().value)
-
- for i in testlist:
- t.remove(i)
- print(t)
+ postOrder (left, right, self)
+ """
+ node_list = []
+ if curr_node is not None:
+ node_list = postorder(curr_node.left) + postorder(curr_node.right) + [curr_node]
+ return node_list
if __name__ == "__main__":
import doctest
- doctest.testmod()
- # binary_search_tree()
+ doctest.testmod(verbose=True)
diff --git a/data_structures/binary_tree/binary_search_tree_recursive.py b/data_structures/binary_tree/binary_search_tree_recursive.py
index f1e46e33cd24..d94ac5253360 100644
--- a/data_structures/binary_tree/binary_search_tree_recursive.py
+++ b/data_structures/binary_tree/binary_search_tree_recursive.py
@@ -7,22 +7,28 @@
To run an example:
python binary_search_tree_recursive.py
"""
+
+from __future__ import annotations
+
import unittest
+from collections.abc import Iterator
+
+import pytest
class Node:
- def __init__(self, label: int, parent):
+ def __init__(self, label: int, parent: Node | None) -> None:
self.label = label
self.parent = parent
- self.left = None
- self.right = None
+ self.left: Node | None = None
+ self.right: Node | None = None
class BinarySearchTree:
- def __init__(self):
- self.root = None
+ def __init__(self) -> None:
+ self.root: Node | None = None
- def empty(self):
+ def empty(self) -> None:
"""
Empties the tree
@@ -46,7 +52,7 @@ def is_empty(self) -> bool:
"""
return self.root is None
- def put(self, label: int):
+ def put(self, label: int) -> None:
"""
Put a new node in the tree
@@ -65,16 +71,16 @@ def put(self, label: int):
"""
self.root = self._put(self.root, label)
- def _put(self, node: Node, label: int, parent: Node = None) -> Node:
+ def _put(self, node: Node | None, label: int, parent: Node | None = None) -> Node:
if node is None:
node = Node(label, parent)
+ elif label < node.label:
+ node.left = self._put(node.left, label, node)
+ elif label > node.label:
+ node.right = self._put(node.right, label, node)
else:
- if label < node.label:
- node.left = self._put(node.left, label, node)
- elif label > node.label:
- node.right = self._put(node.right, label, node)
- else:
- raise Exception(f"Node with label {label} already exists")
+ msg = f"Node with label {label} already exists"
+ raise ValueError(msg)
return node
@@ -91,22 +97,22 @@ def search(self, label: int) -> Node:
>>> node = t.search(3)
Traceback (most recent call last):
...
- Exception: Node with label 3 does not exist
+ ValueError: Node with label 3 does not exist
"""
return self._search(self.root, label)
- def _search(self, node: Node, label: int) -> Node:
+ def _search(self, node: Node | None, label: int) -> Node:
if node is None:
- raise Exception(f"Node with label {label} does not exist")
- else:
- if label < node.label:
- node = self._search(node.left, label)
- elif label > node.label:
- node = self._search(node.right, label)
+ msg = f"Node with label {label} does not exist"
+ raise ValueError(msg)
+ elif label < node.label:
+ node = self._search(node.left, label)
+ elif label > node.label:
+ node = self._search(node.right, label)
return node
- def remove(self, label: int):
+ def remove(self, label: int) -> None:
"""
Removes a node in the tree
@@ -119,16 +125,10 @@ def remove(self, label: int):
>>> t.remove(3)
Traceback (most recent call last):
...
- Exception: Node with label 3 does not exist
+ ValueError: Node with label 3 does not exist
"""
node = self.search(label)
- if not node.right and not node.left:
- self._reassign_nodes(node, None)
- elif not node.right and node.left:
- self._reassign_nodes(node, node.left)
- elif node.right and not node.left:
- self._reassign_nodes(node, node.right)
- else:
+ if node.right and node.left:
lowest_node = self._get_lowest_node(node.right)
lowest_node.left = node.left
lowest_node.right = node.right
@@ -136,8 +136,14 @@ def remove(self, label: int):
if node.right:
node.right.parent = lowest_node
self._reassign_nodes(node, lowest_node)
+ elif not node.right and node.left:
+ self._reassign_nodes(node, node.left)
+ elif node.right and not node.left:
+ self._reassign_nodes(node, node.right)
+ else:
+ self._reassign_nodes(node, None)
- def _reassign_nodes(self, node: Node, new_children: Node):
+ def _reassign_nodes(self, node: Node, new_children: Node | None) -> None:
if new_children:
new_children.parent = node.parent
@@ -174,7 +180,7 @@ def exists(self, label: int) -> bool:
try:
self.search(label)
return True
- except Exception:
+ except ValueError:
return False
def get_max_label(self) -> int:
@@ -185,15 +191,15 @@ def get_max_label(self) -> int:
>>> t.get_max_label()
Traceback (most recent call last):
...
- Exception: Binary search tree is empty
+ ValueError: Binary search tree is empty
>>> t.put(8)
>>> t.put(10)
>>> t.get_max_label()
10
"""
- if self.is_empty():
- raise Exception("Binary search tree is empty")
+ if self.root is None:
+ raise ValueError("Binary search tree is empty")
node = self.root
while node.right is not None:
@@ -209,15 +215,15 @@ def get_min_label(self) -> int:
>>> t.get_min_label()
Traceback (most recent call last):
...
- Exception: Binary search tree is empty
+ ValueError: Binary search tree is empty
>>> t.put(8)
>>> t.put(10)
>>> t.get_min_label()
8
"""
- if self.is_empty():
- raise Exception("Binary search tree is empty")
+ if self.root is None:
+ raise ValueError("Binary search tree is empty")
node = self.root
while node.left is not None:
@@ -225,7 +231,7 @@ def get_min_label(self) -> int:
return node.label
- def inorder_traversal(self) -> list:
+ def inorder_traversal(self) -> Iterator[Node]:
"""
Return the inorder traversal of the tree
@@ -241,13 +247,13 @@ def inorder_traversal(self) -> list:
"""
return self._inorder_traversal(self.root)
- def _inorder_traversal(self, node: Node) -> list:
+ def _inorder_traversal(self, node: Node | None) -> Iterator[Node]:
if node is not None:
yield from self._inorder_traversal(node.left)
yield node
yield from self._inorder_traversal(node.right)
- def preorder_traversal(self) -> list:
+ def preorder_traversal(self) -> Iterator[Node]:
"""
Return the preorder traversal of the tree
@@ -263,7 +269,7 @@ def preorder_traversal(self) -> list:
"""
return self._preorder_traversal(self.root)
- def _preorder_traversal(self, node: Node) -> list:
+ def _preorder_traversal(self, node: Node | None) -> Iterator[Node]:
if node is not None:
yield node
yield from self._preorder_traversal(node.left)
@@ -272,7 +278,7 @@ def _preorder_traversal(self, node: Node) -> list:
class BinarySearchTreeTest(unittest.TestCase):
@staticmethod
- def _get_binary_search_tree():
+ def _get_binary_search_tree() -> BinarySearchTree:
r"""
8
/ \
@@ -298,7 +304,7 @@ def _get_binary_search_tree():
return t
- def test_put(self):
+ def test_put(self) -> None:
t = BinarySearchTree()
assert t.is_empty()
@@ -306,6 +312,7 @@ def test_put(self):
r"""
8
"""
+ assert t.root is not None
assert t.root.parent is None
assert t.root.label == 8
@@ -315,6 +322,7 @@ def test_put(self):
\
10
"""
+ assert t.root.right is not None
assert t.root.right.parent == t.root
assert t.root.right.label == 10
@@ -324,6 +332,7 @@ def test_put(self):
/ \
3 10
"""
+ assert t.root.left is not None
assert t.root.left.parent == t.root
assert t.root.left.label == 3
@@ -335,6 +344,7 @@ def test_put(self):
\
6
"""
+ assert t.root.left.right is not None
assert t.root.left.right.parent == t.root.left
assert t.root.left.right.label == 6
@@ -346,13 +356,14 @@ def test_put(self):
/ \
1 6
"""
+ assert t.root.left.left is not None
assert t.root.left.left.parent == t.root.left
assert t.root.left.left.label == 1
- with self.assertRaises(Exception):
+ with pytest.raises(ValueError):
t.put(1)
- def test_search(self):
+ def test_search(self) -> None:
t = self._get_binary_search_tree()
node = t.search(6)
@@ -361,10 +372,10 @@ def test_search(self):
node = t.search(13)
assert node.label == 13
- with self.assertRaises(Exception):
+ with pytest.raises(ValueError):
t.search(2)
- def test_remove(self):
+ def test_remove(self) -> None:
t = self._get_binary_search_tree()
t.remove(13)
@@ -379,6 +390,9 @@ def test_remove(self):
\
5
"""
+ assert t.root is not None
+ assert t.root.right is not None
+ assert t.root.right.right is not None
assert t.root.right.right.right is None
assert t.root.right.right.left is None
@@ -394,6 +408,9 @@ def test_remove(self):
\
5
"""
+ assert t.root.left is not None
+ assert t.root.left.right is not None
+ assert t.root.left.right.left is not None
assert t.root.left.right.right is None
assert t.root.left.right.left.label == 4
@@ -407,6 +424,8 @@ def test_remove(self):
\
5
"""
+ assert t.root.left.left is not None
+ assert t.root.left.right.right is not None
assert t.root.left.left.label == 1
assert t.root.left.right.label == 4
assert t.root.left.right.right.label == 5
@@ -422,6 +441,7 @@ def test_remove(self):
/ \ \
1 5 14
"""
+ assert t.root is not None
assert t.root.left.label == 4
assert t.root.left.right.label == 5
assert t.root.left.left.label == 1
@@ -437,13 +457,15 @@ def test_remove(self):
/ \
1 14
"""
+ assert t.root.left is not None
+ assert t.root.left.left is not None
assert t.root.left.label == 5
assert t.root.left.right is None
assert t.root.left.left.label == 1
assert t.root.left.parent == t.root
assert t.root.left.left.parent == t.root.left
- def test_remove_2(self):
+ def test_remove_2(self) -> None:
t = self._get_binary_search_tree()
t.remove(3)
@@ -456,6 +478,12 @@ def test_remove_2(self):
/ \ /
5 7 13
"""
+ assert t.root is not None
+ assert t.root.left is not None
+ assert t.root.left.left is not None
+ assert t.root.left.right is not None
+ assert t.root.left.right.left is not None
+ assert t.root.left.right.right is not None
assert t.root.left.label == 4
assert t.root.left.right.label == 6
assert t.root.left.left.label == 1
@@ -466,56 +494,56 @@ def test_remove_2(self):
assert t.root.left.left.parent == t.root.left
assert t.root.left.right.left.parent == t.root.left.right
- def test_empty(self):
+ def test_empty(self) -> None:
t = self._get_binary_search_tree()
t.empty()
assert t.root is None
- def test_is_empty(self):
+ def test_is_empty(self) -> None:
t = self._get_binary_search_tree()
assert not t.is_empty()
t.empty()
assert t.is_empty()
- def test_exists(self):
+ def test_exists(self) -> None:
t = self._get_binary_search_tree()
assert t.exists(6)
assert not t.exists(-1)
- def test_get_max_label(self):
+ def test_get_max_label(self) -> None:
t = self._get_binary_search_tree()
assert t.get_max_label() == 14
t.empty()
- with self.assertRaises(Exception):
+ with pytest.raises(ValueError):
t.get_max_label()
- def test_get_min_label(self):
+ def test_get_min_label(self) -> None:
t = self._get_binary_search_tree()
assert t.get_min_label() == 1
t.empty()
- with self.assertRaises(Exception):
+ with pytest.raises(ValueError):
t.get_min_label()
- def test_inorder_traversal(self):
+ def test_inorder_traversal(self) -> None:
t = self._get_binary_search_tree()
inorder_traversal_nodes = [i.label for i in t.inorder_traversal()]
assert inorder_traversal_nodes == [1, 3, 4, 5, 6, 7, 8, 10, 13, 14]
- def test_preorder_traversal(self):
+ def test_preorder_traversal(self) -> None:
t = self._get_binary_search_tree()
preorder_traversal_nodes = [i.label for i in t.preorder_traversal()]
assert preorder_traversal_nodes == [8, 3, 1, 6, 4, 5, 7, 10, 14, 13]
-def binary_search_tree_example():
+def binary_search_tree_example() -> None:
r"""
Example
8
diff --git a/data_structures/binary_tree/binary_tree_mirror.py b/data_structures/binary_tree/binary_tree_mirror.py
index dc7f657b37c7..b8548f4ec515 100644
--- a/data_structures/binary_tree/binary_tree_mirror.py
+++ b/data_structures/binary_tree/binary_tree_mirror.py
@@ -1,6 +1,6 @@
"""
Problem Description:
-Given a binary tree, return it's mirror.
+Given a binary tree, return its mirror.
"""
@@ -21,17 +21,18 @@ def binary_tree_mirror(binary_tree: dict, root: int = 1) -> dict:
{1: [3, 2], 2: [5, 4], 3: [7, 6], 4: [11, 10]}
>>> binary_tree_mirror({ 1: [2,3], 2: [4,5], 3: [6,7], 4: [10,11]}, 5)
Traceback (most recent call last):
- ...
+ ...
ValueError: root 5 is not present in the binary_tree
>>> binary_tree_mirror({}, 5)
Traceback (most recent call last):
- ...
+ ...
ValueError: binary tree cannot be empty
"""
if not binary_tree:
raise ValueError("binary tree cannot be empty")
if root not in binary_tree:
- raise ValueError(f"root {root} is not present in the binary_tree")
+ msg = f"root {root} is not present in the binary_tree"
+ raise ValueError(msg)
binary_tree_mirror_dictionary = dict(binary_tree)
binary_tree_mirror_dict(binary_tree_mirror_dictionary, root)
return binary_tree_mirror_dictionary
diff --git a/data_structures/binary_tree/binary_tree_node_sum.py b/data_structures/binary_tree/binary_tree_node_sum.py
new file mode 100644
index 000000000000..066617b616c4
--- /dev/null
+++ b/data_structures/binary_tree/binary_tree_node_sum.py
@@ -0,0 +1,75 @@
+"""
+Sum of all nodes in a binary tree.
+
+Python implementation:
+ O(n) time complexity - Recurses through :meth:`depth_first_search`
+ with each element.
+ O(n) space complexity - At any point in time maximum number of stack
+ frames that could be in memory is `n`
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterator
+
+
+class Node:
+ """
+ A Node has a value variable and pointers to Nodes to its left and right.
+ """
+
+ def __init__(self, value: int) -> None:
+ self.value = value
+ self.left: Node | None = None
+ self.right: Node | None = None
+
+
+class BinaryTreeNodeSum:
+ r"""
+ The below tree looks like this
+ 10
+ / \
+ 5 -3
+ / / \
+ 12 8 0
+
+ >>> tree = Node(10)
+ >>> sum(BinaryTreeNodeSum(tree))
+ 10
+
+ >>> tree.left = Node(5)
+ >>> sum(BinaryTreeNodeSum(tree))
+ 15
+
+ >>> tree.right = Node(-3)
+ >>> sum(BinaryTreeNodeSum(tree))
+ 12
+
+ >>> tree.left.left = Node(12)
+ >>> sum(BinaryTreeNodeSum(tree))
+ 24
+
+ >>> tree.right.left = Node(8)
+ >>> tree.right.right = Node(0)
+ >>> sum(BinaryTreeNodeSum(tree))
+ 32
+ """
+
+ def __init__(self, tree: Node) -> None:
+ self.tree = tree
+
+ def depth_first_search(self, node: Node | None) -> int:
+ if node is None:
+ return 0
+ return node.value + (
+ self.depth_first_search(node.left) + self.depth_first_search(node.right)
+ )
+
+ def __iter__(self) -> Iterator[int]:
+ yield self.depth_first_search(self.tree)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/binary_tree_path_sum.py b/data_structures/binary_tree/binary_tree_path_sum.py
new file mode 100644
index 000000000000..a3fe9ca7a7e2
--- /dev/null
+++ b/data_structures/binary_tree/binary_tree_path_sum.py
@@ -0,0 +1,88 @@
+"""
+Given the root of a binary tree and an integer target,
+find the number of paths where the sum of the values
+along the path equals target.
+
+
+Leetcode reference: https://leetcode.com/problems/path-sum-iii/
+"""
+
+from __future__ import annotations
+
+
+class Node:
+ """
+ A Node has value variable and pointers to Nodes to its left and right.
+ """
+
+ def __init__(self, value: int) -> None:
+ self.value = value
+ self.left: Node | None = None
+ self.right: Node | None = None
+
+
+class BinaryTreePathSum:
+ r"""
+ The below tree looks like this
+ 10
+ / \
+ 5 -3
+ / \ \
+ 3 2 11
+ / \ \
+ 3 -2 1
+
+
+ >>> tree = Node(10)
+ >>> tree.left = Node(5)
+ >>> tree.right = Node(-3)
+ >>> tree.left.left = Node(3)
+ >>> tree.left.right = Node(2)
+ >>> tree.right.right = Node(11)
+ >>> tree.left.left.left = Node(3)
+ >>> tree.left.left.right = Node(-2)
+ >>> tree.left.right.right = Node(1)
+
+ >>> BinaryTreePathSum().path_sum(tree, 8)
+ 3
+ >>> BinaryTreePathSum().path_sum(tree, 7)
+ 2
+ >>> tree.right.right = Node(10)
+ >>> BinaryTreePathSum().path_sum(tree, 8)
+ 2
+ """
+
+ target: int
+
+ def __init__(self) -> None:
+ self.paths = 0
+
+ def depth_first_search(self, node: Node | None, path_sum: int) -> None:
+ if node is None:
+ return
+
+ if path_sum == self.target:
+ self.paths += 1
+
+ if node.left:
+ self.depth_first_search(node.left, path_sum + node.left.value)
+ if node.right:
+ self.depth_first_search(node.right, path_sum + node.right.value)
+
+ def path_sum(self, node: Node | None, target: int | None = None) -> int:
+ if node is None:
+ return 0
+ if target is not None:
+ self.target = target
+
+ self.depth_first_search(node, node.value)
+ self.path_sum(node.left)
+ self.path_sum(node.right)
+
+ return self.paths
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/binary_tree_traversals.py b/data_structures/binary_tree/binary_tree_traversals.py
index 7c0ee1dbbc2a..5ba149d0cbc6 100644
--- a/data_structures/binary_tree/binary_tree_traversals.py
+++ b/data_structures/binary_tree/binary_tree_traversals.py
@@ -1,54 +1,88 @@
-# https://en.wikipedia.org/wiki/Tree_traversal
+from __future__ import annotations
+from collections import deque
+from collections.abc import Generator
+from dataclasses import dataclass
-class Node:
- """
- A Node has data variable and pointers to its left and right nodes.
- """
- def __init__(self, data):
- self.left = None
- self.right = None
- self.data = data
+# https://en.wikipedia.org/wiki/Tree_traversal
+@dataclass
+class Node:
+ data: int
+ left: Node | None = None
+ right: Node | None = None
-def make_tree() -> Node:
- root = Node(1)
- root.left = Node(2)
- root.right = Node(3)
- root.left.left = Node(4)
- root.left.right = Node(5)
- return root
+def make_tree() -> Node | None:
+ r"""
+ The below tree
+ 1
+ / \
+ 2 3
+ / \
+ 4 5
+ """
+ tree = Node(1)
+ tree.left = Node(2)
+ tree.right = Node(3)
+ tree.left.left = Node(4)
+ tree.left.right = Node(5)
+ return tree
-def preorder(root: Node):
+def preorder(root: Node | None) -> Generator[int]:
"""
Pre-order traversal visits root node, left subtree, right subtree.
- >>> preorder(make_tree())
+ >>> list(preorder(make_tree()))
[1, 2, 4, 5, 3]
"""
- return [root.data] + preorder(root.left) + preorder(root.right) if root else []
+ if not root:
+ return
+ yield root.data
+ yield from preorder(root.left)
+ yield from preorder(root.right)
-def postorder(root: Node):
+def postorder(root: Node | None) -> Generator[int]:
"""
Post-order traversal visits left subtree, right subtree, root node.
- >>> postorder(make_tree())
+ >>> list(postorder(make_tree()))
[4, 5, 2, 3, 1]
"""
- return postorder(root.left) + postorder(root.right) + [root.data] if root else []
+ if not root:
+ return
+ yield from postorder(root.left)
+ yield from postorder(root.right)
+ yield root.data
-def inorder(root: Node):
+def inorder(root: Node | None) -> Generator[int]:
"""
In-order traversal visits left subtree, root node, right subtree.
- >>> inorder(make_tree())
+ >>> list(inorder(make_tree()))
[4, 2, 5, 1, 3]
"""
- return inorder(root.left) + [root.data] + inorder(root.right) if root else []
+ if not root:
+ return
+ yield from inorder(root.left)
+ yield root.data
+ yield from inorder(root.right)
-def height(root: Node):
+def reverse_inorder(root: Node | None) -> Generator[int]:
+ """
+ Reverse in-order traversal visits right subtree, root node, left subtree.
+ >>> list(reverse_inorder(make_tree()))
+ [3, 1, 5, 2, 4]
+ """
+ if not root:
+ return
+ yield from reverse_inorder(root.right)
+ yield root.data
+ yield from reverse_inorder(root.left)
+
+
+def height(root: Node | None) -> int:
"""
Recursive function for calculating the height of the binary tree.
>>> height(None)
@@ -59,99 +93,117 @@ def height(root: Node):
return (max(height(root.left), height(root.right)) + 1) if root else 0
-def level_order_1(root: Node):
+def level_order(root: Node | None) -> Generator[int]:
"""
- Print whole binary tree in Level Order Traverse.
+ Returns a list of nodes value from a whole binary tree in Level Order Traverse.
Level Order traverse: Visit nodes of the tree level-by-level.
+ >>> list(level_order(make_tree()))
+ [1, 2, 3, 4, 5]
"""
- if not root:
+
+ if root is None:
return
- temp = root
- que = [temp]
- while len(que) > 0:
- print(que[0].data, end=" ")
- temp = que.pop(0)
- if temp.left:
- que.append(temp.left)
- if temp.right:
- que.append(temp.right)
- return que
+ process_queue = deque([root])
-def level_order_2(root: Node, level: int):
- """
- Level-wise traversal: Print all nodes present at the given level of the binary tree
- """
- if not root:
- return root
- if level == 1:
- print(root.data, end=" ")
- elif level > 1:
- level_order_2(root.left, level - 1)
- level_order_2(root.right, level - 1)
+ while process_queue:
+ node = process_queue.popleft()
+ yield node.data
+
+ if node.left:
+ process_queue.append(node.left)
+ if node.right:
+ process_queue.append(node.right)
-def print_left_to_right(root: Node, level: int):
+def get_nodes_from_left_to_right(root: Node | None, level: int) -> Generator[int]:
"""
- Print elements on particular level from left to right direction of the binary tree.
+ Returns a list of nodes value from a particular level:
+ Left to right direction of the binary tree.
+ >>> list(get_nodes_from_left_to_right(make_tree(), 1))
+ [1]
+ >>> list(get_nodes_from_left_to_right(make_tree(), 2))
+ [2, 3]
"""
- if not root:
- return
- if level == 1:
- print(root.data, end=" ")
- elif level > 1:
- print_left_to_right(root.left, level - 1)
- print_left_to_right(root.right, level - 1)
+
+ def populate_output(root: Node | None, level: int) -> Generator[int]:
+ if not root:
+ return
+ if level == 1:
+ yield root.data
+ elif level > 1:
+ yield from populate_output(root.left, level - 1)
+ yield from populate_output(root.right, level - 1)
+
+ yield from populate_output(root, level)
-def print_right_to_left(root: Node, level: int):
+def get_nodes_from_right_to_left(root: Node | None, level: int) -> Generator[int]:
"""
- Print elements on particular level from right to left direction of the binary tree.
+ Returns a list of nodes value from a particular level:
+ Right to left direction of the binary tree.
+ >>> list(get_nodes_from_right_to_left(make_tree(), 1))
+ [1]
+ >>> list(get_nodes_from_right_to_left(make_tree(), 2))
+ [3, 2]
"""
- if not root:
- return
- if level == 1:
- print(root.data, end=" ")
- elif level > 1:
- print_right_to_left(root.right, level - 1)
- print_right_to_left(root.left, level - 1)
+ def populate_output(root: Node | None, level: int) -> Generator[int]:
+ if not root:
+ return
+ if level == 1:
+ yield root.data
+ elif level > 1:
+ yield from populate_output(root.right, level - 1)
+ yield from populate_output(root.left, level - 1)
+
+ yield from populate_output(root, level)
-def zigzag(root: Node):
+
+def zigzag(root: Node | None) -> Generator[int]:
"""
- ZigZag traverse: Print node left to right and right to left, alternatively.
+ ZigZag traverse:
+ Returns a list of nodes value from left to right and right to left, alternatively.
+ >>> list(zigzag(make_tree()))
+ [1, 3, 2, 4, 5]
"""
+ if root is None:
+ return
+
flag = 0
height_tree = height(root)
+
for h in range(1, height_tree + 1):
- if flag == 0:
- print_left_to_right(root, h)
+ if not flag:
+ yield from get_nodes_from_left_to_right(root, h)
flag = 1
else:
- print_right_to_left(root, h)
+ yield from get_nodes_from_right_to_left(root, h)
flag = 0
-def main(): # Main function for testing.
- """
- Create binary tree.
- """
+def main() -> None: # Main function for testing.
+ # Create binary tree.
root = make_tree()
- """
- All Traversals of the binary are as follows:
- """
- print(f" In-order Traversal is {inorder(root)}")
- print(f" Pre-order Traversal is {preorder(root)}")
- print(f"Post-order Traversal is {postorder(root)}")
- print(f"Height of Tree is {height(root)}")
- print("Complete Level Order Traversal is : ")
- level_order_1(root)
- print("\nLevel-wise order Traversal is : ")
- for h in range(1, height(root) + 1):
- level_order_2(root, h)
- print("\nZigZag order Traversal is : ")
- zigzag(root)
- print()
+
+ # All Traversals of the binary are as follows:
+ print(f"In-order Traversal: {list(inorder(root))}")
+ print(f"Reverse In-order Traversal: {list(reverse_inorder(root))}")
+ print(f"Pre-order Traversal: {list(preorder(root))}")
+ print(f"Post-order Traversal: {list(postorder(root))}", "\n")
+
+ print(f"Height of Tree: {height(root)}", "\n")
+
+ print("Complete Level Order Traversal: ")
+ print(f"{list(level_order(root))} \n")
+
+ print("Level-wise order Traversal: ")
+
+ for level in range(1, height(root) + 1):
+ print(f"Level {level}:", list(get_nodes_from_left_to_right(root, level=level)))
+
+ print("\nZigZag order Traversal: ")
+ print(f"{list(zigzag(root))}")
if __name__ == "__main__":
diff --git a/data_structures/binary_tree/diameter_of_binary_tree.py b/data_structures/binary_tree/diameter_of_binary_tree.py
new file mode 100644
index 000000000000..75e5e7373323
--- /dev/null
+++ b/data_structures/binary_tree/diameter_of_binary_tree.py
@@ -0,0 +1,73 @@
+"""
+The diameter/width of a tree is defined as the number of nodes on the longest path
+between two end nodes.
+"""
+
+from __future__ import annotations
+
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ data: int
+ left: Node | None = None
+ right: Node | None = None
+
+ def depth(self) -> int:
+ """
+ >>> root = Node(1)
+ >>> root.depth()
+ 1
+ >>> root.left = Node(2)
+ >>> root.depth()
+ 2
+ >>> root.left.depth()
+ 1
+ >>> root.right = Node(3)
+ >>> root.depth()
+ 2
+ """
+ left_depth = self.left.depth() if self.left else 0
+ right_depth = self.right.depth() if self.right else 0
+ return max(left_depth, right_depth) + 1
+
+ def diameter(self) -> int:
+ """
+ >>> root = Node(1)
+ >>> root.diameter()
+ 1
+ >>> root.left = Node(2)
+ >>> root.diameter()
+ 2
+ >>> root.left.diameter()
+ 1
+ >>> root.right = Node(3)
+ >>> root.diameter()
+ 3
+ """
+ left_depth = self.left.depth() if self.left else 0
+ right_depth = self.right.depth() if self.right else 0
+ return left_depth + right_depth + 1
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ root = Node(1)
+ root.left = Node(2)
+ root.right = Node(3)
+ root.left.left = Node(4)
+ root.left.right = Node(5)
+ r"""
+ Constructed binary tree is
+ 1
+ / \
+ 2 3
+ / \
+ 4 5
+ """
+ print(f"{root.diameter() = }") # 4
+ print(f"{root.left.diameter() = }") # 3
+ print(f"{root.right.diameter() = }") # 1
diff --git a/data_structures/binary_tree/diff_views_of_binary_tree.py b/data_structures/binary_tree/diff_views_of_binary_tree.py
new file mode 100644
index 000000000000..3198d8065918
--- /dev/null
+++ b/data_structures/binary_tree/diff_views_of_binary_tree.py
@@ -0,0 +1,210 @@
+r"""
+Problem: Given root of a binary tree, return the:
+1. binary-tree-right-side-view
+2. binary-tree-left-side-view
+3. binary-tree-top-side-view
+4. binary-tree-bottom-side-view
+"""
+
+from __future__ import annotations
+
+from collections import defaultdict
+from dataclasses import dataclass
+
+
+@dataclass
+class TreeNode:
+ val: int
+ left: TreeNode | None = None
+ right: TreeNode | None = None
+
+
+def make_tree() -> TreeNode:
+ """
+ >>> make_tree().val
+ 3
+ """
+ return TreeNode(3, TreeNode(9), TreeNode(20, TreeNode(15), TreeNode(7)))
+
+
+def binary_tree_right_side_view(root: TreeNode) -> list[int]:
+ r"""
+ Function returns the right side view of binary tree.
+
+ 3 <- 3
+ / \
+ 9 20 <- 20
+ / \
+ 15 7 <- 7
+
+ >>> binary_tree_right_side_view(make_tree())
+ [3, 20, 7]
+ >>> binary_tree_right_side_view(None)
+ []
+ """
+
+ def depth_first_search(
+ root: TreeNode | None, depth: int, right_view: list[int]
+ ) -> None:
+ """
+ A depth first search preorder traversal to append the values at
+ right side of tree.
+ """
+ if not root:
+ return
+
+ if depth == len(right_view):
+ right_view.append(root.val)
+
+ depth_first_search(root.right, depth + 1, right_view)
+ depth_first_search(root.left, depth + 1, right_view)
+
+ right_view: list = []
+ if not root:
+ return right_view
+
+ depth_first_search(root, 0, right_view)
+ return right_view
+
+
+def binary_tree_left_side_view(root: TreeNode) -> list[int]:
+ r"""
+ Function returns the left side view of binary tree.
+
+ 3 -> 3
+ / \
+ 9 -> 9 20
+ / \
+ 15 -> 15 7
+
+ >>> binary_tree_left_side_view(make_tree())
+ [3, 9, 15]
+ >>> binary_tree_left_side_view(None)
+ []
+ """
+
+ def depth_first_search(
+ root: TreeNode | None, depth: int, left_view: list[int]
+ ) -> None:
+ """
+ A depth first search preorder traversal to append the values
+ at left side of tree.
+ """
+ if not root:
+ return
+
+ if depth == len(left_view):
+ left_view.append(root.val)
+
+ depth_first_search(root.left, depth + 1, left_view)
+ depth_first_search(root.right, depth + 1, left_view)
+
+ left_view: list = []
+ if not root:
+ return left_view
+
+ depth_first_search(root, 0, left_view)
+ return left_view
+
+
+def binary_tree_top_side_view(root: TreeNode) -> list[int]:
+ r"""
+ Function returns the top side view of binary tree.
+
+ 9 3 20 7
+ ⬇ ⬇ ⬇ ⬇
+
+ 3
+ / \
+ 9 20
+ / \
+ 15 7
+
+ >>> binary_tree_top_side_view(make_tree())
+ [9, 3, 20, 7]
+ >>> binary_tree_top_side_view(None)
+ []
+ """
+
+ def breadth_first_search(root: TreeNode, top_view: list[int]) -> None:
+ """
+ A breadth first search traversal with defaultdict ds to append
+ the values of tree from top view
+ """
+ queue = [(root, 0)]
+ lookup = defaultdict(list)
+
+ while queue:
+ first = queue.pop(0)
+ node, hd = first
+
+ lookup[hd].append(node.val)
+
+ if node.left:
+ queue.append((node.left, hd - 1))
+ if node.right:
+ queue.append((node.right, hd + 1))
+
+ for pair in sorted(lookup.items(), key=lambda each: each[0]):
+ top_view.append(pair[1][0])
+
+ top_view: list = []
+ if not root:
+ return top_view
+
+ breadth_first_search(root, top_view)
+ return top_view
+
+
+def binary_tree_bottom_side_view(root: TreeNode) -> list[int]:
+ r"""
+ Function returns the bottom side view of binary tree
+
+ 3
+ / \
+ 9 20
+ / \
+ 15 7
+ ↑ ↑ ↑ ↑
+ 9 15 20 7
+
+ >>> binary_tree_bottom_side_view(make_tree())
+ [9, 15, 20, 7]
+ >>> binary_tree_bottom_side_view(None)
+ []
+ """
+ from collections import defaultdict
+
+ def breadth_first_search(root: TreeNode, bottom_view: list[int]) -> None:
+ """
+ A breadth first search traversal with defaultdict ds to append
+ the values of tree from bottom view
+ """
+ queue = [(root, 0)]
+ lookup = defaultdict(list)
+
+ while queue:
+ first = queue.pop(0)
+ node, hd = first
+ lookup[hd].append(node.val)
+
+ if node.left:
+ queue.append((node.left, hd - 1))
+ if node.right:
+ queue.append((node.right, hd + 1))
+
+ for pair in sorted(lookup.items(), key=lambda each: each[0]):
+ bottom_view.append(pair[1][-1])
+
+ bottom_view: list = []
+ if not root:
+ return bottom_view
+
+ breadth_first_search(root, bottom_view)
+ return bottom_view
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/distribute_coins.py b/data_structures/binary_tree/distribute_coins.py
new file mode 100644
index 000000000000..5712604cb87c
--- /dev/null
+++ b/data_structures/binary_tree/distribute_coins.py
@@ -0,0 +1,137 @@
+"""
+Author : Alexander Pantyukhin
+Date : November 7, 2022
+
+Task:
+You are given a tree root of a binary tree with n nodes, where each node has
+node.data coins. There are exactly n coins in whole tree.
+
+In one move, we may choose two adjacent nodes and move one coin from one node
+to another. A move may be from parent to child, or from child to parent.
+
+Return the minimum number of moves required to make every node have exactly one coin.
+
+Example 1:
+
+ 3
+ / \
+ 0 0
+
+Result: 2
+
+Example 2:
+
+ 0
+ / \
+ 3 0
+
+Result 3
+
+leetcode: https://leetcode.com/problems/distribute-coins-in-binary-tree/
+
+Implementation notes:
+User depth-first search approach.
+
+Let n is the number of nodes in tree
+Runtime: O(n)
+Space: O(1)
+"""
+
+from __future__ import annotations
+
+from dataclasses import dataclass
+from typing import NamedTuple
+
+
+@dataclass
+class TreeNode:
+ data: int
+ left: TreeNode | None = None
+ right: TreeNode | None = None
+
+
+class CoinsDistribResult(NamedTuple):
+ moves: int
+ excess: int
+
+
+def distribute_coins(root: TreeNode | None) -> int:
+ """
+ >>> distribute_coins(TreeNode(3, TreeNode(0), TreeNode(0)))
+ 2
+ >>> distribute_coins(TreeNode(0, TreeNode(3), TreeNode(0)))
+ 3
+ >>> distribute_coins(TreeNode(0, TreeNode(0), TreeNode(3)))
+ 3
+ >>> distribute_coins(None)
+ 0
+ >>> distribute_coins(TreeNode(0, TreeNode(0), TreeNode(0)))
+ Traceback (most recent call last):
+ ...
+ ValueError: The nodes number should be same as the number of coins
+ >>> distribute_coins(TreeNode(0, TreeNode(1), TreeNode(1)))
+ Traceback (most recent call last):
+ ...
+ ValueError: The nodes number should be same as the number of coins
+ """
+
+ if root is None:
+ return 0
+
+ # Validation
+ def count_nodes(node: TreeNode | None) -> int:
+ """
+ >>> count_nodes(None)
+ 0
+ """
+ if node is None:
+ return 0
+
+ return count_nodes(node.left) + count_nodes(node.right) + 1
+
+ def count_coins(node: TreeNode | None) -> int:
+ """
+ >>> count_coins(None)
+ 0
+ """
+ if node is None:
+ return 0
+
+ return count_coins(node.left) + count_coins(node.right) + node.data
+
+ if count_nodes(root) != count_coins(root):
+ raise ValueError("The nodes number should be same as the number of coins")
+
+ # Main calculation
+ def get_distrib(node: TreeNode | None) -> CoinsDistribResult:
+ """
+ >>> get_distrib(None)
+ namedtuple("CoinsDistribResult", "0 2")
+ """
+
+ if node is None:
+ return CoinsDistribResult(0, 1)
+
+ left_distrib_moves, left_distrib_excess = get_distrib(node.left)
+ right_distrib_moves, right_distrib_excess = get_distrib(node.right)
+
+ coins_to_left = 1 - left_distrib_excess
+ coins_to_right = 1 - right_distrib_excess
+
+ result_moves = (
+ left_distrib_moves
+ + right_distrib_moves
+ + abs(coins_to_left)
+ + abs(coins_to_right)
+ )
+ result_excess = node.data - coins_to_left - coins_to_right
+
+ return CoinsDistribResult(result_moves, result_excess)
+
+ return get_distrib(root)[0]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/fenwick_tree.py b/data_structures/binary_tree/fenwick_tree.py
index 54f0f07ac68d..88b0873a10fb 100644
--- a/data_structures/binary_tree/fenwick_tree.py
+++ b/data_structures/binary_tree/fenwick_tree.py
@@ -1,28 +1,247 @@
+from copy import deepcopy
+
+
class FenwickTree:
- def __init__(self, SIZE): # create fenwick tree with size SIZE
- self.Size = SIZE
- self.ft = [0 for i in range(0, SIZE)]
+ """
+ Fenwick Tree
+
+ More info: https://en.wikipedia.org/wiki/Fenwick_tree
+ """
+
+ def __init__(self, arr: list[int] | None = None, size: int | None = None) -> None:
+ """
+ Constructor for the Fenwick tree
+
+ Parameters:
+ arr (list): list of elements to initialize the tree with (optional)
+ size (int): size of the Fenwick tree (if arr is None)
+ """
+
+ if arr is None and size is not None:
+ self.size = size
+ self.tree = [0] * size
+ elif arr is not None:
+ self.init(arr)
+ else:
+ raise ValueError("Either arr or size must be specified")
+
+ def init(self, arr: list[int]) -> None:
+ """
+ Initialize the Fenwick tree with arr in O(N)
+
+ Parameters:
+ arr (list): list of elements to initialize the tree with
+
+ Returns:
+ None
+
+ >>> a = [1, 2, 3, 4, 5]
+ >>> f1 = FenwickTree(a)
+ >>> f2 = FenwickTree(size=len(a))
+ >>> for index, value in enumerate(a):
+ ... f2.add(index, value)
+ >>> f1.tree == f2.tree
+ True
+ """
+ self.size = len(arr)
+ self.tree = deepcopy(arr)
+ for i in range(1, self.size):
+ j = self.next_(i)
+ if j < self.size:
+ self.tree[j] += self.tree[i]
+
+ def get_array(self) -> list[int]:
+ """
+ Get the Normal Array of the Fenwick tree in O(N)
+
+ Returns:
+ list: Normal Array of the Fenwick tree
+
+ >>> a = [i for i in range(128)]
+ >>> f = FenwickTree(a)
+ >>> f.get_array() == a
+ True
+ """
+ arr = self.tree[:]
+ for i in range(self.size - 1, 0, -1):
+ j = self.next_(i)
+ if j < self.size:
+ arr[j] -= arr[i]
+ return arr
+
+ @staticmethod
+ def next_(index: int) -> int:
+ return index + (index & (-index))
+
+ @staticmethod
+ def prev(index: int) -> int:
+ return index - (index & (-index))
+
+ def add(self, index: int, value: int) -> None:
+ """
+ Add a value to index in O(lg N)
+
+ Parameters:
+ index (int): index to add value to
+ value (int): value to add to index
+
+ Returns:
+ None
+
+ >>> f = FenwickTree([1, 2, 3, 4, 5])
+ >>> f.add(0, 1)
+ >>> f.add(1, 2)
+ >>> f.add(2, 3)
+ >>> f.add(3, 4)
+ >>> f.add(4, 5)
+ >>> f.get_array()
+ [2, 4, 6, 8, 10]
+ """
+ if index == 0:
+ self.tree[0] += value
+ return
+ while index < self.size:
+ self.tree[index] += value
+ index = self.next_(index)
+
+ def update(self, index: int, value: int) -> None:
+ """
+ Set the value of index in O(lg N)
+
+ Parameters:
+ index (int): index to set value to
+ value (int): value to set in index
- def update(self, i, val): # update data (adding) in index i in O(lg N)
- while i < self.Size:
- self.ft[i] += val
- i += i & (-i)
+ Returns:
+ None
- def query(self, i): # query cumulative data from index 0 to i in O(lg N)
- ret = 0
- while i > 0:
- ret += self.ft[i]
- i -= i & (-i)
- return ret
+ >>> f = FenwickTree([5, 4, 3, 2, 1])
+ >>> f.update(0, 1)
+ >>> f.update(1, 2)
+ >>> f.update(2, 3)
+ >>> f.update(3, 4)
+ >>> f.update(4, 5)
+ >>> f.get_array()
+ [1, 2, 3, 4, 5]
+ """
+ self.add(index, value - self.get(index))
+
+ def prefix(self, right: int) -> int:
+ """
+ Prefix sum of all elements in [0, right) in O(lg N)
+
+ Parameters:
+ right (int): right bound of the query (exclusive)
+
+ Returns:
+ int: sum of all elements in [0, right)
+
+ >>> a = [i for i in range(128)]
+ >>> f = FenwickTree(a)
+ >>> res = True
+ >>> for i in range(len(a)):
+ ... res = res and f.prefix(i) == sum(a[:i])
+ >>> res
+ True
+ """
+ if right == 0:
+ return 0
+ result = self.tree[0]
+ right -= 1 # make right inclusive
+ while right > 0:
+ result += self.tree[right]
+ right = self.prev(right)
+ return result
+
+ def query(self, left: int, right: int) -> int:
+ """
+ Query the sum of all elements in [left, right) in O(lg N)
+
+ Parameters:
+ left (int): left bound of the query (inclusive)
+ right (int): right bound of the query (exclusive)
+
+ Returns:
+ int: sum of all elements in [left, right)
+
+ >>> a = [i for i in range(128)]
+ >>> f = FenwickTree(a)
+ >>> res = True
+ >>> for i in range(len(a)):
+ ... for j in range(i + 1, len(a)):
+ ... res = res and f.query(i, j) == sum(a[i:j])
+ >>> res
+ True
+ """
+ return self.prefix(right) - self.prefix(left)
+
+ def get(self, index: int) -> int:
+ """
+ Get value at index in O(lg N)
+
+ Parameters:
+ index (int): index to get the value
+
+ Returns:
+ int: Value of element at index
+
+ >>> a = [i for i in range(128)]
+ >>> f = FenwickTree(a)
+ >>> res = True
+ >>> for i in range(len(a)):
+ ... res = res and f.get(i) == a[i]
+ >>> res
+ True
+ """
+ return self.query(index, index + 1)
+
+ def rank_query(self, value: int) -> int:
+ """
+ Find the largest index with prefix(i) <= value in O(lg N)
+ NOTE: Requires that all values are non-negative!
+
+ Parameters:
+ value (int): value to find the largest index of
+
+ Returns:
+ -1: if value is smaller than all elements in prefix sum
+ int: largest index with prefix(i) <= value
+
+ >>> f = FenwickTree([1, 2, 0, 3, 0, 5])
+ >>> f.rank_query(0)
+ -1
+ >>> f.rank_query(2)
+ 0
+ >>> f.rank_query(1)
+ 0
+ >>> f.rank_query(3)
+ 2
+ >>> f.rank_query(5)
+ 2
+ >>> f.rank_query(6)
+ 4
+ >>> f.rank_query(11)
+ 5
+ """
+ value -= self.tree[0]
+ if value < 0:
+ return -1
+
+ j = 1 # Largest power of 2 <= size
+ while j * 2 < self.size:
+ j *= 2
+
+ i = 0
+
+ while j > 0:
+ if i + j < self.size and self.tree[i + j] <= value:
+ value -= self.tree[i + j]
+ i += j
+ j //= 2
+ return i
if __name__ == "__main__":
- f = FenwickTree(100)
- f.update(1, 20)
- f.update(4, 4)
- print(f.query(1))
- print(f.query(3))
- print(f.query(4))
- f.update(2, -5)
- print(f.query(1))
- print(f.query(3))
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/flatten_binarytree_to_linkedlist.py b/data_structures/binary_tree/flatten_binarytree_to_linkedlist.py
new file mode 100644
index 000000000000..9b2c7b9af24b
--- /dev/null
+++ b/data_structures/binary_tree/flatten_binarytree_to_linkedlist.py
@@ -0,0 +1,139 @@
+"""
+Binary Tree Flattening Algorithm
+
+This code defines an algorithm to flatten a binary tree into a linked list
+represented using the right pointers of the tree nodes. It uses in-place
+flattening and demonstrates the flattening process along with a display
+function to visualize the flattened linked list.
+https://www.geeksforgeeks.org/flatten-a-binary-tree-into-linked-list
+
+Author: Arunkumar A
+Date: 04/09/2023
+"""
+
+from __future__ import annotations
+
+
+class TreeNode:
+ """
+ A TreeNode has data variable and pointers to TreeNode objects
+ for its left and right children.
+ """
+
+ def __init__(self, data: int) -> None:
+ self.data = data
+ self.left: TreeNode | None = None
+ self.right: TreeNode | None = None
+
+
+def build_tree() -> TreeNode:
+ """
+ Build and return a sample binary tree.
+
+ Returns:
+ TreeNode: The root of the binary tree.
+
+ Examples:
+ >>> root = build_tree()
+ >>> root.data
+ 1
+ >>> root.left.data
+ 2
+ >>> root.right.data
+ 5
+ >>> root.left.left.data
+ 3
+ >>> root.left.right.data
+ 4
+ >>> root.right.right.data
+ 6
+ """
+ root = TreeNode(1)
+ root.left = TreeNode(2)
+ root.right = TreeNode(5)
+ root.left.left = TreeNode(3)
+ root.left.right = TreeNode(4)
+ root.right.right = TreeNode(6)
+ return root
+
+
+def flatten(root: TreeNode | None) -> None:
+ """
+ Flatten a binary tree into a linked list in-place, where the linked list is
+ represented using the right pointers of the tree nodes.
+
+ Args:
+ root (TreeNode): The root of the binary tree to be flattened.
+
+ Examples:
+ >>> root = TreeNode(1)
+ >>> root.left = TreeNode(2)
+ >>> root.right = TreeNode(5)
+ >>> root.left.left = TreeNode(3)
+ >>> root.left.right = TreeNode(4)
+ >>> root.right.right = TreeNode(6)
+ >>> flatten(root)
+ >>> root.data
+ 1
+ >>> root.right.right is None
+ False
+ >>> root.right.right = TreeNode(3)
+ >>> root.right.right.right is None
+ True
+ """
+ if not root:
+ return
+
+ # Flatten the left subtree
+ flatten(root.left)
+
+ # Save the right subtree
+ right_subtree = root.right
+
+ # Make the left subtree the new right subtree
+ root.right = root.left
+ root.left = None
+
+ # Find the end of the new right subtree
+ current = root
+ while current.right:
+ current = current.right
+
+ # Append the original right subtree to the end
+ current.right = right_subtree
+
+ # Flatten the updated right subtree
+ flatten(right_subtree)
+
+
+def display_linked_list(root: TreeNode | None) -> None:
+ """
+ Display the flattened linked list.
+
+ Args:
+ root (TreeNode | None): The root of the flattened linked list.
+
+ Examples:
+ >>> root = TreeNode(1)
+ >>> root.right = TreeNode(2)
+ >>> root.right.right = TreeNode(3)
+ >>> display_linked_list(root)
+ 1 2 3
+ >>> root = None
+ >>> display_linked_list(root)
+
+ """
+ current = root
+ while current:
+ if current.right is None:
+ print(current.data, end="")
+ break
+ print(current.data, end=" ")
+ current = current.right
+
+
+if __name__ == "__main__":
+ print("Flattened Linked List:")
+ root = build_tree()
+ flatten(root)
+ display_linked_list(root)
diff --git a/data_structures/binary_tree/floor_and_ceiling.py b/data_structures/binary_tree/floor_and_ceiling.py
new file mode 100644
index 000000000000..b464aefad3a2
--- /dev/null
+++ b/data_structures/binary_tree/floor_and_ceiling.py
@@ -0,0 +1,88 @@
+"""
+In a binary search tree (BST):
+* The floor of key 'k' is the maximum value that is smaller than or equal to 'k'.
+* The ceiling of key 'k' is the minimum value that is greater than or equal to 'k'.
+
+Reference:
+https://bit.ly/46uB0a2
+
+Author : Arunkumar
+Date : 14th October 2023
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ key: int
+ left: Node | None = None
+ right: Node | None = None
+
+ def __iter__(self) -> Iterator[int]:
+ if self.left:
+ yield from self.left
+ yield self.key
+ if self.right:
+ yield from self.right
+
+ def __len__(self) -> int:
+ return sum(1 for _ in self)
+
+
+def floor_ceiling(root: Node | None, key: int) -> tuple[int | None, int | None]:
+ """
+ Find the floor and ceiling values for a given key in a Binary Search Tree (BST).
+
+ Args:
+ root: The root of the binary search tree.
+ key: The key for which to find the floor and ceiling.
+
+ Returns:
+ A tuple containing the floor and ceiling values, respectively.
+
+ Examples:
+ >>> root = Node(10)
+ >>> root.left = Node(5)
+ >>> root.right = Node(20)
+ >>> root.left.left = Node(3)
+ >>> root.left.right = Node(7)
+ >>> root.right.left = Node(15)
+ >>> root.right.right = Node(25)
+ >>> tuple(root)
+ (3, 5, 7, 10, 15, 20, 25)
+ >>> floor_ceiling(root, 8)
+ (7, 10)
+ >>> floor_ceiling(root, 14)
+ (10, 15)
+ >>> floor_ceiling(root, -1)
+ (None, 3)
+ >>> floor_ceiling(root, 30)
+ (25, None)
+ """
+ floor_val = None
+ ceiling_val = None
+
+ while root:
+ if root.key == key:
+ floor_val = root.key
+ ceiling_val = root.key
+ break
+
+ if key < root.key:
+ ceiling_val = root.key
+ root = root.left
+ else:
+ floor_val = root.key
+ root = root.right
+
+ return floor_val, ceiling_val
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/inorder_tree_traversal_2022.py b/data_structures/binary_tree/inorder_tree_traversal_2022.py
new file mode 100644
index 000000000000..1357527d2953
--- /dev/null
+++ b/data_structures/binary_tree/inorder_tree_traversal_2022.py
@@ -0,0 +1,82 @@
+"""
+Illustrate how to implement inorder traversal in binary search tree.
+Author: Gurneet Singh
+https://www.geeksforgeeks.org/tree-traversals-inorder-preorder-and-postorder/
+"""
+
+
+class BinaryTreeNode:
+ """Defining the structure of BinaryTreeNode"""
+
+ def __init__(self, data: int) -> None:
+ self.data = data
+ self.left_child: BinaryTreeNode | None = None
+ self.right_child: BinaryTreeNode | None = None
+
+
+def insert(node: BinaryTreeNode | None, new_value: int) -> BinaryTreeNode | None:
+ """
+ If the binary search tree is empty, make a new node and declare it as root.
+ >>> node_a = BinaryTreeNode(12345)
+ >>> node_b = insert(node_a, 67890)
+ >>> node_a.left_child == node_b.left_child
+ True
+ >>> node_a.right_child == node_b.right_child
+ True
+ >>> node_a.data == node_b.data
+ True
+ """
+ if node is None:
+ node = BinaryTreeNode(new_value)
+ return node
+
+ # binary search tree is not empty,
+ # so we will insert it into the tree
+ # if new_value is less than value of data in node,
+ # add it to left subtree and proceed recursively
+ if new_value < node.data:
+ node.left_child = insert(node.left_child, new_value)
+ else:
+ # if new_value is greater than value of data in node,
+ # add it to right subtree and proceed recursively
+ node.right_child = insert(node.right_child, new_value)
+ return node
+
+
+def inorder(node: None | BinaryTreeNode) -> list[int]: # if node is None,return
+ """
+ >>> inorder(make_tree())
+ [6, 10, 14, 15, 20, 25, 60]
+ """
+ if node:
+ inorder_array = inorder(node.left_child)
+ inorder_array = [*inorder_array, node.data]
+ inorder_array = inorder_array + inorder(node.right_child)
+ else:
+ inorder_array = []
+ return inorder_array
+
+
+def make_tree() -> BinaryTreeNode | None:
+ root = insert(None, 15)
+ insert(root, 10)
+ insert(root, 25)
+ insert(root, 6)
+ insert(root, 14)
+ insert(root, 20)
+ insert(root, 60)
+ return root
+
+
+def main() -> None:
+ # main function
+ root = make_tree()
+ print("Printing values of binary search tree in Inorder Traversal.")
+ inorder(root)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ main()
diff --git a/data_structures/binary_tree/is_sorted.py b/data_structures/binary_tree/is_sorted.py
new file mode 100644
index 000000000000..91fc8ca82633
--- /dev/null
+++ b/data_structures/binary_tree/is_sorted.py
@@ -0,0 +1,98 @@
+"""
+Given the root of a binary tree, determine if it is a valid binary search tree (BST).
+
+A valid binary search tree is defined as follows:
+- The left subtree of a node contains only nodes with keys less than the node's key.
+- The right subtree of a node contains only nodes with keys greater than the node's key.
+- Both the left and right subtrees must also be binary search trees.
+
+In effect, a binary tree is a valid BST if its nodes are sorted in ascending order.
+leetcode: https://leetcode.com/problems/validate-binary-search-tree/
+
+If n is the number of nodes in the tree then:
+Runtime: O(n)
+Space: O(1)
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ data: float
+ left: Node | None = None
+ right: Node | None = None
+
+ def __iter__(self) -> Iterator[float]:
+ """
+ >>> root = Node(data=2.1)
+ >>> list(root)
+ [2.1]
+ >>> root.left=Node(data=2.0)
+ >>> list(root)
+ [2.0, 2.1]
+ >>> root.right=Node(data=2.2)
+ >>> list(root)
+ [2.0, 2.1, 2.2]
+ """
+ if self.left:
+ yield from self.left
+ yield self.data
+ if self.right:
+ yield from self.right
+
+ @property
+ def is_sorted(self) -> bool:
+ """
+ >>> Node(data='abc').is_sorted
+ True
+ >>> Node(data=2,
+ ... left=Node(data=1.999),
+ ... right=Node(data=3)).is_sorted
+ True
+ >>> Node(data=0,
+ ... left=Node(data=0),
+ ... right=Node(data=0)).is_sorted
+ True
+ >>> Node(data=0,
+ ... left=Node(data=-11),
+ ... right=Node(data=3)).is_sorted
+ True
+ >>> Node(data=5,
+ ... left=Node(data=1),
+ ... right=Node(data=4, left=Node(data=3))).is_sorted
+ False
+ >>> Node(data='a',
+ ... left=Node(data=1),
+ ... right=Node(data=4, left=Node(data=3))).is_sorted
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'str' and 'int'
+ >>> Node(data=2,
+ ... left=Node([]),
+ ... right=Node(data=4, left=Node(data=3))).is_sorted
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'int' and 'list'
+ """
+ if self.left and (self.data < self.left.data or not self.left.is_sorted):
+ return False
+ return not (
+ self.right and (self.data > self.right.data or not self.right.is_sorted)
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ tree = Node(data=2.1, left=Node(data=2.0), right=Node(data=2.2))
+ print(f"Tree {list(tree)} is sorted: {tree.is_sorted = }.")
+ assert tree.right
+ tree.right.data = 2.0
+ print(f"Tree {list(tree)} is sorted: {tree.is_sorted = }.")
+ tree.right.data = 2.1
+ print(f"Tree {list(tree)} is sorted: {tree.is_sorted = }.")
diff --git a/data_structures/binary_tree/is_sum_tree.py b/data_structures/binary_tree/is_sum_tree.py
new file mode 100644
index 000000000000..846bea0fe0f2
--- /dev/null
+++ b/data_structures/binary_tree/is_sum_tree.py
@@ -0,0 +1,162 @@
+"""
+Is a binary tree a sum tree where the value of every non-leaf node is equal to the sum
+of the values of its left and right subtrees?
+https://www.geeksforgeeks.org/check-if-a-given-binary-tree-is-sumtree
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ data: int
+ left: Node | None = None
+ right: Node | None = None
+
+ def __iter__(self) -> Iterator[int]:
+ """
+ >>> root = Node(2)
+ >>> list(root)
+ [2]
+ >>> root.left = Node(1)
+ >>> tuple(root)
+ (1, 2)
+ """
+ if self.left:
+ yield from self.left
+ yield self.data
+ if self.right:
+ yield from self.right
+
+ def __len__(self) -> int:
+ """
+ >>> root = Node(2)
+ >>> len(root)
+ 1
+ >>> root.left = Node(1)
+ >>> len(root)
+ 2
+ """
+ return sum(1 for _ in self)
+
+ @property
+ def is_sum_node(self) -> bool:
+ """
+ >>> root = Node(3)
+ >>> root.is_sum_node
+ True
+ >>> root.left = Node(1)
+ >>> root.is_sum_node
+ False
+ >>> root.right = Node(2)
+ >>> root.is_sum_node
+ True
+ """
+ if not self.left and not self.right:
+ return True # leaf nodes are considered sum nodes
+ left_sum = sum(self.left) if self.left else 0
+ right_sum = sum(self.right) if self.right else 0
+ return all(
+ (
+ self.data == left_sum + right_sum,
+ self.left.is_sum_node if self.left else True,
+ self.right.is_sum_node if self.right else True,
+ )
+ )
+
+
+@dataclass
+class BinaryTree:
+ root: Node
+
+ def __iter__(self) -> Iterator[int]:
+ """
+ >>> list(BinaryTree.build_a_tree())
+ [1, 2, 7, 11, 15, 29, 35, 40]
+ """
+ return iter(self.root)
+
+ def __len__(self) -> int:
+ """
+ >>> len(BinaryTree.build_a_tree())
+ 8
+ """
+ return len(self.root)
+
+ def __str__(self) -> str:
+ """
+ Returns a string representation of the inorder traversal of the binary tree.
+
+ >>> str(list(BinaryTree.build_a_tree()))
+ '[1, 2, 7, 11, 15, 29, 35, 40]'
+ """
+ return str(list(self))
+
+ @property
+ def is_sum_tree(self) -> bool:
+ """
+ >>> BinaryTree.build_a_tree().is_sum_tree
+ False
+ >>> BinaryTree.build_a_sum_tree().is_sum_tree
+ True
+ """
+ return self.root.is_sum_node
+
+ @classmethod
+ def build_a_tree(cls) -> BinaryTree:
+ r"""
+ Create a binary tree with the specified structure:
+ 11
+ / \
+ 2 29
+ / \ / \
+ 1 7 15 40
+ \
+ 35
+ >>> list(BinaryTree.build_a_tree())
+ [1, 2, 7, 11, 15, 29, 35, 40]
+ """
+ tree = BinaryTree(Node(11))
+ root = tree.root
+ root.left = Node(2)
+ root.right = Node(29)
+ root.left.left = Node(1)
+ root.left.right = Node(7)
+ root.right.left = Node(15)
+ root.right.right = Node(40)
+ root.right.right.left = Node(35)
+ return tree
+
+ @classmethod
+ def build_a_sum_tree(cls) -> BinaryTree:
+ r"""
+ Create a binary tree with the specified structure:
+ 26
+ / \
+ 10 3
+ / \ \
+ 4 6 3
+ >>> list(BinaryTree.build_a_sum_tree())
+ [4, 10, 6, 26, 3, 3]
+ """
+ tree = BinaryTree(Node(26))
+ root = tree.root
+ root.left = Node(10)
+ root.right = Node(3)
+ root.left.left = Node(4)
+ root.left.right = Node(6)
+ root.right.right = Node(3)
+ return tree
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ tree = BinaryTree.build_a_tree()
+ print(f"{tree} has {len(tree)} nodes and {tree.is_sum_tree = }.")
+ tree = BinaryTree.build_a_sum_tree()
+ print(f"{tree} has {len(tree)} nodes and {tree.is_sum_tree = }.")
diff --git a/data_structures/binary_tree/lazy_segment_tree.py b/data_structures/binary_tree/lazy_segment_tree.py
index 5bc79e74efcd..c26b0619380c 100644
--- a/data_structures/binary_tree/lazy_segment_tree.py
+++ b/data_structures/binary_tree/lazy_segment_tree.py
@@ -7,10 +7,10 @@ class SegmentTree:
def __init__(self, size: int) -> None:
self.size = size
# approximate the overall size of segment tree with given value
- self.segment_tree = [0 for i in range(0, 4 * size)]
+ self.segment_tree = [0 for i in range(4 * size)]
# create array to store lazy update
- self.lazy = [0 for i in range(0, 4 * size)]
- self.flag = [0 for i in range(0, 4 * size)] # flag for lazy update
+ self.lazy = [0 for i in range(4 * size)]
+ self.flag = [0 for i in range(4 * size)] # flag for lazy update
def left(self, idx: int) -> int:
"""
@@ -37,14 +37,14 @@ def right(self, idx: int) -> int:
return idx * 2 + 1
def build(
- self, idx: int, left_element: int, right_element: int, A: list[int]
+ self, idx: int, left_element: int, right_element: int, a: list[int]
) -> None:
if left_element == right_element:
- self.segment_tree[idx] = A[left_element - 1]
+ self.segment_tree[idx] = a[left_element - 1]
else:
mid = (left_element + right_element) // 2
- self.build(self.left(idx), left_element, mid, A)
- self.build(self.right(idx), mid + 1, right_element, A)
+ self.build(self.left(idx), left_element, mid, a)
+ self.build(self.right(idx), mid + 1, right_element, a)
self.segment_tree[idx] = max(
self.segment_tree[self.left(idx)], self.segment_tree[self.right(idx)]
)
@@ -88,7 +88,7 @@ def update(
# query with O(lg n)
def query(
self, idx: int, left_element: int, right_element: int, a: int, b: int
- ) -> int:
+ ) -> int | float:
"""
query(1, 1, size, a, b) for query max of [a,b]
>>> A = [1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8]
@@ -118,8 +118,8 @@ def query(
q2 = self.query(self.right(idx), mid + 1, right_element, a, b)
return max(q1, q2)
- def __str__(self) -> None:
- return [self.query(1, 1, self.size, i, i) for i in range(1, self.size + 1)]
+ def __str__(self) -> str:
+ return str([self.query(1, 1, self.size, i, i) for i in range(1, self.size + 1)])
if __name__ == "__main__":
diff --git a/data_structures/binary_tree/lowest_common_ancestor.py b/data_structures/binary_tree/lowest_common_ancestor.py
index 2f1e893fcf99..651037703b95 100644
--- a/data_structures/binary_tree/lowest_common_ancestor.py
+++ b/data_structures/binary_tree/lowest_common_ancestor.py
@@ -3,7 +3,7 @@
from __future__ import annotations
-import queue
+from queue import Queue
def swap(a: int, b: int) -> tuple[int, int]:
@@ -37,7 +37,7 @@ def create_sparse(max_node: int, parent: list[list[int]]) -> list[list[int]]:
# returns lca of node u,v
def lowest_common_ancestor(
u: int, v: int, level: list[int], parent: list[list[int]]
-) -> list[list[int]]:
+) -> int:
# u must be deeper in the tree than v
if level[u] < level[v]:
u, v = swap(u, v)
@@ -50,7 +50,7 @@ def lowest_common_ancestor(
return u
# moving both nodes upwards till lca in found
for i in range(18, -1, -1):
- if parent[i][u] != 0 and parent[i][u] != parent[i][v]:
+ if parent[i][u] not in [0, parent[i][v]]:
u, v = parent[i][u], parent[i][v]
# returning longest common ancestor of u,v
return parent[0][u]
@@ -61,8 +61,8 @@ def breadth_first_search(
level: list[int],
parent: list[list[int]],
max_node: int,
- graph: dict[int, int],
- root=1,
+ graph: dict[int, list[int]],
+ root: int = 1,
) -> tuple[list[int], list[list[int]]]:
"""
sets every nodes direct parent
@@ -70,7 +70,7 @@ def breadth_first_search(
calculates depth of each node from root node
"""
level[root] = 0
- q = queue.Queue(maxsize=max_node)
+ q: Queue[int] = Queue(maxsize=max_node)
q.put(root)
while q.qsize() != 0:
u = q.get()
@@ -88,7 +88,7 @@ def main() -> None:
parent = [[0 for _ in range(max_node + 10)] for _ in range(20)]
# initializing with -1 which means every node is unvisited
level = [-1 for _ in range(max_node + 10)]
- graph = {
+ graph: dict[int, list[int]] = {
1: [2, 3, 4],
2: [5],
3: [6, 7],
diff --git a/data_structures/binary_tree/maximum_fenwick_tree.py b/data_structures/binary_tree/maximum_fenwick_tree.py
new file mode 100644
index 000000000000..84967a70cc73
--- /dev/null
+++ b/data_structures/binary_tree/maximum_fenwick_tree.py
@@ -0,0 +1,114 @@
+class MaxFenwickTree:
+ """
+ Maximum Fenwick Tree
+
+ More info: https://cp-algorithms.com/data_structures/fenwick.html
+ ---------
+ >>> ft = MaxFenwickTree(5)
+ >>> ft.query(0, 5)
+ 0
+ >>> ft.update(4, 100)
+ >>> ft.query(0, 5)
+ 100
+ >>> ft.update(4, 0)
+ >>> ft.update(2, 20)
+ >>> ft.query(0, 5)
+ 20
+ >>> ft.update(4, 10)
+ >>> ft.query(2, 5)
+ 20
+ >>> ft.query(1, 5)
+ 20
+ >>> ft.update(2, 0)
+ >>> ft.query(0, 5)
+ 10
+ >>> ft = MaxFenwickTree(10000)
+ >>> ft.update(255, 30)
+ >>> ft.query(0, 10000)
+ 30
+ >>> ft = MaxFenwickTree(6)
+ >>> ft.update(5, 1)
+ >>> ft.query(5, 6)
+ 1
+ >>> ft = MaxFenwickTree(6)
+ >>> ft.update(0, 1000)
+ >>> ft.query(0, 1)
+ 1000
+ """
+
+ def __init__(self, size: int) -> None:
+ """
+ Create empty Maximum Fenwick Tree with specified size
+
+ Parameters:
+ size: size of Array
+
+ Returns:
+ None
+ """
+ self.size = size
+ self.arr = [0] * size
+ self.tree = [0] * size
+
+ @staticmethod
+ def get_next(index: int) -> int:
+ """
+ Get next index in O(1)
+ """
+ return index | (index + 1)
+
+ @staticmethod
+ def get_prev(index: int) -> int:
+ """
+ Get previous index in O(1)
+ """
+ return (index & (index + 1)) - 1
+
+ def update(self, index: int, value: int) -> None:
+ """
+ Set index to value in O(lg^2 N)
+
+ Parameters:
+ index: index to update
+ value: value to set
+
+ Returns:
+ None
+ """
+ self.arr[index] = value
+ while index < self.size:
+ current_left_border = self.get_prev(index) + 1
+ if current_left_border == index:
+ self.tree[index] = value
+ else:
+ self.tree[index] = max(value, current_left_border, index)
+ index = self.get_next(index)
+
+ def query(self, left: int, right: int) -> int:
+ """
+ Answer the query of maximum range [l, r) in O(lg^2 N)
+
+ Parameters:
+ left: left index of query range (inclusive)
+ right: right index of query range (exclusive)
+
+ Returns:
+ Maximum value of range [left, right)
+ """
+ right -= 1 # Because of right is exclusive
+ result = 0
+ while left <= right:
+ current_left = self.get_prev(right)
+ if left <= current_left:
+ result = max(result, self.tree[right])
+ right = current_left
+ else:
+ result = max(result, self.arr[right])
+ right -= 1
+ return result
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/maximum_sum_bst.py b/data_structures/binary_tree/maximum_sum_bst.py
new file mode 100644
index 000000000000..7dadc7b95920
--- /dev/null
+++ b/data_structures/binary_tree/maximum_sum_bst.py
@@ -0,0 +1,78 @@
+from __future__ import annotations
+
+import sys
+from dataclasses import dataclass
+
+INT_MIN = -sys.maxsize + 1
+INT_MAX = sys.maxsize - 1
+
+
+@dataclass
+class TreeNode:
+ val: int = 0
+ left: TreeNode | None = None
+ right: TreeNode | None = None
+
+
+def max_sum_bst(root: TreeNode | None) -> int:
+ """
+ The solution traverses a binary tree to find the maximum sum of
+ keys in any subtree that is a Binary Search Tree (BST). It uses
+ recursion to validate BST properties and calculates sums, returning
+ the highest sum found among all valid BST subtrees.
+
+ >>> t1 = TreeNode(4)
+ >>> t1.left = TreeNode(3)
+ >>> t1.left.left = TreeNode(1)
+ >>> t1.left.right = TreeNode(2)
+ >>> print(max_sum_bst(t1))
+ 2
+ >>> t2 = TreeNode(-4)
+ >>> t2.left = TreeNode(-2)
+ >>> t2.right = TreeNode(-5)
+ >>> print(max_sum_bst(t2))
+ 0
+ >>> t3 = TreeNode(1)
+ >>> t3.left = TreeNode(4)
+ >>> t3.left.left = TreeNode(2)
+ >>> t3.left.right = TreeNode(4)
+ >>> t3.right = TreeNode(3)
+ >>> t3.right.left = TreeNode(2)
+ >>> t3.right.right = TreeNode(5)
+ >>> t3.right.right.left = TreeNode(4)
+ >>> t3.right.right.right = TreeNode(6)
+ >>> print(max_sum_bst(t3))
+ 20
+ """
+ ans: int = 0
+
+ def solver(node: TreeNode | None) -> tuple[bool, int, int, int]:
+ """
+ Returns the maximum sum by making recursive calls
+ >>> t1 = TreeNode(1)
+ >>> print(solver(t1))
+ 1
+ """
+ nonlocal ans
+
+ if not node:
+ return True, INT_MAX, INT_MIN, 0 # Valid BST, min, max, sum
+
+ is_left_valid, min_left, max_left, sum_left = solver(node.left)
+ is_right_valid, min_right, max_right, sum_right = solver(node.right)
+
+ if is_left_valid and is_right_valid and max_left < node.val < min_right:
+ total_sum = sum_left + sum_right + node.val
+ ans = max(ans, total_sum)
+ return True, min(min_left, node.val), max(max_right, node.val), total_sum
+
+ return False, -1, -1, -1 # Not a valid BST
+
+ solver(root)
+ return ans
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/merge_two_binary_trees.py b/data_structures/binary_tree/merge_two_binary_trees.py
new file mode 100644
index 000000000000..6bbb30428704
--- /dev/null
+++ b/data_structures/binary_tree/merge_two_binary_trees.py
@@ -0,0 +1,94 @@
+#!/usr/local/bin/python3
+"""
+Problem Description: Given two binary tree, return the merged tree.
+The rule for merging is that if two nodes overlap, then put the value sum of
+both nodes to the new value of the merged node. Otherwise, the NOT null node
+will be used as the node of new tree.
+"""
+
+from __future__ import annotations
+
+
+class Node:
+ """
+ A binary node has value variable and pointers to its left and right node.
+ """
+
+ def __init__(self, value: int = 0) -> None:
+ self.value = value
+ self.left: Node | None = None
+ self.right: Node | None = None
+
+
+def merge_two_binary_trees(tree1: Node | None, tree2: Node | None) -> Node | None:
+ """
+ Returns root node of the merged tree.
+
+ >>> tree1 = Node(5)
+ >>> tree1.left = Node(6)
+ >>> tree1.right = Node(7)
+ >>> tree1.left.left = Node(2)
+ >>> tree2 = Node(4)
+ >>> tree2.left = Node(5)
+ >>> tree2.right = Node(8)
+ >>> tree2.left.right = Node(1)
+ >>> tree2.right.right = Node(4)
+ >>> merged_tree = merge_two_binary_trees(tree1, tree2)
+ >>> print_preorder(merged_tree)
+ 9
+ 11
+ 2
+ 1
+ 15
+ 4
+ """
+ if tree1 is None:
+ return tree2
+ if tree2 is None:
+ return tree1
+
+ tree1.value = tree1.value + tree2.value
+ tree1.left = merge_two_binary_trees(tree1.left, tree2.left)
+ tree1.right = merge_two_binary_trees(tree1.right, tree2.right)
+ return tree1
+
+
+def print_preorder(root: Node | None) -> None:
+ """
+ Print pre-order traversal of the tree.
+
+ >>> root = Node(1)
+ >>> root.left = Node(2)
+ >>> root.right = Node(3)
+ >>> print_preorder(root)
+ 1
+ 2
+ 3
+ >>> print_preorder(root.right)
+ 3
+ """
+ if root:
+ print(root.value)
+ print_preorder(root.left)
+ print_preorder(root.right)
+
+
+if __name__ == "__main__":
+ tree1 = Node(1)
+ tree1.left = Node(2)
+ tree1.right = Node(3)
+ tree1.left.left = Node(4)
+
+ tree2 = Node(2)
+ tree2.left = Node(4)
+ tree2.right = Node(6)
+ tree2.left.right = Node(9)
+ tree2.right.right = Node(5)
+
+ print("Tree1 is: ")
+ print_preorder(tree1)
+ print("Tree2 is: ")
+ print_preorder(tree2)
+ merged_tree = merge_two_binary_trees(tree1, tree2)
+ print("Merged Tree is: ")
+ print_preorder(merged_tree)
diff --git a/data_structures/binary_tree/mirror_binary_tree.py b/data_structures/binary_tree/mirror_binary_tree.py
new file mode 100644
index 000000000000..f6611d66d676
--- /dev/null
+++ b/data_structures/binary_tree/mirror_binary_tree.py
@@ -0,0 +1,160 @@
+"""
+Given the root of a binary tree, mirror the tree, and return its root.
+
+Leetcode problem reference: https://leetcode.com/problems/mirror-binary-tree/
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ """
+ A Node has value variable and pointers to Nodes to its left and right.
+ """
+
+ value: int
+ left: Node | None = None
+ right: Node | None = None
+
+ def __iter__(self) -> Iterator[int]:
+ if self.left:
+ yield from self.left
+ yield self.value
+ if self.right:
+ yield from self.right
+
+ def __len__(self) -> int:
+ return sum(1 for _ in self)
+
+ def mirror(self) -> Node:
+ """
+ Mirror the binary tree rooted at this node by swapping left and right children.
+
+ >>> tree = Node(0)
+ >>> list(tree)
+ [0]
+ >>> list(tree.mirror())
+ [0]
+ >>> tree = Node(1, Node(0), Node(3, Node(2), Node(4, None, Node(5))))
+ >>> tuple(tree)
+ (0, 1, 2, 3, 4, 5)
+ >>> tuple(tree.mirror())
+ (5, 4, 3, 2, 1, 0)
+ """
+ self.left, self.right = self.right, self.left
+ if self.left:
+ self.left.mirror()
+ if self.right:
+ self.right.mirror()
+ return self
+
+
+def make_tree_seven() -> Node:
+ r"""
+ Return a binary tree with 7 nodes that looks like this:
+ ::
+
+ 1
+ / \
+ 2 3
+ / \ / \
+ 4 5 6 7
+
+ >>> tree_seven = make_tree_seven()
+ >>> len(tree_seven)
+ 7
+ >>> list(tree_seven)
+ [4, 2, 5, 1, 6, 3, 7]
+ """
+ tree = Node(1)
+ tree.left = Node(2)
+ tree.right = Node(3)
+ tree.left.left = Node(4)
+ tree.left.right = Node(5)
+ tree.right.left = Node(6)
+ tree.right.right = Node(7)
+ return tree
+
+
+def make_tree_nine() -> Node:
+ r"""
+ Return a binary tree with 9 nodes that looks like this:
+ ::
+
+ 1
+ / \
+ 2 3
+ / \ \
+ 4 5 6
+ / \ \
+ 7 8 9
+
+ >>> tree_nine = make_tree_nine()
+ >>> len(tree_nine)
+ 9
+ >>> list(tree_nine)
+ [7, 4, 8, 2, 5, 9, 1, 3, 6]
+ """
+ tree = Node(1)
+ tree.left = Node(2)
+ tree.right = Node(3)
+ tree.left.left = Node(4)
+ tree.left.right = Node(5)
+ tree.right.right = Node(6)
+ tree.left.left.left = Node(7)
+ tree.left.left.right = Node(8)
+ tree.left.right.right = Node(9)
+ return tree
+
+
+def main() -> None:
+ r"""
+ Mirror binary trees with the given root and returns the root
+
+ >>> tree = make_tree_nine()
+ >>> tuple(tree)
+ (7, 4, 8, 2, 5, 9, 1, 3, 6)
+ >>> tuple(tree.mirror())
+ (6, 3, 1, 9, 5, 2, 8, 4, 7)
+
+ nine_tree::
+
+ 1
+ / \
+ 2 3
+ / \ \
+ 4 5 6
+ / \ \
+ 7 8 9
+
+ The mirrored tree looks like this::
+
+ 1
+ / \
+ 3 2
+ / / \
+ 6 5 4
+ / / \
+ 9 8 7
+ """
+ trees = {"zero": Node(0), "seven": make_tree_seven(), "nine": make_tree_nine()}
+ for name, tree in trees.items():
+ print(f" The {name} tree: {tuple(tree)}")
+ # (0,)
+ # (4, 2, 5, 1, 6, 3, 7)
+ # (7, 4, 8, 2, 5, 9, 1, 3, 6)
+ print(f"Mirror of {name} tree: {tuple(tree.mirror())}")
+ # (0,)
+ # (7, 3, 6, 1, 5, 2, 4)
+ # (6, 3, 1, 9, 5, 2, 8, 4, 7)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ main()
diff --git a/data_structures/binary_tree/non_recursive_segment_tree.py b/data_structures/binary_tree/non_recursive_segment_tree.py
index 064e5aded7b4..ca0d5c111c4f 100644
--- a/data_structures/binary_tree/non_recursive_segment_tree.py
+++ b/data_structures/binary_tree/non_recursive_segment_tree.py
@@ -35,28 +35,32 @@
>>> st.query(0, 2)
[1, 2, 3]
"""
+
from __future__ import annotations
-from typing import Callable, TypeVar
+from collections.abc import Callable
+from typing import Any, Generic, TypeVar
T = TypeVar("T")
-class SegmentTree:
+class SegmentTree(Generic[T]):
def __init__(self, arr: list[T], fnc: Callable[[T, T], T]) -> None:
"""
Segment Tree constructor, it works just with commutative combiner.
:param arr: list of elements for the segment tree
:param fnc: commutative function for combine two elements
- >>> SegmentTree(['a', 'b', 'c'], lambda a, b: '{}{}'.format(a, b)).query(0, 2)
+ >>> SegmentTree(['a', 'b', 'c'], lambda a, b: f'{a}{b}').query(0, 2)
'abc'
>>> SegmentTree([(1, 2), (2, 3), (3, 4)],
... lambda a, b: (a[0] + b[0], a[1] + b[1])).query(0, 2)
(6, 9)
"""
- self.N = len(arr)
- self.st = [None for _ in range(len(arr))] + arr
+ any_type: Any | T = None
+
+ self.N: int = len(arr)
+ self.st: list[T] = [any_type for _ in range(self.N)] + arr
self.fn = fnc
self.build()
@@ -83,12 +87,12 @@ def update(self, p: int, v: T) -> None:
p = p // 2
self.st[p] = self.fn(self.st[p * 2], self.st[p * 2 + 1])
- def query(self, l: int, r: int) -> T: # noqa: E741
+ def query(self, left: int, right: int) -> T | None:
"""
Get range query value in log(N) time
- :param l: left element index
- :param r: right element index
- :return: element combined in the range [l, r]
+ :param left: left element index
+ :param right: right element index
+ :return: element combined in the range [left, right]
>>> st = SegmentTree([1, 2, 3, 4], lambda a, b: a + b)
>>> st.query(0, 2)
@@ -100,14 +104,15 @@ def query(self, l: int, r: int) -> T: # noqa: E741
>>> st.query(2, 3)
7
"""
- l, r = l + self.N, r + self.N # noqa: E741
- res = None
- while l <= r: # noqa: E741
- if l % 2 == 1:
- res = self.st[l] if res is None else self.fn(res, self.st[l])
- if r % 2 == 0:
- res = self.st[r] if res is None else self.fn(res, self.st[r])
- l, r = (l + 1) // 2, (r - 1) // 2
+ left, right = left + self.N, right + self.N
+
+ res: T | None = None
+ while left <= right:
+ if left % 2 == 1:
+ res = self.st[left] if res is None else self.fn(res, self.st[left])
+ if right % 2 == 0:
+ res = self.st[right] if res is None else self.fn(res, self.st[right])
+ left, right = (left + 1) // 2, (right - 1) // 2
return res
@@ -135,7 +140,7 @@ def query(self, l: int, r: int) -> T: # noqa: E741
max_segment_tree = SegmentTree(test_array, max)
sum_segment_tree = SegmentTree(test_array, lambda a, b: a + b)
- def test_all_segments():
+ def test_all_segments() -> None:
"""
Test all possible segments
"""
diff --git a/data_structures/binary_tree/number_of_possible_binary_trees.py b/data_structures/binary_tree/number_of_possible_binary_trees.py
index 1ad8f2ed4287..b39cbafd0a61 100644
--- a/data_structures/binary_tree/number_of_possible_binary_trees.py
+++ b/data_structures/binary_tree/number_of_possible_binary_trees.py
@@ -6,6 +6,7 @@
Further details at Wikipedia: https://en.wikipedia.org/wiki/Catalan_number
"""
+
"""
Our Contribution:
Basically we Create the 2 function:
@@ -30,8 +31,7 @@ def binomial_coefficient(n: int, k: int) -> int:
"""
result = 1 # To kept the Calculated Value
# Since C(n, k) = C(n, n-k)
- if k > (n - k):
- k = n - k
+ k = min(k, n - k)
# Calculate C(n,k)
for i in range(k):
result *= n - i
@@ -67,7 +67,7 @@ def factorial(n: int) -> int:
True
>>> factorial(-5) # doctest: +ELLIPSIS
Traceback (most recent call last):
- ...
+ ...
ValueError: factorial() not defined for negative values
"""
if n < 0:
diff --git a/data_structures/binary_tree/red_black_tree.py b/data_structures/binary_tree/red_black_tree.py
index 5d721edfa45b..752db1e7026c 100644
--- a/data_structures/binary_tree/red_black_tree.py
+++ b/data_structures/binary_tree/red_black_tree.py
@@ -1,8 +1,6 @@
-"""
-python/black : true
-flake8 : passed
-"""
-from typing import Iterator, Optional
+from __future__ import annotations
+
+from collections.abc import Iterator
class RedBlackTree:
@@ -14,18 +12,18 @@ class RedBlackTree:
and slower for reading in the average case, though, because they're
both balanced binary search trees, both will get the same asymptotic
performance.
- To read more about them, https://en.wikipedia.org/wiki/Red–black_tree
+ To read more about them, https://en.wikipedia.org/wiki/Red-black_tree
Unless otherwise specified, all asymptotic runtimes are specified in
terms of the size of the tree.
"""
def __init__(
self,
- label: Optional[int] = None,
+ label: int | None = None,
color: int = 0,
- parent: Optional["RedBlackTree"] = None,
- left: Optional["RedBlackTree"] = None,
- right: Optional["RedBlackTree"] = None,
+ parent: RedBlackTree | None = None,
+ left: RedBlackTree | None = None,
+ right: RedBlackTree | None = None,
) -> None:
"""Initialize a new Red-Black Tree node with the given values:
label: The value associated with this node
@@ -42,13 +40,15 @@ def __init__(
# Here are functions which are specific to red-black trees
- def rotate_left(self) -> "RedBlackTree":
+ def rotate_left(self) -> RedBlackTree:
"""Rotate the subtree rooted at this node to the left and
returns the new root to this subtree.
Performing one rotation can be done in O(1).
"""
parent = self.parent
right = self.right
+ if right is None:
+ return self
self.right = right.left
if self.right:
self.right.parent = self
@@ -62,11 +62,13 @@ def rotate_left(self) -> "RedBlackTree":
right.parent = parent
return right
- def rotate_right(self) -> "RedBlackTree":
+ def rotate_right(self) -> RedBlackTree:
"""Rotate the subtree rooted at this node to the right and
returns the new root to this subtree.
Performing one rotation can be done in O(1).
"""
+ if self.left is None:
+ return self
parent = self.parent
left = self.left
self.left = left.right
@@ -82,7 +84,7 @@ def rotate_right(self) -> "RedBlackTree":
left.parent = parent
return left
- def insert(self, label: int) -> "RedBlackTree":
+ def insert(self, label: int) -> RedBlackTree:
"""Inserts label into the subtree rooted at self, performs any
rotations necessary to maintain balance, and then returns the
new root to this subtree (likely self).
@@ -100,12 +102,11 @@ def insert(self, label: int) -> "RedBlackTree":
else:
self.left = RedBlackTree(label, 1, self)
self.left._insert_repair()
+ elif self.right:
+ self.right.insert(label)
else:
- if self.right:
- self.right.insert(label)
- else:
- self.right = RedBlackTree(label, 1, self)
- self.right._insert_repair()
+ self.right = RedBlackTree(label, 1, self)
+ self.right._insert_repair()
return self.parent or self
def _insert_repair(self) -> None:
@@ -121,25 +122,32 @@ def _insert_repair(self) -> None:
if color(uncle) == 0:
if self.is_left() and self.parent.is_right():
self.parent.rotate_right()
- self.right._insert_repair()
+ if self.right:
+ self.right._insert_repair()
elif self.is_right() and self.parent.is_left():
self.parent.rotate_left()
- self.left._insert_repair()
+ if self.left:
+ self.left._insert_repair()
elif self.is_left():
- self.grandparent.rotate_right()
- self.parent.color = 0
- self.parent.right.color = 1
+ if self.grandparent:
+ self.grandparent.rotate_right()
+ self.parent.color = 0
+ if self.parent.right:
+ self.parent.right.color = 1
else:
- self.grandparent.rotate_left()
- self.parent.color = 0
- self.parent.left.color = 1
+ if self.grandparent:
+ self.grandparent.rotate_left()
+ self.parent.color = 0
+ if self.parent.left:
+ self.parent.left.color = 1
else:
self.parent.color = 0
- uncle.color = 0
- self.grandparent.color = 1
- self.grandparent._insert_repair()
+ if uncle and self.grandparent:
+ uncle.color = 0
+ self.grandparent.color = 1
+ self.grandparent._insert_repair()
- def remove(self, label: int) -> "RedBlackTree":
+ def remove(self, label: int) -> RedBlackTree:
"""Remove label from this tree."""
if self.label == label:
if self.left and self.right:
@@ -147,8 +155,9 @@ def remove(self, label: int) -> "RedBlackTree":
# so we replace this node with the greatest one less than
# it and remove that.
value = self.left.get_max()
- self.label = value
- self.left.remove(value)
+ if value is not None:
+ self.label = value
+ self.left.remove(value)
else:
# This node has at most one non-None child, so we don't
# need to replace
@@ -158,44 +167,50 @@ def remove(self, label: int) -> "RedBlackTree":
# The only way this happens to a node with one child
# is if both children are None leaves.
# We can just remove this node and call it a day.
- if self.is_left():
- self.parent.left = None
- else:
- self.parent.right = None
- else:
- # The node is black
- if child is None:
- # This node and its child are black
- if self.parent is None:
- # The tree is now empty
- return RedBlackTree(None)
+ if self.parent:
+ if self.is_left():
+ self.parent.left = None
else:
- self._remove_repair()
- if self.is_left():
- self.parent.left = None
- else:
- self.parent.right = None
- self.parent = None
+ self.parent.right = None
+ # The node is black
+ elif child is None:
+ # This node and its child are black
+ if self.parent is None:
+ # The tree is now empty
+ return RedBlackTree(None)
else:
- # This node is black and its child is red
- # Move the child node here and make it black
- self.label = child.label
- self.left = child.left
- self.right = child.right
- if self.left:
- self.left.parent = self
- if self.right:
- self.right.parent = self
- elif self.label > label:
+ self._remove_repair()
+ if self.is_left():
+ self.parent.left = None
+ else:
+ self.parent.right = None
+ self.parent = None
+ else:
+ # This node is black and its child is red
+ # Move the child node here and make it black
+ self.label = child.label
+ self.left = child.left
+ self.right = child.right
+ if self.left:
+ self.left.parent = self
+ if self.right:
+ self.right.parent = self
+ elif self.label is not None and self.label > label:
if self.left:
self.left.remove(label)
- else:
- if self.right:
- self.right.remove(label)
+ elif self.right:
+ self.right.remove(label)
return self.parent or self
def _remove_repair(self) -> None:
"""Repair the coloring of the tree that may have been messed up."""
+ if (
+ self.parent is None
+ or self.sibling is None
+ or self.parent.sibling is None
+ or self.grandparent is None
+ ):
+ return
if color(self.sibling) == 1:
self.sibling.color = 0
self.parent.color = 1
@@ -229,7 +244,8 @@ def _remove_repair(self) -> None:
):
self.sibling.rotate_right()
self.sibling.color = 0
- self.sibling.right.color = 1
+ if self.sibling.right:
+ self.sibling.right.color = 1
if (
self.is_right()
and color(self.sibling) == 0
@@ -238,7 +254,8 @@ def _remove_repair(self) -> None:
):
self.sibling.rotate_left()
self.sibling.color = 0
- self.sibling.left.color = 1
+ if self.sibling.left:
+ self.sibling.left.color = 1
if (
self.is_left()
and color(self.sibling) == 0
@@ -273,21 +290,17 @@ def check_color_properties(self) -> bool:
"""
# I assume property 1 to hold because there is nothing that can
# make the color be anything other than 0 or 1.
-
# Property 2
if self.color:
# The root was red
print("Property 2")
return False
-
# Property 3 does not need to be checked, because None is assumed
# to be black and is all the leaves.
-
# Property 4
if not self.check_coloring():
print("Property 4")
return False
-
# Property 5
if self.black_height() is None:
print("Property 5")
@@ -295,25 +308,22 @@ def check_color_properties(self) -> bool:
# All properties were met
return True
- def check_coloring(self) -> None:
+ def check_coloring(self) -> bool:
"""A helper function to recursively check Property 4 of a
Red-Black Tree. See check_color_properties for more info.
"""
- if self.color == 1:
- if color(self.left) == 1 or color(self.right) == 1:
- return False
- if self.left and not self.left.check_coloring():
+ if self.color == 1 and 1 in (color(self.left), color(self.right)):
return False
- if self.right and not self.right.check_coloring():
+ if self.left and not self.left.check_coloring():
return False
- return True
+ return not (self.right and not self.right.check_coloring())
- def black_height(self) -> int:
+ def black_height(self) -> int | None:
"""Returns the number of black nodes from this node to the
leaves of the tree, or None if there isn't one such value (the
tree is color incorrectly).
"""
- if self is None:
+ if self is None or self.left is None or self.right is None:
# If we're already at a leaf, there is no path
return 1
left = RedBlackTree.black_height(self.left)
@@ -330,37 +340,36 @@ def black_height(self) -> int:
# Here are functions which are general to all binary search trees
- def __contains__(self, label) -> bool:
+ def __contains__(self, label: int) -> bool:
"""Search through the tree for label, returning True iff it is
found somewhere in the tree.
Guaranteed to run in O(log(n)) time.
"""
return self.search(label) is not None
- def search(self, label: int) -> "RedBlackTree":
+ def search(self, label: int) -> RedBlackTree | None:
"""Search through the tree for label, returning its node if
it's found, and None otherwise.
This method is guaranteed to run in O(log(n)) time.
"""
if self.label == label:
return self
- elif label > self.label:
+ elif self.label is not None and label > self.label:
if self.right is None:
return None
else:
return self.right.search(label)
+ elif self.left is None:
+ return None
else:
- if self.left is None:
- return None
- else:
- return self.left.search(label)
+ return self.left.search(label)
- def floor(self, label: int) -> int:
+ def floor(self, label: int) -> int | None:
"""Returns the largest element in this tree which is at most label.
This method is guaranteed to run in O(log(n)) time."""
if self.label == label:
return self.label
- elif self.label > label:
+ elif self.label is not None and self.label > label:
if self.left:
return self.left.floor(label)
else:
@@ -372,13 +381,13 @@ def floor(self, label: int) -> int:
return attempt
return self.label
- def ceil(self, label: int) -> int:
+ def ceil(self, label: int) -> int | None:
"""Returns the smallest element in this tree which is at least label.
This method is guaranteed to run in O(log(n)) time.
"""
if self.label == label:
return self.label
- elif self.label < label:
+ elif self.label is not None and self.label < label:
if self.right:
return self.right.ceil(label)
else:
@@ -390,7 +399,7 @@ def ceil(self, label: int) -> int:
return attempt
return self.label
- def get_max(self) -> int:
+ def get_max(self) -> int | None:
"""Returns the largest element in this tree.
This method is guaranteed to run in O(log(n)) time.
"""
@@ -400,7 +409,7 @@ def get_max(self) -> int:
else:
return self.label
- def get_min(self) -> int:
+ def get_min(self) -> int | None:
"""Returns the smallest element in this tree.
This method is guaranteed to run in O(log(n)) time.
"""
@@ -411,7 +420,7 @@ def get_min(self) -> int:
return self.label
@property
- def grandparent(self) -> "RedBlackTree":
+ def grandparent(self) -> RedBlackTree | None:
"""Get the current node's grandparent, or None if it doesn't exist."""
if self.parent is None:
return None
@@ -419,7 +428,7 @@ def grandparent(self) -> "RedBlackTree":
return self.parent.parent
@property
- def sibling(self) -> "RedBlackTree":
+ def sibling(self) -> RedBlackTree | None:
"""Get the current node's sibling, or None if it doesn't exist."""
if self.parent is None:
return None
@@ -430,11 +439,15 @@ def sibling(self) -> "RedBlackTree":
def is_left(self) -> bool:
"""Returns true iff this node is the left child of its parent."""
- return self.parent and self.parent.left is self
+ if self.parent is None:
+ return False
+ return self.parent.left is self
def is_right(self) -> bool:
"""Returns true iff this node is the right child of its parent."""
- return self.parent and self.parent.right is self
+ if self.parent is None:
+ return False
+ return self.parent.right is self
def __bool__(self) -> bool:
return True
@@ -450,21 +463,21 @@ def __len__(self) -> int:
ln += len(self.right)
return ln
- def preorder_traverse(self) -> Iterator[int]:
+ def preorder_traverse(self) -> Iterator[int | None]:
yield self.label
if self.left:
yield from self.left.preorder_traverse()
if self.right:
yield from self.right.preorder_traverse()
- def inorder_traverse(self) -> Iterator[int]:
+ def inorder_traverse(self) -> Iterator[int | None]:
if self.left:
yield from self.left.inorder_traverse()
yield self.label
if self.right:
yield from self.right.inorder_traverse()
- def postorder_traverse(self) -> Iterator[int]:
+ def postorder_traverse(self) -> Iterator[int | None]:
if self.left:
yield from self.left.postorder_traverse()
if self.right:
@@ -475,24 +488,28 @@ def __repr__(self) -> str:
from pprint import pformat
if self.left is None and self.right is None:
- return "'{} {}'".format(self.label, (self.color and "red") or "blk")
+ return f"'{self.label} {(self.color and 'red') or 'blk'}'"
return pformat(
{
- "%s %s"
- % (self.label, (self.color and "red") or "blk"): (self.left, self.right)
+ f"{self.label} {(self.color and 'red') or 'blk'}": (
+ self.left,
+ self.right,
+ )
},
indent=1,
)
- def __eq__(self, other) -> bool:
+ def __eq__(self, other: object) -> bool:
"""Test if two trees are equal."""
+ if not isinstance(other, RedBlackTree):
+ return NotImplemented
if self.label == other.label:
return self.left == other.left and self.right == other.right
else:
return False
-def color(node) -> int:
+def color(node: RedBlackTree | None) -> int:
"""Returns the color of a node, allowing for None leaves."""
if node is None:
return 0
@@ -537,9 +554,7 @@ def test_rotations() -> bool:
right_rot.right.right = RedBlackTree(10, parent=right_rot.right)
right_rot.right.right.left = RedBlackTree(5, parent=right_rot.right.right)
right_rot.right.right.right = RedBlackTree(20, parent=right_rot.right.right)
- if tree != right_rot:
- return False
- return True
+ return tree == right_rot
def test_insertion_speed() -> bool:
@@ -582,13 +597,11 @@ def test_insert_and_search() -> bool:
tree.insert(12)
tree.insert(10)
tree.insert(11)
- if 5 in tree or -6 in tree or -10 in tree or 13 in tree:
+ if any(i in tree for i in (5, -6, -10, 13)):
# Found something not in there
return False
- if not (11 in tree and 12 in tree and -8 in tree and 0 in tree):
- # Didn't find something in there
- return False
- return True
+ # Find all these things in there
+ return all(i in tree for i in (11, 12, -8, 0))
def test_insert_delete() -> bool:
@@ -610,9 +623,7 @@ def test_insert_delete() -> bool:
tree = tree.remove(9)
if not tree.check_color_properties():
return False
- if list(tree.inorder_traverse()) != [-8, 0, 4, 8, 10, 11, 12]:
- return False
- return True
+ return list(tree.inorder_traverse()) == [-8, 0, 4, 8, 10, 11, 12]
def test_floor_ceil() -> bool:
@@ -640,9 +651,7 @@ def test_min_max() -> bool:
tree.insert(24)
tree.insert(20)
tree.insert(22)
- if tree.get_max() != 22 or tree.get_min() != -16:
- return False
- return True
+ return not (tree.get_max() != 22 or tree.get_min() != -16)
def test_tree_traversal() -> bool:
@@ -658,9 +667,7 @@ def test_tree_traversal() -> bool:
return False
if list(tree.preorder_traverse()) != [0, -16, 16, 8, 22, 20, 24]:
return False
- if list(tree.postorder_traverse()) != [-16, 8, 20, 24, 22, 16, 0]:
- return False
- return True
+ return list(tree.postorder_traverse()) == [-16, 8, 20, 24, 22, 16, 0]
def test_tree_chaining() -> bool:
@@ -671,9 +678,7 @@ def test_tree_chaining() -> bool:
return False
if list(tree.preorder_traverse()) != [0, -16, 16, 8, 22, 20, 24]:
return False
- if list(tree.postorder_traverse()) != [-16, 8, 20, 24, 22, 16, 0]:
- return False
- return True
+ return list(tree.postorder_traverse()) == [-16, 8, 20, 24, 22, 16, 0]
def print_results(msg: str, passes: bool) -> None:
@@ -695,19 +700,12 @@ def main() -> None:
>>> pytests()
"""
print_results("Rotating right and left", test_rotations())
-
print_results("Inserting", test_insert())
-
print_results("Searching", test_insert_and_search())
-
print_results("Deleting", test_insert_delete())
-
print_results("Floor and ceil", test_floor_ceil())
-
print_results("Tree traversal", test_tree_traversal())
-
print_results("Tree traversal", test_tree_chaining())
-
print("Testing tree balancing...")
print("This should only be a few seconds.")
test_insertion_speed()
diff --git a/data_structures/binary_tree/segment_tree.py b/data_structures/binary_tree/segment_tree.py
index 10451ae68bb2..084fcf84955d 100644
--- a/data_structures/binary_tree/segment_tree.py
+++ b/data_structures/binary_tree/segment_tree.py
@@ -2,67 +2,105 @@
class SegmentTree:
- def __init__(self, A):
- self.N = len(A)
+ def __init__(self, a):
+ self.A = a
+ self.N = len(self.A)
self.st = [0] * (
4 * self.N
) # approximate the overall size of segment tree with array N
- self.build(1, 0, self.N - 1)
+ if self.N:
+ self.build(1, 0, self.N - 1)
def left(self, idx):
+ """
+ Returns the left child index for a given index in a binary tree.
+
+ >>> s = SegmentTree([1, 2, 3])
+ >>> s.left(1)
+ 2
+ >>> s.left(2)
+ 4
+ """
return idx * 2
def right(self, idx):
+ """
+ Returns the right child index for a given index in a binary tree.
+
+ >>> s = SegmentTree([1, 2, 3])
+ >>> s.right(1)
+ 3
+ >>> s.right(2)
+ 5
+ """
return idx * 2 + 1
- def build(self, idx, l, r): # noqa: E741
- if l == r: # noqa: E741
- self.st[idx] = A[l]
+ def build(self, idx, left, right):
+ if left == right:
+ self.st[idx] = self.A[left]
else:
- mid = (l + r) // 2
- self.build(self.left(idx), l, mid)
- self.build(self.right(idx), mid + 1, r)
+ mid = (left + right) // 2
+ self.build(self.left(idx), left, mid)
+ self.build(self.right(idx), mid + 1, right)
self.st[idx] = max(self.st[self.left(idx)], self.st[self.right(idx)])
def update(self, a, b, val):
+ """
+ Update the values in the segment tree in the range [a,b] with the given value.
+
+ >>> s = SegmentTree([1, 2, 3, 4, 5])
+ >>> s.update(2, 4, 10)
+ True
+ >>> s.query(1, 5)
+ 10
+ """
return self.update_recursive(1, 0, self.N - 1, a - 1, b - 1, val)
- def update_recursive(self, idx, l, r, a, b, val): # noqa: E741
+ def update_recursive(self, idx, left, right, a, b, val):
"""
update(1, 1, N, a, b, v) for update val v to [a,b]
"""
- if r < a or l > b:
+ if right < a or left > b:
return True
- if l == r: # noqa: E741
+ if left == right:
self.st[idx] = val
return True
- mid = (l + r) // 2
- self.update_recursive(self.left(idx), l, mid, a, b, val)
- self.update_recursive(self.right(idx), mid + 1, r, a, b, val)
+ mid = (left + right) // 2
+ self.update_recursive(self.left(idx), left, mid, a, b, val)
+ self.update_recursive(self.right(idx), mid + 1, right, a, b, val)
self.st[idx] = max(self.st[self.left(idx)], self.st[self.right(idx)])
return True
def query(self, a, b):
+ """
+ Query the maximum value in the range [a,b].
+
+ >>> s = SegmentTree([1, 2, 3, 4, 5])
+ >>> s.query(1, 3)
+ 3
+ >>> s.query(1, 5)
+ 5
+ """
return self.query_recursive(1, 0, self.N - 1, a - 1, b - 1)
- def query_recursive(self, idx, l, r, a, b): # noqa: E741
+ def query_recursive(self, idx, left, right, a, b):
"""
query(1, 1, N, a, b) for query max of [a,b]
"""
- if r < a or l > b:
+ if right < a or left > b:
return -math.inf
- if l >= a and r <= b: # noqa: E741
+ if left >= a and right <= b:
return self.st[idx]
- mid = (l + r) // 2
- q1 = self.query_recursive(self.left(idx), l, mid, a, b)
- q2 = self.query_recursive(self.right(idx), mid + 1, r, a, b)
+ mid = (left + right) // 2
+ q1 = self.query_recursive(self.left(idx), left, mid, a, b)
+ q2 = self.query_recursive(self.right(idx), mid + 1, right, a, b)
return max(q1, q2)
- def showData(self):
- showList = []
- for i in range(1, N + 1):
- showList += [self.query(i, i)]
- print(showList)
+ def show_data(self):
+ show_list = []
+ for i in range(1, self.N + 1):
+ show_list += [self.query(i, i)]
+ print(show_list)
if __name__ == "__main__":
@@ -75,4 +113,4 @@ def showData(self):
segt.update(1, 3, 111)
print(segt.query(1, 15))
segt.update(7, 8, 235)
- segt.showData()
+ segt.show_data()
diff --git a/data_structures/binary_tree/segment_tree_other.py b/data_structures/binary_tree/segment_tree_other.py
index 90afd7ca8b71..95f21ddd4777 100644
--- a/data_structures/binary_tree/segment_tree_other.py
+++ b/data_structures/binary_tree/segment_tree_other.py
@@ -3,6 +3,7 @@
allowing queries to be done later in log(N) time
function takes 2 values and returns a same type value
"""
+
from collections.abc import Sequence
from queue import Queue
@@ -16,40 +17,36 @@ def __init__(self, start, end, val, left=None, right=None):
self.left = left
self.right = right
- def __str__(self):
- return f"val: {self.val}, start: {self.start}, end: {self.end}"
+ def __repr__(self):
+ return f"SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})"
class SegmentTree:
"""
>>> import operator
>>> num_arr = SegmentTree([2, 1, 5, 3, 4], operator.add)
- >>> for node in num_arr.traverse():
- ... print(node)
- ...
- val: 15, start: 0, end: 4
- val: 8, start: 0, end: 2
- val: 7, start: 3, end: 4
- val: 3, start: 0, end: 1
- val: 5, start: 2, end: 2
- val: 3, start: 3, end: 3
- val: 4, start: 4, end: 4
- val: 2, start: 0, end: 0
- val: 1, start: 1, end: 1
+ >>> tuple(num_arr.traverse()) # doctest: +NORMALIZE_WHITESPACE
+ (SegmentTreeNode(start=0, end=4, val=15),
+ SegmentTreeNode(start=0, end=2, val=8),
+ SegmentTreeNode(start=3, end=4, val=7),
+ SegmentTreeNode(start=0, end=1, val=3),
+ SegmentTreeNode(start=2, end=2, val=5),
+ SegmentTreeNode(start=3, end=3, val=3),
+ SegmentTreeNode(start=4, end=4, val=4),
+ SegmentTreeNode(start=0, end=0, val=2),
+ SegmentTreeNode(start=1, end=1, val=1))
>>>
>>> num_arr.update(1, 5)
- >>> for node in num_arr.traverse():
- ... print(node)
- ...
- val: 19, start: 0, end: 4
- val: 12, start: 0, end: 2
- val: 7, start: 3, end: 4
- val: 7, start: 0, end: 1
- val: 5, start: 2, end: 2
- val: 3, start: 3, end: 3
- val: 4, start: 4, end: 4
- val: 2, start: 0, end: 0
- val: 5, start: 1, end: 1
+ >>> tuple(num_arr.traverse()) # doctest: +NORMALIZE_WHITESPACE
+ (SegmentTreeNode(start=0, end=4, val=19),
+ SegmentTreeNode(start=0, end=2, val=12),
+ SegmentTreeNode(start=3, end=4, val=7),
+ SegmentTreeNode(start=0, end=1, val=7),
+ SegmentTreeNode(start=2, end=2, val=5),
+ SegmentTreeNode(start=3, end=3, val=3),
+ SegmentTreeNode(start=4, end=4, val=4),
+ SegmentTreeNode(start=0, end=0, val=2),
+ SegmentTreeNode(start=1, end=1, val=5))
>>>
>>> num_arr.query_range(3, 4)
7
@@ -62,29 +59,29 @@ class SegmentTree:
>>> for node in max_arr.traverse():
... print(node)
...
- val: 5, start: 0, end: 4
- val: 5, start: 0, end: 2
- val: 4, start: 3, end: 4
- val: 2, start: 0, end: 1
- val: 5, start: 2, end: 2
- val: 3, start: 3, end: 3
- val: 4, start: 4, end: 4
- val: 2, start: 0, end: 0
- val: 1, start: 1, end: 1
+ SegmentTreeNode(start=0, end=4, val=5)
+ SegmentTreeNode(start=0, end=2, val=5)
+ SegmentTreeNode(start=3, end=4, val=4)
+ SegmentTreeNode(start=0, end=1, val=2)
+ SegmentTreeNode(start=2, end=2, val=5)
+ SegmentTreeNode(start=3, end=3, val=3)
+ SegmentTreeNode(start=4, end=4, val=4)
+ SegmentTreeNode(start=0, end=0, val=2)
+ SegmentTreeNode(start=1, end=1, val=1)
>>>
>>> max_arr.update(1, 5)
>>> for node in max_arr.traverse():
... print(node)
...
- val: 5, start: 0, end: 4
- val: 5, start: 0, end: 2
- val: 4, start: 3, end: 4
- val: 5, start: 0, end: 1
- val: 5, start: 2, end: 2
- val: 3, start: 3, end: 3
- val: 4, start: 4, end: 4
- val: 2, start: 0, end: 0
- val: 5, start: 1, end: 1
+ SegmentTreeNode(start=0, end=4, val=5)
+ SegmentTreeNode(start=0, end=2, val=5)
+ SegmentTreeNode(start=3, end=4, val=4)
+ SegmentTreeNode(start=0, end=1, val=5)
+ SegmentTreeNode(start=2, end=2, val=5)
+ SegmentTreeNode(start=3, end=3, val=3)
+ SegmentTreeNode(start=4, end=4, val=4)
+ SegmentTreeNode(start=0, end=0, val=2)
+ SegmentTreeNode(start=1, end=1, val=5)
>>>
>>> max_arr.query_range(3, 4)
4
@@ -97,29 +94,29 @@ class SegmentTree:
>>> for node in min_arr.traverse():
... print(node)
...
- val: 1, start: 0, end: 4
- val: 1, start: 0, end: 2
- val: 3, start: 3, end: 4
- val: 1, start: 0, end: 1
- val: 5, start: 2, end: 2
- val: 3, start: 3, end: 3
- val: 4, start: 4, end: 4
- val: 2, start: 0, end: 0
- val: 1, start: 1, end: 1
+ SegmentTreeNode(start=0, end=4, val=1)
+ SegmentTreeNode(start=0, end=2, val=1)
+ SegmentTreeNode(start=3, end=4, val=3)
+ SegmentTreeNode(start=0, end=1, val=1)
+ SegmentTreeNode(start=2, end=2, val=5)
+ SegmentTreeNode(start=3, end=3, val=3)
+ SegmentTreeNode(start=4, end=4, val=4)
+ SegmentTreeNode(start=0, end=0, val=2)
+ SegmentTreeNode(start=1, end=1, val=1)
>>>
>>> min_arr.update(1, 5)
>>> for node in min_arr.traverse():
... print(node)
...
- val: 2, start: 0, end: 4
- val: 2, start: 0, end: 2
- val: 3, start: 3, end: 4
- val: 2, start: 0, end: 1
- val: 5, start: 2, end: 2
- val: 3, start: 3, end: 3
- val: 4, start: 4, end: 4
- val: 2, start: 0, end: 0
- val: 5, start: 1, end: 1
+ SegmentTreeNode(start=0, end=4, val=2)
+ SegmentTreeNode(start=0, end=2, val=2)
+ SegmentTreeNode(start=3, end=4, val=3)
+ SegmentTreeNode(start=0, end=1, val=2)
+ SegmentTreeNode(start=2, end=2, val=5)
+ SegmentTreeNode(start=3, end=3, val=3)
+ SegmentTreeNode(start=4, end=4, val=4)
+ SegmentTreeNode(start=0, end=0, val=2)
+ SegmentTreeNode(start=1, end=1, val=5)
>>>
>>> min_arr.query_range(3, 4)
3
@@ -128,7 +125,6 @@ class SegmentTree:
>>> min_arr.query_range(1, 3)
3
>>>
-
"""
def __init__(self, collection: Sequence, function):
diff --git a/data_structures/binary_tree/serialize_deserialize_binary_tree.py b/data_structures/binary_tree/serialize_deserialize_binary_tree.py
new file mode 100644
index 000000000000..7d3e0c61f96d
--- /dev/null
+++ b/data_structures/binary_tree/serialize_deserialize_binary_tree.py
@@ -0,0 +1,140 @@
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class TreeNode:
+ """
+ A binary tree node has a value, left child, and right child.
+
+ Props:
+ value: The value of the node.
+ left: The left child of the node.
+ right: The right child of the node.
+ """
+
+ value: int = 0
+ left: TreeNode | None = None
+ right: TreeNode | None = None
+
+ def __post_init__(self):
+ if not isinstance(self.value, int):
+ raise TypeError("Value must be an integer.")
+
+ def __iter__(self) -> Iterator[TreeNode]:
+ """
+ Iterate through the tree in preorder.
+
+ Returns:
+ An iterator of the tree nodes.
+
+ >>> list(TreeNode(1))
+ [1,null,null]
+ >>> tuple(TreeNode(1, TreeNode(2), TreeNode(3)))
+ (1,2,null,null,3,null,null, 2,null,null, 3,null,null)
+ """
+ yield self
+ yield from self.left or ()
+ yield from self.right or ()
+
+ def __len__(self) -> int:
+ """
+ Count the number of nodes in the tree.
+
+ Returns:
+ The number of nodes in the tree.
+
+ >>> len(TreeNode(1))
+ 1
+ >>> len(TreeNode(1, TreeNode(2), TreeNode(3)))
+ 3
+ """
+ return sum(1 for _ in self)
+
+ def __repr__(self) -> str:
+ """
+ Represent the tree as a string.
+
+ Returns:
+ A string representation of the tree.
+
+ >>> repr(TreeNode(1))
+ '1,null,null'
+ >>> repr(TreeNode(1, TreeNode(2), TreeNode(3)))
+ '1,2,null,null,3,null,null'
+ >>> repr(TreeNode(1, TreeNode(2), TreeNode(3, TreeNode(4), TreeNode(5))))
+ '1,2,null,null,3,4,null,null,5,null,null'
+ """
+ return f"{self.value},{self.left!r},{self.right!r}".replace("None", "null")
+
+ @classmethod
+ def five_tree(cls) -> TreeNode:
+ """
+ >>> repr(TreeNode.five_tree())
+ '1,2,null,null,3,4,null,null,5,null,null'
+ """
+ root = TreeNode(1)
+ root.left = TreeNode(2)
+ root.right = TreeNode(3)
+ root.right.left = TreeNode(4)
+ root.right.right = TreeNode(5)
+ return root
+
+
+def deserialize(data: str) -> TreeNode | None:
+ """
+ Deserialize a string to a binary tree.
+
+ Args:
+ data(str): The serialized string.
+
+ Returns:
+ The root of the binary tree.
+
+ >>> root = TreeNode.five_tree()
+ >>> serialzed_data = repr(root)
+ >>> deserialized = deserialize(serialzed_data)
+ >>> root == deserialized
+ True
+ >>> root is deserialized # two separate trees
+ False
+ >>> root.right.right.value = 6
+ >>> root == deserialized
+ False
+ >>> serialzed_data = repr(root)
+ >>> deserialized = deserialize(serialzed_data)
+ >>> root == deserialized
+ True
+ >>> deserialize("")
+ Traceback (most recent call last):
+ ...
+ ValueError: Data cannot be empty.
+ """
+
+ if not data:
+ raise ValueError("Data cannot be empty.")
+
+ # Split the serialized string by a comma to get node values
+ nodes = data.split(",")
+
+ def build_tree() -> TreeNode | None:
+ # Get the next value from the list
+ value = nodes.pop(0)
+
+ if value == "null":
+ return None
+
+ node = TreeNode(int(value))
+ node.left = build_tree() # Recursively build left subtree
+ node.right = build_tree() # Recursively build right subtree
+ return node
+
+ return build_tree()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/binary_tree/symmetric_tree.py b/data_structures/binary_tree/symmetric_tree.py
new file mode 100644
index 000000000000..2bfeac98b2c9
--- /dev/null
+++ b/data_structures/binary_tree/symmetric_tree.py
@@ -0,0 +1,159 @@
+"""
+Given the root of a binary tree, check whether it is a mirror of itself
+(i.e., symmetric around its center).
+
+Leetcode reference: https://leetcode.com/problems/symmetric-tree/
+"""
+
+from __future__ import annotations
+
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ """
+ A Node represents an element of a binary tree, which contains:
+
+ Attributes:
+ data: The value stored in the node (int).
+ left: Pointer to the left child node (Node or None).
+ right: Pointer to the right child node (Node or None).
+
+ Example:
+ >>> node = Node(1, Node(2), Node(3))
+ >>> node.data
+ 1
+ >>> node.left.data
+ 2
+ >>> node.right.data
+ 3
+ """
+
+ data: int
+ left: Node | None = None
+ right: Node | None = None
+
+
+def make_symmetric_tree() -> Node:
+ r"""
+ Create a symmetric tree for testing.
+
+ The tree looks like this:
+ 1
+ / \
+ 2 2
+ / \ / \
+ 3 4 4 3
+
+ Returns:
+ Node: Root node of a symmetric tree.
+
+ Example:
+ >>> tree = make_symmetric_tree()
+ >>> tree.data
+ 1
+ >>> tree.left.data == tree.right.data
+ True
+ >>> tree.left.left.data == tree.right.right.data
+ True
+ """
+ root = Node(1)
+ root.left = Node(2)
+ root.right = Node(2)
+ root.left.left = Node(3)
+ root.left.right = Node(4)
+ root.right.left = Node(4)
+ root.right.right = Node(3)
+ return root
+
+
+def make_asymmetric_tree() -> Node:
+ r"""
+ Create an asymmetric tree for testing.
+
+ The tree looks like this:
+ 1
+ / \
+ 2 2
+ / \ / \
+ 3 4 3 4
+
+ Returns:
+ Node: Root node of an asymmetric tree.
+
+ Example:
+ >>> tree = make_asymmetric_tree()
+ >>> tree.data
+ 1
+ >>> tree.left.data == tree.right.data
+ True
+ >>> tree.left.left.data == tree.right.right.data
+ False
+ """
+ root = Node(1)
+ root.left = Node(2)
+ root.right = Node(2)
+ root.left.left = Node(3)
+ root.left.right = Node(4)
+ root.right.left = Node(3)
+ root.right.right = Node(4)
+ return root
+
+
+def is_symmetric_tree(tree: Node) -> bool:
+ """
+ Check if a binary tree is symmetric (i.e., a mirror of itself).
+
+ Parameters:
+ tree: The root node of the binary tree.
+
+ Returns:
+ bool: True if the tree is symmetric, False otherwise.
+
+ Example:
+ >>> is_symmetric_tree(make_symmetric_tree())
+ True
+ >>> is_symmetric_tree(make_asymmetric_tree())
+ False
+ """
+ if tree:
+ return is_mirror(tree.left, tree.right)
+ return True # An empty tree is considered symmetric.
+
+
+def is_mirror(left: Node | None, right: Node | None) -> bool:
+ """
+ Check if two subtrees are mirror images of each other.
+
+ Parameters:
+ left: The root node of the left subtree.
+ right: The root node of the right subtree.
+
+ Returns:
+ bool: True if the two subtrees are mirrors of each other, False otherwise.
+
+ Example:
+ >>> tree1 = make_symmetric_tree()
+ >>> is_mirror(tree1.left, tree1.right)
+ True
+ >>> tree2 = make_asymmetric_tree()
+ >>> is_mirror(tree2.left, tree2.right)
+ False
+ """
+ if left is None and right is None:
+ # Both sides are empty, which is symmetric.
+ return True
+ if left is None or right is None:
+ # One side is empty while the other is not, which is not symmetric.
+ return False
+ if left.data == right.data:
+ # The values match, so check the subtrees recursively.
+ return is_mirror(left.left, right.right) and is_mirror(left.right, right.left)
+ return False
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/data_structures/binary_tree/treap.py b/data_structures/binary_tree/treap.py
index 26648f7aba61..3114c6fa1c26 100644
--- a/data_structures/binary_tree/treap.py
+++ b/data_structures/binary_tree/treap.py
@@ -1,5 +1,3 @@
-# flake8: noqa
-
from __future__ import annotations
from random import random
@@ -11,13 +9,13 @@ class Node:
Treap is a binary tree by value and heap by priority
"""
- def __init__(self, value: int = None):
+ def __init__(self, value: int | None = None):
self.value = value
self.prior = random()
- self.left = None
- self.right = None
+ self.left: Node | None = None
+ self.right: Node | None = None
- def __repr__(self):
+ def __repr__(self) -> str:
from pprint import pformat
if self.left is None and self.right is None:
@@ -27,43 +25,40 @@ def __repr__(self):
{f"{self.value}: {self.prior:.5}": (self.left, self.right)}, indent=1
)
- def __str__(self):
+ def __str__(self) -> str:
value = str(self.value) + " "
left = str(self.left or "")
right = str(self.right or "")
return value + left + right
-def split(root: Node, value: int) -> tuple[Node, Node]:
+def split(root: Node | None, value: int) -> tuple[Node | None, Node | None]:
"""
We split current tree into 2 trees with value:
Left tree contains all values less than split value.
Right tree contains all values greater or equal, than split value
"""
- if root is None: # None tree is split into 2 Nones
- return (None, None)
- elif root.value is None:
- return (None, None)
+ if root is None or root.value is None: # None tree is split into 2 Nones
+ return None, None
+ elif value < root.value:
+ """
+ Right tree's root will be current node.
+ Now we split(with the same value) current node's left son
+ Left tree: left part of that split
+ Right tree's left son: right part of that split
+ """
+ left, root.left = split(root.left, value)
+ return left, root
else:
- if value < root.value:
- """
- Right tree's root will be current node.
- Now we split(with the same value) current node's left son
- Left tree: left part of that split
- Right tree's left son: right part of that split
- """
- left, root.left = split(root.left, value)
- return (left, root)
- else:
- """
- Just symmetric to previous case
- """
- root.right, right = split(root.right, value)
- return (root, right)
+ """
+ Just symmetric to previous case
+ """
+ root.right, right = split(root.right, value)
+ return root, right
-def merge(left: Node, right: Node) -> Node:
+def merge(left: Node | None, right: Node | None) -> Node | None:
"""
We merge 2 trees into one.
Note: all left tree's values must be less than all right tree's
@@ -85,7 +80,7 @@ def merge(left: Node, right: Node) -> Node:
return right
-def insert(root: Node, value: int) -> Node:
+def insert(root: Node | None, value: int) -> Node | None:
"""
Insert element
@@ -98,7 +93,7 @@ def insert(root: Node, value: int) -> Node:
return merge(merge(left, node), right)
-def erase(root: Node, value: int) -> Node:
+def erase(root: Node | None, value: int) -> Node | None:
"""
Erase element
@@ -111,7 +106,7 @@ def erase(root: Node, value: int) -> Node:
return merge(left, right)
-def inorder(root: Node):
+def inorder(root: Node | None) -> None:
"""
Just recursive print of a tree
"""
@@ -123,28 +118,28 @@ def inorder(root: Node):
inorder(root.right)
-def interactTreap(root, args):
+def interact_treap(root: Node | None, args: str) -> Node | None:
"""
Commands:
+ value to add value into treap
- value to erase all nodes with value
- >>> root = interactTreap(None, "+1")
+ >>> root = interact_treap(None, "+1")
>>> inorder(root)
1,
- >>> root = interactTreap(root, "+3 +5 +17 +19 +2 +16 +4 +0")
+ >>> root = interact_treap(root, "+3 +5 +17 +19 +2 +16 +4 +0")
>>> inorder(root)
0,1,2,3,4,5,16,17,19,
- >>> root = interactTreap(root, "+4 +4 +4")
+ >>> root = interact_treap(root, "+4 +4 +4")
>>> inorder(root)
0,1,2,3,4,4,4,4,5,16,17,19,
- >>> root = interactTreap(root, "-0")
+ >>> root = interact_treap(root, "-0")
>>> inorder(root)
1,2,3,4,4,4,4,5,16,17,19,
- >>> root = interactTreap(root, "-4")
+ >>> root = interact_treap(root, "-4")
>>> inorder(root)
1,2,3,5,16,17,19,
- >>> root = interactTreap(root, "=0")
+ >>> root = interact_treap(root, "=0")
Unknown command
"""
for arg in args.split():
@@ -160,7 +155,7 @@ def interactTreap(root, args):
return root
-def main():
+def main() -> None:
"""After each command, program prints treap"""
root = None
print(
@@ -170,7 +165,7 @@ def main():
args = input()
while args != "q":
- root = interactTreap(root, args)
+ root = interact_treap(root, args)
print(root)
args = input()
diff --git a/data_structures/binary_tree/wavelet_tree.py b/data_structures/binary_tree/wavelet_tree.py
new file mode 100644
index 000000000000..2da571e8d326
--- /dev/null
+++ b/data_structures/binary_tree/wavelet_tree.py
@@ -0,0 +1,210 @@
+"""
+Wavelet tree is a data-structure designed to efficiently answer various range queries
+for arrays. Wavelets trees are different from other binary trees in the sense that
+the nodes are split based on the actual values of the elements and not on indices,
+such as the with segment trees or fenwick trees. You can read more about them here:
+1. https://users.dcc.uchile.cl/~jperez/papers/ioiconf16.pdf
+2. https://www.youtube.com/watch?v=4aSv9PcecDw&t=811s
+3. https://www.youtube.com/watch?v=CybAgVF-MMc&t=1178s
+"""
+
+from __future__ import annotations
+
+test_array = [2, 1, 4, 5, 6, 0, 8, 9, 1, 2, 0, 6, 4, 2, 0, 6, 5, 3, 2, 7]
+
+
+class Node:
+ def __init__(self, length: int) -> None:
+ self.minn: int = -1
+ self.maxx: int = -1
+ self.map_left: list[int] = [-1] * length
+ self.left: Node | None = None
+ self.right: Node | None = None
+
+ def __repr__(self) -> str:
+ """
+ >>> node = Node(length=27)
+ >>> repr(node)
+ 'Node(min_value=-1 max_value=-1)'
+ >>> repr(node) == str(node)
+ True
+ """
+ return f"Node(min_value={self.minn} max_value={self.maxx})"
+
+
+def build_tree(arr: list[int]) -> Node | None:
+ """
+ Builds the tree for arr and returns the root
+ of the constructed tree
+
+ >>> build_tree(test_array)
+ Node(min_value=0 max_value=9)
+ """
+ root = Node(len(arr))
+ root.minn, root.maxx = min(arr), max(arr)
+ # Leaf node case where the node contains only one unique value
+ if root.minn == root.maxx:
+ return root
+ """
+ Take the mean of min and max element of arr as the pivot and
+ partition arr into left_arr and right_arr with all elements <= pivot in the
+ left_arr and the rest in right_arr, maintaining the order of the elements,
+ then recursively build trees for left_arr and right_arr
+ """
+ pivot = (root.minn + root.maxx) // 2
+
+ left_arr: list[int] = []
+ right_arr: list[int] = []
+
+ for index, num in enumerate(arr):
+ if num <= pivot:
+ left_arr.append(num)
+ else:
+ right_arr.append(num)
+ root.map_left[index] = len(left_arr)
+ root.left = build_tree(left_arr)
+ root.right = build_tree(right_arr)
+ return root
+
+
+def rank_till_index(node: Node | None, num: int, index: int) -> int:
+ """
+ Returns the number of occurrences of num in interval [0, index] in the list
+
+ >>> root = build_tree(test_array)
+ >>> rank_till_index(root, 6, 6)
+ 1
+ >>> rank_till_index(root, 2, 0)
+ 1
+ >>> rank_till_index(root, 1, 10)
+ 2
+ >>> rank_till_index(root, 17, 7)
+ 0
+ >>> rank_till_index(root, 0, 9)
+ 1
+ """
+ if index < 0 or node is None:
+ return 0
+ # Leaf node cases
+ if node.minn == node.maxx:
+ return index + 1 if node.minn == num else 0
+ pivot = (node.minn + node.maxx) // 2
+ if num <= pivot:
+ # go the left subtree and map index to the left subtree
+ return rank_till_index(node.left, num, node.map_left[index] - 1)
+ else:
+ # go to the right subtree and map index to the right subtree
+ return rank_till_index(node.right, num, index - node.map_left[index])
+
+
+def rank(node: Node | None, num: int, start: int, end: int) -> int:
+ """
+ Returns the number of occurrences of num in interval [start, end] in the list
+
+ >>> root = build_tree(test_array)
+ >>> rank(root, 6, 3, 13)
+ 2
+ >>> rank(root, 2, 0, 19)
+ 4
+ >>> rank(root, 9, 2 ,2)
+ 0
+ >>> rank(root, 0, 5, 10)
+ 2
+ """
+ if start > end:
+ return 0
+ rank_till_end = rank_till_index(node, num, end)
+ rank_before_start = rank_till_index(node, num, start - 1)
+ return rank_till_end - rank_before_start
+
+
+def quantile(node: Node | None, index: int, start: int, end: int) -> int:
+ """
+ Returns the index'th smallest element in interval [start, end] in the list
+ index is 0-indexed
+
+ >>> root = build_tree(test_array)
+ >>> quantile(root, 2, 2, 5)
+ 5
+ >>> quantile(root, 5, 2, 13)
+ 4
+ >>> quantile(root, 0, 6, 6)
+ 8
+ >>> quantile(root, 4, 2, 5)
+ -1
+ """
+ if index > (end - start) or start > end or node is None:
+ return -1
+ # Leaf node case
+ if node.minn == node.maxx:
+ return node.minn
+ # Number of elements in the left subtree in interval [start, end]
+ num_elements_in_left_tree = node.map_left[end] - (
+ node.map_left[start - 1] if start else 0
+ )
+ if num_elements_in_left_tree > index:
+ return quantile(
+ node.left,
+ index,
+ (node.map_left[start - 1] if start else 0),
+ node.map_left[end] - 1,
+ )
+ else:
+ return quantile(
+ node.right,
+ index - num_elements_in_left_tree,
+ start - (node.map_left[start - 1] if start else 0),
+ end - node.map_left[end],
+ )
+
+
+def range_counting(
+ node: Node | None, start: int, end: int, start_num: int, end_num: int
+) -> int:
+ """
+ Returns the number of elements in range [start_num, end_num]
+ in interval [start, end] in the list
+
+ >>> root = build_tree(test_array)
+ >>> range_counting(root, 1, 10, 3, 7)
+ 3
+ >>> range_counting(root, 2, 2, 1, 4)
+ 1
+ >>> range_counting(root, 0, 19, 0, 100)
+ 20
+ >>> range_counting(root, 1, 0, 1, 100)
+ 0
+ >>> range_counting(root, 0, 17, 100, 1)
+ 0
+ """
+ if (
+ start > end
+ or node is None
+ or start_num > end_num
+ or node.minn > end_num
+ or node.maxx < start_num
+ ):
+ return 0
+ if start_num <= node.minn and node.maxx <= end_num:
+ return end - start + 1
+ left = range_counting(
+ node.left,
+ (node.map_left[start - 1] if start else 0),
+ node.map_left[end] - 1,
+ start_num,
+ end_num,
+ )
+ right = range_counting(
+ node.right,
+ start - (node.map_left[start - 1] if start else 0),
+ end - node.map_left[end],
+ start_num,
+ end_num,
+ )
+ return left + right
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/disjoint_set/disjoint_set.py b/data_structures/disjoint_set/disjoint_set.py
index a93b89621c4a..edc4736b6132 100644
--- a/data_structures/disjoint_set/disjoint_set.py
+++ b/data_structures/disjoint_set/disjoint_set.py
@@ -1,17 +1,19 @@
"""
- disjoint set
- Reference: https://en.wikipedia.org/wiki/Disjoint-set_data_structure
+Disjoint set.
+Reference: https://en.wikipedia.org/wiki/Disjoint-set_data_structure
"""
class Node:
- def __init__(self, data):
+ def __init__(self, data: int) -> None:
self.data = data
+ self.rank: int
+ self.parent: Node
-def make_set(x):
+def make_set(x: Node) -> None:
"""
- make x as a set.
+ Make x as a set.
"""
# rank is the distance from x to its' parent
# root's rank is 0
@@ -19,14 +21,17 @@ def make_set(x):
x.parent = x
-def union_set(x, y):
+def union_set(x: Node, y: Node) -> None:
"""
- union two sets.
+ Union of two sets.
set with bigger rank should be parent, so that the
disjoint set tree will be more flat.
"""
x, y = find_set(x), find_set(y)
- if x.rank > y.rank:
+ if x == y:
+ return
+
+ elif x.rank > y.rank:
y.parent = x
else:
x.parent = y
@@ -34,9 +39,9 @@ def union_set(x, y):
y.rank += 1
-def find_set(x):
+def find_set(x: Node) -> Node:
"""
- return the parent of x
+ Return the parent of x
"""
if x != x.parent:
x.parent = find_set(x.parent)
@@ -51,10 +56,11 @@ def find_python_set(node: Node) -> set:
for s in sets:
if node.data in s:
return s
- raise ValueError(f"{node.data} is not in {sets}")
+ msg = f"{node.data} is not in {sets}"
+ raise ValueError(msg)
-def test_disjoint_set():
+def test_disjoint_set() -> None:
"""
>>> test_disjoint_set()
"""
diff --git a/data_structures/hashing/bloom_filter.py b/data_structures/hashing/bloom_filter.py
new file mode 100644
index 000000000000..eb2cb4b79c46
--- /dev/null
+++ b/data_structures/hashing/bloom_filter.py
@@ -0,0 +1,106 @@
+"""
+See https://en.wikipedia.org/wiki/Bloom_filter
+
+The use of this data structure is to test membership in a set.
+Compared to Python's built-in set() it is more space-efficient.
+In the following example, only 8 bits of memory will be used:
+>>> bloom = Bloom(size=8)
+
+Initially, the filter contains all zeros:
+>>> bloom.bitstring
+'00000000'
+
+When an element is added, two bits are set to 1
+since there are 2 hash functions in this implementation:
+>>> "Titanic" in bloom
+False
+>>> bloom.add("Titanic")
+>>> bloom.bitstring
+'01100000'
+>>> "Titanic" in bloom
+True
+
+However, sometimes only one bit is added
+because both hash functions return the same value
+>>> bloom.add("Avatar")
+>>> "Avatar" in bloom
+True
+>>> bloom.format_hash("Avatar")
+'00000100'
+>>> bloom.bitstring
+'01100100'
+
+Not added elements should return False ...
+>>> not_present_films = ("The Godfather", "Interstellar", "Parasite", "Pulp Fiction")
+>>> {
+... film: bloom.format_hash(film) for film in not_present_films
+... } # doctest: +NORMALIZE_WHITESPACE
+{'The Godfather': '00000101',
+ 'Interstellar': '00000011',
+ 'Parasite': '00010010',
+ 'Pulp Fiction': '10000100'}
+>>> any(film in bloom for film in not_present_films)
+False
+
+but sometimes there are false positives:
+>>> "Ratatouille" in bloom
+True
+>>> bloom.format_hash("Ratatouille")
+'01100000'
+
+The probability increases with the number of elements added.
+The probability decreases with the number of bits in the bitarray.
+>>> bloom.estimated_error_rate
+0.140625
+>>> bloom.add("The Godfather")
+>>> bloom.estimated_error_rate
+0.25
+>>> bloom.bitstring
+'01100101'
+"""
+
+from hashlib import md5, sha256
+
+HASH_FUNCTIONS = (sha256, md5)
+
+
+class Bloom:
+ def __init__(self, size: int = 8) -> None:
+ self.bitarray = 0b0
+ self.size = size
+
+ def add(self, value: str) -> None:
+ h = self.hash_(value)
+ self.bitarray |= h
+
+ def exists(self, value: str) -> bool:
+ h = self.hash_(value)
+ return (h & self.bitarray) == h
+
+ def __contains__(self, other: str) -> bool:
+ return self.exists(other)
+
+ def format_bin(self, bitarray: int) -> str:
+ res = bin(bitarray)[2:]
+ return res.zfill(self.size)
+
+ @property
+ def bitstring(self) -> str:
+ return self.format_bin(self.bitarray)
+
+ def hash_(self, value: str) -> int:
+ res = 0b0
+ for func in HASH_FUNCTIONS:
+ position = (
+ int.from_bytes(func(value.encode()).digest(), "little") % self.size
+ )
+ res |= 2**position
+ return res
+
+ def format_hash(self, value: str) -> str:
+ return self.format_bin(self.hash_(value))
+
+ @property
+ def estimated_error_rate(self) -> float:
+ n_ones = bin(self.bitarray).count("1")
+ return (n_ones / self.size) ** len(HASH_FUNCTIONS)
diff --git a/data_structures/hashing/double_hash.py b/data_structures/hashing/double_hash.py
index 57b1ffff4770..324282cbfd8d 100644
--- a/data_structures/hashing/double_hash.py
+++ b/data_structures/hashing/double_hash.py
@@ -1,6 +1,19 @@
#!/usr/bin/env python3
+"""
+Double hashing is a collision resolving technique in Open Addressed Hash tables.
+Double hashing uses the idea of applying a second hash function to key when a collision
+occurs. The advantage of Double hashing is that it is one of the best form of probing,
+producing a uniform distribution of records throughout a hash table. This technique
+does not yield any clusters. It is one of effective method for resolving collisions.
+
+Double hashing can be done using: (hash1(key) + i * hash2(key)) % TABLE_SIZE
+Where hash1() and hash2() are hash functions and TABLE_SIZE is size of hash table.
+
+Reference: https://en.wikipedia.org/wiki/Double_hashing
+"""
+
from .hash_table import HashTable
-from .number_theory.prime_numbers import check_prime, next_prime
+from .number_theory.prime_numbers import is_prime, next_prime
class DoubleHash(HashTable):
@@ -12,10 +25,9 @@ def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
def __hash_function_2(self, value, data):
-
next_prime_gt = (
next_prime(value % self.size_table)
- if not check_prime(value % self.size_table)
+ if not is_prime(value % self.size_table)
else value % self.size_table
) # gt = bigger than
return next_prime_gt - (data % next_prime_gt)
@@ -24,6 +36,33 @@ def __hash_double_function(self, key, data, increment):
return (increment * self.__hash_function_2(key, data)) % self.size_table
def _collision_resolution(self, key, data=None):
+ """
+ Examples:
+
+ 1. Try to add three data elements when the size is three
+ >>> dh = DoubleHash(3)
+ >>> dh.insert_data(10)
+ >>> dh.insert_data(20)
+ >>> dh.insert_data(30)
+ >>> dh.keys()
+ {1: 10, 2: 20, 0: 30}
+
+ 2. Try to add three data elements when the size is two
+ >>> dh = DoubleHash(2)
+ >>> dh.insert_data(10)
+ >>> dh.insert_data(20)
+ >>> dh.insert_data(30)
+ >>> dh.keys()
+ {10: 10, 9: 20, 8: 30}
+
+ 3. Try to add three data elements when the size is four
+ >>> dh = DoubleHash(4)
+ >>> dh.insert_data(10)
+ >>> dh.insert_data(20)
+ >>> dh.insert_data(30)
+ >>> dh.keys()
+ {9: 20, 10: 10, 8: 30}
+ """
i = 1
new_key = self.hash_function(data)
@@ -39,3 +78,9 @@ def _collision_resolution(self, key, data=None):
i += 1
return new_key
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/hashing/hash_map.py b/data_structures/hashing/hash_map.py
new file mode 100644
index 000000000000..9213d6930f67
--- /dev/null
+++ b/data_structures/hashing/hash_map.py
@@ -0,0 +1,305 @@
+"""
+Hash map with open addressing.
+
+https://en.wikipedia.org/wiki/Hash_table
+
+Another hash map implementation, with a good explanation.
+Modern Dictionaries by Raymond Hettinger
+https://www.youtube.com/watch?v=p33CVV29OG8
+"""
+
+from collections.abc import Iterator, MutableMapping
+from dataclasses import dataclass
+from typing import Generic, TypeVar
+
+KEY = TypeVar("KEY")
+VAL = TypeVar("VAL")
+
+
+@dataclass(frozen=True, slots=True)
+class _Item(Generic[KEY, VAL]):
+ key: KEY
+ val: VAL
+
+
+class _DeletedItem(_Item):
+ def __init__(self) -> None:
+ super().__init__(None, None)
+
+ def __bool__(self) -> bool:
+ return False
+
+
+_deleted = _DeletedItem()
+
+
+class HashMap(MutableMapping[KEY, VAL]):
+ """
+ Hash map with open addressing.
+ """
+
+ def __init__(
+ self, initial_block_size: int = 8, capacity_factor: float = 0.75
+ ) -> None:
+ self._initial_block_size = initial_block_size
+ self._buckets: list[_Item | None] = [None] * initial_block_size
+ assert 0.0 < capacity_factor < 1.0
+ self._capacity_factor = capacity_factor
+ self._len = 0
+
+ def _get_bucket_index(self, key: KEY) -> int:
+ return hash(key) % len(self._buckets)
+
+ def _get_next_ind(self, ind: int) -> int:
+ """
+ Get next index.
+
+ Implements linear open addressing.
+ >>> HashMap(5)._get_next_ind(3)
+ 4
+ >>> HashMap(5)._get_next_ind(5)
+ 1
+ >>> HashMap(5)._get_next_ind(6)
+ 2
+ >>> HashMap(5)._get_next_ind(9)
+ 0
+ """
+ return (ind + 1) % len(self._buckets)
+
+ def _try_set(self, ind: int, key: KEY, val: VAL) -> bool:
+ """
+ Try to add value to the bucket.
+
+ If bucket is empty or key is the same, does insert and return True.
+
+ If bucket has another key or deleted placeholder,
+ that means that we need to check next bucket.
+ """
+ stored = self._buckets[ind]
+ if not stored:
+ self._buckets[ind] = _Item(key, val)
+ self._len += 1
+ return True
+ elif stored.key == key:
+ self._buckets[ind] = _Item(key, val)
+ return True
+ else:
+ return False
+
+ def _is_full(self) -> bool:
+ """
+ Return true if we have reached safe capacity.
+
+ So we need to increase the number of buckets to avoid collisions.
+
+ >>> hm = HashMap(2)
+ >>> hm._add_item(1, 10)
+ >>> hm._add_item(2, 20)
+ >>> hm._is_full()
+ True
+ >>> HashMap(2)._is_full()
+ False
+ """
+ limit = len(self._buckets) * self._capacity_factor
+ return len(self) >= int(limit)
+
+ def _is_sparse(self) -> bool:
+ """Return true if we need twice fewer buckets when we have now."""
+ if len(self._buckets) <= self._initial_block_size:
+ return False
+ limit = len(self._buckets) * self._capacity_factor / 2
+ return len(self) < limit
+
+ def _resize(self, new_size: int) -> None:
+ old_buckets = self._buckets
+ self._buckets = [None] * new_size
+ self._len = 0
+ for item in old_buckets:
+ if item:
+ self._add_item(item.key, item.val)
+
+ def _size_up(self) -> None:
+ self._resize(len(self._buckets) * 2)
+
+ def _size_down(self) -> None:
+ self._resize(len(self._buckets) // 2)
+
+ def _iterate_buckets(self, key: KEY) -> Iterator[int]:
+ ind = self._get_bucket_index(key)
+ for _ in range(len(self._buckets)):
+ yield ind
+ ind = self._get_next_ind(ind)
+
+ def _add_item(self, key: KEY, val: VAL) -> None:
+ """
+ Try to add 3 elements when the size is 5
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm._add_item(2, 20)
+ >>> hm._add_item(3, 30)
+ >>> hm
+ HashMap(1: 10, 2: 20, 3: 30)
+
+ Try to add 3 elements when the size is 5
+ >>> hm = HashMap(5)
+ >>> hm._add_item(-5, 10)
+ >>> hm._add_item(6, 30)
+ >>> hm._add_item(-7, 20)
+ >>> hm
+ HashMap(-5: 10, 6: 30, -7: 20)
+
+ Try to add 3 elements when size is 1
+ >>> hm = HashMap(1)
+ >>> hm._add_item(10, 13.2)
+ >>> hm._add_item(6, 5.26)
+ >>> hm._add_item(7, 5.155)
+ >>> hm
+ HashMap(10: 13.2)
+
+ Trying to add an element with a key that is a floating point value
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1.5, 10)
+ >>> hm
+ HashMap(1.5: 10)
+
+ 5. Trying to add an item with the same key
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm._add_item(1, 20)
+ >>> hm
+ HashMap(1: 20)
+ """
+ for ind in self._iterate_buckets(key):
+ if self._try_set(ind, key, val):
+ break
+
+ def __setitem__(self, key: KEY, val: VAL) -> None:
+ """
+ 1. Changing value of item whose key is present
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm.__setitem__(1, 20)
+ >>> hm
+ HashMap(1: 20)
+
+ 2. Changing value of item whose key is not present
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm.__setitem__(0, 20)
+ >>> hm
+ HashMap(0: 20, 1: 10)
+
+ 3. Changing the value of the same item multiple times
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm.__setitem__(1, 20)
+ >>> hm.__setitem__(1, 30)
+ >>> hm
+ HashMap(1: 30)
+ """
+ if self._is_full():
+ self._size_up()
+
+ self._add_item(key, val)
+
+ def __delitem__(self, key: KEY) -> None:
+ """
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm._add_item(2, 20)
+ >>> hm._add_item(3, 30)
+ >>> hm.__delitem__(3)
+ >>> hm
+ HashMap(1: 10, 2: 20)
+ >>> hm = HashMap(5)
+ >>> hm._add_item(-5, 10)
+ >>> hm._add_item(6, 30)
+ >>> hm._add_item(-7, 20)
+ >>> hm.__delitem__(-5)
+ >>> hm
+ HashMap(6: 30, -7: 20)
+
+ # Trying to remove a non-existing item
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm._add_item(2, 20)
+ >>> hm._add_item(3, 30)
+ >>> hm.__delitem__(4)
+ Traceback (most recent call last):
+ ...
+ KeyError: 4
+ """
+ for ind in self._iterate_buckets(key):
+ item = self._buckets[ind]
+ if item is None:
+ raise KeyError(key)
+ if item is _deleted:
+ continue
+ if item.key == key:
+ self._buckets[ind] = _deleted
+ self._len -= 1
+ break
+ if self._is_sparse():
+ self._size_down()
+
+ def __getitem__(self, key: KEY) -> VAL:
+ """
+ Returns the item at the given key
+
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm.__getitem__(1)
+ 10
+
+ >>> hm = HashMap(5)
+ >>> hm._add_item(10, -10)
+ >>> hm._add_item(20, -20)
+ >>> hm.__getitem__(20)
+ -20
+
+ >>> hm = HashMap(5)
+ >>> hm._add_item(-1, 10)
+ >>> hm.__getitem__(-1)
+ 10
+ """
+ for ind in self._iterate_buckets(key):
+ item = self._buckets[ind]
+ if item is None:
+ break
+ if item is _deleted:
+ continue
+ if item.key == key:
+ return item.val
+ raise KeyError(key)
+
+ def __len__(self) -> int:
+ """
+ Returns the number of items present in hashmap
+
+ >>> hm = HashMap(5)
+ >>> hm._add_item(1, 10)
+ >>> hm._add_item(2, 20)
+ >>> hm._add_item(3, 30)
+ >>> hm.__len__()
+ 3
+
+ >>> hm = HashMap(5)
+ >>> hm.__len__()
+ 0
+ """
+ return self._len
+
+ def __iter__(self) -> Iterator[KEY]:
+ yield from (item.key for item in self._buckets if item)
+
+ def __repr__(self) -> str:
+ val_string = ", ".join(
+ f"{item.key}: {item.val}" for item in self._buckets if item
+ )
+ return f"HashMap({val_string})"
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/hashing/hash_table.py b/data_structures/hashing/hash_table.py
index fd9e6eec134c..40fcad9a3dab 100644
--- a/data_structures/hashing/hash_table.py
+++ b/data_structures/hashing/hash_table.py
@@ -1,4 +1,6 @@
#!/usr/bin/env python3
+from abc import abstractmethod
+
from .number_theory.prime_numbers import next_prime
@@ -7,32 +9,120 @@ class HashTable:
Basic Hash Table example with open addressing and linear probing
"""
- def __init__(self, size_table, charge_factor=None, lim_charge=None):
+ def __init__(
+ self,
+ size_table: int,
+ charge_factor: int | None = None,
+ lim_charge: float | None = None,
+ ) -> None:
self.size_table = size_table
self.values = [None] * self.size_table
self.lim_charge = 0.75 if lim_charge is None else lim_charge
self.charge_factor = 1 if charge_factor is None else charge_factor
- self.__aux_list = []
- self._keys = {}
+ self.__aux_list: list = []
+ self._keys: dict = {}
def keys(self):
+ """
+ The keys function returns a dictionary containing the key value pairs.
+ key being the index number in hash table and value being the data value.
+
+ Examples:
+ 1. creating HashTable with size 10 and inserting 3 elements
+ >>> ht = HashTable(10)
+ >>> ht.insert_data(10)
+ >>> ht.insert_data(20)
+ >>> ht.insert_data(30)
+ >>> ht.keys()
+ {0: 10, 1: 20, 2: 30}
+
+ 2. creating HashTable with size 5 and inserting 5 elements
+ >>> ht = HashTable(5)
+ >>> ht.insert_data(5)
+ >>> ht.insert_data(4)
+ >>> ht.insert_data(3)
+ >>> ht.insert_data(2)
+ >>> ht.insert_data(1)
+ >>> ht.keys()
+ {0: 5, 4: 4, 3: 3, 2: 2, 1: 1}
+ """
return self._keys
def balanced_factor(self):
- return sum([1 for slot in self.values if slot is not None]) / (
+ return sum(1 for slot in self.values if slot is not None) / (
self.size_table * self.charge_factor
)
def hash_function(self, key):
+ """
+ Generates hash for the given key value
+
+ Examples:
+
+ Creating HashTable with size 5
+ >>> ht = HashTable(5)
+ >>> ht.hash_function(10)
+ 0
+ >>> ht.hash_function(20)
+ 0
+ >>> ht.hash_function(4)
+ 4
+ >>> ht.hash_function(18)
+ 3
+ >>> ht.hash_function(-18)
+ 2
+ >>> ht.hash_function(18.5)
+ 3.5
+ >>> ht.hash_function(0)
+ 0
+ >>> ht.hash_function(-0)
+ 0
+ """
return key % self.size_table
def _step_by_step(self, step_ord):
-
print(f"step {step_ord}")
- print([i for i in range(len(self.values))])
+ print(list(range(len(self.values))))
print(self.values)
def bulk_insert(self, values):
+ """
+ bulk_insert is used for entering more than one element at a time
+ in the HashTable.
+
+ Examples:
+ 1.
+ >>> ht = HashTable(5)
+ >>> ht.bulk_insert((10,20,30))
+ step 1
+ [0, 1, 2, 3, 4]
+ [10, None, None, None, None]
+ step 2
+ [0, 1, 2, 3, 4]
+ [10, 20, None, None, None]
+ step 3
+ [0, 1, 2, 3, 4]
+ [10, 20, 30, None, None]
+
+ 2.
+ >>> ht = HashTable(5)
+ >>> ht.bulk_insert([5,4,3,2,1])
+ step 1
+ [0, 1, 2, 3, 4]
+ [5, None, None, None, None]
+ step 2
+ [0, 1, 2, 3, 4]
+ [5, None, None, None, 4]
+ step 3
+ [0, 1, 2, 3, 4]
+ [5, None, None, 3, 4]
+ step 4
+ [0, 1, 2, 3, 4]
+ [5, None, 2, 3, 4]
+ step 5
+ [0, 1, 2, 3, 4]
+ [5, 1, 2, 3, 4]
+ """
i = 1
self.__aux_list = values
for value in values:
@@ -41,14 +131,103 @@ def bulk_insert(self, values):
i += 1
def _set_value(self, key, data):
+ """
+ _set_value functions allows to update value at a particular hash
+
+ Examples:
+ 1. _set_value in HashTable of size 5
+ >>> ht = HashTable(5)
+ >>> ht.insert_data(10)
+ >>> ht.insert_data(20)
+ >>> ht.insert_data(30)
+ >>> ht._set_value(0,15)
+ >>> ht.keys()
+ {0: 15, 1: 20, 2: 30}
+
+ 2. _set_value in HashTable of size 2
+ >>> ht = HashTable(2)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99)
+ >>> ht._set_value(3,15)
+ >>> ht.keys()
+ {3: 15, 2: 17, 4: 99}
+
+ 3. _set_value in HashTable when hash is not present
+ >>> ht = HashTable(2)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99)
+ >>> ht._set_value(0,15)
+ >>> ht.keys()
+ {3: 18, 2: 17, 4: 99, 0: 15}
+
+ 4. _set_value in HashTable when multiple hash are not present
+ >>> ht = HashTable(2)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99)
+ >>> ht._set_value(0,15)
+ >>> ht._set_value(1,20)
+ >>> ht.keys()
+ {3: 18, 2: 17, 4: 99, 0: 15, 1: 20}
+ """
self.values[key] = data
self._keys[key] = data
+ @abstractmethod
def _collision_resolution(self, key, data=None):
+ """
+ This method is a type of open addressing which is used for handling collision.
+
+ In this implementation the concept of linear probing has been used.
+
+ The hash table is searched sequentially from the original location of the
+ hash, if the new hash/location we get is already occupied we check for the next
+ hash/location.
+
+ references:
+ - https://en.wikipedia.org/wiki/Linear_probing
+
+ Examples:
+ 1. The collision will be with keys 18 & 99, so new hash will be created for 99
+ >>> ht = HashTable(3)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99)
+ >>> ht.keys()
+ {2: 17, 0: 18, 1: 99}
+
+ 2. The collision will be with keys 17 & 101, so new hash
+ will be created for 101
+ >>> ht = HashTable(4)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99)
+ >>> ht.insert_data(101)
+ >>> ht.keys()
+ {1: 17, 2: 18, 3: 99, 0: 101}
+
+ 2. The collision will be with all keys, so new hash will be created for all
+ >>> ht = HashTable(1)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99)
+ >>> ht.keys()
+ {2: 17, 3: 18, 4: 99}
+
+ 3. Trying to insert float key in hash
+ >>> ht = HashTable(1)
+ >>> ht.insert_data(17)
+ >>> ht.insert_data(18)
+ >>> ht.insert_data(99.99)
+ Traceback (most recent call last):
+ ...
+ TypeError: list indices must be integers or slices, not float
+ """
new_key = self.hash_function(key + 1)
while self.values[new_key] is not None and self.values[new_key] != key:
-
if self.values.count(None) > 0:
new_key = self.hash_function(new_key + 1)
else:
@@ -66,6 +245,21 @@ def rehashing(self):
self.insert_data(value)
def insert_data(self, data):
+ """
+ insert_data is used for inserting a single element at a time in the HashTable.
+
+ Examples:
+
+ >>> ht = HashTable(3)
+ >>> ht.insert_data(5)
+ >>> ht.keys()
+ {2: 5}
+ >>> ht = HashTable(5)
+ >>> ht.insert_data(30)
+ >>> ht.insert_data(50)
+ >>> ht.keys()
+ {0: 30, 1: 50}
+ """
key = self.hash_function(data)
if self.values[key] is None:
@@ -81,3 +275,9 @@ def insert_data(self, data):
else:
self.rehashing()
self.insert_data(data)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/hashing/hash_table_with_linked_list.py b/data_structures/hashing/hash_table_with_linked_list.py
index fe838268fce8..f404c5251246 100644
--- a/data_structures/hashing/hash_table_with_linked_list.py
+++ b/data_structures/hashing/hash_table_with_linked_list.py
@@ -14,7 +14,7 @@ def _set_value(self, key, data):
def balanced_factor(self):
return (
- sum([self.charge_factor - len(slot) for slot in self.values])
+ sum(self.charge_factor - len(slot) for slot in self.values)
/ self.size_table
* self.charge_factor
)
diff --git a/data_structures/hashing/number_theory/prime_numbers.py b/data_structures/hashing/number_theory/prime_numbers.py
index db4d40f475b2..82071b5e9f09 100644
--- a/data_structures/hashing/number_theory/prime_numbers.py
+++ b/data_structures/hashing/number_theory/prime_numbers.py
@@ -1,28 +1,58 @@
#!/usr/bin/env python3
"""
- module to operations with prime numbers
+module to operations with prime numbers
"""
+import math
-def check_prime(number):
- """
- it's not the best solution
+
+def is_prime(number: int) -> bool:
+ """Checks to see if a number is a prime in O(sqrt(n)).
+
+ A number is prime if it has exactly two factors: 1 and itself.
+
+ >>> is_prime(0)
+ False
+ >>> is_prime(1)
+ False
+ >>> is_prime(2)
+ True
+ >>> is_prime(3)
+ True
+ >>> is_prime(27)
+ False
+ >>> is_prime(87)
+ False
+ >>> is_prime(563)
+ True
+ >>> is_prime(2999)
+ True
+ >>> is_prime(67483)
+ False
"""
- special_non_primes = [0, 1, 2]
- if number in special_non_primes[:2]:
- return 2
- elif number == special_non_primes[-1]:
- return 3
- return all([number % i for i in range(2, number)])
+ # precondition
+ assert isinstance(number, int) and (number >= 0), (
+ "'number' must been an int and positive"
+ )
+
+ if 1 < number < 4:
+ # 2 and 3 are primes
+ return True
+ elif number < 2 or not number % 2:
+ # Negatives, 0, 1 and all even numbers are not primes
+ return False
+
+ odd_numbers = range(3, int(math.sqrt(number) + 1), 2)
+ return not any(not number % i for i in odd_numbers)
def next_prime(value, factor=1, **kwargs):
value = factor * value
first_value_val = value
- while not check_prime(value):
- value += 1 if not ("desc" in kwargs.keys() and kwargs["desc"] is True) else -1
+ while not is_prime(value):
+ value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1
if value == first_value_val:
return next_prime(value + 1, **kwargs)
diff --git a/data_structures/hashing/quadratic_probing.py b/data_structures/hashing/quadratic_probing.py
index 0930340a347f..56d4926eee9b 100644
--- a/data_structures/hashing/quadratic_probing.py
+++ b/data_structures/hashing/quadratic_probing.py
@@ -11,7 +11,56 @@ class QuadraticProbing(HashTable):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
- def _collision_resolution(self, key, data=None):
+ def _collision_resolution(self, key, data=None): # noqa: ARG002
+ """
+ Quadratic probing is an open addressing scheme used for resolving
+ collisions in hash table.
+
+ It works by taking the original hash index and adding successive
+ values of an arbitrary quadratic polynomial until open slot is found.
+
+ Hash + 1², Hash + 2², Hash + 3² .... Hash + n²
+
+ reference:
+ - https://en.wikipedia.org/wiki/Quadratic_probing
+ e.g:
+ 1. Create hash table with size 7
+ >>> qp = QuadraticProbing(7)
+ >>> qp.insert_data(90)
+ >>> qp.insert_data(340)
+ >>> qp.insert_data(24)
+ >>> qp.insert_data(45)
+ >>> qp.insert_data(99)
+ >>> qp.insert_data(73)
+ >>> qp.insert_data(7)
+ >>> qp.keys()
+ {11: 45, 14: 99, 7: 24, 0: 340, 5: 73, 6: 90, 8: 7}
+
+ 2. Create hash table with size 8
+ >>> qp = QuadraticProbing(8)
+ >>> qp.insert_data(0)
+ >>> qp.insert_data(999)
+ >>> qp.insert_data(111)
+ >>> qp.keys()
+ {0: 0, 7: 999, 3: 111}
+
+ 3. Try to add three data elements when the size is two
+ >>> qp = QuadraticProbing(2)
+ >>> qp.insert_data(0)
+ >>> qp.insert_data(999)
+ >>> qp.insert_data(111)
+ >>> qp.keys()
+ {0: 0, 4: 999, 1: 111}
+
+ 4. Try to add three data elements when the size is one
+ >>> qp = QuadraticProbing(1)
+ >>> qp.insert_data(0)
+ >>> qp.insert_data(999)
+ >>> qp.insert_data(111)
+ >>> qp.keys()
+ {4: 999, 1: 111}
+ """
+
i = 1
new_key = self.hash_function(key + i * i)
@@ -27,3 +76,9 @@ def _collision_resolution(self, key, data=None):
break
return new_key
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/queue/__init__.py b/data_structures/hashing/tests/__init__.py
similarity index 100%
rename from data_structures/queue/__init__.py
rename to data_structures/hashing/tests/__init__.py
diff --git a/data_structures/hashing/tests/test_hash_map.py b/data_structures/hashing/tests/test_hash_map.py
new file mode 100644
index 000000000000..4292c0178b7b
--- /dev/null
+++ b/data_structures/hashing/tests/test_hash_map.py
@@ -0,0 +1,97 @@
+from operator import delitem, getitem, setitem
+
+import pytest
+
+from data_structures.hashing.hash_map import HashMap
+
+
+def _get(k):
+ return getitem, k
+
+
+def _set(k, v):
+ return setitem, k, v
+
+
+def _del(k):
+ return delitem, k
+
+
+def _run_operation(obj, fun, *args):
+ try:
+ return fun(obj, *args), None
+ except Exception as e:
+ return None, e
+
+
+_add_items = (
+ _set("key_a", "val_a"),
+ _set("key_b", "val_b"),
+)
+
+_overwrite_items = [
+ _set("key_a", "val_a"),
+ _set("key_a", "val_b"),
+]
+
+_delete_items = [
+ _set("key_a", "val_a"),
+ _set("key_b", "val_b"),
+ _del("key_a"),
+ _del("key_b"),
+ _set("key_a", "val_a"),
+ _del("key_a"),
+]
+
+_access_absent_items = [
+ _get("key_a"),
+ _del("key_a"),
+ _set("key_a", "val_a"),
+ _del("key_a"),
+ _del("key_a"),
+ _get("key_a"),
+]
+
+_add_with_resize_up = [
+ *[_set(x, x) for x in range(5)], # guaranteed upsize
+]
+
+_add_with_resize_down = [
+ *[_set(x, x) for x in range(5)], # guaranteed upsize
+ *[_del(x) for x in range(5)],
+ _set("key_a", "val_b"),
+]
+
+
+@pytest.mark.parametrize(
+ "operations",
+ [
+ pytest.param(_add_items, id="add items"),
+ pytest.param(_overwrite_items, id="overwrite items"),
+ pytest.param(_delete_items, id="delete items"),
+ pytest.param(_access_absent_items, id="access absent items"),
+ pytest.param(_add_with_resize_up, id="add with resize up"),
+ pytest.param(_add_with_resize_down, id="add with resize down"),
+ ],
+)
+def test_hash_map_is_the_same_as_dict(operations):
+ my = HashMap(initial_block_size=4)
+ py = {}
+ for _, (fun, *args) in enumerate(operations):
+ my_res, my_exc = _run_operation(my, fun, *args)
+ py_res, py_exc = _run_operation(py, fun, *args)
+ assert my_res == py_res
+ assert str(my_exc) == str(py_exc)
+ assert set(py) == set(my)
+ assert len(py) == len(my)
+ assert set(my.items()) == set(py.items())
+
+
+def test_no_new_methods_was_added_to_api():
+ def is_public(name: str) -> bool:
+ return not name.startswith("_")
+
+ dict_public_names = {name for name in dir({}) if is_public(name)}
+ hash_public_names = {name for name in dir(HashMap()) if is_public(name)}
+
+ assert dict_public_names > hash_public_names
diff --git a/data_structures/heap/binomial_heap.py b/data_structures/heap/binomial_heap.py
index 334b444eaaff..9cfdf0c12fe0 100644
--- a/data_structures/heap/binomial_heap.py
+++ b/data_structures/heap/binomial_heap.py
@@ -1,5 +1,3 @@
-# flake8: noqa
-
"""
Binomial Heap
Reference: Advanced Data Structures, Peter Brass
@@ -22,7 +20,7 @@ def __init__(self, val):
self.right = None
self.parent = None
- def mergeTrees(self, other):
+ def merge_trees(self, other):
"""
In-place merge of two binomial trees of equal size.
Returns the root of the resulting tree
@@ -71,13 +69,12 @@ class BinomialHeap:
... first_heap.insert(number)
Size test
- >>> print(first_heap.size)
+ >>> first_heap.size
30
Deleting - delete() test
- >>> for i in range(25):
- ... print(first_heap.deleteMin(), end=" ")
- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
+ >>> [int(first_heap.delete_min()) for _ in range(20)]
+ [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19]
Create a new Heap
>>> second_heap = BinomialHeap()
@@ -97,8 +94,8 @@ class BinomialHeap:
# # # #
preOrder() test
- >>> print(second_heap.preOrder())
- [(17, 0), ('#', 1), (31, 1), (20, 2), ('#', 3), ('#', 3), (34, 2), ('#', 3), ('#', 3)]
+ >>> " ".join(str(x) for x in second_heap.pre_order())
+ "(17, 0) ('#', 1) (31, 1) (20, 2) ('#', 3) ('#', 3) (34, 2) ('#', 3) ('#', 3)"
printing Heap - __str__() test
>>> print(second_heap)
@@ -113,14 +110,17 @@ class BinomialHeap:
---#
mergeHeaps() test
- >>> merged = second_heap.mergeHeaps(first_heap)
+ >>>
+ >>> merged = second_heap.merge_heaps(first_heap)
>>> merged.peek()
17
values in merged heap; (merge is inplace)
- >>> while not first_heap.isEmpty():
- ... print(first_heap.deleteMin(), end=" ")
- 17 20 25 26 27 28 29 31 34
+ >>> results = []
+ >>> while not first_heap.is_empty():
+ ... results.append(int(first_heap.delete_min()))
+ >>> results
+ [17, 20, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 31, 34]
"""
def __init__(self, bottom_root=None, min_node=None, heap_size=0):
@@ -128,7 +128,7 @@ def __init__(self, bottom_root=None, min_node=None, heap_size=0):
self.bottom_root = bottom_root
self.min_node = min_node
- def mergeHeaps(self, other):
+ def merge_heaps(self, other):
"""
In-place merge of two binomial heaps.
Both of them become the resulting merged heap
@@ -136,12 +136,12 @@ def mergeHeaps(self, other):
# Empty heaps corner cases
if other.size == 0:
- return
+ return None
if self.size == 0:
self.size = other.size
self.bottom_root = other.bottom_root
self.min_node = other.min_node
- return
+ return None
# Update size
self.size = self.size + other.size
@@ -174,13 +174,12 @@ def mergeHeaps(self, other):
i.left_tree_size == i.parent.left_tree_size
and i.left_tree_size != i.parent.parent.left_tree_size
):
-
# Neighbouring Nodes
previous_node = i.left
next_node = i.parent.parent
# Merging trees
- i = i.mergeTrees(i.parent)
+ i = i.merge_trees(i.parent)
# Updating links
i.left = previous_node
@@ -233,12 +232,11 @@ def insert(self, val):
and self.bottom_root.left_tree_size
== self.bottom_root.parent.left_tree_size
):
-
# Next node
next_node = self.bottom_root.parent.parent
# Merge
- self.bottom_root = self.bottom_root.mergeTrees(self.bottom_root.parent)
+ self.bottom_root = self.bottom_root.merge_trees(self.bottom_root.parent)
# Update Links
self.bottom_root.parent = next_node
@@ -252,10 +250,10 @@ def peek(self):
"""
return self.min_node.val
- def isEmpty(self):
+ def is_empty(self):
return self.size == 0
- def deleteMin(self):
+ def delete_min(self):
"""
delete min element and return it
"""
@@ -317,7 +315,7 @@ def deleteMin(self):
return min_value
# Remaining cases
# Construct heap of right subtree
- newHeap = BinomialHeap(
+ new_heap = BinomialHeap(
bottom_root=bottom_of_new, min_node=min_of_new, heap_size=size_of_new
)
@@ -354,11 +352,11 @@ def deleteMin(self):
self.min_node = i
i = i.parent
# Merge heaps
- self.mergeHeaps(newHeap)
+ self.merge_heaps(new_heap)
- return min_value
+ return int(min_value)
- def preOrder(self):
+ def pre_order(self):
"""
Returns the Pre-order representation of the heap including
values of nodes plus their level distance from the root;
@@ -369,9 +367,9 @@ def preOrder(self):
while top_root.parent:
top_root = top_root.parent
# preorder
- heap_preOrder = []
- self.__traversal(top_root, heap_preOrder)
- return heap_preOrder
+ heap_pre_order = []
+ self.__traversal(top_root, heap_pre_order)
+ return heap_pre_order
def __traversal(self, curr_node, preorder, level=0):
"""
@@ -389,9 +387,9 @@ def __str__(self):
Overwriting str for a pre-order print of nodes in heap;
Performance is poor, so use only for small examples
"""
- if self.isEmpty():
+ if self.is_empty():
return ""
- preorder_heap = self.preOrder()
+ preorder_heap = self.pre_order()
return "\n".join(("-" * level + str(value)) for value, level in preorder_heap)
diff --git a/data_structures/heap/heap.py b/data_structures/heap/heap.py
index 2dc047436a77..7b15e69f13ca 100644
--- a/data_structures/heap/heap.py
+++ b/data_structures/heap/heap.py
@@ -1,101 +1,251 @@
-#!/usr/bin/python3
+from __future__ import annotations
+from abc import abstractmethod
+from collections.abc import Iterable
+from typing import Generic, Protocol, TypeVar
+
+
+class Comparable(Protocol):
+ @abstractmethod
+ def __lt__(self: T, other: T) -> bool:
+ pass
+
+ @abstractmethod
+ def __gt__(self: T, other: T) -> bool:
+ pass
+
+ @abstractmethod
+ def __eq__(self: T, other: object) -> bool:
+ pass
+
+
+T = TypeVar("T", bound=Comparable)
+
+
+class Heap(Generic[T]):
+ """A Max Heap Implementation
-class Heap:
- """
>>> unsorted = [103, 9, 1, 7, 11, 15, 25, 201, 209, 107, 5]
>>> h = Heap()
- >>> h.build_heap(unsorted)
- >>> h.display()
+ >>> h.build_max_heap(unsorted)
+ >>> h
[209, 201, 25, 103, 107, 15, 1, 9, 7, 11, 5]
>>>
- >>> h.get_max()
+ >>> h.extract_max()
209
- >>> h.display()
+ >>> h
[201, 107, 25, 103, 11, 15, 1, 9, 7, 5]
>>>
>>> h.insert(100)
- >>> h.display()
+ >>> h
[201, 107, 25, 103, 100, 15, 1, 9, 7, 5, 11]
>>>
>>> h.heap_sort()
- >>> h.display()
+ >>> h
[1, 5, 7, 9, 11, 15, 25, 100, 103, 107, 201]
- >>>
"""
- def __init__(self):
- self.h = []
- self.curr_size = 0
+ def __init__(self) -> None:
+ self.h: list[T] = []
+ self.heap_size: int = 0
+
+ def __repr__(self) -> str:
+ return str(self.h)
+
+ def parent_index(self, child_idx: int) -> int | None:
+ """
+ returns the parent index based on the given child index
+
+ >>> h = Heap()
+ >>> h.build_max_heap([103, 9, 1, 7, 11, 15, 25, 201, 209, 107, 5])
+ >>> h
+ [209, 201, 25, 103, 107, 15, 1, 9, 7, 11, 5]
+
+ >>> h.parent_index(-1) # returns none if index is <=0
- def get_left_child_index(self, i):
- left_child_index = 2 * i + 1
- if left_child_index < self.curr_size:
+ >>> h.parent_index(0) # returns none if index is <=0
+
+ >>> h.parent_index(1)
+ 0
+ >>> h.parent_index(2)
+ 0
+ >>> h.parent_index(3)
+ 1
+ >>> h.parent_index(4)
+ 1
+ >>> h.parent_index(5)
+ 2
+ >>> h.parent_index(10.5)
+ 4.0
+ >>> h.parent_index(209.0)
+ 104.0
+ >>> h.parent_index("Test")
+ Traceback (most recent call last):
+ ...
+ TypeError: '>' not supported between instances of 'str' and 'int'
+ """
+ if child_idx > 0:
+ return (child_idx - 1) // 2
+ return None
+
+ def left_child_idx(self, parent_idx: int) -> int | None:
+ """
+ return the left child index if the left child exists.
+ if not, return None.
+ """
+ left_child_index = 2 * parent_idx + 1
+ if left_child_index < self.heap_size:
return left_child_index
return None
- def get_right_child(self, i):
- right_child_index = 2 * i + 2
- if right_child_index < self.curr_size:
+ def right_child_idx(self, parent_idx: int) -> int | None:
+ """
+ return the right child index if the right child exists.
+ if not, return None.
+ """
+ right_child_index = 2 * parent_idx + 2
+ if right_child_index < self.heap_size:
return right_child_index
return None
- def max_heapify(self, index):
- if index < self.curr_size:
- largest = index
- lc = self.get_left_child_index(index)
- rc = self.get_right_child(index)
- if lc is not None and self.h[lc] > self.h[largest]:
- largest = lc
- if rc is not None and self.h[rc] > self.h[largest]:
- largest = rc
- if largest != index:
- self.h[largest], self.h[index] = self.h[index], self.h[largest]
- self.max_heapify(largest)
-
- def build_heap(self, collection):
- self.curr_size = len(collection)
+ def max_heapify(self, index: int) -> None:
+ """
+ correct a single violation of the heap property in a subtree's root.
+
+ It is the function that is responsible for restoring the property
+ of Max heap i.e the maximum element is always at top.
+ """
+ if index < self.heap_size:
+ violation: int = index
+ left_child = self.left_child_idx(index)
+ right_child = self.right_child_idx(index)
+ # check which child is larger than its parent
+ if left_child is not None and self.h[left_child] > self.h[violation]:
+ violation = left_child
+ if right_child is not None and self.h[right_child] > self.h[violation]:
+ violation = right_child
+ # if violation indeed exists
+ if violation != index:
+ # swap to fix the violation
+ self.h[violation], self.h[index] = self.h[index], self.h[violation]
+ # fix the subsequent violation recursively if any
+ self.max_heapify(violation)
+
+ def build_max_heap(self, collection: Iterable[T]) -> None:
+ """
+ build max heap from an unsorted array
+
+ >>> h = Heap()
+ >>> h.build_max_heap([20,40,50,20,10])
+ >>> h
+ [50, 40, 20, 20, 10]
+
+ >>> h = Heap()
+ >>> h.build_max_heap([1,2,3,4,5,6,7,8,9,0])
+ >>> h
+ [9, 8, 7, 4, 5, 6, 3, 2, 1, 0]
+
+ >>> h = Heap()
+ >>> h.build_max_heap([514,5,61,57,8,99,105])
+ >>> h
+ [514, 57, 105, 5, 8, 99, 61]
+
+ >>> h = Heap()
+ >>> h.build_max_heap([514,5,61.6,57,8,9.9,105])
+ >>> h
+ [514, 57, 105, 5, 8, 9.9, 61.6]
+ """
self.h = list(collection)
- if self.curr_size <= 1:
- return
- for i in range(self.curr_size // 2 - 1, -1, -1):
- self.max_heapify(i)
+ self.heap_size = len(self.h)
+ if self.heap_size > 1:
+ # max_heapify from right to left but exclude leaves (last level)
+ for i in range(self.heap_size // 2 - 1, -1, -1):
+ self.max_heapify(i)
- def get_max(self):
- if self.curr_size >= 2:
+ def extract_max(self) -> T:
+ """
+ get and remove max from heap
+
+ >>> h = Heap()
+ >>> h.build_max_heap([20,40,50,20,10])
+ >>> h.extract_max()
+ 50
+
+ >>> h = Heap()
+ >>> h.build_max_heap([514,5,61,57,8,99,105])
+ >>> h.extract_max()
+ 514
+
+ >>> h = Heap()
+ >>> h.build_max_heap([1,2,3,4,5,6,7,8,9,0])
+ >>> h.extract_max()
+ 9
+ """
+ if self.heap_size >= 2:
me = self.h[0]
self.h[0] = self.h.pop(-1)
- self.curr_size -= 1
+ self.heap_size -= 1
self.max_heapify(0)
return me
- elif self.curr_size == 1:
- self.curr_size -= 1
+ elif self.heap_size == 1:
+ self.heap_size -= 1
return self.h.pop(-1)
- return None
+ else:
+ raise Exception("Empty heap")
- def heap_sort(self):
- size = self.curr_size
+ def insert(self, value: T) -> None:
+ """
+ insert a new value into the max heap
+
+ >>> h = Heap()
+ >>> h.insert(10)
+ >>> h
+ [10]
+
+ >>> h = Heap()
+ >>> h.insert(10)
+ >>> h.insert(10)
+ >>> h
+ [10, 10]
+
+ >>> h = Heap()
+ >>> h.insert(10)
+ >>> h.insert(10.1)
+ >>> h
+ [10.1, 10]
+
+ >>> h = Heap()
+ >>> h.insert(0.1)
+ >>> h.insert(0)
+ >>> h.insert(9)
+ >>> h.insert(5)
+ >>> h
+ [9, 5, 0.1, 0]
+ """
+ self.h.append(value)
+ idx = (self.heap_size - 1) // 2
+ self.heap_size += 1
+ while idx >= 0:
+ self.max_heapify(idx)
+ idx = (idx - 1) // 2
+
+ def heap_sort(self) -> None:
+ size = self.heap_size
for j in range(size - 1, 0, -1):
self.h[0], self.h[j] = self.h[j], self.h[0]
- self.curr_size -= 1
+ self.heap_size -= 1
self.max_heapify(0)
- self.curr_size = size
+ self.heap_size = size
- def insert(self, data):
- self.h.append(data)
- curr = (self.curr_size - 1) // 2
- self.curr_size += 1
- while curr >= 0:
- self.max_heapify(curr)
- curr = (curr - 1) // 2
- def display(self):
- print(self.h)
+if __name__ == "__main__":
+ import doctest
+ # run doc test
+ doctest.testmod()
-def main():
+ # demo
for unsorted in [
- [],
[0],
[2],
[3, 5],
@@ -110,26 +260,17 @@ def main():
[103, 9, 1, 7, 11, 15, 25, 201, 209, 107, 5],
[-45, -2, -5],
]:
- print("source unsorted list: %s" % unsorted)
-
- h = Heap()
- h.build_heap(unsorted)
- print("after build heap: ", end=" ")
- h.display()
+ print(f"unsorted array: {unsorted}")
- print("max value: %s" % h.get_max())
- print("delete max value: ", end=" ")
- h.display()
+ heap: Heap[int] = Heap()
+ heap.build_max_heap(unsorted)
+ print(f"after build heap: {heap}")
- h.insert(100)
- print("after insert new value 100: ", end=" ")
- h.display()
+ print(f"max value: {heap.extract_max()}")
+ print(f"after max value removed: {heap}")
- h.heap_sort()
- print("heap sort: ", end=" ")
- h.display()
- print()
+ heap.insert(100)
+ print(f"after new value 100 inserted: {heap}")
-
-if __name__ == "__main__":
- main()
+ heap.heap_sort()
+ print(f"heap-sorted array: {heap}\n")
diff --git a/data_structures/heap/heap_generic.py b/data_structures/heap/heap_generic.py
index 553cb94518c4..ee92149e25a9 100644
--- a/data_structures/heap/heap_generic.py
+++ b/data_structures/heap/heap_generic.py
@@ -1,35 +1,38 @@
+from collections.abc import Callable
+
+
class Heap:
"""
A generic Heap class, can be used as min or max by passing the key function
accordingly.
"""
- def __init__(self, key=None):
+ def __init__(self, key: Callable | None = None) -> None:
# Stores actual heap items.
- self.arr = list()
+ self.arr: list = []
# Stores indexes of each item for supporting updates and deletion.
- self.pos_map = {}
+ self.pos_map: dict = {}
# Stores current size of heap.
self.size = 0
# Stores function used to evaluate the score of an item on which basis ordering
# will be done.
self.key = key or (lambda x: x)
- def _parent(self, i):
+ def _parent(self, i: int) -> int | None:
"""Returns parent index of given index if exists else None"""
return int((i - 1) / 2) if i > 0 else None
- def _left(self, i):
+ def _left(self, i: int) -> int | None:
"""Returns left-child-index of given index if exists else None"""
left = int(2 * i + 1)
return left if 0 < left < self.size else None
- def _right(self, i):
+ def _right(self, i: int) -> int | None:
"""Returns right-child-index of given index if exists else None"""
right = int(2 * i + 2)
return right if 0 < right < self.size else None
- def _swap(self, i, j):
+ def _swap(self, i: int, j: int) -> None:
"""Performs changes required for swapping two elements in the heap"""
# First update the indexes of the items in index map.
self.pos_map[self.arr[i][0]], self.pos_map[self.arr[j][0]] = (
@@ -39,11 +42,11 @@ def _swap(self, i, j):
# Then swap the items in the list.
self.arr[i], self.arr[j] = self.arr[j], self.arr[i]
- def _cmp(self, i, j):
+ def _cmp(self, i: int, j: int) -> bool:
"""Compares the two items using default comparison"""
return self.arr[i][1] < self.arr[j][1]
- def _get_valid_parent(self, i):
+ def _get_valid_parent(self, i: int) -> int:
"""
Returns index of valid parent as per desired ordering among given index and
both it's children
@@ -59,21 +62,21 @@ def _get_valid_parent(self, i):
return valid_parent
- def _heapify_up(self, index):
+ def _heapify_up(self, index: int) -> None:
"""Fixes the heap in upward direction of given index"""
parent = self._parent(index)
while parent is not None and not self._cmp(index, parent):
self._swap(index, parent)
index, parent = parent, self._parent(parent)
- def _heapify_down(self, index):
+ def _heapify_down(self, index: int) -> None:
"""Fixes the heap in downward direction of given index"""
valid_parent = self._get_valid_parent(index)
while valid_parent != index:
self._swap(index, valid_parent)
index, valid_parent = valid_parent, self._get_valid_parent(valid_parent)
- def update_item(self, item, item_value):
+ def update_item(self, item: int, item_value: int) -> None:
"""Updates given item value in heap if present"""
if item not in self.pos_map:
return
@@ -84,7 +87,7 @@ def update_item(self, item, item_value):
self._heapify_up(index)
self._heapify_down(index)
- def delete_item(self, item):
+ def delete_item(self, item: int) -> None:
"""Deletes given item from heap if present"""
if item not in self.pos_map:
return
@@ -99,7 +102,7 @@ def delete_item(self, item):
self._heapify_up(index)
self._heapify_down(index)
- def insert_item(self, item, item_value):
+ def insert_item(self, item: int, item_value: int) -> None:
"""Inserts given item with given value in heap"""
arr_len = len(self.arr)
if arr_len == self.size:
@@ -110,11 +113,11 @@ def insert_item(self, item, item_value):
self.size += 1
self._heapify_up(self.size - 1)
- def get_top(self):
+ def get_top(self) -> tuple | None:
"""Returns top item tuple (Calculated value, item) from heap if present"""
return self.arr[0] if self.size else None
- def extract_top(self):
+ def extract_top(self) -> tuple | None:
"""
Return top item tuple (Calculated value, item) from heap and removes it as well
if present
@@ -163,7 +166,6 @@ def test_heap() -> None:
>>> h.get_top()
[9, -40]
"""
- pass
if __name__ == "__main__":
diff --git a/data_structures/heap/max_heap.py b/data_structures/heap/max_heap.py
index 2a08f8fa2cd1..589f2595a8da 100644
--- a/data_structures/heap/max_heap.py
+++ b/data_structures/heap/max_heap.py
@@ -21,7 +21,7 @@ def __init__(self):
self.__size = 0
def __swap_up(self, i: int) -> None:
- """ Swap the element up """
+ """Swap the element up"""
temporary = self.__heap[i]
while i // 2 > 0:
if self.__heap[i] > self.__heap[i // 2]:
@@ -30,21 +30,20 @@ def __swap_up(self, i: int) -> None:
i //= 2
def insert(self, value: int) -> None:
- """ Insert new element """
+ """Insert new element"""
self.__heap.append(value)
self.__size += 1
self.__swap_up(self.__size)
def __swap_down(self, i: int) -> None:
- """ Swap the element down """
+ """Swap the element down"""
while self.__size >= 2 * i:
- if 2 * i + 1 > self.__size:
+ if 2 * i + 1 > self.__size: # noqa: SIM114
+ bigger_child = 2 * i
+ elif self.__heap[2 * i] > self.__heap[2 * i + 1]:
bigger_child = 2 * i
else:
- if self.__heap[2 * i] > self.__heap[2 * i + 1]:
- bigger_child = 2 * i
- else:
- bigger_child = 2 * i + 1
+ bigger_child = 2 * i + 1
temporary = self.__heap[i]
if self.__heap[i] < self.__heap[bigger_child]:
self.__heap[i] = self.__heap[bigger_child]
@@ -52,7 +51,7 @@ def __swap_down(self, i: int) -> None:
i = bigger_child
def pop(self) -> int:
- """ Pop the root element """
+ """Pop the root element"""
max_value = self.__heap[1]
self.__heap[1] = self.__heap[self.__size]
self.__size -= 1
@@ -65,7 +64,7 @@ def get_list(self):
return self.__heap[1:]
def __len__(self):
- """ Length of the array """
+ """Length of the array"""
return self.__size
diff --git a/data_structures/heap/min_heap.py b/data_structures/heap/min_heap.py
index 9265c4839536..577b98d788a1 100644
--- a/data_structures/heap/min_heap.py
+++ b/data_structures/heap/min_heap.py
@@ -27,7 +27,7 @@ class MinHeap:
>>> myMinHeap.decrease_key(b, -17)
>>> print(b)
Node(B, -17)
- >>> print(myMinHeap["B"])
+ >>> myMinHeap["B"]
-17
"""
@@ -52,28 +52,28 @@ def get_value(self, key):
return self.heap_dict[key]
def build_heap(self, array):
- lastIdx = len(array) - 1
- startFrom = self.get_parent_idx(lastIdx)
+ last_idx = len(array) - 1
+ start_from = self.get_parent_idx(last_idx)
for idx, i in enumerate(array):
self.idx_of_element[i] = idx
self.heap_dict[i.name] = i.val
- for i in range(startFrom, -1, -1):
+ for i in range(start_from, -1, -1):
self.sift_down(i, array)
return array
# this is min-heapify method
def sift_down(self, idx, array):
while True:
- l = self.get_left_child_idx(idx) # noqa: E741
- r = self.get_right_child_idx(idx)
+ left = self.get_left_child_idx(idx)
+ right = self.get_right_child_idx(idx)
smallest = idx
- if l < len(array) and array[l] < array[idx]:
- smallest = l
- if r < len(array) and array[r] < array[smallest]:
- smallest = r
+ if left < len(array) and array[left] < array[idx]:
+ smallest = left
+ if right < len(array) and array[right] < array[smallest]:
+ smallest = right
if smallest != idx:
array[idx], array[smallest] = array[smallest], array[idx]
@@ -121,14 +121,14 @@ def insert(self, node):
self.sift_up(len(self.heap) - 1)
def is_empty(self):
- return True if len(self.heap) == 0 else False
-
- def decrease_key(self, node, newValue):
- assert (
- self.heap[self.idx_of_element[node]].val > newValue
- ), "newValue must be less that current value"
- node.val = newValue
- self.heap_dict[node.name] = newValue
+ return len(self.heap) == 0
+
+ def decrease_key(self, node, new_value):
+ assert self.heap[self.idx_of_element[node]].val > new_value, (
+ "newValue must be less that current value"
+ )
+ node.val = new_value
+ self.heap_dict[node.name] = new_value
self.sift_up(self.idx_of_element[node])
@@ -143,7 +143,7 @@ def decrease_key(self, node, newValue):
# Use one of these two ways to generate Min-Heap
# Generating Min-Heap from array
-myMinHeap = MinHeap([r, b, a, x, e])
+my_min_heap = MinHeap([r, b, a, x, e])
# Generating Min-Heap by Insert method
# myMinHeap.insert(a)
@@ -154,14 +154,14 @@ def decrease_key(self, node, newValue):
# Before
print("Min Heap - before decrease key")
-for i in myMinHeap.heap:
+for i in my_min_heap.heap:
print(i)
print("Min Heap - After decrease key of node [B -> -17]")
-myMinHeap.decrease_key(b, -17)
+my_min_heap.decrease_key(b, -17)
# After
-for i in myMinHeap.heap:
+for i in my_min_heap.heap:
print(i)
if __name__ == "__main__":
diff --git a/data_structures/heap/randomized_heap.py b/data_structures/heap/randomized_heap.py
index 0ddc2272efe8..12888c1f4089 100644
--- a/data_structures/heap/randomized_heap.py
+++ b/data_structures/heap/randomized_heap.py
@@ -3,9 +3,10 @@
from __future__ import annotations
import random
-from typing import Generic, Iterable, List, Optional, TypeVar
+from collections.abc import Iterable
+from typing import Any, Generic, TypeVar
-T = TypeVar("T")
+T = TypeVar("T", bound=bool)
class RandomizedHeapNode(Generic[T]):
@@ -16,19 +17,45 @@ class RandomizedHeapNode(Generic[T]):
def __init__(self, value: T) -> None:
self._value: T = value
- self.left: Optional[RandomizedHeapNode[T]] = None
- self.right: Optional[RandomizedHeapNode[T]] = None
+ self.left: RandomizedHeapNode[T] | None = None
+ self.right: RandomizedHeapNode[T] | None = None
@property
def value(self) -> T:
- """Return the value of the node."""
+ """
+ Return the value of the node.
+
+ >>> rhn = RandomizedHeapNode(10)
+ >>> rhn.value
+ 10
+ >>> rhn = RandomizedHeapNode(-10)
+ >>> rhn.value
+ -10
+ """
return self._value
@staticmethod
def merge(
- root1: Optional[RandomizedHeapNode[T]], root2: Optional[RandomizedHeapNode[T]]
- ) -> Optional[RandomizedHeapNode[T]]:
- """Merge 2 nodes together."""
+ root1: RandomizedHeapNode[T] | None, root2: RandomizedHeapNode[T] | None
+ ) -> RandomizedHeapNode[T] | None:
+ """
+ Merge 2 nodes together.
+
+ >>> rhn1 = RandomizedHeapNode(10)
+ >>> rhn2 = RandomizedHeapNode(20)
+ >>> RandomizedHeapNode.merge(rhn1, rhn2).value
+ 10
+
+ >>> rhn1 = RandomizedHeapNode(20)
+ >>> rhn2 = RandomizedHeapNode(10)
+ >>> RandomizedHeapNode.merge(rhn1, rhn2).value
+ 10
+
+ >>> rhn1 = RandomizedHeapNode(5)
+ >>> rhn2 = RandomizedHeapNode(0)
+ >>> RandomizedHeapNode.merge(rhn1, rhn2).value
+ 0
+ """
if not root1:
return root2
@@ -69,15 +96,17 @@ class RandomizedHeap(Generic[T]):
[-1, 0, 1]
"""
- def __init__(self, data: Optional[Iterable[T]] = ()) -> None:
+ def __init__(self, data: Iterable[T] | None = ()) -> None:
"""
>>> rh = RandomizedHeap([3, 1, 3, 7])
>>> rh.to_sorted_list()
[1, 3, 3, 7]
"""
- self._root: Optional[RandomizedHeapNode[T]] = None
- for item in data:
- self.insert(item)
+ self._root: RandomizedHeapNode[T] | None = None
+
+ if data:
+ for item in data:
+ self.insert(item)
def insert(self, value: T) -> None:
"""
@@ -93,7 +122,7 @@ def insert(self, value: T) -> None:
"""
self._root = RandomizedHeapNode.merge(self._root, RandomizedHeapNode(value))
- def pop(self) -> T:
+ def pop(self) -> T | None:
"""
Pop the smallest value from the heap and return it.
@@ -111,7 +140,12 @@ def pop(self) -> T:
...
IndexError: Can't get top element for the empty heap.
"""
+
result = self.top()
+
+ if self._root is None:
+ return None
+
self._root = RandomizedHeapNode.merge(self._root.left, self._root.right)
return result
@@ -138,7 +172,7 @@ def top(self) -> T:
raise IndexError("Can't get top element for the empty heap.")
return self._root.value
- def clear(self):
+ def clear(self) -> None:
"""
Clear the heap.
@@ -151,7 +185,7 @@ def clear(self):
"""
self._root = None
- def to_sorted_list(self) -> List[T]:
+ def to_sorted_list(self) -> list[Any]:
"""
Returns sorted list containing all the values in the heap.
diff --git a/data_structures/heap/skew_heap.py b/data_structures/heap/skew_heap.py
index 417a383f733e..0839db711cb1 100644
--- a/data_structures/heap/skew_heap.py
+++ b/data_structures/heap/skew_heap.py
@@ -2,9 +2,10 @@
from __future__ import annotations
-from typing import Generic, Iterable, Iterator, Optional, TypeVar
+from collections.abc import Iterable, Iterator
+from typing import Any, Generic, TypeVar
-T = TypeVar("T")
+T = TypeVar("T", bound=bool)
class SkewNode(Generic[T]):
@@ -15,19 +16,60 @@ class SkewNode(Generic[T]):
def __init__(self, value: T) -> None:
self._value: T = value
- self.left: Optional[SkewNode[T]] = None
- self.right: Optional[SkewNode[T]] = None
+ self.left: SkewNode[T] | None = None
+ self.right: SkewNode[T] | None = None
@property
def value(self) -> T:
- """Return the value of the node."""
+ """
+ Return the value of the node.
+
+ >>> SkewNode(0).value
+ 0
+ >>> SkewNode(3.14159).value
+ 3.14159
+ >>> SkewNode("hello").value
+ 'hello'
+ >>> SkewNode(None).value
+
+ >>> SkewNode(True).value
+ True
+ >>> SkewNode([]).value
+ []
+ >>> SkewNode({}).value
+ {}
+ >>> SkewNode(set()).value
+ set()
+ >>> SkewNode(0.0).value
+ 0.0
+ >>> SkewNode(-1e-10).value
+ -1e-10
+ >>> SkewNode(10).value
+ 10
+ >>> SkewNode(-10.5).value
+ -10.5
+ >>> SkewNode().value
+ Traceback (most recent call last):
+ ...
+ TypeError: SkewNode.__init__() missing 1 required positional argument: 'value'
+ """
return self._value
@staticmethod
def merge(
- root1: Optional[SkewNode[T]], root2: Optional[SkewNode[T]]
- ) -> Optional[SkewNode[T]]:
- """Merge 2 nodes together."""
+ root1: SkewNode[T] | None, root2: SkewNode[T] | None
+ ) -> SkewNode[T] | None:
+ """
+ Merge 2 nodes together.
+ >>> SkewNode.merge(SkewNode(10),SkewNode(-10.5)).value
+ -10.5
+ >>> SkewNode.merge(SkewNode(10),SkewNode(10.5)).value
+ 10
+ >>> SkewNode.merge(SkewNode(10),SkewNode(10)).value
+ 10
+ >>> SkewNode.merge(SkewNode(-100),SkewNode(-10.5)).value
+ -100
+ """
if not root1:
return root2
@@ -51,7 +93,7 @@ class SkewHeap(Generic[T]):
values. Both operations take O(logN) time where N is the size of the
structure.
Wiki: https://en.wikipedia.org/wiki/Skew_heap
- Visualisation: https://www.cs.usfca.edu/~galles/visualization/SkewHeap.html
+ Visualization: https://www.cs.usfca.edu/~galles/visualization/SkewHeap.html
>>> list(SkewHeap([2, 3, 1, 5, 1, 7]))
[1, 1, 2, 3, 5, 7]
@@ -69,15 +111,16 @@ class SkewHeap(Generic[T]):
[-1, 0, 1]
"""
- def __init__(self, data: Optional[Iterable[T]] = ()) -> None:
+ def __init__(self, data: Iterable[T] | None = ()) -> None:
"""
>>> sh = SkewHeap([3, 1, 3, 7])
>>> list(sh)
[1, 3, 3, 7]
"""
- self._root: Optional[SkewNode[T]] = None
- for item in data:
- self.insert(item)
+ self._root: SkewNode[T] | None = None
+ if data:
+ for item in data:
+ self.insert(item)
def __bool__(self) -> bool:
"""
@@ -103,7 +146,7 @@ def __iter__(self) -> Iterator[T]:
>>> list(sh)
[1, 3, 3, 7]
"""
- result = []
+ result: list[Any] = []
while self:
result.append(self.pop())
@@ -127,7 +170,7 @@ def insert(self, value: T) -> None:
"""
self._root = SkewNode.merge(self._root, SkewNode(value))
- def pop(self) -> T:
+ def pop(self) -> T | None:
"""
Pop the smallest value from the heap and return it.
@@ -146,7 +189,9 @@ def pop(self) -> T:
IndexError: Can't get top element for the empty heap.
"""
result = self.top()
- self._root = SkewNode.merge(self._root.left, self._root.right)
+ self._root = (
+ SkewNode.merge(self._root.left, self._root.right) if self._root else None
+ )
return result
@@ -172,7 +217,7 @@ def top(self) -> T:
raise IndexError("Can't get top element for the empty heap.")
return self._root.value
- def clear(self):
+ def clear(self) -> None:
"""
Clear the heap.
diff --git a/greedy_method/__init__.py b/data_structures/kd_tree/__init__.py
similarity index 100%
rename from greedy_method/__init__.py
rename to data_structures/kd_tree/__init__.py
diff --git a/data_structures/kd_tree/build_kdtree.py b/data_structures/kd_tree/build_kdtree.py
new file mode 100644
index 000000000000..074a5dac4d42
--- /dev/null
+++ b/data_structures/kd_tree/build_kdtree.py
@@ -0,0 +1,43 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11532
+# https://github.com/TheAlgorithms/Python/pull/11532
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+from data_structures.kd_tree.kd_node import KDNode
+
+
+def build_kdtree(points: list[list[float]], depth: int = 0) -> KDNode | None:
+ """
+ Builds a KD-Tree from a list of points.
+
+ Args:
+ points: The list of points to build the KD-Tree from.
+ depth: The current depth in the tree
+ (used to determine axis for splitting).
+
+ Returns:
+ The root node of the KD-Tree,
+ or None if no points are provided.
+ """
+ if not points:
+ return None
+
+ k = len(points[0]) # Dimensionality of the points
+ axis = depth % k
+
+ # Sort point list and choose median as pivot element
+ points.sort(key=lambda point: point[axis])
+ median_idx = len(points) // 2
+
+ # Create node and construct subtrees
+ left_points = points[:median_idx]
+ right_points = points[median_idx + 1 :]
+
+ return KDNode(
+ point=points[median_idx],
+ left=build_kdtree(left_points, depth + 1),
+ right=build_kdtree(right_points, depth + 1),
+ )
diff --git a/images/__init__.py b/data_structures/kd_tree/example/__init__.py
similarity index 100%
rename from images/__init__.py
rename to data_structures/kd_tree/example/__init__.py
diff --git a/data_structures/kd_tree/example/example_usage.py b/data_structures/kd_tree/example/example_usage.py
new file mode 100644
index 000000000000..892c3b8c4a2a
--- /dev/null
+++ b/data_structures/kd_tree/example/example_usage.py
@@ -0,0 +1,46 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11532
+# https://github.com/TheAlgorithms/Python/pull/11532
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+import numpy as np
+
+from data_structures.kd_tree.build_kdtree import build_kdtree
+from data_structures.kd_tree.example.hypercube_points import hypercube_points
+from data_structures.kd_tree.nearest_neighbour_search import nearest_neighbour_search
+
+
+def main() -> None:
+ """
+ Demonstrates the use of KD-Tree by building it from random points
+ in a 10-dimensional hypercube and performing a nearest neighbor search.
+ """
+ num_points: int = 5000
+ cube_size: float = 10.0 # Size of the hypercube (edge length)
+ num_dimensions: int = 10
+
+ # Generate random points within the hypercube
+ points: np.ndarray = hypercube_points(num_points, cube_size, num_dimensions)
+ hypercube_kdtree = build_kdtree(points.tolist())
+
+ # Generate a random query point within the same space
+ rng = np.random.default_rng()
+ query_point: list[float] = rng.random(num_dimensions).tolist()
+
+ # Perform nearest neighbor search
+ nearest_point, nearest_dist, nodes_visited = nearest_neighbour_search(
+ hypercube_kdtree, query_point
+ )
+
+ # Print the results
+ print(f"Query point: {query_point}")
+ print(f"Nearest point: {nearest_point}")
+ print(f"Distance: {nearest_dist:.4f}")
+ print(f"Nodes visited: {nodes_visited}")
+
+
+if __name__ == "__main__":
+ main()
diff --git a/data_structures/kd_tree/example/hypercube_points.py b/data_structures/kd_tree/example/hypercube_points.py
new file mode 100644
index 000000000000..66744856e6d5
--- /dev/null
+++ b/data_structures/kd_tree/example/hypercube_points.py
@@ -0,0 +1,29 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11532
+# https://github.com/TheAlgorithms/Python/pull/11532
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+import numpy as np
+
+
+def hypercube_points(
+ num_points: int, hypercube_size: float, num_dimensions: int
+) -> np.ndarray:
+ """
+ Generates random points uniformly distributed within an n-dimensional hypercube.
+
+ Args:
+ num_points: Number of points to generate.
+ hypercube_size: Size of the hypercube.
+ num_dimensions: Number of dimensions of the hypercube.
+
+ Returns:
+ An array of shape (num_points, num_dimensions)
+ with generated points.
+ """
+ rng = np.random.default_rng()
+ shape = (num_points, num_dimensions)
+ return hypercube_size * rng.random(shape)
diff --git a/data_structures/kd_tree/kd_node.py b/data_structures/kd_tree/kd_node.py
new file mode 100644
index 000000000000..5a22ef609077
--- /dev/null
+++ b/data_structures/kd_tree/kd_node.py
@@ -0,0 +1,38 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11532
+# https://github.com/TheAlgorithms/Python/pull/11532
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+from __future__ import annotations
+
+
+class KDNode:
+ """
+ Represents a node in a KD-Tree.
+
+ Attributes:
+ point: The point stored in this node.
+ left: The left child node.
+ right: The right child node.
+ """
+
+ def __init__(
+ self,
+ point: list[float],
+ left: KDNode | None = None,
+ right: KDNode | None = None,
+ ) -> None:
+ """
+ Initializes a KDNode with the given point and child nodes.
+
+ Args:
+ point (list[float]): The point stored in this node.
+ left (Optional[KDNode]): The left child node.
+ right (Optional[KDNode]): The right child node.
+ """
+ self.point = point
+ self.left = left
+ self.right = right
diff --git a/data_structures/kd_tree/nearest_neighbour_search.py b/data_structures/kd_tree/nearest_neighbour_search.py
new file mode 100644
index 000000000000..8104944c08f0
--- /dev/null
+++ b/data_structures/kd_tree/nearest_neighbour_search.py
@@ -0,0 +1,79 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11532
+# https://github.com/TheAlgorithms/Python/pull/11532
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+from data_structures.kd_tree.kd_node import KDNode
+
+
+def nearest_neighbour_search(
+ root: KDNode | None, query_point: list[float]
+) -> tuple[list[float] | None, float, int]:
+ """
+ Performs a nearest neighbor search in a KD-Tree for a given query point.
+
+ Args:
+ root (KDNode | None): The root node of the KD-Tree.
+ query_point (list[float]): The point for which the nearest neighbor
+ is being searched.
+
+ Returns:
+ tuple[list[float] | None, float, int]:
+ - The nearest point found in the KD-Tree to the query point,
+ or None if no point is found.
+ - The squared distance to the nearest point.
+ - The number of nodes visited during the search.
+ """
+ nearest_point: list[float] | None = None
+ nearest_dist: float = float("inf")
+ nodes_visited: int = 0
+
+ def search(node: KDNode | None, depth: int = 0) -> None:
+ """
+ Recursively searches for the nearest neighbor in the KD-Tree.
+
+ Args:
+ node: The current node in the KD-Tree.
+ depth: The current depth in the KD-Tree.
+ """
+ nonlocal nearest_point, nearest_dist, nodes_visited
+ if node is None:
+ return
+
+ nodes_visited += 1
+
+ # Calculate the current distance (squared distance)
+ current_point = node.point
+ current_dist = sum(
+ (query_coord - point_coord) ** 2
+ for query_coord, point_coord in zip(query_point, current_point)
+ )
+
+ # Update nearest point if the current node is closer
+ if nearest_point is None or current_dist < nearest_dist:
+ nearest_point = current_point
+ nearest_dist = current_dist
+
+ # Determine which subtree to search first (based on axis and query point)
+ k = len(query_point) # Dimensionality of points
+ axis = depth % k
+
+ if query_point[axis] <= current_point[axis]:
+ nearer_subtree = node.left
+ further_subtree = node.right
+ else:
+ nearer_subtree = node.right
+ further_subtree = node.left
+
+ # Search the nearer subtree first
+ search(nearer_subtree, depth + 1)
+
+ # If the further subtree has a closer point
+ if (query_point[axis] - current_point[axis]) ** 2 < nearest_dist:
+ search(further_subtree, depth + 1)
+
+ search(root, 0)
+ return nearest_point, nearest_dist, nodes_visited
diff --git a/traversals/__init__.py b/data_structures/kd_tree/tests/__init__.py
similarity index 100%
rename from traversals/__init__.py
rename to data_structures/kd_tree/tests/__init__.py
diff --git a/data_structures/kd_tree/tests/test_kdtree.py b/data_structures/kd_tree/tests/test_kdtree.py
new file mode 100644
index 000000000000..d6a4a66dd24d
--- /dev/null
+++ b/data_structures/kd_tree/tests/test_kdtree.py
@@ -0,0 +1,108 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11532
+# https://github.com/TheAlgorithms/Python/pull/11532
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+import numpy as np
+import pytest
+
+from data_structures.kd_tree.build_kdtree import build_kdtree
+from data_structures.kd_tree.example.hypercube_points import hypercube_points
+from data_structures.kd_tree.kd_node import KDNode
+from data_structures.kd_tree.nearest_neighbour_search import nearest_neighbour_search
+
+
+@pytest.mark.parametrize(
+ ("num_points", "cube_size", "num_dimensions", "depth", "expected_result"),
+ [
+ (0, 10.0, 2, 0, None), # Empty points list
+ (10, 10.0, 2, 2, KDNode), # Depth = 2, 2D points
+ (10, 10.0, 3, -2, KDNode), # Depth = -2, 3D points
+ ],
+)
+def test_build_kdtree(num_points, cube_size, num_dimensions, depth, expected_result):
+ """
+ Test that KD-Tree is built correctly.
+
+ Cases:
+ - Empty points list.
+ - Positive depth value.
+ - Negative depth value.
+ """
+ points = (
+ hypercube_points(num_points, cube_size, num_dimensions).tolist()
+ if num_points > 0
+ else []
+ )
+
+ kdtree = build_kdtree(points, depth=depth)
+
+ if expected_result is None:
+ # Empty points list case
+ assert kdtree is None, f"Expected None for empty points list, got {kdtree}"
+ else:
+ # Check if root node is not None
+ assert kdtree is not None, "Expected a KDNode, got None"
+
+ # Check if root has correct dimensions
+ assert len(kdtree.point) == num_dimensions, (
+ f"Expected point dimension {num_dimensions}, got {len(kdtree.point)}"
+ )
+
+ # Check that the tree is balanced to some extent (simplistic check)
+ assert isinstance(kdtree, KDNode), (
+ f"Expected KDNode instance, got {type(kdtree)}"
+ )
+
+
+def test_nearest_neighbour_search():
+ """
+ Test the nearest neighbor search function.
+ """
+ num_points = 10
+ cube_size = 10.0
+ num_dimensions = 2
+ points = hypercube_points(num_points, cube_size, num_dimensions)
+ kdtree = build_kdtree(points.tolist())
+
+ rng = np.random.default_rng()
+ query_point = rng.random(num_dimensions).tolist()
+
+ nearest_point, nearest_dist, nodes_visited = nearest_neighbour_search(
+ kdtree, query_point
+ )
+
+ # Check that nearest point is not None
+ assert nearest_point is not None
+
+ # Check that distance is a non-negative number
+ assert nearest_dist >= 0
+
+ # Check that nodes visited is a non-negative integer
+ assert nodes_visited >= 0
+
+
+def test_edge_cases():
+ """
+ Test edge cases such as an empty KD-Tree.
+ """
+ empty_kdtree = build_kdtree([])
+ query_point = [0.0] * 2 # Using a default 2D query point
+
+ nearest_point, nearest_dist, nodes_visited = nearest_neighbour_search(
+ empty_kdtree, query_point
+ )
+
+ # With an empty KD-Tree, nearest_point should be None
+ assert nearest_point is None
+ assert nearest_dist == float("inf")
+ assert nodes_visited == 0
+
+
+if __name__ == "__main__":
+ import pytest
+
+ pytest.main()
diff --git a/data_structures/linked_list/__init__.py b/data_structures/linked_list/__init__.py
index 3ddfea5c5abf..00ef337a1211 100644
--- a/data_structures/linked_list/__init__.py
+++ b/data_structures/linked_list/__init__.py
@@ -1,20 +1,84 @@
+"""
+Linked Lists consists of Nodes.
+Nodes contain data and also may link to other nodes:
+ - Head Node: First node, the address of the
+ head node gives us access of the complete list
+ - Last node: points to null
+"""
+
+from __future__ import annotations
+
+from typing import Any
+
+
class Node:
- def __init__(self, item, next):
+ def __init__(self, item: Any, next: Any) -> None: # noqa: A002
self.item = item
self.next = next
class LinkedList:
- def __init__(self):
- self.head = None
+ def __init__(self) -> None:
+ self.head: Node | None = None
self.size = 0
- def add(self, item):
- self.head = Node(item, self.head)
+ def add(self, item: Any, position: int = 0) -> None:
+ """
+ Add an item to the LinkedList at the specified position.
+ Default position is 0 (the head).
+
+ Args:
+ item (Any): The item to add to the LinkedList.
+ position (int, optional): The position at which to add the item.
+ Defaults to 0.
+
+ Raises:
+ ValueError: If the position is negative or out of bounds.
+
+ >>> linked_list = LinkedList()
+ >>> linked_list.add(1)
+ >>> linked_list.add(2)
+ >>> linked_list.add(3)
+ >>> linked_list.add(4, 2)
+ >>> print(linked_list)
+ 3 --> 2 --> 4 --> 1
+
+ # Test adding to a negative position
+ >>> linked_list.add(5, -3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Position must be non-negative
+
+ # Test adding to an out-of-bounds position
+ >>> linked_list.add(5,7)
+ Traceback (most recent call last):
+ ...
+ ValueError: Out of bounds
+ >>> linked_list.add(5, 4)
+ >>> print(linked_list)
+ 3 --> 2 --> 4 --> 1 --> 5
+ """
+ if position < 0:
+ raise ValueError("Position must be non-negative")
+
+ if position == 0 or self.head is None:
+ new_node = Node(item, self.head)
+ self.head = new_node
+ else:
+ current = self.head
+ for _ in range(position - 1):
+ current = current.next
+ if current is None:
+ raise ValueError("Out of bounds")
+ new_node = Node(item, current.next)
+ current.next = new_node
self.size += 1
- def remove(self):
- if self.is_empty():
+ def remove(self) -> Any:
+ # Switched 'self.is_empty()' to 'self.head is None'
+ # because mypy was considering the possibility that 'self.head'
+ # can be None in below else part and giving error
+ if self.head is None:
return None
else:
item = self.head.item
@@ -22,10 +86,33 @@ def remove(self):
self.size -= 1
return item
- def is_empty(self):
+ def is_empty(self) -> bool:
return self.head is None
- def __len__(self):
+ def __str__(self) -> str:
+ """
+ >>> linked_list = LinkedList()
+ >>> linked_list.add(23)
+ >>> linked_list.add(14)
+ >>> linked_list.add(9)
+ >>> print(linked_list)
+ 9 --> 14 --> 23
+ """
+ if self.is_empty():
+ return ""
+ else:
+ iterate = self.head
+ item_str = ""
+ item_list: list[str] = []
+ while iterate:
+ item_list.append(str(iterate.item))
+ iterate = iterate.next
+
+ item_str = " --> ".join(item_list)
+
+ return item_str
+
+ def __len__(self) -> int:
"""
>>> linked_list = LinkedList()
>>> len(linked_list)
diff --git a/data_structures/linked_list/circular_linked_list.py b/data_structures/linked_list/circular_linked_list.py
index f67c1e8f2cf7..bb64441d4560 100644
--- a/data_structures/linked_list/circular_linked_list.py
+++ b/data_structures/linked_list/circular_linked_list.py
@@ -1,87 +1,156 @@
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
from typing import Any
+@dataclass
class Node:
- def __init__(self, data: Any):
- self.data = data
- self.next = None
+ data: Any
+ next_node: Node | None = None
+@dataclass
class CircularLinkedList:
- def __init__(self):
- self.head = None
- self.tail = None
-
- def __iter__(self):
+ head: Node | None = None # Reference to the head (first node)
+ tail: Node | None = None # Reference to the tail (last node)
+
+ def __iter__(self) -> Iterator[Any]:
+ """
+ Iterate through all nodes in the Circular Linked List yielding their data.
+ Yields:
+ The data of each node in the linked list.
+ """
node = self.head
- while self.head:
+ while node:
yield node.data
- node = node.next
+ node = node.next_node
if node == self.head:
break
def __len__(self) -> int:
- return len(tuple(iter(self)))
-
- def __repr__(self):
+ """
+ Get the length (number of nodes) in the Circular Linked List.
+ """
+ return sum(1 for _ in self)
+
+ def __repr__(self) -> str:
+ """
+ Generate a string representation of the Circular Linked List.
+ Returns:
+ A string of the format "1->2->....->N".
+ """
return "->".join(str(item) for item in iter(self))
def insert_tail(self, data: Any) -> None:
+ """
+ Insert a node with the given data at the end of the Circular Linked List.
+ """
self.insert_nth(len(self), data)
def insert_head(self, data: Any) -> None:
+ """
+ Insert a node with the given data at the beginning of the Circular Linked List.
+ """
self.insert_nth(0, data)
def insert_nth(self, index: int, data: Any) -> None:
+ """
+ Insert the data of the node at the nth pos in the Circular Linked List.
+ Args:
+ index: The index at which the data should be inserted.
+ data: The data to be inserted.
+
+ Raises:
+ IndexError: If the index is out of range.
+ """
if index < 0 or index > len(self):
raise IndexError("list index out of range.")
- new_node = Node(data)
+ new_node: Node = Node(data)
if self.head is None:
- new_node.next = new_node # first node points itself
+ new_node.next_node = new_node # First node points to itself
self.tail = self.head = new_node
- elif index == 0: # insert at head
- new_node.next = self.head
- self.head = self.tail.next = new_node
+ elif index == 0: # Insert at the head
+ new_node.next_node = self.head
+ assert self.tail is not None # List is not empty, tail exists
+ self.head = self.tail.next_node = new_node
else:
- temp = self.head
+ temp: Node | None = self.head
for _ in range(index - 1):
- temp = temp.next
- new_node.next = temp.next
- temp.next = new_node
- if index == len(self) - 1: # insert at tail
+ assert temp is not None
+ temp = temp.next_node
+ assert temp is not None
+ new_node.next_node = temp.next_node
+ temp.next_node = new_node
+ if index == len(self) - 1: # Insert at the tail
self.tail = new_node
- def delete_front(self):
+ def delete_front(self) -> Any:
+ """
+ Delete and return the data of the node at the front of the Circular Linked List.
+ Raises:
+ IndexError: If the list is empty.
+ """
return self.delete_nth(0)
- def delete_tail(self) -> None:
+ def delete_tail(self) -> Any:
+ """
+ Delete and return the data of the node at the end of the Circular Linked List.
+ Returns:
+ Any: The data of the deleted node.
+ Raises:
+ IndexError: If the index is out of range.
+ """
return self.delete_nth(len(self) - 1)
- def delete_nth(self, index: int = 0):
+ def delete_nth(self, index: int = 0) -> Any:
+ """
+ Delete and return the data of the node at the nth pos in Circular Linked List.
+ Args:
+ index (int): The index of the node to be deleted. Defaults to 0.
+ Returns:
+ Any: The data of the deleted node.
+ Raises:
+ IndexError: If the index is out of range.
+ """
if not 0 <= index < len(self):
raise IndexError("list index out of range.")
- delete_node = self.head
- if self.head == self.tail: # just one node
+
+ assert self.head is not None
+ assert self.tail is not None
+ delete_node: Node = self.head
+ if self.head == self.tail: # Just one node
self.head = self.tail = None
- elif index == 0: # delete head node
- self.tail.next = self.tail.next.next
- self.head = self.head.next
+ elif index == 0: # Delete head node
+ assert self.tail.next_node is not None
+ self.tail.next_node = self.tail.next_node.next_node
+ self.head = self.head.next_node
else:
- temp = self.head
+ temp: Node | None = self.head
for _ in range(index - 1):
- temp = temp.next
- delete_node = temp.next
- temp.next = temp.next.next
- if index == len(self) - 1: # delete at tail
+ assert temp is not None
+ temp = temp.next_node
+ assert temp is not None
+ assert temp.next_node is not None
+ delete_node = temp.next_node
+ temp.next_node = temp.next_node.next_node
+ if index == len(self) - 1: # Delete at tail
self.tail = temp
return delete_node.data
- def is_empty(self):
+ def is_empty(self) -> bool:
+ """
+ Check if the Circular Linked List is empty.
+ Returns:
+ bool: True if the list is empty, False otherwise.
+ """
return len(self) == 0
def test_circular_linked_list() -> None:
"""
+ Test cases for the CircularLinkedList class.
>>> test_circular_linked_list()
"""
circular_linked_list = CircularLinkedList()
@@ -91,25 +160,25 @@ def test_circular_linked_list() -> None:
try:
circular_linked_list.delete_front()
- assert False # This should not happen
+ raise AssertionError # This should not happen
except IndexError:
assert True # This should happen
try:
circular_linked_list.delete_tail()
- assert False # This should not happen
+ raise AssertionError # This should not happen
except IndexError:
assert True # This should happen
try:
circular_linked_list.delete_nth(-1)
- assert False
+ raise AssertionError
except IndexError:
assert True
try:
circular_linked_list.delete_nth(0)
- assert False
+ raise AssertionError
except IndexError:
assert True
@@ -122,7 +191,7 @@ def test_circular_linked_list() -> None:
circular_linked_list.insert_tail(6)
assert str(circular_linked_list) == "->".join(str(i) for i in range(1, 7))
circular_linked_list.insert_head(0)
- assert str(circular_linked_list) == "->".join(str(i) for i in range(0, 7))
+ assert str(circular_linked_list) == "->".join(str(i) for i in range(7))
assert circular_linked_list.delete_front() == 0
assert circular_linked_list.delete_tail() == 6
diff --git a/data_structures/linked_list/deque_doubly.py b/data_structures/linked_list/deque_doubly.py
index 894f91d561cc..e554ead91c5a 100644
--- a/data_structures/linked_list/deque_doubly.py
+++ b/data_structures/linked_list/deque_doubly.py
@@ -9,10 +9,10 @@
class _DoublyLinkedBase:
- """ A Private class (to be inherited) """
+ """A Private class (to be inherited)"""
class _Node:
- __slots__ = "_prev", "_data", "_next"
+ __slots__ = "_data", "_next", "_prev"
def __init__(self, link_p, element, link_n):
self._prev = link_p
@@ -20,8 +20,8 @@ def __init__(self, link_p, element, link_n):
self._next = link_n
def has_next_and_prev(self):
- return " Prev -> {}, Next -> {}".format(
- self._prev is not None, self._next is not None
+ return (
+ f" Prev -> {self._prev is not None}, Next -> {self._next is not None}"
)
def __init__(self):
diff --git a/data_structures/linked_list/doubly_linked_list.py b/data_structures/linked_list/doubly_linked_list.py
index 0eb3cf101a3e..bd3445f9f6c5 100644
--- a/data_structures/linked_list/doubly_linked_list.py
+++ b/data_structures/linked_list/doubly_linked_list.py
@@ -51,7 +51,7 @@ def __len__(self):
>>> len(linked_list) == 5
True
"""
- return len(tuple(iter(self)))
+ return sum(1 for _ in self)
def insert_at_head(self, data):
self.insert_at_nth(0, data)
@@ -64,11 +64,11 @@ def insert_at_nth(self, index: int, data):
>>> linked_list = DoublyLinkedList()
>>> linked_list.insert_at_nth(-1, 666)
Traceback (most recent call last):
- ....
+ ....
IndexError: list index out of range
>>> linked_list.insert_at_nth(1, 666)
Traceback (most recent call last):
- ....
+ ....
IndexError: list index out of range
>>> linked_list.insert_at_nth(0, 2)
>>> linked_list.insert_at_nth(0, 1)
@@ -78,10 +78,12 @@ def insert_at_nth(self, index: int, data):
'1->2->3->4'
>>> linked_list.insert_at_nth(5, 5)
Traceback (most recent call last):
- ....
+ ....
IndexError: list index out of range
"""
- if not 0 <= index <= len(self):
+ length = len(self)
+
+ if not 0 <= index <= length:
raise IndexError("list index out of range")
new_node = Node(data)
if self.head is None:
@@ -90,13 +92,13 @@ def insert_at_nth(self, index: int, data):
self.head.previous = new_node
new_node.next = self.head
self.head = new_node
- elif index == len(self):
+ elif index == length:
self.tail.next = new_node
new_node.previous = self.tail
self.tail = new_node
else:
temp = self.head
- for i in range(0, index):
+ for _ in range(index):
temp = temp.next
temp.previous.next = new_node
new_node.previous = temp.previous
@@ -114,7 +116,7 @@ def delete_at_nth(self, index: int):
>>> linked_list = DoublyLinkedList()
>>> linked_list.delete_at_nth(0)
Traceback (most recent call last):
- ....
+ ....
IndexError: list index out of range
>>> for i in range(0, 5):
... linked_list.insert_at_nth(i, i + 1)
@@ -128,24 +130,26 @@ def delete_at_nth(self, index: int):
'2->4'
>>> linked_list.delete_at_nth(2)
Traceback (most recent call last):
- ....
+ ....
IndexError: list index out of range
"""
- if not 0 <= index <= len(self) - 1:
+ length = len(self)
+
+ if not 0 <= index <= length - 1:
raise IndexError("list index out of range")
delete_node = self.head # default first node
- if len(self) == 1:
+ if length == 1:
self.head = self.tail = None
elif index == 0:
self.head = self.head.next
self.head.previous = None
- elif index == len(self) - 1:
+ elif index == length - 1:
delete_node = self.tail
self.tail = self.tail.previous
self.tail.next = None
else:
temp = self.head
- for i in range(0, index):
+ for _ in range(index):
temp = temp.next
delete_node = temp
temp.next.previous = temp.previous
@@ -159,7 +163,7 @@ def delete(self, data) -> str:
if current.next:
current = current.next
else: # We have reached the end an no value matches
- return "No data matching given value"
+ raise ValueError("No data matching given value")
if current == self.head:
self.delete_head()
@@ -194,13 +198,13 @@ def test_doubly_linked_list() -> None:
try:
linked_list.delete_head()
- assert False # This should not happen.
+ raise AssertionError # This should not happen.
except IndexError:
assert True # This should happen.
try:
linked_list.delete_tail()
- assert False # This should not happen.
+ raise AssertionError # This should not happen.
except IndexError:
assert True # This should happen.
@@ -211,7 +215,7 @@ def test_doubly_linked_list() -> None:
linked_list.insert_at_head(0)
linked_list.insert_at_tail(11)
- assert str(linked_list) == "->".join(str(i) for i in range(0, 12))
+ assert str(linked_list) == "->".join(str(i) for i in range(12))
assert linked_list.delete_head() == 0
assert linked_list.delete_at_nth(9) == 10
diff --git a/data_structures/linked_list/doubly_linked_list_two.py b/data_structures/linked_list/doubly_linked_list_two.py
index 184b6966b5a9..8c93cddd5d31 100644
--- a/data_structures/linked_list/doubly_linked_list_two.py
+++ b/data_structures/linked_list/doubly_linked_list_two.py
@@ -9,25 +9,19 @@
Delete operation is more efficient
"""
+from dataclasses import dataclass
+from typing import Self
+
+@dataclass
class Node:
- def __init__(self, data: int, previous=None, next_node=None):
- self.data = data
- self.previous = previous
- self.next = next_node
+ data: int
+ previous: Self | None = None
+ next: Self | None = None
def __str__(self) -> str:
return f"{self.data}"
- def get_data(self) -> int:
- return self.data
-
- def get_next(self):
- return self.next
-
- def get_previous(self):
- return self.previous
-
class LinkedListIterator:
def __init__(self, head):
@@ -40,30 +34,30 @@ def __next__(self):
if not self.current:
raise StopIteration
else:
- value = self.current.get_data()
- self.current = self.current.get_next()
+ value = self.current.data
+ self.current = self.current.next
return value
+@dataclass
class LinkedList:
- def __init__(self):
- self.head = None # First node in list
- self.tail = None # Last node in list
+ head: Node | None = None # First node in list
+ tail: Node | None = None # Last node in list
def __str__(self):
current = self.head
nodes = []
while current is not None:
- nodes.append(current.get_data())
- current = current.get_next()
+ nodes.append(current.data)
+ current = current.next
return " ".join(str(node) for node in nodes)
def __contains__(self, value: int):
current = self.head
while current:
- if current.get_data() == value:
+ if current.data == value:
return True
- current = current.get_next()
+ current = current.next
return False
def __iter__(self):
@@ -71,16 +65,15 @@ def __iter__(self):
def get_head_data(self):
if self.head:
- return self.head.get_data()
+ return self.head.data
return None
def get_tail_data(self):
if self.tail:
- return self.tail.get_data()
+ return self.tail.data
return None
def set_head(self, node: Node) -> None:
-
if self.head is None:
self.head = node
self.tail = node
@@ -88,8 +81,9 @@ def set_head(self, node: Node) -> None:
self.insert_before_node(self.head, node)
def set_tail(self, node: Node) -> None:
- if self.head is None:
- self.set_head(node)
+ if self.tail is None:
+ self.head = node
+ self.tail = node
else:
self.insert_after_node(self.tail, node)
@@ -104,7 +98,7 @@ def insert_before_node(self, node: Node, node_to_insert: Node) -> None:
node_to_insert.next = node
node_to_insert.previous = node.previous
- if node.get_previous() is None:
+ if node.previous is None:
self.head = node_to_insert
else:
node.previous.next = node_to_insert
@@ -115,7 +109,7 @@ def insert_after_node(self, node: Node, node_to_insert: Node) -> None:
node_to_insert.previous = node
node_to_insert.next = node.next
- if node.get_next() is None:
+ if node.next is None:
self.tail = node_to_insert
else:
node.next.previous = node_to_insert
@@ -129,37 +123,35 @@ def insert_at_position(self, position: int, value: int) -> None:
while node:
if current_position == position:
self.insert_before_node(node, new_node)
- return None
+ return
current_position += 1
node = node.next
- self.insert_after_node(self.tail, new_node)
+ self.set_tail(new_node)
def get_node(self, item: int) -> Node:
node = self.head
while node:
- if node.get_data() == item:
+ if node.data == item:
return node
- node = node.get_next()
+ node = node.next
raise Exception("Node not found")
def delete_value(self, value):
- node = self.get_node(value)
-
- if node is not None:
+ if (node := self.get_node(value)) is not None:
if node == self.head:
- self.head = self.head.get_next()
+ self.head = self.head.next
if node == self.tail:
- self.tail = self.tail.get_previous()
+ self.tail = self.tail.previous
self.remove_node_pointers(node)
@staticmethod
def remove_node_pointers(node: Node) -> None:
- if node.get_next():
+ if node.next:
node.next.previous = node.previous
- if node.get_previous():
+ if node.previous:
node.previous.next = node.next
node.next = None
@@ -244,6 +236,22 @@ def create_linked_list() -> None:
7
8
9
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_at_position(position=1, value=10)
+ >>> str(linked_list)
+ '10'
+ >>> linked_list.insert_at_position(position=2, value=20)
+ >>> str(linked_list)
+ '10 20'
+ >>> linked_list.insert_at_position(position=1, value=30)
+ >>> str(linked_list)
+ '30 10 20'
+ >>> linked_list.insert_at_position(position=3, value=40)
+ >>> str(linked_list)
+ '30 10 40 20'
+ >>> linked_list.insert_at_position(position=5, value=50)
+ >>> str(linked_list)
+ '30 10 40 20 50'
"""
diff --git a/data_structures/linked_list/floyds_cycle_detection.py b/data_structures/linked_list/floyds_cycle_detection.py
new file mode 100644
index 000000000000..6c3f13760260
--- /dev/null
+++ b/data_structures/linked_list/floyds_cycle_detection.py
@@ -0,0 +1,150 @@
+"""
+Floyd's cycle detection algorithm is a popular algorithm used to detect cycles
+in a linked list. It uses two pointers, a slow pointer and a fast pointer,
+to traverse the linked list. The slow pointer moves one node at a time while the fast
+pointer moves two nodes at a time. If there is a cycle in the linked list,
+the fast pointer will eventually catch up to the slow pointer and they will
+meet at the same node. If there is no cycle, the fast pointer will reach the end of
+the linked list and the algorithm will terminate.
+
+For more information: https://en.wikipedia.org/wiki/Cycle_detection#Floyd's_tortoise_and_hare
+"""
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+from typing import Any, Self
+
+
+@dataclass
+class Node:
+ """
+ A class representing a node in a singly linked list.
+ """
+
+ data: Any
+ next_node: Self | None = None
+
+
+@dataclass
+class LinkedList:
+ """
+ A class representing a singly linked list.
+ """
+
+ head: Node | None = None
+
+ def __iter__(self) -> Iterator:
+ """
+ Iterates through the linked list.
+
+ Returns:
+ Iterator: An iterator over the linked list.
+
+ Examples:
+ >>> linked_list = LinkedList()
+ >>> list(linked_list)
+ []
+ >>> linked_list.add_node(1)
+ >>> tuple(linked_list)
+ (1,)
+ """
+ visited = []
+ node = self.head
+ while node:
+ # Avoid infinite loop in there's a cycle
+ if node in visited:
+ return
+ visited.append(node)
+ yield node.data
+ node = node.next_node
+
+ def add_node(self, data: Any) -> None:
+ """
+ Adds a new node to the end of the linked list.
+
+ Args:
+ data (Any): The data to be stored in the new node.
+
+ Examples:
+ >>> linked_list = LinkedList()
+ >>> linked_list.add_node(1)
+ >>> linked_list.add_node(2)
+ >>> linked_list.add_node(3)
+ >>> linked_list.add_node(4)
+ >>> tuple(linked_list)
+ (1, 2, 3, 4)
+ """
+ new_node = Node(data)
+
+ if self.head is None:
+ self.head = new_node
+ return
+
+ current_node = self.head
+ while current_node.next_node is not None:
+ current_node = current_node.next_node
+
+ current_node.next_node = new_node
+
+ def detect_cycle(self) -> bool:
+ """
+ Detects if there is a cycle in the linked list using
+ Floyd's cycle detection algorithm.
+
+ Returns:
+ bool: True if there is a cycle, False otherwise.
+
+ Examples:
+ >>> linked_list = LinkedList()
+ >>> linked_list.add_node(1)
+ >>> linked_list.add_node(2)
+ >>> linked_list.add_node(3)
+ >>> linked_list.add_node(4)
+
+ >>> linked_list.detect_cycle()
+ False
+
+ # Create a cycle in the linked list
+ >>> linked_list.head.next_node.next_node.next_node = linked_list.head.next_node
+
+ >>> linked_list.detect_cycle()
+ True
+ """
+ if self.head is None:
+ return False
+
+ slow_pointer: Node | None = self.head
+ fast_pointer: Node | None = self.head
+
+ while fast_pointer is not None and fast_pointer.next_node is not None:
+ slow_pointer = slow_pointer.next_node if slow_pointer else None
+ fast_pointer = fast_pointer.next_node.next_node
+ if slow_pointer == fast_pointer:
+ return True
+
+ return False
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ linked_list = LinkedList()
+ linked_list.add_node(1)
+ linked_list.add_node(2)
+ linked_list.add_node(3)
+ linked_list.add_node(4)
+
+ # Create a cycle in the linked list
+ # It first checks if the head, next_node, and next_node.next_node attributes of the
+ # linked list are not None to avoid any potential type errors.
+ if (
+ linked_list.head
+ and linked_list.head.next_node
+ and linked_list.head.next_node.next_node
+ ):
+ linked_list.head.next_node.next_node.next_node = linked_list.head.next_node
+
+ has_cycle = linked_list.detect_cycle()
+ print(has_cycle) # Output: True
diff --git a/data_structures/linked_list/from_sequence.py b/data_structures/linked_list/from_sequence.py
index 94b44f15037f..fa43f4d10e08 100644
--- a/data_structures/linked_list/from_sequence.py
+++ b/data_structures/linked_list/from_sequence.py
@@ -1,4 +1,4 @@
-# Recursive Prorgam to create a Linked List from a sequence and
+# Recursive Program to create a Linked List from a sequence and
# print a string representation of it.
diff --git a/data_structures/linked_list/has_loop.py b/data_structures/linked_list/has_loop.py
index 405ece7e27c8..f49e01579adc 100644
--- a/data_structures/linked_list/has_loop.py
+++ b/data_structures/linked_list/has_loop.py
@@ -1,3 +1,5 @@
+from __future__ import annotations
+
from typing import Any
@@ -7,16 +9,16 @@ class ContainsLoopError(Exception):
class Node:
def __init__(self, data: Any) -> None:
- self.data = data
- self.next_node = None
+ self.data: Any = data
+ self.next_node: Node | None = None
def __iter__(self):
node = self
- visited = []
+ visited = set()
while node:
if node in visited:
raise ContainsLoopError
- visited.append(node)
+ visited.add(node)
yield node.data
node = node.next_node
diff --git a/data_structures/linked_list/is_palindrome.py b/data_structures/linked_list/is_palindrome.py
index acc87c1c272b..da788e3e5045 100644
--- a/data_structures/linked_list/is_palindrome.py
+++ b/data_structures/linked_list/is_palindrome.py
@@ -1,65 +1,169 @@
-def is_palindrome(head):
+from __future__ import annotations
+
+from dataclasses import dataclass
+
+
+@dataclass
+class ListNode:
+ val: int = 0
+ next_node: ListNode | None = None
+
+
+def is_palindrome(head: ListNode | None) -> bool:
+ """
+ Check if a linked list is a palindrome.
+
+ Args:
+ head: The head of the linked list.
+
+ Returns:
+ bool: True if the linked list is a palindrome, False otherwise.
+
+ Examples:
+ >>> is_palindrome(None)
+ True
+
+ >>> is_palindrome(ListNode(1))
+ True
+
+ >>> is_palindrome(ListNode(1, ListNode(2)))
+ False
+
+ >>> is_palindrome(ListNode(1, ListNode(2, ListNode(1))))
+ True
+
+ >>> is_palindrome(ListNode(1, ListNode(2, ListNode(2, ListNode(1)))))
+ True
+ """
if not head:
return True
# split the list to two parts
- fast, slow = head.next, head
- while fast and fast.next:
- fast = fast.next.next
- slow = slow.next
- second = slow.next
- slow.next = None # Don't forget here! But forget still works!
+ fast: ListNode | None = head.next_node
+ slow: ListNode | None = head
+ while fast and fast.next_node:
+ fast = fast.next_node.next_node
+ slow = slow.next_node if slow else None
+ if slow:
+ # slow will always be defined,
+ # adding this check to resolve mypy static check
+ second = slow.next_node
+ slow.next_node = None # Don't forget here! But forget still works!
# reverse the second part
- node = None
+ node: ListNode | None = None
while second:
- nxt = second.next
- second.next = node
+ nxt = second.next_node
+ second.next_node = node
node = second
second = nxt
# compare two parts
# second part has the same or one less node
- while node:
+ while node and head:
if node.val != head.val:
return False
- node = node.next
- head = head.next
+ node = node.next_node
+ head = head.next_node
return True
-def is_palindrome_stack(head):
- if not head or not head.next:
+def is_palindrome_stack(head: ListNode | None) -> bool:
+ """
+ Check if a linked list is a palindrome using a stack.
+
+ Args:
+ head (ListNode): The head of the linked list.
+
+ Returns:
+ bool: True if the linked list is a palindrome, False otherwise.
+
+ Examples:
+ >>> is_palindrome_stack(None)
+ True
+
+ >>> is_palindrome_stack(ListNode(1))
+ True
+
+ >>> is_palindrome_stack(ListNode(1, ListNode(2)))
+ False
+
+ >>> is_palindrome_stack(ListNode(1, ListNode(2, ListNode(1))))
+ True
+
+ >>> is_palindrome_stack(ListNode(1, ListNode(2, ListNode(2, ListNode(1)))))
+ True
+ """
+ if not head or not head.next_node:
return True
# 1. Get the midpoint (slow)
- slow = fast = cur = head
- while fast and fast.next:
- fast, slow = fast.next.next, slow.next
-
- # 2. Push the second half into the stack
- stack = [slow.val]
- while slow.next:
- slow = slow.next
- stack.append(slow.val)
-
- # 3. Comparison
- while stack:
- if stack.pop() != cur.val:
- return False
- cur = cur.next
+ slow: ListNode | None = head
+ fast: ListNode | None = head
+ while fast and fast.next_node:
+ fast = fast.next_node.next_node
+ slow = slow.next_node if slow else None
+
+ # slow will always be defined,
+ # adding this check to resolve mypy static check
+ if slow:
+ stack = [slow.val]
+
+ # 2. Push the second half into the stack
+ while slow.next_node:
+ slow = slow.next_node
+ stack.append(slow.val)
+
+ # 3. Comparison
+ cur: ListNode | None = head
+ while stack and cur:
+ if stack.pop() != cur.val:
+ return False
+ cur = cur.next_node
return True
-def is_palindrome_dict(head):
- if not head or not head.next:
+def is_palindrome_dict(head: ListNode | None) -> bool:
+ """
+ Check if a linked list is a palindrome using a dictionary.
+
+ Args:
+ head (ListNode): The head of the linked list.
+
+ Returns:
+ bool: True if the linked list is a palindrome, False otherwise.
+
+ Examples:
+ >>> is_palindrome_dict(None)
+ True
+
+ >>> is_palindrome_dict(ListNode(1))
+ True
+
+ >>> is_palindrome_dict(ListNode(1, ListNode(2)))
+ False
+
+ >>> is_palindrome_dict(ListNode(1, ListNode(2, ListNode(1))))
+ True
+
+ >>> is_palindrome_dict(ListNode(1, ListNode(2, ListNode(2, ListNode(1)))))
+ True
+
+ >>> is_palindrome_dict(
+ ... ListNode(
+ ... 1, ListNode(2, ListNode(1, ListNode(3, ListNode(2, ListNode(1)))))
+ ... )
+ ... )
+ False
+ """
+ if not head or not head.next_node:
return True
- d = {}
+ d: dict[int, list[int]] = {}
pos = 0
while head:
- if head.val in d.keys():
+ if head.val in d:
d[head.val].append(pos)
else:
d[head.val] = [pos]
- head = head.next
+ head = head.next_node
pos += 1
checksum = pos - 1
middle = 0
@@ -67,11 +171,15 @@ def is_palindrome_dict(head):
if len(v) % 2 != 0:
middle += 1
else:
- step = 0
- for i in range(0, len(v)):
+ for step, i in enumerate(range(len(v))):
if v[i] + v[len(v) - 1 - step] != checksum:
return False
- step += 1
if middle > 1:
return False
return True
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/linked_list/merge_two_lists.py b/data_structures/linked_list/merge_two_lists.py
new file mode 100644
index 000000000000..e47dbdadcf39
--- /dev/null
+++ b/data_structures/linked_list/merge_two_lists.py
@@ -0,0 +1,83 @@
+"""
+Algorithm that merges two sorted linked lists into one sorted linked list.
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterable, Iterator
+from dataclasses import dataclass
+
+test_data_odd = (3, 9, -11, 0, 7, 5, 1, -1)
+test_data_even = (4, 6, 2, 0, 8, 10, 3, -2)
+
+
+@dataclass
+class Node:
+ data: int
+ next_node: Node | None
+
+
+class SortedLinkedList:
+ def __init__(self, ints: Iterable[int]) -> None:
+ self.head: Node | None = None
+ for i in sorted(ints, reverse=True):
+ self.head = Node(i, self.head)
+
+ def __iter__(self) -> Iterator[int]:
+ """
+ >>> tuple(SortedLinkedList(test_data_odd)) == tuple(sorted(test_data_odd))
+ True
+ >>> tuple(SortedLinkedList(test_data_even)) == tuple(sorted(test_data_even))
+ True
+ """
+ node = self.head
+ while node:
+ yield node.data
+ node = node.next_node
+
+ def __len__(self) -> int:
+ """
+ >>> for i in range(3):
+ ... len(SortedLinkedList(range(i))) == i
+ True
+ True
+ True
+ >>> len(SortedLinkedList(test_data_odd))
+ 8
+ """
+ return sum(1 for _ in self)
+
+ def __str__(self) -> str:
+ """
+ >>> str(SortedLinkedList([]))
+ ''
+ >>> str(SortedLinkedList(test_data_odd))
+ '-11 -> -1 -> 0 -> 1 -> 3 -> 5 -> 7 -> 9'
+ >>> str(SortedLinkedList(test_data_even))
+ '-2 -> 0 -> 2 -> 3 -> 4 -> 6 -> 8 -> 10'
+ """
+ return " -> ".join([str(node) for node in self])
+
+
+def merge_lists(
+ sll_one: SortedLinkedList, sll_two: SortedLinkedList
+) -> SortedLinkedList:
+ """
+ >>> SSL = SortedLinkedList
+ >>> merged = merge_lists(SSL(test_data_odd), SSL(test_data_even))
+ >>> len(merged)
+ 16
+ >>> str(merged)
+ '-11 -> -2 -> -1 -> 0 -> 0 -> 1 -> 2 -> 3 -> 3 -> 4 -> 5 -> 6 -> 7 -> 8 -> 9 -> 10'
+ >>> list(merged) == list(sorted(test_data_odd + test_data_even))
+ True
+ """
+ return SortedLinkedList(list(sll_one) + list(sll_two))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ SSL = SortedLinkedList
+ print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
diff --git a/data_structures/linked_list/middle_element_of_linked_list.py b/data_structures/linked_list/middle_element_of_linked_list.py
index 185c4ccbbb0a..86dad6b41d73 100644
--- a/data_structures/linked_list/middle_element_of_linked_list.py
+++ b/data_structures/linked_list/middle_element_of_linked_list.py
@@ -1,5 +1,8 @@
+from __future__ import annotations
+
+
class Node:
- def __init__(self, data: int) -> int:
+ def __init__(self, data: int) -> None:
self.data = data
self.next = None
@@ -14,7 +17,7 @@ def push(self, new_data: int) -> int:
self.head = new_node
return self.head.data
- def middle_element(self) -> int:
+ def middle_element(self) -> int | None:
"""
>>> link = LinkedList()
>>> link.middle_element()
@@ -54,11 +57,12 @@ def middle_element(self) -> int:
return slow_pointer.data
else:
print("No element found.")
+ return None
if __name__ == "__main__":
link = LinkedList()
- for i in range(int(input().strip())):
+ for _ in range(int(input().strip())):
data = int(input().strip())
link.push(data)
print(link.middle_element())
diff --git a/data_structures/linked_list/print_reverse.py b/data_structures/linked_list/print_reverse.py
index c46f228e7260..a023745dee69 100644
--- a/data_structures/linked_list/print_reverse.py
+++ b/data_structures/linked_list/print_reverse.py
@@ -1,22 +1,91 @@
-from typing import List
+from __future__ import annotations
+from collections.abc import Iterable, Iterator
+from dataclasses import dataclass
+
+@dataclass
class Node:
- def __init__(self, data=None):
- self.data = data
- self.next = None
+ data: int
+ next_node: Node | None = None
+
+
+class LinkedList:
+ """A class to represent a Linked List.
+ Use a tail pointer to speed up the append() operation.
+ """
+
+ def __init__(self) -> None:
+ """Initialize a LinkedList with the head node set to None.
+ >>> linked_list = LinkedList()
+ >>> (linked_list.head, linked_list.tail)
+ (None, None)
+ """
+ self.head: Node | None = None
+ self.tail: Node | None = None # Speeds up the append() operation
+
+ def __iter__(self) -> Iterator[int]:
+ """Iterate the LinkedList yielding each Node's data.
+ >>> linked_list = LinkedList()
+ >>> items = (1, 2, 3, 4, 5)
+ >>> linked_list.extend(items)
+ >>> tuple(linked_list) == items
+ True
+ """
+ node = self.head
+ while node:
+ yield node.data
+ node = node.next_node
+
+ def __repr__(self) -> str:
+ """Returns a string representation of the LinkedList.
+ >>> linked_list = LinkedList()
+ >>> str(linked_list)
+ ''
+ >>> linked_list.append(1)
+ >>> str(linked_list)
+ '1'
+ >>> linked_list.extend([2, 3, 4, 5])
+ >>> str(linked_list)
+ '1 -> 2 -> 3 -> 4 -> 5'
+ """
+ return " -> ".join([str(data) for data in self])
- def __repr__(self):
- """Returns a visual representation of the node and all its following nodes."""
- string_rep = []
- temp = self
- while temp:
- string_rep.append(f"{temp.data}")
- temp = temp.next
- return "->".join(string_rep)
+ def append(self, data: int) -> None:
+ """Appends a new node with the given data to the end of the LinkedList.
+ >>> linked_list = LinkedList()
+ >>> str(linked_list)
+ ''
+ >>> linked_list.append(1)
+ >>> str(linked_list)
+ '1'
+ >>> linked_list.append(2)
+ >>> str(linked_list)
+ '1 -> 2'
+ """
+ if self.tail:
+ self.tail.next_node = self.tail = Node(data)
+ else:
+ self.head = self.tail = Node(data)
+ def extend(self, items: Iterable[int]) -> None:
+ """Appends each item to the end of the LinkedList.
+ >>> linked_list = LinkedList()
+ >>> linked_list.extend([])
+ >>> str(linked_list)
+ ''
+ >>> linked_list.extend([1, 2])
+ >>> str(linked_list)
+ '1 -> 2'
+ >>> linked_list.extend([3,4])
+ >>> str(linked_list)
+ '1 -> 2 -> 3 -> 4'
+ """
+ for item in items:
+ self.append(item)
-def make_linked_list(elements_list: List):
+
+def make_linked_list(elements_list: Iterable[int]) -> LinkedList:
"""Creates a Linked List from the elements of the given sequence
(list/tuple) and returns the head of the Linked List.
>>> make_linked_list([])
@@ -28,43 +97,30 @@ def make_linked_list(elements_list: List):
>>> make_linked_list(['abc'])
abc
>>> make_linked_list([7, 25])
- 7->25
+ 7 -> 25
"""
if not elements_list:
raise Exception("The Elements List is empty")
- current = head = Node(elements_list[0])
- for i in range(1, len(elements_list)):
- current.next = Node(elements_list[i])
- current = current.next
- return head
+ linked_list = LinkedList()
+ linked_list.extend(elements_list)
+ return linked_list
-def print_reverse(head_node: Node) -> None:
+def in_reverse(linked_list: LinkedList) -> str:
"""Prints the elements of the given Linked List in reverse order
- >>> print_reverse([])
- >>> linked_list = make_linked_list([69, 88, 73])
- >>> print_reverse(linked_list)
- 73
- 88
- 69
+ >>> in_reverse(LinkedList())
+ ''
+ >>> in_reverse(make_linked_list([69, 88, 73]))
+ '73 <- 88 <- 69'
"""
- if head_node is not None and isinstance(head_node, Node):
- print_reverse(head_node.next)
- print(head_node.data)
+ return " <- ".join(str(line) for line in reversed(tuple(linked_list)))
-def main():
+if __name__ == "__main__":
from doctest import testmod
testmod()
-
- linked_list = make_linked_list([14, 52, 14, 12, 43])
- print("Linked List:")
- print(linked_list)
- print("Elements in Reverse:")
- print_reverse(linked_list)
-
-
-if __name__ == "__main__":
- main()
+ linked_list = make_linked_list((14, 52, 14, 12, 43))
+ print(f"Linked List: {linked_list}")
+ print(f"Reverse List: {in_reverse(linked_list)}")
diff --git a/data_structures/linked_list/reverse_k_group.py b/data_structures/linked_list/reverse_k_group.py
new file mode 100644
index 000000000000..5fc45491a540
--- /dev/null
+++ b/data_structures/linked_list/reverse_k_group.py
@@ -0,0 +1,118 @@
+from __future__ import annotations
+
+from collections.abc import Iterable, Iterator
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ data: int
+ next_node: Node | None = None
+
+
+class LinkedList:
+ def __init__(self, ints: Iterable[int]) -> None:
+ self.head: Node | None = None
+ for i in ints:
+ self.append(i)
+
+ def __iter__(self) -> Iterator[int]:
+ """
+ >>> ints = []
+ >>> list(LinkedList(ints)) == ints
+ True
+ >>> ints = tuple(range(5))
+ >>> tuple(LinkedList(ints)) == ints
+ True
+ """
+ node = self.head
+ while node:
+ yield node.data
+ node = node.next_node
+
+ def __len__(self) -> int:
+ """
+ >>> for i in range(3):
+ ... len(LinkedList(range(i))) == i
+ True
+ True
+ True
+ >>> len(LinkedList("abcdefgh"))
+ 8
+ """
+ return sum(1 for _ in self)
+
+ def __str__(self) -> str:
+ """
+ >>> str(LinkedList([]))
+ ''
+ >>> str(LinkedList(range(5)))
+ '0 -> 1 -> 2 -> 3 -> 4'
+ """
+ return " -> ".join([str(node) for node in self])
+
+ def append(self, data: int) -> None:
+ """
+ >>> ll = LinkedList([1, 2])
+ >>> tuple(ll)
+ (1, 2)
+ >>> ll.append(3)
+ >>> tuple(ll)
+ (1, 2, 3)
+ >>> ll.append(4)
+ >>> tuple(ll)
+ (1, 2, 3, 4)
+ >>> len(ll)
+ 4
+ """
+ if not self.head:
+ self.head = Node(data)
+ return
+ node = self.head
+ while node.next_node:
+ node = node.next_node
+ node.next_node = Node(data)
+
+ def reverse_k_nodes(self, group_size: int) -> None:
+ """
+ reverse nodes within groups of size k
+ >>> ll = LinkedList([1, 2, 3, 4, 5])
+ >>> ll.reverse_k_nodes(2)
+ >>> tuple(ll)
+ (2, 1, 4, 3, 5)
+ >>> str(ll)
+ '2 -> 1 -> 4 -> 3 -> 5'
+ """
+ if self.head is None or self.head.next_node is None:
+ return
+
+ length = len(self)
+ dummy_head = Node(0)
+ dummy_head.next_node = self.head
+ previous_node = dummy_head
+
+ while length >= group_size:
+ current_node = previous_node.next_node
+ assert current_node
+ next_node = current_node.next_node
+ for _ in range(1, group_size):
+ assert next_node, current_node
+ current_node.next_node = next_node.next_node
+ assert previous_node
+ next_node.next_node = previous_node.next_node
+ previous_node.next_node = next_node
+ next_node = current_node.next_node
+ previous_node = current_node
+ length -= group_size
+ self.head = dummy_head.next_node
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ ll = LinkedList([1, 2, 3, 4, 5])
+ print(f"Original Linked List: {ll}")
+ k = 2
+ ll.reverse_k_nodes(k)
+ print(f"After reversing groups of size {k}: {ll}")
diff --git a/data_structures/linked_list/rotate_to_the_right.py b/data_structures/linked_list/rotate_to_the_right.py
new file mode 100644
index 000000000000..6b1c54f4be4d
--- /dev/null
+++ b/data_structures/linked_list/rotate_to_the_right.py
@@ -0,0 +1,156 @@
+from __future__ import annotations
+
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ data: int
+ next_node: Node | None = None
+
+
+def print_linked_list(head: Node | None) -> None:
+ """
+ Print the entire linked list iteratively.
+
+ This function prints the elements of a linked list separated by '->'.
+
+ Parameters:
+ head (Node | None): The head of the linked list to be printed,
+ or None if the linked list is empty.
+
+ >>> head = insert_node(None, 0)
+ >>> head = insert_node(head, 2)
+ >>> head = insert_node(head, 1)
+ >>> print_linked_list(head)
+ 0->2->1
+ >>> head = insert_node(head, 4)
+ >>> head = insert_node(head, 5)
+ >>> print_linked_list(head)
+ 0->2->1->4->5
+ """
+ if head is None:
+ return
+ while head.next_node is not None:
+ print(head.data, end="->")
+ head = head.next_node
+ print(head.data)
+
+
+def insert_node(head: Node | None, data: int) -> Node:
+ """
+ Insert a new node at the end of a linked list and return the new head.
+
+ Parameters:
+ head (Node | None): The head of the linked list.
+ data (int): The data to be inserted into the new node.
+
+ Returns:
+ Node: The new head of the linked list.
+
+ >>> head = insert_node(None, 10)
+ >>> head = insert_node(head, 9)
+ >>> head = insert_node(head, 8)
+ >>> print_linked_list(head)
+ 10->9->8
+ """
+ new_node = Node(data)
+ # If the linked list is empty, the new_node becomes the head
+ if head is None:
+ return new_node
+
+ temp_node = head
+ while temp_node.next_node:
+ temp_node = temp_node.next_node
+
+ temp_node.next_node = new_node
+ return head
+
+
+def rotate_to_the_right(head: Node, places: int) -> Node:
+ """
+ Rotate a linked list to the right by places times.
+
+ Parameters:
+ head: The head of the linked list.
+ places: The number of places to rotate.
+
+ Returns:
+ Node: The head of the rotated linked list.
+
+ >>> rotate_to_the_right(None, places=1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The linked list is empty.
+ >>> head = insert_node(None, 1)
+ >>> rotate_to_the_right(head, places=1) == head
+ True
+ >>> head = insert_node(None, 1)
+ >>> head = insert_node(head, 2)
+ >>> head = insert_node(head, 3)
+ >>> head = insert_node(head, 4)
+ >>> head = insert_node(head, 5)
+ >>> new_head = rotate_to_the_right(head, places=2)
+ >>> print_linked_list(new_head)
+ 4->5->1->2->3
+ """
+ # Check if the list is empty or has only one element
+ if not head:
+ raise ValueError("The linked list is empty.")
+
+ if head.next_node is None:
+ return head
+
+ # Calculate the length of the linked list
+ length = 1
+ temp_node = head
+ while temp_node.next_node is not None:
+ length += 1
+ temp_node = temp_node.next_node
+
+ # Adjust the value of places to avoid places longer than the list.
+ places %= length
+
+ if places == 0:
+ return head # As no rotation is needed.
+
+ # Find the new head position after rotation.
+ new_head_index = length - places
+
+ # Traverse to the new head position
+ temp_node = head
+ for _ in range(new_head_index - 1):
+ assert temp_node.next_node
+ temp_node = temp_node.next_node
+
+ # Update pointers to perform rotation
+ assert temp_node.next_node
+ new_head = temp_node.next_node
+ temp_node.next_node = None
+ temp_node = new_head
+ while temp_node.next_node:
+ temp_node = temp_node.next_node
+ temp_node.next_node = head
+
+ assert new_head
+ return new_head
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ head = insert_node(None, 5)
+ head = insert_node(head, 1)
+ head = insert_node(head, 2)
+ head = insert_node(head, 4)
+ head = insert_node(head, 3)
+
+ print("Original list: ", end="")
+ print_linked_list(head)
+
+ places = 3
+ new_head = rotate_to_the_right(head, places)
+
+ print(f"After {places} iterations: ", end="")
+ print_linked_list(new_head)
diff --git a/data_structures/linked_list/singly_linked_list.py b/data_structures/linked_list/singly_linked_list.py
index e45a210a1785..2c6713a47ad9 100644
--- a/data_structures/linked_list/singly_linked_list.py
+++ b/data_structures/linked_list/singly_linked_list.py
@@ -1,21 +1,70 @@
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
+from typing import Any
+
+
+@dataclass
class Node:
- def __init__(self, data):
- self.data = data
- self.next = None
+ """
+ Create and initialize Node class instance.
+ >>> Node(20)
+ Node(20)
+ >>> Node("Hello, world!")
+ Node(Hello, world!)
+ >>> Node(None)
+ Node(None)
+ >>> Node(True)
+ Node(True)
+ """
+
+ data: Any
+ next_node: Node | None = None
- def __repr__(self):
+ def __repr__(self) -> str:
+ """
+ Get the string representation of this node.
+ >>> Node(10).__repr__()
+ 'Node(10)'
+ >>> repr(Node(10))
+ 'Node(10)'
+ >>> str(Node(10))
+ 'Node(10)'
+ >>> Node(10)
+ Node(10)
+ """
return f"Node({self.data})"
class LinkedList:
def __init__(self):
+ """
+ Create and initialize LinkedList class instance.
+ >>> linked_list = LinkedList()
+ >>> linked_list.head is None
+ True
+ """
self.head = None
- def __iter__(self):
+ def __iter__(self) -> Iterator[Any]:
+ """
+ This function is intended for iterators to access
+ and iterate through data inside linked list.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("tail")
+ >>> linked_list.insert_tail("tail_1")
+ >>> linked_list.insert_tail("tail_2")
+ >>> for node in linked_list: # __iter__ used here.
+ ... node
+ 'tail'
+ 'tail_1'
+ 'tail_2'
+ """
node = self.head
while node:
yield node.data
- node = node.next
+ node = node.next_node
def __len__(self) -> int:
"""
@@ -23,7 +72,7 @@ def __len__(self) -> int:
>>> linked_list = LinkedList()
>>> len(linked_list)
0
- >>> linked_list.insert_tail("head")
+ >>> linked_list.insert_tail("tail")
>>> len(linked_list)
1
>>> linked_list.insert_head("head")
@@ -36,15 +85,27 @@ def __len__(self) -> int:
>>> len(linked_list)
0
"""
- return len(tuple(iter(self)))
+ return sum(1 for _ in self)
- def __repr__(self):
+ def __repr__(self) -> str:
"""
String representation/visualization of a Linked Lists
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail(1)
+ >>> linked_list.insert_tail(3)
+ >>> linked_list.__repr__()
+ '1 -> 3'
+ >>> repr(linked_list)
+ '1 -> 3'
+ >>> str(linked_list)
+ '1 -> 3'
+ >>> linked_list.insert_tail(5)
+ >>> f"{linked_list}"
+ '1 -> 3 -> 5'
"""
- return "->".join([str(item) for item in self])
+ return " -> ".join([str(item) for item in self])
- def __getitem__(self, index):
+ def __getitem__(self, index: int) -> Any:
"""
Indexing Support. Used to get a node at particular position
>>> linked_list = LinkedList()
@@ -54,11 +115,11 @@ def __getitem__(self, index):
True
>>> linked_list[-10]
Traceback (most recent call last):
- ...
+ ...
ValueError: list index out of range.
>>> linked_list[len(linked_list)]
Traceback (most recent call last):
- ...
+ ...
ValueError: list index out of range.
"""
if not 0 <= index < len(self):
@@ -66,9 +127,10 @@ def __getitem__(self, index):
for i, node in enumerate(self):
if i == index:
return node
+ return None
# Used to change the data of a particular node
- def __setitem__(self, index, data):
+ def __setitem__(self, index: int, data: Any) -> None:
"""
>>> linked_list = LinkedList()
>>> for i in range(0, 10):
@@ -81,80 +143,221 @@ def __setitem__(self, index, data):
-666
>>> linked_list[-10] = 666
Traceback (most recent call last):
- ...
+ ...
ValueError: list index out of range.
>>> linked_list[len(linked_list)] = 666
Traceback (most recent call last):
- ...
+ ...
ValueError: list index out of range.
"""
if not 0 <= index < len(self):
raise ValueError("list index out of range.")
current = self.head
- for i in range(index):
- current = current.next
+ for _ in range(index):
+ current = current.next_node
current.data = data
- def insert_tail(self, data) -> None:
+ def insert_tail(self, data: Any) -> None:
+ """
+ Insert data to the end of linked list.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("tail")
+ >>> linked_list
+ tail
+ >>> linked_list.insert_tail("tail_2")
+ >>> linked_list
+ tail -> tail_2
+ >>> linked_list.insert_tail("tail_3")
+ >>> linked_list
+ tail -> tail_2 -> tail_3
+ """
self.insert_nth(len(self), data)
- def insert_head(self, data) -> None:
+ def insert_head(self, data: Any) -> None:
+ """
+ Insert data to the beginning of linked list.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_head("head")
+ >>> linked_list
+ head
+ >>> linked_list.insert_head("head_2")
+ >>> linked_list
+ head_2 -> head
+ >>> linked_list.insert_head("head_3")
+ >>> linked_list
+ head_3 -> head_2 -> head
+ """
self.insert_nth(0, data)
- def insert_nth(self, index: int, data) -> None:
+ def insert_nth(self, index: int, data: Any) -> None:
+ """
+ Insert data at given index.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("first")
+ >>> linked_list.insert_tail("second")
+ >>> linked_list.insert_tail("third")
+ >>> linked_list
+ first -> second -> third
+ >>> linked_list.insert_nth(1, "fourth")
+ >>> linked_list
+ first -> fourth -> second -> third
+ >>> linked_list.insert_nth(3, "fifth")
+ >>> linked_list
+ first -> fourth -> second -> fifth -> third
+ """
if not 0 <= index <= len(self):
raise IndexError("list index out of range")
new_node = Node(data)
if self.head is None:
self.head = new_node
elif index == 0:
- new_node.next = self.head # link new_node to head
+ new_node.next_node = self.head # link new_node to head
self.head = new_node
else:
temp = self.head
for _ in range(index - 1):
- temp = temp.next
- new_node.next = temp.next
- temp.next = new_node
+ temp = temp.next_node
+ new_node.next_node = temp.next_node
+ temp.next_node = new_node
def print_list(self) -> None: # print every node data
+ """
+ This method prints every node data.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("first")
+ >>> linked_list.insert_tail("second")
+ >>> linked_list.insert_tail("third")
+ >>> linked_list
+ first -> second -> third
+ """
print(self)
- def delete_head(self):
+ def delete_head(self) -> Any:
+ """
+ Delete the first node and return the
+ node's data.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("first")
+ >>> linked_list.insert_tail("second")
+ >>> linked_list.insert_tail("third")
+ >>> linked_list
+ first -> second -> third
+ >>> linked_list.delete_head()
+ 'first'
+ >>> linked_list
+ second -> third
+ >>> linked_list.delete_head()
+ 'second'
+ >>> linked_list
+ third
+ >>> linked_list.delete_head()
+ 'third'
+ >>> linked_list.delete_head()
+ Traceback (most recent call last):
+ ...
+ IndexError: List index out of range.
+ """
return self.delete_nth(0)
- def delete_tail(self): # delete from tail
+ def delete_tail(self) -> Any: # delete from tail
+ """
+ Delete the tail end node and return the
+ node's data.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("first")
+ >>> linked_list.insert_tail("second")
+ >>> linked_list.insert_tail("third")
+ >>> linked_list
+ first -> second -> third
+ >>> linked_list.delete_tail()
+ 'third'
+ >>> linked_list
+ first -> second
+ >>> linked_list.delete_tail()
+ 'second'
+ >>> linked_list
+ first
+ >>> linked_list.delete_tail()
+ 'first'
+ >>> linked_list.delete_tail()
+ Traceback (most recent call last):
+ ...
+ IndexError: List index out of range.
+ """
return self.delete_nth(len(self) - 1)
- def delete_nth(self, index: int = 0):
+ def delete_nth(self, index: int = 0) -> Any:
+ """
+ Delete node at given index and return the
+ node's data.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("first")
+ >>> linked_list.insert_tail("second")
+ >>> linked_list.insert_tail("third")
+ >>> linked_list
+ first -> second -> third
+ >>> linked_list.delete_nth(1) # delete middle
+ 'second'
+ >>> linked_list
+ first -> third
+ >>> linked_list.delete_nth(5) # this raises error
+ Traceback (most recent call last):
+ ...
+ IndexError: List index out of range.
+ >>> linked_list.delete_nth(-1) # this also raises error
+ Traceback (most recent call last):
+ ...
+ IndexError: List index out of range.
+ """
if not 0 <= index <= len(self) - 1: # test if index is valid
- raise IndexError("list index out of range")
+ raise IndexError("List index out of range.")
delete_node = self.head # default first node
if index == 0:
- self.head = self.head.next
+ self.head = self.head.next_node
else:
temp = self.head
for _ in range(index - 1):
- temp = temp.next
- delete_node = temp.next
- temp.next = temp.next.next
+ temp = temp.next_node
+ delete_node = temp.next_node
+ temp.next_node = temp.next_node.next_node
return delete_node.data
def is_empty(self) -> bool:
+ """
+ Check if linked list is empty.
+ >>> linked_list = LinkedList()
+ >>> linked_list.is_empty()
+ True
+ >>> linked_list.insert_head("first")
+ >>> linked_list.is_empty()
+ False
+ """
return self.head is None
- def reverse(self):
+ def reverse(self) -> None:
+ """
+ This reverses the linked list order.
+ >>> linked_list = LinkedList()
+ >>> linked_list.insert_tail("first")
+ >>> linked_list.insert_tail("second")
+ >>> linked_list.insert_tail("third")
+ >>> linked_list
+ first -> second -> third
+ >>> linked_list.reverse()
+ >>> linked_list
+ third -> second -> first
+ """
prev = None
current = self.head
while current:
# Store the current node's next node.
- next_node = current.next
- # Make the current node's next point backwards
- current.next = prev
+ next_node = current.next_node
+ # Make the current node's next_node point backwards
+ current.next_node = prev
# Make the previous node be the current node
prev = current
- # Make the current node the next node (to progress iteration)
+ # Make the current node the next_node node (to progress iteration)
current = next_node
# Return prev in order to put the head at the end
self.head = prev
@@ -170,36 +373,122 @@ def test_singly_linked_list() -> None:
try:
linked_list.delete_head()
- assert False # This should not happen.
+ raise AssertionError # This should not happen.
except IndexError:
assert True # This should happen.
try:
linked_list.delete_tail()
- assert False # This should not happen.
+ raise AssertionError # This should not happen.
except IndexError:
assert True # This should happen.
for i in range(10):
assert len(linked_list) == i
linked_list.insert_nth(i, i + 1)
- assert str(linked_list) == "->".join(str(i) for i in range(1, 11))
+ assert str(linked_list) == " -> ".join(str(i) for i in range(1, 11))
linked_list.insert_head(0)
linked_list.insert_tail(11)
- assert str(linked_list) == "->".join(str(i) for i in range(0, 12))
+ assert str(linked_list) == " -> ".join(str(i) for i in range(12))
assert linked_list.delete_head() == 0
assert linked_list.delete_nth(9) == 10
assert linked_list.delete_tail() == 11
assert len(linked_list) == 9
- assert str(linked_list) == "->".join(str(i) for i in range(1, 10))
+ assert str(linked_list) == " -> ".join(str(i) for i in range(1, 10))
- assert all(linked_list[i] == i + 1 for i in range(0, 9)) is True
+ assert all(linked_list[i] == i + 1 for i in range(9)) is True
- for i in range(0, 9):
+ for i in range(9):
linked_list[i] = -i
- assert all(linked_list[i] == -i for i in range(0, 9)) is True
+ assert all(linked_list[i] == -i for i in range(9)) is True
+
+ linked_list.reverse()
+ assert str(linked_list) == " -> ".join(str(i) for i in range(-8, 1))
+
+
+def test_singly_linked_list_2() -> None:
+ """
+ This section of the test used varying data types for input.
+ >>> test_singly_linked_list_2()
+ """
+ test_input = [
+ -9,
+ 100,
+ Node(77345112),
+ "dlrow olleH",
+ 7,
+ 5555,
+ 0,
+ -192.55555,
+ "Hello, world!",
+ 77.9,
+ Node(10),
+ None,
+ None,
+ 12.20,
+ ]
+ linked_list = LinkedList()
+
+ for i in test_input:
+ linked_list.insert_tail(i)
+
+ # Check if it's empty or not
+ assert linked_list.is_empty() is False
+ assert (
+ str(linked_list)
+ == "-9 -> 100 -> Node(77345112) -> dlrow olleH -> 7 -> 5555 -> "
+ "0 -> -192.55555 -> Hello, world! -> 77.9 -> Node(10) -> None -> None -> 12.2"
+ )
+
+ # Delete the head
+ result = linked_list.delete_head()
+ assert result == -9
+ assert (
+ str(linked_list) == "100 -> Node(77345112) -> dlrow olleH -> 7 -> 5555 -> 0 -> "
+ "-192.55555 -> Hello, world! -> 77.9 -> Node(10) -> None -> None -> 12.2"
+ )
+
+ # Delete the tail
+ result = linked_list.delete_tail()
+ assert result == 12.2
+ assert (
+ str(linked_list) == "100 -> Node(77345112) -> dlrow olleH -> 7 -> 5555 -> 0 -> "
+ "-192.55555 -> Hello, world! -> 77.9 -> Node(10) -> None -> None"
+ )
+
+ # Delete a node in specific location in linked list
+ result = linked_list.delete_nth(10)
+ assert result is None
+ assert (
+ str(linked_list) == "100 -> Node(77345112) -> dlrow olleH -> 7 -> 5555 -> 0 -> "
+ "-192.55555 -> Hello, world! -> 77.9 -> Node(10) -> None"
+ )
+
+ # Add a Node instance to its head
+ linked_list.insert_head(Node("Hello again, world!"))
+ assert (
+ str(linked_list)
+ == "Node(Hello again, world!) -> 100 -> Node(77345112) -> dlrow olleH -> "
+ "7 -> 5555 -> 0 -> -192.55555 -> Hello, world! -> 77.9 -> Node(10) -> None"
+ )
+
+ # Add None to its tail
+ linked_list.insert_tail(None)
+ assert (
+ str(linked_list)
+ == "Node(Hello again, world!) -> 100 -> Node(77345112) -> dlrow olleH -> 7 -> "
+ "5555 -> 0 -> -192.55555 -> Hello, world! -> 77.9 -> Node(10) -> None -> None"
+ )
+
+ # Reverse the linked list
+ linked_list.reverse()
+ assert (
+ str(linked_list)
+ == "None -> None -> Node(10) -> 77.9 -> Hello, world! -> -192.55555 -> 0 -> "
+ "5555 -> 7 -> dlrow olleH -> Node(77345112) -> 100 -> Node(Hello again, world!)"
+ )
def main():
diff --git a/data_structures/linked_list/skip_list.py b/data_structures/linked_list/skip_list.py
index 8f06e6193d52..13e9a94a8698 100644
--- a/data_structures/linked_list/skip_list.py
+++ b/data_structures/linked_list/skip_list.py
@@ -5,15 +5,16 @@
from __future__ import annotations
+from itertools import pairwise
from random import random
-from typing import Generic, Optional, TypeVar
+from typing import Generic, TypeVar
KT = TypeVar("KT")
VT = TypeVar("VT")
class Node(Generic[KT, VT]):
- def __init__(self, key: KT, value: VT):
+ def __init__(self, key: KT | str = "root", value: VT | None = None):
self.key = key
self.value = value
self.forward: list[Node[KT, VT]] = []
@@ -50,7 +51,7 @@ def level(self) -> int:
class SkipList(Generic[KT, VT]):
def __init__(self, p: float = 0.5, max_level: int = 16):
- self.head = Node("root", None)
+ self.head: Node[KT, VT] = Node[KT, VT]()
self.level = 0
self.p = p
self.max_level = max_level
@@ -124,7 +125,7 @@ def random_level(self) -> int:
return level
- def _locate_node(self, key) -> tuple[Optional[Node[KT, VT]], list[Node[KT, VT]]]:
+ def _locate_node(self, key) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]:
"""
:param key: Searched key,
:return: Tuple with searched node (or None if given key is not present)
@@ -206,7 +207,7 @@ def insert(self, key: KT, value: VT):
if level > self.level:
# After level increase we have to add additional nodes to head.
- for i in range(self.level - 1, level):
+ for _ in range(self.level - 1, level):
update_vector.append(self.head)
self.level = level
@@ -222,7 +223,7 @@ def insert(self, key: KT, value: VT):
else:
update_node.forward[i] = new_node
- def find(self, key: VT) -> Optional[VT]:
+ def find(self, key: VT) -> VT | None:
"""
:param key: Search key.
:return: Value associated with given key or None if given key is not present.
@@ -389,10 +390,7 @@ def traverse_keys(node):
def test_iter_always_yields_sorted_values():
def is_sorted(lst):
- for item, next_item in zip(lst, lst[1:]):
- if next_item < item:
- return False
- return True
+ return all(next_item >= item for item, next_item in pairwise(lst))
skip_list = SkipList()
for i in range(10):
@@ -408,7 +406,7 @@ def is_sorted(lst):
def pytests():
- for i in range(100):
+ for _ in range(100):
# Repeat test 100 times due to the probabilistic nature of skip list
# random values == random bugs
test_insert()
@@ -444,4 +442,7 @@ def main():
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
main()
diff --git a/data_structures/linked_list/swap_nodes.py b/data_structures/linked_list/swap_nodes.py
index 3f825756b3d2..d66512087d2d 100644
--- a/data_structures/linked_list/swap_nodes.py
+++ b/data_structures/linked_list/swap_nodes.py
@@ -1,55 +1,148 @@
+from __future__ import annotations
+
+from collections.abc import Iterator
+from dataclasses import dataclass
from typing import Any
+@dataclass
class Node:
- def __init__(self, data: Any):
- self.data = data
- self.next = None
+ data: Any
+ next_node: Node | None = None
+@dataclass
class LinkedList:
- def __init__(self):
- self.head = None
-
- def print_list(self):
- temp = self.head
- while temp is not None:
- print(temp.data, end=" ")
- temp = temp.next
- print()
-
- # adding nodes
- def push(self, new_data: Any):
+ head: Node | None = None
+
+ def __iter__(self) -> Iterator:
+ """
+ >>> linked_list = LinkedList()
+ >>> list(linked_list)
+ []
+ >>> linked_list.push(0)
+ >>> tuple(linked_list)
+ (0,)
+ """
+ node = self.head
+ while node:
+ yield node.data
+ node = node.next_node
+
+ def __len__(self) -> int:
+ """
+ >>> linked_list = LinkedList()
+ >>> len(linked_list)
+ 0
+ >>> linked_list.push(0)
+ >>> len(linked_list)
+ 1
+ """
+ return sum(1 for _ in self)
+
+ def push(self, new_data: Any) -> None:
+ """
+ Add a new node with the given data to the beginning of the Linked List.
+
+ Args:
+ new_data (Any): The data to be added to the new node.
+
+ Returns:
+ None
+
+ Examples:
+ >>> linked_list = LinkedList()
+ >>> linked_list.push(5)
+ >>> linked_list.push(4)
+ >>> linked_list.push(3)
+ >>> linked_list.push(2)
+ >>> linked_list.push(1)
+ >>> list(linked_list)
+ [1, 2, 3, 4, 5]
+ """
new_node = Node(new_data)
- new_node.next = self.head
+ new_node.next_node = self.head
self.head = new_node
- # swapping nodes
- def swap_nodes(self, node_data_1, node_data_2):
- if node_data_1 == node_data_2:
- return
- else:
- node_1 = self.head
- while node_1 is not None and node_1.data != node_data_1:
- node_1 = node_1.next
+ def swap_nodes(self, node_data_1: Any, node_data_2: Any) -> None:
+ """
+ Swap the positions of two nodes in the Linked List based on their data values.
+
+ Args:
+ node_data_1: Data value of the first node to be swapped.
+ node_data_2: Data value of the second node to be swapped.
- node_2 = self.head
- while node_2 is not None and node_2.data != node_data_2:
- node_2 = node_2.next
- if node_1 is None or node_2 is None:
- return
+ Note:
+ If either of the specified data values isn't found then, no swapping occurs.
- node_1.data, node_2.data = node_2.data, node_1.data
+ Examples:
+ When both values are present in a linked list.
+ >>> linked_list = LinkedList()
+ >>> linked_list.push(5)
+ >>> linked_list.push(4)
+ >>> linked_list.push(3)
+ >>> linked_list.push(2)
+ >>> linked_list.push(1)
+ >>> list(linked_list)
+ [1, 2, 3, 4, 5]
+ >>> linked_list.swap_nodes(1, 5)
+ >>> tuple(linked_list)
+ (5, 2, 3, 4, 1)
+
+ When one value is present and the other isn't in the linked list.
+ >>> second_list = LinkedList()
+ >>> second_list.push(6)
+ >>> second_list.push(7)
+ >>> second_list.push(8)
+ >>> second_list.push(9)
+ >>> second_list.swap_nodes(1, 6) is None
+ True
+
+ When both values are absent in the linked list.
+ >>> second_list = LinkedList()
+ >>> second_list.push(10)
+ >>> second_list.push(9)
+ >>> second_list.push(8)
+ >>> second_list.push(7)
+ >>> second_list.swap_nodes(1, 3) is None
+ True
+
+ When linkedlist is empty.
+ >>> second_list = LinkedList()
+ >>> second_list.swap_nodes(1, 3) is None
+ True
+
+ Returns:
+ None
+ """
+ if node_data_1 == node_data_2:
+ return
+
+ node_1 = self.head
+ while node_1 and node_1.data != node_data_1:
+ node_1 = node_1.next_node
+ node_2 = self.head
+ while node_2 and node_2.data != node_data_2:
+ node_2 = node_2.next_node
+ if node_1 is None or node_2 is None:
+ return
+ # Swap the data values of the two nodes
+ node_1.data, node_2.data = node_2.data, node_1.data
if __name__ == "__main__":
- ll = LinkedList()
- for i in range(5, 0, -1):
- ll.push(i)
+ """
+ Python script that outputs the swap of nodes in a linked list.
+ """
+ from doctest import testmod
- ll.print_list()
+ testmod()
+ linked_list = LinkedList()
+ for i in range(5, 0, -1):
+ linked_list.push(i)
- ll.swap_nodes(1, 4)
- print("After swapping")
- ll.print_list()
+ print(f"Original Linked List: {list(linked_list)}")
+ linked_list.swap_nodes(1, 4)
+ print(f"Modified Linked List: {list(linked_list)}")
+ print("After swapping the nodes whose data is 1 and 4.")
diff --git a/data_structures/queue/double_ended_queue.py b/data_structures/queue/double_ended_queue.py
deleted file mode 100644
index dd003b7c98ac..000000000000
--- a/data_structures/queue/double_ended_queue.py
+++ /dev/null
@@ -1,57 +0,0 @@
-# Python code to demonstrate working of
-# extend(), extendleft(), rotate(), reverse()
-
-# importing "collections" for deque operations
-import collections
-
-# initializing deque
-de = collections.deque([1, 2, 3])
-
-# using extend() to add numbers to right end
-# adds 4,5,6 to right end
-de.extend([4, 5, 6])
-
-# printing modified deque
-print("The deque after extending deque at end is : ")
-print(de)
-
-# using extendleft() to add numbers to left end
-# adds 7,8,9 to right end
-de.extendleft([7, 8, 9])
-
-# printing modified deque
-print("The deque after extending deque at beginning is : ")
-print(de)
-
-# using rotate() to rotate the deque
-# rotates by 3 to left
-de.rotate(-3)
-
-# printing modified deque
-print("The deque after rotating deque is : ")
-print(de)
-
-# using reverse() to reverse the deque
-de.reverse()
-
-# printing modified deque
-print("The deque after reversing deque is : ")
-print(de)
-
-# get right-end value and eliminate
-startValue = de.pop()
-
-print("The deque after popping value at end is : ")
-print(de)
-
-# get left-end value and eliminate
-endValue = de.popleft()
-
-print("The deque after popping value at start is : ")
-print(de)
-
-# eliminate element searched by value
-de.remove(5)
-
-print("The deque after eliminating element searched by value : ")
-print(de)
diff --git a/data_structures/queue/queue_on_list.py b/data_structures/queue/queue_on_list.py
deleted file mode 100644
index 485cf0b6f7a3..000000000000
--- a/data_structures/queue/queue_on_list.py
+++ /dev/null
@@ -1,52 +0,0 @@
-"""Queue represented by a Python list"""
-
-
-class Queue:
- def __init__(self):
- self.entries = []
- self.length = 0
- self.front = 0
-
- def __str__(self):
- printed = "<" + str(self.entries)[1:-1] + ">"
- return printed
-
- """Enqueues {@code item}
- @param item
- item to enqueue"""
-
- def put(self, item):
- self.entries.append(item)
- self.length = self.length + 1
-
- """Dequeues {@code item}
- @requirement: |self.length| > 0
- @return dequeued
- item that was dequeued"""
-
- def get(self):
- self.length = self.length - 1
- dequeued = self.entries[self.front]
- # self.front-=1
- # self.entries = self.entries[self.front:]
- self.entries = self.entries[1:]
- return dequeued
-
- """Rotates the queue {@code rotation} times
- @param rotation
- number of times to rotate queue"""
-
- def rotate(self, rotation):
- for i in range(rotation):
- self.put(self.get())
-
- """Enqueues {@code item}
- @return item at front of self.entries"""
-
- def get_front(self):
- return self.entries[0]
-
- """Returns the length of this.entries"""
-
- def size(self):
- return self.length
diff --git a/data_structures/queues/__init__.py b/data_structures/queues/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/data_structures/queue/circular_queue.py b/data_structures/queues/circular_queue.py
similarity index 79%
rename from data_structures/queue/circular_queue.py
rename to data_structures/queues/circular_queue.py
index 93a6ef805c7c..efbf1efdc42d 100644
--- a/data_structures/queue/circular_queue.py
+++ b/data_structures/queues/circular_queue.py
@@ -17,7 +17,9 @@ def __len__(self) -> int:
>>> len(cq)
0
>>> cq.enqueue("A") # doctest: +ELLIPSIS
- >> cq.array
+ ['A', None, None, None, None]
>>> len(cq)
1
"""
@@ -25,6 +27,7 @@ def __len__(self) -> int:
def is_empty(self) -> bool:
"""
+ Checks whether the queue is empty or not
>>> cq = CircularQueue(5)
>>> cq.is_empty()
True
@@ -35,6 +38,7 @@ def is_empty(self) -> bool:
def first(self):
"""
+ Returns the first element of the queue
>>> cq = CircularQueue(5)
>>> cq.first()
False
@@ -45,14 +49,17 @@ def first(self):
def enqueue(self, data):
"""
- This function insert an element in the queue using self.rear value as an index
+ This function inserts an element at the end of the queue using self.rear value
+ as an index.
>>> cq = CircularQueue(5)
>>> cq.enqueue("A") # doctest: +ELLIPSIS
- >> (cq.size, cq.first())
(1, 'A')
>>> cq.enqueue("B") # doctest: +ELLIPSIS
- >> cq.array
+ ['A', 'B', None, None, None]
>>> (cq.size, cq.first())
(2, 'A')
"""
@@ -67,7 +74,7 @@ def enqueue(self, data):
def dequeue(self):
"""
This function removes an element from the queue using on self.front value as an
- index
+ index and returns it
>>> cq = CircularQueue(5)
>>> cq.dequeue()
Traceback (most recent call last):
diff --git a/data_structures/queues/circular_queue_linked_list.py b/data_structures/queues/circular_queue_linked_list.py
new file mode 100644
index 000000000000..da8629678e52
--- /dev/null
+++ b/data_structures/queues/circular_queue_linked_list.py
@@ -0,0 +1,161 @@
+# Implementation of Circular Queue using linked lists
+# https://en.wikipedia.org/wiki/Circular_buffer
+
+from __future__ import annotations
+
+from typing import Any
+
+
+class CircularQueueLinkedList:
+ """
+ Circular FIFO list with the given capacity (default queue length : 6)
+
+ >>> cq = CircularQueueLinkedList(2)
+ >>> cq.enqueue('a')
+ >>> cq.enqueue('b')
+ >>> cq.enqueue('c')
+ Traceback (most recent call last):
+ ...
+ Exception: Full Queue
+ """
+
+ def __init__(self, initial_capacity: int = 6) -> None:
+ self.front: Node | None = None
+ self.rear: Node | None = None
+ self.create_linked_list(initial_capacity)
+
+ def create_linked_list(self, initial_capacity: int) -> None:
+ current_node = Node()
+ self.front = current_node
+ self.rear = current_node
+ previous_node = current_node
+ for _ in range(1, initial_capacity):
+ current_node = Node()
+ previous_node.next = current_node
+ current_node.prev = previous_node
+ previous_node = current_node
+ previous_node.next = self.front
+ self.front.prev = previous_node
+
+ def is_empty(self) -> bool:
+ """
+ Checks whether the queue is empty or not
+ >>> cq = CircularQueueLinkedList()
+ >>> cq.is_empty()
+ True
+ >>> cq.enqueue('a')
+ >>> cq.is_empty()
+ False
+ >>> cq.dequeue()
+ 'a'
+ >>> cq.is_empty()
+ True
+ """
+
+ return (
+ self.front == self.rear
+ and self.front is not None
+ and self.front.data is None
+ )
+
+ def first(self) -> Any | None:
+ """
+ Returns the first element of the queue
+ >>> cq = CircularQueueLinkedList()
+ >>> cq.first()
+ Traceback (most recent call last):
+ ...
+ Exception: Empty Queue
+ >>> cq.enqueue('a')
+ >>> cq.first()
+ 'a'
+ >>> cq.dequeue()
+ 'a'
+ >>> cq.first()
+ Traceback (most recent call last):
+ ...
+ Exception: Empty Queue
+ >>> cq.enqueue('b')
+ >>> cq.enqueue('c')
+ >>> cq.first()
+ 'b'
+ """
+ self.check_can_perform_operation()
+ return self.front.data if self.front else None
+
+ def enqueue(self, data: Any) -> None:
+ """
+ Saves data at the end of the queue
+
+ >>> cq = CircularQueueLinkedList()
+ >>> cq.enqueue('a')
+ >>> cq.enqueue('b')
+ >>> cq.dequeue()
+ 'a'
+ >>> cq.dequeue()
+ 'b'
+ >>> cq.dequeue()
+ Traceback (most recent call last):
+ ...
+ Exception: Empty Queue
+ """
+ if self.rear is None:
+ return
+
+ self.check_is_full()
+ if not self.is_empty():
+ self.rear = self.rear.next
+ if self.rear:
+ self.rear.data = data
+
+ def dequeue(self) -> Any:
+ """
+ Removes and retrieves the first element of the queue
+
+ >>> cq = CircularQueueLinkedList()
+ >>> cq.dequeue()
+ Traceback (most recent call last):
+ ...
+ Exception: Empty Queue
+ >>> cq.enqueue('a')
+ >>> cq.dequeue()
+ 'a'
+ >>> cq.dequeue()
+ Traceback (most recent call last):
+ ...
+ Exception: Empty Queue
+ """
+ self.check_can_perform_operation()
+ if self.rear is None or self.front is None:
+ return None
+ if self.front == self.rear:
+ data = self.front.data
+ self.front.data = None
+ return data
+
+ old_front = self.front
+ self.front = old_front.next
+ data = old_front.data
+ old_front.data = None
+ return data
+
+ def check_can_perform_operation(self) -> None:
+ if self.is_empty():
+ raise Exception("Empty Queue")
+
+ def check_is_full(self) -> None:
+ if self.rear and self.rear.next == self.front:
+ raise Exception("Full Queue")
+
+
+class Node:
+ def __init__(self) -> None:
+ self.data: Any | None = None
+ self.next: Node | None = None
+ self.prev: Node | None = None
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/queues/double_ended_queue.py b/data_structures/queues/double_ended_queue.py
new file mode 100644
index 000000000000..c28d46c65168
--- /dev/null
+++ b/data_structures/queues/double_ended_queue.py
@@ -0,0 +1,463 @@
+"""
+Implementation of double ended queue.
+"""
+
+from __future__ import annotations
+
+from collections.abc import Iterable
+from dataclasses import dataclass
+from typing import Any
+
+
+class Deque:
+ """
+ Deque data structure.
+ Operations
+ ----------
+ append(val: Any) -> None
+ appendleft(val: Any) -> None
+ extend(iterable: Iterable) -> None
+ extendleft(iterable: Iterable) -> None
+ pop() -> Any
+ popleft() -> Any
+ Observers
+ ---------
+ is_empty() -> bool
+ Attributes
+ ----------
+ _front: _Node
+ front of the deque a.k.a. the first element
+ _back: _Node
+ back of the element a.k.a. the last element
+ _len: int
+ the number of nodes
+ """
+
+ __slots__ = ("_back", "_front", "_len")
+
+ @dataclass
+ class _Node:
+ """
+ Representation of a node.
+ Contains a value and a pointer to the next node as well as to the previous one.
+ """
+
+ val: Any = None
+ next_node: Deque._Node | None = None
+ prev_node: Deque._Node | None = None
+
+ class _Iterator:
+ """
+ Helper class for iteration. Will be used to implement iteration.
+ Attributes
+ ----------
+ _cur: _Node
+ the current node of the iteration.
+ """
+
+ __slots__ = ("_cur",)
+
+ def __init__(self, cur: Deque._Node | None) -> None:
+ self._cur = cur
+
+ def __iter__(self) -> Deque._Iterator:
+ """
+ >>> our_deque = Deque([1, 2, 3])
+ >>> iterator = iter(our_deque)
+ """
+ return self
+
+ def __next__(self) -> Any:
+ """
+ >>> our_deque = Deque([1, 2, 3])
+ >>> iterator = iter(our_deque)
+ >>> next(iterator)
+ 1
+ >>> next(iterator)
+ 2
+ >>> next(iterator)
+ 3
+ """
+ if self._cur is None:
+ # finished iterating
+ raise StopIteration
+ val = self._cur.val
+ self._cur = self._cur.next_node
+
+ return val
+
+ def __init__(self, iterable: Iterable[Any] | None = None) -> None:
+ self._front: Any = None
+ self._back: Any = None
+ self._len: int = 0
+
+ if iterable is not None:
+ # append every value to the deque
+ for val in iterable:
+ self.append(val)
+
+ def append(self, val: Any) -> None:
+ """
+ Adds val to the end of the deque.
+ Time complexity: O(1)
+ >>> our_deque_1 = Deque([1, 2, 3])
+ >>> our_deque_1.append(4)
+ >>> our_deque_1
+ [1, 2, 3, 4]
+ >>> our_deque_2 = Deque('ab')
+ >>> our_deque_2.append('c')
+ >>> our_deque_2
+ ['a', 'b', 'c']
+ >>> from collections import deque
+ >>> deque_collections_1 = deque([1, 2, 3])
+ >>> deque_collections_1.append(4)
+ >>> deque_collections_1
+ deque([1, 2, 3, 4])
+ >>> deque_collections_2 = deque('ab')
+ >>> deque_collections_2.append('c')
+ >>> deque_collections_2
+ deque(['a', 'b', 'c'])
+ >>> list(our_deque_1) == list(deque_collections_1)
+ True
+ >>> list(our_deque_2) == list(deque_collections_2)
+ True
+ """
+ node = self._Node(val, None, None)
+ if self.is_empty():
+ # front = back
+ self._front = self._back = node
+ self._len = 1
+ else:
+ # connect nodes
+ self._back.next_node = node
+ node.prev_node = self._back
+ self._back = node # assign new back to the new node
+
+ self._len += 1
+
+ # make sure there were no errors
+ assert not self.is_empty(), "Error on appending value."
+
+ def appendleft(self, val: Any) -> None:
+ """
+ Adds val to the beginning of the deque.
+ Time complexity: O(1)
+ >>> our_deque_1 = Deque([2, 3])
+ >>> our_deque_1.appendleft(1)
+ >>> our_deque_1
+ [1, 2, 3]
+ >>> our_deque_2 = Deque('bc')
+ >>> our_deque_2.appendleft('a')
+ >>> our_deque_2
+ ['a', 'b', 'c']
+ >>> from collections import deque
+ >>> deque_collections_1 = deque([2, 3])
+ >>> deque_collections_1.appendleft(1)
+ >>> deque_collections_1
+ deque([1, 2, 3])
+ >>> deque_collections_2 = deque('bc')
+ >>> deque_collections_2.appendleft('a')
+ >>> deque_collections_2
+ deque(['a', 'b', 'c'])
+ >>> list(our_deque_1) == list(deque_collections_1)
+ True
+ >>> list(our_deque_2) == list(deque_collections_2)
+ True
+ """
+ node = self._Node(val, None, None)
+ if self.is_empty():
+ # front = back
+ self._front = self._back = node
+ self._len = 1
+ else:
+ # connect nodes
+ node.next_node = self._front
+ self._front.prev_node = node
+ self._front = node # assign new front to the new node
+
+ self._len += 1
+
+ # make sure there were no errors
+ assert not self.is_empty(), "Error on appending value."
+
+ def extend(self, iterable: Iterable[Any]) -> None:
+ """
+ Appends every value of iterable to the end of the deque.
+ Time complexity: O(n)
+ >>> our_deque_1 = Deque([1, 2, 3])
+ >>> our_deque_1.extend([4, 5])
+ >>> our_deque_1
+ [1, 2, 3, 4, 5]
+ >>> our_deque_2 = Deque('ab')
+ >>> our_deque_2.extend('cd')
+ >>> our_deque_2
+ ['a', 'b', 'c', 'd']
+ >>> from collections import deque
+ >>> deque_collections_1 = deque([1, 2, 3])
+ >>> deque_collections_1.extend([4, 5])
+ >>> deque_collections_1
+ deque([1, 2, 3, 4, 5])
+ >>> deque_collections_2 = deque('ab')
+ >>> deque_collections_2.extend('cd')
+ >>> deque_collections_2
+ deque(['a', 'b', 'c', 'd'])
+ >>> list(our_deque_1) == list(deque_collections_1)
+ True
+ >>> list(our_deque_2) == list(deque_collections_2)
+ True
+ """
+ for val in iterable:
+ self.append(val)
+
+ def extendleft(self, iterable: Iterable[Any]) -> None:
+ """
+ Appends every value of iterable to the beginning of the deque.
+ Time complexity: O(n)
+ >>> our_deque_1 = Deque([1, 2, 3])
+ >>> our_deque_1.extendleft([0, -1])
+ >>> our_deque_1
+ [-1, 0, 1, 2, 3]
+ >>> our_deque_2 = Deque('cd')
+ >>> our_deque_2.extendleft('ba')
+ >>> our_deque_2
+ ['a', 'b', 'c', 'd']
+ >>> from collections import deque
+ >>> deque_collections_1 = deque([1, 2, 3])
+ >>> deque_collections_1.extendleft([0, -1])
+ >>> deque_collections_1
+ deque([-1, 0, 1, 2, 3])
+ >>> deque_collections_2 = deque('cd')
+ >>> deque_collections_2.extendleft('ba')
+ >>> deque_collections_2
+ deque(['a', 'b', 'c', 'd'])
+ >>> list(our_deque_1) == list(deque_collections_1)
+ True
+ >>> list(our_deque_2) == list(deque_collections_2)
+ True
+ """
+ for val in iterable:
+ self.appendleft(val)
+
+ def pop(self) -> Any:
+ """
+ Removes the last element of the deque and returns it.
+ Time complexity: O(1)
+ @returns topop.val: the value of the node to pop.
+ >>> our_deque1 = Deque([1])
+ >>> our_popped1 = our_deque1.pop()
+ >>> our_popped1
+ 1
+ >>> our_deque1
+ []
+
+ >>> our_deque2 = Deque([1, 2, 3, 15182])
+ >>> our_popped2 = our_deque2.pop()
+ >>> our_popped2
+ 15182
+ >>> our_deque2
+ [1, 2, 3]
+
+ >>> from collections import deque
+ >>> deque_collections = deque([1, 2, 3, 15182])
+ >>> collections_popped = deque_collections.pop()
+ >>> collections_popped
+ 15182
+ >>> deque_collections
+ deque([1, 2, 3])
+ >>> list(our_deque2) == list(deque_collections)
+ True
+ >>> our_popped2 == collections_popped
+ True
+ """
+ # make sure the deque has elements to pop
+ assert not self.is_empty(), "Deque is empty."
+
+ topop = self._back
+ # if only one element in the queue: point the front and back to None
+ # else remove one element from back
+ if self._front == self._back:
+ self._front = None
+ self._back = None
+ else:
+ self._back = self._back.prev_node # set new back
+ # drop the last node, python will deallocate memory automatically
+ self._back.next_node = None
+
+ self._len -= 1
+
+ return topop.val
+
+ def popleft(self) -> Any:
+ """
+ Removes the first element of the deque and returns it.
+ Time complexity: O(1)
+ @returns topop.val: the value of the node to pop.
+ >>> our_deque1 = Deque([1])
+ >>> our_popped1 = our_deque1.pop()
+ >>> our_popped1
+ 1
+ >>> our_deque1
+ []
+ >>> our_deque2 = Deque([15182, 1, 2, 3])
+ >>> our_popped2 = our_deque2.popleft()
+ >>> our_popped2
+ 15182
+ >>> our_deque2
+ [1, 2, 3]
+ >>> from collections import deque
+ >>> deque_collections = deque([15182, 1, 2, 3])
+ >>> collections_popped = deque_collections.popleft()
+ >>> collections_popped
+ 15182
+ >>> deque_collections
+ deque([1, 2, 3])
+ >>> list(our_deque2) == list(deque_collections)
+ True
+ >>> our_popped2 == collections_popped
+ True
+ """
+ # make sure the deque has elements to pop
+ assert not self.is_empty(), "Deque is empty."
+
+ topop = self._front
+ # if only one element in the queue: point the front and back to None
+ # else remove one element from front
+ if self._front == self._back:
+ self._front = None
+ self._back = None
+ else:
+ self._front = self._front.next_node # set new front and drop the first node
+ self._front.prev_node = None
+
+ self._len -= 1
+
+ return topop.val
+
+ def is_empty(self) -> bool:
+ """
+ Checks if the deque is empty.
+ Time complexity: O(1)
+ >>> our_deque = Deque([1, 2, 3])
+ >>> our_deque.is_empty()
+ False
+ >>> our_empty_deque = Deque()
+ >>> our_empty_deque.is_empty()
+ True
+ >>> from collections import deque
+ >>> empty_deque_collections = deque()
+ >>> list(our_empty_deque) == list(empty_deque_collections)
+ True
+ """
+ return self._front is None
+
+ def __len__(self) -> int:
+ """
+ Implements len() function. Returns the length of the deque.
+ Time complexity: O(1)
+ >>> our_deque = Deque([1, 2, 3])
+ >>> len(our_deque)
+ 3
+ >>> our_empty_deque = Deque()
+ >>> len(our_empty_deque)
+ 0
+ >>> from collections import deque
+ >>> deque_collections = deque([1, 2, 3])
+ >>> len(deque_collections)
+ 3
+ >>> empty_deque_collections = deque()
+ >>> len(empty_deque_collections)
+ 0
+ >>> len(our_empty_deque) == len(empty_deque_collections)
+ True
+ """
+ return self._len
+
+ def __eq__(self, other: object) -> bool:
+ """
+ Implements "==" operator. Returns if *self* is equal to *other*.
+ Time complexity: O(n)
+ >>> our_deque_1 = Deque([1, 2, 3])
+ >>> our_deque_2 = Deque([1, 2, 3])
+ >>> our_deque_1 == our_deque_2
+ True
+ >>> our_deque_3 = Deque([1, 2])
+ >>> our_deque_1 == our_deque_3
+ False
+ >>> from collections import deque
+ >>> deque_collections_1 = deque([1, 2, 3])
+ >>> deque_collections_2 = deque([1, 2, 3])
+ >>> deque_collections_1 == deque_collections_2
+ True
+ >>> deque_collections_3 = deque([1, 2])
+ >>> deque_collections_1 == deque_collections_3
+ False
+ >>> (our_deque_1 == our_deque_2) == (deque_collections_1 == deque_collections_2)
+ True
+ >>> (our_deque_1 == our_deque_3) == (deque_collections_1 == deque_collections_3)
+ True
+ """
+
+ if not isinstance(other, Deque):
+ return NotImplemented
+
+ me = self._front
+ oth = other._front
+
+ # if the length of the dequeues are not the same, they are not equal
+ if len(self) != len(other):
+ return False
+
+ while me is not None and oth is not None:
+ # compare every value
+ if me.val != oth.val:
+ return False
+ me = me.next_node
+ oth = oth.next_node
+
+ return True
+
+ def __iter__(self) -> Deque._Iterator:
+ """
+ Implements iteration.
+ Time complexity: O(1)
+ >>> our_deque = Deque([1, 2, 3])
+ >>> for v in our_deque:
+ ... print(v)
+ 1
+ 2
+ 3
+ >>> from collections import deque
+ >>> deque_collections = deque([1, 2, 3])
+ >>> for v in deque_collections:
+ ... print(v)
+ 1
+ 2
+ 3
+ """
+ return Deque._Iterator(self._front)
+
+ def __repr__(self) -> str:
+ """
+ Implements representation of the deque.
+ Represents it as a list, with its values between '[' and ']'.
+ Time complexity: O(n)
+ >>> our_deque = Deque([1, 2, 3])
+ >>> our_deque
+ [1, 2, 3]
+ """
+ values_list = []
+ aux = self._front
+ while aux is not None:
+ # append the values in a list to display
+ values_list.append(aux.val)
+ aux = aux.next_node
+
+ return f"[{', '.join(repr(val) for val in values_list)}]"
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ dq = Deque([3])
+ dq.pop()
diff --git a/data_structures/queue/linked_queue.py b/data_structures/queues/linked_queue.py
similarity index 89%
rename from data_structures/queue/linked_queue.py
rename to data_structures/queues/linked_queue.py
index 8526ad311ed0..80f6d309af9a 100644
--- a/data_structures/queue/linked_queue.py
+++ b/data_structures/queues/linked_queue.py
@@ -1,11 +1,15 @@
-""" A Queue using a linked list like structure """
+"""A Queue using a linked list like structure"""
+
+from __future__ import annotations
+
+from collections.abc import Iterator
from typing import Any
class Node:
def __init__(self, data: Any) -> None:
- self.data = data
- self.next = None
+ self.data: Any = data
+ self.next: Node | None = None
def __str__(self) -> str:
return f"{self.data}"
@@ -19,7 +23,7 @@ class LinkedQueue:
>>> queue.put(5)
>>> queue.put(9)
>>> queue.put('python')
- >>> queue.is_empty();
+ >>> queue.is_empty()
False
>>> queue.get()
5
@@ -39,9 +43,10 @@ class LinkedQueue:
"""
def __init__(self) -> None:
- self.front = self.rear = None
+ self.front: Node | None = None
+ self.rear: Node | None = None
- def __iter__(self):
+ def __iter__(self) -> Iterator[Any]:
node = self.front
while node:
yield node.data
@@ -87,12 +92,12 @@ def is_empty(self) -> bool:
"""
return len(self) == 0
- def put(self, item) -> None:
+ def put(self, item: Any) -> None:
"""
>>> queue = LinkedQueue()
>>> queue.get()
Traceback (most recent call last):
- ...
+ ...
IndexError: dequeue from empty queue
>>> for i in range(1, 6):
... queue.put(i)
@@ -112,7 +117,7 @@ def get(self) -> Any:
>>> queue = LinkedQueue()
>>> queue.get()
Traceback (most recent call last):
- ...
+ ...
IndexError: dequeue from empty queue
>>> queue = LinkedQueue()
>>> for i in range(1, 6):
diff --git a/data_structures/queue/priority_queue_using_list.py b/data_structures/queues/priority_queue_using_list.py
similarity index 96%
rename from data_structures/queue/priority_queue_using_list.py
rename to data_structures/queues/priority_queue_using_list.py
index c5cf26433fff..15e56c557069 100644
--- a/data_structures/queue/priority_queue_using_list.py
+++ b/data_structures/queues/priority_queue_using_list.py
@@ -58,13 +58,13 @@ class FixedPriorityQueue:
4
>>> fpq.dequeue()
Traceback (most recent call last):
- ...
- data_structures.queue.priority_queue_using_list.UnderFlowError: All queues are empty
+ ...
+ data_structures.queues.priority_queue_using_list.UnderFlowError: All queues are empty
>>> print(fpq)
Priority 0: []
Priority 1: []
Priority 2: []
- """
+ """ # noqa: E501
def __init__(self):
self.queues = [
@@ -141,7 +141,7 @@ class ElementPriorityQueue:
>>> epq.dequeue()
Traceback (most recent call last):
...
- data_structures.queue.priority_queue_using_list.UnderFlowError: The queue is empty
+ data_structures.queues.priority_queue_using_list.UnderFlowError: The queue is empty
>>> print(epq)
[]
"""
diff --git a/data_structures/queues/queue_by_list.py b/data_structures/queues/queue_by_list.py
new file mode 100644
index 000000000000..4b05be9fd08e
--- /dev/null
+++ b/data_structures/queues/queue_by_list.py
@@ -0,0 +1,141 @@
+"""Queue represented by a Python list"""
+
+from collections.abc import Iterable
+from typing import Generic, TypeVar
+
+_T = TypeVar("_T")
+
+
+class QueueByList(Generic[_T]):
+ def __init__(self, iterable: Iterable[_T] | None = None) -> None:
+ """
+ >>> QueueByList()
+ Queue(())
+ >>> QueueByList([10, 20, 30])
+ Queue((10, 20, 30))
+ >>> QueueByList((i**2 for i in range(1, 4)))
+ Queue((1, 4, 9))
+ """
+ self.entries: list[_T] = list(iterable or [])
+
+ def __len__(self) -> int:
+ """
+ >>> len(QueueByList())
+ 0
+ >>> from string import ascii_lowercase
+ >>> len(QueueByList(ascii_lowercase))
+ 26
+ >>> queue = QueueByList()
+ >>> for i in range(1, 11):
+ ... queue.put(i)
+ >>> len(queue)
+ 10
+ >>> for i in range(2):
+ ... queue.get()
+ 1
+ 2
+ >>> len(queue)
+ 8
+ """
+
+ return len(self.entries)
+
+ def __repr__(self) -> str:
+ """
+ >>> queue = QueueByList()
+ >>> queue
+ Queue(())
+ >>> str(queue)
+ 'Queue(())'
+ >>> queue.put(10)
+ >>> queue
+ Queue((10,))
+ >>> queue.put(20)
+ >>> queue.put(30)
+ >>> queue
+ Queue((10, 20, 30))
+ """
+
+ return f"Queue({tuple(self.entries)})"
+
+ def put(self, item: _T) -> None:
+ """Put `item` to the Queue
+
+ >>> queue = QueueByList()
+ >>> queue.put(10)
+ >>> queue.put(20)
+ >>> len(queue)
+ 2
+ >>> queue
+ Queue((10, 20))
+ """
+
+ self.entries.append(item)
+
+ def get(self) -> _T:
+ """
+ Get `item` from the Queue
+
+ >>> queue = QueueByList((10, 20, 30))
+ >>> queue.get()
+ 10
+ >>> queue.put(40)
+ >>> queue.get()
+ 20
+ >>> queue.get()
+ 30
+ >>> len(queue)
+ 1
+ >>> queue.get()
+ 40
+ >>> queue.get()
+ Traceback (most recent call last):
+ ...
+ IndexError: Queue is empty
+ """
+
+ if not self.entries:
+ raise IndexError("Queue is empty")
+ return self.entries.pop(0)
+
+ def rotate(self, rotation: int) -> None:
+ """Rotate the items of the Queue `rotation` times
+
+ >>> queue = QueueByList([10, 20, 30, 40])
+ >>> queue
+ Queue((10, 20, 30, 40))
+ >>> queue.rotate(1)
+ >>> queue
+ Queue((20, 30, 40, 10))
+ >>> queue.rotate(2)
+ >>> queue
+ Queue((40, 10, 20, 30))
+ """
+
+ put = self.entries.append
+ get = self.entries.pop
+
+ for _ in range(rotation):
+ put(get(0))
+
+ def get_front(self) -> _T:
+ """Get the front item from the Queue
+
+ >>> queue = QueueByList((10, 20, 30))
+ >>> queue.get_front()
+ 10
+ >>> queue
+ Queue((10, 20, 30))
+ >>> queue.get()
+ 10
+ >>> queue.get_front()
+ 20
+ """
+
+ return self.entries[0]
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/data_structures/queues/queue_by_two_stacks.py b/data_structures/queues/queue_by_two_stacks.py
new file mode 100644
index 000000000000..cd62f155a63b
--- /dev/null
+++ b/data_structures/queues/queue_by_two_stacks.py
@@ -0,0 +1,115 @@
+"""Queue implementation using two stacks"""
+
+from collections.abc import Iterable
+from typing import Generic, TypeVar
+
+_T = TypeVar("_T")
+
+
+class QueueByTwoStacks(Generic[_T]):
+ def __init__(self, iterable: Iterable[_T] | None = None) -> None:
+ """
+ >>> QueueByTwoStacks()
+ Queue(())
+ >>> QueueByTwoStacks([10, 20, 30])
+ Queue((10, 20, 30))
+ >>> QueueByTwoStacks((i**2 for i in range(1, 4)))
+ Queue((1, 4, 9))
+ """
+ self._stack1: list[_T] = list(iterable or [])
+ self._stack2: list[_T] = []
+
+ def __len__(self) -> int:
+ """
+ >>> len(QueueByTwoStacks())
+ 0
+ >>> from string import ascii_lowercase
+ >>> len(QueueByTwoStacks(ascii_lowercase))
+ 26
+ >>> queue = QueueByTwoStacks()
+ >>> for i in range(1, 11):
+ ... queue.put(i)
+ ...
+ >>> len(queue)
+ 10
+ >>> for i in range(2):
+ ... queue.get()
+ 1
+ 2
+ >>> len(queue)
+ 8
+ """
+
+ return len(self._stack1) + len(self._stack2)
+
+ def __repr__(self) -> str:
+ """
+ >>> queue = QueueByTwoStacks()
+ >>> queue
+ Queue(())
+ >>> str(queue)
+ 'Queue(())'
+ >>> queue.put(10)
+ >>> queue
+ Queue((10,))
+ >>> queue.put(20)
+ >>> queue.put(30)
+ >>> queue
+ Queue((10, 20, 30))
+ """
+ return f"Queue({tuple(self._stack2[::-1] + self._stack1)})"
+
+ def put(self, item: _T) -> None:
+ """
+ Put `item` into the Queue
+
+ >>> queue = QueueByTwoStacks()
+ >>> queue.put(10)
+ >>> queue.put(20)
+ >>> len(queue)
+ 2
+ >>> queue
+ Queue((10, 20))
+ """
+
+ self._stack1.append(item)
+
+ def get(self) -> _T:
+ """
+ Get `item` from the Queue
+
+ >>> queue = QueueByTwoStacks((10, 20, 30))
+ >>> queue.get()
+ 10
+ >>> queue.put(40)
+ >>> queue.get()
+ 20
+ >>> queue.get()
+ 30
+ >>> len(queue)
+ 1
+ >>> queue.get()
+ 40
+ >>> queue.get()
+ Traceback (most recent call last):
+ ...
+ IndexError: Queue is empty
+ """
+
+ # To reduce number of attribute look-ups in `while` loop.
+ stack1_pop = self._stack1.pop
+ stack2_append = self._stack2.append
+
+ if not self._stack2:
+ while self._stack1:
+ stack2_append(stack1_pop())
+
+ if not self._stack2:
+ raise IndexError("Queue is empty")
+ return self._stack2.pop()
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/data_structures/queue/queue_on_pseudo_stack.py b/data_structures/queues/queue_on_pseudo_stack.py
similarity index 81%
rename from data_structures/queue/queue_on_pseudo_stack.py
rename to data_structures/queues/queue_on_pseudo_stack.py
index 7fa2fb2566af..2da67ecc263c 100644
--- a/data_structures/queue/queue_on_pseudo_stack.py
+++ b/data_structures/queues/queue_on_pseudo_stack.py
@@ -1,5 +1,7 @@
"""Queue represented by a pseudo stack (represented by a list with pop and append)"""
+from typing import Any
+
class Queue:
def __init__(self):
@@ -14,7 +16,7 @@ def __str__(self):
@param item
item to enqueue"""
- def put(self, item):
+ def put(self, item: Any) -> None:
self.stack.append(item)
self.length = self.length + 1
@@ -23,7 +25,7 @@ def put(self, item):
@return dequeued
item that was dequeued"""
- def get(self):
+ def get(self) -> Any:
self.rotate(1)
dequeued = self.stack[self.length - 1]
self.stack = self.stack[:-1]
@@ -35,8 +37,8 @@ def get(self):
@param rotation
number of times to rotate queue"""
- def rotate(self, rotation):
- for i in range(rotation):
+ def rotate(self, rotation: int) -> None:
+ for _ in range(rotation):
temp = self.stack[0]
self.stack = self.stack[1:]
self.put(temp)
@@ -45,7 +47,7 @@ def rotate(self, rotation):
"""Reports item at the front of self
@return item at front of self.stack"""
- def front(self):
+ def front(self) -> Any:
front = self.get()
self.put(front)
self.rotate(self.length - 1)
@@ -53,5 +55,5 @@ def front(self):
"""Returns the length of this.stack"""
- def size(self):
+ def size(self) -> int:
return self.length
diff --git a/data_structures/stacks/balanced_parentheses.py b/data_structures/stacks/balanced_parentheses.py
index 674f7ea436ed..928815bb2111 100644
--- a/data_structures/stacks/balanced_parentheses.py
+++ b/data_structures/stacks/balanced_parentheses.py
@@ -14,14 +14,15 @@ def balanced_parentheses(parentheses: str) -> bool:
>>> balanced_parentheses("")
True
"""
- stack = Stack()
+ stack: Stack[str] = Stack()
bracket_pairs = {"(": ")", "[": "]", "{": "}"}
for bracket in parentheses:
if bracket in bracket_pairs:
stack.push(bracket)
- elif bracket in (")", "]", "}"):
- if stack.is_empty() or bracket_pairs[stack.pop()] != bracket:
- return False
+ elif bracket in (")", "]", "}") and (
+ stack.is_empty() or bracket_pairs[stack.pop()] != bracket
+ ):
+ return False
return stack.is_empty()
diff --git a/data_structures/stacks/dijkstras_two_stack_algorithm.py b/data_structures/stacks/dijkstras_two_stack_algorithm.py
index 8b4668f9f839..94d19156f1c3 100644
--- a/data_structures/stacks/dijkstras_two_stack_algorithm.py
+++ b/data_structures/stacks/dijkstras_two_stack_algorithm.py
@@ -10,7 +10,7 @@
THESE ARE THE ALGORITHM'S RULES:
RULE 1: Scan the expression from left to right. When an operand is encountered,
- push it onto the the operand stack.
+ push it onto the operand stack.
RULE 2: When an operator is encountered in the expression,
push it onto the operator stack.
@@ -29,6 +29,7 @@
NOTE: It only works with whole numbers.
"""
+
__author__ = "Alexander Joslin"
import operator as op
@@ -51,8 +52,8 @@ def dijkstras_two_stack_algorithm(equation: str) -> int:
"""
operators = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub}
- operand_stack = Stack()
- operator_stack = Stack()
+ operand_stack: Stack[int] = Stack()
+ operator_stack: Stack[str] = Stack()
for i in equation:
if i.isdigit():
diff --git a/data_structures/stacks/evaluate_postfix_notations.py b/data_structures/stacks/evaluate_postfix_notations.py
deleted file mode 100644
index a03cb43bb020..000000000000
--- a/data_structures/stacks/evaluate_postfix_notations.py
+++ /dev/null
@@ -1,49 +0,0 @@
-"""
-The Reverse Polish Nation also known as Polish postfix notation
-or simply postfix notation.
-https://en.wikipedia.org/wiki/Reverse_Polish_notation
-Classic examples of simple stack implementations
-Valid operators are +, -, *, /.
-Each operand may be an integer or another expression.
-"""
-
-
-def evaluate_postfix(postfix_notation: list) -> int:
- """
- >>> evaluate_postfix(["2", "1", "+", "3", "*"])
- 9
- >>> evaluate_postfix(["4", "13", "5", "/", "+"])
- 6
- >>> evaluate_postfix([])
- 0
- """
- if not postfix_notation:
- return 0
-
- operations = {"+", "-", "*", "/"}
- stack = []
-
- for token in postfix_notation:
- if token in operations:
- b, a = stack.pop(), stack.pop()
- if token == "+":
- stack.append(a + b)
- elif token == "-":
- stack.append(a - b)
- elif token == "*":
- stack.append(a * b)
- else:
- if a * b < 0 and a % b != 0:
- stack.append(a // b + 1)
- else:
- stack.append(a // b)
- else:
- stack.append(int(token))
-
- return stack.pop()
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/data_structures/stacks/infix_to_postfix_conversion.py b/data_structures/stacks/infix_to_postfix_conversion.py
index dedba8479ac8..e697061937c9 100644
--- a/data_structures/stacks/infix_to_postfix_conversion.py
+++ b/data_structures/stacks/infix_to_postfix_conversion.py
@@ -4,9 +4,26 @@
https://en.wikipedia.org/wiki/Shunting-yard_algorithm
"""
+from typing import Literal
+
from .balanced_parentheses import balanced_parentheses
from .stack import Stack
+PRECEDENCES: dict[str, int] = {
+ "+": 1,
+ "-": 1,
+ "*": 2,
+ "/": 2,
+ "^": 3,
+}
+ASSOCIATIVITIES: dict[str, Literal["LR", "RL"]] = {
+ "+": "LR",
+ "-": "LR",
+ "*": "LR",
+ "/": "LR",
+ "^": "RL",
+}
+
def precedence(char: str) -> int:
"""
@@ -14,14 +31,22 @@ def precedence(char: str) -> int:
order of operation.
https://en.wikipedia.org/wiki/Order_of_operations
"""
- return {"+": 1, "-": 1, "*": 2, "/": 2, "^": 3}.get(char, -1)
+ return PRECEDENCES.get(char, -1)
+
+
+def associativity(char: str) -> Literal["LR", "RL"]:
+ """
+ Return the associativity of the operator `char`.
+ https://en.wikipedia.org/wiki/Operator_associativity
+ """
+ return ASSOCIATIVITIES[char]
def infix_to_postfix(expression_str: str) -> str:
"""
>>> infix_to_postfix("(1*(2+3)+4))")
Traceback (most recent call last):
- ...
+ ...
ValueError: Mismatched parentheses
>>> infix_to_postfix("")
''
@@ -35,10 +60,12 @@ def infix_to_postfix(expression_str: str) -> str:
'a b c * + d e * f + g * +'
>>> infix_to_postfix("x^y/(5*z)+2")
'x y ^ 5 z * / 2 +'
+ >>> infix_to_postfix("2^3^2")
+ '2 3 2 ^ ^'
"""
if not balanced_parentheses(expression_str):
raise ValueError("Mismatched parentheses")
- stack = Stack()
+ stack: Stack[str] = Stack()
postfix = []
for char in expression_str:
if char.isalpha() or char.isdigit():
@@ -50,9 +77,26 @@ def infix_to_postfix(expression_str: str) -> str:
postfix.append(stack.pop())
stack.pop()
else:
- while not stack.is_empty() and precedence(char) <= precedence(stack.peek()):
+ while True:
+ if stack.is_empty():
+ stack.push(char)
+ break
+
+ char_precedence = precedence(char)
+ tos_precedence = precedence(stack.peek())
+
+ if char_precedence > tos_precedence:
+ stack.push(char)
+ break
+ if char_precedence < tos_precedence:
+ postfix.append(stack.pop())
+ continue
+ # Precedences are equal
+ if associativity(char) == "RL":
+ stack.push(char)
+ break
postfix.append(stack.pop())
- stack.push(char)
+
while not stack.is_empty():
postfix.append(stack.pop())
return " ".join(postfix)
diff --git a/data_structures/stacks/infix_to_prefix_conversion.py b/data_structures/stacks/infix_to_prefix_conversion.py
index d3dc9e3e9c73..878473b93c19 100644
--- a/data_structures/stacks/infix_to_prefix_conversion.py
+++ b/data_structures/stacks/infix_to_prefix_conversion.py
@@ -15,9 +15,55 @@
"""
-def infix_2_postfix(Infix):
- Stack = []
- Postfix = []
+def infix_2_postfix(infix: str) -> str:
+ """
+ >>> infix_2_postfix("a+b^c") # doctest: +NORMALIZE_WHITESPACE
+ Symbol | Stack | Postfix
+ ----------------------------
+ a | | a
+ + | + | a
+ b | + | ab
+ ^ | +^ | ab
+ c | +^ | abc
+ | + | abc^
+ | | abc^+
+ 'abc^+'
+
+ >>> infix_2_postfix("1*((-a)*2+b)") # doctest: +NORMALIZE_WHITESPACE
+ Symbol | Stack | Postfix
+ -------------------------------------------
+ 1 | | 1
+ * | * | 1
+ ( | *( | 1
+ ( | *(( | 1
+ - | *((- | 1
+ a | *((- | 1a
+ ) | *( | 1a-
+ * | *(* | 1a-
+ 2 | *(* | 1a-2
+ + | *(+ | 1a-2*
+ b | *(+ | 1a-2*b
+ ) | * | 1a-2*b+
+ | | 1a-2*b+*
+ '1a-2*b+*'
+
+ >>> infix_2_postfix("")
+ Symbol | Stack | Postfix
+ ----------------------------
+ ''
+
+ >>> infix_2_postfix("(()")
+ Traceback (most recent call last):
+ ...
+ ValueError: invalid expression
+
+ >>> infix_2_postfix("())")
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ """
+ stack = []
+ post_fix = []
priority = {
"^": 3,
"*": 2,
@@ -26,7 +72,7 @@ def infix_2_postfix(Infix):
"+": 1,
"-": 1,
} # Priority of each operator
- print_width = len(Infix) if (len(Infix) > 7) else 7
+ print_width = max(len(infix), 7)
# Print table header for output
print(
@@ -37,57 +83,110 @@ def infix_2_postfix(Infix):
)
print("-" * (print_width * 3 + 7))
- for x in Infix:
+ for x in infix:
if x.isalpha() or x.isdigit():
- Postfix.append(x) # if x is Alphabet / Digit, add it to Postfix
+ post_fix.append(x) # if x is Alphabet / Digit, add it to Postfix
elif x == "(":
- Stack.append(x) # if x is "(" push to Stack
+ stack.append(x) # if x is "(" push to Stack
elif x == ")": # if x is ")" pop stack until "(" is encountered
- while Stack[-1] != "(":
- Postfix.append(Stack.pop()) # Pop stack & add the content to Postfix
- Stack.pop()
- else:
- if len(Stack) == 0:
- Stack.append(x) # If stack is empty, push x to stack
- else: # while priority of x is not > priority of element in the stack
- while len(Stack) > 0 and priority[x] <= priority[Stack[-1]]:
- Postfix.append(Stack.pop()) # pop stack & add to Postfix
- Stack.append(x) # push x to stack
+ if len(stack) == 0: # close bracket without open bracket
+ raise IndexError("list index out of range")
+
+ while stack[-1] != "(":
+ post_fix.append(stack.pop()) # Pop stack & add the content to Postfix
+ stack.pop()
+ elif len(stack) == 0:
+ stack.append(x) # If stack is empty, push x to stack
+ else: # while priority of x is not > priority of element in the stack
+ while stack and stack[-1] != "(" and priority[x] <= priority[stack[-1]]:
+ post_fix.append(stack.pop()) # pop stack & add to Postfix
+ stack.append(x) # push x to stack
print(
x.center(8),
- ("".join(Stack)).ljust(print_width),
- ("".join(Postfix)).ljust(print_width),
+ ("".join(stack)).ljust(print_width),
+ ("".join(post_fix)).ljust(print_width),
sep=" | ",
) # Output in tabular format
- while len(Stack) > 0: # while stack is not empty
- Postfix.append(Stack.pop()) # pop stack & add to Postfix
+ while len(stack) > 0: # while stack is not empty
+ if stack[-1] == "(": # open bracket with no close bracket
+ raise ValueError("invalid expression")
+
+ post_fix.append(stack.pop()) # pop stack & add to Postfix
print(
" ".center(8),
- ("".join(Stack)).ljust(print_width),
- ("".join(Postfix)).ljust(print_width),
+ ("".join(stack)).ljust(print_width),
+ ("".join(post_fix)).ljust(print_width),
sep=" | ",
) # Output in tabular format
- return "".join(Postfix) # return Postfix as str
+ return "".join(post_fix) # return Postfix as str
-def infix_2_prefix(Infix):
- Infix = list(Infix[::-1]) # reverse the infix equation
+def infix_2_prefix(infix: str) -> str:
+ """
+ >>> infix_2_prefix("a+b^c") # doctest: +NORMALIZE_WHITESPACE
+ Symbol | Stack | Postfix
+ ----------------------------
+ c | | c
+ ^ | ^ | c
+ b | ^ | cb
+ + | + | cb^
+ a | + | cb^a
+ | | cb^a+
+ '+a^bc'
- for i in range(len(Infix)):
- if Infix[i] == "(":
- Infix[i] = ")" # change "(" to ")"
- elif Infix[i] == ")":
- Infix[i] = "(" # change ")" to "("
+ >>> infix_2_prefix("1*((-a)*2+b)") # doctest: +NORMALIZE_WHITESPACE
+ Symbol | Stack | Postfix
+ -------------------------------------------
+ ( | ( |
+ b | ( | b
+ + | (+ | b
+ 2 | (+ | b2
+ * | (+* | b2
+ ( | (+*( | b2
+ a | (+*( | b2a
+ - | (+*(- | b2a
+ ) | (+* | b2a-
+ ) | | b2a-*+
+ * | * | b2a-*+
+ 1 | * | b2a-*+1
+ | | b2a-*+1*
+ '*1+*-a2b'
- return (infix_2_postfix("".join(Infix)))[
- ::-1
- ] # call infix_2_postfix on Infix, return reverse of Postfix
+ >>> infix_2_prefix('')
+ Symbol | Stack | Postfix
+ ----------------------------
+ ''
+
+ >>> infix_2_prefix('(()')
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+
+ >>> infix_2_prefix('())')
+ Traceback (most recent call last):
+ ...
+ ValueError: invalid expression
+ """
+ reversed_infix = list(infix[::-1]) # reverse the infix equation
+
+ for i in range(len(reversed_infix)):
+ if reversed_infix[i] == "(":
+ reversed_infix[i] = ")" # change "(" to ")"
+ elif reversed_infix[i] == ")":
+ reversed_infix[i] = "(" # change ")" to "("
+
+ # call infix_2_postfix on Infix, return reverse of Postfix
+ return (infix_2_postfix("".join(reversed_infix)))[::-1]
if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+
Infix = input("\nEnter an Infix Equation = ") # Input an Infix equation
Infix = "".join(Infix.split()) # Remove spaces from the input
print("\n\t", Infix, "(Infix) -> ", infix_2_prefix(Infix), "(Prefix)")
diff --git a/data_structures/stacks/largest_rectangle_histogram.py b/data_structures/stacks/largest_rectangle_histogram.py
new file mode 100644
index 000000000000..7575bd9f628d
--- /dev/null
+++ b/data_structures/stacks/largest_rectangle_histogram.py
@@ -0,0 +1,39 @@
+def largest_rectangle_area(heights: list[int]) -> int:
+ """
+ Inputs an array of integers representing the heights of bars,
+ and returns the area of the largest rectangle that can be formed
+
+ >>> largest_rectangle_area([2, 1, 5, 6, 2, 3])
+ 10
+
+ >>> largest_rectangle_area([2, 4])
+ 4
+
+ >>> largest_rectangle_area([6, 2, 5, 4, 5, 1, 6])
+ 12
+
+ >>> largest_rectangle_area([1])
+ 1
+ """
+ stack: list[int] = []
+ max_area = 0
+ heights = [*heights, 0] # make a new list by appending the sentinel 0
+ n = len(heights)
+
+ for i in range(n):
+ # make sure the stack remains in increasing order
+ while stack and heights[i] < heights[stack[-1]]:
+ h = heights[stack.pop()] # height of the bar
+ # if stack is empty, it means entire width can be taken from index 0 to i-1
+ w = i if not stack else i - stack[-1] - 1 # calculate width
+ max_area = max(max_area, h * w)
+
+ stack.append(i)
+
+ return max_area
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/stacks/lexicographical_numbers.py b/data_structures/stacks/lexicographical_numbers.py
new file mode 100644
index 000000000000..6a174e7d9e95
--- /dev/null
+++ b/data_structures/stacks/lexicographical_numbers.py
@@ -0,0 +1,38 @@
+from collections.abc import Iterator
+
+
+def lexical_order(max_number: int) -> Iterator[int]:
+ """
+ Generate numbers in lexical order from 1 to max_number.
+
+ >>> " ".join(map(str, lexical_order(13)))
+ '1 10 11 12 13 2 3 4 5 6 7 8 9'
+ >>> list(lexical_order(1))
+ [1]
+ >>> " ".join(map(str, lexical_order(20)))
+ '1 10 11 12 13 14 15 16 17 18 19 2 20 3 4 5 6 7 8 9'
+ >>> " ".join(map(str, lexical_order(25)))
+ '1 10 11 12 13 14 15 16 17 18 19 2 20 21 22 23 24 25 3 4 5 6 7 8 9'
+ >>> list(lexical_order(12))
+ [1, 10, 11, 12, 2, 3, 4, 5, 6, 7, 8, 9]
+ """
+
+ stack = [1]
+
+ while stack:
+ num = stack.pop()
+ if num > max_number:
+ continue
+
+ yield num
+ if (num % 10) != 9:
+ stack.append(num + 1)
+
+ stack.append(num * 10)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ print(f"Numbers from 1 to 25 in lexical order: {list(lexical_order(26))}")
diff --git a/data_structures/stacks/next_greater_element.py b/data_structures/stacks/next_greater_element.py
index d8c7ed17317b..216850b4b894 100644
--- a/data_structures/stacks/next_greater_element.py
+++ b/data_structures/stacks/next_greater_element.py
@@ -1,73 +1,108 @@
+from __future__ import annotations
+
arr = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0]
expect = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1]
-def next_greatest_element_slow(arr: list) -> list:
+def next_greatest_element_slow(arr: list[float]) -> list[float]:
"""
- Get the Next Greatest Element (NGE) for all elements in a list.
- Maximum element present after the current one which is also greater than the
- current one.
+ Get the Next Greatest Element (NGE) for each element in the array
+ by checking all subsequent elements to find the next greater one.
+
+ This is a brute-force implementation, and it has a time complexity
+ of O(n^2), where n is the size of the array.
+
+ Args:
+ arr: List of numbers for which the NGE is calculated.
+
+ Returns:
+ List containing the next greatest elements. If no
+ greater element is found, -1 is placed in the result.
+
+ Example:
>>> next_greatest_element_slow(arr) == expect
True
"""
+
result = []
- for i in range(0, len(arr), 1):
- next = -1
- for j in range(i + 1, len(arr), 1):
+ arr_size = len(arr)
+
+ for i in range(arr_size):
+ next_element: float = -1
+ for j in range(i + 1, arr_size):
if arr[i] < arr[j]:
- next = arr[j]
+ next_element = arr[j]
break
- result.append(next)
+ result.append(next_element)
return result
-def next_greatest_element_fast(arr: list) -> list:
+def next_greatest_element_fast(arr: list[float]) -> list[float]:
"""
- Like next_greatest_element_slow() but changes the loops to use
- enumerate() instead of range(len()) for the outer loop and
- for in a slice of arr for the inner loop.
+ Find the Next Greatest Element (NGE) for each element in the array
+ using a more readable approach. This implementation utilizes
+ enumerate() for the outer loop and slicing for the inner loop.
+
+ While this improves readability over next_greatest_element_slow(),
+ it still has a time complexity of O(n^2).
+
+ Args:
+ arr: List of numbers for which the NGE is calculated.
+
+ Returns:
+ List containing the next greatest elements. If no
+ greater element is found, -1 is placed in the result.
+
+ Example:
>>> next_greatest_element_fast(arr) == expect
True
"""
result = []
for i, outer in enumerate(arr):
- next = -1
+ next_item: float = -1
for inner in arr[i + 1 :]:
if outer < inner:
- next = inner
+ next_item = inner
break
- result.append(next)
+ result.append(next_item)
return result
-def next_greatest_element(arr: list) -> list:
+def next_greatest_element(arr: list[float]) -> list[float]:
"""
- Get the Next Greatest Element (NGE) for all elements in a list.
- Maximum element present after the current one which is also greater than the
- current one.
-
- A naive way to solve this is to take two loops and check for the next bigger
- number but that will make the time complexity as O(n^2). The better way to solve
- this would be to use a stack to keep track of maximum number giving a linear time
- solution.
+ Efficient solution to find the Next Greatest Element (NGE) for all elements
+ using a stack. The time complexity is reduced to O(n), making it suitable
+ for larger arrays.
+
+ The stack keeps track of elements for which the next greater element hasn't
+ been found yet. By iterating through the array in reverse (from the last
+ element to the first), the stack is used to efficiently determine the next
+ greatest element for each element.
+
+ Args:
+ arr: List of numbers for which the NGE is calculated.
+
+ Returns:
+ List containing the next greatest elements. If no
+ greater element is found, -1 is placed in the result.
+
+ Example:
>>> next_greatest_element(arr) == expect
True
"""
- stack = []
- result = [-1] * len(arr)
+ arr_size = len(arr)
+ stack: list[float] = []
+ result: list[float] = [-1] * arr_size
- for index in reversed(range(len(arr))):
- if len(stack):
+ for index in reversed(range(arr_size)):
+ if stack:
while stack[-1] <= arr[index]:
stack.pop()
- if len(stack) == 0:
+ if not stack:
break
-
- if len(stack) != 0:
+ if stack:
result[index] = stack[-1]
-
stack.append(arr[index])
-
return result
diff --git a/data_structures/stacks/postfix_evaluation.py b/data_structures/stacks/postfix_evaluation.py
index 574acac71c43..03a87b9e0fa3 100644
--- a/data_structures/stacks/postfix_evaluation.py
+++ b/data_structures/stacks/postfix_evaluation.py
@@ -1,4 +1,11 @@
"""
+Reverse Polish Nation is also known as Polish postfix notation or simply postfix
+notation.
+https://en.wikipedia.org/wiki/Reverse_Polish_notation
+Classic examples of simple stack implementations.
+Valid operators are +, -, *, /.
+Each operand may be an integer or another expression.
+
Output:
Enter a Postfix Equation (space separated) = 5 6 9 * +
@@ -17,52 +24,177 @@
Result = 59
"""
-import operator as op
+# Defining valid unary operator symbols
+UNARY_OP_SYMBOLS = ("-", "+")
+# operators & their respective operation
+OPERATORS = {
+ "^": lambda p, q: p**q,
+ "*": lambda p, q: p * q,
+ "/": lambda p, q: p / q,
+ "+": lambda p, q: p + q,
+ "-": lambda p, q: p - q,
+}
-def Solve(Postfix):
- Stack = []
- Div = lambda x, y: int(x / y) # noqa: E731 integer division operation
- Opr = {
- "^": op.pow,
- "*": op.mul,
- "/": Div,
- "+": op.add,
- "-": op.sub,
- } # operators & their respective operation
- # print table header
- print("Symbol".center(8), "Action".center(12), "Stack", sep=" | ")
- print("-" * (30 + len(Postfix)))
+def parse_token(token: str | float) -> float | str:
+ """
+ Converts the given data to the appropriate number if it is indeed a number, else
+ returns the data as it is with a False flag. This function also serves as a check
+ of whether the input is a number or not.
- for x in Postfix:
- if x.isdigit(): # if x in digit
- Stack.append(x) # append x to stack
- # output in tabular format
- print(x.rjust(8), ("push(" + x + ")").ljust(12), ",".join(Stack), sep=" | ")
- else:
- B = Stack.pop() # pop stack
- # output in tabular format
- print("".rjust(8), ("pop(" + B + ")").ljust(12), ",".join(Stack), sep=" | ")
+ Parameters
+ ----------
+ token: The data that needs to be converted to the appropriate operator or number.
- A = Stack.pop() # pop stack
- # output in tabular format
- print("".rjust(8), ("pop(" + A + ")").ljust(12), ",".join(Stack), sep=" | ")
+ Returns
+ -------
+ float or str
+ Returns a float if `token` is a number or a str if `token` is an operator
+ """
+ if token in OPERATORS:
+ return token
+ try:
+ return float(token)
+ except ValueError:
+ msg = f"{token} is neither a number nor a valid operator"
+ raise ValueError(msg)
+
+
+def evaluate(post_fix: list[str], verbose: bool = False) -> float:
+ """
+ Evaluate postfix expression using a stack.
+ >>> evaluate(["0"])
+ 0.0
+ >>> evaluate(["-0"])
+ -0.0
+ >>> evaluate(["1"])
+ 1.0
+ >>> evaluate(["-1"])
+ -1.0
+ >>> evaluate(["-1.1"])
+ -1.1
+ >>> evaluate(["2", "1", "+", "3", "*"])
+ 9.0
+ >>> evaluate(["2", "1.9", "+", "3", "*"])
+ 11.7
+ >>> evaluate(["2", "-1.9", "+", "3", "*"])
+ 0.30000000000000027
+ >>> evaluate(["4", "13", "5", "/", "+"])
+ 6.6
+ >>> evaluate(["2", "-", "3", "+"])
+ 1.0
+ >>> evaluate(["-4", "5", "*", "6", "-"])
+ -26.0
+ >>> evaluate([])
+ 0
+ >>> evaluate(["4", "-", "6", "7", "/", "9", "8"])
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: Input is not a valid postfix expression
+
+ Parameters
+ ----------
+ post_fix:
+ The postfix expression is tokenized into operators and operands and stored
+ as a Python list
+
+ verbose:
+ Display stack contents while evaluating the expression if verbose is True
- Stack.append(
- str(Opr[x](int(A), int(B)))
- ) # evaluate the 2 values popped from stack & push result to stack
+ Returns
+ -------
+ float
+ The evaluated value
+ """
+ if not post_fix:
+ return 0
+ # Checking the list to find out whether the postfix expression is valid
+ valid_expression = [parse_token(token) for token in post_fix]
+ if verbose:
+ # print table header
+ print("Symbol".center(8), "Action".center(12), "Stack", sep=" | ")
+ print("-" * (30 + len(post_fix)))
+ stack = []
+ for x in valid_expression:
+ if x not in OPERATORS:
+ stack.append(x) # append x to stack
+ if verbose:
+ # output in tabular format
+ print(
+ f"{x}".rjust(8),
+ f"push({x})".ljust(12),
+ stack,
+ sep=" | ",
+ )
+ continue
+ # If x is operator
+ # If only 1 value is inside the stack and + or - is encountered
+ # then this is unary + or - case
+ if x in UNARY_OP_SYMBOLS and len(stack) < 2:
+ b = stack.pop() # pop stack
+ if x == "-":
+ b *= -1 # negate b
+ stack.append(b)
+ if verbose:
+ # output in tabular format
+ print(
+ "".rjust(8),
+ f"pop({b})".ljust(12),
+ stack,
+ sep=" | ",
+ )
+ print(
+ str(x).rjust(8),
+ f"push({x}{b})".ljust(12),
+ stack,
+ sep=" | ",
+ )
+ continue
+ b = stack.pop() # pop stack
+ if verbose:
# output in tabular format
print(
- x.rjust(8),
- ("push(" + A + x + B + ")").ljust(12),
- ",".join(Stack),
+ "".rjust(8),
+ f"pop({b})".ljust(12),
+ stack,
sep=" | ",
)
- return int(Stack[0])
+ a = stack.pop() # pop stack
+ if verbose:
+ # output in tabular format
+ print(
+ "".rjust(8),
+ f"pop({a})".ljust(12),
+ stack,
+ sep=" | ",
+ )
+ # evaluate the 2 values popped from stack & push result to stack
+ stack.append(OPERATORS[x](a, b)) # type: ignore[index]
+ if verbose:
+ # output in tabular format
+ print(
+ f"{x}".rjust(8),
+ f"push({a}{x}{b})".ljust(12),
+ stack,
+ sep=" | ",
+ )
+ # If everything is executed correctly, the stack will contain
+ # only one element which is the result
+ if len(stack) != 1:
+ raise ArithmeticError("Input is not a valid postfix expression")
+ return float(stack[0])
if __name__ == "__main__":
- Postfix = input("\n\nEnter a Postfix Equation (space separated) = ").split(" ")
- print("\n\tResult = ", Solve(Postfix))
+ # Create a loop so that the user can evaluate postfix expressions multiple times
+ while True:
+ expression = input("Enter a Postfix Expression (space separated): ").split(" ")
+ prompt = "Do you want to see stack contents while evaluating? [y/N]: "
+ verbose = input(prompt).strip().lower() == "y"
+ output = evaluate(expression, verbose)
+ print("Result = ", output)
+ prompt = "Do you want to enter another expression? [y/N]: "
+ if input(prompt).strip().lower() != "y":
+ break
diff --git a/data_structures/stacks/prefix_evaluation.py b/data_structures/stacks/prefix_evaluation.py
index 00df2c1e63b0..03a70d884725 100644
--- a/data_structures/stacks/prefix_evaluation.py
+++ b/data_structures/stacks/prefix_evaluation.py
@@ -1,8 +1,9 @@
"""
-Python3 program to evaluate a prefix expression.
+Program to evaluate a prefix expression.
+https://en.wikipedia.org/wiki/Polish_notation
"""
-calc = {
+operators = {
"+": lambda x, y: x + y,
"-": lambda x, y: x - y,
"*": lambda x, y: x * y,
@@ -31,12 +32,15 @@ def evaluate(expression):
21
>>> evaluate("/ * 10 2 + 4 1 ")
4.0
+ >>> evaluate("2")
+ 2
+ >>> evaluate("+ * 2 3 / 8 4")
+ 8.0
"""
stack = []
# iterate over the string in reverse order
for c in expression.split()[::-1]:
-
# push operand to stack
if is_operand(c):
stack.append(int(c))
@@ -46,11 +50,39 @@ def evaluate(expression):
# push the result onto the stack again
o1 = stack.pop()
o2 = stack.pop()
- stack.append(calc[c](o1, o2))
+ stack.append(operators[c](o1, o2))
return stack.pop()
+def evaluate_recursive(expression: list[str]):
+ """
+ Alternative recursive implementation
+
+ >>> evaluate_recursive(['2'])
+ 2
+ >>> expression = ['+', '*', '2', '3', '/', '8', '4']
+ >>> evaluate_recursive(expression)
+ 8.0
+ >>> expression
+ []
+ >>> evaluate_recursive(['+', '9', '*', '2', '6'])
+ 21
+ >>> evaluate_recursive(['/', '*', '10', '2', '+', '4', '1'])
+ 4.0
+ """
+
+ op = expression.pop(0)
+ if is_operand(op):
+ return int(op)
+
+ operation = operators[op]
+
+ a = evaluate_recursive(expression)
+ b = evaluate_recursive(expression)
+ return operation(a, b)
+
+
# Driver code
if __name__ == "__main__":
test_expression = "+ 9 * 2 6"
diff --git a/data_structures/stacks/stack.py b/data_structures/stacks/stack.py
index 840cde099d38..93698f5aa116 100644
--- a/data_structures/stacks/stack.py
+++ b/data_structures/stacks/stack.py
@@ -1,8 +1,19 @@
+from __future__ import annotations
+
+from typing import Generic, TypeVar
+
+T = TypeVar("T")
+
+
class StackOverflowError(BaseException):
pass
-class Stack:
+class StackUnderflowError(BaseException):
+ pass
+
+
+class Stack(Generic[T]):
"""A stack is an abstract data type that serves as a collection of
elements with two principal operations: push() and pop(). push() adds an
element to the top of the stack, and pop() removes an element from the top
@@ -12,7 +23,7 @@ class Stack:
"""
def __init__(self, limit: int = 10):
- self.stack = []
+ self.stack: list[T] = []
self.limit = limit
def __bool__(self) -> bool:
@@ -21,33 +32,129 @@ def __bool__(self) -> bool:
def __str__(self) -> str:
return str(self.stack)
- def push(self, data):
- """ Push an element to the top of the stack."""
+ def push(self, data: T) -> None:
+ """
+ Push an element to the top of the stack.
+
+ >>> S = Stack(2) # stack size = 2
+ >>> S.push(10)
+ >>> S.push(20)
+ >>> print(S)
+ [10, 20]
+
+ >>> S = Stack(1) # stack size = 1
+ >>> S.push(10)
+ >>> S.push(20)
+ Traceback (most recent call last):
+ ...
+ data_structures.stacks.stack.StackOverflowError
+
+ """
if len(self.stack) >= self.limit:
raise StackOverflowError
self.stack.append(data)
- def pop(self):
- """ Pop an element off of the top of the stack."""
+ def pop(self) -> T:
+ """
+ Pop an element off of the top of the stack.
+
+ >>> S = Stack()
+ >>> S.push(-5)
+ >>> S.push(10)
+ >>> S.pop()
+ 10
+
+ >>> Stack().pop()
+ Traceback (most recent call last):
+ ...
+ data_structures.stacks.stack.StackUnderflowError
+ """
+ if not self.stack:
+ raise StackUnderflowError
return self.stack.pop()
- def peek(self):
- """ Peek at the top-most element of the stack."""
+ def peek(self) -> T:
+ """
+ Peek at the top-most element of the stack.
+
+ >>> S = Stack()
+ >>> S.push(-5)
+ >>> S.push(10)
+ >>> S.peek()
+ 10
+
+ >>> Stack().peek()
+ Traceback (most recent call last):
+ ...
+ data_structures.stacks.stack.StackUnderflowError
+ """
+ if not self.stack:
+ raise StackUnderflowError
return self.stack[-1]
def is_empty(self) -> bool:
- """ Check if a stack is empty."""
+ """
+ Check if a stack is empty.
+
+ >>> S = Stack()
+ >>> S.is_empty()
+ True
+
+ >>> S = Stack()
+ >>> S.push(10)
+ >>> S.is_empty()
+ False
+ """
return not bool(self.stack)
def is_full(self) -> bool:
+ """
+ >>> S = Stack()
+ >>> S.is_full()
+ False
+
+ >>> S = Stack(1)
+ >>> S.push(10)
+ >>> S.is_full()
+ True
+ """
return self.size() == self.limit
def size(self) -> int:
- """ Return the size of the stack."""
+ """
+ Return the size of the stack.
+
+ >>> S = Stack(3)
+ >>> S.size()
+ 0
+
+ >>> S = Stack(3)
+ >>> S.push(10)
+ >>> S.size()
+ 1
+
+ >>> S = Stack(3)
+ >>> S.push(10)
+ >>> S.push(20)
+ >>> S.size()
+ 2
+ """
return len(self.stack)
- def __contains__(self, item) -> bool:
- """Check if item is in stack"""
+ def __contains__(self, item: T) -> bool:
+ """
+ Check if item is in stack
+
+ >>> S = Stack(3)
+ >>> S.push(10)
+ >>> 10 in S
+ True
+
+ >>> S = Stack(3)
+ >>> S.push(10)
+ >>> 20 in S
+ False
+ """
return item in self.stack
@@ -55,7 +162,7 @@ def test_stack() -> None:
"""
>>> test_stack()
"""
- stack = Stack(10)
+ stack: Stack[int] = Stack(10)
assert bool(stack) is False
assert stack.is_empty() is True
assert stack.is_full() is False
@@ -63,23 +170,23 @@ def test_stack() -> None:
try:
_ = stack.pop()
- assert False # This should not happen
- except IndexError:
+ raise AssertionError # This should not happen
+ except StackUnderflowError:
assert True # This should happen
try:
_ = stack.peek()
- assert False # This should not happen
- except IndexError:
+ raise AssertionError # This should not happen
+ except StackUnderflowError:
assert True # This should happen
for i in range(10):
assert stack.size() == i
stack.push(i)
- assert bool(stack) is True
- assert stack.is_empty() is False
- assert stack.is_full() is True
+ assert bool(stack)
+ assert not stack.is_empty()
+ assert stack.is_full()
assert str(stack) == str(list(range(10)))
assert stack.pop() == 9
assert stack.peek() == 8
@@ -89,11 +196,11 @@ def test_stack() -> None:
try:
stack.push(200)
- assert False # This should not happen
+ raise AssertionError # This should not happen
except StackOverflowError:
assert True # This should happen
- assert stack.is_empty() is False
+ assert not stack.is_empty()
assert stack.size() == 10
assert 5 in stack
@@ -102,3 +209,7 @@ def test_stack() -> None:
if __name__ == "__main__":
test_stack()
+
+ import doctest
+
+ doctest.testmod()
diff --git a/data_structures/stacks/stack_using_two_queues.py b/data_structures/stacks/stack_using_two_queues.py
new file mode 100644
index 000000000000..4b73246a045c
--- /dev/null
+++ b/data_structures/stacks/stack_using_two_queues.py
@@ -0,0 +1,85 @@
+from __future__ import annotations
+
+from collections import deque
+from dataclasses import dataclass, field
+
+
+@dataclass
+class StackWithQueues:
+ """
+ https://www.geeksforgeeks.org/implement-stack-using-queue/
+
+ >>> stack = StackWithQueues()
+ >>> stack.push(1)
+ >>> stack.push(2)
+ >>> stack.push(3)
+ >>> stack.peek()
+ 3
+ >>> stack.pop()
+ 3
+ >>> stack.peek()
+ 2
+ >>> stack.pop()
+ 2
+ >>> stack.pop()
+ 1
+ >>> stack.peek() is None
+ True
+ >>> stack.pop()
+ Traceback (most recent call last):
+ ...
+ IndexError: pop from an empty deque
+ """
+
+ main_queue: deque[int] = field(default_factory=deque)
+ temp_queue: deque[int] = field(default_factory=deque)
+
+ def push(self, item: int) -> None:
+ self.temp_queue.append(item)
+ while self.main_queue:
+ self.temp_queue.append(self.main_queue.popleft())
+ self.main_queue, self.temp_queue = self.temp_queue, self.main_queue
+
+ def pop(self) -> int:
+ return self.main_queue.popleft()
+
+ def peek(self) -> int | None:
+ return self.main_queue[0] if self.main_queue else None
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ stack: StackWithQueues | None = StackWithQueues()
+ while stack:
+ print("\nChoose operation:")
+ print("1. Push")
+ print("2. Pop")
+ print("3. Peek")
+ print("4. Quit")
+
+ choice = input("Enter choice (1/2/3/4): ")
+
+ if choice == "1":
+ element = int(input("Enter an integer to push: ").strip())
+ stack.push(element)
+ print(f"{element} pushed onto the stack.")
+ elif choice == "2":
+ popped_element = stack.pop()
+ if popped_element is not None:
+ print(f"Popped element: {popped_element}")
+ else:
+ print("Stack is empty.")
+ elif choice == "3":
+ peeked_element = stack.peek()
+ if peeked_element is not None:
+ print(f"Top element: {peeked_element}")
+ else:
+ print("Stack is empty.")
+ elif choice == "4":
+ del stack
+ stack = None
+ else:
+ print("Invalid choice. Please try again.")
diff --git a/data_structures/stacks/stack_using_dll.py b/data_structures/stacks/stack_with_doubly_linked_list.py
similarity index 72%
rename from data_structures/stacks/stack_using_dll.py
rename to data_structures/stacks/stack_with_doubly_linked_list.py
index 75e0cd20640d..50c5236e073c 100644
--- a/data_structures/stacks/stack_using_dll.py
+++ b/data_structures/stacks/stack_with_doubly_linked_list.py
@@ -1,15 +1,21 @@
# A complete working Python program to demonstrate all
# stack operations using a doubly linked list
+from __future__ import annotations
-class Node:
- def __init__(self, data):
+from typing import Generic, TypeVar
+
+T = TypeVar("T")
+
+
+class Node(Generic[T]):
+ def __init__(self, data: T):
self.data = data # Assign data
- self.next = None # Initialize next as null
- self.prev = None # Initialize prev as null
+ self.next: Node[T] | None = None # Initialize next as null
+ self.prev: Node[T] | None = None # Initialize prev as null
-class Stack:
+class Stack(Generic[T]):
"""
>>> stack = Stack()
>>> stack.is_empty()
@@ -35,10 +41,10 @@ class Stack:
2->1->0->
"""
- def __init__(self):
- self.head = None
+ def __init__(self) -> None:
+ self.head: Node[T] | None = None
- def push(self, data):
+ def push(self, data: T) -> None:
"""add a Node to the stack"""
if self.head is None:
self.head = Node(data)
@@ -49,21 +55,23 @@ def push(self, data):
new_node.prev = None
self.head = new_node
- def pop(self):
+ def pop(self) -> T | None:
"""pop the top element off the stack"""
if self.head is None:
return None
else:
+ assert self.head is not None
temp = self.head.data
self.head = self.head.next
- self.head.prev = None
+ if self.head is not None:
+ self.head.prev = None
return temp
- def top(self):
+ def top(self) -> T | None:
"""return the top element of the stack"""
- return self.head.data
+ return self.head.data if self.head is not None else None
- def __len__(self):
+ def __len__(self) -> int:
temp = self.head
count = 0
while temp is not None:
@@ -71,10 +79,10 @@ def __len__(self):
temp = temp.next
return count
- def is_empty(self):
+ def is_empty(self) -> bool:
return self.head is None
- def print_stack(self):
+ def print_stack(self) -> None:
print("stack elements are:")
temp = self.head
while temp is not None:
@@ -84,9 +92,8 @@ def print_stack(self):
# Code execution starts here
if __name__ == "__main__":
-
# Start with the empty stack
- stack = Stack()
+ stack: Stack[int] = Stack()
# Insert 4 at the beginning. So stack becomes 4->None
print("Stack operations using Doubly LinkedList")
diff --git a/data_structures/stacks/linked_stack.py b/data_structures/stacks/stack_with_singly_linked_list.py
similarity index 79%
rename from data_structures/stacks/linked_stack.py
rename to data_structures/stacks/stack_with_singly_linked_list.py
index 1a2d07f20e7c..8e77c2b967ef 100644
--- a/data_structures/stacks/linked_stack.py
+++ b/data_structures/stacks/stack_with_singly_linked_list.py
@@ -1,17 +1,23 @@
-""" A Stack using a linked list like structure """
-from typing import Any
+"""A Stack using a linked list like structure"""
+from __future__ import annotations
-class Node:
- def __init__(self, data):
+from collections.abc import Iterator
+from typing import Generic, TypeVar
+
+T = TypeVar("T")
+
+
+class Node(Generic[T]):
+ def __init__(self, data: T):
self.data = data
- self.next = None
+ self.next: Node[T] | None = None
- def __str__(self):
+ def __str__(self) -> str:
return f"{self.data}"
-class LinkedStack:
+class LinkedStack(Generic[T]):
"""
Linked List Stack implementing push (to top),
pop (from top) and is_empty
@@ -42,15 +48,15 @@ class LinkedStack:
"""
def __init__(self) -> None:
- self.top = None
+ self.top: Node[T] | None = None
- def __iter__(self):
+ def __iter__(self) -> Iterator[T]:
node = self.top
while node:
yield node.data
node = node.next
- def __str__(self):
+ def __str__(self) -> str:
"""
>>> stack = LinkedStack()
>>> stack.push("c")
@@ -61,7 +67,7 @@ def __str__(self):
"""
return "->".join([str(item) for item in self])
- def __len__(self):
+ def __len__(self) -> int:
"""
>>> stack = LinkedStack()
>>> len(stack) == 0
@@ -85,7 +91,7 @@ def is_empty(self) -> bool:
"""
return self.top is None
- def push(self, item: Any) -> None:
+ def push(self, item: T) -> None:
"""
>>> stack = LinkedStack()
>>> stack.push("Python")
@@ -99,12 +105,12 @@ def push(self, item: Any) -> None:
node.next = self.top
self.top = node
- def pop(self) -> Any:
+ def pop(self) -> T:
"""
>>> stack = LinkedStack()
>>> stack.pop()
Traceback (most recent call last):
- ...
+ ...
IndexError: pop from empty stack
>>> stack.push("c")
>>> stack.push("b")
@@ -123,7 +129,7 @@ def pop(self) -> Any:
self.top = self.top.next
return pop_node.data
- def peek(self) -> Any:
+ def peek(self) -> T:
"""
>>> stack = LinkedStack()
>>> stack.push("Java")
@@ -134,6 +140,8 @@ def peek(self) -> Any:
"""
if self.is_empty():
raise IndexError("peek from empty stack")
+
+ assert self.top is not None
return self.top.data
def clear(self) -> None:
diff --git a/data_structures/stacks/stock_span_problem.py b/data_structures/stacks/stock_span_problem.py
index cc2adfdd6c21..5efe58d25798 100644
--- a/data_structures/stacks/stock_span_problem.py
+++ b/data_structures/stacks/stock_span_problem.py
@@ -8,19 +8,17 @@
"""
-def calculateSpan(price, S):
-
+def calculation_span(price, s):
n = len(price)
# Create a stack and push index of fist element to it
st = []
st.append(0)
# Span value of first element is always 1
- S[0] = 1
+ s[0] = 1
# Calculate span values for rest of the elements
for i in range(1, n):
-
# Pop elements from stack while stack is not
# empty and top of stack is smaller than price[i]
while len(st) > 0 and price[st[0]] <= price[i]:
@@ -30,15 +28,15 @@ def calculateSpan(price, S):
# than all elements on left of it, i.e. price[0],
# price[1], ..price[i-1]. Else the price[i] is
# greater than elements after top of stack
- S[i] = i + 1 if len(st) <= 0 else (i - st[0])
+ s[i] = i + 1 if len(st) <= 0 else (i - st[0])
# Push this element to stack
st.append(i)
# A utility function to print elements of array
-def printArray(arr, n):
- for i in range(0, n):
+def print_array(arr, n):
+ for i in range(n):
print(arr[i], end=" ")
@@ -47,7 +45,7 @@ def printArray(arr, n):
S = [0 for i in range(len(price) + 1)]
# Fill the span values in array S[]
-calculateSpan(price, S)
+calculation_span(price, S)
# Print the calculated span values
-printArray(S, len(price))
+print_array(S, len(price))
diff --git a/data_structures/suffix_tree/__init__.py b/data_structures/suffix_tree/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/data_structures/suffix_tree/example/__init__.py b/data_structures/suffix_tree/example/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/data_structures/suffix_tree/example/example_usage.py b/data_structures/suffix_tree/example/example_usage.py
new file mode 100644
index 000000000000..724ac57e8bfb
--- /dev/null
+++ b/data_structures/suffix_tree/example/example_usage.py
@@ -0,0 +1,37 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11554
+# https://github.com/TheAlgorithms/Python/pull/11554
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+from data_structures.suffix_tree.suffix_tree import SuffixTree
+
+
+def main() -> None:
+ """
+ Demonstrate the usage of the SuffixTree class.
+
+ - Initializes a SuffixTree with a predefined text.
+ - Defines a list of patterns to search for within the suffix tree.
+ - Searches for each pattern in the suffix tree.
+
+ Patterns tested:
+ - "ana" (found) --> True
+ - "ban" (found) --> True
+ - "na" (found) --> True
+ - "xyz" (not found) --> False
+ - "mon" (found) --> True
+ """
+ text = "monkey banana"
+ suffix_tree = SuffixTree(text)
+
+ patterns = ["ana", "ban", "na", "xyz", "mon"]
+ for pattern in patterns:
+ found = suffix_tree.search(pattern)
+ print(f"Pattern '{pattern}' found: {found}")
+
+
+if __name__ == "__main__":
+ main()
diff --git a/data_structures/suffix_tree/suffix_tree.py b/data_structures/suffix_tree/suffix_tree.py
new file mode 100644
index 000000000000..ad54fb0ba009
--- /dev/null
+++ b/data_structures/suffix_tree/suffix_tree.py
@@ -0,0 +1,66 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11554
+# https://github.com/TheAlgorithms/Python/pull/11554
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+from data_structures.suffix_tree.suffix_tree_node import SuffixTreeNode
+
+
+class SuffixTree:
+ def __init__(self, text: str) -> None:
+ """
+ Initializes the suffix tree with the given text.
+
+ Args:
+ text (str): The text for which the suffix tree is to be built.
+ """
+ self.text: str = text
+ self.root: SuffixTreeNode = SuffixTreeNode()
+ self.build_suffix_tree()
+
+ def build_suffix_tree(self) -> None:
+ """
+ Builds the suffix tree for the given text by adding all suffixes.
+ """
+ text = self.text
+ n = len(text)
+ for i in range(n):
+ suffix = text[i:]
+ self._add_suffix(suffix, i)
+
+ def _add_suffix(self, suffix: str, index: int) -> None:
+ """
+ Adds a suffix to the suffix tree.
+
+ Args:
+ suffix (str): The suffix to add.
+ index (int): The starting index of the suffix in the original text.
+ """
+ node = self.root
+ for char in suffix:
+ if char not in node.children:
+ node.children[char] = SuffixTreeNode()
+ node = node.children[char]
+ node.is_end_of_string = True
+ node.start = index
+ node.end = index + len(suffix) - 1
+
+ def search(self, pattern: str) -> bool:
+ """
+ Searches for a pattern in the suffix tree.
+
+ Args:
+ pattern (str): The pattern to search for.
+
+ Returns:
+ bool: True if the pattern is found, False otherwise.
+ """
+ node = self.root
+ for char in pattern:
+ if char not in node.children:
+ return False
+ node = node.children[char]
+ return True
diff --git a/data_structures/suffix_tree/suffix_tree_node.py b/data_structures/suffix_tree/suffix_tree_node.py
new file mode 100644
index 000000000000..e5b628645063
--- /dev/null
+++ b/data_structures/suffix_tree/suffix_tree_node.py
@@ -0,0 +1,36 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11554
+# https://github.com/TheAlgorithms/Python/pull/11554
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+from __future__ import annotations
+
+
+class SuffixTreeNode:
+ def __init__(
+ self,
+ children: dict[str, SuffixTreeNode] | None = None,
+ is_end_of_string: bool = False,
+ start: int | None = None,
+ end: int | None = None,
+ suffix_link: SuffixTreeNode | None = None,
+ ) -> None:
+ """
+ Initializes a suffix tree node.
+
+ Parameters:
+ children (dict[str, SuffixTreeNode] | None): The children of this node.
+ is_end_of_string (bool): Indicates if this node represents
+ the end of a string.
+ start (int | None): The start index of the suffix in the text.
+ end (int | None): The end index of the suffix in the text.
+ suffix_link (SuffixTreeNode | None): Link to another suffix tree node.
+ """
+ self.children = children or {}
+ self.is_end_of_string = is_end_of_string
+ self.start = start
+ self.end = end
+ self.suffix_link = suffix_link
diff --git a/data_structures/suffix_tree/tests/__init__.py b/data_structures/suffix_tree/tests/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/data_structures/suffix_tree/tests/test_suffix_tree.py b/data_structures/suffix_tree/tests/test_suffix_tree.py
new file mode 100644
index 000000000000..c9dbe199d19d
--- /dev/null
+++ b/data_structures/suffix_tree/tests/test_suffix_tree.py
@@ -0,0 +1,59 @@
+# Created by: Ramy-Badr-Ahmed (https://github.com/Ramy-Badr-Ahmed)
+# in Pull Request: #11554
+# https://github.com/TheAlgorithms/Python/pull/11554
+#
+# Please mention me (@Ramy-Badr-Ahmed) in any issue or pull request
+# addressing bugs/corrections to this file.
+# Thank you!
+
+import unittest
+
+from data_structures.suffix_tree.suffix_tree import SuffixTree
+
+
+class TestSuffixTree(unittest.TestCase):
+ def setUp(self) -> None:
+ """Set up the initial conditions for each test."""
+ self.text = "banana"
+ self.suffix_tree = SuffixTree(self.text)
+
+ def test_search_existing_patterns(self) -> None:
+ """Test searching for patterns that exist in the suffix tree."""
+ patterns = ["ana", "ban", "na"]
+ for pattern in patterns:
+ with self.subTest(pattern=pattern):
+ assert self.suffix_tree.search(pattern), (
+ f"Pattern '{pattern}' should be found."
+ )
+
+ def test_search_non_existing_patterns(self) -> None:
+ """Test searching for patterns that do not exist in the suffix tree."""
+ patterns = ["xyz", "apple", "cat"]
+ for pattern in patterns:
+ with self.subTest(pattern=pattern):
+ assert not self.suffix_tree.search(pattern), (
+ f"Pattern '{pattern}' should not be found."
+ )
+
+ def test_search_empty_pattern(self) -> None:
+ """Test searching for an empty pattern."""
+ assert self.suffix_tree.search(""), "An empty pattern should be found."
+
+ def test_search_full_text(self) -> None:
+ """Test searching for the full text."""
+ assert self.suffix_tree.search(self.text), (
+ "The full text should be found in the suffix tree."
+ )
+
+ def test_search_substrings(self) -> None:
+ """Test searching for substrings of the full text."""
+ substrings = ["ban", "ana", "a", "na"]
+ for substring in substrings:
+ with self.subTest(substring=substring):
+ assert self.suffix_tree.search(substring), (
+ f"Substring '{substring}' should be found."
+ )
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/data_structures/trie/radix_tree.py b/data_structures/trie/radix_tree.py
new file mode 100644
index 000000000000..caf566a6ce30
--- /dev/null
+++ b/data_structures/trie/radix_tree.py
@@ -0,0 +1,229 @@
+"""
+A Radix Tree is a data structure that represents a space-optimized
+trie (prefix tree) in whicheach node that is the only child is merged
+with its parent [https://en.wikipedia.org/wiki/Radix_tree]
+"""
+
+
+class RadixNode:
+ def __init__(self, prefix: str = "", is_leaf: bool = False) -> None:
+ # Mapping from the first character of the prefix of the node
+ self.nodes: dict[str, RadixNode] = {}
+
+ # A node will be a leaf if the tree contains its word
+ self.is_leaf = is_leaf
+
+ self.prefix = prefix
+
+ def match(self, word: str) -> tuple[str, str, str]:
+ """Compute the common substring of the prefix of the node and a word
+
+ Args:
+ word (str): word to compare
+
+ Returns:
+ (str, str, str): common substring, remaining prefix, remaining word
+
+ >>> RadixNode("myprefix").match("mystring")
+ ('my', 'prefix', 'string')
+ """
+ x = 0
+ for q, w in zip(self.prefix, word):
+ if q != w:
+ break
+
+ x += 1
+
+ return self.prefix[:x], self.prefix[x:], word[x:]
+
+ def insert_many(self, words: list[str]) -> None:
+ """Insert many words in the tree
+
+ Args:
+ words (list[str]): list of words
+
+ >>> RadixNode("myprefix").insert_many(["mystring", "hello"])
+ """
+ for word in words:
+ self.insert(word)
+
+ def insert(self, word: str) -> None:
+ """Insert a word into the tree
+
+ Args:
+ word (str): word to insert
+
+ >>> RadixNode("myprefix").insert("mystring")
+
+ >>> root = RadixNode()
+ >>> root.insert_many(['myprefix', 'myprefixA', 'myprefixAA'])
+ >>> root.print_tree()
+ - myprefix (leaf)
+ -- A (leaf)
+ --- A (leaf)
+ """
+ # Case 1: If the word is the prefix of the node
+ # Solution: We set the current node as leaf
+ if self.prefix == word and not self.is_leaf:
+ self.is_leaf = True
+
+ # Case 2: The node has no edges that have a prefix to the word
+ # Solution: We create an edge from the current node to a new one
+ # containing the word
+ elif word[0] not in self.nodes:
+ self.nodes[word[0]] = RadixNode(prefix=word, is_leaf=True)
+
+ else:
+ incoming_node = self.nodes[word[0]]
+ matching_string, remaining_prefix, remaining_word = incoming_node.match(
+ word
+ )
+
+ # Case 3: The node prefix is equal to the matching
+ # Solution: We insert remaining word on the next node
+ if remaining_prefix == "":
+ self.nodes[matching_string[0]].insert(remaining_word)
+
+ # Case 4: The word is greater equal to the matching
+ # Solution: Create a node in between both nodes, change
+ # prefixes and add the new node for the remaining word
+ else:
+ incoming_node.prefix = remaining_prefix
+
+ aux_node = self.nodes[matching_string[0]]
+ self.nodes[matching_string[0]] = RadixNode(matching_string, False)
+ self.nodes[matching_string[0]].nodes[remaining_prefix[0]] = aux_node
+
+ if remaining_word == "":
+ self.nodes[matching_string[0]].is_leaf = True
+ else:
+ self.nodes[matching_string[0]].insert(remaining_word)
+
+ def find(self, word: str) -> bool:
+ """Returns if the word is on the tree
+
+ Args:
+ word (str): word to check
+
+ Returns:
+ bool: True if the word appears on the tree
+
+ >>> RadixNode("myprefix").find("mystring")
+ False
+ """
+ incoming_node = self.nodes.get(word[0], None)
+ if not incoming_node:
+ return False
+ else:
+ matching_string, remaining_prefix, remaining_word = incoming_node.match(
+ word
+ )
+ # If there is remaining prefix, the word can't be on the tree
+ if remaining_prefix != "":
+ return False
+ # This applies when the word and the prefix are equal
+ elif remaining_word == "":
+ return incoming_node.is_leaf
+ # We have word remaining so we check the next node
+ else:
+ return incoming_node.find(remaining_word)
+
+ def delete(self, word: str) -> bool:
+ """Deletes a word from the tree if it exists
+
+ Args:
+ word (str): word to be deleted
+
+ Returns:
+ bool: True if the word was found and deleted. False if word is not found
+
+ >>> RadixNode("myprefix").delete("mystring")
+ False
+ """
+ incoming_node = self.nodes.get(word[0], None)
+ if not incoming_node:
+ return False
+ else:
+ matching_string, remaining_prefix, remaining_word = incoming_node.match(
+ word
+ )
+ # If there is remaining prefix, the word can't be on the tree
+ if remaining_prefix != "":
+ return False
+ # We have word remaining so we check the next node
+ elif remaining_word != "":
+ return incoming_node.delete(remaining_word)
+ # If it is not a leaf, we don't have to delete
+ elif not incoming_node.is_leaf:
+ return False
+ else:
+ # We delete the nodes if no edges go from it
+ if len(incoming_node.nodes) == 0:
+ del self.nodes[word[0]]
+ # We merge the current node with its only child
+ if len(self.nodes) == 1 and not self.is_leaf:
+ merging_node = next(iter(self.nodes.values()))
+ self.is_leaf = merging_node.is_leaf
+ self.prefix += merging_node.prefix
+ self.nodes = merging_node.nodes
+ # If there is more than 1 edge, we just mark it as non-leaf
+ elif len(incoming_node.nodes) > 1:
+ incoming_node.is_leaf = False
+ # If there is 1 edge, we merge it with its child
+ else:
+ merging_node = next(iter(incoming_node.nodes.values()))
+ incoming_node.is_leaf = merging_node.is_leaf
+ incoming_node.prefix += merging_node.prefix
+ incoming_node.nodes = merging_node.nodes
+
+ return True
+
+ def print_tree(self, height: int = 0) -> None:
+ """Print the tree
+
+ Args:
+ height (int, optional): Height of the printed node
+ """
+ if self.prefix != "":
+ print("-" * height, self.prefix, " (leaf)" if self.is_leaf else "")
+
+ for value in self.nodes.values():
+ value.print_tree(height + 1)
+
+
+def test_trie() -> bool:
+ words = "banana bananas bandana band apple all beast".split()
+ root = RadixNode()
+ root.insert_many(words)
+
+ assert all(root.find(word) for word in words)
+ assert not root.find("bandanas")
+ assert not root.find("apps")
+ root.delete("all")
+ assert not root.find("all")
+ root.delete("banana")
+ assert not root.find("banana")
+ assert root.find("bananas")
+
+ return True
+
+
+def pytests() -> None:
+ assert test_trie()
+
+
+def main() -> None:
+ """
+ >>> pytests()
+ """
+ root = RadixNode()
+ words = "banana bananas bandanas bandana band apple all beast".split()
+ root.insert_many(words)
+
+ print("Words:", words)
+ print("Tree:")
+ root.print_tree()
+
+
+if __name__ == "__main__":
+ main()
diff --git a/data_structures/trie/trie.py b/data_structures/trie/trie.py
index 6582be24fd0c..46b93a499d14 100644
--- a/data_structures/trie/trie.py
+++ b/data_structures/trie/trie.py
@@ -7,11 +7,11 @@
class TrieNode:
- def __init__(self):
- self.nodes = dict() # Mapping from char to TrieNode
+ def __init__(self) -> None:
+ self.nodes: dict[str, TrieNode] = {} # Mapping from char to TrieNode
self.is_leaf = False
- def insert_many(self, words: [str]):
+ def insert_many(self, words: list[str]) -> None:
"""
Inserts a list of words into the Trie
:param words: list of string words
@@ -20,7 +20,7 @@ def insert_many(self, words: [str]):
for word in words:
self.insert(word)
- def insert(self, word: str):
+ def insert(self, word: str) -> None:
"""
Inserts a word into the Trie
:param word: word to be inserted
@@ -46,14 +46,14 @@ def find(self, word: str) -> bool:
curr = curr.nodes[char]
return curr.is_leaf
- def delete(self, word: str):
+ def delete(self, word: str) -> None:
"""
Deletes a word in a Trie
:param word: word to delete
:return: None
"""
- def _delete(curr: TrieNode, word: str, index: int):
+ def _delete(curr: TrieNode, word: str, index: int) -> bool:
if index == len(word):
# If word does not exist
if not curr.is_leaf:
@@ -75,7 +75,7 @@ def _delete(curr: TrieNode, word: str, index: int):
_delete(self, word, 0)
-def print_words(node: TrieNode, word: str):
+def print_words(node: TrieNode, word: str) -> None:
"""
Prints all the words in a Trie
:param node: root node of Trie
@@ -89,7 +89,7 @@ def print_words(node: TrieNode, word: str):
print_words(value, word + key)
-def test_trie():
+def test_trie() -> bool:
words = "banana bananas bandana band apple all beast".split()
root = TrieNode()
root.insert_many(words)
@@ -112,11 +112,11 @@ def print_results(msg: str, passes: bool) -> None:
print(str(msg), "works!" if passes else "doesn't work :(")
-def pytests():
+def pytests() -> None:
assert test_trie()
-def main():
+def main() -> None:
"""
>>> pytests()
"""
diff --git a/digital_image_processing/change_contrast.py b/digital_image_processing/change_contrast.py
index 6a150400249f..7e49694708f8 100644
--- a/digital_image_processing/change_contrast.py
+++ b/digital_image_processing/change_contrast.py
@@ -4,8 +4,8 @@
This algorithm is used in
https://noivce.pythonanywhere.com/ Python web app.
-python/black: True
-flake8 : True
+psf/black: True
+ruff : True
"""
from PIL import Image
diff --git a/digital_image_processing/convert_to_negative.py b/digital_image_processing/convert_to_negative.py
index 7df44138973c..9bf2d8f2c075 100644
--- a/digital_image_processing/convert_to_negative.py
+++ b/digital_image_processing/convert_to_negative.py
@@ -1,6 +1,7 @@
"""
- Implemented an algorithm using opencv to convert a colored image into its negative
+Implemented an algorithm using opencv to convert a colored image into its negative
"""
+
from cv2 import destroyAllWindows, imread, imshow, waitKey
diff --git a/digital_image_processing/dithering/burkes.py b/digital_image_processing/dithering/burkes.py
index 2bf0bbe03225..4b59356d8f08 100644
--- a/digital_image_processing/dithering/burkes.py
+++ b/digital_image_processing/dithering/burkes.py
@@ -1,6 +1,7 @@
"""
Implementation Burke's algorithm (dithering)
"""
+
import numpy as np
from cv2 import destroyAllWindows, imread, imshow, waitKey
@@ -21,7 +22,8 @@ def __init__(self, input_img, threshold: int):
self.max_threshold = int(self.get_greyscale(255, 255, 255))
if not self.min_threshold < threshold < self.max_threshold:
- raise ValueError(f"Factor value should be from 0 to {self.max_threshold}")
+ msg = f"Factor value should be from 0 to {self.max_threshold}"
+ raise ValueError(msg)
self.input_img = input_img
self.threshold = threshold
@@ -38,9 +40,18 @@ def __init__(self, input_img, threshold: int):
def get_greyscale(cls, blue: int, green: int, red: int) -> float:
"""
>>> Burkes.get_greyscale(3, 4, 5)
- 3.753
+ 4.185
+ >>> Burkes.get_greyscale(0, 0, 0)
+ 0.0
+ >>> Burkes.get_greyscale(255, 255, 255)
+ 255.0
+ """
+ """
+ Formula from https://en.wikipedia.org/wiki/HSL_and_HSV
+ cf Lightness section, and Fig 13c.
+ We use the first of four possible.
"""
- return 0.114 * blue + 0.587 * green + 0.2126 * red
+ return 0.114 * blue + 0.587 * green + 0.299 * red
def process(self) -> None:
for y in range(self.height):
@@ -48,10 +59,10 @@ def process(self) -> None:
greyscale = int(self.get_greyscale(*self.input_img[y][x]))
if self.threshold > greyscale + self.error_table[y][x]:
self.output_img[y][x] = (0, 0, 0)
- current_error = greyscale + self.error_table[x][y]
+ current_error = greyscale + self.error_table[y][x]
else:
self.output_img[y][x] = (255, 255, 255)
- current_error = greyscale + self.error_table[x][y] - 255
+ current_error = greyscale + self.error_table[y][x] - 255
"""
Burkes error propagation (`*` is current pixel):
diff --git a/digital_image_processing/edge_detection/canny.py b/digital_image_processing/edge_detection/canny.py
index 295b4d825c12..944161c31cfc 100644
--- a/digital_image_processing/edge_detection/canny.py
+++ b/digital_image_processing/edge_detection/canny.py
@@ -18,105 +18,126 @@ def gen_gaussian_kernel(k_size, sigma):
return g
-def canny(image, threshold_low=15, threshold_high=30, weak=128, strong=255):
- image_row, image_col = image.shape[0], image.shape[1]
- # gaussian_filter
- gaussian_out = img_convolve(image, gen_gaussian_kernel(9, sigma=1.4))
- # get the gradient and degree by sobel_filter
- sobel_grad, sobel_theta = sobel_filter(gaussian_out)
- gradient_direction = np.rad2deg(sobel_theta)
- gradient_direction += PI
-
- dst = np.zeros((image_row, image_col))
-
+def suppress_non_maximum(image_shape, gradient_direction, sobel_grad):
"""
Non-maximum suppression. If the edge strength of the current pixel is the largest
compared to the other pixels in the mask with the same direction, the value will be
preserved. Otherwise, the value will be suppressed.
"""
- for row in range(1, image_row - 1):
- for col in range(1, image_col - 1):
+ destination = np.zeros(image_shape)
+
+ for row in range(1, image_shape[0] - 1):
+ for col in range(1, image_shape[1] - 1):
direction = gradient_direction[row, col]
if (
- 0 <= direction < 22.5
+ 0 <= direction < PI / 8
or 15 * PI / 8 <= direction <= 2 * PI
or 7 * PI / 8 <= direction <= 9 * PI / 8
):
- W = sobel_grad[row, col - 1]
- E = sobel_grad[row, col + 1]
- if sobel_grad[row, col] >= W and sobel_grad[row, col] >= E:
- dst[row, col] = sobel_grad[row, col]
-
- elif (PI / 8 <= direction < 3 * PI / 8) or (
- 9 * PI / 8 <= direction < 11 * PI / 8
+ w = sobel_grad[row, col - 1]
+ e = sobel_grad[row, col + 1]
+ if sobel_grad[row, col] >= w and sobel_grad[row, col] >= e:
+ destination[row, col] = sobel_grad[row, col]
+
+ elif (
+ PI / 8 <= direction < 3 * PI / 8
+ or 9 * PI / 8 <= direction < 11 * PI / 8
):
- SW = sobel_grad[row + 1, col - 1]
- NE = sobel_grad[row - 1, col + 1]
- if sobel_grad[row, col] >= SW and sobel_grad[row, col] >= NE:
- dst[row, col] = sobel_grad[row, col]
-
- elif (3 * PI / 8 <= direction < 5 * PI / 8) or (
- 11 * PI / 8 <= direction < 13 * PI / 8
+ sw = sobel_grad[row + 1, col - 1]
+ ne = sobel_grad[row - 1, col + 1]
+ if sobel_grad[row, col] >= sw and sobel_grad[row, col] >= ne:
+ destination[row, col] = sobel_grad[row, col]
+
+ elif (
+ 3 * PI / 8 <= direction < 5 * PI / 8
+ or 11 * PI / 8 <= direction < 13 * PI / 8
):
- N = sobel_grad[row - 1, col]
- S = sobel_grad[row + 1, col]
- if sobel_grad[row, col] >= N and sobel_grad[row, col] >= S:
- dst[row, col] = sobel_grad[row, col]
-
- elif (5 * PI / 8 <= direction < 7 * PI / 8) or (
- 13 * PI / 8 <= direction < 15 * PI / 8
+ n = sobel_grad[row - 1, col]
+ s = sobel_grad[row + 1, col]
+ if sobel_grad[row, col] >= n and sobel_grad[row, col] >= s:
+ destination[row, col] = sobel_grad[row, col]
+
+ elif (
+ 5 * PI / 8 <= direction < 7 * PI / 8
+ or 13 * PI / 8 <= direction < 15 * PI / 8
):
- NW = sobel_grad[row - 1, col - 1]
- SE = sobel_grad[row + 1, col + 1]
- if sobel_grad[row, col] >= NW and sobel_grad[row, col] >= SE:
- dst[row, col] = sobel_grad[row, col]
-
- """
- High-Low threshold detection. If an edge pixel’s gradient value is higher
- than the high threshold value, it is marked as a strong edge pixel. If an
- edge pixel’s gradient value is smaller than the high threshold value and
- larger than the low threshold value, it is marked as a weak edge pixel. If
- an edge pixel's value is smaller than the low threshold value, it will be
- suppressed.
- """
- if dst[row, col] >= threshold_high:
- dst[row, col] = strong
- elif dst[row, col] <= threshold_low:
- dst[row, col] = 0
+ nw = sobel_grad[row - 1, col - 1]
+ se = sobel_grad[row + 1, col + 1]
+ if sobel_grad[row, col] >= nw and sobel_grad[row, col] >= se:
+ destination[row, col] = sobel_grad[row, col]
+
+ return destination
+
+
+def detect_high_low_threshold(
+ image_shape, destination, threshold_low, threshold_high, weak, strong
+):
+ """
+ High-Low threshold detection. If an edge pixel's gradient value is higher
+ than the high threshold value, it is marked as a strong edge pixel. If an
+ edge pixel's gradient value is smaller than the high threshold value and
+ larger than the low threshold value, it is marked as a weak edge pixel. If
+ an edge pixel's value is smaller than the low threshold value, it will be
+ suppressed.
+ """
+ for row in range(1, image_shape[0] - 1):
+ for col in range(1, image_shape[1] - 1):
+ if destination[row, col] >= threshold_high:
+ destination[row, col] = strong
+ elif destination[row, col] <= threshold_low:
+ destination[row, col] = 0
else:
- dst[row, col] = weak
+ destination[row, col] = weak
+
+def track_edge(image_shape, destination, weak, strong):
"""
Edge tracking. Usually a weak edge pixel caused from true edges will be connected
to a strong edge pixel while noise responses are unconnected. As long as there is
one strong edge pixel that is involved in its 8-connected neighborhood, that weak
edge point can be identified as one that should be preserved.
"""
- for row in range(1, image_row):
- for col in range(1, image_col):
- if dst[row, col] == weak:
+ for row in range(1, image_shape[0]):
+ for col in range(1, image_shape[1]):
+ if destination[row, col] == weak:
if 255 in (
- dst[row, col + 1],
- dst[row, col - 1],
- dst[row - 1, col],
- dst[row + 1, col],
- dst[row - 1, col - 1],
- dst[row + 1, col - 1],
- dst[row - 1, col + 1],
- dst[row + 1, col + 1],
+ destination[row, col + 1],
+ destination[row, col - 1],
+ destination[row - 1, col],
+ destination[row + 1, col],
+ destination[row - 1, col - 1],
+ destination[row + 1, col - 1],
+ destination[row - 1, col + 1],
+ destination[row + 1, col + 1],
):
- dst[row, col] = strong
+ destination[row, col] = strong
else:
- dst[row, col] = 0
+ destination[row, col] = 0
+
+
+def canny(image, threshold_low=15, threshold_high=30, weak=128, strong=255):
+ # gaussian_filter
+ gaussian_out = img_convolve(image, gen_gaussian_kernel(9, sigma=1.4))
+ # get the gradient and degree by sobel_filter
+ sobel_grad, sobel_theta = sobel_filter(gaussian_out)
+ gradient_direction = PI + np.rad2deg(sobel_theta)
+
+ destination = suppress_non_maximum(image.shape, gradient_direction, sobel_grad)
+
+ detect_high_low_threshold(
+ image.shape, destination, threshold_low, threshold_high, weak, strong
+ )
+
+ track_edge(image.shape, destination, weak, strong)
- return dst
+ return destination
if __name__ == "__main__":
# read original image in gray mode
lena = cv2.imread(r"../image_data/lena.jpg", 0)
# canny edge detection
- canny_dst = canny(lena)
- cv2.imshow("canny", canny_dst)
+ canny_destination = canny(lena)
+ cv2.imshow("canny", canny_destination)
cv2.waitKey(0)
diff --git a/digital_image_processing/filters/bilateral_filter.py b/digital_image_processing/filters/bilateral_filter.py
index 76ae4dd20345..6ef4434d959c 100644
--- a/digital_image_processing/filters/bilateral_filter.py
+++ b/digital_image_processing/filters/bilateral_filter.py
@@ -9,6 +9,7 @@
Output:
img:A 2d zero padded image with values in between 0 and 1
"""
+
import math
import sys
@@ -31,8 +32,8 @@ def get_slice(img: np.ndarray, x: int, y: int, kernel_size: int) -> np.ndarray:
def get_gauss_kernel(kernel_size: int, spatial_variance: float) -> np.ndarray:
# Creates a gaussian kernel of given dimension.
arr = np.zeros((kernel_size, kernel_size))
- for i in range(0, kernel_size):
- for j in range(0, kernel_size):
+ for i in range(kernel_size):
+ for j in range(kernel_size):
arr[i, j] = math.sqrt(
abs(i - kernel_size // 2) ** 2 + abs(j - kernel_size // 2) ** 2
)
@@ -46,16 +47,15 @@ def bilateral_filter(
kernel_size: int,
) -> np.ndarray:
img2 = np.zeros(img.shape)
- gaussKer = get_gauss_kernel(kernel_size, spatial_variance)
- sizeX, sizeY = img.shape
- for i in range(kernel_size // 2, sizeX - kernel_size // 2):
- for j in range(kernel_size // 2, sizeY - kernel_size // 2):
-
- imgS = get_slice(img, i, j, kernel_size)
- imgI = imgS - imgS[kernel_size // 2, kernel_size // 2]
- imgIG = vec_gaussian(imgI, intensity_variance)
- weights = np.multiply(gaussKer, imgIG)
- vals = np.multiply(imgS, weights)
+ gauss_ker = get_gauss_kernel(kernel_size, spatial_variance)
+ size_x, size_y = img.shape
+ for i in range(kernel_size // 2, size_x - kernel_size // 2):
+ for j in range(kernel_size // 2, size_y - kernel_size // 2):
+ img_s = get_slice(img, i, j, kernel_size)
+ img_i = img_s - img_s[kernel_size // 2, kernel_size // 2]
+ img_ig = vec_gaussian(img_i, intensity_variance)
+ weights = np.multiply(gauss_ker, img_ig)
+ vals = np.multiply(img_s, weights)
val = np.sum(vals) / np.sum(weights)
img2[i, j] = val
return img2
diff --git a/digital_image_processing/filters/convolve.py b/digital_image_processing/filters/convolve.py
index 299682010da6..004402f29ba9 100644
--- a/digital_image_processing/filters/convolve.py
+++ b/digital_image_processing/filters/convolve.py
@@ -11,8 +11,8 @@ def im2col(image, block_size):
dst_width = rows - block_size[0] + 1
image_array = zeros((dst_height * dst_width, block_size[1] * block_size[0]))
row = 0
- for i in range(0, dst_height):
- for j in range(0, dst_width):
+ for i in range(dst_height):
+ for j in range(dst_width):
window = ravel(image[i : i + block_size[0], j : j + block_size[1]])
image_array[row, :] = window
row += 1
diff --git a/digital_image_processing/filters/gabor_filter.py b/digital_image_processing/filters/gabor_filter.py
new file mode 100644
index 000000000000..aaec567f4c99
--- /dev/null
+++ b/digital_image_processing/filters/gabor_filter.py
@@ -0,0 +1,85 @@
+# Implementation of the Gaborfilter
+# https://en.wikipedia.org/wiki/Gabor_filter
+import numpy as np
+from cv2 import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filter2D, imread, imshow, waitKey
+
+
+def gabor_filter_kernel(
+ ksize: int, sigma: int, theta: int, lambd: int, gamma: int, psi: int
+) -> np.ndarray:
+ """
+ :param ksize: The kernelsize of the convolutional filter (ksize x ksize)
+ :param sigma: standard deviation of the gaussian bell curve
+ :param theta: The orientation of the normal to the parallel stripes
+ of Gabor function.
+ :param lambd: Wavelength of the sinusoidal component.
+ :param gamma: The spatial aspect ratio and specifies the ellipticity
+ of the support of Gabor function.
+ :param psi: The phase offset of the sinusoidal function.
+
+ >>> gabor_filter_kernel(3, 8, 0, 10, 0, 0).tolist()
+ [[0.8027212023735046, 1.0, 0.8027212023735046], [0.8027212023735046, 1.0, \
+0.8027212023735046], [0.8027212023735046, 1.0, 0.8027212023735046]]
+
+ """
+
+ # prepare kernel
+ # the kernel size have to be odd
+ if (ksize % 2) == 0:
+ ksize = ksize + 1
+ gabor = np.zeros((ksize, ksize), dtype=np.float32)
+
+ # each value
+ for y in range(ksize):
+ for x in range(ksize):
+ # distance from center
+ px = x - ksize // 2
+ py = y - ksize // 2
+
+ # degree to radiant
+ _theta = theta / 180 * np.pi
+ cos_theta = np.cos(_theta)
+ sin_theta = np.sin(_theta)
+
+ # get kernel x
+ _x = cos_theta * px + sin_theta * py
+
+ # get kernel y
+ _y = -sin_theta * px + cos_theta * py
+
+ # fill kernel
+ gabor[y, x] = np.exp(-(_x**2 + gamma**2 * _y**2) / (2 * sigma**2)) * np.cos(
+ 2 * np.pi * _x / lambd + psi
+ )
+
+ return gabor
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ # read original image
+ img = imread("../image_data/lena.jpg")
+ # turn image in gray scale value
+ gray = cvtColor(img, COLOR_BGR2GRAY)
+
+ # Apply multiple Kernel to detect edges
+ out = np.zeros(gray.shape[:2])
+ for theta in [0, 30, 60, 90, 120, 150]:
+ """
+ ksize = 10
+ sigma = 8
+ lambd = 10
+ gamma = 0
+ psi = 0
+ """
+ kernel_10 = gabor_filter_kernel(10, 8, theta, 10, 0, 0)
+ out += filter2D(gray, CV_8UC3, kernel_10)
+ out = out / out.max() * 255
+ out = out.astype(np.uint8)
+
+ imshow("Original", gray)
+ imshow("Gabor filter with 20x20 mask and 6 directions", out)
+
+ waitKey(0)
diff --git a/digital_image_processing/filters/gaussian_filter.py b/digital_image_processing/filters/gaussian_filter.py
index 87fa67fb65ea..0c34e59fafe5 100644
--- a/digital_image_processing/filters/gaussian_filter.py
+++ b/digital_image_processing/filters/gaussian_filter.py
@@ -1,6 +1,7 @@
"""
Implementation of gaussian filter algorithm
"""
+
from itertools import product
from cv2 import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey
@@ -22,11 +23,9 @@ def gaussian_filter(image, k_size, sigma):
# im2col, turn the k_size*k_size pixels into a row and np.vstack all rows
image_array = zeros((dst_height * dst_width, k_size * k_size))
- row = 0
- for i, j in product(range(dst_height), range(dst_width)):
+ for row, (i, j) in enumerate(product(range(dst_height), range(dst_width))):
window = ravel(image[i : i + k_size, j : j + k_size])
image_array[row, :] = window
- row += 1
# turn the kernel into shape(k*k, 1)
gaussian_kernel = gen_gaussian_kernel(k_size, sigma)
diff --git a/digital_image_processing/filters/laplacian_filter.py b/digital_image_processing/filters/laplacian_filter.py
new file mode 100644
index 000000000000..69b9616e4d30
--- /dev/null
+++ b/digital_image_processing/filters/laplacian_filter.py
@@ -0,0 +1,81 @@
+# @Author : ojas-wani
+# @File : laplacian_filter.py
+# @Date : 10/04/2023
+
+import numpy as np
+from cv2 import (
+ BORDER_DEFAULT,
+ COLOR_BGR2GRAY,
+ CV_64F,
+ cvtColor,
+ filter2D,
+ imread,
+ imshow,
+ waitKey,
+)
+
+from digital_image_processing.filters.gaussian_filter import gaussian_filter
+
+
+def my_laplacian(src: np.ndarray, ksize: int) -> np.ndarray:
+ """
+ :param src: the source image, which should be a grayscale or color image.
+ :param ksize: the size of the kernel used to compute the Laplacian filter,
+ which can be 1, 3, 5, or 7.
+
+ >>> my_laplacian(src=np.array([]), ksize=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: ksize must be in (1, 3, 5, 7)
+ """
+ kernels = {
+ 1: np.array([[0, -1, 0], [-1, 4, -1], [0, -1, 0]]),
+ 3: np.array([[0, 1, 0], [1, -4, 1], [0, 1, 0]]),
+ 5: np.array(
+ [
+ [0, 0, -1, 0, 0],
+ [0, -1, -2, -1, 0],
+ [-1, -2, 16, -2, -1],
+ [0, -1, -2, -1, 0],
+ [0, 0, -1, 0, 0],
+ ]
+ ),
+ 7: np.array(
+ [
+ [0, 0, 0, -1, 0, 0, 0],
+ [0, 0, -2, -3, -2, 0, 0],
+ [0, -2, -7, -10, -7, -2, 0],
+ [-1, -3, -10, 68, -10, -3, -1],
+ [0, -2, -7, -10, -7, -2, 0],
+ [0, 0, -2, -3, -2, 0, 0],
+ [0, 0, 0, -1, 0, 0, 0],
+ ]
+ ),
+ }
+ if ksize not in kernels:
+ msg = f"ksize must be in {tuple(kernels)}"
+ raise ValueError(msg)
+
+ # Apply the Laplacian kernel using convolution
+ return filter2D(
+ src, CV_64F, kernels[ksize], 0, borderType=BORDER_DEFAULT, anchor=(0, 0)
+ )
+
+
+if __name__ == "__main__":
+ # read original image
+ img = imread(r"../image_data/lena.jpg")
+
+ # turn image in gray scale value
+ gray = cvtColor(img, COLOR_BGR2GRAY)
+
+ # Applying gaussian filter
+ blur_image = gaussian_filter(gray, 3, sigma=1)
+
+ # Apply multiple Kernel to detect edges
+ laplacian_image = my_laplacian(ksize=3, src=blur_image)
+
+ imshow("Original image", img)
+ imshow("Detected edges using laplacian filter", laplacian_image)
+
+ waitKey(0)
diff --git a/digital_image_processing/filters/local_binary_pattern.py b/digital_image_processing/filters/local_binary_pattern.py
new file mode 100644
index 000000000000..861369ba6a32
--- /dev/null
+++ b/digital_image_processing/filters/local_binary_pattern.py
@@ -0,0 +1,80 @@
+import cv2
+import numpy as np
+
+
+def get_neighbors_pixel(
+ image: np.ndarray, x_coordinate: int, y_coordinate: int, center: int
+) -> int:
+ """
+ Comparing local neighborhood pixel value with threshold value of centre pixel.
+ Exception is required when neighborhood value of a center pixel value is null.
+ i.e. values present at boundaries.
+
+ :param image: The image we're working with
+ :param x_coordinate: x-coordinate of the pixel
+ :param y_coordinate: The y coordinate of the pixel
+ :param center: center pixel value
+ :return: The value of the pixel is being returned.
+ """
+
+ try:
+ return int(image[x_coordinate][y_coordinate] >= center)
+ except (IndexError, TypeError):
+ return 0
+
+
+def local_binary_value(image: np.ndarray, x_coordinate: int, y_coordinate: int) -> int:
+ """
+ It takes an image, an x and y coordinate, and returns the
+ decimal value of the local binary patternof the pixel
+ at that coordinate
+
+ :param image: the image to be processed
+ :param x_coordinate: x coordinate of the pixel
+ :param y_coordinate: the y coordinate of the pixel
+ :return: The decimal value of the binary value of the pixels
+ around the center pixel.
+ """
+ center = image[x_coordinate][y_coordinate]
+ powers = [1, 2, 4, 8, 16, 32, 64, 128]
+
+ # skip get_neighbors_pixel if center is null
+ if center is None:
+ return 0
+
+ # Starting from the top right, assigning value to pixels clockwise
+ binary_values = [
+ get_neighbors_pixel(image, x_coordinate - 1, y_coordinate + 1, center),
+ get_neighbors_pixel(image, x_coordinate, y_coordinate + 1, center),
+ get_neighbors_pixel(image, x_coordinate - 1, y_coordinate, center),
+ get_neighbors_pixel(image, x_coordinate + 1, y_coordinate + 1, center),
+ get_neighbors_pixel(image, x_coordinate + 1, y_coordinate, center),
+ get_neighbors_pixel(image, x_coordinate + 1, y_coordinate - 1, center),
+ get_neighbors_pixel(image, x_coordinate, y_coordinate - 1, center),
+ get_neighbors_pixel(image, x_coordinate - 1, y_coordinate - 1, center),
+ ]
+
+ # Converting the binary value to decimal.
+ return sum(
+ binary_value * power for binary_value, power in zip(binary_values, powers)
+ )
+
+
+if __name__ == "__main__":
+ # Reading the image and converting it to grayscale.
+ image = cv2.imread(
+ "digital_image_processing/image_data/lena.jpg", cv2.IMREAD_GRAYSCALE
+ )
+
+ # Create a numpy array as the same height and width of read image
+ lbp_image = np.zeros((image.shape[0], image.shape[1]))
+
+ # Iterating through the image and calculating the
+ # local binary pattern value for each pixel.
+ for i in range(image.shape[0]):
+ for j in range(image.shape[1]):
+ lbp_image[i][j] = local_binary_value(image, i, j)
+
+ cv2.imshow("local binary pattern", lbp_image)
+ cv2.waitKey(0)
+ cv2.destroyAllWindows()
diff --git a/digital_image_processing/filters/median_filter.py b/digital_image_processing/filters/median_filter.py
index 174018569d62..fc8b582ef67a 100644
--- a/digital_image_processing/filters/median_filter.py
+++ b/digital_image_processing/filters/median_filter.py
@@ -1,6 +1,7 @@
"""
Implementation of median filter algorithm
"""
+
from cv2 import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey
from numpy import divide, int8, multiply, ravel, sort, zeros_like
diff --git a/digital_image_processing/histogram_equalization/histogram_stretch.py b/digital_image_processing/histogram_equalization/histogram_stretch.py
index 0288a2c1fcf5..1270c964dee6 100644
--- a/digital_image_processing/histogram_equalization/histogram_stretch.py
+++ b/digital_image_processing/histogram_equalization/histogram_stretch.py
@@ -3,6 +3,7 @@
@author: Binish125
"""
+
import copy
import os
@@ -11,7 +12,7 @@
from matplotlib import pyplot as plt
-class contrastStretch:
+class ConstantStretch:
def __init__(self):
self.img = ""
self.original_image = ""
@@ -45,10 +46,10 @@ def stretch(self, input_image):
self.img[j][i] = self.last_list[num]
cv2.imwrite("output_data/output.jpg", self.img)
- def plotHistogram(self):
+ def plot_histogram(self):
plt.hist(self.img.ravel(), 256, [0, 256])
- def showImage(self):
+ def show_image(self):
cv2.imshow("Output-Image", self.img)
cv2.imshow("Input-Image", self.original_image)
cv2.waitKey(5000)
@@ -57,7 +58,7 @@ def showImage(self):
if __name__ == "__main__":
file_path = os.path.join(os.path.basename(__file__), "image_data/input.jpg")
- stretcher = contrastStretch()
+ stretcher = ConstantStretch()
stretcher.stretch(file_path)
- stretcher.plotHistogram()
- stretcher.showImage()
+ stretcher.plot_histogram()
+ stretcher.show_image()
diff --git a/digital_image_processing/index_calculation.py b/digital_image_processing/index_calculation.py
index 4350b8603390..988f8e72b9a8 100644
--- a/digital_image_processing/index_calculation.py
+++ b/digital_image_processing/index_calculation.py
@@ -104,72 +104,71 @@ class IndexCalculation:
#RGBIndex = ["GLI", "CI", "Hue", "I", "NGRDI", "RI", "S", "IF"]
"""
- def __init__(self, red=None, green=None, blue=None, redEdge=None, nir=None):
- # print("Numpy version: " + np.__version__)
- self.setMatrices(red=red, green=green, blue=blue, redEdge=redEdge, nir=nir)
+ def __init__(self, red=None, green=None, blue=None, red_edge=None, nir=None):
+ self.set_matricies(red=red, green=green, blue=blue, red_edge=red_edge, nir=nir)
- def setMatrices(self, red=None, green=None, blue=None, redEdge=None, nir=None):
+ def set_matricies(self, red=None, green=None, blue=None, red_edge=None, nir=None):
if red is not None:
self.red = red
if green is not None:
self.green = green
if blue is not None:
self.blue = blue
- if redEdge is not None:
- self.redEdge = redEdge
+ if red_edge is not None:
+ self.redEdge = red_edge
if nir is not None:
self.nir = nir
return True
def calculation(
- self, index="", red=None, green=None, blue=None, redEdge=None, nir=None
+ self, index="", red=None, green=None, blue=None, red_edge=None, nir=None
):
"""
performs the calculation of the index with the values instantiated in the class
:str index: abbreviation of index name to perform
"""
- self.setMatrices(red=red, green=green, blue=blue, redEdge=redEdge, nir=nir)
+ self.set_matricies(red=red, green=green, blue=blue, red_edge=red_edge, nir=nir)
funcs = {
- "ARVI2": self.ARVI2,
- "CCCI": self.CCCI,
- "CVI": self.CVI,
- "GLI": self.GLI,
- "NDVI": self.NDVI,
- "BNDVI": self.BNDVI,
- "redEdgeNDVI": self.redEdgeNDVI,
- "GNDVI": self.GNDVI,
- "GBNDVI": self.GBNDVI,
- "GRNDVI": self.GRNDVI,
- "RBNDVI": self.RBNDVI,
- "PNDVI": self.PNDVI,
- "ATSAVI": self.ATSAVI,
- "BWDRVI": self.BWDRVI,
- "CIgreen": self.CIgreen,
- "CIrededge": self.CIrededge,
- "CI": self.CI,
- "CTVI": self.CTVI,
- "GDVI": self.GDVI,
- "EVI": self.EVI,
- "GEMI": self.GEMI,
- "GOSAVI": self.GOSAVI,
- "GSAVI": self.GSAVI,
- "Hue": self.Hue,
- "IVI": self.IVI,
- "IPVI": self.IPVI,
- "I": self.I,
- "RVI": self.RVI,
- "MRVI": self.MRVI,
- "MSAVI": self.MSAVI,
- "NormG": self.NormG,
- "NormNIR": self.NormNIR,
- "NormR": self.NormR,
- "NGRDI": self.NGRDI,
- "RI": self.RI,
- "S": self.S,
- "IF": self.IF,
- "DVI": self.DVI,
- "TVI": self.TVI,
- "NDRE": self.NDRE,
+ "ARVI2": self.arv12,
+ "CCCI": self.ccci,
+ "CVI": self.cvi,
+ "GLI": self.gli,
+ "NDVI": self.ndvi,
+ "BNDVI": self.bndvi,
+ "redEdgeNDVI": self.red_edge_ndvi,
+ "GNDVI": self.gndvi,
+ "GBNDVI": self.gbndvi,
+ "GRNDVI": self.grndvi,
+ "RBNDVI": self.rbndvi,
+ "PNDVI": self.pndvi,
+ "ATSAVI": self.atsavi,
+ "BWDRVI": self.bwdrvi,
+ "CIgreen": self.ci_green,
+ "CIrededge": self.ci_rededge,
+ "CI": self.ci,
+ "CTVI": self.ctvi,
+ "GDVI": self.gdvi,
+ "EVI": self.evi,
+ "GEMI": self.gemi,
+ "GOSAVI": self.gosavi,
+ "GSAVI": self.gsavi,
+ "Hue": self.hue,
+ "IVI": self.ivi,
+ "IPVI": self.ipvi,
+ "I": self.i,
+ "RVI": self.rvi,
+ "MRVI": self.mrvi,
+ "MSAVI": self.m_savi,
+ "NormG": self.norm_g,
+ "NormNIR": self.norm_nir,
+ "NormR": self.norm_r,
+ "NGRDI": self.ngrdi,
+ "RI": self.ri,
+ "S": self.s,
+ "IF": self._if,
+ "DVI": self.dvi,
+ "TVI": self.tvi,
+ "NDRE": self.ndre,
}
try:
@@ -178,16 +177,16 @@ def calculation(
print("Index not in the list!")
return False
- def ARVI2(self):
+ def arv12(self):
"""
Atmospherically Resistant Vegetation Index 2
https://www.indexdatabase.de/db/i-single.php?id=396
:return: index
- −0.18+1.17*(self.nir−self.red)/(self.nir+self.red)
+ -0.18+1.17*(self.nir-self.red)/(self.nir+self.red)
"""
return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red)))
- def CCCI(self):
+ def ccci(self):
"""
Canopy Chlorophyll Content Index
https://www.indexdatabase.de/db/i-single.php?id=224
@@ -197,15 +196,15 @@ def CCCI(self):
(self.nir - self.red) / (self.nir + self.red)
)
- def CVI(self):
+ def cvi(self):
"""
Chlorophyll vegetation index
https://www.indexdatabase.de/db/i-single.php?id=391
:return: index
"""
- return self.nir * (self.red / (self.green ** 2))
+ return self.nir * (self.red / (self.green**2))
- def GLI(self):
+ def gli(self):
"""
self.green leaf index
https://www.indexdatabase.de/db/i-single.php?id=375
@@ -215,7 +214,7 @@ def GLI(self):
2 * self.green + self.red + self.blue
)
- def NDVI(self):
+ def ndvi(self):
"""
Normalized Difference self.nir/self.red Normalized Difference Vegetation
Index, Calibrated NDVI - CDVI
@@ -224,7 +223,7 @@ def NDVI(self):
"""
return (self.nir - self.red) / (self.nir + self.red)
- def BNDVI(self):
+ def bndvi(self):
"""
Normalized Difference self.nir/self.blue self.blue-normalized difference
vegetation index
@@ -233,7 +232,7 @@ def BNDVI(self):
"""
return (self.nir - self.blue) / (self.nir + self.blue)
- def redEdgeNDVI(self):
+ def red_edge_ndvi(self):
"""
Normalized Difference self.rededge/self.red
https://www.indexdatabase.de/db/i-single.php?id=235
@@ -241,7 +240,7 @@ def redEdgeNDVI(self):
"""
return (self.redEdge - self.red) / (self.redEdge + self.red)
- def GNDVI(self):
+ def gndvi(self):
"""
Normalized Difference self.nir/self.green self.green NDVI
https://www.indexdatabase.de/db/i-single.php?id=401
@@ -249,7 +248,7 @@ def GNDVI(self):
"""
return (self.nir - self.green) / (self.nir + self.green)
- def GBNDVI(self):
+ def gbndvi(self):
"""
self.green-self.blue NDVI
https://www.indexdatabase.de/db/i-single.php?id=186
@@ -259,7 +258,7 @@ def GBNDVI(self):
self.nir + (self.green + self.blue)
)
- def GRNDVI(self):
+ def grndvi(self):
"""
self.green-self.red NDVI
https://www.indexdatabase.de/db/i-single.php?id=185
@@ -269,7 +268,7 @@ def GRNDVI(self):
self.nir + (self.green + self.red)
)
- def RBNDVI(self):
+ def rbndvi(self):
"""
self.red-self.blue NDVI
https://www.indexdatabase.de/db/i-single.php?id=187
@@ -277,7 +276,7 @@ def RBNDVI(self):
"""
return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red))
- def PNDVI(self):
+ def pndvi(self):
"""
Pan NDVI
https://www.indexdatabase.de/db/i-single.php?id=188
@@ -287,7 +286,7 @@ def PNDVI(self):
self.nir + (self.green + self.red + self.blue)
)
- def ATSAVI(self, X=0.08, a=1.22, b=0.03):
+ def atsavi(self, x=0.08, a=1.22, b=0.03):
"""
Adjusted transformed soil-adjusted VI
https://www.indexdatabase.de/db/i-single.php?id=209
@@ -295,10 +294,10 @@ def ATSAVI(self, X=0.08, a=1.22, b=0.03):
"""
return a * (
(self.nir - a * self.red - b)
- / (a * self.nir + self.red - a * b + X * (1 + a ** 2))
+ / (a * self.nir + self.red - a * b + x * (1 + a**2))
)
- def BWDRVI(self):
+ def bwdrvi(self):
"""
self.blue-wide dynamic range vegetation index
https://www.indexdatabase.de/db/i-single.php?id=136
@@ -306,7 +305,7 @@ def BWDRVI(self):
"""
return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue)
- def CIgreen(self):
+ def ci_green(self):
"""
Chlorophyll Index self.green
https://www.indexdatabase.de/db/i-single.php?id=128
@@ -314,7 +313,7 @@ def CIgreen(self):
"""
return (self.nir / self.green) - 1
- def CIrededge(self):
+ def ci_rededge(self):
"""
Chlorophyll Index self.redEdge
https://www.indexdatabase.de/db/i-single.php?id=131
@@ -322,7 +321,7 @@ def CIrededge(self):
"""
return (self.nir / self.redEdge) - 1
- def CI(self):
+ def ci(self):
"""
Coloration Index
https://www.indexdatabase.de/db/i-single.php?id=11
@@ -330,16 +329,16 @@ def CI(self):
"""
return (self.red - self.blue) / self.red
- def CTVI(self):
+ def ctvi(self):
"""
Corrected Transformed Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=244
:return: index
"""
- ndvi = self.NDVI()
+ ndvi = self.ndvi()
return ((ndvi + 0.5) / (abs(ndvi + 0.5))) * (abs(ndvi + 0.5) ** (1 / 2))
- def GDVI(self):
+ def gdvi(self):
"""
Difference self.nir/self.green self.green Difference Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=27
@@ -347,7 +346,7 @@ def GDVI(self):
"""
return self.nir - self.green
- def EVI(self):
+ def evi(self):
"""
Enhanced Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=16
@@ -357,36 +356,36 @@ def EVI(self):
(self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1)
)
- def GEMI(self):
+ def gemi(self):
"""
Global Environment Monitoring Index
https://www.indexdatabase.de/db/i-single.php?id=25
:return: index
"""
- n = (2 * (self.nir ** 2 - self.red ** 2) + 1.5 * self.nir + 0.5 * self.red) / (
+ n = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / (
self.nir + self.red + 0.5
)
return n * (1 - 0.25 * n) - (self.red - 0.125) / (1 - self.red)
- def GOSAVI(self, Y=0.16):
+ def gosavi(self, y=0.16):
"""
self.green Optimized Soil Adjusted Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=29
mit Y = 0,16
:return: index
"""
- return (self.nir - self.green) / (self.nir + self.green + Y)
+ return (self.nir - self.green) / (self.nir + self.green + y)
- def GSAVI(self, L=0.5):
+ def gsavi(self, n=0.5):
"""
self.green Soil Adjusted Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=31
- mit L = 0,5
+ mit N = 0,5
:return: index
"""
- return ((self.nir - self.green) / (self.nir + self.green + L)) * (1 + L)
+ return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n)
- def Hue(self):
+ def hue(self):
"""
Hue
https://www.indexdatabase.de/db/i-single.php?id=34
@@ -396,7 +395,7 @@ def Hue(self):
((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue)
)
- def IVI(self, a=None, b=None):
+ def ivi(self, a=None, b=None):
"""
Ideal vegetation index
https://www.indexdatabase.de/db/i-single.php?id=276
@@ -406,15 +405,15 @@ def IVI(self, a=None, b=None):
"""
return (self.nir - b) / (a * self.red)
- def IPVI(self):
+ def ipvi(self):
"""
Infraself.red percentage vegetation index
https://www.indexdatabase.de/db/i-single.php?id=35
:return: index
"""
- return (self.nir / ((self.nir + self.red) / 2)) * (self.NDVI() + 1)
+ return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1)
- def I(self): # noqa: E741,E743
+ def i(self):
"""
Intensity
https://www.indexdatabase.de/db/i-single.php?id=36
@@ -422,7 +421,7 @@ def I(self): # noqa: E741,E743
"""
return (self.red + self.green + self.blue) / 30.5
- def RVI(self):
+ def rvi(self):
"""
Ratio-Vegetation-Index
http://www.seos-project.eu/modules/remotesensing/remotesensing-c03-s01-p01.html
@@ -430,15 +429,15 @@ def RVI(self):
"""
return self.nir / self.red
- def MRVI(self):
+ def mrvi(self):
"""
Modified Normalized Difference Vegetation Index RVI
https://www.indexdatabase.de/db/i-single.php?id=275
:return: index
"""
- return (self.RVI() - 1) / (self.RVI() + 1)
+ return (self.rvi() - 1) / (self.rvi() + 1)
- def MSAVI(self):
+ def m_savi(self):
"""
Modified Soil Adjusted Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=44
@@ -449,7 +448,7 @@ def MSAVI(self):
- ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2)
) / 2
- def NormG(self):
+ def norm_g(self):
"""
Norm G
https://www.indexdatabase.de/db/i-single.php?id=50
@@ -457,7 +456,7 @@ def NormG(self):
"""
return self.green / (self.nir + self.red + self.green)
- def NormNIR(self):
+ def norm_nir(self):
"""
Norm self.nir
https://www.indexdatabase.de/db/i-single.php?id=51
@@ -465,7 +464,7 @@ def NormNIR(self):
"""
return self.nir / (self.nir + self.red + self.green)
- def NormR(self):
+ def norm_r(self):
"""
Norm R
https://www.indexdatabase.de/db/i-single.php?id=52
@@ -473,7 +472,7 @@ def NormR(self):
"""
return self.red / (self.nir + self.red + self.green)
- def NGRDI(self):
+ def ngrdi(self):
"""
Normalized Difference self.green/self.red Normalized self.green self.red
difference index, Visible Atmospherically Resistant Indices self.green
@@ -483,7 +482,7 @@ def NGRDI(self):
"""
return (self.green - self.red) / (self.green + self.red)
- def RI(self):
+ def ri(self):
"""
Normalized Difference self.red/self.green self.redness Index
https://www.indexdatabase.de/db/i-single.php?id=74
@@ -491,17 +490,17 @@ def RI(self):
"""
return (self.red - self.green) / (self.red + self.green)
- def S(self):
+ def s(self):
"""
Saturation
https://www.indexdatabase.de/db/i-single.php?id=77
:return: index
"""
- max = np.max([np.max(self.red), np.max(self.green), np.max(self.blue)])
- min = np.min([np.min(self.red), np.min(self.green), np.min(self.blue)])
- return (max - min) / max
+ max_value = np.max([np.max(self.red), np.max(self.green), np.max(self.blue)])
+ min_value = np.min([np.min(self.red), np.min(self.green), np.min(self.blue)])
+ return (max_value - min_value) / max_value
- def IF(self):
+ def _if(self):
"""
Shape Index
https://www.indexdatabase.de/db/i-single.php?id=79
@@ -509,7 +508,7 @@ def IF(self):
"""
return (2 * self.red - self.green - self.blue) / (self.green - self.blue)
- def DVI(self):
+ def dvi(self):
"""
Simple Ratio self.nir/self.red Difference Vegetation Index, Vegetation Index
Number (VIN)
@@ -518,15 +517,15 @@ def DVI(self):
"""
return self.nir / self.red
- def TVI(self):
+ def tvi(self):
"""
Transformed Vegetation Index
https://www.indexdatabase.de/db/i-single.php?id=98
:return: index
"""
- return (self.NDVI() + 0.5) ** (1 / 2)
+ return (self.ndvi() + 0.5) ** (1 / 2)
- def NDRE(self):
+ def ndre(self):
return (self.nir - self.redEdge) / (self.nir + self.redEdge)
diff --git a/digital_image_processing/morphological_operations/__init__.py b/digital_image_processing/morphological_operations/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/digital_image_processing/morphological_operations/dilation_operation.py b/digital_image_processing/morphological_operations/dilation_operation.py
new file mode 100644
index 000000000000..e49b955c1480
--- /dev/null
+++ b/digital_image_processing/morphological_operations/dilation_operation.py
@@ -0,0 +1,75 @@
+from pathlib import Path
+
+import numpy as np
+from PIL import Image
+
+
+def rgb_to_gray(rgb: np.ndarray) -> np.ndarray:
+ """
+ Return gray image from rgb image
+ >>> rgb_to_gray(np.array([[[127, 255, 0]]]))
+ array([[187.6453]])
+ >>> rgb_to_gray(np.array([[[0, 0, 0]]]))
+ array([[0.]])
+ >>> rgb_to_gray(np.array([[[2, 4, 1]]]))
+ array([[3.0598]])
+ >>> rgb_to_gray(np.array([[[26, 255, 14], [5, 147, 20], [1, 200, 0]]]))
+ array([[159.0524, 90.0635, 117.6989]])
+ """
+ r, g, b = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2]
+ return 0.2989 * r + 0.5870 * g + 0.1140 * b
+
+
+def gray_to_binary(gray: np.ndarray) -> np.ndarray:
+ """
+ Return binary image from gray image
+ >>> gray_to_binary(np.array([[127, 255, 0]]))
+ array([[False, True, False]])
+ >>> gray_to_binary(np.array([[0]]))
+ array([[False]])
+ >>> gray_to_binary(np.array([[26.2409, 4.9315, 1.4729]]))
+ array([[False, False, False]])
+ >>> gray_to_binary(np.array([[26, 255, 14], [5, 147, 20], [1, 200, 0]]))
+ array([[False, True, False],
+ [False, True, False],
+ [False, True, False]])
+ """
+ return (gray > 127) & (gray <= 255)
+
+
+def dilation(image: np.ndarray, kernel: np.ndarray) -> np.ndarray:
+ """
+ Return dilated image
+ >>> dilation(np.array([[True, False, True]]), np.array([[0, 1, 0]]))
+ array([[False, False, False]])
+ >>> dilation(np.array([[False, False, True]]), np.array([[1, 0, 1]]))
+ array([[False, False, False]])
+ """
+ output = np.zeros_like(image)
+ image_padded = np.zeros(
+ (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1)
+ )
+
+ # Copy image to padded image
+ image_padded[kernel.shape[0] - 2 : -1 :, kernel.shape[1] - 2 : -1 :] = image
+
+ # Iterate over image & apply kernel
+ for x in range(image.shape[1]):
+ for y in range(image.shape[0]):
+ summation = (
+ kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]]
+ ).sum()
+ output[y, x] = int(summation > 0)
+ return output
+
+
+if __name__ == "__main__":
+ # read original image
+ lena_path = Path(__file__).resolve().parent / "image_data" / "lena.jpg"
+ lena = np.array(Image.open(lena_path))
+ # kernel to be applied
+ structuring_element = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]])
+ output = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element)
+ # Save the output image
+ pil_img = Image.fromarray(output).convert("RGB")
+ pil_img.save("result_dilation.png")
diff --git a/digital_image_processing/morphological_operations/erosion_operation.py b/digital_image_processing/morphological_operations/erosion_operation.py
new file mode 100644
index 000000000000..53001da83468
--- /dev/null
+++ b/digital_image_processing/morphological_operations/erosion_operation.py
@@ -0,0 +1,82 @@
+from pathlib import Path
+
+import numpy as np
+from PIL import Image
+
+
+def rgb_to_gray(rgb: np.ndarray) -> np.ndarray:
+ """
+ Return gray image from rgb image
+
+ >>> rgb_to_gray(np.array([[[127, 255, 0]]]))
+ array([[187.6453]])
+ >>> rgb_to_gray(np.array([[[0, 0, 0]]]))
+ array([[0.]])
+ >>> rgb_to_gray(np.array([[[2, 4, 1]]]))
+ array([[3.0598]])
+ >>> rgb_to_gray(np.array([[[26, 255, 14], [5, 147, 20], [1, 200, 0]]]))
+ array([[159.0524, 90.0635, 117.6989]])
+ """
+ r, g, b = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2]
+ return 0.2989 * r + 0.5870 * g + 0.1140 * b
+
+
+def gray_to_binary(gray: np.ndarray) -> np.ndarray:
+ """
+ Return binary image from gray image
+
+ >>> gray_to_binary(np.array([[127, 255, 0]]))
+ array([[False, True, False]])
+ >>> gray_to_binary(np.array([[0]]))
+ array([[False]])
+ >>> gray_to_binary(np.array([[26.2409, 4.9315, 1.4729]]))
+ array([[False, False, False]])
+ >>> gray_to_binary(np.array([[26, 255, 14], [5, 147, 20], [1, 200, 0]]))
+ array([[False, True, False],
+ [False, True, False],
+ [False, True, False]])
+ """
+ return (gray > 127) & (gray <= 255)
+
+
+def erosion(image: np.ndarray, kernel: np.ndarray) -> np.ndarray:
+ """
+ Return eroded image
+
+ >>> erosion(np.array([[True, True, False]]), np.array([[0, 1, 0]]))
+ array([[False, False, False]])
+ >>> erosion(np.array([[True, False, False]]), np.array([[1, 1, 0]]))
+ array([[False, False, False]])
+ """
+ output = np.zeros_like(image)
+ image_padded = np.zeros(
+ (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1)
+ )
+
+ # Copy image to padded image
+ image_padded[kernel.shape[0] - 2 : -1 :, kernel.shape[1] - 2 : -1 :] = image
+
+ # Iterate over image & apply kernel
+ for x in range(image.shape[1]):
+ for y in range(image.shape[0]):
+ summation = (
+ kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]]
+ ).sum()
+ output[y, x] = int(summation == 5)
+ return output
+
+
+if __name__ == "__main__":
+ # read original image
+ lena_path = Path(__file__).resolve().parent / "image_data" / "lena.jpg"
+ lena = np.array(Image.open(lena_path))
+
+ # kernel to be applied
+ structuring_element = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]])
+
+ # Apply erosion operation to a binary image
+ output = erosion(gray_to_binary(rgb_to_gray(lena)), structuring_element)
+
+ # Save the output image
+ pil_img = Image.fromarray(output).convert("RGB")
+ pil_img.save("result_erosion.png")
diff --git a/digital_image_processing/resize/resize.py b/digital_image_processing/resize/resize.py
index 4836521f9f58..7bde118da69b 100644
--- a/digital_image_processing/resize/resize.py
+++ b/digital_image_processing/resize/resize.py
@@ -1,4 +1,5 @@
-""" Multiple image resizing techniques """
+"""Multiple image resizing techniques"""
+
import numpy as np
from cv2 import destroyAllWindows, imread, imshow, waitKey
diff --git a/digital_image_processing/rotation/rotation.py b/digital_image_processing/rotation/rotation.py
index 2951f18fc0ec..0f5e36ddd5be 100644
--- a/digital_image_processing/rotation/rotation.py
+++ b/digital_image_processing/rotation/rotation.py
@@ -1,19 +1,21 @@
+from pathlib import Path
+
import cv2
import numpy as np
from matplotlib import pyplot as plt
def get_rotation(
- img: np.array, pt1: np.float32, pt2: np.float32, rows: int, cols: int
-) -> np.array:
+ img: np.ndarray, pt1: np.ndarray, pt2: np.ndarray, rows: int, cols: int
+) -> np.ndarray:
"""
Get image rotation
- :param img: np.array
+ :param img: np.ndarray
:param pt1: 3x2 list
:param pt2: 3x2 list
:param rows: columns image shape
:param cols: rows image shape
- :return: np.array
+ :return: np.ndarray
"""
matrix = cv2.getAffineTransform(pt1, pt2)
return cv2.warpAffine(img, matrix, (rows, cols))
@@ -21,17 +23,19 @@ def get_rotation(
if __name__ == "__main__":
# read original image
- image = cv2.imread("lena.jpg")
+ image = cv2.imread(
+ str(Path(__file__).resolve().parent.parent / "image_data" / "lena.jpg")
+ )
# turn image in gray scale value
gray_img = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
# get image shape
img_rows, img_cols = gray_img.shape
# set different points to rotate image
- pts1 = np.float32([[50, 50], [200, 50], [50, 200]])
- pts2 = np.float32([[10, 100], [200, 50], [100, 250]])
- pts3 = np.float32([[50, 50], [150, 50], [120, 200]])
- pts4 = np.float32([[10, 100], [80, 50], [180, 250]])
+ pts1 = np.array([[50, 50], [200, 50], [50, 200]], np.float32)
+ pts2 = np.array([[10, 100], [200, 50], [100, 250]], np.float32)
+ pts3 = np.array([[50, 50], [150, 50], [120, 200]], np.float32)
+ pts4 = np.array([[10, 100], [80, 50], [180, 250]], np.float32)
# add all rotated images in a list
images = [
diff --git a/digital_image_processing/sepia.py b/digital_image_processing/sepia.py
index dfb5951676aa..1924a80451e5 100644
--- a/digital_image_processing/sepia.py
+++ b/digital_image_processing/sepia.py
@@ -1,6 +1,7 @@
"""
- Implemented an algorithm using opencv to tone an image with sepia technique
+Implemented an algorithm using opencv to tone an image with sepia technique
"""
+
from cv2 import destroyAllWindows, imread, imshow, waitKey
@@ -19,7 +20,7 @@ def to_grayscale(blue, green, red):
return 0.2126 * red + 0.587 * green + 0.114 * blue
def normalize(value):
- """ Helper function to normalize R/G/B value -> return 255 if value > 255"""
+ """Helper function to normalize R/G/B value -> return 255 if value > 255"""
return min(value, 255)
for i in range(pixel_h):
diff --git a/digital_image_processing/test_digital_image_processing.py b/digital_image_processing/test_digital_image_processing.py
index 40f2f7b83b6d..d1200f4d65ca 100644
--- a/digital_image_processing/test_digital_image_processing.py
+++ b/digital_image_processing/test_digital_image_processing.py
@@ -1,6 +1,8 @@
"""
PyTest's for Digital Image Processing
"""
+
+import numpy as np
from cv2 import COLOR_BGR2GRAY, cvtColor, imread
from numpy import array, uint8
from PIL import Image
@@ -9,9 +11,10 @@
from digital_image_processing import convert_to_negative as cn
from digital_image_processing import sepia as sp
from digital_image_processing.dithering import burkes as bs
-from digital_image_processing.edge_detection import canny as canny
+from digital_image_processing.edge_detection import canny
from digital_image_processing.filters import convolve as conv
from digital_image_processing.filters import gaussian_filter as gg
+from digital_image_processing.filters import local_binary_pattern as lbp
from digital_image_processing.filters import median_filter as med
from digital_image_processing.filters import sobel_filter as sob
from digital_image_processing.resize import resize as rs
@@ -60,8 +63,8 @@ def test_gen_gaussian_kernel_filter():
def test_convolve_filter():
# laplace diagonals
- Laplace = array([[0.25, 0.5, 0.25], [0.5, -3, 0.5], [0.25, 0.5, 0.25]])
- res = conv.img_convolve(gray, Laplace).astype(uint8)
+ laplace = array([[0.25, 0.5, 0.25], [0.5, -3, 0.5], [0.25, 0.5, 0.25]])
+ res = conv.img_convolve(gray, laplace).astype(uint8)
assert res.any()
@@ -71,7 +74,8 @@ def test_median_filter():
def test_sobel_filter():
grad, theta = sob.sobel_filter(gray)
- assert grad.any() and theta.any()
+ assert grad.any()
+ assert theta.any()
def test_sepia():
@@ -91,3 +95,40 @@ def test_nearest_neighbour(
nn = rs.NearestNeighbour(imread(file_path, 1), 400, 200)
nn.process()
assert nn.output.any()
+
+
+def test_local_binary_pattern():
+ # pull request 10161 before:
+ # "digital_image_processing/image_data/lena.jpg"
+ # after: "digital_image_processing/image_data/lena_small.jpg"
+
+ from os import getenv # Speed up our Continuous Integration tests
+
+ file_name = "lena_small.jpg" if getenv("CI") else "lena.jpg"
+ file_path = f"digital_image_processing/image_data/{file_name}"
+
+ # Reading the image and converting it to grayscale
+ image = imread(file_path, 0)
+
+ # Test for get_neighbors_pixel function() return not None
+ x_coordinate = 0
+ y_coordinate = 0
+ center = image[x_coordinate][y_coordinate]
+
+ neighbors_pixels = lbp.get_neighbors_pixel(
+ image, x_coordinate, y_coordinate, center
+ )
+
+ assert neighbors_pixels is not None
+
+ # Test for local_binary_pattern function()
+ # Create a numpy array as the same height and width of read image
+ lbp_image = np.zeros((image.shape[0], image.shape[1]))
+
+ # Iterating through the image and calculating the local binary pattern value
+ # for each pixel.
+ for i in range(image.shape[0]):
+ for j in range(image.shape[1]):
+ lbp_image[i][j] = lbp.local_binary_value(image, i, j)
+
+ assert lbp_image.any()
diff --git a/divide_and_conquer/closest_pair_of_points.py b/divide_and_conquer/closest_pair_of_points.py
index cb7fa00d1c8f..534cbba9b718 100644
--- a/divide_and_conquer/closest_pair_of_points.py
+++ b/divide_and_conquer/closest_pair_of_points.py
@@ -54,8 +54,7 @@ def dis_between_closest_pair(points, points_counts, min_dis=float("inf")):
for i in range(points_counts - 1):
for j in range(i + 1, points_counts):
current_dis = euclidean_distance_sqr(points[i], points[j])
- if current_dis < min_dis:
- min_dis = current_dis
+ min_dis = min(min_dis, current_dis)
return min_dis
@@ -76,8 +75,7 @@ def dis_between_closest_in_strip(points, points_counts, min_dis=float("inf")):
for i in range(min(6, points_counts - 1), points_counts):
for j in range(max(0, i - 6), i):
current_dis = euclidean_distance_sqr(points[i], points[j])
- if current_dis < min_dis:
- min_dis = current_dis
+ min_dis = min(min_dis, current_dis)
return min_dis
diff --git a/divide_and_conquer/convex_hull.py b/divide_and_conquer/convex_hull.py
index 9c096f671385..93f6daf1f88c 100644
--- a/divide_and_conquer/convex_hull.py
+++ b/divide_and_conquer/convex_hull.py
@@ -13,7 +13,9 @@
"""
-from typing import Iterable, List, Set, Union
+from __future__ import annotations
+
+from collections.abc import Iterable
class Point:
@@ -84,8 +86,8 @@ def __hash__(self):
def _construct_points(
- list_of_tuples: Union[List[Point], List[List[float]], Iterable[List[float]]]
-) -> List[Point]:
+ list_of_tuples: list[Point] | list[list[float]] | Iterable[list[float]],
+) -> list[Point]:
"""
constructs a list of points from an array-like object of numbers
@@ -114,7 +116,7 @@ def _construct_points(
[]
"""
- points: List[Point] = []
+ points: list[Point] = []
if list_of_tuples:
for p in list_of_tuples:
if isinstance(p, Point):
@@ -130,7 +132,7 @@ def _construct_points(
return points
-def _validate_input(points: Union[List[Point], List[List[float]]]) -> List[Point]:
+def _validate_input(points: list[Point] | list[list[float]]) -> list[Point]:
"""
validates an input instance before a convex-hull algorithms uses it
@@ -173,12 +175,12 @@ def _validate_input(points: Union[List[Point], List[List[float]]]) -> List[Point
"""
if not hasattr(points, "__iter__"):
- raise ValueError(
- f"Expecting an iterable object but got an non-iterable type {points}"
- )
+ msg = f"Expecting an iterable object but got an non-iterable type {points}"
+ raise ValueError(msg)
if not points:
- raise ValueError(f"Expecting a list of points but got {points}")
+ msg = f"Expecting a list of points but got {points}"
+ raise ValueError(msg)
return _construct_points(points)
@@ -218,7 +220,7 @@ def _det(a: Point, b: Point, c: Point) -> float:
return det
-def convex_hull_bf(points: List[Point]) -> List[Point]:
+def convex_hull_bf(points: list[Point]) -> list[Point]:
"""
Constructs the convex hull of a set of 2D points using a brute force algorithm.
The algorithm basically considers all combinations of points (i, j) and uses the
@@ -265,21 +267,20 @@ def convex_hull_bf(points: List[Point]) -> List[Point]:
points_left_of_ij = points_right_of_ij = False
ij_part_of_convex_hull = True
for k in range(n):
- if k != i and k != j:
+ if k not in {i, j}:
det_k = _det(points[i], points[j], points[k])
if det_k > 0:
points_left_of_ij = True
elif det_k < 0:
points_right_of_ij = True
- else:
- # point[i], point[j], point[k] all lie on a straight line
- # if point[k] is to the left of point[i] or it's to the
- # right of point[j], then point[i], point[j] cannot be
- # part of the convex hull of A
- if points[k] < points[i] or points[k] > points[j]:
- ij_part_of_convex_hull = False
- break
+ # point[i], point[j], point[k] all lie on a straight line
+ # if point[k] is to the left of point[i] or it's to the
+ # right of point[j], then point[i], point[j] cannot be
+ # part of the convex hull of A
+ elif points[k] < points[i] or points[k] > points[j]:
+ ij_part_of_convex_hull = False
+ break
if points_left_of_ij and points_right_of_ij:
ij_part_of_convex_hull = False
@@ -291,7 +292,7 @@ def convex_hull_bf(points: List[Point]) -> List[Point]:
return sorted(convex_set)
-def convex_hull_recursive(points: List[Point]) -> List[Point]:
+def convex_hull_recursive(points: list[Point]) -> list[Point]:
"""
Constructs the convex hull of a set of 2D points using a divide-and-conquer strategy
The algorithm exploits the geometric properties of the problem by repeatedly
@@ -362,7 +363,7 @@ def convex_hull_recursive(points: List[Point]) -> List[Point]:
def _construct_hull(
- points: List[Point], left: Point, right: Point, convex_set: Set[Point]
+ points: list[Point], left: Point, right: Point, convex_set: set[Point]
) -> None:
"""
@@ -405,7 +406,7 @@ def _construct_hull(
_construct_hull(candidate_points, extreme_point, right, convex_set)
-def convex_hull_melkman(points: List[Point]) -> List[Point]:
+def convex_hull_melkman(points: list[Point]) -> list[Point]:
"""
Constructs the convex hull of a set of 2D points using the melkman algorithm.
The algorithm works by iteratively inserting points of a simple polygonal chain
@@ -457,16 +458,16 @@ def convex_hull_melkman(points: List[Point]) -> List[Point]:
convex_hull[1] = points[i]
i += 1
- for i in range(i, n):
+ for j in range(i, n):
if (
- _det(convex_hull[0], convex_hull[-1], points[i]) > 0
+ _det(convex_hull[0], convex_hull[-1], points[j]) > 0
and _det(convex_hull[-1], convex_hull[0], points[1]) < 0
):
# The point lies within the convex hull
continue
- convex_hull.insert(0, points[i])
- convex_hull.append(points[i])
+ convex_hull.insert(0, points[j])
+ convex_hull.append(points[j])
while _det(convex_hull[0], convex_hull[1], convex_hull[2]) >= 0:
del convex_hull[1]
while _det(convex_hull[-1], convex_hull[-2], convex_hull[-3]) <= 0:
diff --git a/divide_and_conquer/inversions.py b/divide_and_conquer/inversions.py
index 9bb656229321..35f78fe5cf1e 100644
--- a/divide_and_conquer/inversions.py
+++ b/divide_and_conquer/inversions.py
@@ -2,31 +2,25 @@
Given an array-like data structure A[1..n], how many pairs
(i, j) for all 1 <= i < j <= n such that A[i] > A[j]? These pairs are
called inversions. Counting the number of such inversions in an array-like
-object is the important. Among other things, counting inversions can help
-us determine how close a given array is to being sorted
-
+object is the important. Among other things, counting inversions can help
+us determine how close a given array is to being sorted.
In this implementation, I provide two algorithms, a divide-and-conquer
algorithm which runs in nlogn and the brute-force n^2 algorithm.
-
"""
def count_inversions_bf(arr):
"""
- Counts the number of inversions using a a naive brute-force algorithm
-
+ Counts the number of inversions using a naive brute-force algorithm
Parameters
----------
arr: arr: array-like, the list containing the items for which the number
of inversions is desired. The elements of `arr` must be comparable.
-
Returns
-------
num_inversions: The total number of inversions in `arr`
-
Examples
---------
-
>>> count_inversions_bf([1, 4, 2, 4, 1])
4
>>> count_inversions_bf([1, 1, 2, 4, 4])
@@ -49,20 +43,16 @@ def count_inversions_bf(arr):
def count_inversions_recursive(arr):
"""
Counts the number of inversions using a divide-and-conquer algorithm
-
Parameters
-----------
arr: array-like, the list containing the items for which the number
of inversions is desired. The elements of `arr` must be comparable.
-
Returns
-------
C: a sorted copy of `arr`.
num_inversions: int, the total number of inversions in 'arr'
-
Examples
--------
-
>>> count_inversions_recursive([1, 4, 2, 4, 1])
([1, 1, 2, 4, 4], 4)
>>> count_inversions_recursive([1, 1, 2, 4, 4])
@@ -72,66 +62,60 @@ def count_inversions_recursive(arr):
"""
if len(arr) <= 1:
return arr, 0
- else:
- mid = len(arr) // 2
- P = arr[0:mid]
- Q = arr[mid:]
+ mid = len(arr) // 2
+ p = arr[0:mid]
+ q = arr[mid:]
- A, inversion_p = count_inversions_recursive(P)
- B, inversions_q = count_inversions_recursive(Q)
- C, cross_inversions = _count_cross_inversions(A, B)
+ a, inversion_p = count_inversions_recursive(p)
+ b, inversions_q = count_inversions_recursive(q)
+ c, cross_inversions = _count_cross_inversions(a, b)
- num_inversions = inversion_p + inversions_q + cross_inversions
- return C, num_inversions
+ num_inversions = inversion_p + inversions_q + cross_inversions
+ return c, num_inversions
-def _count_cross_inversions(P, Q):
+def _count_cross_inversions(p, q):
"""
Counts the inversions across two sorted arrays.
And combine the two arrays into one sorted array
-
For all 1<= i<=len(P) and for all 1 <= j <= len(Q),
if P[i] > Q[j], then (i, j) is a cross inversion
-
Parameters
----------
P: array-like, sorted in non-decreasing order
Q: array-like, sorted in non-decreasing order
-
Returns
------
R: array-like, a sorted array of the elements of `P` and `Q`
num_inversion: int, the number of inversions across `P` and `Q`
-
Examples
--------
-
>>> _count_cross_inversions([1, 2, 3], [0, 2, 5])
([0, 1, 2, 2, 3, 5], 4)
>>> _count_cross_inversions([1, 2, 3], [3, 4, 5])
([1, 2, 3, 3, 4, 5], 0)
"""
- R = []
+ r = []
i = j = num_inversion = 0
- while i < len(P) and j < len(Q):
- if P[i] > Q[j]:
+ while i < len(p) and j < len(q):
+ if p[i] > q[j]:
# if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P)
# These are all inversions. The claim emerges from the
# property that P is sorted.
- num_inversion += len(P) - i
- R.append(Q[j])
+ num_inversion += len(p) - i
+ r.append(q[j])
j += 1
else:
- R.append(P[i])
+ r.append(p[i])
i += 1
- if i < len(P):
- R.extend(P[i:])
+ if i < len(p):
+ r.extend(p[i:])
else:
- R.extend(Q[j:])
+ r.extend(q[j:])
- return R, num_inversion
+ return r, num_inversion
def main():
diff --git a/divide_and_conquer/kth_order_statistic.py b/divide_and_conquer/kth_order_statistic.py
index f6e81a306bff..23fd8be5ea47 100644
--- a/divide_and_conquer/kth_order_statistic.py
+++ b/divide_and_conquer/kth_order_statistic.py
@@ -8,8 +8,10 @@
For more information of this algorithm:
https://web.stanford.edu/class/archive/cs/cs161/cs161.1138/lectures/08/Small08.pdf
"""
+
+from __future__ import annotations
+
from random import choice
-from typing import List
def random_pivot(lst):
@@ -21,7 +23,7 @@ def random_pivot(lst):
return choice(lst)
-def kth_number(lst: List[int], k: int) -> int:
+def kth_number(lst: list[int], k: int) -> int:
"""
Return the kth smallest number in lst.
>>> kth_number([2, 1, 3, 4, 5], 3)
diff --git a/divide_and_conquer/max_difference_pair.py b/divide_and_conquer/max_difference_pair.py
new file mode 100644
index 000000000000..ffc4b76a7154
--- /dev/null
+++ b/divide_and_conquer/max_difference_pair.py
@@ -0,0 +1,44 @@
+def max_difference(a: list[int]) -> tuple[int, int]:
+ """
+ We are given an array A[1..n] of integers, n >= 1. We want to
+ find a pair of indices (i, j) such that
+ 1 <= i <= j <= n and A[j] - A[i] is as large as possible.
+
+ Explanation:
+ https://www.geeksforgeeks.org/maximum-difference-between-two-elements/
+
+ >>> max_difference([5, 11, 2, 1, 7, 9, 0, 7])
+ (1, 9)
+ """
+ # base case
+ if len(a) == 1:
+ return a[0], a[0]
+ else:
+ # split A into half.
+ first = a[: len(a) // 2]
+ second = a[len(a) // 2 :]
+
+ # 2 sub problems, 1/2 of original size.
+ small1, big1 = max_difference(first)
+ small2, big2 = max_difference(second)
+
+ # get min of first and max of second
+ # linear time
+ min_first = min(first)
+ max_second = max(second)
+
+ # 3 cases, either (small1, big1),
+ # (min_first, max_second), (small2, big2)
+ # constant comparisons
+ if big2 - small2 > max_second - min_first and big2 - small2 > big1 - small1:
+ return small2, big2
+ elif big1 - small1 > max_second - min_first:
+ return small1, big1
+ else:
+ return min_first, max_second
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/divide_and_conquer/max_subarray.py b/divide_and_conquer/max_subarray.py
new file mode 100644
index 000000000000..0fad7ab5d920
--- /dev/null
+++ b/divide_and_conquer/max_subarray.py
@@ -0,0 +1,113 @@
+"""
+The maximum subarray problem is the task of finding the continuous subarray that has the
+maximum sum within a given array of numbers. For example, given the array
+[-2, 1, -3, 4, -1, 2, 1, -5, 4], the contiguous subarray with the maximum sum is
+[4, -1, 2, 1], which has a sum of 6.
+
+This divide-and-conquer algorithm finds the maximum subarray in O(n log n) time.
+"""
+
+from __future__ import annotations
+
+import time
+from collections.abc import Sequence
+from random import randint
+
+from matplotlib import pyplot as plt
+
+
+def max_subarray(
+ arr: Sequence[float], low: int, high: int
+) -> tuple[int | None, int | None, float]:
+ """
+ Solves the maximum subarray problem using divide and conquer.
+ :param arr: the given array of numbers
+ :param low: the start index
+ :param high: the end index
+ :return: the start index of the maximum subarray, the end index of the
+ maximum subarray, and the maximum subarray sum
+
+ >>> nums = [-2, 1, -3, 4, -1, 2, 1, -5, 4]
+ >>> max_subarray(nums, 0, len(nums) - 1)
+ (3, 6, 6)
+ >>> nums = [2, 8, 9]
+ >>> max_subarray(nums, 0, len(nums) - 1)
+ (0, 2, 19)
+ >>> nums = [0, 0]
+ >>> max_subarray(nums, 0, len(nums) - 1)
+ (0, 0, 0)
+ >>> nums = [-1.0, 0.0, 1.0]
+ >>> max_subarray(nums, 0, len(nums) - 1)
+ (2, 2, 1.0)
+ >>> nums = [-2, -3, -1, -4, -6]
+ >>> max_subarray(nums, 0, len(nums) - 1)
+ (2, 2, -1)
+ >>> max_subarray([], 0, 0)
+ (None, None, 0)
+ """
+ if not arr:
+ return None, None, 0
+ if low == high:
+ return low, high, arr[low]
+
+ mid = (low + high) // 2
+ left_low, left_high, left_sum = max_subarray(arr, low, mid)
+ right_low, right_high, right_sum = max_subarray(arr, mid + 1, high)
+ cross_left, cross_right, cross_sum = max_cross_sum(arr, low, mid, high)
+ if left_sum >= right_sum and left_sum >= cross_sum:
+ return left_low, left_high, left_sum
+ elif right_sum >= left_sum and right_sum >= cross_sum:
+ return right_low, right_high, right_sum
+ return cross_left, cross_right, cross_sum
+
+
+def max_cross_sum(
+ arr: Sequence[float], low: int, mid: int, high: int
+) -> tuple[int, int, float]:
+ left_sum, max_left = float("-inf"), -1
+ right_sum, max_right = float("-inf"), -1
+
+ summ: int | float = 0
+ for i in range(mid, low - 1, -1):
+ summ += arr[i]
+ if summ > left_sum:
+ left_sum = summ
+ max_left = i
+
+ summ = 0
+ for i in range(mid + 1, high + 1):
+ summ += arr[i]
+ if summ > right_sum:
+ right_sum = summ
+ max_right = i
+
+ return max_left, max_right, (left_sum + right_sum)
+
+
+def time_max_subarray(input_size: int) -> float:
+ arr = [randint(1, input_size) for _ in range(input_size)]
+ start = time.time()
+ max_subarray(arr, 0, input_size - 1)
+ end = time.time()
+ return end - start
+
+
+def plot_runtimes() -> None:
+ input_sizes = [10, 100, 1000, 10000, 50000, 100000, 200000, 300000, 400000, 500000]
+ runtimes = [time_max_subarray(input_size) for input_size in input_sizes]
+ print("No of Inputs\t\tTime Taken")
+ for input_size, runtime in zip(input_sizes, runtimes):
+ print(input_size, "\t\t", runtime)
+ plt.plot(input_sizes, runtimes)
+ plt.xlabel("Number of Inputs")
+ plt.ylabel("Time taken in seconds")
+ plt.show()
+
+
+if __name__ == "__main__":
+ """
+ A random simulation of this algorithm.
+ """
+ from doctest import testmod
+
+ testmod()
diff --git a/divide_and_conquer/max_subarray_sum.py b/divide_and_conquer/max_subarray_sum.py
deleted file mode 100644
index 43f58086e078..000000000000
--- a/divide_and_conquer/max_subarray_sum.py
+++ /dev/null
@@ -1,76 +0,0 @@
-"""
-Given a array of length n, max_subarray_sum() finds
-the maximum of sum of contiguous sub-array using divide and conquer method.
-
-Time complexity : O(n log n)
-
-Ref : INTRODUCTION TO ALGORITHMS THIRD EDITION
-(section : 4, sub-section : 4.1, page : 70)
-
-"""
-
-
-def max_sum_from_start(array):
- """This function finds the maximum contiguous sum of array from 0 index
-
- Parameters :
- array (list[int]) : given array
-
- Returns :
- max_sum (int) : maximum contiguous sum of array from 0 index
-
- """
- array_sum = 0
- max_sum = float("-inf")
- for num in array:
- array_sum += num
- if array_sum > max_sum:
- max_sum = array_sum
- return max_sum
-
-
-def max_cross_array_sum(array, left, mid, right):
- """This function finds the maximum contiguous sum of left and right arrays
-
- Parameters :
- array, left, mid, right (list[int], int, int, int)
-
- Returns :
- (int) : maximum of sum of contiguous sum of left and right arrays
-
- """
-
- max_sum_of_left = max_sum_from_start(array[left : mid + 1][::-1])
- max_sum_of_right = max_sum_from_start(array[mid + 1 : right + 1])
- return max_sum_of_left + max_sum_of_right
-
-
-def max_subarray_sum(array, left, right):
- """Maximum contiguous sub-array sum, using divide and conquer method
-
- Parameters :
- array, left, right (list[int], int, int) :
- given array, current left index and current right index
-
- Returns :
- int : maximum of sum of contiguous sub-array
-
- """
-
- # base case: array has only one element
- if left == right:
- return array[right]
-
- # Recursion
- mid = (left + right) // 2
- left_half_sum = max_subarray_sum(array, left, mid)
- right_half_sum = max_subarray_sum(array, mid + 1, right)
- cross_sum = max_cross_array_sum(array, left, mid, right)
- return max(left_half_sum, right_half_sum, cross_sum)
-
-
-array = [-2, -5, 6, -2, -3, 1, 5, -6]
-array_length = len(array)
-print(
- "Maximum sum of contiguous subarray:", max_subarray_sum(array, 0, array_length - 1)
-)
diff --git a/divide_and_conquer/mergesort.py b/divide_and_conquer/mergesort.py
index 46a46941cab3..628080cefc9b 100644
--- a/divide_and_conquer/mergesort.py
+++ b/divide_and_conquer/mergesort.py
@@ -1,7 +1,7 @@
-from typing import List
+from __future__ import annotations
-def merge(left_half: List, right_half: List) -> List:
+def merge(left_half: list, right_half: list) -> list:
"""Helper function for mergesort.
>>> left_half = [-2]
@@ -57,7 +57,7 @@ def merge(left_half: List, right_half: List) -> List:
return sorted_array
-def merge_sort(array: List) -> List:
+def merge_sort(array: list) -> list:
"""Returns a list of sorted array elements using merge sort.
>>> from random import shuffle
diff --git a/divide_and_conquer/peak.py b/divide_and_conquer/peak.py
index f94f83ed3fcb..71ab5ac86574 100644
--- a/divide_and_conquer/peak.py
+++ b/divide_and_conquer/peak.py
@@ -7,10 +7,11 @@
(From Kleinberg and Tardos. Algorithm Design.
Addison Wesley 2006: Chapter 5 Solved Exercise 1)
"""
-from typing import List
+from __future__ import annotations
-def peak(lst: List[int]) -> int:
+
+def peak(lst: list[int]) -> int:
"""
Return the peak value of `lst`.
>>> peak([1, 2, 3, 4, 5, 4, 3, 2, 1])
diff --git a/divide_and_conquer/power.py b/divide_and_conquer/power.py
index f2e023afd536..492ee6dd12f0 100644
--- a/divide_and_conquer/power.py
+++ b/divide_and_conquer/power.py
@@ -1,18 +1,38 @@
-def actual_power(a: int, b: int):
+def actual_power(a: int, b: int) -> int:
"""
Function using divide and conquer to calculate a^b.
It only works for integer a,b.
+
+ :param a: The base of the power operation, an integer.
+ :param b: The exponent of the power operation, a non-negative integer.
+ :return: The result of a^b.
+
+ Examples:
+ >>> actual_power(3, 2)
+ 9
+ >>> actual_power(5, 3)
+ 125
+ >>> actual_power(2, 5)
+ 32
+ >>> actual_power(7, 0)
+ 1
"""
if b == 0:
return 1
+ half = actual_power(a, b // 2)
+
if (b % 2) == 0:
- return actual_power(a, int(b / 2)) * actual_power(a, int(b / 2))
+ return half * half
else:
- return a * actual_power(a, int(b / 2)) * actual_power(a, int(b / 2))
+ return a * half * half
def power(a: int, b: int) -> float:
"""
+ :param a: The base (integer).
+ :param b: The exponent (integer).
+ :return: The result of a^b, as a float for negative exponents.
+
>>> power(4,6)
4096
>>> power(2,3)
@@ -25,9 +45,9 @@ def power(a: int, b: int) -> float:
-0.125
"""
if b < 0:
- return 1 / actual_power(a, b)
+ return 1 / actual_power(a, -b)
return actual_power(a, b)
if __name__ == "__main__":
- print(power(-2, -3))
+ print(power(-2, -3)) # output -0.125
diff --git a/divide_and_conquer/strassen_matrix_multiplication.py b/divide_and_conquer/strassen_matrix_multiplication.py
index 29a174daebf9..f529a255d2ef 100644
--- a/divide_and_conquer/strassen_matrix_multiplication.py
+++ b/divide_and_conquer/strassen_matrix_multiplication.py
@@ -68,14 +68,13 @@ def matrix_dimensions(matrix: list) -> tuple[int, int]:
def print_matrix(matrix: list) -> None:
- for i in range(len(matrix)):
- print(matrix[i])
+ print("\n".join(str(line) for line in matrix))
def actual_strassen(matrix_a: list, matrix_b: list) -> list:
"""
Recursive function to calculate the product of two matrices, using the Strassen
- Algorithm. It only supports even length matrices.
+ Algorithm. It only supports square matrices of any size that is a power of 2.
"""
if matrix_dimensions(matrix_a) == (2, 2):
return default_matrix_multiplication(matrix_a, matrix_b)
@@ -113,31 +112,33 @@ def strassen(matrix1: list, matrix2: list) -> list:
[[139, 163], [121, 134], [100, 121]]
"""
if matrix_dimensions(matrix1)[1] != matrix_dimensions(matrix2)[0]:
- raise Exception(
- f"Unable to multiply these matrices, please check the dimensions. \n"
- f"Matrix A:{matrix1} \nMatrix B:{matrix2}"
+ msg = (
+ "Unable to multiply these matrices, please check the dimensions.\n"
+ f"Matrix A: {matrix1}\n"
+ f"Matrix B: {matrix2}"
)
+ raise Exception(msg)
dimension1 = matrix_dimensions(matrix1)
dimension2 = matrix_dimensions(matrix2)
if dimension1[0] == dimension1[1] and dimension2[0] == dimension2[1]:
- return matrix1, matrix2
+ return [matrix1, matrix2]
- maximum = max(max(dimension1), max(dimension2))
+ maximum = max(*dimension1, *dimension2)
maxim = int(math.pow(2, math.ceil(math.log2(maximum))))
new_matrix1 = matrix1
new_matrix2 = matrix2
- # Adding zeros to the matrices so that the arrays dimensions are the same and also
- # power of 2
- for i in range(0, maxim):
+ # Adding zeros to the matrices to convert them both into square matrices of equal
+ # dimensions that are a power of 2
+ for i in range(maxim):
if i < dimension1[0]:
- for j in range(dimension1[1], maxim):
+ for _ in range(dimension1[1], maxim):
new_matrix1[i].append(0)
else:
new_matrix1.append([0] * maxim)
if i < dimension2[0]:
- for j in range(dimension2[1], maxim):
+ for _ in range(dimension2[1], maxim):
new_matrix2[i].append(0)
else:
new_matrix2.append([0] * maxim)
@@ -145,9 +146,9 @@ def strassen(matrix1: list, matrix2: list) -> list:
final_matrix = actual_strassen(new_matrix1, new_matrix2)
# Removing the additional zeros
- for i in range(0, maxim):
+ for i in range(maxim):
if i < dimension1[0]:
- for j in range(dimension2[1], maxim):
+ for _ in range(dimension2[1], maxim):
final_matrix[i].pop()
else:
final_matrix.pop()
diff --git a/docs/__init__.py b/docs/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/docs/conf.py b/docs/conf.py
new file mode 100644
index 000000000000..f2481f107267
--- /dev/null
+++ b/docs/conf.py
@@ -0,0 +1,3 @@
+from sphinx_pyproject import SphinxConfig
+
+project = SphinxConfig("../pyproject.toml", globalns=globals()).name
diff --git a/docs/source/__init__.py b/docs/source/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/dynamic_programming/all_construct.py b/dynamic_programming/all_construct.py
new file mode 100644
index 000000000000..ca00f2beb06a
--- /dev/null
+++ b/dynamic_programming/all_construct.py
@@ -0,0 +1,60 @@
+"""
+Program to list all the ways a target string can be
+constructed from the given list of substrings
+"""
+
+from __future__ import annotations
+
+
+def all_construct(target: str, word_bank: list[str] | None = None) -> list[list[str]]:
+ """
+ returns the list containing all the possible
+ combinations a string(`target`) can be constructed from
+ the given list of substrings(`word_bank`)
+
+ >>> all_construct("hello", ["he", "l", "o"])
+ [['he', 'l', 'l', 'o']]
+ >>> all_construct("purple",["purp","p","ur","le","purpl"])
+ [['purp', 'le'], ['p', 'ur', 'p', 'le']]
+ """
+
+ word_bank = word_bank or []
+ # create a table
+ table_size: int = len(target) + 1
+
+ table: list[list[list[str]]] = []
+ for _ in range(table_size):
+ table.append([])
+ # seed value
+ table[0] = [[]] # because empty string has empty combination
+
+ # iterate through the indices
+ for i in range(table_size):
+ # condition
+ if table[i] != []:
+ for word in word_bank:
+ # slice condition
+ if target[i : i + len(word)] == word:
+ new_combinations: list[list[str]] = [
+ [word, *way] for way in table[i]
+ ]
+ # adds the word to every combination the current position holds
+ # now,push that combination to the table[i+len(word)]
+ table[i + len(word)] += new_combinations
+
+ # combinations are in reverse order so reverse for better output
+ for combination in table[len(target)]:
+ combination.reverse()
+
+ return table[len(target)]
+
+
+if __name__ == "__main__":
+ print(all_construct("jwajalapa", ["jwa", "j", "w", "a", "la", "lapa"]))
+ print(all_construct("rajamati", ["s", "raj", "amat", "raja", "ma", "i", "t"]))
+ print(
+ all_construct(
+ "hexagonosaurus",
+ ["h", "ex", "hex", "ag", "ago", "ru", "auru", "rus", "go", "no", "o", "s"],
+ )
+ )
diff --git a/dynamic_programming/bitmask.py b/dynamic_programming/bitmask.py
index 2994db5b5e1e..a6e6a0cda7bf 100644
--- a/dynamic_programming/bitmask.py
+++ b/dynamic_programming/bitmask.py
@@ -8,12 +8,12 @@
a person can do only one task and a task is performed only by one person.
Find the total no of ways in which the tasks can be distributed.
"""
+
from collections import defaultdict
class AssignmentUsingBitmask:
def __init__(self, task_performed, total):
-
self.total_tasks = total # total no of tasks (N)
# DP table will have a dimension of (2^M)*N
@@ -28,8 +28,7 @@ def __init__(self, task_performed, total):
# to 1
self.final_mask = (1 << len(task_performed)) - 1
- def CountWaysUtil(self, mask, task_no):
-
+ def count_ways_until(self, mask, task_no):
# if mask == self.finalmask all persons are distributed tasks, return 1
if mask == self.final_mask:
return 1
@@ -43,45 +42,42 @@ def CountWaysUtil(self, mask, task_no):
return self.dp[mask][task_no]
# Number of ways when we don't this task in the arrangement
- total_ways_util = self.CountWaysUtil(mask, task_no + 1)
+ total_ways_util = self.count_ways_until(mask, task_no + 1)
# now assign the tasks one by one to all possible persons and recursively
# assign for the remaining tasks.
if task_no in self.task:
for p in self.task[task_no]:
-
# if p is already given a task
if mask & (1 << p):
continue
# assign this task to p and change the mask value. And recursively
# assign tasks with the new mask value.
- total_ways_util += self.CountWaysUtil(mask | (1 << p), task_no + 1)
+ total_ways_util += self.count_ways_until(mask | (1 << p), task_no + 1)
# save the value.
self.dp[mask][task_no] = total_ways_util
return self.dp[mask][task_no]
- def countNoOfWays(self, task_performed):
-
+ def count_no_of_ways(self, task_performed):
# Store the list of persons for each task
for i in range(len(task_performed)):
for j in task_performed[i]:
self.task[j].append(i)
# call the function to fill the DP table, final answer is stored in dp[0][1]
- return self.CountWaysUtil(0, 1)
+ return self.count_ways_until(0, 1)
if __name__ == "__main__":
-
total_tasks = 5 # total no of tasks (the value of N)
# the list of tasks that can be done by M persons.
task_performed = [[1, 3, 4], [1, 2, 5], [3, 4]]
print(
- AssignmentUsingBitmask(task_performed, total_tasks).countNoOfWays(
+ AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways(
task_performed
)
)
diff --git a/dynamic_programming/catalan_numbers.py b/dynamic_programming/catalan_numbers.py
new file mode 100644
index 000000000000..7b74f2763d43
--- /dev/null
+++ b/dynamic_programming/catalan_numbers.py
@@ -0,0 +1,79 @@
+"""
+Print all the Catalan numbers from 0 to n, n being the user input.
+
+ * The Catalan numbers are a sequence of positive integers that
+ * appear in many counting problems in combinatorics [1]. Such
+ * problems include counting [2]:
+ * - The number of Dyck words of length 2n
+ * - The number well-formed expressions with n pairs of parentheses
+ * (e.g., `()()` is valid but `())(` is not)
+ * - The number of different ways n + 1 factors can be completely
+ * parenthesized (e.g., for n = 2, C(n) = 2 and (ab)c and a(bc)
+ * are the two valid ways to parenthesize.
+ * - The number of full binary trees with n + 1 leaves
+
+ * A Catalan number satisfies the following recurrence relation
+ * which we will use in this algorithm [1].
+ * C(0) = C(1) = 1
+ * C(n) = sum(C(i).C(n-i-1)), from i = 0 to n-1
+
+ * In addition, the n-th Catalan number can be calculated using
+ * the closed form formula below [1]:
+ * C(n) = (1 / (n + 1)) * (2n choose n)
+
+ * Sources:
+ * [1] https://brilliant.org/wiki/catalan-numbers/
+ * [2] https://en.wikipedia.org/wiki/Catalan_number
+"""
+
+
+def catalan_numbers(upper_limit: int) -> "list[int]":
+ """
+ Return a list of the Catalan number sequence from 0 through `upper_limit`.
+
+ >>> catalan_numbers(5)
+ [1, 1, 2, 5, 14, 42]
+ >>> catalan_numbers(2)
+ [1, 1, 2]
+ >>> catalan_numbers(-1)
+ Traceback (most recent call last):
+ ValueError: Limit for the Catalan sequence must be ≥ 0
+ """
+ if upper_limit < 0:
+ raise ValueError("Limit for the Catalan sequence must be ≥ 0")
+
+ catalan_list = [0] * (upper_limit + 1)
+
+ # Base case: C(0) = C(1) = 1
+ catalan_list[0] = 1
+ if upper_limit > 0:
+ catalan_list[1] = 1
+
+ # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i
+ for i in range(2, upper_limit + 1):
+ for j in range(i):
+ catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1]
+
+ return catalan_list
+
+
+if __name__ == "__main__":
+ print("\n********* Catalan Numbers Using Dynamic Programming ************\n")
+ print("\n*** Enter -1 at any time to quit ***")
+ print("\nEnter the upper limit (≥ 0) for the Catalan number sequence: ", end="")
+ try:
+ while True:
+ N = int(input().strip())
+ if N < 0:
+ print("\n********* Goodbye!! ************")
+ break
+ else:
+ print(f"The Catalan numbers from 0 through {N} are:")
+ print(catalan_numbers(N))
+ print("Try another upper limit for the sequence: ", end="")
+ except (NameError, ValueError):
+ print("\n********* Invalid input, goodbye! ************\n")
+
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/climbing_stairs.py b/dynamic_programming/climbing_stairs.py
index 79605261f981..38bdb427eedc 100644
--- a/dynamic_programming/climbing_stairs.py
+++ b/dynamic_programming/climbing_stairs.py
@@ -1,20 +1,20 @@
#!/usr/bin/env python3
-def climb_stairs(n: int) -> int:
+def climb_stairs(number_of_steps: int) -> int:
"""
LeetCdoe No.70: Climbing Stairs
- Distinct ways to climb a n step staircase where
- each time you can either climb 1 or 2 steps.
+ Distinct ways to climb a number_of_steps staircase where each time you can either
+ climb 1 or 2 steps.
Args:
- n: number of steps of staircase
+ number_of_steps: number of steps on the staircase
Returns:
- Distinct ways to climb a n step staircase
+ Distinct ways to climb a number_of_steps staircase
Raises:
- AssertionError: n not positive integer
+ AssertionError: number_of_steps not positive integer
>>> climb_stairs(3)
3
@@ -23,17 +23,17 @@ def climb_stairs(n: int) -> int:
>>> climb_stairs(-7) # doctest: +ELLIPSIS
Traceback (most recent call last):
...
- AssertionError: n needs to be positive integer, your input -7
+ AssertionError: number_of_steps needs to be positive integer, your input -7
"""
- fmt = "n needs to be positive integer, your input {}"
- assert isinstance(n, int) and n > 0, fmt.format(n)
- if n == 1:
+ assert isinstance(number_of_steps, int) and number_of_steps > 0, (
+ f"number_of_steps needs to be positive integer, your input {number_of_steps}"
+ )
+ if number_of_steps == 1:
return 1
- dp = [0] * (n + 1)
- dp[0], dp[1] = (1, 1)
- for i in range(2, n + 1):
- dp[i] = dp[i - 1] + dp[i - 2]
- return dp[n]
+ previous, current = 1, 1
+ for _ in range(number_of_steps - 1):
+ current, previous = current + previous, current
+ return current
if __name__ == "__main__":
diff --git a/dynamic_programming/combination_sum_iv.py b/dynamic_programming/combination_sum_iv.py
new file mode 100644
index 000000000000..ed8dcd88e6fd
--- /dev/null
+++ b/dynamic_programming/combination_sum_iv.py
@@ -0,0 +1,102 @@
+"""
+Question:
+ You are given an array of distinct integers and you have to tell how many
+ different ways of selecting the elements from the array are there such that
+ the sum of chosen elements is equal to the target number tar.
+
+Example
+
+Input:
+ * N = 3
+ * target = 5
+ * array = [1, 2, 5]
+
+Output:
+ 9
+
+Approach:
+ The basic idea is to go over recursively to find the way such that the sum
+ of chosen elements is `target`. For every element, we have two choices
+
+ 1. Include the element in our set of chosen elements.
+ 2. Don't include the element in our set of chosen elements.
+"""
+
+
+def combination_sum_iv(array: list[int], target: int) -> int:
+ """
+ Function checks the all possible combinations, and returns the count
+ of possible combination in exponential Time Complexity.
+
+ >>> combination_sum_iv([1,2,5], 5)
+ 9
+ """
+
+ def count_of_possible_combinations(target: int) -> int:
+ if target < 0:
+ return 0
+ if target == 0:
+ return 1
+ return sum(count_of_possible_combinations(target - item) for item in array)
+
+ return count_of_possible_combinations(target)
+
+
+def combination_sum_iv_dp_array(array: list[int], target: int) -> int:
+ """
+ Function checks the all possible combinations, and returns the count
+ of possible combination in O(N^2) Time Complexity as we are using Dynamic
+ programming array here.
+
+ >>> combination_sum_iv_dp_array([1,2,5], 5)
+ 9
+ """
+
+ def count_of_possible_combinations_with_dp_array(
+ target: int, dp_array: list[int]
+ ) -> int:
+ if target < 0:
+ return 0
+ if target == 0:
+ return 1
+ if dp_array[target] != -1:
+ return dp_array[target]
+ answer = sum(
+ count_of_possible_combinations_with_dp_array(target - item, dp_array)
+ for item in array
+ )
+ dp_array[target] = answer
+ return answer
+
+ dp_array = [-1] * (target + 1)
+ return count_of_possible_combinations_with_dp_array(target, dp_array)
+
+
+def combination_sum_iv_bottom_up(n: int, array: list[int], target: int) -> int:
+ """
+ Function checks the all possible combinations with using bottom up approach,
+ and returns the count of possible combination in O(N^2) Time Complexity
+ as we are using Dynamic programming array here.
+
+ >>> combination_sum_iv_bottom_up(3, [1,2,5], 5)
+ 9
+ """
+
+ dp_array = [0] * (target + 1)
+ dp_array[0] = 1
+
+ for i in range(1, target + 1):
+ for j in range(n):
+ if i - array[j] >= 0:
+ dp_array[i] += dp_array[i - array[j]]
+
+ return dp_array[target]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ target = 5
+ array = [1, 2, 5]
+ print(combination_sum_iv(array, target))
diff --git a/dynamic_programming/edit_distance.py b/dynamic_programming/edit_distance.py
index 56877e0c50a2..774aa047326e 100644
--- a/dynamic_programming/edit_distance.py
+++ b/dynamic_programming/edit_distance.py
@@ -19,74 +19,72 @@ class EditDistance:
"""
def __init__(self):
- self.__prepare__()
-
- def __prepare__(self, N=0, M=0):
- self.dp = [[-1 for y in range(0, M)] for x in range(0, N)]
-
- def __solveDP(self, x, y):
- if x == -1:
- return y + 1
- elif y == -1:
- return x + 1
- elif self.dp[x][y] > -1:
- return self.dp[x][y]
+ self.word1 = ""
+ self.word2 = ""
+ self.dp = []
+
+ def __min_dist_top_down_dp(self, m: int, n: int) -> int:
+ if m == -1:
+ return n + 1
+ elif n == -1:
+ return m + 1
+ elif self.dp[m][n] > -1:
+ return self.dp[m][n]
else:
- if self.A[x] == self.B[y]:
- self.dp[x][y] = self.__solveDP(x - 1, y - 1)
+ if self.word1[m] == self.word2[n]:
+ self.dp[m][n] = self.__min_dist_top_down_dp(m - 1, n - 1)
else:
- self.dp[x][y] = 1 + min(
- self.__solveDP(x, y - 1),
- self.__solveDP(x - 1, y),
- self.__solveDP(x - 1, y - 1),
- )
-
- return self.dp[x][y]
-
- def solve(self, A, B):
- if isinstance(A, bytes):
- A = A.decode("ascii")
-
- if isinstance(B, bytes):
- B = B.decode("ascii")
-
- self.A = str(A)
- self.B = str(B)
-
- self.__prepare__(len(A), len(B))
-
- return self.__solveDP(len(A) - 1, len(B) - 1)
-
-
-def min_distance_bottom_up(word1: str, word2: str) -> int:
- """
- >>> min_distance_bottom_up("intention", "execution")
- 5
- >>> min_distance_bottom_up("intention", "")
- 9
- >>> min_distance_bottom_up("", "")
- 0
- """
- m = len(word1)
- n = len(word2)
- dp = [[0 for _ in range(n + 1)] for _ in range(m + 1)]
- for i in range(m + 1):
- for j in range(n + 1):
-
- if i == 0: # first string is empty
- dp[i][j] = j
- elif j == 0: # second string is empty
- dp[i][j] = i
- elif (
- word1[i - 1] == word2[j - 1]
- ): # last character of both substing is equal
- dp[i][j] = dp[i - 1][j - 1]
- else:
- insert = dp[i][j - 1]
- delete = dp[i - 1][j]
- replace = dp[i - 1][j - 1]
- dp[i][j] = 1 + min(insert, delete, replace)
- return dp[m][n]
+ insert = self.__min_dist_top_down_dp(m, n - 1)
+ delete = self.__min_dist_top_down_dp(m - 1, n)
+ replace = self.__min_dist_top_down_dp(m - 1, n - 1)
+ self.dp[m][n] = 1 + min(insert, delete, replace)
+
+ return self.dp[m][n]
+
+ def min_dist_top_down(self, word1: str, word2: str) -> int:
+ """
+ >>> EditDistance().min_dist_top_down("intention", "execution")
+ 5
+ >>> EditDistance().min_dist_top_down("intention", "")
+ 9
+ >>> EditDistance().min_dist_top_down("", "")
+ 0
+ """
+ self.word1 = word1
+ self.word2 = word2
+ self.dp = [[-1 for _ in range(len(word2))] for _ in range(len(word1))]
+
+ return self.__min_dist_top_down_dp(len(word1) - 1, len(word2) - 1)
+
+ def min_dist_bottom_up(self, word1: str, word2: str) -> int:
+ """
+ >>> EditDistance().min_dist_bottom_up("intention", "execution")
+ 5
+ >>> EditDistance().min_dist_bottom_up("intention", "")
+ 9
+ >>> EditDistance().min_dist_bottom_up("", "")
+ 0
+ """
+ self.word1 = word1
+ self.word2 = word2
+ m = len(word1)
+ n = len(word2)
+ self.dp = [[0 for _ in range(n + 1)] for _ in range(m + 1)]
+
+ for i in range(m + 1):
+ for j in range(n + 1):
+ if i == 0: # first string is empty
+ self.dp[i][j] = j
+ elif j == 0: # second string is empty
+ self.dp[i][j] = i
+ elif word1[i - 1] == word2[j - 1]: # last characters are equal
+ self.dp[i][j] = self.dp[i - 1][j - 1]
+ else:
+ insert = self.dp[i][j - 1]
+ delete = self.dp[i - 1][j]
+ replace = self.dp[i - 1][j - 1]
+ self.dp[i][j] = 1 + min(insert, delete, replace)
+ return self.dp[m][n]
if __name__ == "__main__":
@@ -99,7 +97,7 @@ def min_distance_bottom_up(word1: str, word2: str) -> int:
S2 = input("Enter the second string: ").strip()
print()
- print("The minimum Edit Distance is: %d" % (solver.solve(S1, S2)))
- print("The minimum Edit Distance is: %d" % (min_distance_bottom_up(S1, S2)))
+ print(f"The minimum edit distance is: {solver.min_dist_top_down(S1, S2)}")
+ print(f"The minimum edit distance is: {solver.min_dist_bottom_up(S1, S2)}")
print()
print("*************** End of Testing Edit Distance DP Algorithm ***************")
diff --git a/dynamic_programming/fast_fibonacci.py b/dynamic_programming/fast_fibonacci.py
index f48186a34c25..d04a5ac8249b 100644
--- a/dynamic_programming/fast_fibonacci.py
+++ b/dynamic_programming/fast_fibonacci.py
@@ -4,6 +4,7 @@
This program calculates the nth Fibonacci number in O(log(n)).
It's possible to calculate F(1_000_000) in less than a second.
"""
+
from __future__ import annotations
import sys
@@ -25,7 +26,7 @@ def _fib(n: int) -> tuple[int, int]:
if n == 0: # (F(0), F(1))
return (0, 1)
- # F(2n) = F(n)[2F(n+1) − F(n)]
+ # F(2n) = F(n)[2F(n+1) - F(n)]
# F(2n+1) = F(n+1)^2+F(n)^2
a, b = _fib(n // 2)
c = a * (b * 2 - a)
diff --git a/dynamic_programming/fibonacci.py b/dynamic_programming/fibonacci.py
index cab1358ddea1..c102493aa00b 100644
--- a/dynamic_programming/fibonacci.py
+++ b/dynamic_programming/fibonacci.py
@@ -5,61 +5,47 @@
class Fibonacci:
- def __init__(self, N=None):
- self.fib_array = []
- if N:
- N = int(N)
- self.fib_array.append(0)
- self.fib_array.append(1)
- for i in range(2, N + 1):
- self.fib_array.append(self.fib_array[i - 1] + self.fib_array[i - 2])
- elif N == 0:
- self.fib_array.append(0)
- print(self.fib_array)
+ def __init__(self) -> None:
+ self.sequence = [0, 1]
- def get(self, sequence_no=None):
+ def get(self, index: int) -> list:
"""
- >>> Fibonacci(5).get(3)
- [0, 1, 1, 2, 3, 5]
- [0, 1, 1, 2]
- >>> Fibonacci(5).get(6)
- [0, 1, 1, 2, 3, 5]
- Out of bound.
- >>> Fibonacci(5).get(-1)
- [0, 1, 1, 2, 3, 5]
- []
+ Get the Fibonacci number of `index`. If the number does not exist,
+ calculate all missing numbers leading up to the number of `index`.
+
+ >>> Fibonacci().get(10)
+ [0, 1, 1, 2, 3, 5, 8, 13, 21, 34]
+ >>> Fibonacci().get(5)
+ [0, 1, 1, 2, 3]
"""
- if sequence_no is not None:
- if sequence_no < len(self.fib_array):
- return print(self.fib_array[: sequence_no + 1])
- else:
- print("Out of bound.")
- else:
- print("Please specify a value")
+ if (difference := index - (len(self.sequence) - 2)) >= 1:
+ for _ in range(difference):
+ self.sequence.append(self.sequence[-1] + self.sequence[-2])
+ return self.sequence[:index]
-if __name__ == "__main__":
- print("\n********* Fibonacci Series Using Dynamic Programming ************\n")
- print("\n Enter the upper limit for the fibonacci sequence: ", end="")
- try:
- N = int(input().strip())
- fib = Fibonacci(N)
- print(
- "\n********* Enter different values to get the corresponding fibonacci "
- "sequence, enter any negative number to exit. ************\n"
- )
- while True:
- try:
- i = int(input("Enter value: ").strip())
- if i < 0:
- print("\n********* Good Bye!! ************\n")
- break
- fib.get(i)
- except NameError:
- print("\nInvalid input, please try again.")
- except NameError:
- print("\n********* Invalid input, good bye!! ************\n")
+def main() -> None:
+ print(
+ "Fibonacci Series Using Dynamic Programming\n",
+ "Enter the index of the Fibonacci number you want to calculate ",
+ "in the prompt below. (To exit enter exit or Ctrl-C)\n",
+ sep="",
+ )
+ fibonacci = Fibonacci()
+
+ while True:
+ prompt: str = input(">> ")
+ if prompt in {"exit", "quit"}:
+ break
- import doctest
+ try:
+ index: int = int(prompt)
+ except ValueError:
+ print("Enter a number or 'exit'")
+ continue
- doctest.testmod()
+ print(fibonacci.get(index))
+
+
+if __name__ == "__main__":
+ main()
diff --git a/dynamic_programming/fizz_buzz.py b/dynamic_programming/fizz_buzz.py
new file mode 100644
index 000000000000..0cb48897875b
--- /dev/null
+++ b/dynamic_programming/fizz_buzz.py
@@ -0,0 +1,65 @@
+# https://en.wikipedia.org/wiki/Fizz_buzz#Programming
+
+
+def fizz_buzz(number: int, iterations: int) -> str:
+ """
+ | Plays FizzBuzz.
+ | Prints Fizz if number is a multiple of ``3``.
+ | Prints Buzz if its a multiple of ``5``.
+ | Prints FizzBuzz if its a multiple of both ``3`` and ``5`` or ``15``.
+ | Else Prints The Number Itself.
+
+ >>> fizz_buzz(1,7)
+ '1 2 Fizz 4 Buzz Fizz 7 '
+ >>> fizz_buzz(1,0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Iterations must be done more than 0 times to play FizzBuzz
+ >>> fizz_buzz(-5,5)
+ Traceback (most recent call last):
+ ...
+ ValueError: starting number must be
+ and integer and be more than 0
+ >>> fizz_buzz(10,-5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Iterations must be done more than 0 times to play FizzBuzz
+ >>> fizz_buzz(1.5,5)
+ Traceback (most recent call last):
+ ...
+ ValueError: starting number must be
+ and integer and be more than 0
+ >>> fizz_buzz(1,5.5)
+ Traceback (most recent call last):
+ ...
+ ValueError: iterations must be defined as integers
+ """
+ if not isinstance(iterations, int):
+ raise ValueError("iterations must be defined as integers")
+ if not isinstance(number, int) or not number >= 1:
+ raise ValueError(
+ """starting number must be
+ and integer and be more than 0"""
+ )
+ if not iterations >= 1:
+ raise ValueError("Iterations must be done more than 0 times to play FizzBuzz")
+
+ out = ""
+ while number <= iterations:
+ if number % 3 == 0:
+ out += "Fizz"
+ if number % 5 == 0:
+ out += "Buzz"
+ if 0 not in (number % 3, number % 5):
+ out += str(number)
+
+ # print(out)
+ number += 1
+ out += " "
+ return out
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/floyd_warshall.py b/dynamic_programming/floyd_warshall.py
index a4b6c6a82568..b92c6667fb5c 100644
--- a/dynamic_programming/floyd_warshall.py
+++ b/dynamic_programming/floyd_warshall.py
@@ -2,41 +2,84 @@
class Graph:
- def __init__(self, N=0): # a graph with Node 0,1,...,N-1
- self.N = N
- self.W = [
- [math.inf for j in range(0, N)] for i in range(0, N)
+ def __init__(self, n=0): # a graph with Node 0,1,...,N-1
+ self.n = n
+ self.w = [
+ [math.inf for j in range(n)] for i in range(n)
] # adjacency matrix for weight
self.dp = [
- [math.inf for j in range(0, N)] for i in range(0, N)
+ [math.inf for j in range(n)] for i in range(n)
] # dp[i][j] stores minimum distance from i to j
- def addEdge(self, u, v, w):
+ def add_edge(self, u, v, w):
+ """
+ Adds a directed edge from node u
+ to node v with weight w.
+
+ >>> g = Graph(3)
+ >>> g.add_edge(0, 1, 5)
+ >>> g.dp[0][1]
+ 5
+ """
self.dp[u][v] = w
def floyd_warshall(self):
- for k in range(0, self.N):
- for i in range(0, self.N):
- for j in range(0, self.N):
+ """
+ Computes the shortest paths between all pairs of
+ nodes using the Floyd-Warshall algorithm.
+
+ >>> g = Graph(3)
+ >>> g.add_edge(0, 1, 1)
+ >>> g.add_edge(1, 2, 2)
+ >>> g.floyd_warshall()
+ >>> g.show_min(0, 2)
+ 3
+ >>> g.show_min(2, 0)
+ inf
+ """
+ for k in range(self.n):
+ for i in range(self.n):
+ for j in range(self.n):
self.dp[i][j] = min(self.dp[i][j], self.dp[i][k] + self.dp[k][j])
- def showMin(self, u, v):
+ def show_min(self, u, v):
+ """
+ Returns the minimum distance from node u to node v.
+
+ >>> g = Graph(3)
+ >>> g.add_edge(0, 1, 3)
+ >>> g.add_edge(1, 2, 4)
+ >>> g.floyd_warshall()
+ >>> g.show_min(0, 2)
+ 7
+ >>> g.show_min(1, 0)
+ inf
+ """
return self.dp[u][v]
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ # Example usage
graph = Graph(5)
- graph.addEdge(0, 2, 9)
- graph.addEdge(0, 4, 10)
- graph.addEdge(1, 3, 5)
- graph.addEdge(2, 3, 7)
- graph.addEdge(3, 0, 10)
- graph.addEdge(3, 1, 2)
- graph.addEdge(3, 2, 1)
- graph.addEdge(3, 4, 6)
- graph.addEdge(4, 1, 3)
- graph.addEdge(4, 2, 4)
- graph.addEdge(4, 3, 9)
+ graph.add_edge(0, 2, 9)
+ graph.add_edge(0, 4, 10)
+ graph.add_edge(1, 3, 5)
+ graph.add_edge(2, 3, 7)
+ graph.add_edge(3, 0, 10)
+ graph.add_edge(3, 1, 2)
+ graph.add_edge(3, 2, 1)
+ graph.add_edge(3, 4, 6)
+ graph.add_edge(4, 1, 3)
+ graph.add_edge(4, 2, 4)
+ graph.add_edge(4, 3, 9)
graph.floyd_warshall()
- graph.showMin(1, 4)
- graph.showMin(0, 3)
+ print(
+ graph.show_min(1, 4)
+ ) # Should output the minimum distance from node 1 to node 4
+ print(
+ graph.show_min(0, 3)
+ ) # Should output the minimum distance from node 0 to node 3
diff --git a/dynamic_programming/fractional_knapsack.py b/dynamic_programming/fractional_knapsack.py
deleted file mode 100644
index c74af7ef8fc5..000000000000
--- a/dynamic_programming/fractional_knapsack.py
+++ /dev/null
@@ -1,27 +0,0 @@
-from bisect import bisect
-from itertools import accumulate
-
-
-def fracKnapsack(vl, wt, W, n):
- """
- >>> fracKnapsack([60, 100, 120], [10, 20, 30], 50, 3)
- 240.0
- """
-
- r = list(sorted(zip(vl, wt), key=lambda x: x[0] / x[1], reverse=True))
- vl, wt = [i[0] for i in r], [i[1] for i in r]
- acc = list(accumulate(wt))
- k = bisect(acc, W)
- return (
- 0
- if k == 0
- else sum(vl[:k]) + (W - acc[k - 1]) * (vl[k]) / (wt[k])
- if k != n
- else sum(vl[:k])
- )
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/dynamic_programming/integer_partition.py b/dynamic_programming/integer_partition.py
index 4eb06348ce84..145bc29d0fca 100644
--- a/dynamic_programming/integer_partition.py
+++ b/dynamic_programming/integer_partition.py
@@ -3,11 +3,35 @@
partitions into exactly k parts plus the number of partitions into at least k-1 parts.
Subtracting 1 from each part of a partition of n into k parts gives a partition of n-k
into k parts. These two facts together are used for this algorithm.
+* https://en.wikipedia.org/wiki/Partition_(number_theory)
+* https://en.wikipedia.org/wiki/Partition_function_(number_theory)
"""
-def partition(m):
- memo = [[0 for _ in range(m)] for _ in range(m + 1)]
+def partition(m: int) -> int:
+ """
+ >>> partition(5)
+ 7
+ >>> partition(7)
+ 15
+ >>> partition(100)
+ 190569292
+ >>> partition(1_000)
+ 24061467864032622473692149727991
+ >>> partition(-7)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> partition(0)
+ Traceback (most recent call last):
+ ...
+ IndexError: list assignment index out of range
+ >>> partition(7.8)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'float' object cannot be interpreted as an integer
+ """
+ memo: list[list[int]] = [[0 for _ in range(m)] for _ in range(m + 1)]
for i in range(m + 1):
memo[i][0] = 1
diff --git a/dynamic_programming/iterating_through_submasks.py b/dynamic_programming/iterating_through_submasks.py
index 855af61d6707..efab6dacff3f 100644
--- a/dynamic_programming/iterating_through_submasks.py
+++ b/dynamic_programming/iterating_through_submasks.py
@@ -5,11 +5,11 @@
its submasks. The mask s is submask of m if only bits that were included in
bitmask are set
"""
+
from __future__ import annotations
def list_of_submasks(mask: int) -> list[int]:
-
"""
Args:
mask : number which shows mask ( always integer > 0, zero does not have any
@@ -37,8 +37,9 @@ def list_of_submasks(mask: int) -> list[int]:
"""
- fmt = "mask needs to be positive integer, your input {}"
- assert isinstance(mask, int) and mask > 0, fmt.format(mask)
+ assert isinstance(mask, int) and mask > 0, (
+ f"mask needs to be positive integer, your input {mask}"
+ )
"""
first submask iterated will be mask itself then operation will be performed
diff --git a/dynamic_programming/k_means_clustering_tensorflow.py_tf b/dynamic_programming/k_means_clustering_tensorflow.py
similarity index 98%
rename from dynamic_programming/k_means_clustering_tensorflow.py_tf
rename to dynamic_programming/k_means_clustering_tensorflow.py
index 4fbcedeaa0dc..8d3f6f0dfbcb 100644
--- a/dynamic_programming/k_means_clustering_tensorflow.py_tf
+++ b/dynamic_programming/k_means_clustering_tensorflow.py
@@ -1,9 +1,10 @@
-import tensorflow as tf
from random import shuffle
+
+import tensorflow as tf
from numpy import array
-def TFKMeansCluster(vectors, noofclusters):
+def tf_k_means_cluster(vectors, noofclusters):
"""
K-Means Clustering using TensorFlow.
'vectors' should be a n*k 2-D NumPy array, where n is the number
@@ -30,7 +31,6 @@ def TFKMeansCluster(vectors, noofclusters):
graph = tf.Graph()
with graph.as_default():
-
# SESSION OF COMPUTATION
sess = tf.Session()
@@ -95,8 +95,7 @@ def TFKMeansCluster(vectors, noofclusters):
# iterations. To keep things simple, we will only do a set number of
# iterations, instead of using a Stopping Criterion.
noofiterations = 100
- for iteration_n in range(noofiterations):
-
+ for _ in range(noofiterations):
##EXPECTATION STEP
##Based on the centroid locations till last iteration, compute
##the _expected_ centroid assignments.
diff --git a/dynamic_programming/knapsack.py b/dynamic_programming/knapsack.py
index 69e54c00aa4e..28c5b19dbe36 100644
--- a/dynamic_programming/knapsack.py
+++ b/dynamic_programming/knapsack.py
@@ -1,66 +1,68 @@
"""
Given weights and values of n items, put these items in a knapsack of
- capacity W to get the maximum total value in the knapsack.
+capacity W to get the maximum total value in the knapsack.
Note that only the integer weights 0-1 knapsack problem is solvable
- using dynamic programming.
+using dynamic programming.
"""
-def MF_knapsack(i, wt, val, j):
+def mf_knapsack(i, wt, val, j):
"""
This code involves the concept of memory functions. Here we solve the subproblems
which are needed unlike the below example
- F is a 2D array with -1s filled up
+ F is a 2D array with ``-1`` s filled up
"""
- global F # a global dp table for knapsack
- if F[i][j] < 0:
+ global f # a global dp table for knapsack
+ if f[i][j] < 0:
if j < wt[i - 1]:
- val = MF_knapsack(i - 1, wt, val, j)
+ val = mf_knapsack(i - 1, wt, val, j)
else:
val = max(
- MF_knapsack(i - 1, wt, val, j),
- MF_knapsack(i - 1, wt, val, j - wt[i - 1]) + val[i - 1],
+ mf_knapsack(i - 1, wt, val, j),
+ mf_knapsack(i - 1, wt, val, j - wt[i - 1]) + val[i - 1],
)
- F[i][j] = val
- return F[i][j]
+ f[i][j] = val
+ return f[i][j]
-def knapsack(W, wt, val, n):
- dp = [[0 for i in range(W + 1)] for j in range(n + 1)]
+def knapsack(w, wt, val, n):
+ dp = [[0] * (w + 1) for _ in range(n + 1)]
for i in range(1, n + 1):
- for w in range(1, W + 1):
- if wt[i - 1] <= w:
- dp[i][w] = max(val[i - 1] + dp[i - 1][w - wt[i - 1]], dp[i - 1][w])
+ for w_ in range(1, w + 1):
+ if wt[i - 1] <= w_:
+ dp[i][w_] = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]], dp[i - 1][w_])
else:
- dp[i][w] = dp[i - 1][w]
+ dp[i][w_] = dp[i - 1][w_]
- return dp[n][W], dp
+ return dp[n][w_], dp
-def knapsack_with_example_solution(W: int, wt: list, val: list):
+def knapsack_with_example_solution(w: int, wt: list, val: list):
"""
Solves the integer weights knapsack problem returns one of
the several possible optimal subsets.
Parameters
- ---------
+ ----------
- W: int, the total maximum weight for the given knapsack problem.
- wt: list, the vector of weights for all items where wt[i] is the weight
- of the i-th item.
- val: list, the vector of values for all items where val[i] is the value
- of the i-th item
+ * `w`: int, the total maximum weight for the given knapsack problem.
+ * `wt`: list, the vector of weights for all items where ``wt[i]`` is the weight
+ of the ``i``-th item.
+ * `val`: list, the vector of values for all items where ``val[i]`` is the value
+ of the ``i``-th item
Returns
-------
- optimal_val: float, the optimal value for the given knapsack problem
- example_optional_set: set, the indices of one of the optimal subsets
- which gave rise to the optimal value.
+
+ * `optimal_val`: float, the optimal value for the given knapsack problem
+ * `example_optional_set`: set, the indices of one of the optimal subsets
+ which gave rise to the optimal value.
Examples
- -------
+ --------
+
>>> knapsack_with_example_solution(10, [1, 3, 5, 2], [10, 20, 100, 22])
(142, {2, 3, 4})
>>> knapsack_with_example_solution(6, [4, 3, 2, 3], [3, 2, 4, 4])
@@ -78,21 +80,22 @@ def knapsack_with_example_solution(W: int, wt: list, val: list):
num_items = len(wt)
if num_items != len(val):
- raise ValueError(
- "The number of weights must be the "
- "same as the number of values.\nBut "
- f"got {num_items} weights and {len(val)} values"
+ msg = (
+ "The number of weights must be the same as the number of values.\n"
+ f"But got {num_items} weights and {len(val)} values"
)
+ raise ValueError(msg)
for i in range(num_items):
if not isinstance(wt[i], int):
- raise TypeError(
- "All weights must be integers but "
- f"got weight of type {type(wt[i])} at index {i}"
+ msg = (
+ "All weights must be integers but got weight of "
+ f"type {type(wt[i])} at index {i}"
)
+ raise TypeError(msg)
- optimal_val, dp_table = knapsack(W, wt, val, num_items)
- example_optional_set = set()
- _construct_solution(dp_table, wt, num_items, W, example_optional_set)
+ optimal_val, dp_table = knapsack(w, wt, val, num_items)
+ example_optional_set: set = set()
+ _construct_solution(dp_table, wt, num_items, w, example_optional_set)
return optimal_val, example_optional_set
@@ -103,19 +106,19 @@ def _construct_solution(dp: list, wt: list, i: int, j: int, optimal_set: set):
a filled DP table and the vector of weights
Parameters
- ---------
-
- dp: list of list, the table of a solved integer weight dynamic programming problem
+ ----------
- wt: list or tuple, the vector of weights of the items
- i: int, the index of the item under consideration
- j: int, the current possible maximum weight
- optimal_set: set, the optimal subset so far. This gets modified by the function.
+ * `dp`: list of list, the table of a solved integer weight dynamic programming
+ problem
+ * `wt`: list or tuple, the vector of weights of the items
+ * `i`: int, the index of the item under consideration
+ * `j`: int, the current possible maximum weight
+ * `optimal_set`: set, the optimal subset so far. This gets modified by the function.
Returns
-------
- None
+ ``None``
"""
# for the current item i at a maximum weight j to be part of an optimal subset,
# the optimal value at (i, j) must be greater than the optimal value at (i-1, j).
@@ -136,10 +139,10 @@ def _construct_solution(dp: list, wt: list, i: int, j: int, optimal_set: set):
wt = [4, 3, 2, 3]
n = 4
w = 6
- F = [[0] * (w + 1)] + [[0] + [-1 for i in range(w + 1)] for j in range(n + 1)]
+ f = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)]
optimal_solution, _ = knapsack(w, wt, val, n)
print(optimal_solution)
- print(MF_knapsack(n, wt, val, w)) # switched the n and w
+ print(mf_knapsack(n, wt, val, w)) # switched the n and w
# testing the dynamic programming problem with example
# the optimal subset for the above example are items 3 and 4
diff --git a/dynamic_programming/largest_divisible_subset.py b/dynamic_programming/largest_divisible_subset.py
new file mode 100644
index 000000000000..db38636e29db
--- /dev/null
+++ b/dynamic_programming/largest_divisible_subset.py
@@ -0,0 +1,74 @@
+from __future__ import annotations
+
+
+def largest_divisible_subset(items: list[int]) -> list[int]:
+ """
+ Algorithm to find the biggest subset in the given array such that for any 2 elements
+ x and y in the subset, either x divides y or y divides x.
+ >>> largest_divisible_subset([1, 16, 7, 8, 4])
+ [16, 8, 4, 1]
+ >>> largest_divisible_subset([1, 2, 3])
+ [2, 1]
+ >>> largest_divisible_subset([-1, -2, -3])
+ [-3]
+ >>> largest_divisible_subset([1, 2, 4, 8])
+ [8, 4, 2, 1]
+ >>> largest_divisible_subset((1, 2, 4, 8))
+ [8, 4, 2, 1]
+ >>> largest_divisible_subset([1, 1, 1])
+ [1, 1, 1]
+ >>> largest_divisible_subset([0, 0, 0])
+ [0, 0, 0]
+ >>> largest_divisible_subset([-1, -1, -1])
+ [-1, -1, -1]
+ >>> largest_divisible_subset([])
+ []
+ """
+ # Sort the array in ascending order as the sequence does not matter we only have to
+ # pick up a subset.
+ items = sorted(items)
+
+ number_of_items = len(items)
+
+ # Initialize memo with 1s and hash with increasing numbers
+ memo = [1] * number_of_items
+ hash_array = list(range(number_of_items))
+
+ # Iterate through the array
+ for i, item in enumerate(items):
+ for prev_index in range(i):
+ if ((items[prev_index] != 0 and item % items[prev_index]) == 0) and (
+ (1 + memo[prev_index]) > memo[i]
+ ):
+ memo[i] = 1 + memo[prev_index]
+ hash_array[i] = prev_index
+
+ ans = -1
+ last_index = -1
+
+ # Find the maximum length and its corresponding index
+ for i, memo_item in enumerate(memo):
+ if memo_item > ans:
+ ans = memo_item
+ last_index = i
+
+ # Reconstruct the divisible subset
+ if last_index == -1:
+ return []
+ result = [items[last_index]]
+ while hash_array[last_index] != last_index:
+ last_index = hash_array[last_index]
+ result.append(items[last_index])
+
+ return result
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+
+ items = [1, 16, 7, 8, 4]
+ print(
+ f"The longest divisible subset of {items} is {largest_divisible_subset(items)}."
+ )
diff --git a/dynamic_programming/longest_common_subsequence.py b/dynamic_programming/longest_common_subsequence.py
index fdcf3311a017..4a6c880aff61 100644
--- a/dynamic_programming/longest_common_subsequence.py
+++ b/dynamic_programming/longest_common_subsequence.py
@@ -28,6 +28,24 @@ def longest_common_subsequence(x: str, y: str):
(2, 'ph')
>>> longest_common_subsequence("computer", "food")
(1, 'o')
+ >>> longest_common_subsequence("", "abc") # One string is empty
+ (0, '')
+ >>> longest_common_subsequence("abc", "") # Other string is empty
+ (0, '')
+ >>> longest_common_subsequence("", "") # Both strings are empty
+ (0, '')
+ >>> longest_common_subsequence("abc", "def") # No common subsequence
+ (0, '')
+ >>> longest_common_subsequence("abc", "abc") # Identical strings
+ (3, 'abc')
+ >>> longest_common_subsequence("a", "a") # Single character match
+ (1, 'a')
+ >>> longest_common_subsequence("a", "b") # Single character no match
+ (0, '')
+ >>> longest_common_subsequence("abcdef", "ace") # Interleaved subsequence
+ (3, 'ace')
+ >>> longest_common_subsequence("ABCD", "ACBD") # No repeated characters
+ (3, 'ABD')
"""
# find the length of strings
@@ -38,36 +56,30 @@ def longest_common_subsequence(x: str, y: str):
n = len(y)
# declaring the array for storing the dp values
- L = [[0] * (n + 1) for _ in range(m + 1)]
+ dp = [[0] * (n + 1) for _ in range(m + 1)]
for i in range(1, m + 1):
for j in range(1, n + 1):
- if x[i - 1] == y[j - 1]:
- match = 1
- else:
- match = 0
+ match = 1 if x[i - 1] == y[j - 1] else 0
- L[i][j] = max(L[i - 1][j], L[i][j - 1], L[i - 1][j - 1] + match)
+ dp[i][j] = max(dp[i - 1][j], dp[i][j - 1], dp[i - 1][j - 1] + match)
seq = ""
i, j = m, n
while i > 0 and j > 0:
- if x[i - 1] == y[j - 1]:
- match = 1
- else:
- match = 0
+ match = 1 if x[i - 1] == y[j - 1] else 0
- if L[i][j] == L[i - 1][j - 1] + match:
+ if dp[i][j] == dp[i - 1][j - 1] + match:
if match == 1:
seq = x[i - 1] + seq
i -= 1
j -= 1
- elif L[i][j] == L[i - 1][j]:
+ elif dp[i][j] == dp[i - 1][j]:
i -= 1
else:
j -= 1
- return L[m][n], seq
+ return dp[m][n], seq
if __name__ == "__main__":
diff --git a/dynamic_programming/longest_common_substring.py b/dynamic_programming/longest_common_substring.py
new file mode 100644
index 000000000000..ea5233eb2d17
--- /dev/null
+++ b/dynamic_programming/longest_common_substring.py
@@ -0,0 +1,67 @@
+"""
+Longest Common Substring Problem Statement:
+ Given two sequences, find the
+ longest common substring present in both of them. A substring is
+ necessarily continuous.
+
+Example:
+ ``abcdef`` and ``xabded`` have two longest common substrings, ``ab`` or ``de``.
+ Therefore, algorithm should return any one of them.
+"""
+
+
+def longest_common_substring(text1: str, text2: str) -> str:
+ """
+ Finds the longest common substring between two strings.
+
+ >>> longest_common_substring("", "")
+ ''
+ >>> longest_common_substring("a","")
+ ''
+ >>> longest_common_substring("", "a")
+ ''
+ >>> longest_common_substring("a", "a")
+ 'a'
+ >>> longest_common_substring("abcdef", "bcd")
+ 'bcd'
+ >>> longest_common_substring("abcdef", "xabded")
+ 'ab'
+ >>> longest_common_substring("GeeksforGeeks", "GeeksQuiz")
+ 'Geeks'
+ >>> longest_common_substring("abcdxyz", "xyzabcd")
+ 'abcd'
+ >>> longest_common_substring("zxabcdezy", "yzabcdezx")
+ 'abcdez'
+ >>> longest_common_substring("OldSite:GeeksforGeeks.org", "NewSite:GeeksQuiz.com")
+ 'Site:Geeks'
+ >>> longest_common_substring(1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: longest_common_substring() takes two strings for inputs
+ """
+
+ if not (isinstance(text1, str) and isinstance(text2, str)):
+ raise ValueError("longest_common_substring() takes two strings for inputs")
+
+ text1_length = len(text1)
+ text2_length = len(text2)
+
+ dp = [[0] * (text2_length + 1) for _ in range(text1_length + 1)]
+ ans_index = 0
+ ans_length = 0
+
+ for i in range(1, text1_length + 1):
+ for j in range(1, text2_length + 1):
+ if text1[i - 1] == text2[j - 1]:
+ dp[i][j] = 1 + dp[i - 1][j - 1]
+ if dp[i][j] > ans_length:
+ ans_index = i
+ ans_length = dp[i][j]
+
+ return text1[ans_index - ans_length : ans_index]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/longest_increasing_subsequence.py b/dynamic_programming/longest_increasing_subsequence.py
index f5ca8a2b5cdc..1863a882c41e 100644
--- a/dynamic_programming/longest_increasing_subsequence.py
+++ b/dynamic_programming/longest_increasing_subsequence.py
@@ -4,24 +4,30 @@
This is a pure Python implementation of Dynamic Programming solution to the longest
increasing subsequence of a given sequence.
-The problem is :
-Given an array, to find the longest and increasing sub-array in that given array and
-return it.
-Example: [10, 22, 9, 33, 21, 50, 41, 60, 80] as input will return
- [10, 22, 33, 41, 60, 80] as output
+The problem is:
+ Given an array, to find the longest and increasing sub-array in that given array and
+ return it.
+
+Example:
+ ``[10, 22, 9, 33, 21, 50, 41, 60, 80]`` as input will return
+ ``[10, 22, 33, 41, 60, 80]`` as output
"""
+
from __future__ import annotations
def longest_subsequence(array: list[int]) -> list[int]: # This function is recursive
"""
Some examples
+
>>> longest_subsequence([10, 22, 9, 33, 21, 50, 41, 60, 80])
[10, 22, 33, 41, 60, 80]
>>> longest_subsequence([4, 8, 7, 5, 1, 12, 2, 3, 9])
[1, 2, 3, 9]
+ >>> longest_subsequence([28, 26, 12, 23, 35, 39])
+ [12, 23, 35, 39]
>>> longest_subsequence([9, 8, 7, 6, 5, 7])
- [8]
+ [5, 7]
>>> longest_subsequence([1, 1, 1])
[1, 1, 1]
>>> longest_subsequence([])
@@ -34,13 +40,13 @@ def longest_subsequence(array: list[int]) -> list[int]: # This function is recu
return array
# Else
pivot = array[0]
- isFound = False
+ is_found = False
i = 1
- longest_subseq = []
- while not isFound and i < array_length:
+ longest_subseq: list[int] = []
+ while not is_found and i < array_length:
if array[i] < pivot:
- isFound = True
- temp_array = [element for element in array[i:] if element >= array[i]]
+ is_found = True
+ temp_array = array[i:]
temp_array = longest_subsequence(temp_array)
if len(temp_array) > len(longest_subseq):
longest_subseq = temp_array
@@ -48,7 +54,7 @@ def longest_subsequence(array: list[int]) -> list[int]: # This function is recu
i += 1
temp_array = [element for element in array[1:] if element >= pivot]
- temp_array = [pivot] + longest_subsequence(temp_array)
+ temp_array = [pivot, *longest_subsequence(temp_array)]
if len(temp_array) > len(longest_subseq):
return temp_array
else:
diff --git a/dynamic_programming/longest_increasing_subsequence_iterative.py b/dynamic_programming/longest_increasing_subsequence_iterative.py
new file mode 100644
index 000000000000..665c86a35d2e
--- /dev/null
+++ b/dynamic_programming/longest_increasing_subsequence_iterative.py
@@ -0,0 +1,72 @@
+"""
+Author : Sanjay Muthu
+
+This is a pure Python implementation of Dynamic Programming solution to the longest
+increasing subsequence of a given sequence.
+
+The problem is:
+ Given an array, to find the longest and increasing sub-array in that given array and
+ return it.
+
+Example:
+ ``[10, 22, 9, 33, 21, 50, 41, 60, 80]`` as input will return
+ ``[10, 22, 33, 50, 60, 80]`` as output
+"""
+
+from __future__ import annotations
+
+import copy
+
+
+def longest_subsequence(array: list[int]) -> list[int]:
+ """
+ Some examples
+
+ >>> longest_subsequence([10, 22, 9, 33, 21, 50, 41, 60, 80])
+ [10, 22, 33, 50, 60, 80]
+ >>> longest_subsequence([4, 8, 7, 5, 1, 12, 2, 3, 9])
+ [1, 2, 3, 9]
+ >>> longest_subsequence([9, 8, 7, 6, 5, 7])
+ [7, 7]
+ >>> longest_subsequence([28, 26, 12, 23, 35, 39])
+ [12, 23, 35, 39]
+ >>> longest_subsequence([1, 1, 1])
+ [1, 1, 1]
+ >>> longest_subsequence([])
+ []
+ """
+ n = len(array)
+ # The longest increasing subsequence ending at array[i]
+ longest_increasing_subsequence = []
+ for i in range(n):
+ longest_increasing_subsequence.append([array[i]])
+
+ for i in range(1, n):
+ for prev in range(i):
+ # If array[prev] is less than or equal to array[i], then
+ # longest_increasing_subsequence[prev] + array[i]
+ # is a valid increasing subsequence
+
+ # longest_increasing_subsequence[i] is only set to
+ # longest_increasing_subsequence[prev] + array[i] if the length is longer.
+
+ if array[prev] <= array[i] and len(
+ longest_increasing_subsequence[prev]
+ ) + 1 > len(longest_increasing_subsequence[i]):
+ longest_increasing_subsequence[i] = copy.copy(
+ longest_increasing_subsequence[prev]
+ )
+ longest_increasing_subsequence[i].append(array[i])
+
+ result: list[int] = []
+ for i in range(n):
+ if len(longest_increasing_subsequence[i]) > len(result):
+ result = longest_increasing_subsequence[i]
+
+ return result
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/longest_increasing_subsequence_o(nlogn).py b/dynamic_programming/longest_increasing_subsequence_o_nlogn.py
similarity index 52%
rename from dynamic_programming/longest_increasing_subsequence_o(nlogn).py
rename to dynamic_programming/longest_increasing_subsequence_o_nlogn.py
index af536f8bbd01..bbc7a62b6b5c 100644
--- a/dynamic_programming/longest_increasing_subsequence_o(nlogn).py
+++ b/dynamic_programming/longest_increasing_subsequence_o_nlogn.py
@@ -7,26 +7,26 @@
from __future__ import annotations
-def CeilIndex(v, l, r, key): # noqa: E741
- while r - l > 1:
- m = (l + r) // 2
- if v[m] >= key:
- r = m
+def ceil_index(v, left, right, key):
+ while right - left > 1:
+ middle = (left + right) // 2
+ if v[middle] >= key:
+ right = middle
else:
- l = m # noqa: E741
- return r
+ left = middle
+ return right
-def LongestIncreasingSubsequenceLength(v: list[int]) -> int:
+def longest_increasing_subsequence_length(v: list[int]) -> int:
"""
- >>> LongestIncreasingSubsequenceLength([2, 5, 3, 7, 11, 8, 10, 13, 6])
+ >>> longest_increasing_subsequence_length([2, 5, 3, 7, 11, 8, 10, 13, 6])
6
- >>> LongestIncreasingSubsequenceLength([])
+ >>> longest_increasing_subsequence_length([])
0
- >>> LongestIncreasingSubsequenceLength([0, 8, 4, 12, 2, 10, 6, 14, 1, 9, 5, 13, 3,
- ... 11, 7, 15])
+ >>> longest_increasing_subsequence_length([0, 8, 4, 12, 2, 10, 6, 14, 1, 9, 5, 13,
+ ... 3, 11, 7, 15])
6
- >>> LongestIncreasingSubsequenceLength([5, 4, 3, 2, 1])
+ >>> longest_increasing_subsequence_length([5, 4, 3, 2, 1])
1
"""
if len(v) == 0:
@@ -44,7 +44,7 @@ def LongestIncreasingSubsequenceLength(v: list[int]) -> int:
tail[length] = v[i]
length += 1
else:
- tail[CeilIndex(tail, -1, length - 1, v[i])] = v[i]
+ tail[ceil_index(tail, -1, length - 1, v[i])] = v[i]
return length
diff --git a/dynamic_programming/longest_palindromic_subsequence.py b/dynamic_programming/longest_palindromic_subsequence.py
new file mode 100644
index 000000000000..a60d95e460e6
--- /dev/null
+++ b/dynamic_programming/longest_palindromic_subsequence.py
@@ -0,0 +1,44 @@
+"""
+author: Sanket Kittad
+Given a string s, find the longest palindromic subsequence's length in s.
+Input: s = "bbbab"
+Output: 4
+Explanation: One possible longest palindromic subsequence is "bbbb".
+Leetcode link: https://leetcode.com/problems/longest-palindromic-subsequence/description/
+"""
+
+
+def longest_palindromic_subsequence(input_string: str) -> int:
+ """
+ This function returns the longest palindromic subsequence in a string
+ >>> longest_palindromic_subsequence("bbbab")
+ 4
+ >>> longest_palindromic_subsequence("bbabcbcab")
+ 7
+ """
+ n = len(input_string)
+ rev = input_string[::-1]
+ m = len(rev)
+ dp = [[-1] * (m + 1) for i in range(n + 1)]
+ for i in range(n + 1):
+ dp[i][0] = 0
+ for i in range(m + 1):
+ dp[0][i] = 0
+
+ # create and initialise dp array
+ for i in range(1, n + 1):
+ for j in range(1, m + 1):
+ # If characters at i and j are the same
+ # include them in the palindromic subsequence
+ if input_string[i - 1] == rev[j - 1]:
+ dp[i][j] = 1 + dp[i - 1][j - 1]
+ else:
+ dp[i][j] = max(dp[i - 1][j], dp[i][j - 1])
+
+ return dp[n][m]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/longest_sub_array.py b/dynamic_programming/longest_sub_array.py
deleted file mode 100644
index 30159a1386c3..000000000000
--- a/dynamic_programming/longest_sub_array.py
+++ /dev/null
@@ -1,34 +0,0 @@
-"""
-Author : Yvonne
-
-This is a pure Python implementation of Dynamic Programming solution to the
- longest_sub_array problem.
-
-The problem is :
-Given an array, to find the longest and continuous sub array and get the max sum of the
- sub array in the given array.
-"""
-
-
-class SubArray:
- def __init__(self, arr):
- # we need a list not a string, so do something to change the type
- self.array = arr.split(",")
- print(("the input array is:", self.array))
-
- def solve_sub_array(self):
- rear = [int(self.array[0])] * len(self.array)
- sum_value = [int(self.array[0])] * len(self.array)
- for i in range(1, len(self.array)):
- sum_value[i] = max(
- int(self.array[i]) + sum_value[i - 1], int(self.array[i])
- )
- rear[i] = max(sum_value[i], rear[i - 1])
- return rear[len(self.array) - 1]
-
-
-if __name__ == "__main__":
- whole_array = input("please input some numbers:")
- array = SubArray(whole_array)
- re = array.solve_sub_array()
- print(("the results is:", re))
diff --git a/dynamic_programming/matrix_chain_multiplication.py b/dynamic_programming/matrix_chain_multiplication.py
new file mode 100644
index 000000000000..4c0c771f9092
--- /dev/null
+++ b/dynamic_programming/matrix_chain_multiplication.py
@@ -0,0 +1,151 @@
+"""
+| Find the minimum number of multiplications needed to multiply chain of matrices.
+| Reference: https://www.geeksforgeeks.org/matrix-chain-multiplication-dp-8/
+
+The algorithm has interesting real-world applications.
+
+Example:
+ 1. Image transformations in Computer Graphics as images are composed of matrix.
+ 2. Solve complex polynomial equations in the field of algebra using least processing
+ power.
+ 3. Calculate overall impact of macroeconomic decisions as economic equations involve a
+ number of variables.
+ 4. Self-driving car navigation can be made more accurate as matrix multiplication can
+ accurately determine position and orientation of obstacles in short time.
+
+Python doctests can be run with the following command::
+
+ python -m doctest -v matrix_chain_multiply.py
+
+Given a sequence ``arr[]`` that represents chain of 2D matrices such that the dimension
+of the ``i`` th matrix is ``arr[i-1]*arr[i]``.
+So suppose ``arr = [40, 20, 30, 10, 30]`` means we have ``4`` matrices of dimensions
+``40*20``, ``20*30``, ``30*10`` and ``10*30``.
+
+``matrix_chain_multiply()`` returns an integer denoting minimum number of
+multiplications to multiply the chain.
+
+We do not need to perform actual multiplication here.
+We only need to decide the order in which to perform the multiplication.
+
+Hints:
+ 1. Number of multiplications (ie cost) to multiply ``2`` matrices
+ of size ``m*p`` and ``p*n`` is ``m*p*n``.
+ 2. Cost of matrix multiplication is not associative ie ``(M1*M2)*M3 != M1*(M2*M3)``
+ 3. Matrix multiplication is not commutative. So, ``M1*M2`` does not mean ``M2*M1``
+ can be done.
+ 4. To determine the required order, we can try different combinations.
+
+So, this problem has overlapping sub-problems and can be solved using recursion.
+We use Dynamic Programming for optimal time complexity.
+
+Example input:
+ ``arr = [40, 20, 30, 10, 30]``
+output:
+ ``26000``
+"""
+
+from collections.abc import Iterator
+from contextlib import contextmanager
+from functools import cache
+from sys import maxsize
+
+
+def matrix_chain_multiply(arr: list[int]) -> int:
+ """
+ Find the minimum number of multiplcations required to multiply the chain of matrices
+
+ Args:
+ `arr`: The input array of integers.
+
+ Returns:
+ Minimum number of multiplications needed to multiply the chain
+
+ Examples:
+
+ >>> matrix_chain_multiply([1, 2, 3, 4, 3])
+ 30
+ >>> matrix_chain_multiply([10])
+ 0
+ >>> matrix_chain_multiply([10, 20])
+ 0
+ >>> matrix_chain_multiply([19, 2, 19])
+ 722
+ >>> matrix_chain_multiply(list(range(1, 100)))
+ 323398
+ >>> # matrix_chain_multiply(list(range(1, 251)))
+ # 2626798
+ """
+ if len(arr) < 2:
+ return 0
+ # initialising 2D dp matrix
+ n = len(arr)
+ dp = [[maxsize for j in range(n)] for i in range(n)]
+ # we want minimum cost of multiplication of matrices
+ # of dimension (i*k) and (k*j). This cost is arr[i-1]*arr[k]*arr[j].
+ for i in range(n - 1, 0, -1):
+ for j in range(i, n):
+ if i == j:
+ dp[i][j] = 0
+ continue
+ for k in range(i, j):
+ dp[i][j] = min(
+ dp[i][j], dp[i][k] + dp[k + 1][j] + arr[i - 1] * arr[k] * arr[j]
+ )
+
+ return dp[1][n - 1]
+
+
+def matrix_chain_order(dims: list[int]) -> int:
+ """
+ Source: https://en.wikipedia.org/wiki/Matrix_chain_multiplication
+
+ The dynamic programming solution is faster than cached the recursive solution and
+ can handle larger inputs.
+
+ >>> matrix_chain_order([1, 2, 3, 4, 3])
+ 30
+ >>> matrix_chain_order([10])
+ 0
+ >>> matrix_chain_order([10, 20])
+ 0
+ >>> matrix_chain_order([19, 2, 19])
+ 722
+ >>> matrix_chain_order(list(range(1, 100)))
+ 323398
+ >>> # matrix_chain_order(list(range(1, 251))) # Max before RecursionError is raised
+ # 2626798
+ """
+
+ @cache
+ def a(i: int, j: int) -> int:
+ return min(
+ (a(i, k) + dims[i] * dims[k] * dims[j] + a(k, j) for k in range(i + 1, j)),
+ default=0,
+ )
+
+ return a(0, len(dims) - 1)
+
+
+@contextmanager
+def elapsed_time(msg: str) -> Iterator:
+ # print(f"Starting: {msg}")
+ from time import perf_counter_ns
+
+ start = perf_counter_ns()
+ yield
+ print(f"Finished: {msg} in {(perf_counter_ns() - start) / 10**9} seconds.")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ with elapsed_time("matrix_chain_order"):
+ print(f"{matrix_chain_order(list(range(1, 251))) = }")
+ with elapsed_time("matrix_chain_multiply"):
+ print(f"{matrix_chain_multiply(list(range(1, 251))) = }")
+ with elapsed_time("matrix_chain_order"):
+ print(f"{matrix_chain_order(list(range(1, 251))) = }")
+ with elapsed_time("matrix_chain_multiply"):
+ print(f"{matrix_chain_multiply(list(range(1, 251))) = }")
diff --git a/dynamic_programming/matrix_chain_order.py b/dynamic_programming/matrix_chain_order.py
index 9411bc704f1c..d612aea7b99d 100644
--- a/dynamic_programming/matrix_chain_order.py
+++ b/dynamic_programming/matrix_chain_order.py
@@ -8,34 +8,34 @@
"""
-def MatrixChainOrder(array):
- N = len(array)
- Matrix = [[0 for x in range(N)] for x in range(N)]
- Sol = [[0 for x in range(N)] for x in range(N)]
+def matrix_chain_order(array):
+ n = len(array)
+ matrix = [[0 for x in range(n)] for x in range(n)]
+ sol = [[0 for x in range(n)] for x in range(n)]
- for ChainLength in range(2, N):
- for a in range(1, N - ChainLength + 1):
- b = a + ChainLength - 1
+ for chain_length in range(2, n):
+ for a in range(1, n - chain_length + 1):
+ b = a + chain_length - 1
- Matrix[a][b] = sys.maxsize
+ matrix[a][b] = sys.maxsize
for c in range(a, b):
cost = (
- Matrix[a][c] + Matrix[c + 1][b] + array[a - 1] * array[c] * array[b]
+ matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b]
)
- if cost < Matrix[a][b]:
- Matrix[a][b] = cost
- Sol[a][b] = c
- return Matrix, Sol
+ if cost < matrix[a][b]:
+ matrix[a][b] = cost
+ sol[a][b] = c
+ return matrix, sol
# Print order of matrix with Ai as Matrix
-def PrintOptimalSolution(OptimalSolution, i, j):
+def print_optiomal_solution(optimal_solution, i, j):
if i == j:
print("A" + str(i), end=" ")
else:
print("(", end=" ")
- PrintOptimalSolution(OptimalSolution, i, OptimalSolution[i][j])
- PrintOptimalSolution(OptimalSolution, OptimalSolution[i][j] + 1, j)
+ print_optiomal_solution(optimal_solution, i, optimal_solution[i][j])
+ print_optiomal_solution(optimal_solution, optimal_solution[i][j] + 1, j)
print(")", end=" ")
@@ -44,10 +44,10 @@ def main():
n = len(array)
# Size of matrix created from above array will be
# 30*35 35*15 15*5 5*10 10*20 20*25
- Matrix, OptimalSolution = MatrixChainOrder(array)
+ matrix, optimal_solution = matrix_chain_order(array)
- print("No. of Operation required: " + str(Matrix[1][n - 1]))
- PrintOptimalSolution(OptimalSolution, 1, n - 1)
+ print("No. of Operation required: " + str(matrix[1][n - 1]))
+ print_optiomal_solution(optimal_solution, 1, n - 1)
if __name__ == "__main__":
diff --git a/dynamic_programming/max_non_adjacent_sum.py b/dynamic_programming/max_non_adjacent_sum.py
index 5362b22ca9dc..e3cc23f4983e 100644
--- a/dynamic_programming/max_non_adjacent_sum.py
+++ b/dynamic_programming/max_non_adjacent_sum.py
@@ -7,7 +7,7 @@ def maximum_non_adjacent_sum(nums: list[int]) -> int:
"""
Find the maximum non-adjacent sum of the integers in the nums input list
- >>> print(maximum_non_adjacent_sum([1, 2, 3]))
+ >>> maximum_non_adjacent_sum([1, 2, 3])
4
>>> maximum_non_adjacent_sum([1, 5, 3, 7, 2, 2, 6])
18
diff --git a/dynamic_programming/max_product_subarray.py b/dynamic_programming/max_product_subarray.py
new file mode 100644
index 000000000000..6f4f38e38942
--- /dev/null
+++ b/dynamic_programming/max_product_subarray.py
@@ -0,0 +1,54 @@
+def max_product_subarray(numbers: list[int]) -> int:
+ """
+ Returns the maximum product that can be obtained by multiplying a
+ contiguous subarray of the given integer list `numbers`.
+
+ Example:
+
+ >>> max_product_subarray([2, 3, -2, 4])
+ 6
+ >>> max_product_subarray((-2, 0, -1))
+ 0
+ >>> max_product_subarray([2, 3, -2, 4, -1])
+ 48
+ >>> max_product_subarray([-1])
+ -1
+ >>> max_product_subarray([0])
+ 0
+ >>> max_product_subarray([])
+ 0
+ >>> max_product_subarray("")
+ 0
+ >>> max_product_subarray(None)
+ 0
+ >>> max_product_subarray([2, 3, -2, 4.5, -1])
+ Traceback (most recent call last):
+ ...
+ ValueError: numbers must be an iterable of integers
+ >>> max_product_subarray("ABC")
+ Traceback (most recent call last):
+ ...
+ ValueError: numbers must be an iterable of integers
+ """
+ if not numbers:
+ return 0
+
+ if not isinstance(numbers, (list, tuple)) or not all(
+ isinstance(number, int) for number in numbers
+ ):
+ raise ValueError("numbers must be an iterable of integers")
+
+ max_till_now = min_till_now = max_prod = numbers[0]
+
+ for i in range(1, len(numbers)):
+ # update the maximum and minimum subarray products
+ number = numbers[i]
+ if number < 0:
+ max_till_now, min_till_now = min_till_now, max_till_now
+ max_till_now = max(number, max_till_now * number)
+ min_till_now = min(number, min_till_now * number)
+
+ # update the maximum product found till now
+ max_prod = max(max_prod, max_till_now)
+
+ return max_prod
diff --git a/dynamic_programming/max_sub_array.py b/dynamic_programming/max_sub_array.py
deleted file mode 100644
index 3060010ef7c6..000000000000
--- a/dynamic_programming/max_sub_array.py
+++ /dev/null
@@ -1,94 +0,0 @@
-"""
-author : Mayank Kumar Jha (mk9440)
-"""
-from __future__ import annotations
-
-
-def find_max_sub_array(A, low, high):
- if low == high:
- return low, high, A[low]
- else:
- mid = (low + high) // 2
- left_low, left_high, left_sum = find_max_sub_array(A, low, mid)
- right_low, right_high, right_sum = find_max_sub_array(A, mid + 1, high)
- cross_left, cross_right, cross_sum = find_max_cross_sum(A, low, mid, high)
- if left_sum >= right_sum and left_sum >= cross_sum:
- return left_low, left_high, left_sum
- elif right_sum >= left_sum and right_sum >= cross_sum:
- return right_low, right_high, right_sum
- else:
- return cross_left, cross_right, cross_sum
-
-
-def find_max_cross_sum(A, low, mid, high):
- left_sum, max_left = -999999999, -1
- right_sum, max_right = -999999999, -1
- summ = 0
- for i in range(mid, low - 1, -1):
- summ += A[i]
- if summ > left_sum:
- left_sum = summ
- max_left = i
- summ = 0
- for i in range(mid + 1, high + 1):
- summ += A[i]
- if summ > right_sum:
- right_sum = summ
- max_right = i
- return max_left, max_right, (left_sum + right_sum)
-
-
-def max_sub_array(nums: list[int]) -> int:
- """
- Finds the contiguous subarray which has the largest sum and return its sum.
-
- >>> max_sub_array([-2, 1, -3, 4, -1, 2, 1, -5, 4])
- 6
-
- An empty (sub)array has sum 0.
- >>> max_sub_array([])
- 0
-
- If all elements are negative, the largest subarray would be the empty array,
- having the sum 0.
- >>> max_sub_array([-1, -2, -3])
- 0
- >>> max_sub_array([5, -2, -3])
- 5
- >>> max_sub_array([31, -41, 59, 26, -53, 58, 97, -93, -23, 84])
- 187
- """
- best = 0
- current = 0
- for i in nums:
- current += i
- if current < 0:
- current = 0
- best = max(best, current)
- return best
-
-
-if __name__ == "__main__":
- """
- A random simulation of this algorithm.
- """
- import time
- from random import randint
-
- from matplotlib import pyplot as plt
-
- inputs = [10, 100, 1000, 10000, 50000, 100000, 200000, 300000, 400000, 500000]
- tim = []
- for i in inputs:
- li = [randint(1, i) for j in range(i)]
- strt = time.time()
- (find_max_sub_array(li, 0, len(li) - 1))
- end = time.time()
- tim.append(end - strt)
- print("No of Inputs Time Taken")
- for i in range(len(inputs)):
- print(inputs[i], "\t\t", tim[i])
- plt.plot(inputs, tim)
- plt.xlabel("Number of Inputs")
- plt.ylabel("Time taken in seconds ")
- plt.show()
diff --git a/dynamic_programming/max_subarray_sum.py b/dynamic_programming/max_subarray_sum.py
new file mode 100644
index 000000000000..8c1dc0889a85
--- /dev/null
+++ b/dynamic_programming/max_subarray_sum.py
@@ -0,0 +1,61 @@
+"""
+The maximum subarray sum problem is the task of finding the maximum sum that can be
+obtained from a contiguous subarray within a given array of numbers. For example, given
+the array [-2, 1, -3, 4, -1, 2, 1, -5, 4], the contiguous subarray with the maximum sum
+is [4, -1, 2, 1], so the maximum subarray sum is 6.
+
+Kadane's algorithm is a simple dynamic programming algorithm that solves the maximum
+subarray sum problem in O(n) time and O(1) space.
+
+Reference: https://en.wikipedia.org/wiki/Maximum_subarray_problem
+"""
+
+from collections.abc import Sequence
+
+
+def max_subarray_sum(
+ arr: Sequence[float], allow_empty_subarrays: bool = False
+) -> float:
+ """
+ Solves the maximum subarray sum problem using Kadane's algorithm.
+ :param arr: the given array of numbers
+ :param allow_empty_subarrays: if True, then the algorithm considers empty subarrays
+
+ >>> max_subarray_sum([2, 8, 9])
+ 19
+ >>> max_subarray_sum([0, 0])
+ 0
+ >>> max_subarray_sum([-1.0, 0.0, 1.0])
+ 1.0
+ >>> max_subarray_sum([1, 2, 3, 4, -2])
+ 10
+ >>> max_subarray_sum([-2, 1, -3, 4, -1, 2, 1, -5, 4])
+ 6
+ >>> max_subarray_sum([2, 3, -9, 8, -2])
+ 8
+ >>> max_subarray_sum([-2, -3, -1, -4, -6])
+ -1
+ >>> max_subarray_sum([-2, -3, -1, -4, -6], allow_empty_subarrays=True)
+ 0
+ >>> max_subarray_sum([])
+ 0
+ """
+ if not arr:
+ return 0
+
+ max_sum = 0 if allow_empty_subarrays else float("-inf")
+ curr_sum = 0.0
+ for num in arr:
+ curr_sum = max(0 if allow_empty_subarrays else num, curr_sum + num)
+ max_sum = max(max_sum, curr_sum)
+
+ return max_sum
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+
+ nums = [-2, 1, -3, 4, -1, 2, 1, -5, 4]
+ print(f"{max_subarray_sum(nums) = }")
diff --git a/dynamic_programming/max_sum_contiguous_subsequence.py b/dynamic_programming/max_sum_contiguous_subsequence.py
deleted file mode 100644
index bac592370c5d..000000000000
--- a/dynamic_programming/max_sum_contiguous_subsequence.py
+++ /dev/null
@@ -1,20 +0,0 @@
-def max_subarray_sum(nums: list) -> int:
- """
- >>> max_subarray_sum([6 , 9, -1, 3, -7, -5, 10])
- 17
- """
- if not nums:
- return 0
- n = len(nums)
-
- res, s, s_pre = nums[0], nums[0], nums[0]
- for i in range(1, n):
- s = max(nums[i], s_pre + nums[i])
- s_pre = s
- res = max(res, s)
- return res
-
-
-if __name__ == "__main__":
- nums = [6, 9, -1, 3, -7, -5, 10]
- print(max_subarray_sum(nums))
diff --git a/dynamic_programming/min_distance_up_bottom.py b/dynamic_programming/min_distance_up_bottom.py
new file mode 100644
index 000000000000..6b38a41a1c0a
--- /dev/null
+++ b/dynamic_programming/min_distance_up_bottom.py
@@ -0,0 +1,49 @@
+"""
+Author : Alexander Pantyukhin
+Date : October 14, 2022
+This is an implementation of the up-bottom approach to find edit distance.
+The implementation was tested on Leetcode: https://leetcode.com/problems/edit-distance/
+
+Levinstein distance
+Dynamic Programming: up -> down.
+"""
+
+import functools
+
+
+def min_distance_up_bottom(word1: str, word2: str) -> int:
+ """
+ >>> min_distance_up_bottom("intention", "execution")
+ 5
+ >>> min_distance_up_bottom("intention", "")
+ 9
+ >>> min_distance_up_bottom("", "")
+ 0
+ >>> min_distance_up_bottom("zooicoarchaeologist", "zoologist")
+ 10
+ """
+ len_word1 = len(word1)
+ len_word2 = len(word2)
+
+ @functools.cache
+ def min_distance(index1: int, index2: int) -> int:
+ # if first word index overflows - delete all from the second word
+ if index1 >= len_word1:
+ return len_word2 - index2
+ # if second word index overflows - delete all from the first word
+ if index2 >= len_word2:
+ return len_word1 - index1
+ diff = int(word1[index1] != word2[index2]) # current letters not identical
+ return min(
+ 1 + min_distance(index1 + 1, index2),
+ 1 + min_distance(index1, index2 + 1),
+ diff + min_distance(index1 + 1, index2 + 1),
+ )
+
+ return min_distance(0, 0)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/coin_change.py b/dynamic_programming/minimum_coin_change.py
similarity index 96%
rename from dynamic_programming/coin_change.py
rename to dynamic_programming/minimum_coin_change.py
index 2869b5857be1..848bd654d3b9 100644
--- a/dynamic_programming/coin_change.py
+++ b/dynamic_programming/minimum_coin_change.py
@@ -7,7 +7,7 @@
"""
-def dp_count(S, n):
+def dp_count(s, n):
"""
>>> dp_count([1, 2, 3], 4)
4
@@ -33,7 +33,7 @@ def dp_count(S, n):
# Pick all coins one by one and update table[] values
# after the index greater than or equal to the value of the
# picked coin
- for coin_val in S:
+ for coin_val in s:
for j in range(coin_val, n + 1):
table[j] += table[j - coin_val]
diff --git a/dynamic_programming/minimum_partition.py b/dynamic_programming/minimum_partition.py
index 8fad4ef3072f..748c0599efb0 100644
--- a/dynamic_programming/minimum_partition.py
+++ b/dynamic_programming/minimum_partition.py
@@ -3,13 +3,53 @@
"""
-def findMin(arr):
- n = len(arr)
- s = sum(arr)
+def find_min(numbers: list[int]) -> int:
+ """
+ >>> find_min([1, 2, 3, 4, 5])
+ 1
+ >>> find_min([5, 5, 5, 5, 5])
+ 5
+ >>> find_min([5, 5, 5, 5])
+ 0
+ >>> find_min([3])
+ 3
+ >>> find_min([])
+ 0
+ >>> find_min([1, 2, 3, 4])
+ 0
+ >>> find_min([0, 0, 0, 0])
+ 0
+ >>> find_min([-1, -5, 5, 1])
+ 0
+ >>> find_min([-1, -5, 5, 1])
+ 0
+ >>> find_min([9, 9, 9, 9, 9])
+ 9
+ >>> find_min([1, 5, 10, 3])
+ 1
+ >>> find_min([-1, 0, 1])
+ 0
+ >>> find_min(range(10, 0, -1))
+ 1
+ >>> find_min([-1])
+ Traceback (most recent call last):
+ --
+ IndexError: list assignment index out of range
+ >>> find_min([0, 0, 0, 1, 2, -4])
+ Traceback (most recent call last):
+ ...
+ IndexError: list assignment index out of range
+ >>> find_min([-1, -5, -10, -3])
+ Traceback (most recent call last):
+ ...
+ IndexError: list assignment index out of range
+ """
+ n = len(numbers)
+ s = sum(numbers)
dp = [[False for x in range(s + 1)] for y in range(n + 1)]
- for i in range(1, n + 1):
+ for i in range(n + 1):
dp[i][0] = True
for i in range(1, s + 1):
@@ -17,10 +57,10 @@ def findMin(arr):
for i in range(1, n + 1):
for j in range(1, s + 1):
- dp[i][j] = dp[i][j - 1]
+ dp[i][j] = dp[i - 1][j]
- if arr[i - 1] <= j:
- dp[i][j] = dp[i][j] or dp[i - 1][j - arr[i - 1]]
+ if numbers[i - 1] <= j:
+ dp[i][j] = dp[i][j] or dp[i - 1][j - numbers[i - 1]]
for j in range(int(s / 2), -1, -1):
if dp[n][j] is True:
@@ -28,3 +68,9 @@ def findMin(arr):
break
return diff
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/dynamic_programming/minimum_size_subarray_sum.py b/dynamic_programming/minimum_size_subarray_sum.py
new file mode 100644
index 000000000000..3868d73535fb
--- /dev/null
+++ b/dynamic_programming/minimum_size_subarray_sum.py
@@ -0,0 +1,62 @@
+import sys
+
+
+def minimum_subarray_sum(target: int, numbers: list[int]) -> int:
+ """
+ Return the length of the shortest contiguous subarray in a list of numbers whose sum
+ is at least target. Reference: https://stackoverflow.com/questions/8269916
+
+ >>> minimum_subarray_sum(7, [2, 3, 1, 2, 4, 3])
+ 2
+ >>> minimum_subarray_sum(7, [2, 3, -1, 2, 4, -3])
+ 4
+ >>> minimum_subarray_sum(11, [1, 1, 1, 1, 1, 1, 1, 1])
+ 0
+ >>> minimum_subarray_sum(10, [1, 2, 3, 4, 5, 6, 7])
+ 2
+ >>> minimum_subarray_sum(5, [1, 1, 1, 1, 1, 5])
+ 1
+ >>> minimum_subarray_sum(0, [])
+ 0
+ >>> minimum_subarray_sum(0, [1, 2, 3])
+ 1
+ >>> minimum_subarray_sum(10, [10, 20, 30])
+ 1
+ >>> minimum_subarray_sum(7, [1, 1, 1, 1, 1, 1, 10])
+ 1
+ >>> minimum_subarray_sum(6, [])
+ 0
+ >>> minimum_subarray_sum(2, [1, 2, 3])
+ 1
+ >>> minimum_subarray_sum(-6, [])
+ 0
+ >>> minimum_subarray_sum(-6, [3, 4, 5])
+ 1
+ >>> minimum_subarray_sum(8, None)
+ 0
+ >>> minimum_subarray_sum(2, "ABC")
+ Traceback (most recent call last):
+ ...
+ ValueError: numbers must be an iterable of integers
+ """
+ if not numbers:
+ return 0
+ if target == 0 and target in numbers:
+ return 0
+ if not isinstance(numbers, (list, tuple)) or not all(
+ isinstance(number, int) for number in numbers
+ ):
+ raise ValueError("numbers must be an iterable of integers")
+
+ left = right = curr_sum = 0
+ min_len = sys.maxsize
+
+ while right < len(numbers):
+ curr_sum += numbers[right]
+ while curr_sum >= target and left <= right:
+ min_len = min(min_len, right - left + 1)
+ curr_sum -= numbers[left]
+ left += 1
+ right += 1
+
+ return 0 if min_len == sys.maxsize else min_len
diff --git a/dynamic_programming/minimum_squares_to_represent_a_number.py b/dynamic_programming/minimum_squares_to_represent_a_number.py
new file mode 100644
index 000000000000..98c0602fa831
--- /dev/null
+++ b/dynamic_programming/minimum_squares_to_represent_a_number.py
@@ -0,0 +1,49 @@
+import math
+import sys
+
+
+def minimum_squares_to_represent_a_number(number: int) -> int:
+ """
+ Count the number of minimum squares to represent a number
+
+ >>> minimum_squares_to_represent_a_number(25)
+ 1
+ >>> minimum_squares_to_represent_a_number(37)
+ 2
+ >>> minimum_squares_to_represent_a_number(21)
+ 3
+ >>> minimum_squares_to_represent_a_number(58)
+ 2
+ >>> minimum_squares_to_represent_a_number(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: the value of input must not be a negative number
+ >>> minimum_squares_to_represent_a_number(0)
+ 1
+ >>> minimum_squares_to_represent_a_number(12.34)
+ Traceback (most recent call last):
+ ...
+ ValueError: the value of input must be a natural number
+ """
+ if number != int(number):
+ raise ValueError("the value of input must be a natural number")
+ if number < 0:
+ raise ValueError("the value of input must not be a negative number")
+ if number == 0:
+ return 1
+ answers = [-1] * (number + 1)
+ answers[0] = 0
+ for i in range(1, number + 1):
+ answer = sys.maxsize
+ root = int(math.sqrt(i))
+ for j in range(1, root + 1):
+ current_answer = 1 + answers[i - (j**2)]
+ answer = min(answer, current_answer)
+ answers[i] = answer
+ return answers[number]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/minimum_steps_to_one.py b/dynamic_programming/minimum_steps_to_one.py
new file mode 100644
index 000000000000..68eaf56e21a7
--- /dev/null
+++ b/dynamic_programming/minimum_steps_to_one.py
@@ -0,0 +1,66 @@
+"""
+YouTube Explanation: https://www.youtube.com/watch?v=f2xi3c1S95M
+
+Given an integer n, return the minimum steps from n to 1
+
+AVAILABLE STEPS:
+ * Decrement by 1
+ * if n is divisible by 2, divide by 2
+ * if n is divisible by 3, divide by 3
+
+
+Example 1: n = 10
+10 -> 9 -> 3 -> 1
+Result: 3 steps
+
+Example 2: n = 15
+15 -> 5 -> 4 -> 2 -> 1
+Result: 4 steps
+
+Example 3: n = 6
+6 -> 2 -> 1
+Result: 2 step
+"""
+
+from __future__ import annotations
+
+__author__ = "Alexander Joslin"
+
+
+def min_steps_to_one(number: int) -> int:
+ """
+ Minimum steps to 1 implemented using tabulation.
+ >>> min_steps_to_one(10)
+ 3
+ >>> min_steps_to_one(15)
+ 4
+ >>> min_steps_to_one(6)
+ 2
+
+ :param number:
+ :return int:
+ """
+
+ if number <= 0:
+ msg = f"n must be greater than 0. Got n = {number}"
+ raise ValueError(msg)
+
+ table = [number + 1] * (number + 1)
+
+ # starting position
+ table[1] = 0
+ for i in range(1, number):
+ table[i + 1] = min(table[i + 1], table[i] + 1)
+ # check if out of bounds
+ if i * 2 <= number:
+ table[i * 2] = min(table[i * 2], table[i] + 1)
+ # check if out of bounds
+ if i * 3 <= number:
+ table[i * 3] = min(table[i * 3], table[i] + 1)
+ return table[number]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/minimum_tickets_cost.py b/dynamic_programming/minimum_tickets_cost.py
new file mode 100644
index 000000000000..6790c21f16ed
--- /dev/null
+++ b/dynamic_programming/minimum_tickets_cost.py
@@ -0,0 +1,129 @@
+"""
+Author : Alexander Pantyukhin
+Date : November 1, 2022
+
+Task:
+Given a list of days when you need to travel. Each day is integer from 1 to 365.
+You are able to use tickets for 1 day, 7 days and 30 days.
+Each ticket has a cost.
+
+Find the minimum cost you need to travel every day in the given list of days.
+
+Implementation notes:
+implementation Dynamic Programming up bottom approach.
+
+Runtime complexity: O(n)
+
+The implementation was tested on the
+leetcode: https://leetcode.com/problems/minimum-cost-for-tickets/
+
+
+Minimum Cost For Tickets
+Dynamic Programming: up -> down.
+"""
+
+import functools
+
+
+def mincost_tickets(days: list[int], costs: list[int]) -> int:
+ """
+ >>> mincost_tickets([1, 4, 6, 7, 8, 20], [2, 7, 15])
+ 11
+
+ >>> mincost_tickets([1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [2, 7, 15])
+ 17
+
+ >>> mincost_tickets([1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [2, 90, 150])
+ 24
+
+ >>> mincost_tickets([2], [2, 90, 150])
+ 2
+
+ >>> mincost_tickets([], [2, 90, 150])
+ 0
+
+ >>> mincost_tickets('hello', [2, 90, 150])
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter days should be a list of integers
+
+ >>> mincost_tickets([], 'world')
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter costs should be a list of three integers
+
+ >>> mincost_tickets([0.25, 2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [2, 90, 150])
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter days should be a list of integers
+
+ >>> mincost_tickets([1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [2, 0.9, 150])
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter costs should be a list of three integers
+
+ >>> mincost_tickets([-1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [2, 90, 150])
+ Traceback (most recent call last):
+ ...
+ ValueError: All days elements should be greater than 0
+
+ >>> mincost_tickets([2, 367], [2, 90, 150])
+ Traceback (most recent call last):
+ ...
+ ValueError: All days elements should be less than 366
+
+ >>> mincost_tickets([2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [])
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter costs should be a list of three integers
+
+ >>> mincost_tickets([], [])
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter costs should be a list of three integers
+
+ >>> mincost_tickets([2, 3, 4, 5, 6, 7, 8, 9, 10, 30, 31], [1, 2, 3, 4])
+ Traceback (most recent call last):
+ ...
+ ValueError: The parameter costs should be a list of three integers
+ """
+
+ # Validation
+ if not isinstance(days, list) or not all(isinstance(day, int) for day in days):
+ raise ValueError("The parameter days should be a list of integers")
+
+ if len(costs) != 3 or not all(isinstance(cost, int) for cost in costs):
+ raise ValueError("The parameter costs should be a list of three integers")
+
+ if len(days) == 0:
+ return 0
+
+ if min(days) <= 0:
+ raise ValueError("All days elements should be greater than 0")
+
+ if max(days) >= 366:
+ raise ValueError("All days elements should be less than 366")
+
+ days_set = set(days)
+
+ @functools.cache
+ def dynamic_programming(index: int) -> int:
+ if index > 365:
+ return 0
+
+ if index not in days_set:
+ return dynamic_programming(index + 1)
+
+ return min(
+ costs[0] + dynamic_programming(index + 1),
+ costs[1] + dynamic_programming(index + 7),
+ costs[2] + dynamic_programming(index + 30),
+ )
+
+ return dynamic_programming(1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/optimal_binary_search_tree.py b/dynamic_programming/optimal_binary_search_tree.py
index 0d94c1b61d39..b4f1181ac11c 100644
--- a/dynamic_programming/optimal_binary_search_tree.py
+++ b/dynamic_programming/optimal_binary_search_tree.py
@@ -104,7 +104,7 @@ def find_optimal_binary_search_tree(nodes):
dp = [[freqs[i] if i == j else 0 for j in range(n)] for i in range(n)]
# sum[i][j] stores the sum of key frequencies between i and j inclusive in nodes
# array
- sum = [[freqs[i] if i == j else 0 for j in range(n)] for i in range(n)]
+ total = [[freqs[i] if i == j else 0 for j in range(n)] for i in range(n)]
# stores tree roots that will be used later for constructing binary search tree
root = [[i if i == j else 0 for j in range(n)] for i in range(n)]
@@ -113,14 +113,14 @@ def find_optimal_binary_search_tree(nodes):
j = i + interval_length - 1
dp[i][j] = sys.maxsize # set the value to "infinity"
- sum[i][j] = sum[i][j - 1] + freqs[j]
+ total[i][j] = total[i][j - 1] + freqs[j]
# Apply Knuth's optimization
# Loop without optimization: for r in range(i, j + 1):
for r in range(root[i][j - 1], root[i + 1][j] + 1): # r is a temporal root
left = dp[i][r - 1] if r != i else 0 # optimal cost for left subtree
right = dp[r + 1][j] if r != j else 0 # optimal cost for right subtree
- cost = left + sum[i][j] + right
+ cost = left + total[i][j] + right
if dp[i][j] > cost:
dp[i][j] = cost
diff --git a/dynamic_programming/palindrome_partitioning.py b/dynamic_programming/palindrome_partitioning.py
new file mode 100644
index 000000000000..c1629440ef2e
--- /dev/null
+++ b/dynamic_programming/palindrome_partitioning.py
@@ -0,0 +1,39 @@
+"""
+Given a string s, partition s such that every substring of the
+partition is a palindrome.
+Find the minimum cuts needed for a palindrome partitioning of s.
+
+Time Complexity: O(n^2)
+Space Complexity: O(n^2)
+For other explanations refer to: https://www.youtube.com/watch?v=_H8V5hJUGd0
+"""
+
+
+def find_minimum_partitions(string: str) -> int:
+ """
+ Returns the minimum cuts needed for a palindrome partitioning of string
+
+ >>> find_minimum_partitions("aab")
+ 1
+ >>> find_minimum_partitions("aaa")
+ 0
+ >>> find_minimum_partitions("ababbbabbababa")
+ 3
+ """
+ length = len(string)
+ cut = [0] * length
+ is_palindromic = [[False for i in range(length)] for j in range(length)]
+ for i, c in enumerate(string):
+ mincut = i
+ for j in range(i + 1):
+ if c == string[j] and (i - j < 2 or is_palindromic[j + 1][i - 1]):
+ is_palindromic[j][i] = True
+ mincut = min(mincut, 0 if j == 0 else (cut[j - 1] + 1))
+ cut[i] = mincut
+ return cut[length - 1]
+
+
+if __name__ == "__main__":
+ s = input("Enter the string: ").strip()
+ ans = find_minimum_partitions(s)
+ print(f"Minimum number of partitions required for the '{s}' is {ans}")
diff --git a/dynamic_programming/range_sum_query.py b/dynamic_programming/range_sum_query.py
new file mode 100644
index 000000000000..484fcf785fda
--- /dev/null
+++ b/dynamic_programming/range_sum_query.py
@@ -0,0 +1,92 @@
+"""
+Author: Sanjay Muthu
+
+This is an implementation of the Dynamic Programming solution to the Range Sum Query.
+
+The problem statement is:
+ Given an array and q queries,
+ each query stating you to find the sum of elements from l to r (inclusive)
+
+Example:
+ arr = [1, 4, 6, 2, 61, 12]
+ queries = 3
+ l_1 = 2, r_1 = 5
+ l_2 = 1, r_2 = 5
+ l_3 = 3, r_3 = 4
+
+ as input will return
+
+ [81, 85, 63]
+
+ as output
+
+0-indexing:
+NOTE: 0-indexing means the indexing of the array starts from 0
+Example: a = [1, 2, 3, 4, 5, 6]
+ Here, the 0th index of a is 1,
+ the 1st index of a is 2,
+ and so forth
+
+Time Complexity: O(N + Q)
+* O(N) pre-calculation time to calculate the prefix sum array
+* and O(1) time per each query = O(1 * Q) = O(Q) time
+
+Space Complexity: O(N)
+* O(N) to store the prefix sum
+
+Algorithm:
+So, first we calculate the prefix sum (dp) of the array.
+The prefix sum of the index i is the sum of all elements indexed
+from 0 to i (inclusive).
+The prefix sum of the index i is the prefix sum of index (i - 1) + the current element.
+So, the state of the dp is dp[i] = dp[i - 1] + a[i].
+
+After we calculate the prefix sum,
+for each query [l, r]
+the answer is dp[r] - dp[l - 1] (we need to be careful because l might be 0).
+For example take this array:
+ [4, 2, 1, 6, 3]
+The prefix sum calculated for this array would be:
+ [4, 4 + 2, 4 + 2 + 1, 4 + 2 + 1 + 6, 4 + 2 + 1 + 6 + 3]
+ ==> [4, 6, 7, 13, 16]
+If the query was l = 3, r = 4,
+the answer would be 6 + 3 = 9 but this would require O(r - l + 1) time ≈ O(N) time
+
+If we use prefix sums we can find it in O(1) by using the formula
+prefix[r] - prefix[l - 1].
+This formula works because prefix[r] is the sum of elements from [0, r]
+and prefix[l - 1] is the sum of elements from [0, l - 1],
+so if we do prefix[r] - prefix[l - 1] it will be
+[0, r] - [0, l - 1] = [0, l - 1] + [l, r] - [0, l - 1] = [l, r]
+"""
+
+
+def prefix_sum(array: list[int], queries: list[tuple[int, int]]) -> list[int]:
+ """
+ >>> prefix_sum([1, 4, 6, 2, 61, 12], [(2, 5), (1, 5), (3, 4)])
+ [81, 85, 63]
+ >>> prefix_sum([4, 2, 1, 6, 3], [(3, 4), (1, 3), (0, 2)])
+ [9, 9, 7]
+ """
+ # The prefix sum array
+ dp = [0] * len(array)
+ dp[0] = array[0]
+ for i in range(1, len(array)):
+ dp[i] = dp[i - 1] + array[i]
+
+ # See Algorithm section (Line 44)
+ result = []
+ for query in queries:
+ left, right = query
+ res = dp[right]
+ if left > 0:
+ res -= dp[left - 1]
+ result.append(res)
+
+ return result
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/regex_match.py b/dynamic_programming/regex_match.py
new file mode 100644
index 000000000000..e94d82093c8b
--- /dev/null
+++ b/dynamic_programming/regex_match.py
@@ -0,0 +1,99 @@
+"""
+Regex matching check if a text matches pattern or not.
+Pattern:
+
+ 1. ``.`` Matches any single character.
+ 2. ``*`` Matches zero or more of the preceding element.
+
+More info:
+ https://medium.com/trick-the-interviwer/regular-expression-matching-9972eb74c03
+"""
+
+
+def recursive_match(text: str, pattern: str) -> bool:
+ r"""
+ Recursive matching algorithm.
+
+ | Time complexity: O(2^(\|text\| + \|pattern\|))
+ | Space complexity: Recursion depth is O(\|text\| + \|pattern\|).
+
+ :param text: Text to match.
+ :param pattern: Pattern to match.
+ :return: ``True`` if `text` matches `pattern`, ``False`` otherwise.
+
+ >>> recursive_match('abc', 'a.c')
+ True
+ >>> recursive_match('abc', 'af*.c')
+ True
+ >>> recursive_match('abc', 'a.c*')
+ True
+ >>> recursive_match('abc', 'a.c*d')
+ False
+ >>> recursive_match('aa', '.*')
+ True
+ """
+ if not pattern:
+ return not text
+
+ if not text:
+ return pattern[-1] == "*" and recursive_match(text, pattern[:-2])
+
+ if text[-1] == pattern[-1] or pattern[-1] == ".":
+ return recursive_match(text[:-1], pattern[:-1])
+
+ if pattern[-1] == "*":
+ return recursive_match(text[:-1], pattern) or recursive_match(
+ text, pattern[:-2]
+ )
+
+ return False
+
+
+def dp_match(text: str, pattern: str) -> bool:
+ r"""
+ Dynamic programming matching algorithm.
+
+ | Time complexity: O(\|text\| * \|pattern\|)
+ | Space complexity: O(\|text\| * \|pattern\|)
+
+ :param text: Text to match.
+ :param pattern: Pattern to match.
+ :return: ``True`` if `text` matches `pattern`, ``False`` otherwise.
+
+ >>> dp_match('abc', 'a.c')
+ True
+ >>> dp_match('abc', 'af*.c')
+ True
+ >>> dp_match('abc', 'a.c*')
+ True
+ >>> dp_match('abc', 'a.c*d')
+ False
+ >>> dp_match('aa', '.*')
+ True
+ """
+ m = len(text)
+ n = len(pattern)
+ dp = [[False for _ in range(n + 1)] for _ in range(m + 1)]
+ dp[0][0] = True
+
+ for j in range(1, n + 1):
+ dp[0][j] = pattern[j - 1] == "*" and dp[0][j - 2]
+
+ for i in range(1, m + 1):
+ for j in range(1, n + 1):
+ if pattern[j - 1] in {".", text[i - 1]}:
+ dp[i][j] = dp[i - 1][j - 1]
+ elif pattern[j - 1] == "*":
+ dp[i][j] = dp[i][j - 2]
+ if pattern[j - 2] in {".", text[i - 1]}:
+ dp[i][j] |= dp[i - 1][j]
+ else:
+ dp[i][j] = False
+
+ return dp[m][n]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/rod_cutting.py b/dynamic_programming/rod_cutting.py
index 442a39cb1616..d12c759dc928 100644
--- a/dynamic_programming/rod_cutting.py
+++ b/dynamic_programming/rod_cutting.py
@@ -1,7 +1,7 @@
"""
This module provides two implementations for the rod-cutting problem:
-1. A naive recursive implementation which has an exponential runtime
-2. Two dynamic programming implementations which have quadratic runtime
+ 1. A naive recursive implementation which has an exponential runtime
+ 2. Two dynamic programming implementations which have quadratic runtime
The rod-cutting problem is the problem of finding the maximum possible revenue
obtainable from a rod of length ``n`` given a list of prices for each integral piece
@@ -20,18 +20,21 @@ def naive_cut_rod_recursive(n: int, prices: list):
Runtime: O(2^n)
Arguments
- -------
- n: int, the length of the rod
- prices: list, the prices for each piece of rod. ``p[i-i]`` is the
- price for a rod of length ``i``
+ ---------
+
+ * `n`: int, the length of the rod
+ * `prices`: list, the prices for each piece of rod. ``p[i-i]`` is the
+ price for a rod of length ``i``
Returns
-------
- The maximum revenue obtainable for a rod of length n given the list of prices
+
+ The maximum revenue obtainable for a rod of length `n` given the list of prices
for each piece.
Examples
--------
+
>>> naive_cut_rod_recursive(4, [1, 5, 8, 9])
10
>>> naive_cut_rod_recursive(10, [1, 5, 8, 9, 10, 17, 17, 20, 24, 30])
@@ -54,28 +57,30 @@ def top_down_cut_rod(n: int, prices: list):
"""
Constructs a top-down dynamic programming solution for the rod-cutting
problem via memoization. This function serves as a wrapper for
- _top_down_cut_rod_recursive
+ ``_top_down_cut_rod_recursive``
Runtime: O(n^2)
Arguments
- --------
- n: int, the length of the rod
- prices: list, the prices for each piece of rod. ``p[i-i]`` is the
- price for a rod of length ``i``
+ ---------
- Note
- ----
- For convenience and because Python's lists using 0-indexing, length(max_rev) =
- n + 1, to accommodate for the revenue obtainable from a rod of length 0.
+ * `n`: int, the length of the rod
+ * `prices`: list, the prices for each piece of rod. ``p[i-i]`` is the
+ price for a rod of length ``i``
+
+ .. note::
+ For convenience and because Python's lists using ``0``-indexing, ``length(max_rev)
+ = n + 1``, to accommodate for the revenue obtainable from a rod of length ``0``.
Returns
-------
- The maximum revenue obtainable for a rod of length n given the list of prices
+
+ The maximum revenue obtainable for a rod of length `n` given the list of prices
for each piece.
Examples
- -------
+ --------
+
>>> top_down_cut_rod(4, [1, 5, 8, 9])
10
>>> top_down_cut_rod(10, [1, 5, 8, 9, 10, 17, 17, 20, 24, 30])
@@ -94,16 +99,18 @@ def _top_down_cut_rod_recursive(n: int, prices: list, max_rev: list):
Runtime: O(n^2)
Arguments
- --------
- n: int, the length of the rod
- prices: list, the prices for each piece of rod. ``p[i-i]`` is the
- price for a rod of length ``i``
- max_rev: list, the computed maximum revenue for a piece of rod.
- ``max_rev[i]`` is the maximum revenue obtainable for a rod of length ``i``
+ ---------
+
+ * `n`: int, the length of the rod
+ * `prices`: list, the prices for each piece of rod. ``p[i-i]`` is the
+ price for a rod of length ``i``
+ * `max_rev`: list, the computed maximum revenue for a piece of rod.
+ ``max_rev[i]`` is the maximum revenue obtainable for a rod of length ``i``
Returns
-------
- The maximum revenue obtainable for a rod of length n given the list of prices
+
+ The maximum revenue obtainable for a rod of length `n` given the list of prices
for each piece.
"""
if max_rev[n] >= 0:
@@ -130,18 +137,21 @@ def bottom_up_cut_rod(n: int, prices: list):
Runtime: O(n^2)
Arguments
- ----------
- n: int, the maximum length of the rod.
- prices: list, the prices for each piece of rod. ``p[i-i]`` is the
- price for a rod of length ``i``
+ ---------
+
+ * `n`: int, the maximum length of the rod.
+ * `prices`: list, the prices for each piece of rod. ``p[i-i]`` is the
+ price for a rod of length ``i``
Returns
-------
- The maximum revenue obtainable from cutting a rod of length n given
+
+ The maximum revenue obtainable from cutting a rod of length `n` given
the prices for each piece of rod p.
Examples
- -------
+ --------
+
>>> bottom_up_cut_rod(4, [1, 5, 8, 9])
10
>>> bottom_up_cut_rod(10, [1, 5, 8, 9, 10, 17, 17, 20, 24, 30])
@@ -168,22 +178,23 @@ def _enforce_args(n: int, prices: list):
"""
Basic checks on the arguments to the rod-cutting algorithms
- n: int, the length of the rod
- prices: list, the price list for each piece of rod.
-
- Throws ValueError:
+ * `n`: int, the length of the rod
+ * `prices`: list, the price list for each piece of rod.
- if n is negative or there are fewer items in the price list than the length of
- the rod
+ Throws ``ValueError``:
+ if `n` is negative or there are fewer items in the price list than the length of
+ the rod
"""
if n < 0:
- raise ValueError(f"n must be greater than or equal to 0. Got n = {n}")
+ msg = f"n must be greater than or equal to 0. Got n = {n}"
+ raise ValueError(msg)
if n > len(prices):
- raise ValueError(
- f"Each integral piece of rod must have a corresponding "
- f"price. Got n = {n} but length of prices = {len(prices)}"
+ msg = (
+ "Each integral piece of rod must have a corresponding price. "
+ f"Got n = {n} but length of prices = {len(prices)}"
)
+ raise ValueError(msg)
def main():
diff --git a/dynamic_programming/smith_waterman.py b/dynamic_programming/smith_waterman.py
new file mode 100644
index 000000000000..4c5d58379f07
--- /dev/null
+++ b/dynamic_programming/smith_waterman.py
@@ -0,0 +1,193 @@
+"""
+https://en.wikipedia.org/wiki/Smith%E2%80%93Waterman_algorithm
+The Smith-Waterman algorithm is a dynamic programming algorithm used for sequence
+alignment. It is particularly useful for finding similarities between two sequences,
+such as DNA or protein sequences. In this implementation, gaps are penalized
+linearly, meaning that the score is reduced by a fixed amount for each gap introduced
+in the alignment. However, it's important to note that the Smith-Waterman algorithm
+supports other gap penalty methods as well.
+"""
+
+
+def score_function(
+ source_char: str,
+ target_char: str,
+ match: int = 1,
+ mismatch: int = -1,
+ gap: int = -2,
+) -> int:
+ """
+ Calculate the score for a character pair based on whether they match or mismatch.
+ Returns 1 if the characters match, -1 if they mismatch, and -2 if either of the
+ characters is a gap.
+ >>> score_function('A', 'A')
+ 1
+ >>> score_function('A', 'C')
+ -1
+ >>> score_function('-', 'A')
+ -2
+ >>> score_function('A', '-')
+ -2
+ >>> score_function('-', '-')
+ -2
+ """
+ if "-" in (source_char, target_char):
+ return gap
+ return match if source_char == target_char else mismatch
+
+
+def smith_waterman(
+ query: str,
+ subject: str,
+ match: int = 1,
+ mismatch: int = -1,
+ gap: int = -2,
+) -> list[list[int]]:
+ """
+ Perform the Smith-Waterman local sequence alignment algorithm.
+ Returns a 2D list representing the score matrix. Each value in the matrix
+ corresponds to the score of the best local alignment ending at that point.
+ >>> smith_waterman('ACAC', 'CA')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+ >>> smith_waterman('acac', 'ca')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+ >>> smith_waterman('ACAC', 'ca')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+ >>> smith_waterman('acac', 'CA')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+ >>> smith_waterman('ACAC', '')
+ [[0], [0], [0], [0], [0]]
+ >>> smith_waterman('', 'CA')
+ [[0, 0, 0]]
+ >>> smith_waterman('ACAC', 'CA')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+
+ >>> smith_waterman('acac', 'ca')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+
+ >>> smith_waterman('ACAC', 'ca')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+
+ >>> smith_waterman('acac', 'CA')
+ [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]]
+
+ >>> smith_waterman('ACAC', '')
+ [[0], [0], [0], [0], [0]]
+
+ >>> smith_waterman('', 'CA')
+ [[0, 0, 0]]
+
+ >>> smith_waterman('AGT', 'AGT')
+ [[0, 0, 0, 0], [0, 1, 0, 0], [0, 0, 2, 0], [0, 0, 0, 3]]
+
+ >>> smith_waterman('AGT', 'GTA')
+ [[0, 0, 0, 0], [0, 0, 0, 1], [0, 1, 0, 0], [0, 0, 2, 0]]
+
+ >>> smith_waterman('AGT', 'GTC')
+ [[0, 0, 0, 0], [0, 0, 0, 0], [0, 1, 0, 0], [0, 0, 2, 0]]
+
+ >>> smith_waterman('AGT', 'G')
+ [[0, 0], [0, 0], [0, 1], [0, 0]]
+
+ >>> smith_waterman('G', 'AGT')
+ [[0, 0, 0, 0], [0, 0, 1, 0]]
+
+ >>> smith_waterman('AGT', 'AGTCT')
+ [[0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], [0, 0, 2, 0, 0, 0], [0, 0, 0, 3, 1, 1]]
+
+ >>> smith_waterman('AGTCT', 'AGT')
+ [[0, 0, 0, 0], [0, 1, 0, 0], [0, 0, 2, 0], [0, 0, 0, 3], [0, 0, 0, 1], [0, 0, 0, 1]]
+
+ >>> smith_waterman('AGTCT', 'GTC')
+ [[0, 0, 0, 0], [0, 0, 0, 0], [0, 1, 0, 0], [0, 0, 2, 0], [0, 0, 0, 3], [0, 0, 1, 1]]
+ """
+ # make both query and subject uppercase
+ query = query.upper()
+ subject = subject.upper()
+
+ # Initialize score matrix
+ m = len(query)
+ n = len(subject)
+ score = [[0] * (n + 1) for _ in range(m + 1)]
+ kwargs = {"match": match, "mismatch": mismatch, "gap": gap}
+
+ for i in range(1, m + 1):
+ for j in range(1, n + 1):
+ # Calculate scores for each cell
+ match = score[i - 1][j - 1] + score_function(
+ query[i - 1], subject[j - 1], **kwargs
+ )
+ delete = score[i - 1][j] + gap
+ insert = score[i][j - 1] + gap
+
+ # Take maximum score
+ score[i][j] = max(0, match, delete, insert)
+
+ return score
+
+
+def traceback(score: list[list[int]], query: str, subject: str) -> str:
+ r"""
+ Perform traceback to find the optimal local alignment.
+ Starts from the highest scoring cell in the matrix and traces back recursively
+ until a 0 score is found. Returns the alignment strings.
+ >>> traceback([[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]], 'ACAC', 'CA')
+ 'CA\nCA'
+ >>> traceback([[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]], 'acac', 'ca')
+ 'CA\nCA'
+ >>> traceback([[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]], 'ACAC', 'ca')
+ 'CA\nCA'
+ >>> traceback([[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 0, 2], [0, 1, 0]], 'acac', 'CA')
+ 'CA\nCA'
+ >>> traceback([[0, 0, 0]], 'ACAC', '')
+ ''
+ """
+ # make both query and subject uppercase
+ query = query.upper()
+ subject = subject.upper()
+ # find the indices of the maximum value in the score matrix
+ max_value = float("-inf")
+ i_max = j_max = 0
+ for i, row in enumerate(score):
+ for j, value in enumerate(row):
+ if value > max_value:
+ max_value = value
+ i_max, j_max = i, j
+ # Traceback logic to find optimal alignment
+ i = i_max
+ j = j_max
+ align1 = ""
+ align2 = ""
+ gap = score_function("-", "-")
+ # guard against empty query or subject
+ if i == 0 or j == 0:
+ return ""
+ while i > 0 and j > 0:
+ if score[i][j] == score[i - 1][j - 1] + score_function(
+ query[i - 1], subject[j - 1]
+ ):
+ # optimal path is a diagonal take both letters
+ align1 = query[i - 1] + align1
+ align2 = subject[j - 1] + align2
+ i -= 1
+ j -= 1
+ elif score[i][j] == score[i - 1][j] + gap:
+ # optimal path is a vertical
+ align1 = query[i - 1] + align1
+ align2 = f"-{align2}"
+ i -= 1
+ else:
+ # optimal path is a horizontal
+ align1 = f"-{align1}"
+ align2 = subject[j - 1] + align2
+ j -= 1
+
+ return f"{align1}\n{align2}"
+
+
+if __name__ == "__main__":
+ query = "HEAGAWGHEE"
+ subject = "PAWHEAE"
+
+ score = smith_waterman(query, subject, match=1, mismatch=-1, gap=-2)
+ print(traceback(score, query, subject))
diff --git a/dynamic_programming/subset_generation.py b/dynamic_programming/subset_generation.py
index 4781b23b32eb..08daaac6f88a 100644
--- a/dynamic_programming/subset_generation.py
+++ b/dynamic_programming/subset_generation.py
@@ -1,43 +1,63 @@
-# Print all subset combinations of n element in given set of r element.
+def subset_combinations(elements: list[int], n: int) -> list:
+ """
+ Compute n-element combinations from a given list using dynamic programming.
+ Args:
+ * `elements`: The list of elements from which combinations will be generated.
+ * `n`: The number of elements in each combination.
-def combination_util(arr, n, r, index, data, i):
- """
- Current combination is ready to be printed, print it
- arr[] ---> Input Array
- data[] ---> Temporary array to store current combination
- start & end ---> Staring and Ending indexes in arr[]
- index ---> Current index in data[]
- r ---> Size of a combination to be printed
+ Returns:
+ A list of tuples, each representing a combination of `n` elements.
+
+ >>> subset_combinations(elements=[10, 20, 30, 40], n=2)
+ [(10, 20), (10, 30), (10, 40), (20, 30), (20, 40), (30, 40)]
+ >>> subset_combinations(elements=[1, 2, 3], n=1)
+ [(1,), (2,), (3,)]
+ >>> subset_combinations(elements=[1, 2, 3], n=3)
+ [(1, 2, 3)]
+ >>> subset_combinations(elements=[42], n=1)
+ [(42,)]
+ >>> subset_combinations(elements=[6, 7, 8, 9], n=4)
+ [(6, 7, 8, 9)]
+ >>> subset_combinations(elements=[10, 20, 30, 40, 50], n=0)
+ [()]
+ >>> subset_combinations(elements=[1, 2, 3, 4], n=2)
+ [(1, 2), (1, 3), (1, 4), (2, 3), (2, 4), (3, 4)]
+ >>> subset_combinations(elements=[1, 'apple', 3.14], n=2)
+ [(1, 'apple'), (1, 3.14), ('apple', 3.14)]
+ >>> subset_combinations(elements=['single'], n=0)
+ [()]
+ >>> subset_combinations(elements=[], n=9)
+ []
+ >>> from itertools import combinations
+ >>> all(subset_combinations(items, n) == list(combinations(items, n))
+ ... for items, n in (
+ ... ([10, 20, 30, 40], 2), ([1, 2, 3], 1), ([1, 2, 3], 3), ([42], 1),
+ ... ([6, 7, 8, 9], 4), ([10, 20, 30, 40, 50], 1), ([1, 2, 3, 4], 2),
+ ... ([1, 'apple', 3.14], 2), (['single'], 0), ([], 9)))
+ True
"""
- if index == r:
- for j in range(r):
- print(data[j], end=" ")
- print(" ")
- return
- # When no more elements are there to put in data[]
- if i >= n:
- return
- # current is included, put next at next location
- data[index] = arr[i]
- combination_util(arr, n, r, index + 1, data, i + 1)
- # current is excluded, replace it with
- # next (Note that i+1 is passed, but
- # index is not changed)
- combination_util(arr, n, r, index, data, i + 1)
- # The main function that prints all combinations
- # of size r in arr[] of size n. This function
- # mainly uses combinationUtil()
-
-
-def print_combination(arr, n, r):
- # A temporary array to store all combination one by one
- data = [0] * r
- # Print all combination using temporary array 'data[]'
- combination_util(arr, n, r, 0, data, 0)
-
-
-# Driver function to check for above function
-arr = [10, 20, 30, 40, 50]
-print_combination(arr, len(arr), 3)
-# This code is contributed by Ambuj sahu
+ r = len(elements)
+ if n > r:
+ return []
+
+ dp: list[list[tuple]] = [[] for _ in range(r + 1)]
+
+ dp[0].append(())
+
+ for i in range(1, r + 1):
+ for j in range(i, 0, -1):
+ for prev_combination in dp[j - 1]:
+ dp[j].append((*prev_combination, elements[i - 1]))
+
+ try:
+ return sorted(dp[n])
+ except TypeError:
+ return dp[n]
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ print(f"{subset_combinations(elements=[10, 20, 30, 40], n=2) = }")
diff --git a/dynamic_programming/sum_of_subset.py b/dynamic_programming/sum_of_subset.py
index a12177b57c74..96ebcf583a4b 100644
--- a/dynamic_programming/sum_of_subset.py
+++ b/dynamic_programming/sum_of_subset.py
@@ -1,34 +1,32 @@
-def isSumSubset(arr, arrLen, requiredSum):
+def is_sum_subset(arr: list[int], required_sum: int) -> bool:
"""
- >>> isSumSubset([2, 4, 6, 8], 4, 5)
+ >>> is_sum_subset([2, 4, 6, 8], 5)
False
- >>> isSumSubset([2, 4, 6, 8], 4, 14)
+ >>> is_sum_subset([2, 4, 6, 8], 14)
True
"""
# a subset value says 1 if that subset sum can be formed else 0
# initially no subsets can be formed hence False/0
- subset = [[False for i in range(requiredSum + 1)] for i in range(arrLen + 1)]
+ arr_len = len(arr)
+ subset = [[False] * (required_sum + 1) for _ in range(arr_len + 1)]
# for each arr value, a sum of zero(0) can be formed by not taking any element
# hence True/1
- for i in range(arrLen + 1):
+ for i in range(arr_len + 1):
subset[i][0] = True
# sum is not zero and set is empty then false
- for i in range(1, requiredSum + 1):
+ for i in range(1, required_sum + 1):
subset[0][i] = False
- for i in range(1, arrLen + 1):
- for j in range(1, requiredSum + 1):
+ for i in range(1, arr_len + 1):
+ for j in range(1, required_sum + 1):
if arr[i - 1] > j:
subset[i][j] = subset[i - 1][j]
if arr[i - 1] <= j:
subset[i][j] = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]]
- # uncomment to print the subset
- # for i in range(arrLen+1):
- # print(subset[i])
- print(subset[arrLen][requiredSum])
+ return subset[arr_len][required_sum]
if __name__ == "__main__":
diff --git a/dynamic_programming/trapped_water.py b/dynamic_programming/trapped_water.py
new file mode 100644
index 000000000000..8bec9fac5fef
--- /dev/null
+++ b/dynamic_programming/trapped_water.py
@@ -0,0 +1,60 @@
+"""
+Given an array of non-negative integers representing an elevation map where the width
+of each bar is 1, this program calculates how much rainwater can be trapped.
+
+Example - height = (0, 1, 0, 2, 1, 0, 1, 3, 2, 1, 2, 1)
+Output: 6
+This problem can be solved using the concept of "DYNAMIC PROGRAMMING".
+
+We calculate the maximum height of bars on the left and right of every bar in array.
+Then iterate over the width of structure and at each index.
+The amount of water that will be stored is equal to minimum of maximum height of bars
+on both sides minus height of bar at current position.
+"""
+
+
+def trapped_rainwater(heights: tuple[int, ...]) -> int:
+ """
+ The trapped_rainwater function calculates the total amount of rainwater that can be
+ trapped given an array of bar heights.
+ It uses a dynamic programming approach, determining the maximum height of bars on
+ both sides for each bar, and then computing the trapped water above each bar.
+ The function returns the total trapped water.
+
+ >>> trapped_rainwater((0, 1, 0, 2, 1, 0, 1, 3, 2, 1, 2, 1))
+ 6
+ >>> trapped_rainwater((7, 1, 5, 3, 6, 4))
+ 9
+ >>> trapped_rainwater((7, 1, 5, 3, 6, -1))
+ Traceback (most recent call last):
+ ...
+ ValueError: No height can be negative
+ """
+ if not heights:
+ return 0
+ if any(h < 0 for h in heights):
+ raise ValueError("No height can be negative")
+ length = len(heights)
+
+ left_max = [0] * length
+ left_max[0] = heights[0]
+ for i, height in enumerate(heights[1:], start=1):
+ left_max[i] = max(height, left_max[i - 1])
+
+ right_max = [0] * length
+ right_max[-1] = heights[-1]
+ for i in range(length - 2, -1, -1):
+ right_max[i] = max(heights[i], right_max[i + 1])
+
+ return sum(
+ min(left, right) - height
+ for left, right, height in zip(left_max, right_max, heights)
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(f"{trapped_rainwater((0, 1, 0, 2, 1, 0, 1, 3, 2, 1, 2, 1)) = }")
+ print(f"{trapped_rainwater((7, 1, 5, 3, 6, 4)) = }")
diff --git a/dynamic_programming/tribonacci.py b/dynamic_programming/tribonacci.py
new file mode 100644
index 000000000000..58e15da918e2
--- /dev/null
+++ b/dynamic_programming/tribonacci.py
@@ -0,0 +1,24 @@
+# Tribonacci sequence using Dynamic Programming
+
+
+def tribonacci(num: int) -> list[int]:
+ """
+ Given a number, return first n Tribonacci Numbers.
+ >>> tribonacci(5)
+ [0, 0, 1, 1, 2]
+ >>> tribonacci(8)
+ [0, 0, 1, 1, 2, 4, 7, 13]
+ """
+ dp = [0] * num
+ dp[2] = 1
+
+ for i in range(3, num):
+ dp[i] = dp[i - 1] + dp[i - 2] + dp[i - 3]
+
+ return dp
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/viterbi.py b/dynamic_programming/viterbi.py
new file mode 100644
index 000000000000..5b78fa9e46d0
--- /dev/null
+++ b/dynamic_programming/viterbi.py
@@ -0,0 +1,377 @@
+from typing import Any
+
+
+def viterbi(
+ observations_space: list,
+ states_space: list,
+ initial_probabilities: dict,
+ transition_probabilities: dict,
+ emission_probabilities: dict,
+) -> list:
+ """
+ Viterbi Algorithm, to find the most likely path of
+ states from the start and the expected output.
+
+ https://en.wikipedia.org/wiki/Viterbi_algorithm
+
+ Wikipedia example
+
+ >>> observations = ["normal", "cold", "dizzy"]
+ >>> states = ["Healthy", "Fever"]
+ >>> start_p = {"Healthy": 0.6, "Fever": 0.4}
+ >>> trans_p = {
+ ... "Healthy": {"Healthy": 0.7, "Fever": 0.3},
+ ... "Fever": {"Healthy": 0.4, "Fever": 0.6},
+ ... }
+ >>> emit_p = {
+ ... "Healthy": {"normal": 0.5, "cold": 0.4, "dizzy": 0.1},
+ ... "Fever": {"normal": 0.1, "cold": 0.3, "dizzy": 0.6},
+ ... }
+ >>> viterbi(observations, states, start_p, trans_p, emit_p)
+ ['Healthy', 'Healthy', 'Fever']
+ >>> viterbi((), states, start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ >>> viterbi(observations, (), start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ >>> viterbi(observations, states, {}, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ >>> viterbi(observations, states, start_p, {}, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ >>> viterbi(observations, states, start_p, trans_p, {})
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ >>> viterbi("invalid", states, start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: observations_space must be a list
+ >>> viterbi(["valid", 123], states, start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: observations_space must be a list of strings
+ >>> viterbi(observations, "invalid", start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: states_space must be a list
+ >>> viterbi(observations, ["valid", 123], start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: states_space must be a list of strings
+ >>> viterbi(observations, states, "invalid", trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: initial_probabilities must be a dict
+ >>> viterbi(observations, states, {2:2}, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: initial_probabilities all keys must be strings
+ >>> viterbi(observations, states, {"a":2}, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: initial_probabilities all values must be float
+ >>> viterbi(observations, states, start_p, "invalid", emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: transition_probabilities must be a dict
+ >>> viterbi(observations, states, start_p, {"a":2}, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: transition_probabilities all values must be dict
+ >>> viterbi(observations, states, start_p, {2:{2:2}}, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: transition_probabilities all keys must be strings
+ >>> viterbi(observations, states, start_p, {"a":{2:2}}, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: transition_probabilities all keys must be strings
+ >>> viterbi(observations, states, start_p, {"a":{"b":2}}, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: transition_probabilities nested dictionary all values must be float
+ >>> viterbi(observations, states, start_p, trans_p, "invalid")
+ Traceback (most recent call last):
+ ...
+ ValueError: emission_probabilities must be a dict
+ >>> viterbi(observations, states, start_p, trans_p, None)
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+
+ """
+ _validation(
+ observations_space,
+ states_space,
+ initial_probabilities,
+ transition_probabilities,
+ emission_probabilities,
+ )
+ # Creates data structures and fill initial step
+ probabilities: dict = {}
+ pointers: dict = {}
+ for state in states_space:
+ observation = observations_space[0]
+ probabilities[(state, observation)] = (
+ initial_probabilities[state] * emission_probabilities[state][observation]
+ )
+ pointers[(state, observation)] = None
+
+ # Fills the data structure with the probabilities of
+ # different transitions and pointers to previous states
+ for o in range(1, len(observations_space)):
+ observation = observations_space[o]
+ prior_observation = observations_space[o - 1]
+ for state in states_space:
+ # Calculates the argmax for probability function
+ arg_max = ""
+ max_probability = -1
+ for k_state in states_space:
+ probability = (
+ probabilities[(k_state, prior_observation)]
+ * transition_probabilities[k_state][state]
+ * emission_probabilities[state][observation]
+ )
+ if probability > max_probability:
+ max_probability = probability
+ arg_max = k_state
+
+ # Update probabilities and pointers dicts
+ probabilities[(state, observation)] = (
+ probabilities[(arg_max, prior_observation)]
+ * transition_probabilities[arg_max][state]
+ * emission_probabilities[state][observation]
+ )
+
+ pointers[(state, observation)] = arg_max
+
+ # The final observation
+ final_observation = observations_space[len(observations_space) - 1]
+
+ # argmax for given final observation
+ arg_max = ""
+ max_probability = -1
+ for k_state in states_space:
+ probability = probabilities[(k_state, final_observation)]
+ if probability > max_probability:
+ max_probability = probability
+ arg_max = k_state
+ last_state = arg_max
+
+ # Process pointers backwards
+ previous = last_state
+ result = []
+ for o in range(len(observations_space) - 1, -1, -1):
+ result.append(previous)
+ previous = pointers[previous, observations_space[o]]
+ result.reverse()
+
+ return result
+
+
+def _validation(
+ observations_space: Any,
+ states_space: Any,
+ initial_probabilities: Any,
+ transition_probabilities: Any,
+ emission_probabilities: Any,
+) -> None:
+ """
+ >>> observations = ["normal", "cold", "dizzy"]
+ >>> states = ["Healthy", "Fever"]
+ >>> start_p = {"Healthy": 0.6, "Fever": 0.4}
+ >>> trans_p = {
+ ... "Healthy": {"Healthy": 0.7, "Fever": 0.3},
+ ... "Fever": {"Healthy": 0.4, "Fever": 0.6},
+ ... }
+ >>> emit_p = {
+ ... "Healthy": {"normal": 0.5, "cold": 0.4, "dizzy": 0.1},
+ ... "Fever": {"normal": 0.1, "cold": 0.3, "dizzy": 0.6},
+ ... }
+ >>> _validation(observations, states, start_p, trans_p, emit_p)
+ >>> _validation([], states, start_p, trans_p, emit_p)
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ """
+ _validate_not_empty(
+ observations_space,
+ states_space,
+ initial_probabilities,
+ transition_probabilities,
+ emission_probabilities,
+ )
+ _validate_lists(observations_space, states_space)
+ _validate_dicts(
+ initial_probabilities, transition_probabilities, emission_probabilities
+ )
+
+
+def _validate_not_empty(
+ observations_space: Any,
+ states_space: Any,
+ initial_probabilities: Any,
+ transition_probabilities: Any,
+ emission_probabilities: Any,
+) -> None:
+ """
+ >>> _validate_not_empty(["a"], ["b"], {"c":0.5},
+ ... {"d": {"e": 0.6}}, {"f": {"g": 0.7}})
+ >>> _validate_not_empty(["a"], ["b"], {"c":0.5}, {}, {"f": {"g": 0.7}})
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ >>> _validate_not_empty(["a"], ["b"], None, {"d": {"e": 0.6}}, {"f": {"g": 0.7}})
+ Traceback (most recent call last):
+ ...
+ ValueError: There's an empty parameter
+ """
+ if not all(
+ [
+ observations_space,
+ states_space,
+ initial_probabilities,
+ transition_probabilities,
+ emission_probabilities,
+ ]
+ ):
+ raise ValueError("There's an empty parameter")
+
+
+def _validate_lists(observations_space: Any, states_space: Any) -> None:
+ """
+ >>> _validate_lists(["a"], ["b"])
+ >>> _validate_lists(1234, ["b"])
+ Traceback (most recent call last):
+ ...
+ ValueError: observations_space must be a list
+ >>> _validate_lists(["a"], [3])
+ Traceback (most recent call last):
+ ...
+ ValueError: states_space must be a list of strings
+ """
+ _validate_list(observations_space, "observations_space")
+ _validate_list(states_space, "states_space")
+
+
+def _validate_list(_object: Any, var_name: str) -> None:
+ """
+ >>> _validate_list(["a"], "mock_name")
+ >>> _validate_list("a", "mock_name")
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name must be a list
+ >>> _validate_list([0.5], "mock_name")
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name must be a list of strings
+ """
+ if not isinstance(_object, list):
+ msg = f"{var_name} must be a list"
+ raise ValueError(msg)
+ else:
+ for x in _object:
+ if not isinstance(x, str):
+ msg = f"{var_name} must be a list of strings"
+ raise ValueError(msg)
+
+
+def _validate_dicts(
+ initial_probabilities: Any,
+ transition_probabilities: Any,
+ emission_probabilities: Any,
+) -> None:
+ """
+ >>> _validate_dicts({"c":0.5}, {"d": {"e": 0.6}}, {"f": {"g": 0.7}})
+ >>> _validate_dicts("invalid", {"d": {"e": 0.6}}, {"f": {"g": 0.7}})
+ Traceback (most recent call last):
+ ...
+ ValueError: initial_probabilities must be a dict
+ >>> _validate_dicts({"c":0.5}, {2: {"e": 0.6}}, {"f": {"g": 0.7}})
+ Traceback (most recent call last):
+ ...
+ ValueError: transition_probabilities all keys must be strings
+ >>> _validate_dicts({"c":0.5}, {"d": {"e": 0.6}}, {"f": {2: 0.7}})
+ Traceback (most recent call last):
+ ...
+ ValueError: emission_probabilities all keys must be strings
+ >>> _validate_dicts({"c":0.5}, {"d": {"e": 0.6}}, {"f": {"g": "h"}})
+ Traceback (most recent call last):
+ ...
+ ValueError: emission_probabilities nested dictionary all values must be float
+ """
+ _validate_dict(initial_probabilities, "initial_probabilities", float)
+ _validate_nested_dict(transition_probabilities, "transition_probabilities")
+ _validate_nested_dict(emission_probabilities, "emission_probabilities")
+
+
+def _validate_nested_dict(_object: Any, var_name: str) -> None:
+ """
+ >>> _validate_nested_dict({"a":{"b": 0.5}}, "mock_name")
+ >>> _validate_nested_dict("invalid", "mock_name")
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name must be a dict
+ >>> _validate_nested_dict({"a": 8}, "mock_name")
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name all values must be dict
+ >>> _validate_nested_dict({"a":{2: 0.5}}, "mock_name")
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name all keys must be strings
+ >>> _validate_nested_dict({"a":{"b": 4}}, "mock_name")
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name nested dictionary all values must be float
+ """
+ _validate_dict(_object, var_name, dict)
+ for x in _object.values():
+ _validate_dict(x, var_name, float, True)
+
+
+def _validate_dict(
+ _object: Any, var_name: str, value_type: type, nested: bool = False
+) -> None:
+ """
+ >>> _validate_dict({"b": 0.5}, "mock_name", float)
+ >>> _validate_dict("invalid", "mock_name", float)
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name must be a dict
+ >>> _validate_dict({"a": 8}, "mock_name", dict)
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name all values must be dict
+ >>> _validate_dict({2: 0.5}, "mock_name",float, True)
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name all keys must be strings
+ >>> _validate_dict({"b": 4}, "mock_name", float,True)
+ Traceback (most recent call last):
+ ...
+ ValueError: mock_name nested dictionary all values must be float
+ """
+ if not isinstance(_object, dict):
+ msg = f"{var_name} must be a dict"
+ raise ValueError(msg)
+ if not all(isinstance(x, str) for x in _object):
+ msg = f"{var_name} all keys must be strings"
+ raise ValueError(msg)
+ if not all(isinstance(x, value_type) for x in _object.values()):
+ nested_text = "nested dictionary " if nested else ""
+ msg = f"{var_name} {nested_text}all values must be {value_type.__name__}"
+ raise ValueError(msg)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/dynamic_programming/wildcard_matching.py b/dynamic_programming/wildcard_matching.py
new file mode 100644
index 000000000000..d9a1392720bd
--- /dev/null
+++ b/dynamic_programming/wildcard_matching.py
@@ -0,0 +1,68 @@
+"""
+Author : ilyas dahhou
+Date : Oct 7, 2023
+
+Task:
+Given an input string and a pattern, implement wildcard pattern matching with support
+for '?' and '*' where:
+'?' matches any single character.
+'*' matches any sequence of characters (including the empty sequence).
+The matching should cover the entire input string (not partial).
+
+Runtime complexity: O(m * n)
+
+The implementation was tested on the
+leetcode: https://leetcode.com/problems/wildcard-matching/
+"""
+
+
+def is_match(string: str, pattern: str) -> bool:
+ """
+ >>> is_match("", "")
+ True
+ >>> is_match("aa", "a")
+ False
+ >>> is_match("abc", "abc")
+ True
+ >>> is_match("abc", "*c")
+ True
+ >>> is_match("abc", "a*")
+ True
+ >>> is_match("abc", "*a*")
+ True
+ >>> is_match("abc", "?b?")
+ True
+ >>> is_match("abc", "*?")
+ True
+ >>> is_match("abc", "a*d")
+ False
+ >>> is_match("abc", "a*c?")
+ False
+ >>> is_match('baaabab','*****ba*****ba')
+ False
+ >>> is_match('baaabab','*****ba*****ab')
+ True
+ >>> is_match('aa','*')
+ True
+ """
+ dp = [[False] * (len(pattern) + 1) for _ in string + "1"]
+ dp[0][0] = True
+ # Fill in the first row
+ for j, char in enumerate(pattern, 1):
+ if char == "*":
+ dp[0][j] = dp[0][j - 1]
+ # Fill in the rest of the DP table
+ for i, s_char in enumerate(string, 1):
+ for j, p_char in enumerate(pattern, 1):
+ if p_char in (s_char, "?"):
+ dp[i][j] = dp[i - 1][j - 1]
+ elif pattern[j - 1] == "*":
+ dp[i][j] = dp[i - 1][j] or dp[i][j - 1]
+ return dp[len(string)][len(pattern)]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(f"{is_match('baaabab','*****ba*****ab') = }")
diff --git a/dynamic_programming/word_break.py b/dynamic_programming/word_break.py
new file mode 100644
index 000000000000..4d7ac869080c
--- /dev/null
+++ b/dynamic_programming/word_break.py
@@ -0,0 +1,111 @@
+"""
+Author : Alexander Pantyukhin
+Date : December 12, 2022
+
+Task:
+Given a string and a list of words, return true if the string can be
+segmented into a space-separated sequence of one or more words.
+
+Note that the same word may be reused
+multiple times in the segmentation.
+
+Implementation notes: Trie + Dynamic programming up -> down.
+The Trie will be used to store the words. It will be useful for scanning
+available words for the current position in the string.
+
+Leetcode:
+https://leetcode.com/problems/word-break/description/
+
+Runtime: O(n * n)
+Space: O(n)
+"""
+
+import functools
+from typing import Any
+
+
+def word_break(string: str, words: list[str]) -> bool:
+ """
+ Return True if numbers have opposite signs False otherwise.
+
+ >>> word_break("applepenapple", ["apple","pen"])
+ True
+ >>> word_break("catsandog", ["cats","dog","sand","and","cat"])
+ False
+ >>> word_break("cars", ["car","ca","rs"])
+ True
+ >>> word_break('abc', [])
+ False
+ >>> word_break(123, ['a'])
+ Traceback (most recent call last):
+ ...
+ ValueError: the string should be not empty string
+ >>> word_break('', ['a'])
+ Traceback (most recent call last):
+ ...
+ ValueError: the string should be not empty string
+ >>> word_break('abc', [123])
+ Traceback (most recent call last):
+ ...
+ ValueError: the words should be a list of non-empty strings
+ >>> word_break('abc', [''])
+ Traceback (most recent call last):
+ ...
+ ValueError: the words should be a list of non-empty strings
+ """
+
+ # Validation
+ if not isinstance(string, str) or len(string) == 0:
+ raise ValueError("the string should be not empty string")
+
+ if not isinstance(words, list) or not all(
+ isinstance(item, str) and len(item) > 0 for item in words
+ ):
+ raise ValueError("the words should be a list of non-empty strings")
+
+ # Build trie
+ trie: dict[str, Any] = {}
+ word_keeper_key = "WORD_KEEPER"
+
+ for word in words:
+ trie_node = trie
+ for c in word:
+ if c not in trie_node:
+ trie_node[c] = {}
+
+ trie_node = trie_node[c]
+
+ trie_node[word_keeper_key] = True
+
+ len_string = len(string)
+
+ # Dynamic programming method
+ @functools.cache
+ def is_breakable(index: int) -> bool:
+ """
+ >>> string = 'a'
+ >>> is_breakable(1)
+ True
+ """
+ if index == len_string:
+ return True
+
+ trie_node = trie
+ for i in range(index, len_string):
+ trie_node = trie_node.get(string[i], None)
+
+ if trie_node is None:
+ return False
+
+ if trie_node.get(word_keeper_key, False) and is_breakable(i + 1):
+ return True
+
+ return False
+
+ return is_breakable(0)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/__init__.py b/electronics/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/electronics/apparent_power.py b/electronics/apparent_power.py
new file mode 100644
index 000000000000..0ce1c2aa95b9
--- /dev/null
+++ b/electronics/apparent_power.py
@@ -0,0 +1,37 @@
+import cmath
+import math
+
+
+def apparent_power(
+ voltage: float, current: float, voltage_angle: float, current_angle: float
+) -> complex:
+ """
+ Calculate the apparent power in a single-phase AC circuit.
+
+ Reference: https://en.wikipedia.org/wiki/AC_power#Apparent_power
+
+ >>> apparent_power(100, 5, 0, 0)
+ (500+0j)
+ >>> apparent_power(100, 5, 90, 0)
+ (3.061616997868383e-14+500j)
+ >>> apparent_power(100, 5, -45, -60)
+ (-129.40952255126027-482.9629131445341j)
+ >>> apparent_power(200, 10, -30, -90)
+ (-999.9999999999998-1732.0508075688776j)
+ """
+ # Convert angles from degrees to radians
+ voltage_angle_rad = math.radians(voltage_angle)
+ current_angle_rad = math.radians(current_angle)
+
+ # Convert voltage and current to rectangular form
+ voltage_rect = cmath.rect(voltage, voltage_angle_rad)
+ current_rect = cmath.rect(current, current_angle_rad)
+
+ # Calculate apparent power
+ return voltage_rect * current_rect
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/builtin_voltage.py b/electronics/builtin_voltage.py
new file mode 100644
index 000000000000..38fde4524d1a
--- /dev/null
+++ b/electronics/builtin_voltage.py
@@ -0,0 +1,67 @@
+from math import log
+
+from scipy.constants import Boltzmann, physical_constants
+
+T = 300 # TEMPERATURE (unit = K)
+
+
+def builtin_voltage(
+ donor_conc: float, # donor concentration
+ acceptor_conc: float, # acceptor concentration
+ intrinsic_conc: float, # intrinsic concentration
+) -> float:
+ """
+ This function can calculate the Builtin Voltage of a pn junction diode.
+ This is calculated from the given three values.
+ Examples -
+ >>> builtin_voltage(donor_conc=1e17, acceptor_conc=1e17, intrinsic_conc=1e10)
+ 0.833370010652644
+ >>> builtin_voltage(donor_conc=0, acceptor_conc=1600, intrinsic_conc=200)
+ Traceback (most recent call last):
+ ...
+ ValueError: Donor concentration should be positive
+ >>> builtin_voltage(donor_conc=1000, acceptor_conc=0, intrinsic_conc=1200)
+ Traceback (most recent call last):
+ ...
+ ValueError: Acceptor concentration should be positive
+ >>> builtin_voltage(donor_conc=1000, acceptor_conc=1000, intrinsic_conc=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Intrinsic concentration should be positive
+ >>> builtin_voltage(donor_conc=1000, acceptor_conc=3000, intrinsic_conc=2000)
+ Traceback (most recent call last):
+ ...
+ ValueError: Donor concentration should be greater than intrinsic concentration
+ >>> builtin_voltage(donor_conc=3000, acceptor_conc=1000, intrinsic_conc=2000)
+ Traceback (most recent call last):
+ ...
+ ValueError: Acceptor concentration should be greater than intrinsic concentration
+ """
+
+ if donor_conc <= 0:
+ raise ValueError("Donor concentration should be positive")
+ elif acceptor_conc <= 0:
+ raise ValueError("Acceptor concentration should be positive")
+ elif intrinsic_conc <= 0:
+ raise ValueError("Intrinsic concentration should be positive")
+ elif donor_conc <= intrinsic_conc:
+ raise ValueError(
+ "Donor concentration should be greater than intrinsic concentration"
+ )
+ elif acceptor_conc <= intrinsic_conc:
+ raise ValueError(
+ "Acceptor concentration should be greater than intrinsic concentration"
+ )
+ else:
+ return (
+ Boltzmann
+ * T
+ * log((donor_conc * acceptor_conc) / intrinsic_conc**2)
+ / physical_constants["electron volt"][0]
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/capacitor_equivalence.py b/electronics/capacitor_equivalence.py
new file mode 100644
index 000000000000..274b18afb3ef
--- /dev/null
+++ b/electronics/capacitor_equivalence.py
@@ -0,0 +1,53 @@
+# https://farside.ph.utexas.edu/teaching/316/lectures/node46.html
+
+from __future__ import annotations
+
+
+def capacitor_parallel(capacitors: list[float]) -> float:
+ """
+ Ceq = C1 + C2 + ... + Cn
+ Calculate the equivalent resistance for any number of capacitors in parallel.
+ >>> capacitor_parallel([5.71389, 12, 3])
+ 20.71389
+ >>> capacitor_parallel([5.71389, 12, -3])
+ Traceback (most recent call last):
+ ...
+ ValueError: Capacitor at index 2 has a negative value!
+ """
+ sum_c = 0.0
+ for index, capacitor in enumerate(capacitors):
+ if capacitor < 0:
+ msg = f"Capacitor at index {index} has a negative value!"
+ raise ValueError(msg)
+ sum_c += capacitor
+ return sum_c
+
+
+def capacitor_series(capacitors: list[float]) -> float:
+ """
+ Ceq = 1/ (1/C1 + 1/C2 + ... + 1/Cn)
+ >>> capacitor_series([5.71389, 12, 3])
+ 1.6901062252507735
+ >>> capacitor_series([5.71389, 12, -3])
+ Traceback (most recent call last):
+ ...
+ ValueError: Capacitor at index 2 has a negative or zero value!
+ >>> capacitor_series([5.71389, 12, 0.000])
+ Traceback (most recent call last):
+ ...
+ ValueError: Capacitor at index 2 has a negative or zero value!
+ """
+
+ first_sum = 0.0
+ for index, capacitor in enumerate(capacitors):
+ if capacitor <= 0:
+ msg = f"Capacitor at index {index} has a negative or zero value!"
+ raise ValueError(msg)
+ first_sum += 1 / capacitor
+ return 1 / first_sum
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/carrier_concentration.py b/electronics/carrier_concentration.py
new file mode 100644
index 000000000000..1fb9f2430dcd
--- /dev/null
+++ b/electronics/carrier_concentration.py
@@ -0,0 +1,75 @@
+# https://en.wikipedia.org/wiki/Charge_carrier_density
+# https://www.pveducation.org/pvcdrom/pn-junctions/equilibrium-carrier-concentration
+# http://www.ece.utep.edu/courses/ee3329/ee3329/Studyguide/ToC/Fundamentals/Carriers/concentrations.html
+
+from __future__ import annotations
+
+
+def carrier_concentration(
+ electron_conc: float,
+ hole_conc: float,
+ intrinsic_conc: float,
+) -> tuple:
+ """
+ This function can calculate any one of the three -
+ 1. Electron Concentration
+ 2, Hole Concentration
+ 3. Intrinsic Concentration
+ given the other two.
+ Examples -
+ >>> carrier_concentration(electron_conc=25, hole_conc=100, intrinsic_conc=0)
+ ('intrinsic_conc', 50.0)
+ >>> carrier_concentration(electron_conc=0, hole_conc=1600, intrinsic_conc=200)
+ ('electron_conc', 25.0)
+ >>> carrier_concentration(electron_conc=1000, hole_conc=0, intrinsic_conc=1200)
+ ('hole_conc', 1440.0)
+ >>> carrier_concentration(electron_conc=1000, hole_conc=400, intrinsic_conc=1200)
+ Traceback (most recent call last):
+ ...
+ ValueError: You cannot supply more or less than 2 values
+ >>> carrier_concentration(electron_conc=-1000, hole_conc=0, intrinsic_conc=1200)
+ Traceback (most recent call last):
+ ...
+ ValueError: Electron concentration cannot be negative in a semiconductor
+ >>> carrier_concentration(electron_conc=0, hole_conc=-400, intrinsic_conc=1200)
+ Traceback (most recent call last):
+ ...
+ ValueError: Hole concentration cannot be negative in a semiconductor
+ >>> carrier_concentration(electron_conc=0, hole_conc=400, intrinsic_conc=-1200)
+ Traceback (most recent call last):
+ ...
+ ValueError: Intrinsic concentration cannot be negative in a semiconductor
+ """
+ if (electron_conc, hole_conc, intrinsic_conc).count(0) != 1:
+ raise ValueError("You cannot supply more or less than 2 values")
+ elif electron_conc < 0:
+ raise ValueError("Electron concentration cannot be negative in a semiconductor")
+ elif hole_conc < 0:
+ raise ValueError("Hole concentration cannot be negative in a semiconductor")
+ elif intrinsic_conc < 0:
+ raise ValueError(
+ "Intrinsic concentration cannot be negative in a semiconductor"
+ )
+ elif electron_conc == 0:
+ return (
+ "electron_conc",
+ intrinsic_conc**2 / hole_conc,
+ )
+ elif hole_conc == 0:
+ return (
+ "hole_conc",
+ intrinsic_conc**2 / electron_conc,
+ )
+ elif intrinsic_conc == 0:
+ return (
+ "intrinsic_conc",
+ (electron_conc * hole_conc) ** 0.5,
+ )
+ else:
+ return (-1, -1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/charging_capacitor.py b/electronics/charging_capacitor.py
new file mode 100644
index 000000000000..0021e4e345e0
--- /dev/null
+++ b/electronics/charging_capacitor.py
@@ -0,0 +1,72 @@
+# source - The ARRL Handbook for Radio Communications
+# https://en.wikipedia.org/wiki/RC_time_constant
+
+"""
+Description
+-----------
+When a capacitor is connected with a potential source (AC or DC). It starts to charge
+at a general speed but when a resistor is connected in the circuit with in series to
+a capacitor then the capacitor charges slowly means it will take more time than usual.
+while the capacitor is being charged, the voltage is in exponential function with time.
+
+'resistance(ohms) * capacitance(farads)' is called RC-timeconstant which may also be
+represented as τ (tau). By using this RC-timeconstant we can find the voltage at any
+time 't' from the initiation of charging a capacitor with the help of the exponential
+function containing RC. Both at charging and discharging of a capacitor.
+"""
+
+from math import exp # value of exp = 2.718281828459…
+
+
+def charging_capacitor(
+ source_voltage: float, # voltage in volts.
+ resistance: float, # resistance in ohms.
+ capacitance: float, # capacitance in farads.
+ time_sec: float, # time in seconds after charging initiation of capacitor.
+) -> float:
+ """
+ Find capacitor voltage at any nth second after initiating its charging.
+
+ Examples
+ --------
+ >>> charging_capacitor(source_voltage=.2,resistance=.9,capacitance=8.4,time_sec=.5)
+ 0.013
+
+ >>> charging_capacitor(source_voltage=2.2,resistance=3.5,capacitance=2.4,time_sec=9)
+ 1.446
+
+ >>> charging_capacitor(source_voltage=15,resistance=200,capacitance=20,time_sec=2)
+ 0.007
+
+ >>> charging_capacitor(20, 2000, 30*pow(10,-5), 4)
+ 19.975
+
+ >>> charging_capacitor(source_voltage=0,resistance=10.0,capacitance=.30,time_sec=3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Source voltage must be positive.
+
+ >>> charging_capacitor(source_voltage=20,resistance=-2000,capacitance=30,time_sec=4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistance must be positive.
+
+ >>> charging_capacitor(source_voltage=30,resistance=1500,capacitance=0,time_sec=4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Capacitance must be positive.
+ """
+
+ if source_voltage <= 0:
+ raise ValueError("Source voltage must be positive.")
+ if resistance <= 0:
+ raise ValueError("Resistance must be positive.")
+ if capacitance <= 0:
+ raise ValueError("Capacitance must be positive.")
+ return round(source_voltage * (1 - exp(-time_sec / (resistance * capacitance))), 3)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/charging_inductor.py b/electronics/charging_inductor.py
new file mode 100644
index 000000000000..8a3bbc0bbfcd
--- /dev/null
+++ b/electronics/charging_inductor.py
@@ -0,0 +1,97 @@
+# source - The ARRL Handbook for Radio Communications
+# https://en.wikipedia.org/wiki/RL_circuit
+
+"""
+Description
+-----------
+Inductor is a passive electronic device which stores energy but unlike capacitor, it
+stores energy in its 'magnetic field' or 'magnetostatic field'.
+
+When inductor is connected to 'DC' current source nothing happens it just works like a
+wire because it's real effect cannot be seen while 'DC' is connected, its not even
+going to store energy. Inductor stores energy only when it is working on 'AC' current.
+
+Connecting a inductor in series with a resistor(when R = 0) to a 'AC' potential source,
+from zero to a finite value causes a sudden voltage to induced in inductor which
+opposes the current. which results in initially slowly current rise. However it would
+cease if there is no further changes in current. With resistance zero current will never
+stop rising.
+
+'Resistance(ohms) / Inductance(henrys)' is known as RL-timeconstant. It also represents
+as τ (tau). While the charging of a inductor with a resistor results in
+a exponential function.
+
+when inductor is connected across 'AC' potential source. It starts to store the energy
+in its 'magnetic field'.with the help 'RL-time-constant' we can find current at any time
+in inductor while it is charging.
+"""
+
+from math import exp # value of exp = 2.718281828459…
+
+
+def charging_inductor(
+ source_voltage: float, # source_voltage should be in volts.
+ resistance: float, # resistance should be in ohms.
+ inductance: float, # inductance should be in henrys.
+ time: float, # time should in seconds.
+) -> float:
+ """
+ Find inductor current at any nth second after initiating its charging.
+
+ Examples
+ --------
+ >>> charging_inductor(source_voltage=5.8,resistance=1.5,inductance=2.3,time=2)
+ 2.817
+
+ >>> charging_inductor(source_voltage=8,resistance=5,inductance=3,time=2)
+ 1.543
+
+ >>> charging_inductor(source_voltage=8,resistance=5*pow(10,2),inductance=3,time=2)
+ 0.016
+
+ >>> charging_inductor(source_voltage=-8,resistance=100,inductance=15,time=12)
+ Traceback (most recent call last):
+ ...
+ ValueError: Source voltage must be positive.
+
+ >>> charging_inductor(source_voltage=80,resistance=-15,inductance=100,time=5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistance must be positive.
+
+ >>> charging_inductor(source_voltage=12,resistance=200,inductance=-20,time=5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inductance must be positive.
+
+ >>> charging_inductor(source_voltage=0,resistance=200,inductance=20,time=5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Source voltage must be positive.
+
+ >>> charging_inductor(source_voltage=10,resistance=0,inductance=20,time=5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistance must be positive.
+
+ >>> charging_inductor(source_voltage=15, resistance=25, inductance=0, time=5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inductance must be positive.
+ """
+
+ if source_voltage <= 0:
+ raise ValueError("Source voltage must be positive.")
+ if resistance <= 0:
+ raise ValueError("Resistance must be positive.")
+ if inductance <= 0:
+ raise ValueError("Inductance must be positive.")
+ return round(
+ source_voltage / resistance * (1 - exp((-time * resistance) / inductance)), 3
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/circular_convolution.py b/electronics/circular_convolution.py
new file mode 100644
index 000000000000..d06e76be759b
--- /dev/null
+++ b/electronics/circular_convolution.py
@@ -0,0 +1,98 @@
+# https://en.wikipedia.org/wiki/Circular_convolution
+
+"""
+Circular convolution, also known as cyclic convolution,
+is a special case of periodic convolution, which is the convolution of two
+periodic functions that have the same period. Periodic convolution arises,
+for example, in the context of the discrete-time Fourier transform (DTFT).
+In particular, the DTFT of the product of two discrete sequences is the periodic
+convolution of the DTFTs of the individual sequences. And each DTFT is a periodic
+summation of a continuous Fourier transform function.
+
+Source: https://en.wikipedia.org/wiki/Circular_convolution
+"""
+
+import doctest
+from collections import deque
+
+import numpy as np
+
+
+class CircularConvolution:
+ """
+ This class stores the first and second signal and performs the circular convolution
+ """
+
+ def __init__(self) -> None:
+ """
+ First signal and second signal are stored as 1-D array
+ """
+
+ self.first_signal = [2, 1, 2, -1]
+ self.second_signal = [1, 2, 3, 4]
+
+ def circular_convolution(self) -> list[float]:
+ """
+ This function performs the circular convolution of the first and second signal
+ using matrix method
+
+ Usage:
+ >>> convolution = CircularConvolution()
+ >>> convolution.circular_convolution()
+ [10.0, 10.0, 6.0, 14.0]
+
+ >>> convolution.first_signal = [0.2, 0.4, 0.6, 0.8, 1.0, 1.2, 1.4, 1.6]
+ >>> convolution.second_signal = [0.1, 0.3, 0.5, 0.7, 0.9, 1.1, 1.3, 1.5]
+ >>> convolution.circular_convolution()
+ [5.2, 6.0, 6.48, 6.64, 6.48, 6.0, 5.2, 4.08]
+
+ >>> convolution.first_signal = [-1, 1, 2, -2]
+ >>> convolution.second_signal = [0.5, 1, -1, 2, 0.75]
+ >>> convolution.circular_convolution()
+ [6.25, -3.0, 1.5, -2.0, -2.75]
+
+ >>> convolution.first_signal = [1, -1, 2, 3, -1]
+ >>> convolution.second_signal = [1, 2, 3]
+ >>> convolution.circular_convolution()
+ [8.0, -2.0, 3.0, 4.0, 11.0]
+
+ """
+
+ length_first_signal = len(self.first_signal)
+ length_second_signal = len(self.second_signal)
+
+ max_length = max(length_first_signal, length_second_signal)
+
+ # create a zero matrix of max_length x max_length
+ matrix = [[0] * max_length for i in range(max_length)]
+
+ # fills the smaller signal with zeros to make both signals of same length
+ if length_first_signal < length_second_signal:
+ self.first_signal += [0] * (max_length - length_first_signal)
+ elif length_first_signal > length_second_signal:
+ self.second_signal += [0] * (max_length - length_second_signal)
+
+ """
+ Fills the matrix in the following way assuming 'x' is the signal of length 4
+ [
+ [x[0], x[3], x[2], x[1]],
+ [x[1], x[0], x[3], x[2]],
+ [x[2], x[1], x[0], x[3]],
+ [x[3], x[2], x[1], x[0]]
+ ]
+ """
+ for i in range(max_length):
+ rotated_signal = deque(self.second_signal)
+ rotated_signal.rotate(i)
+ for j, item in enumerate(rotated_signal):
+ matrix[i][j] += item
+
+ # multiply the matrix with the first signal
+ final_signal = np.matmul(np.transpose(matrix), np.transpose(self.first_signal))
+
+ # rounding-off to two decimal places
+ return [float(round(i, 2)) for i in final_signal]
+
+
+if __name__ == "__main__":
+ doctest.testmod()
diff --git a/electronics/coulombs_law.py b/electronics/coulombs_law.py
new file mode 100644
index 000000000000..74bbea5ea8ec
--- /dev/null
+++ b/electronics/coulombs_law.py
@@ -0,0 +1,85 @@
+# https://en.wikipedia.org/wiki/Coulomb%27s_law
+
+from __future__ import annotations
+
+COULOMBS_CONSTANT = 8.988e9 # units = N * m^s * C^-2
+
+
+def couloumbs_law(
+ force: float, charge1: float, charge2: float, distance: float
+) -> dict[str, float]:
+ """
+ Apply Coulomb's Law on any three given values. These can be force, charge1,
+ charge2, or distance, and then in a Python dict return name/value pair of
+ the zero value.
+
+ Coulomb's Law states that the magnitude of the electrostatic force of
+ attraction or repulsion between two point charges is directly proportional
+ to the product of the magnitudes of charges and inversely proportional to
+ the square of the distance between them.
+
+ Reference
+ ----------
+ Coulomb (1785) "Premier mémoire sur l'électricité et le magnétisme,"
+ Histoire de l'Académie Royale des Sciences, pp. 569-577.
+
+ Parameters
+ ----------
+ force : float with units in Newtons
+
+ charge1 : float with units in Coulombs
+
+ charge2 : float with units in Coulombs
+
+ distance : float with units in meters
+
+ Returns
+ -------
+ result : dict name/value pair of the zero value
+
+ >>> couloumbs_law(force=0, charge1=3, charge2=5, distance=2000)
+ {'force': 33705.0}
+
+ >>> couloumbs_law(force=10, charge1=3, charge2=5, distance=0)
+ {'distance': 116112.01488218177}
+
+ >>> couloumbs_law(force=10, charge1=0, charge2=5, distance=2000)
+ {'charge1': 0.0008900756564307966}
+
+ >>> couloumbs_law(force=0, charge1=0, charge2=5, distance=2000)
+ Traceback (most recent call last):
+ ...
+ ValueError: One and only one argument must be 0
+
+ >>> couloumbs_law(force=0, charge1=3, charge2=5, distance=-2000)
+ Traceback (most recent call last):
+ ...
+ ValueError: Distance cannot be negative
+
+ """
+
+ charge_product = abs(charge1 * charge2)
+
+ if (force, charge1, charge2, distance).count(0) != 1:
+ raise ValueError("One and only one argument must be 0")
+ if distance < 0:
+ raise ValueError("Distance cannot be negative")
+ if force == 0:
+ force = COULOMBS_CONSTANT * charge_product / (distance**2)
+ return {"force": force}
+ elif charge1 == 0:
+ charge1 = abs(force) * (distance**2) / (COULOMBS_CONSTANT * charge2)
+ return {"charge1": charge1}
+ elif charge2 == 0:
+ charge2 = abs(force) * (distance**2) / (COULOMBS_CONSTANT * charge1)
+ return {"charge2": charge2}
+ elif distance == 0:
+ distance = (COULOMBS_CONSTANT * charge_product / abs(force)) ** 0.5
+ return {"distance": distance}
+ raise ValueError("Exactly one argument must be 0")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/electric_conductivity.py b/electronics/electric_conductivity.py
new file mode 100644
index 000000000000..65bb6c5ceaf0
--- /dev/null
+++ b/electronics/electric_conductivity.py
@@ -0,0 +1,73 @@
+from __future__ import annotations
+
+ELECTRON_CHARGE = 1.6021e-19 # units = C
+
+
+def electric_conductivity(
+ conductivity: float,
+ electron_conc: float,
+ mobility: float,
+) -> tuple[str, float]:
+ """
+ This function can calculate any one of the three -
+ 1. Conductivity
+ 2. Electron Concentration
+ 3. Electron Mobility
+ This is calculated from the other two provided values
+ Examples -
+ >>> electric_conductivity(conductivity=25, electron_conc=100, mobility=0)
+ ('mobility', 1.5604519068722301e+18)
+ >>> electric_conductivity(conductivity=0, electron_conc=1600, mobility=200)
+ ('conductivity', 5.12672e-14)
+ >>> electric_conductivity(conductivity=1000, electron_conc=0, mobility=1200)
+ ('electron_conc', 5.201506356240767e+18)
+ >>> electric_conductivity(conductivity=-10, electron_conc=100, mobility=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Conductivity cannot be negative
+ >>> electric_conductivity(conductivity=50, electron_conc=-10, mobility=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Electron concentration cannot be negative
+ >>> electric_conductivity(conductivity=50, electron_conc=0, mobility=-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: mobility cannot be negative
+ >>> electric_conductivity(conductivity=50, electron_conc=0, mobility=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: You cannot supply more or less than 2 values
+ >>> electric_conductivity(conductivity=50, electron_conc=200, mobility=300)
+ Traceback (most recent call last):
+ ...
+ ValueError: You cannot supply more or less than 2 values
+ """
+ if (conductivity, electron_conc, mobility).count(0) != 1:
+ raise ValueError("You cannot supply more or less than 2 values")
+ elif conductivity < 0:
+ raise ValueError("Conductivity cannot be negative")
+ elif electron_conc < 0:
+ raise ValueError("Electron concentration cannot be negative")
+ elif mobility < 0:
+ raise ValueError("mobility cannot be negative")
+ elif conductivity == 0:
+ return (
+ "conductivity",
+ mobility * electron_conc * ELECTRON_CHARGE,
+ )
+ elif electron_conc == 0:
+ return (
+ "electron_conc",
+ conductivity / (mobility * ELECTRON_CHARGE),
+ )
+ else:
+ return (
+ "mobility",
+ conductivity / (electron_conc * ELECTRON_CHARGE),
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/electric_power.py b/electronics/electric_power.py
new file mode 100644
index 000000000000..8e3454e39c3f
--- /dev/null
+++ b/electronics/electric_power.py
@@ -0,0 +1,59 @@
+# https://en.m.wikipedia.org/wiki/Electric_power
+from __future__ import annotations
+
+from typing import NamedTuple
+
+
+class Result(NamedTuple):
+ name: str
+ value: float
+
+
+def electric_power(voltage: float, current: float, power: float) -> tuple:
+ """
+ This function can calculate any one of the three (voltage, current, power),
+ fundamental value of electrical system.
+ examples are below:
+ >>> electric_power(voltage=0, current=2, power=5)
+ Result(name='voltage', value=2.5)
+ >>> electric_power(voltage=2, current=2, power=0)
+ Result(name='power', value=4.0)
+ >>> electric_power(voltage=-2, current=3, power=0)
+ Result(name='power', value=6.0)
+ >>> electric_power(voltage=2, current=4, power=2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Exactly one argument must be 0
+ >>> electric_power(voltage=0, current=0, power=2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Exactly one argument must be 0
+ >>> electric_power(voltage=0, current=2, power=-4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Power cannot be negative in any electrical/electronics system
+ >>> electric_power(voltage=2.2, current=2.2, power=0)
+ Result(name='power', value=4.84)
+ >>> electric_power(current=0, power=6, voltage=2)
+ Result(name='current', value=3.0)
+ """
+ if (voltage, current, power).count(0) != 1:
+ raise ValueError("Exactly one argument must be 0")
+ elif power < 0:
+ raise ValueError(
+ "Power cannot be negative in any electrical/electronics system"
+ )
+ elif voltage == 0:
+ return Result("voltage", power / current)
+ elif current == 0:
+ return Result("current", power / voltage)
+ elif power == 0:
+ return Result("power", float(round(abs(voltage * current), 2)))
+ else:
+ raise AssertionError
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/electrical_impedance.py b/electronics/electrical_impedance.py
new file mode 100644
index 000000000000..4f4f1d308293
--- /dev/null
+++ b/electronics/electrical_impedance.py
@@ -0,0 +1,46 @@
+"""Electrical impedance is the measure of the opposition that a
+circuit presents to a current when a voltage is applied.
+Impedance extends the concept of resistance to alternating current (AC) circuits.
+Source: https://en.wikipedia.org/wiki/Electrical_impedance
+"""
+
+from __future__ import annotations
+
+from math import pow, sqrt # noqa: A004
+
+
+def electrical_impedance(
+ resistance: float, reactance: float, impedance: float
+) -> dict[str, float]:
+ """
+ Apply Electrical Impedance formula, on any two given electrical values,
+ which can be resistance, reactance, and impedance, and then in a Python dict
+ return name/value pair of the zero value.
+
+ >>> electrical_impedance(3,4,0)
+ {'impedance': 5.0}
+ >>> electrical_impedance(0,4,5)
+ {'resistance': 3.0}
+ >>> electrical_impedance(3,0,5)
+ {'reactance': 4.0}
+ >>> electrical_impedance(3,4,5)
+ Traceback (most recent call last):
+ ...
+ ValueError: One and only one argument must be 0
+ """
+ if (resistance, reactance, impedance).count(0) != 1:
+ raise ValueError("One and only one argument must be 0")
+ if resistance == 0:
+ return {"resistance": sqrt(pow(impedance, 2) - pow(reactance, 2))}
+ elif reactance == 0:
+ return {"reactance": sqrt(pow(impedance, 2) - pow(resistance, 2))}
+ elif impedance == 0:
+ return {"impedance": sqrt(pow(resistance, 2) + pow(reactance, 2))}
+ else:
+ raise ValueError("Exactly one argument must be 0")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/ic_555_timer.py b/electronics/ic_555_timer.py
new file mode 100644
index 000000000000..e187e1928dca
--- /dev/null
+++ b/electronics/ic_555_timer.py
@@ -0,0 +1,75 @@
+from __future__ import annotations
+
+"""
+ Calculate the frequency and/or duty cycle of an astable 555 timer.
+ * https://en.wikipedia.org/wiki/555_timer_IC#Astable
+
+ These functions take in the value of the external resistances (in ohms)
+ and capacitance (in Microfarad), and calculates the following:
+
+ -------------------------------------
+ | Freq = 1.44 /[( R1+ 2 x R2) x C1] | ... in Hz
+ -------------------------------------
+ where Freq is the frequency,
+ R1 is the first resistance in ohms,
+ R2 is the second resistance in ohms,
+ C1 is the capacitance in Microfarads.
+
+ ------------------------------------------------
+ | Duty Cycle = (R1 + R2) / (R1 + 2 x R2) x 100 | ... in %
+ ------------------------------------------------
+ where R1 is the first resistance in ohms,
+ R2 is the second resistance in ohms.
+"""
+
+
+def astable_frequency(
+ resistance_1: float, resistance_2: float, capacitance: float
+) -> float:
+ """
+ Usage examples:
+ >>> astable_frequency(resistance_1=45, resistance_2=45, capacitance=7)
+ 1523.8095238095239
+ >>> astable_frequency(resistance_1=356, resistance_2=234, capacitance=976)
+ 1.7905459175553078
+ >>> astable_frequency(resistance_1=2, resistance_2=-1, capacitance=2)
+ Traceback (most recent call last):
+ ...
+ ValueError: All values must be positive
+ >>> astable_frequency(resistance_1=45, resistance_2=45, capacitance=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: All values must be positive
+ """
+
+ if resistance_1 <= 0 or resistance_2 <= 0 or capacitance <= 0:
+ raise ValueError("All values must be positive")
+ return (1.44 / ((resistance_1 + 2 * resistance_2) * capacitance)) * 10**6
+
+
+def astable_duty_cycle(resistance_1: float, resistance_2: float) -> float:
+ """
+ Usage examples:
+ >>> astable_duty_cycle(resistance_1=45, resistance_2=45)
+ 66.66666666666666
+ >>> astable_duty_cycle(resistance_1=356, resistance_2=234)
+ 71.60194174757282
+ >>> astable_duty_cycle(resistance_1=2, resistance_2=-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: All values must be positive
+ >>> astable_duty_cycle(resistance_1=0, resistance_2=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: All values must be positive
+ """
+
+ if resistance_1 <= 0 or resistance_2 <= 0:
+ raise ValueError("All values must be positive")
+ return (resistance_1 + resistance_2) / (resistance_1 + 2 * resistance_2) * 100
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/ind_reactance.py b/electronics/ind_reactance.py
new file mode 100644
index 000000000000..3f77ef628203
--- /dev/null
+++ b/electronics/ind_reactance.py
@@ -0,0 +1,69 @@
+# https://en.wikipedia.org/wiki/Electrical_reactance#Inductive_reactance
+from __future__ import annotations
+
+from math import pi
+
+
+def ind_reactance(
+ inductance: float, frequency: float, reactance: float
+) -> dict[str, float]:
+ """
+ Calculate inductive reactance, frequency or inductance from two given electrical
+ properties then return name/value pair of the zero value in a Python dict.
+
+ Parameters
+ ----------
+ inductance : float with units in Henries
+
+ frequency : float with units in Hertz
+
+ reactance : float with units in Ohms
+
+ >>> ind_reactance(-35e-6, 1e3, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inductance cannot be negative
+
+ >>> ind_reactance(35e-6, -1e3, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Frequency cannot be negative
+
+ >>> ind_reactance(35e-6, 0, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inductive reactance cannot be negative
+
+ >>> ind_reactance(0, 10e3, 50)
+ {'inductance': 0.0007957747154594767}
+
+ >>> ind_reactance(35e-3, 0, 50)
+ {'frequency': 227.36420441699332}
+
+ >>> ind_reactance(35e-6, 1e3, 0)
+ {'reactance': 0.2199114857512855}
+
+ """
+
+ if (inductance, frequency, reactance).count(0) != 1:
+ raise ValueError("One and only one argument must be 0")
+ if inductance < 0:
+ raise ValueError("Inductance cannot be negative")
+ if frequency < 0:
+ raise ValueError("Frequency cannot be negative")
+ if reactance < 0:
+ raise ValueError("Inductive reactance cannot be negative")
+ if inductance == 0:
+ return {"inductance": reactance / (2 * pi * frequency)}
+ elif frequency == 0:
+ return {"frequency": reactance / (2 * pi * inductance)}
+ elif reactance == 0:
+ return {"reactance": 2 * pi * frequency * inductance}
+ else:
+ raise ValueError("Exactly one argument must be 0")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/ohms_law.py b/electronics/ohms_law.py
new file mode 100644
index 000000000000..66e737c1f909
--- /dev/null
+++ b/electronics/ohms_law.py
@@ -0,0 +1,42 @@
+# https://en.wikipedia.org/wiki/Ohm%27s_law
+from __future__ import annotations
+
+
+def ohms_law(voltage: float, current: float, resistance: float) -> dict[str, float]:
+ """
+ Apply Ohm's Law, on any two given electrical values, which can be voltage, current,
+ and resistance, and then in a Python dict return name/value pair of the zero value.
+
+ >>> ohms_law(voltage=10, resistance=5, current=0)
+ {'current': 2.0}
+ >>> ohms_law(voltage=0, current=0, resistance=10)
+ Traceback (most recent call last):
+ ...
+ ValueError: One and only one argument must be 0
+ >>> ohms_law(voltage=0, current=1, resistance=-2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistance cannot be negative
+ >>> ohms_law(resistance=0, voltage=-10, current=1)
+ {'resistance': -10.0}
+ >>> ohms_law(voltage=0, current=-1.5, resistance=2)
+ {'voltage': -3.0}
+ """
+ if (voltage, current, resistance).count(0) != 1:
+ raise ValueError("One and only one argument must be 0")
+ if resistance < 0:
+ raise ValueError("Resistance cannot be negative")
+ if voltage == 0:
+ return {"voltage": float(current * resistance)}
+ elif current == 0:
+ return {"current": voltage / resistance}
+ elif resistance == 0:
+ return {"resistance": voltage / current}
+ else:
+ raise ValueError("Exactly one argument must be 0")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/real_and_reactive_power.py b/electronics/real_and_reactive_power.py
new file mode 100644
index 000000000000..81dcba800e82
--- /dev/null
+++ b/electronics/real_and_reactive_power.py
@@ -0,0 +1,49 @@
+import math
+
+
+def real_power(apparent_power: float, power_factor: float) -> float:
+ """
+ Calculate real power from apparent power and power factor.
+
+ Examples:
+ >>> real_power(100, 0.9)
+ 90.0
+ >>> real_power(0, 0.8)
+ 0.0
+ >>> real_power(100, -0.9)
+ -90.0
+ """
+ if (
+ not isinstance(power_factor, (int, float))
+ or power_factor < -1
+ or power_factor > 1
+ ):
+ raise ValueError("power_factor must be a valid float value between -1 and 1.")
+ return apparent_power * power_factor
+
+
+def reactive_power(apparent_power: float, power_factor: float) -> float:
+ """
+ Calculate reactive power from apparent power and power factor.
+
+ Examples:
+ >>> reactive_power(100, 0.9)
+ 43.58898943540673
+ >>> reactive_power(0, 0.8)
+ 0.0
+ >>> reactive_power(100, -0.9)
+ 43.58898943540673
+ """
+ if (
+ not isinstance(power_factor, (int, float))
+ or power_factor < -1
+ or power_factor > 1
+ ):
+ raise ValueError("power_factor must be a valid float value between -1 and 1.")
+ return apparent_power * math.sqrt(1 - power_factor**2)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/resistor_color_code.py b/electronics/resistor_color_code.py
new file mode 100644
index 000000000000..189d19946d9d
--- /dev/null
+++ b/electronics/resistor_color_code.py
@@ -0,0 +1,374 @@
+"""
+Title : Calculating the resistance of a n band resistor using the color codes
+
+Description :
+ Resistors resist the flow of electrical current.Each one has a value that tells how
+ strongly it resists current flow.This value's unit is the ohm, often noted with the
+ Greek letter omega: Ω.
+
+ The colored bands on a resistor can tell you everything you need to know about its
+ value and tolerance, as long as you understand how to read them. The order in which
+ the colors are arranged is very important, and each value of resistor has its own
+ unique combination.
+
+ The color coding for resistors is an international standard that is defined in IEC
+ 60062.
+
+ The number of bands present in a resistor varies from three to six. These represent
+ significant figures, multiplier, tolerance, reliability, and temperature coefficient
+ Each color used for a type of band has a value assigned to it. It is read from left
+ to right.
+ All resistors will have significant figures and multiplier bands. In a three band
+ resistor first two bands from the left represent significant figures and the third
+ represents the multiplier band.
+
+ Significant figures - The number of significant figures band in a resistor can vary
+ from two to three.
+ Colors and values associated with significant figure bands -
+ (Black = 0, Brown = 1, Red = 2, Orange = 3, Yellow = 4, Green = 5, Blue = 6,
+ Violet = 7, Grey = 8, White = 9)
+
+ Multiplier - There will be one multiplier band in a resistor. It is multiplied with
+ the significant figures obtained from previous bands.
+ Colors and values associated with multiplier band -
+ (Black = 100, Brown = 10^1, Red = 10^2, Orange = 10^3, Yellow = 10^4, Green = 10^5,
+ Blue = 10^6, Violet = 10^7, Grey = 10^8, White = 10^9, Gold = 10^-1, Silver = 10^-2)
+ Note that multiplier bands use Gold and Silver which are not used for significant
+ figure bands.
+
+ Tolerance - The tolerance band is not always present. It can be seen in four band
+ resistors and above. This is a percentage by which the resistor value can vary.
+ Colors and values associated with tolerance band -
+ (Brown = 1%, Red = 2%, Orange = 0.05%, Yellow = 0.02%, Green = 0.5%,Blue = 0.25%,
+ Violet = 0.1%, Grey = 0.01%, Gold = 5%, Silver = 10%)
+ If no color is mentioned then by default tolerance is 20%
+ Note that tolerance band does not use Black and White colors.
+
+ Temperature Coeffecient - Indicates the change in resistance of the component as
+ a function of ambient temperature in terms of ppm/K.
+ It is present in six band resistors.
+ Colors and values associated with Temperature coeffecient -
+ (Black = 250 ppm/K, Brown = 100 ppm/K, Red = 50 ppm/K, Orange = 15 ppm/K,
+ Yellow = 25 ppm/K, Green = 20 ppm/K, Blue = 10 ppm/K, Violet = 5 ppm/K,
+ Grey = 1 ppm/K)
+ Note that temperature coeffecient band does not use White, Gold, Silver colors.
+
+Sources :
+ https://www.calculator.net/resistor-calculator.html
+ https://learn.parallax.com/support/reference/resistor-color-codes
+ https://byjus.com/physics/resistor-colour-codes/
+"""
+
+valid_colors: list = [
+ "Black",
+ "Brown",
+ "Red",
+ "Orange",
+ "Yellow",
+ "Green",
+ "Blue",
+ "Violet",
+ "Grey",
+ "White",
+ "Gold",
+ "Silver",
+]
+
+significant_figures_color_values: dict[str, int] = {
+ "Black": 0,
+ "Brown": 1,
+ "Red": 2,
+ "Orange": 3,
+ "Yellow": 4,
+ "Green": 5,
+ "Blue": 6,
+ "Violet": 7,
+ "Grey": 8,
+ "White": 9,
+}
+
+multiplier_color_values: dict[str, float] = {
+ "Black": 10**0,
+ "Brown": 10**1,
+ "Red": 10**2,
+ "Orange": 10**3,
+ "Yellow": 10**4,
+ "Green": 10**5,
+ "Blue": 10**6,
+ "Violet": 10**7,
+ "Grey": 10**8,
+ "White": 10**9,
+ "Gold": 10**-1,
+ "Silver": 10**-2,
+}
+
+tolerance_color_values: dict[str, float] = {
+ "Brown": 1,
+ "Red": 2,
+ "Orange": 0.05,
+ "Yellow": 0.02,
+ "Green": 0.5,
+ "Blue": 0.25,
+ "Violet": 0.1,
+ "Grey": 0.01,
+ "Gold": 5,
+ "Silver": 10,
+}
+
+temperature_coeffecient_color_values: dict[str, int] = {
+ "Black": 250,
+ "Brown": 100,
+ "Red": 50,
+ "Orange": 15,
+ "Yellow": 25,
+ "Green": 20,
+ "Blue": 10,
+ "Violet": 5,
+ "Grey": 1,
+}
+
+band_types: dict[int, dict[str, int]] = {
+ 3: {"significant": 2, "multiplier": 1},
+ 4: {"significant": 2, "multiplier": 1, "tolerance": 1},
+ 5: {"significant": 3, "multiplier": 1, "tolerance": 1},
+ 6: {"significant": 3, "multiplier": 1, "tolerance": 1, "temp_coeffecient": 1},
+}
+
+
+def get_significant_digits(colors: list) -> str:
+ """
+ Function returns the digit associated with the color. Function takes a
+ list containing colors as input and returns digits as string
+
+ >>> get_significant_digits(['Black','Blue'])
+ '06'
+
+ >>> get_significant_digits(['Aqua','Blue'])
+ Traceback (most recent call last):
+ ...
+ ValueError: Aqua is not a valid color for significant figure bands
+
+ """
+ digit = ""
+ for color in colors:
+ if color not in significant_figures_color_values:
+ msg = f"{color} is not a valid color for significant figure bands"
+ raise ValueError(msg)
+ digit = digit + str(significant_figures_color_values[color])
+ return str(digit)
+
+
+def get_multiplier(color: str) -> float:
+ """
+ Function returns the multiplier value associated with the color.
+ Function takes color as input and returns multiplier value
+
+ >>> get_multiplier('Gold')
+ 0.1
+
+ >>> get_multiplier('Ivory')
+ Traceback (most recent call last):
+ ...
+ ValueError: Ivory is not a valid color for multiplier band
+
+ """
+ if color not in multiplier_color_values:
+ msg = f"{color} is not a valid color for multiplier band"
+ raise ValueError(msg)
+ return multiplier_color_values[color]
+
+
+def get_tolerance(color: str) -> float:
+ """
+ Function returns the tolerance value associated with the color.
+ Function takes color as input and returns tolerance value.
+
+ >>> get_tolerance('Green')
+ 0.5
+
+ >>> get_tolerance('Indigo')
+ Traceback (most recent call last):
+ ...
+ ValueError: Indigo is not a valid color for tolerance band
+
+ """
+ if color not in tolerance_color_values:
+ msg = f"{color} is not a valid color for tolerance band"
+ raise ValueError(msg)
+ return tolerance_color_values[color]
+
+
+def get_temperature_coeffecient(color: str) -> int:
+ """
+ Function returns the temperature coeffecient value associated with the color.
+ Function takes color as input and returns temperature coeffecient value.
+
+ >>> get_temperature_coeffecient('Yellow')
+ 25
+
+ >>> get_temperature_coeffecient('Cyan')
+ Traceback (most recent call last):
+ ...
+ ValueError: Cyan is not a valid color for temperature coeffecient band
+
+ """
+ if color not in temperature_coeffecient_color_values:
+ msg = f"{color} is not a valid color for temperature coeffecient band"
+ raise ValueError(msg)
+ return temperature_coeffecient_color_values[color]
+
+
+def get_band_type_count(total_number_of_bands: int, type_of_band: str) -> int:
+ """
+ Function returns the number of bands of a given type in a resistor with n bands
+ Function takes total_number_of_bands and type_of_band as input and returns
+ number of bands belonging to that type in the given resistor
+
+ >>> get_band_type_count(3,'significant')
+ 2
+
+ >>> get_band_type_count(2,'significant')
+ Traceback (most recent call last):
+ ...
+ ValueError: 2 is not a valid number of bands
+
+ >>> get_band_type_count(3,'sign')
+ Traceback (most recent call last):
+ ...
+ ValueError: sign is not valid for a 3 band resistor
+
+ >>> get_band_type_count(3,'tolerance')
+ Traceback (most recent call last):
+ ...
+ ValueError: tolerance is not valid for a 3 band resistor
+
+ >>> get_band_type_count(5,'temp_coeffecient')
+ Traceback (most recent call last):
+ ...
+ ValueError: temp_coeffecient is not valid for a 5 band resistor
+
+ """
+ if total_number_of_bands not in band_types:
+ msg = f"{total_number_of_bands} is not a valid number of bands"
+ raise ValueError(msg)
+ if type_of_band not in band_types[total_number_of_bands]:
+ msg = f"{type_of_band} is not valid for a {total_number_of_bands} band resistor"
+ raise ValueError(msg)
+ return band_types[total_number_of_bands][type_of_band]
+
+
+def check_validity(number_of_bands: int, colors: list) -> bool:
+ """
+ Function checks if the input provided is valid or not.
+ Function takes number_of_bands and colors as input and returns
+ True if it is valid
+
+ >>> check_validity(3, ["Black","Blue","Orange"])
+ True
+
+ >>> check_validity(4, ["Black","Blue","Orange"])
+ Traceback (most recent call last):
+ ...
+ ValueError: Expecting 4 colors, provided 3 colors
+
+ >>> check_validity(3, ["Cyan","Red","Yellow"])
+ Traceback (most recent call last):
+ ...
+ ValueError: Cyan is not a valid color
+
+ """
+ if number_of_bands >= 3 and number_of_bands <= 6:
+ if number_of_bands == len(colors):
+ for color in colors:
+ if color not in valid_colors:
+ msg = f"{color} is not a valid color"
+ raise ValueError(msg)
+ return True
+ else:
+ msg = f"Expecting {number_of_bands} colors, provided {len(colors)} colors"
+ raise ValueError(msg)
+ else:
+ msg = "Invalid number of bands. Resistor bands must be 3 to 6"
+ raise ValueError(msg)
+
+
+def calculate_resistance(number_of_bands: int, color_code_list: list) -> dict:
+ """
+ Function calculates the total resistance of the resistor using the color codes.
+ Function takes number_of_bands, color_code_list as input and returns
+ resistance
+
+ >>> calculate_resistance(3, ["Black","Blue","Orange"])
+ {'resistance': '6000Ω ±20% '}
+
+ >>> calculate_resistance(4, ["Orange","Green","Blue","Gold"])
+ {'resistance': '35000000Ω ±5% '}
+
+ >>> calculate_resistance(5, ["Violet","Brown","Grey","Silver","Green"])
+ {'resistance': '7.18Ω ±0.5% '}
+
+ >>> calculate_resistance(6, ["Red","Green","Blue","Yellow","Orange","Grey"])
+ {'resistance': '2560000Ω ±0.05% 1 ppm/K'}
+
+ >>> calculate_resistance(0, ["Violet","Brown","Grey","Silver","Green"])
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid number of bands. Resistor bands must be 3 to 6
+
+ >>> calculate_resistance(4, ["Violet","Brown","Grey","Silver","Green"])
+ Traceback (most recent call last):
+ ...
+ ValueError: Expecting 4 colors, provided 5 colors
+
+ >>> calculate_resistance(4, ["Violet","Silver","Brown","Grey"])
+ Traceback (most recent call last):
+ ...
+ ValueError: Silver is not a valid color for significant figure bands
+
+ >>> calculate_resistance(4, ["Violet","Blue","Lime","Grey"])
+ Traceback (most recent call last):
+ ...
+ ValueError: Lime is not a valid color
+
+ """
+ is_valid = check_validity(number_of_bands, color_code_list)
+ if is_valid:
+ number_of_significant_bands = get_band_type_count(
+ number_of_bands, "significant"
+ )
+ significant_colors = color_code_list[:number_of_significant_bands]
+ significant_digits = int(get_significant_digits(significant_colors))
+ multiplier_color = color_code_list[number_of_significant_bands]
+ multiplier = get_multiplier(multiplier_color)
+ if number_of_bands == 3:
+ tolerance_color = None
+ else:
+ tolerance_color = color_code_list[number_of_significant_bands + 1]
+ tolerance = (
+ 20 if tolerance_color is None else get_tolerance(str(tolerance_color))
+ )
+ if number_of_bands != 6:
+ temperature_coeffecient_color = None
+ else:
+ temperature_coeffecient_color = color_code_list[
+ number_of_significant_bands + 2
+ ]
+ temperature_coeffecient = (
+ 0
+ if temperature_coeffecient_color is None
+ else get_temperature_coeffecient(str(temperature_coeffecient_color))
+ )
+ resisitance = significant_digits * multiplier
+ if temperature_coeffecient == 0:
+ answer = f"{resisitance}Ω ±{tolerance}% "
+ else:
+ answer = f"{resisitance}Ω ±{tolerance}% {temperature_coeffecient} ppm/K"
+ return {"resistance": answer}
+ else:
+ raise ValueError("Input is invalid")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/resistor_equivalence.py b/electronics/resistor_equivalence.py
new file mode 100644
index 000000000000..c4ea7d4b757e
--- /dev/null
+++ b/electronics/resistor_equivalence.py
@@ -0,0 +1,56 @@
+# https://byjus.com/equivalent-resistance-formula/
+
+from __future__ import annotations
+
+
+def resistor_parallel(resistors: list[float]) -> float:
+ """
+ Req = 1/ (1/R1 + 1/R2 + ... + 1/Rn)
+
+ >>> resistor_parallel([3.21389, 2, 3])
+ 0.8737571620498019
+ >>> resistor_parallel([3.21389, 2, -3])
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistor at index 2 has a negative or zero value!
+ >>> resistor_parallel([3.21389, 2, 0.000])
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistor at index 2 has a negative or zero value!
+ """
+
+ first_sum = 0.00
+ for index, resistor in enumerate(resistors):
+ if resistor <= 0:
+ msg = f"Resistor at index {index} has a negative or zero value!"
+ raise ValueError(msg)
+ first_sum += 1 / float(resistor)
+ return 1 / first_sum
+
+
+def resistor_series(resistors: list[float]) -> float:
+ """
+ Req = R1 + R2 + ... + Rn
+
+ Calculate the equivalent resistance for any number of resistors in parallel.
+
+ >>> resistor_series([3.21389, 2, 3])
+ 8.21389
+ >>> resistor_series([3.21389, 2, -3])
+ Traceback (most recent call last):
+ ...
+ ValueError: Resistor at index 2 has a negative value!
+ """
+ sum_r = 0.00
+ for index, resistor in enumerate(resistors):
+ sum_r += resistor
+ if resistor < 0:
+ msg = f"Resistor at index {index} has a negative value!"
+ raise ValueError(msg)
+ return sum_r
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/resonant_frequency.py b/electronics/resonant_frequency.py
new file mode 100644
index 000000000000..4f95043b600a
--- /dev/null
+++ b/electronics/resonant_frequency.py
@@ -0,0 +1,50 @@
+# https://en.wikipedia.org/wiki/LC_circuit
+
+"""An LC circuit, also called a resonant circuit, tank circuit, or tuned circuit,
+is an electric circuit consisting of an inductor, represented by the letter L,
+and a capacitor, represented by the letter C, connected together.
+The circuit can act as an electrical resonator, an electrical analogue of a
+tuning fork, storing energy oscillating at the circuit's resonant frequency.
+Source: https://en.wikipedia.org/wiki/LC_circuit
+"""
+
+from __future__ import annotations
+
+from math import pi, sqrt
+
+
+def resonant_frequency(inductance: float, capacitance: float) -> tuple:
+ """
+ This function can calculate the resonant frequency of LC circuit,
+ for the given value of inductance and capacitnace.
+
+ Examples are given below:
+ >>> resonant_frequency(inductance=10, capacitance=5)
+ ('Resonant frequency', 0.022507907903927652)
+ >>> resonant_frequency(inductance=0, capacitance=5)
+ Traceback (most recent call last):
+ ...
+ ValueError: Inductance cannot be 0 or negative
+ >>> resonant_frequency(inductance=10, capacitance=0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Capacitance cannot be 0 or negative
+ """
+
+ if inductance <= 0:
+ raise ValueError("Inductance cannot be 0 or negative")
+
+ elif capacitance <= 0:
+ raise ValueError("Capacitance cannot be 0 or negative")
+
+ else:
+ return (
+ "Resonant frequency",
+ float(1 / (2 * pi * (sqrt(inductance * capacitance)))),
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/electronics/wheatstone_bridge.py b/electronics/wheatstone_bridge.py
new file mode 100644
index 000000000000..3529a09339c4
--- /dev/null
+++ b/electronics/wheatstone_bridge.py
@@ -0,0 +1,41 @@
+# https://en.wikipedia.org/wiki/Wheatstone_bridge
+from __future__ import annotations
+
+
+def wheatstone_solver(
+ resistance_1: float, resistance_2: float, resistance_3: float
+) -> float:
+ """
+ This function can calculate the unknown resistance in an wheatstone network,
+ given that the three other resistances in the network are known.
+ The formula to calculate the same is:
+
+ ---------------
+ |Rx=(R2/R1)*R3|
+ ---------------
+
+ Usage examples:
+ >>> wheatstone_solver(resistance_1=2, resistance_2=4, resistance_3=5)
+ 10.0
+ >>> wheatstone_solver(resistance_1=356, resistance_2=234, resistance_3=976)
+ 641.5280898876405
+ >>> wheatstone_solver(resistance_1=2, resistance_2=-1, resistance_3=2)
+ Traceback (most recent call last):
+ ...
+ ValueError: All resistance values must be positive
+ >>> wheatstone_solver(resistance_1=0, resistance_2=0, resistance_3=2)
+ Traceback (most recent call last):
+ ...
+ ValueError: All resistance values must be positive
+ """
+
+ if resistance_1 <= 0 or resistance_2 <= 0 or resistance_3 <= 0:
+ raise ValueError("All resistance values must be positive")
+ else:
+ return float((resistance_2 / resistance_1) * resistance_3)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/file_transfer/receive_file.py b/file_transfer/receive_file.py
index cfba6ed88484..f50ad9fe1107 100644
--- a/file_transfer/receive_file.py
+++ b/file_transfer/receive_file.py
@@ -1,8 +1,9 @@
-if __name__ == "__main__":
- import socket # Import socket module
+import socket
+
- sock = socket.socket() # Create a socket object
- host = socket.gethostname() # Get local machine name
+def main():
+ sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
+ host = socket.gethostname()
port = 12312
sock.connect((host, port))
@@ -13,11 +14,14 @@
print("Receiving data...")
while True:
data = sock.recv(1024)
- print(f"data={data}")
if not data:
break
- out_file.write(data) # Write data to a file
+ out_file.write(data)
- print("Successfully got the file")
+ print("Successfully received the file")
sock.close()
print("Connection closed")
+
+
+if __name__ == "__main__":
+ main()
diff --git a/file_transfer/send_file.py b/file_transfer/send_file.py
index 5b53471dfb50..1c56e48f47a1 100644
--- a/file_transfer/send_file.py
+++ b/file_transfer/send_file.py
@@ -13,7 +13,7 @@ def send_file(filename: str = "mytext.txt", testing: bool = False) -> None:
conn, addr = sock.accept() # Establish connection with client.
print(f"Got connection from {addr}")
data = conn.recv(1024)
- print(f"Server received {data}")
+ print(f"Server received: {data = }")
with open(filename, "rb") as in_file:
data = in_file.read(1024)
diff --git a/financial/README.md b/financial/README.md
new file mode 100644
index 000000000000..e5d3a84c8381
--- /dev/null
+++ b/financial/README.md
@@ -0,0 +1,4 @@
+# Interest
+
+* Compound Interest: "Compound interest is calculated by multiplying the initial principal amount by one plus the annual interest rate raised to the number of compound periods minus one." [Compound Interest](https://www.investopedia.com/)
+* Simple Interest: "Simple interest paid or received over a certain period is a fixed percentage of the principal amount that was borrowed or lent. " [Simple Interest](https://www.investopedia.com/)
diff --git a/financial/__init__.py b/financial/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/financial/equated_monthly_installments.py b/financial/equated_monthly_installments.py
new file mode 100644
index 000000000000..3af9224930b5
--- /dev/null
+++ b/financial/equated_monthly_installments.py
@@ -0,0 +1,61 @@
+"""
+Program to calculate the amortization amount per month, given
+- Principal borrowed
+- Rate of interest per annum
+- Years to repay the loan
+
+Wikipedia Reference: https://en.wikipedia.org/wiki/Equated_monthly_installment
+"""
+
+
+def equated_monthly_installments(
+ principal: float, rate_per_annum: float, years_to_repay: int
+) -> float:
+ """
+ Formula for amortization amount per month:
+ A = p * r * (1 + r)^n / ((1 + r)^n - 1)
+ where p is the principal, r is the rate of interest per month
+ and n is the number of payments
+
+ >>> equated_monthly_installments(25000, 0.12, 3)
+ 830.3577453212793
+ >>> equated_monthly_installments(25000, 0.12, 10)
+ 358.67737100646826
+ >>> equated_monthly_installments(0, 0.12, 3)
+ Traceback (most recent call last):
+ ...
+ Exception: Principal borrowed must be > 0
+ >>> equated_monthly_installments(25000, -1, 3)
+ Traceback (most recent call last):
+ ...
+ Exception: Rate of interest must be >= 0
+ >>> equated_monthly_installments(25000, 0.12, 0)
+ Traceback (most recent call last):
+ ...
+ Exception: Years to repay must be an integer > 0
+ """
+ if principal <= 0:
+ raise Exception("Principal borrowed must be > 0")
+ if rate_per_annum < 0:
+ raise Exception("Rate of interest must be >= 0")
+ if years_to_repay <= 0 or not isinstance(years_to_repay, int):
+ raise Exception("Years to repay must be an integer > 0")
+
+ # Yearly rate is divided by 12 to get monthly rate
+ rate_per_month = rate_per_annum / 12
+
+ # Years to repay is multiplied by 12 to get number of payments as payment is monthly
+ number_of_payments = years_to_repay * 12
+
+ return (
+ principal
+ * rate_per_month
+ * (1 + rate_per_month) ** number_of_payments
+ / ((1 + rate_per_month) ** number_of_payments - 1)
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/financial/exponential_moving_average.py b/financial/exponential_moving_average.py
new file mode 100644
index 000000000000..b56eb2712415
--- /dev/null
+++ b/financial/exponential_moving_average.py
@@ -0,0 +1,73 @@
+"""
+Calculate the exponential moving average (EMA) on the series of stock prices.
+Wikipedia Reference: https://en.wikipedia.org/wiki/Exponential_smoothing
+https://www.investopedia.com/terms/e/ema.asp#toc-what-is-an-exponential
+-moving-average-ema
+
+Exponential moving average is used in finance to analyze changes stock prices.
+EMA is used in conjunction with Simple moving average (SMA), EMA reacts to the
+changes in the value quicker than SMA, which is one of the advantages of using EMA.
+"""
+
+from collections.abc import Iterator
+
+
+def exponential_moving_average(
+ stock_prices: Iterator[float], window_size: int
+) -> Iterator[float]:
+ """
+ Yields exponential moving averages of the given stock prices.
+ >>> tuple(exponential_moving_average(iter([2, 5, 3, 8.2, 6, 9, 10]), 3))
+ (2, 3.5, 3.25, 5.725, 5.8625, 7.43125, 8.715625)
+
+ :param stock_prices: A stream of stock prices
+ :param window_size: The number of stock prices that will trigger a new calculation
+ of the exponential average (window_size > 0)
+ :return: Yields a sequence of exponential moving averages
+
+ Formula:
+
+ st = alpha * xt + (1 - alpha) * st_prev
+
+ Where,
+ st : Exponential moving average at timestamp t
+ xt : stock price in from the stock prices at timestamp t
+ st_prev : Exponential moving average at timestamp t-1
+ alpha : 2/(1 + window_size) - smoothing factor
+
+ Exponential moving average (EMA) is a rule of thumb technique for
+ smoothing time series data using an exponential window function.
+ """
+
+ if window_size <= 0:
+ raise ValueError("window_size must be > 0")
+
+ # Calculating smoothing factor
+ alpha = 2 / (1 + window_size)
+
+ # Exponential average at timestamp t
+ moving_average = 0.0
+
+ for i, stock_price in enumerate(stock_prices):
+ if i <= window_size:
+ # Assigning simple moving average till the window_size for the first time
+ # is reached
+ moving_average = (moving_average + stock_price) * 0.5 if i else stock_price
+ else:
+ # Calculating exponential moving average based on current timestamp data
+ # point and previous exponential average value
+ moving_average = (alpha * stock_price) + ((1 - alpha) * moving_average)
+ yield moving_average
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ stock_prices = [2.0, 5, 3, 8.2, 6, 9, 10]
+ window_size = 3
+ result = tuple(exponential_moving_average(iter(stock_prices), window_size))
+ print(f"{stock_prices = }")
+ print(f"{window_size = }")
+ print(f"{result = }")
diff --git a/financial/interest.py b/financial/interest.py
new file mode 100644
index 000000000000..33d02e27ccb3
--- /dev/null
+++ b/financial/interest.py
@@ -0,0 +1,120 @@
+# https://www.investopedia.com
+
+from __future__ import annotations
+
+
+def simple_interest(
+ principal: float, daily_interest_rate: float, days_between_payments: float
+) -> float:
+ """
+ >>> simple_interest(18000.0, 0.06, 3)
+ 3240.0
+ >>> simple_interest(0.5, 0.06, 3)
+ 0.09
+ >>> simple_interest(18000.0, 0.01, 10)
+ 1800.0
+ >>> simple_interest(18000.0, 0.0, 3)
+ 0.0
+ >>> simple_interest(5500.0, 0.01, 100)
+ 5500.0
+ >>> simple_interest(10000.0, -0.06, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: daily_interest_rate must be >= 0
+ >>> simple_interest(-10000.0, 0.06, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: principal must be > 0
+ >>> simple_interest(5500.0, 0.01, -5)
+ Traceback (most recent call last):
+ ...
+ ValueError: days_between_payments must be > 0
+ """
+ if days_between_payments <= 0:
+ raise ValueError("days_between_payments must be > 0")
+ if daily_interest_rate < 0:
+ raise ValueError("daily_interest_rate must be >= 0")
+ if principal <= 0:
+ raise ValueError("principal must be > 0")
+ return principal * daily_interest_rate * days_between_payments
+
+
+def compound_interest(
+ principal: float,
+ nominal_annual_interest_rate_percentage: float,
+ number_of_compounding_periods: float,
+) -> float:
+ """
+ >>> compound_interest(10000.0, 0.05, 3)
+ 1576.2500000000014
+ >>> compound_interest(10000.0, 0.05, 1)
+ 500.00000000000045
+ >>> compound_interest(0.5, 0.05, 3)
+ 0.07881250000000006
+ >>> compound_interest(10000.0, 0.06, -4)
+ Traceback (most recent call last):
+ ...
+ ValueError: number_of_compounding_periods must be > 0
+ >>> compound_interest(10000.0, -3.5, 3.0)
+ Traceback (most recent call last):
+ ...
+ ValueError: nominal_annual_interest_rate_percentage must be >= 0
+ >>> compound_interest(-5500.0, 0.01, 5)
+ Traceback (most recent call last):
+ ...
+ ValueError: principal must be > 0
+ """
+ if number_of_compounding_periods <= 0:
+ raise ValueError("number_of_compounding_periods must be > 0")
+ if nominal_annual_interest_rate_percentage < 0:
+ raise ValueError("nominal_annual_interest_rate_percentage must be >= 0")
+ if principal <= 0:
+ raise ValueError("principal must be > 0")
+
+ return principal * (
+ (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods
+ - 1
+ )
+
+
+def apr_interest(
+ principal: float,
+ nominal_annual_percentage_rate: float,
+ number_of_years: float,
+) -> float:
+ """
+ >>> apr_interest(10000.0, 0.05, 3)
+ 1618.223072263547
+ >>> apr_interest(10000.0, 0.05, 1)
+ 512.6749646744732
+ >>> apr_interest(0.5, 0.05, 3)
+ 0.08091115361317736
+ >>> apr_interest(10000.0, 0.06, -4)
+ Traceback (most recent call last):
+ ...
+ ValueError: number_of_years must be > 0
+ >>> apr_interest(10000.0, -3.5, 3.0)
+ Traceback (most recent call last):
+ ...
+ ValueError: nominal_annual_percentage_rate must be >= 0
+ >>> apr_interest(-5500.0, 0.01, 5)
+ Traceback (most recent call last):
+ ...
+ ValueError: principal must be > 0
+ """
+ if number_of_years <= 0:
+ raise ValueError("number_of_years must be > 0")
+ if nominal_annual_percentage_rate < 0:
+ raise ValueError("nominal_annual_percentage_rate must be >= 0")
+ if principal <= 0:
+ raise ValueError("principal must be > 0")
+
+ return compound_interest(
+ principal, nominal_annual_percentage_rate / 365, number_of_years * 365
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/financial/present_value.py b/financial/present_value.py
new file mode 100644
index 000000000000..f74612b923af
--- /dev/null
+++ b/financial/present_value.py
@@ -0,0 +1,42 @@
+"""
+Reference: https://www.investopedia.com/terms/p/presentvalue.asp
+
+An algorithm that calculates the present value of a stream of yearly cash flows given...
+1. The discount rate (as a decimal, not a percent)
+2. An array of cash flows, with the index of the cash flow being the associated year
+
+Note: This algorithm assumes that cash flows are paid at the end of the specified year
+"""
+
+
+def present_value(discount_rate: float, cash_flows: list[float]) -> float:
+ """
+ >>> present_value(0.13, [10, 20.70, -293, 297])
+ 4.69
+ >>> present_value(0.07, [-109129.39, 30923.23, 15098.93, 29734,39])
+ -42739.63
+ >>> present_value(0.07, [109129.39, 30923.23, 15098.93, 29734,39])
+ 175519.15
+ >>> present_value(-1, [109129.39, 30923.23, 15098.93, 29734,39])
+ Traceback (most recent call last):
+ ...
+ ValueError: Discount rate cannot be negative
+ >>> present_value(0.03, [])
+ Traceback (most recent call last):
+ ...
+ ValueError: Cash flows list cannot be empty
+ """
+ if discount_rate < 0:
+ raise ValueError("Discount rate cannot be negative")
+ if not cash_flows:
+ raise ValueError("Cash flows list cannot be empty")
+ present_value = sum(
+ cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(cash_flows)
+ )
+ return round(present_value, ndigits=2)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/financial/price_plus_tax.py b/financial/price_plus_tax.py
new file mode 100644
index 000000000000..43876d35e57c
--- /dev/null
+++ b/financial/price_plus_tax.py
@@ -0,0 +1,18 @@
+"""
+Calculate price plus tax of a good or service given its price and a tax rate.
+"""
+
+
+def price_plus_tax(price: float, tax_rate: float) -> float:
+ """
+ >>> price_plus_tax(100, 0.25)
+ 125.0
+ >>> price_plus_tax(125.50, 0.05)
+ 131.775
+ """
+ return price * (1 + tax_rate)
+
+
+if __name__ == "__main__":
+ print(f"{price_plus_tax(100, 0.25) = }")
+ print(f"{price_plus_tax(125.50, 0.05) = }")
diff --git a/financial/simple_moving_average.py b/financial/simple_moving_average.py
new file mode 100644
index 000000000000..f5ae444fd027
--- /dev/null
+++ b/financial/simple_moving_average.py
@@ -0,0 +1,69 @@
+"""
+The Simple Moving Average (SMA) is a statistical calculation used to analyze data points
+by creating a constantly updated average price over a specific time period.
+In finance, SMA is often used in time series analysis to smooth out price data
+and identify trends.
+
+Reference: https://en.wikipedia.org/wiki/Moving_average
+"""
+
+from collections.abc import Sequence
+
+
+def simple_moving_average(
+ data: Sequence[float], window_size: int
+) -> list[float | None]:
+ """
+ Calculate the simple moving average (SMA) for some given time series data.
+
+ :param data: A list of numerical data points.
+ :param window_size: An integer representing the size of the SMA window.
+ :return: A list of SMA values with the same length as the input data.
+
+ Examples:
+ >>> sma = simple_moving_average([10, 12, 15, 13, 14, 16, 18, 17, 19, 21], 3)
+ >>> [round(value, 2) if value is not None else None for value in sma]
+ [None, None, 12.33, 13.33, 14.0, 14.33, 16.0, 17.0, 18.0, 19.0]
+ >>> simple_moving_average([10, 12, 15], 5)
+ [None, None, None]
+ >>> simple_moving_average([10, 12, 15, 13, 14, 16, 18, 17, 19, 21], 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Window size must be a positive integer
+ """
+ if window_size < 1:
+ raise ValueError("Window size must be a positive integer")
+
+ sma: list[float | None] = []
+
+ for i in range(len(data)):
+ if i < window_size - 1:
+ sma.append(None) # SMA not available for early data points
+ else:
+ window = data[i - window_size + 1 : i + 1]
+ sma_value = sum(window) / window_size
+ sma.append(sma_value)
+ return sma
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ # Example data (replace with your own time series data)
+ data = [10, 12, 15, 13, 14, 16, 18, 17, 19, 21]
+
+ # Specify the window size for the SMA
+ window_size = 3
+
+ # Calculate the Simple Moving Average
+ sma_values = simple_moving_average(data, window_size)
+
+ # Print the SMA values
+ print("Simple Moving Average (SMA) Values:")
+ for i, value in enumerate(sma_values):
+ if value is not None:
+ print(f"Day {i + 1}: {value:.2f}")
+ else:
+ print(f"Day {i + 1}: Not enough data for SMA")
diff --git a/fractals/__init__.py b/fractals/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/fractals/julia_sets.py b/fractals/julia_sets.py
new file mode 100644
index 000000000000..bea599d44339
--- /dev/null
+++ b/fractals/julia_sets.py
@@ -0,0 +1,217 @@
+"""Author Alexandre De Zotti
+
+Draws Julia sets of quadratic polynomials and exponential maps.
+ More specifically, this iterates the function a fixed number of times
+ then plots whether the absolute value of the last iterate is greater than
+ a fixed threshold (named "escape radius"). For the exponential map this is not
+ really an escape radius but rather a convenient way to approximate the Julia
+ set with bounded orbits.
+
+The examples presented here are:
+- The Cauliflower Julia set, see e.g.
+https://en.wikipedia.org/wiki/File:Julia_z2%2B0,25.png
+- Other examples from https://en.wikipedia.org/wiki/Julia_set
+- An exponential map Julia set, ambiantly homeomorphic to the examples in
+https://www.math.univ-toulouse.fr/~cheritat/GalII/galery.html
+ and
+https://ddd.uab.cat/pub/pubmat/02141493v43n1/02141493v43n1p27.pdf
+
+Remark: Some overflow runtime warnings are suppressed. This is because of the
+ way the iteration loop is implemented, using numpy's efficient computations.
+ Overflows and infinites are replaced after each step by a large number.
+"""
+
+import warnings
+from collections.abc import Callable
+from typing import Any
+
+import matplotlib.pyplot as plt
+import numpy as np
+
+c_cauliflower = 0.25 + 0.0j
+c_polynomial_1 = -0.4 + 0.6j
+c_polynomial_2 = -0.1 + 0.651j
+c_exponential = -2.0
+nb_iterations = 56
+window_size = 2.0
+nb_pixels = 666
+
+
+def eval_exponential(c_parameter: complex, z_values: np.ndarray) -> np.ndarray:
+ """
+ Evaluate $e^z + c$.
+ >>> float(eval_exponential(0, 0))
+ 1.0
+ >>> bool(abs(eval_exponential(1, np.pi*1.j)) < 1e-15)
+ True
+ >>> bool(abs(eval_exponential(1.j, 0)-1-1.j) < 1e-15)
+ True
+ """
+ return np.exp(z_values) + c_parameter
+
+
+def eval_quadratic_polynomial(c_parameter: complex, z_values: np.ndarray) -> np.ndarray:
+ """
+ >>> eval_quadratic_polynomial(0, 2)
+ 4
+ >>> eval_quadratic_polynomial(-1, 1)
+ 0
+ >>> round(eval_quadratic_polynomial(1.j, 0).imag)
+ 1
+ >>> round(eval_quadratic_polynomial(1.j, 0).real)
+ 0
+ """
+ return z_values * z_values + c_parameter
+
+
+def prepare_grid(window_size: float, nb_pixels: int) -> np.ndarray:
+ """
+ Create a grid of complex values of size nb_pixels*nb_pixels with real and
+ imaginary parts ranging from -window_size to window_size (inclusive).
+ Returns a numpy array.
+
+ >>> prepare_grid(1,3)
+ array([[-1.-1.j, -1.+0.j, -1.+1.j],
+ [ 0.-1.j, 0.+0.j, 0.+1.j],
+ [ 1.-1.j, 1.+0.j, 1.+1.j]])
+ """
+ x = np.linspace(-window_size, window_size, nb_pixels)
+ x = x.reshape((nb_pixels, 1))
+ y = np.linspace(-window_size, window_size, nb_pixels)
+ y = y.reshape((1, nb_pixels))
+ return x + 1.0j * y
+
+
+def iterate_function(
+ eval_function: Callable[[Any, np.ndarray], np.ndarray],
+ function_params: Any,
+ nb_iterations: int,
+ z_0: np.ndarray,
+ infinity: float | None = None,
+) -> np.ndarray:
+ """
+ Iterate the function "eval_function" exactly nb_iterations times.
+ The first argument of the function is a parameter which is contained in
+ function_params. The variable z_0 is an array that contains the initial
+ values to iterate from.
+ This function returns the final iterates.
+
+ >>> iterate_function(eval_quadratic_polynomial, 0, 3, np.array([0,1,2])).shape
+ (3,)
+ >>> complex(np.round(iterate_function(eval_quadratic_polynomial,
+ ... 0,
+ ... 3,
+ ... np.array([0,1,2]))[0]))
+ 0j
+ >>> complex(np.round(iterate_function(eval_quadratic_polynomial,
+ ... 0,
+ ... 3,
+ ... np.array([0,1,2]))[1]))
+ (1+0j)
+ >>> complex(np.round(iterate_function(eval_quadratic_polynomial,
+ ... 0,
+ ... 3,
+ ... np.array([0,1,2]))[2]))
+ (256+0j)
+ """
+
+ z_n = z_0.astype("complex64")
+ for _ in range(nb_iterations):
+ z_n = eval_function(function_params, z_n)
+ if infinity is not None:
+ np.nan_to_num(z_n, copy=False, nan=infinity)
+ z_n[abs(z_n) == np.inf] = infinity
+ return z_n
+
+
+def show_results(
+ function_label: str,
+ function_params: Any,
+ escape_radius: float,
+ z_final: np.ndarray,
+) -> None:
+ """
+ Plots of whether the absolute value of z_final is greater than
+ the value of escape_radius. Adds the function_label and function_params to
+ the title.
+
+ >>> show_results('80', 0, 1, np.array([[0,1,.5],[.4,2,1.1],[.2,1,1.3]]))
+ """
+
+ abs_z_final = (abs(z_final)).transpose()
+ abs_z_final[:, :] = abs_z_final[::-1, :]
+ plt.matshow(abs_z_final < escape_radius)
+ plt.title(f"Julia set of ${function_label}$, $c={function_params}$")
+ plt.show()
+
+
+def ignore_overflow_warnings() -> None:
+ """
+ Ignore some overflow and invalid value warnings.
+
+ >>> ignore_overflow_warnings()
+ """
+ warnings.filterwarnings(
+ "ignore", category=RuntimeWarning, message="overflow encountered in multiply"
+ )
+ warnings.filterwarnings(
+ "ignore",
+ category=RuntimeWarning,
+ message="invalid value encountered in multiply",
+ )
+ warnings.filterwarnings(
+ "ignore", category=RuntimeWarning, message="overflow encountered in absolute"
+ )
+ warnings.filterwarnings(
+ "ignore", category=RuntimeWarning, message="overflow encountered in exp"
+ )
+
+
+if __name__ == "__main__":
+ z_0 = prepare_grid(window_size, nb_pixels)
+
+ ignore_overflow_warnings() # See file header for explanations
+
+ nb_iterations = 24
+ escape_radius = 2 * abs(c_cauliflower) + 1
+ z_final = iterate_function(
+ eval_quadratic_polynomial,
+ c_cauliflower,
+ nb_iterations,
+ z_0,
+ infinity=1.1 * escape_radius,
+ )
+ show_results("z^2+c", c_cauliflower, escape_radius, z_final)
+
+ nb_iterations = 64
+ escape_radius = 2 * abs(c_polynomial_1) + 1
+ z_final = iterate_function(
+ eval_quadratic_polynomial,
+ c_polynomial_1,
+ nb_iterations,
+ z_0,
+ infinity=1.1 * escape_radius,
+ )
+ show_results("z^2+c", c_polynomial_1, escape_radius, z_final)
+
+ nb_iterations = 161
+ escape_radius = 2 * abs(c_polynomial_2) + 1
+ z_final = iterate_function(
+ eval_quadratic_polynomial,
+ c_polynomial_2,
+ nb_iterations,
+ z_0,
+ infinity=1.1 * escape_radius,
+ )
+ show_results("z^2+c", c_polynomial_2, escape_radius, z_final)
+
+ nb_iterations = 12
+ escape_radius = 10000.0
+ z_final = iterate_function(
+ eval_exponential,
+ c_exponential,
+ nb_iterations,
+ z_0 + 2,
+ infinity=1.0e10,
+ )
+ show_results("e^z+c", c_exponential, escape_radius, z_final)
diff --git a/fractals/koch_snowflake.py b/fractals/koch_snowflake.py
new file mode 100644
index 000000000000..724b78f41a69
--- /dev/null
+++ b/fractals/koch_snowflake.py
@@ -0,0 +1,115 @@
+"""
+Description
+ The Koch snowflake is a fractal curve and one of the earliest fractals to
+ have been described. The Koch snowflake can be built up iteratively, in a
+ sequence of stages. The first stage is an equilateral triangle, and each
+ successive stage is formed by adding outward bends to each side of the
+ previous stage, making smaller equilateral triangles.
+ This can be achieved through the following steps for each line:
+ 1. divide the line segment into three segments of equal length.
+ 2. draw an equilateral triangle that has the middle segment from step 1
+ as its base and points outward.
+ 3. remove the line segment that is the base of the triangle from step 2.
+ (description adapted from https://en.wikipedia.org/wiki/Koch_snowflake )
+ (for a more detailed explanation and an implementation in the
+ Processing language, see https://natureofcode.com/book/chapter-8-fractals/
+ #84-the-koch-curve-and-the-arraylist-technique )
+
+Requirements (pip):
+ - matplotlib
+ - numpy
+"""
+
+from __future__ import annotations
+
+import matplotlib.pyplot as plt
+import numpy as np
+
+# initial triangle of Koch snowflake
+VECTOR_1 = np.array([0, 0])
+VECTOR_2 = np.array([0.5, 0.8660254])
+VECTOR_3 = np.array([1, 0])
+INITIAL_VECTORS = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1]
+
+# uncomment for simple Koch curve instead of Koch snowflake
+# INITIAL_VECTORS = [VECTOR_1, VECTOR_3]
+
+
+def iterate(initial_vectors: list[np.ndarray], steps: int) -> list[np.ndarray]:
+ """
+ Go through the number of iterations determined by the argument "steps".
+ Be careful with high values (above 5) since the time to calculate increases
+ exponentially.
+ >>> iterate([np.array([0, 0]), np.array([1, 0])], 1)
+ [array([0, 0]), array([0.33333333, 0. ]), array([0.5 , \
+0.28867513]), array([0.66666667, 0. ]), array([1, 0])]
+ """
+ vectors = initial_vectors
+ for _ in range(steps):
+ vectors = iteration_step(vectors)
+ return vectors
+
+
+def iteration_step(vectors: list[np.ndarray]) -> list[np.ndarray]:
+ """
+ Loops through each pair of adjacent vectors. Each line between two adjacent
+ vectors is divided into 4 segments by adding 3 additional vectors in-between
+ the original two vectors. The vector in the middle is constructed through a
+ 60 degree rotation so it is bent outwards.
+ >>> iteration_step([np.array([0, 0]), np.array([1, 0])])
+ [array([0, 0]), array([0.33333333, 0. ]), array([0.5 , \
+0.28867513]), array([0.66666667, 0. ]), array([1, 0])]
+ """
+ new_vectors = []
+ for i, start_vector in enumerate(vectors[:-1]):
+ end_vector = vectors[i + 1]
+ new_vectors.append(start_vector)
+ difference_vector = end_vector - start_vector
+ new_vectors.append(start_vector + difference_vector / 3)
+ new_vectors.append(
+ start_vector + difference_vector / 3 + rotate(difference_vector / 3, 60)
+ )
+ new_vectors.append(start_vector + difference_vector * 2 / 3)
+ new_vectors.append(vectors[-1])
+ return new_vectors
+
+
+def rotate(vector: np.ndarray, angle_in_degrees: float) -> np.ndarray:
+ """
+ Standard rotation of a 2D vector with a rotation matrix
+ (see https://en.wikipedia.org/wiki/Rotation_matrix )
+ >>> rotate(np.array([1, 0]), 60)
+ array([0.5 , 0.8660254])
+ >>> rotate(np.array([1, 0]), 90)
+ array([6.123234e-17, 1.000000e+00])
+ """
+ theta = np.radians(angle_in_degrees)
+ c, s = np.cos(theta), np.sin(theta)
+ rotation_matrix = np.array(((c, -s), (s, c)))
+ return np.dot(rotation_matrix, vector)
+
+
+def plot(vectors: list[np.ndarray]) -> None:
+ """
+ Utility function to plot the vectors using matplotlib.pyplot
+ No doctest was implemented since this function does not have a return value
+ """
+ # avoid stretched display of graph
+ axes = plt.gca()
+ axes.set_aspect("equal")
+
+ # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all
+ # y-coordinates as inputs, which are constructed from the vector-list using
+ # zip()
+ x_coordinates, y_coordinates = zip(*vectors)
+ plt.plot(x_coordinates, y_coordinates)
+ plt.show()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ processed_vectors = iterate(INITIAL_VECTORS, 5)
+ plot(processed_vectors)
diff --git a/fractals/mandelbrot.py b/fractals/mandelbrot.py
new file mode 100644
index 000000000000..359d965a882d
--- /dev/null
+++ b/fractals/mandelbrot.py
@@ -0,0 +1,150 @@
+"""
+The Mandelbrot set is the set of complex numbers "c" for which the series
+"z_(n+1) = z_n * z_n + c" does not diverge, i.e. remains bounded. Thus, a
+complex number "c" is a member of the Mandelbrot set if, when starting with
+"z_0 = 0" and applying the iteration repeatedly, the absolute value of
+"z_n" remains bounded for all "n > 0". Complex numbers can be written as
+"a + b*i": "a" is the real component, usually drawn on the x-axis, and "b*i"
+is the imaginary component, usually drawn on the y-axis. Most visualizations
+of the Mandelbrot set use a color-coding to indicate after how many steps in
+the series the numbers outside the set diverge. Images of the Mandelbrot set
+exhibit an elaborate and infinitely complicated boundary that reveals
+progressively ever-finer recursive detail at increasing magnifications, making
+the boundary of the Mandelbrot set a fractal curve.
+(description adapted from https://en.wikipedia.org/wiki/Mandelbrot_set )
+(see also https://en.wikipedia.org/wiki/Plotting_algorithms_for_the_Mandelbrot_set )
+"""
+
+import colorsys
+
+from PIL import Image
+
+
+def get_distance(x: float, y: float, max_step: int) -> float:
+ """
+ Return the relative distance (= step/max_step) after which the complex number
+ constituted by this x-y-pair diverges. Members of the Mandelbrot set do not
+ diverge so their distance is 1.
+
+ >>> get_distance(0, 0, 50)
+ 1.0
+ >>> get_distance(0.5, 0.5, 50)
+ 0.061224489795918366
+ >>> get_distance(2, 0, 50)
+ 0.0
+ """
+ a = x
+ b = y
+ for step in range(max_step): # noqa: B007
+ a_new = a * a - b * b + x
+ b = 2 * a * b + y
+ a = a_new
+
+ # divergence happens for all complex number with an absolute value
+ # greater than 4
+ if a * a + b * b > 4:
+ break
+ return step / (max_step - 1)
+
+
+def get_black_and_white_rgb(distance: float) -> tuple:
+ """
+ Black&white color-coding that ignores the relative distance. The Mandelbrot
+ set is black, everything else is white.
+
+ >>> get_black_and_white_rgb(0)
+ (255, 255, 255)
+ >>> get_black_and_white_rgb(0.5)
+ (255, 255, 255)
+ >>> get_black_and_white_rgb(1)
+ (0, 0, 0)
+ """
+ if distance == 1:
+ return (0, 0, 0)
+ else:
+ return (255, 255, 255)
+
+
+def get_color_coded_rgb(distance: float) -> tuple:
+ """
+ Color-coding taking the relative distance into account. The Mandelbrot set
+ is black.
+
+ >>> get_color_coded_rgb(0)
+ (255, 0, 0)
+ >>> get_color_coded_rgb(0.5)
+ (0, 255, 255)
+ >>> get_color_coded_rgb(1)
+ (0, 0, 0)
+ """
+ if distance == 1:
+ return (0, 0, 0)
+ else:
+ return tuple(round(i * 255) for i in colorsys.hsv_to_rgb(distance, 1, 1))
+
+
+def get_image(
+ image_width: int = 800,
+ image_height: int = 600,
+ figure_center_x: float = -0.6,
+ figure_center_y: float = 0,
+ figure_width: float = 3.2,
+ max_step: int = 50,
+ use_distance_color_coding: bool = True,
+) -> Image.Image:
+ """
+ Function to generate the image of the Mandelbrot set. Two types of coordinates
+ are used: image-coordinates that refer to the pixels and figure-coordinates
+ that refer to the complex numbers inside and outside the Mandelbrot set. The
+ figure-coordinates in the arguments of this function determine which section
+ of the Mandelbrot set is viewed. The main area of the Mandelbrot set is
+ roughly between "-1.5 < x < 0.5" and "-1 < y < 1" in the figure-coordinates.
+
+ Commenting out tests that slow down pytest...
+ # 13.35s call fractals/mandelbrot.py::mandelbrot.get_image
+ # >>> get_image().load()[0,0]
+ (255, 0, 0)
+ # >>> get_image(use_distance_color_coding = False).load()[0,0]
+ (255, 255, 255)
+ """
+ img = Image.new("RGB", (image_width, image_height))
+ pixels = img.load()
+
+ # loop through the image-coordinates
+ for image_x in range(image_width):
+ for image_y in range(image_height):
+ # determine the figure-coordinates based on the image-coordinates
+ figure_height = figure_width / image_width * image_height
+ figure_x = figure_center_x + (image_x / image_width - 0.5) * figure_width
+ figure_y = figure_center_y + (image_y / image_height - 0.5) * figure_height
+
+ distance = get_distance(figure_x, figure_y, max_step)
+
+ # color the corresponding pixel based on the selected coloring-function
+ if use_distance_color_coding:
+ pixels[image_x, image_y] = get_color_coded_rgb(distance)
+ else:
+ pixels[image_x, image_y] = get_black_and_white_rgb(distance)
+
+ return img
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ # colored version, full figure
+ img = get_image()
+
+ # uncomment for colored version, different section, zoomed in
+ # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4,
+ # figure_width = 0.8)
+
+ # uncomment for black and white version, full figure
+ # img = get_image(use_distance_color_coding = False)
+
+ # uncomment to save the image
+ # img.save("mandelbrot.png")
+
+ img.show()
diff --git a/fractals/sierpinski_triangle.py b/fractals/sierpinski_triangle.py
new file mode 100644
index 000000000000..ceb2001b681d
--- /dev/null
+++ b/fractals/sierpinski_triangle.py
@@ -0,0 +1,86 @@
+"""
+Author Anurag Kumar | anuragkumarak95@gmail.com | git/anuragkumarak95
+
+Simple example of fractal generation using recursion.
+
+What is the Sierpiński Triangle?
+ The Sierpiński triangle (sometimes spelled Sierpinski), also called the
+Sierpiński gasket or Sierpiński sieve, is a fractal attractive fixed set with
+the overall shape of an equilateral triangle, subdivided recursively into
+smaller equilateral triangles. Originally constructed as a curve, this is one of
+the basic examples of self-similar sets—that is, it is a mathematically
+generated pattern that is reproducible at any magnification or reduction. It is
+named after the Polish mathematician Wacław Sierpiński, but appeared as a
+decorative pattern many centuries before the work of Sierpiński.
+
+
+Usage: python sierpinski_triangle.py
+
+Credits:
+ The above description is taken from
+ https://en.wikipedia.org/wiki/Sierpi%C5%84ski_triangle
+ This code was written by editing the code from
+ https://www.riannetrujillo.com/blog/python-fractal/
+"""
+
+import sys
+import turtle
+
+
+def get_mid(p1: tuple[float, float], p2: tuple[float, float]) -> tuple[float, float]:
+ """
+ Find the midpoint of two points
+
+ >>> get_mid((0, 0), (2, 2))
+ (1.0, 1.0)
+ >>> get_mid((-3, -3), (3, 3))
+ (0.0, 0.0)
+ >>> get_mid((1, 0), (3, 2))
+ (2.0, 1.0)
+ >>> get_mid((0, 0), (1, 1))
+ (0.5, 0.5)
+ >>> get_mid((0, 0), (0, 0))
+ (0.0, 0.0)
+ """
+ return (p1[0] + p2[0]) / 2, (p1[1] + p2[1]) / 2
+
+
+def triangle(
+ vertex1: tuple[float, float],
+ vertex2: tuple[float, float],
+ vertex3: tuple[float, float],
+ depth: int,
+) -> None:
+ """
+ Recursively draw the Sierpinski triangle given the vertices of the triangle
+ and the recursion depth
+ """
+ my_pen.up()
+ my_pen.goto(vertex1[0], vertex1[1])
+ my_pen.down()
+ my_pen.goto(vertex2[0], vertex2[1])
+ my_pen.goto(vertex3[0], vertex3[1])
+ my_pen.goto(vertex1[0], vertex1[1])
+
+ if depth == 0:
+ return
+
+ triangle(vertex1, get_mid(vertex1, vertex2), get_mid(vertex1, vertex3), depth - 1)
+ triangle(vertex2, get_mid(vertex1, vertex2), get_mid(vertex2, vertex3), depth - 1)
+ triangle(vertex3, get_mid(vertex3, vertex2), get_mid(vertex1, vertex3), depth - 1)
+
+
+if __name__ == "__main__":
+ if len(sys.argv) != 2:
+ raise ValueError(
+ "Correct format for using this script: "
+ "python fractals.py "
+ )
+ my_pen = turtle.Turtle()
+ my_pen.ht()
+ my_pen.speed(5)
+ my_pen.pencolor("red")
+
+ vertices = [(-175, -125), (0, 175), (175, -125)] # vertices of triangle
+ triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
+ turtle.Screen().exitonclick()
diff --git a/fractals/vicsek.py b/fractals/vicsek.py
new file mode 100644
index 000000000000..290fe95b79b4
--- /dev/null
+++ b/fractals/vicsek.py
@@ -0,0 +1,76 @@
+"""Authors Bastien Capiaux & Mehdi Oudghiri
+
+The Vicsek fractal algorithm is a recursive algorithm that creates a
+pattern known as the Vicsek fractal or the Vicsek square.
+It is based on the concept of self-similarity, where the pattern at each
+level of recursion resembles the overall pattern.
+The algorithm involves dividing a square into 9 equal smaller squares,
+removing the center square, and then repeating this process on the remaining 8 squares.
+This results in a pattern that exhibits self-similarity and has a
+square-shaped outline with smaller squares within it.
+
+Source: https://en.wikipedia.org/wiki/Vicsek_fractal
+"""
+
+import turtle
+
+
+def draw_cross(x: float, y: float, length: float):
+ """
+ Draw a cross at the specified position and with the specified length.
+ """
+ turtle.up()
+ turtle.goto(x - length / 2, y - length / 6)
+ turtle.down()
+ turtle.seth(0)
+ turtle.begin_fill()
+ for _ in range(4):
+ turtle.fd(length / 3)
+ turtle.right(90)
+ turtle.fd(length / 3)
+ turtle.left(90)
+ turtle.fd(length / 3)
+ turtle.left(90)
+ turtle.end_fill()
+
+
+def draw_fractal_recursive(x: float, y: float, length: float, depth: float):
+ """
+ Recursively draw the Vicsek fractal at the specified position, with the
+ specified length and depth.
+ """
+ if depth == 0:
+ draw_cross(x, y, length)
+ return
+
+ draw_fractal_recursive(x, y, length / 3, depth - 1)
+ draw_fractal_recursive(x + length / 3, y, length / 3, depth - 1)
+ draw_fractal_recursive(x - length / 3, y, length / 3, depth - 1)
+ draw_fractal_recursive(x, y + length / 3, length / 3, depth - 1)
+ draw_fractal_recursive(x, y - length / 3, length / 3, depth - 1)
+
+
+def set_color(rgb: str):
+ turtle.color(rgb)
+
+
+def draw_vicsek_fractal(x: float, y: float, length: float, depth: float, color="blue"):
+ """
+ Draw the Vicsek fractal at the specified position, with the specified
+ length and depth.
+ """
+ turtle.speed(0)
+ turtle.hideturtle()
+ set_color(color)
+ draw_fractal_recursive(x, y, length, depth)
+ turtle.Screen().update()
+
+
+def main():
+ draw_vicsek_fractal(0, 0, 800, 4)
+
+ turtle.done()
+
+
+if __name__ == "__main__":
+ main()
diff --git a/fuzzy_logic/fuzzy_operations.py b/fuzzy_logic/fuzzy_operations.py
index 0f573f158663..c5e4cbde019d 100644
--- a/fuzzy_logic/fuzzy_operations.py
+++ b/fuzzy_logic/fuzzy_operations.py
@@ -1,102 +1,195 @@
-"""README, Author - Jigyasa Gandhi(mailto:jigsgandhi97@gmail.com)
-Requirements:
- - scikit-fuzzy
- - numpy
- - matplotlib
-Python:
- - 3.5
"""
+By @Shreya123714
+
+https://en.wikipedia.org/wiki/Fuzzy_set
+"""
+
+from __future__ import annotations
+
+from dataclasses import dataclass
+
+import matplotlib.pyplot as plt
import numpy as np
-import skfuzzy as fuzz
+
+
+@dataclass
+class FuzzySet:
+ """
+ A class for representing and manipulating triangular fuzzy sets.
+ Attributes:
+ name: The name or label of the fuzzy set.
+ left_boundary: The left boundary of the fuzzy set.
+ peak: The peak (central) value of the fuzzy set.
+ right_boundary: The right boundary of the fuzzy set.
+ Methods:
+ membership(x): Calculate the membership value of an input 'x' in the fuzzy set.
+ union(other): Calculate the union of this fuzzy set with another fuzzy set.
+ intersection(other): Calculate the intersection of this fuzzy set with another.
+ complement(): Calculate the complement (negation) of this fuzzy set.
+ plot(): Plot the membership function of the fuzzy set.
+
+ >>> sheru = FuzzySet("Sheru", 0.4, 1, 0.6)
+ >>> sheru
+ FuzzySet(name='Sheru', left_boundary=0.4, peak=1, right_boundary=0.6)
+ >>> str(sheru)
+ 'Sheru: [0.4, 1, 0.6]'
+
+ >>> siya = FuzzySet("Siya", 0.5, 1, 0.7)
+ >>> siya
+ FuzzySet(name='Siya', left_boundary=0.5, peak=1, right_boundary=0.7)
+
+ # Complement Operation
+ >>> sheru.complement()
+ FuzzySet(name='¬Sheru', left_boundary=0.4, peak=0.6, right_boundary=0)
+ >>> siya.complement() # doctest: +NORMALIZE_WHITESPACE
+ FuzzySet(name='¬Siya', left_boundary=0.30000000000000004, peak=0.5,
+ right_boundary=0)
+
+ # Intersection Operation
+ >>> siya.intersection(sheru)
+ FuzzySet(name='Siya ∩ Sheru', left_boundary=0.5, peak=0.6, right_boundary=1.0)
+
+ # Membership Operation
+ >>> sheru.membership(0.5)
+ 0.16666666666666663
+ >>> sheru.membership(0.6)
+ 0.0
+
+ # Union Operations
+ >>> siya.union(sheru)
+ FuzzySet(name='Siya U Sheru', left_boundary=0.4, peak=0.7, right_boundary=1.0)
+ """
+
+ name: str
+ left_boundary: float
+ peak: float
+ right_boundary: float
+
+ def __str__(self) -> str:
+ """
+ >>> FuzzySet("fuzzy_set", 0.1, 0.2, 0.3)
+ FuzzySet(name='fuzzy_set', left_boundary=0.1, peak=0.2, right_boundary=0.3)
+ """
+ return (
+ f"{self.name}: [{self.left_boundary}, {self.peak}, {self.right_boundary}]"
+ )
+
+ def complement(self) -> FuzzySet:
+ """
+ Calculate the complement (negation) of this fuzzy set.
+ Returns:
+ FuzzySet: A new fuzzy set representing the complement.
+
+ >>> FuzzySet("fuzzy_set", 0.1, 0.2, 0.3).complement()
+ FuzzySet(name='¬fuzzy_set', left_boundary=0.7, peak=0.9, right_boundary=0.8)
+ """
+ return FuzzySet(
+ f"¬{self.name}",
+ 1 - self.right_boundary,
+ 1 - self.left_boundary,
+ 1 - self.peak,
+ )
+
+ def intersection(self, other) -> FuzzySet:
+ """
+ Calculate the intersection of this fuzzy set
+ with another fuzzy set.
+ Args:
+ other: Another fuzzy set to intersect with.
+ Returns:
+ A new fuzzy set representing the intersection.
+
+ >>> FuzzySet("a", 0.1, 0.2, 0.3).intersection(FuzzySet("b", 0.4, 0.5, 0.6))
+ FuzzySet(name='a ∩ b', left_boundary=0.4, peak=0.3, right_boundary=0.35)
+ """
+ return FuzzySet(
+ f"{self.name} ∩ {other.name}",
+ max(self.left_boundary, other.left_boundary),
+ min(self.right_boundary, other.right_boundary),
+ (self.peak + other.peak) / 2,
+ )
+
+ def membership(self, x: float) -> float:
+ """
+ Calculate the membership value of an input 'x' in the fuzzy set.
+ Returns:
+ The membership value of 'x' in the fuzzy set.
+
+ >>> a = FuzzySet("a", 0.1, 0.2, 0.3)
+ >>> a.membership(0.09)
+ 0.0
+ >>> a.membership(0.1)
+ 0.0
+ >>> a.membership(0.11)
+ 0.09999999999999995
+ >>> a.membership(0.4)
+ 0.0
+ >>> FuzzySet("A", 0, 0.5, 1).membership(0.1)
+ 0.2
+ >>> FuzzySet("B", 0.2, 0.7, 1).membership(0.6)
+ 0.8
+ """
+ if x <= self.left_boundary or x >= self.right_boundary:
+ return 0.0
+ elif self.left_boundary < x <= self.peak:
+ return (x - self.left_boundary) / (self.peak - self.left_boundary)
+ elif self.peak < x < self.right_boundary:
+ return (self.right_boundary - x) / (self.right_boundary - self.peak)
+ msg = f"Invalid value {x} for fuzzy set {self}"
+ raise ValueError(msg)
+
+ def union(self, other) -> FuzzySet:
+ """
+ Calculate the union of this fuzzy set with another fuzzy set.
+ Args:
+ other (FuzzySet): Another fuzzy set to union with.
+ Returns:
+ FuzzySet: A new fuzzy set representing the union.
+
+ >>> FuzzySet("a", 0.1, 0.2, 0.3).union(FuzzySet("b", 0.4, 0.5, 0.6))
+ FuzzySet(name='a U b', left_boundary=0.1, peak=0.6, right_boundary=0.35)
+ """
+ return FuzzySet(
+ f"{self.name} U {other.name}",
+ min(self.left_boundary, other.left_boundary),
+ max(self.right_boundary, other.right_boundary),
+ (self.peak + other.peak) / 2,
+ )
+
+ def plot(self):
+ """
+ Plot the membership function of the fuzzy set.
+ """
+ x = np.linspace(0, 1, 1000)
+ y = [self.membership(xi) for xi in x]
+
+ plt.plot(x, y, label=self.name)
+
if __name__ == "__main__":
- # Create universe of discourse in Python using linspace ()
- X = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False)
-
- # Create two fuzzy sets by defining any membership function
- # (trapmf(), gbellmf(), gaussmf(), etc).
- abc1 = [0, 25, 50]
- abc2 = [25, 50, 75]
- young = fuzz.membership.trimf(X, abc1)
- middle_aged = fuzz.membership.trimf(X, abc2)
-
- # Compute the different operations using inbuilt functions.
- one = np.ones(75)
- zero = np.zeros((75,))
- # 1. Union = max(µA(x), µB(x))
- union = fuzz.fuzzy_or(X, young, X, middle_aged)[1]
- # 2. Intersection = min(µA(x), µB(x))
- intersection = fuzz.fuzzy_and(X, young, X, middle_aged)[1]
- # 3. Complement (A) = (1- min(µA(x))
- complement_a = fuzz.fuzzy_not(young)
- # 4. Difference (A/B) = min(µA(x),(1- µB(x)))
- difference = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1]
- # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))]
- alg_sum = young + middle_aged - (young * middle_aged)
- # 6. Algebraic Product = (µA(x) * µB(x))
- alg_product = young * middle_aged
- # 7. Bounded Sum = min[1,(µA(x), µB(x))]
- bdd_sum = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1]
- # 8. Bounded difference = min[0,(µA(x), µB(x))]
- bdd_difference = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1]
-
- # max-min composition
- # max-product composition
-
- # Plot each set A, set B and each operation result using plot() and subplot().
- from matplotlib import pyplot as plt
-
- plt.figure()
-
- plt.subplot(4, 3, 1)
- plt.plot(X, young)
- plt.title("Young")
- plt.grid(True)
-
- plt.subplot(4, 3, 2)
- plt.plot(X, middle_aged)
- plt.title("Middle aged")
- plt.grid(True)
-
- plt.subplot(4, 3, 3)
- plt.plot(X, union)
- plt.title("union")
- plt.grid(True)
-
- plt.subplot(4, 3, 4)
- plt.plot(X, intersection)
- plt.title("intersection")
- plt.grid(True)
-
- plt.subplot(4, 3, 5)
- plt.plot(X, complement_a)
- plt.title("complement_a")
- plt.grid(True)
-
- plt.subplot(4, 3, 6)
- plt.plot(X, difference)
- plt.title("difference a/b")
- plt.grid(True)
-
- plt.subplot(4, 3, 7)
- plt.plot(X, alg_sum)
- plt.title("alg_sum")
- plt.grid(True)
-
- plt.subplot(4, 3, 8)
- plt.plot(X, alg_product)
- plt.title("alg_product")
- plt.grid(True)
-
- plt.subplot(4, 3, 9)
- plt.plot(X, bdd_sum)
- plt.title("bdd_sum")
- plt.grid(True)
-
- plt.subplot(4, 3, 10)
- plt.plot(X, bdd_difference)
- plt.title("bdd_difference")
- plt.grid(True)
-
- plt.subplots_adjust(hspace=0.5)
+ from doctest import testmod
+
+ testmod()
+ a = FuzzySet("A", 0, 0.5, 1)
+ b = FuzzySet("B", 0.2, 0.7, 1)
+
+ a.plot()
+ b.plot()
+
+ plt.xlabel("x")
+ plt.ylabel("Membership")
+ plt.legend()
+ plt.show()
+
+ union_ab = a.union(b)
+ intersection_ab = a.intersection(b)
+ complement_a = a.complement()
+
+ union_ab.plot()
+ intersection_ab.plot()
+ complement_a.plot()
+
+ plt.xlabel("x")
+ plt.ylabel("Membership")
+ plt.legend()
plt.show()
diff --git a/fuzzy_logic/fuzzy_operations.py.DISABLED.txt b/fuzzy_logic/fuzzy_operations.py.DISABLED.txt
new file mode 100644
index 000000000000..67fd587f4baf
--- /dev/null
+++ b/fuzzy_logic/fuzzy_operations.py.DISABLED.txt
@@ -0,0 +1,103 @@
+"""
+README, Author - Jigyasa Gandhi(mailto:jigsgandhi97@gmail.com)
+Requirements:
+ - scikit-fuzzy
+ - numpy
+ - matplotlib
+Python:
+ - 3.5
+"""
+import numpy as np
+import skfuzzy as fuzz
+
+if __name__ == "__main__":
+ # Create universe of discourse in Python using linspace ()
+ X = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False)
+
+ # Create two fuzzy sets by defining any membership function
+ # (trapmf(), gbellmf(), gaussmf(), etc).
+ abc1 = [0, 25, 50]
+ abc2 = [25, 50, 75]
+ young = fuzz.membership.trimf(X, abc1)
+ middle_aged = fuzz.membership.trimf(X, abc2)
+
+ # Compute the different operations using inbuilt functions.
+ one = np.ones(75)
+ zero = np.zeros((75,))
+ # 1. Union = max(µA(x), µB(x))
+ union = fuzz.fuzzy_or(X, young, X, middle_aged)[1]
+ # 2. Intersection = min(µA(x), µB(x))
+ intersection = fuzz.fuzzy_and(X, young, X, middle_aged)[1]
+ # 3. Complement (A) = (1 - min(µA(x)))
+ complement_a = fuzz.fuzzy_not(young)
+ # 4. Difference (A/B) = min(µA(x),(1- µB(x)))
+ difference = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1]
+ # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))]
+ alg_sum = young + middle_aged - (young * middle_aged)
+ # 6. Algebraic Product = (µA(x) * µB(x))
+ alg_product = young * middle_aged
+ # 7. Bounded Sum = min[1,(µA(x), µB(x))]
+ bdd_sum = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1]
+ # 8. Bounded difference = min[0,(µA(x), µB(x))]
+ bdd_difference = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1]
+
+ # max-min composition
+ # max-product composition
+
+ # Plot each set A, set B and each operation result using plot() and subplot().
+ from matplotlib import pyplot as plt
+
+ plt.figure()
+
+ plt.subplot(4, 3, 1)
+ plt.plot(X, young)
+ plt.title("Young")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 2)
+ plt.plot(X, middle_aged)
+ plt.title("Middle aged")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 3)
+ plt.plot(X, union)
+ plt.title("union")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 4)
+ plt.plot(X, intersection)
+ plt.title("intersection")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 5)
+ plt.plot(X, complement_a)
+ plt.title("complement_a")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 6)
+ plt.plot(X, difference)
+ plt.title("difference a/b")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 7)
+ plt.plot(X, alg_sum)
+ plt.title("alg_sum")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 8)
+ plt.plot(X, alg_product)
+ plt.title("alg_product")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 9)
+ plt.plot(X, bdd_sum)
+ plt.title("bdd_sum")
+ plt.grid(True)
+
+ plt.subplot(4, 3, 10)
+ plt.plot(X, bdd_difference)
+ plt.title("bdd_difference")
+ plt.grid(True)
+
+ plt.subplots_adjust(hspace=0.5)
+ plt.show()
diff --git a/genetic_algorithm/basic_string.py b/genetic_algorithm/basic_string.py
index 97dbe182bc82..b75491d9a949 100644
--- a/genetic_algorithm/basic_string.py
+++ b/genetic_algorithm/basic_string.py
@@ -9,18 +9,91 @@
import random
-# Maximum size of the population. bigger could be faster but is more memory expensive
+# Maximum size of the population. Bigger could be faster but is more memory expensive.
N_POPULATION = 200
-# Number of elements selected in every generation for evolution the selection takes
-# place from the best to the worst of that generation must be smaller than N_POPULATION
+# Number of elements selected in every generation of evolution. The selection takes
+# place from best to worst of that generation and must be smaller than N_POPULATION.
N_SELECTED = 50
-# Probability that an element of a generation can mutate changing one of its genes this
-# guarantees that all genes will be used during evolution
+# Probability that an element of a generation can mutate, changing one of its genes.
+# This will guarantee that all genes will be used during evolution.
MUTATION_PROBABILITY = 0.4
-# just a seed to improve randomness required by the algorithm
+# Just a seed to improve randomness required by the algorithm.
random.seed(random.randint(0, 1000))
+def evaluate(item: str, main_target: str) -> tuple[str, float]:
+ """
+ Evaluate how similar the item is with the target by just
+ counting each char in the right position
+ >>> evaluate("Helxo Worlx", "Hello World")
+ ('Helxo Worlx', 9.0)
+ """
+ score = len([g for position, g in enumerate(item) if g == main_target[position]])
+ return (item, float(score))
+
+
+def crossover(parent_1: str, parent_2: str) -> tuple[str, str]:
+ """
+ Slice and combine two strings at a random point.
+ >>> random.seed(42)
+ >>> crossover("123456", "abcdef")
+ ('12345f', 'abcde6')
+ """
+ random_slice = random.randint(0, len(parent_1) - 1)
+ child_1 = parent_1[:random_slice] + parent_2[random_slice:]
+ child_2 = parent_2[:random_slice] + parent_1[random_slice:]
+ return (child_1, child_2)
+
+
+def mutate(child: str, genes: list[str]) -> str:
+ """
+ Mutate a random gene of a child with another one from the list.
+ >>> random.seed(123)
+ >>> mutate("123456", list("ABCDEF"))
+ '12345A'
+ """
+ child_list = list(child)
+ if random.uniform(0, 1) < MUTATION_PROBABILITY:
+ child_list[random.randint(0, len(child)) - 1] = random.choice(genes)
+ return "".join(child_list)
+
+
+# Select, crossover and mutate a new population.
+def select(
+ parent_1: tuple[str, float],
+ population_score: list[tuple[str, float]],
+ genes: list[str],
+) -> list[str]:
+ """
+ Select the second parent and generate new population
+
+ >>> random.seed(42)
+ >>> parent_1 = ("123456", 8.0)
+ >>> population_score = [("abcdef", 4.0), ("ghijkl", 5.0), ("mnopqr", 7.0)]
+ >>> genes = list("ABCDEF")
+ >>> child_n = int(min(parent_1[1] + 1, 10))
+ >>> population = []
+ >>> for _ in range(child_n):
+ ... parent_2 = population_score[random.randrange(len(population_score))][0]
+ ... child_1, child_2 = crossover(parent_1[0], parent_2)
+ ... population.extend((mutate(child_1, genes), mutate(child_2, genes)))
+ >>> len(population) == (int(parent_1[1]) + 1) * 2
+ True
+ """
+ pop = []
+ # Generate more children proportionally to the fitness score.
+ child_n = int(parent_1[1] * 100) + 1
+ child_n = 10 if child_n >= 10 else child_n
+ for _ in range(child_n):
+ parent_2 = population_score[random.randint(0, N_SELECTED)][0]
+
+ child_1, child_2 = crossover(parent_1[0], parent_2)
+ # Append new string to the population list.
+ pop.append(mutate(child_1, genes))
+ pop.append(mutate(child_2, genes))
+ return pop
+
+
def basic(target: str, genes: list[str], debug: bool = True) -> tuple[int, int, str]:
"""
Verify that the target contains no genes besides the ones inside genes variable.
@@ -32,77 +105,66 @@ def basic(target: str, genes: list[str], debug: bool = True) -> tuple[int, int,
>>> genes.remove("e")
>>> basic("test", genes)
Traceback (most recent call last):
- ...
+ ...
ValueError: ['e'] is not in genes list, evolution cannot converge
>>> genes.remove("s")
>>> basic("test", genes)
Traceback (most recent call last):
- ...
+ ...
ValueError: ['e', 's'] is not in genes list, evolution cannot converge
>>> genes.remove("t")
>>> basic("test", genes)
Traceback (most recent call last):
- ...
+ ...
ValueError: ['e', 's', 't'] is not in genes list, evolution cannot converge
"""
# Verify if N_POPULATION is bigger than N_SELECTED
if N_POPULATION < N_SELECTED:
- raise ValueError(f"{N_POPULATION} must be bigger than {N_SELECTED}")
+ msg = f"{N_POPULATION} must be bigger than {N_SELECTED}"
+ raise ValueError(msg)
# Verify that the target contains no genes besides the ones inside genes variable.
not_in_genes_list = sorted({c for c in target if c not in genes})
if not_in_genes_list:
- raise ValueError(
- f"{not_in_genes_list} is not in genes list, evolution cannot converge"
- )
+ msg = f"{not_in_genes_list} is not in genes list, evolution cannot converge"
+ raise ValueError(msg)
- # Generate random starting population
+ # Generate random starting population.
population = []
for _ in range(N_POPULATION):
population.append("".join([random.choice(genes) for i in range(len(target))]))
- # Just some logs to know what the algorithms is doing
+ # Just some logs to know what the algorithms is doing.
generation, total_population = 0, 0
- # This loop will end when we will find a perfect match for our target
+ # This loop will end when we find a perfect match for our target.
while True:
generation += 1
total_population += len(population)
- # Random population created now it's time to evaluate
- def evaluate(item: str, main_target: str = target) -> tuple[str, float]:
- """
- Evaluate how similar the item is with the target by just
- counting each char in the right position
- >>> evaluate("Helxo Worlx", Hello World)
- ["Helxo Worlx", 9]
- """
- score = len(
- [g for position, g in enumerate(item) if g == main_target[position]]
- )
- return (item, float(score))
+ # Random population created. Now it's time to evaluate.
- # Adding a bit of concurrency can make everything faster,
+ # (Option 1) Adding a bit of concurrency can make everything faster,
#
# import concurrent.futures
# population_score: list[tuple[str, float]] = []
# with concurrent.futures.ThreadPoolExecutor(
# max_workers=NUM_WORKERS) as executor:
- # futures = {executor.submit(evaluate, item) for item in population}
+ # futures = {executor.submit(evaluate, item, target) for item in population}
# concurrent.futures.wait(futures)
# population_score = [item.result() for item in futures]
#
- # but with a simple algorithm like this will probably be slower
- # we just need to call evaluate for every item inside population
- population_score = [evaluate(item) for item in population]
+ # but with a simple algorithm like this, it will probably be slower.
+ # (Option 2) We just need to call evaluate for every item inside the population.
+ population_score = [evaluate(item, target) for item in population]
- # Check if there is a matching evolution
+ # Check if there is a matching evolution.
population_score = sorted(population_score, key=lambda x: x[1], reverse=True)
if population_score[0][0] == target:
return (generation, total_population, population_score[0][0])
- # Print the Best result every 10 generation
- # just to know that the algorithm is working
+ # Print the best result every 10 generation.
+ # Just to know that the algorithm is working.
if debug and generation % 10 == 0:
print(
f"\nGeneration: {generation}"
@@ -111,52 +173,23 @@ def evaluate(item: str, main_target: str = target) -> tuple[str, float]:
f"\nBest string: {population_score[0][0]}"
)
- # Flush the old population keeping some of the best evolutions
- # Keeping this avoid regression of evolution
+ # Flush the old population, keeping some of the best evolutions.
+ # Keeping this avoid regression of evolution.
population_best = population[: int(N_POPULATION / 3)]
population.clear()
population.extend(population_best)
- # Normalize population score from 0 to 1
+ # Normalize population score to be between 0 and 1.
population_score = [
(item, score / len(target)) for item, score in population_score
]
- # Select, Crossover and Mutate a new population
- def select(parent_1: tuple[str, float]) -> list[str]:
- """Select the second parent and generate new population"""
- pop = []
- # Generate more child proportionally to the fitness score
- child_n = int(parent_1[1] * 100) + 1
- child_n = 10 if child_n >= 10 else child_n
- for _ in range(child_n):
- parent_2 = population_score[random.randint(0, N_SELECTED)][0]
- child_1, child_2 = crossover(parent_1[0], parent_2)
- # Append new string to the population list
- pop.append(mutate(child_1))
- pop.append(mutate(child_2))
- return pop
-
- def crossover(parent_1: str, parent_2: str) -> tuple[str, str]:
- """Slice and combine two string in a random point"""
- random_slice = random.randint(0, len(parent_1) - 1)
- child_1 = parent_1[:random_slice] + parent_2[random_slice:]
- child_2 = parent_2[:random_slice] + parent_1[random_slice:]
- return (child_1, child_2)
-
- def mutate(child: str) -> str:
- """Mutate a random gene of a child with another one from the list"""
- child_list = list(child)
- if random.uniform(0, 1) < MUTATION_PROBABILITY:
- child_list[random.randint(0, len(child)) - 1] = random.choice(genes)
- return "".join(child_list)
-
- # This is Selection
+ # This is selection
for i in range(N_SELECTED):
- population.extend(select(population_score[int(i)]))
+ population.extend(select(population_score[int(i)], population_score, genes))
# Check if the population has already reached the maximum value and if so,
- # break the cycle. if this check is disabled the algorithm will take
- # forever to compute large strings but will also calculate small string in
- # a lot fewer generations
+ # break the cycle. If this check is disabled, the algorithm will take
+ # forever to compute large strings, but will also calculate small strings in
+ # a far fewer generations.
if len(population) > N_POPULATION:
break
@@ -169,7 +202,7 @@ def mutate(child: str) -> str:
" ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm"
"nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\"
)
+ generation, population, target = basic(target_str, genes_list)
print(
- "\nGeneration: %s\nTotal Population: %s\nTarget: %s"
- % basic(target_str, genes_list)
+ f"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}"
)
diff --git a/geodesy/haversine_distance.py b/geodesy/haversine_distance.py
index de8ac7f88302..39cd250af965 100644
--- a/geodesy/haversine_distance.py
+++ b/geodesy/haversine_distance.py
@@ -1,5 +1,9 @@
from math import asin, atan, cos, radians, sin, sqrt, tan
+AXIS_A = 6378137.0
+AXIS_B = 6356752.314245
+RADIUS = 6378137
+
def haversine_distance(lat1: float, lon1: float, lat2: float, lon2: float) -> float:
"""
@@ -17,10 +21,11 @@ def haversine_distance(lat1: float, lon1: float, lat2: float, lon2: float) -> fl
computation like Haversine can be handy for shorter range distances.
Args:
- lat1, lon1: latitude and longitude of coordinate 1
- lat2, lon2: latitude and longitude of coordinate 2
+ * `lat1`, `lon1`: latitude and longitude of coordinate 1
+ * `lat2`, `lon2`: latitude and longitude of coordinate 2
Returns:
geographical distance between two points in metres
+
>>> from collections import namedtuple
>>> point_2d = namedtuple("point_2d", "lat lon")
>>> SAN_FRANCISCO = point_2d(37.774856, -122.424227)
@@ -30,9 +35,6 @@ def haversine_distance(lat1: float, lon1: float, lat2: float, lon2: float) -> fl
"""
# CONSTANTS per WGS84 https://en.wikipedia.org/wiki/World_Geodetic_System
# Distance in metres(m)
- AXIS_A = 6378137.0
- AXIS_B = 6356752.314245
- RADIUS = 6378137
# Equation parameters
# Equation https://en.wikipedia.org/wiki/Haversine_formula#Formulation
flattening = (AXIS_A - AXIS_B) / AXIS_A
diff --git a/geodesy/lamberts_ellipsoidal_distance.py b/geodesy/lamberts_ellipsoidal_distance.py
index bf8f1b9a5080..4805674e51ab 100644
--- a/geodesy/lamberts_ellipsoidal_distance.py
+++ b/geodesy/lamberts_ellipsoidal_distance.py
@@ -2,11 +2,14 @@
from .haversine_distance import haversine_distance
+AXIS_A = 6378137.0
+AXIS_B = 6356752.314245
+EQUATORIAL_RADIUS = 6378137
+
def lamberts_ellipsoidal_distance(
lat1: float, lon1: float, lat2: float, lon2: float
) -> float:
-
"""
Calculate the shortest distance along the surface of an ellipsoid between
two points on the surface of earth given longitudes and latitudes
@@ -45,10 +48,6 @@ def lamberts_ellipsoidal_distance(
# CONSTANTS per WGS84 https://en.wikipedia.org/wiki/World_Geodetic_System
# Distance in metres(m)
- AXIS_A = 6378137.0
- AXIS_B = 6356752.314245
- EQUATORIAL_RADIUS = 6378137
-
# Equation Parameters
# https://en.wikipedia.org/wiki/Geographical_distance#Lambert's_formula_for_long_lines
flattening = (AXIS_A - AXIS_B) / AXIS_A
@@ -62,22 +61,22 @@ def lamberts_ellipsoidal_distance(
sigma = haversine_distance(lat1, lon1, lat2, lon2) / EQUATORIAL_RADIUS
# Intermediate P and Q values
- P_value = (b_lat1 + b_lat2) / 2
- Q_value = (b_lat2 - b_lat1) / 2
+ p_value = (b_lat1 + b_lat2) / 2
+ q_value = (b_lat2 - b_lat1) / 2
# Intermediate X value
# X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2)
- X_numerator = (sin(P_value) ** 2) * (cos(Q_value) ** 2)
- X_demonimator = cos(sigma / 2) ** 2
- X_value = (sigma - sin(sigma)) * (X_numerator / X_demonimator)
+ x_numerator = (sin(p_value) ** 2) * (cos(q_value) ** 2)
+ x_demonimator = cos(sigma / 2) ** 2
+ x_value = (sigma - sin(sigma)) * (x_numerator / x_demonimator)
# Intermediate Y value
# Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2)
- Y_numerator = (cos(P_value) ** 2) * (sin(Q_value) ** 2)
- Y_denominator = sin(sigma / 2) ** 2
- Y_value = (sigma + sin(sigma)) * (Y_numerator / Y_denominator)
+ y_numerator = (cos(p_value) ** 2) * (sin(q_value) ** 2)
+ y_denominator = sin(sigma / 2) ** 2
+ y_value = (sigma + sin(sigma)) * (y_numerator / y_denominator)
- return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (X_value + Y_value)))
+ return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value)))
if __name__ == "__main__":
diff --git a/geometry/__init__.py b/geometry/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/geometry/geometry.py b/geometry/geometry.py
new file mode 100644
index 000000000000..a0be8eb3befc
--- /dev/null
+++ b/geometry/geometry.py
@@ -0,0 +1,288 @@
+from __future__ import annotations
+
+import math
+from dataclasses import dataclass, field
+from types import NoneType
+from typing import Self
+
+# Building block classes
+
+
+@dataclass
+class Angle:
+ """
+ An Angle in degrees (unit of measurement)
+
+ >>> Angle()
+ Angle(degrees=90)
+ >>> Angle(45.5)
+ Angle(degrees=45.5)
+ >>> Angle(-1)
+ Traceback (most recent call last):
+ ...
+ TypeError: degrees must be a numeric value between 0 and 360.
+ >>> Angle(361)
+ Traceback (most recent call last):
+ ...
+ TypeError: degrees must be a numeric value between 0 and 360.
+ """
+
+ degrees: float = 90
+
+ def __post_init__(self) -> None:
+ if not isinstance(self.degrees, (int, float)) or not 0 <= self.degrees <= 360:
+ raise TypeError("degrees must be a numeric value between 0 and 360.")
+
+
+@dataclass
+class Side:
+ """
+ A side of a two dimensional Shape such as Polygon, etc.
+ adjacent_sides: a list of sides which are adjacent to the current side
+ angle: the angle in degrees between each adjacent side
+ length: the length of the current side in meters
+
+ >>> Side(5)
+ Side(length=5, angle=Angle(degrees=90), next_side=None)
+ >>> Side(5, Angle(45.6))
+ Side(length=5, angle=Angle(degrees=45.6), next_side=None)
+ >>> Side(5, Angle(45.6), Side(1, Angle(2))) # doctest: +ELLIPSIS
+ Side(length=5, angle=Angle(degrees=45.6), next_side=Side(length=1, angle=Angle(d...
+ >>> Side(-1)
+ Traceback (most recent call last):
+ ...
+ TypeError: length must be a positive numeric value.
+ >>> Side(5, None)
+ Traceback (most recent call last):
+ ...
+ TypeError: angle must be an Angle object.
+ >>> Side(5, Angle(90), "Invalid next_side")
+ Traceback (most recent call last):
+ ...
+ TypeError: next_side must be a Side or None.
+ """
+
+ length: float
+ angle: Angle = field(default_factory=Angle)
+ next_side: Side | None = None
+
+ def __post_init__(self) -> None:
+ if not isinstance(self.length, (int, float)) or self.length <= 0:
+ raise TypeError("length must be a positive numeric value.")
+ if not isinstance(self.angle, Angle):
+ raise TypeError("angle must be an Angle object.")
+ if not isinstance(self.next_side, (Side, NoneType)):
+ raise TypeError("next_side must be a Side or None.")
+
+
+@dataclass
+class Ellipse:
+ """
+ A geometric Ellipse on a 2D surface
+
+ >>> Ellipse(5, 10)
+ Ellipse(major_radius=5, minor_radius=10)
+ >>> Ellipse(5, 10) is Ellipse(5, 10)
+ False
+ >>> Ellipse(5, 10) == Ellipse(5, 10)
+ True
+ """
+
+ major_radius: float
+ minor_radius: float
+
+ @property
+ def area(self) -> float:
+ """
+ >>> Ellipse(5, 10).area
+ 157.07963267948966
+ """
+ return math.pi * self.major_radius * self.minor_radius
+
+ @property
+ def perimeter(self) -> float:
+ """
+ >>> Ellipse(5, 10).perimeter
+ 47.12388980384689
+ """
+ return math.pi * (self.major_radius + self.minor_radius)
+
+
+class Circle(Ellipse):
+ """
+ A geometric Circle on a 2D surface
+
+ >>> Circle(5)
+ Circle(radius=5)
+ >>> Circle(5) is Circle(5)
+ False
+ >>> Circle(5) == Circle(5)
+ True
+ >>> Circle(5).area
+ 78.53981633974483
+ >>> Circle(5).perimeter
+ 31.41592653589793
+ """
+
+ def __init__(self, radius: float) -> None:
+ super().__init__(radius, radius)
+ self.radius = radius
+
+ def __repr__(self) -> str:
+ return f"Circle(radius={self.radius})"
+
+ @property
+ def diameter(self) -> float:
+ """
+ >>> Circle(5).diameter
+ 10
+ """
+ return self.radius * 2
+
+ def max_parts(self, num_cuts: float) -> float:
+ """
+ Return the maximum number of parts that circle can be divided into if cut
+ 'num_cuts' times.
+
+ >>> circle = Circle(5)
+ >>> circle.max_parts(0)
+ 1.0
+ >>> circle.max_parts(7)
+ 29.0
+ >>> circle.max_parts(54)
+ 1486.0
+ >>> circle.max_parts(22.5)
+ 265.375
+ >>> circle.max_parts(-222)
+ Traceback (most recent call last):
+ ...
+ TypeError: num_cuts must be a positive numeric value.
+ >>> circle.max_parts("-222")
+ Traceback (most recent call last):
+ ...
+ TypeError: num_cuts must be a positive numeric value.
+ """
+ if not isinstance(num_cuts, (int, float)) or num_cuts < 0:
+ raise TypeError("num_cuts must be a positive numeric value.")
+ return (num_cuts + 2 + num_cuts**2) * 0.5
+
+
+@dataclass
+class Polygon:
+ """
+ An abstract class which represents Polygon on a 2D surface.
+
+ >>> Polygon()
+ Polygon(sides=[])
+ >>> polygon = Polygon()
+ >>> polygon.add_side(Side(5)).get_side(0)
+ Side(length=5, angle=Angle(degrees=90), next_side=None)
+ >>> polygon.get_side(1)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> polygon.set_side(0, Side(10)).get_side(0)
+ Side(length=10, angle=Angle(degrees=90), next_side=None)
+ >>> polygon.set_side(1, Side(10))
+ Traceback (most recent call last):
+ ...
+ IndexError: list assignment index out of range
+ """
+
+ sides: list[Side] = field(default_factory=list)
+
+ def add_side(self, side: Side) -> Self:
+ """
+ >>> Polygon().add_side(Side(5))
+ Polygon(sides=[Side(length=5, angle=Angle(degrees=90), next_side=None)])
+ """
+ self.sides.append(side)
+ return self
+
+ def get_side(self, index: int) -> Side:
+ """
+ >>> Polygon().get_side(0)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> Polygon().add_side(Side(5)).get_side(-1)
+ Side(length=5, angle=Angle(degrees=90), next_side=None)
+ """
+ return self.sides[index]
+
+ def set_side(self, index: int, side: Side) -> Self:
+ """
+ >>> Polygon().set_side(0, Side(5))
+ Traceback (most recent call last):
+ ...
+ IndexError: list assignment index out of range
+ >>> Polygon().add_side(Side(5)).set_side(0, Side(10))
+ Polygon(sides=[Side(length=10, angle=Angle(degrees=90), next_side=None)])
+ """
+ self.sides[index] = side
+ return self
+
+
+class Rectangle(Polygon):
+ """
+ A geometric rectangle on a 2D surface.
+
+ >>> rectangle_one = Rectangle(5, 10)
+ >>> rectangle_one.perimeter()
+ 30
+ >>> rectangle_one.area()
+ 50
+ >>> Rectangle(-5, 10)
+ Traceback (most recent call last):
+ ...
+ TypeError: length must be a positive numeric value.
+ """
+
+ def __init__(self, short_side_length: float, long_side_length: float) -> None:
+ super().__init__()
+ self.short_side_length = short_side_length
+ self.long_side_length = long_side_length
+ self.post_init()
+
+ def post_init(self) -> None:
+ """
+ >>> Rectangle(5, 10) # doctest: +NORMALIZE_WHITESPACE
+ Rectangle(sides=[Side(length=5, angle=Angle(degrees=90), next_side=None),
+ Side(length=10, angle=Angle(degrees=90), next_side=None)])
+ """
+ self.short_side = Side(self.short_side_length)
+ self.long_side = Side(self.long_side_length)
+ super().add_side(self.short_side)
+ super().add_side(self.long_side)
+
+ def perimeter(self) -> float:
+ return (self.short_side.length + self.long_side.length) * 2
+
+ def area(self) -> float:
+ return self.short_side.length * self.long_side.length
+
+
+@dataclass
+class Square(Rectangle):
+ """
+ a structure which represents a
+ geometrical square on a 2D surface
+ >>> square_one = Square(5)
+ >>> square_one.perimeter()
+ 20
+ >>> square_one.area()
+ 25
+ """
+
+ def __init__(self, side_length: float) -> None:
+ super().__init__(side_length, side_length)
+
+ def perimeter(self) -> float:
+ return super().perimeter()
+
+ def area(self) -> float:
+ return super().area()
+
+
+if __name__ == "__main__":
+ __import__("doctest").testmod()
diff --git a/graphics/bezier_curve.py b/graphics/bezier_curve.py
index 295ff47e8cdc..6c7dcd4f06e7 100644
--- a/graphics/bezier_curve.py
+++ b/graphics/bezier_curve.py
@@ -30,9 +30,9 @@ def basis_function(self, t: float) -> list[float]:
returns the x, y values of basis function at time t
>>> curve = BezierCurve([(1,1), (1,2)])
- >>> curve.basis_function(0)
+ >>> [float(x) for x in curve.basis_function(0)]
[1.0, 0.0]
- >>> curve.basis_function(1)
+ >>> [float(x) for x in curve.basis_function(1)]
[0.0, 1.0]
"""
assert 0 <= t <= 1, "Time t must be between 0 and 1."
@@ -40,7 +40,7 @@ def basis_function(self, t: float) -> list[float]:
for i in range(len(self.list_of_points)):
# basis function for each i
output_values.append(
- comb(self.degree, i) * ((1 - t) ** (self.degree - i)) * (t ** i)
+ comb(self.degree, i) * ((1 - t) ** (self.degree - i)) * (t**i)
)
# the basis must sum up to 1 for it to produce a valid Bezier curve.
assert round(sum(output_values), 5) == 1
@@ -55,9 +55,9 @@ def bezier_curve_function(self, t: float) -> tuple[float, float]:
The last point in the curve is when t = 1.
>>> curve = BezierCurve([(1,1), (1,2)])
- >>> curve.bezier_curve_function(0)
+ >>> tuple(float(x) for x in curve.bezier_curve_function(0))
(1.0, 1.0)
- >>> curve.bezier_curve_function(1)
+ >>> tuple(float(x) for x in curve.bezier_curve_function(1))
(1.0, 2.0)
"""
diff --git a/graphics/butterfly_pattern.py b/graphics/butterfly_pattern.py
new file mode 100644
index 000000000000..7913b03a7e95
--- /dev/null
+++ b/graphics/butterfly_pattern.py
@@ -0,0 +1,46 @@
+def butterfly_pattern(n: int) -> str:
+ """
+ Creates a butterfly pattern of size n and returns it as a string.
+
+ >>> print(butterfly_pattern(3))
+ * *
+ ** **
+ *****
+ ** **
+ * *
+ >>> print(butterfly_pattern(5))
+ * *
+ ** **
+ *** ***
+ **** ****
+ *********
+ **** ****
+ *** ***
+ ** **
+ * *
+ """
+ result = []
+
+ # Upper part
+ for i in range(1, n):
+ left_stars = "*" * i
+ spaces = " " * (2 * (n - i) - 1)
+ right_stars = "*" * i
+ result.append(left_stars + spaces + right_stars)
+
+ # Middle part
+ result.append("*" * (2 * n - 1))
+
+ # Lower part
+ for i in range(n - 1, 0, -1):
+ left_stars = "*" * i
+ spaces = " " * (2 * (n - i) - 1)
+ right_stars = "*" * i
+ result.append(left_stars + spaces + right_stars)
+
+ return "\n".join(result)
+
+
+if __name__ == "__main__":
+ n = int(input("Enter the size of the butterfly pattern: "))
+ print(butterfly_pattern(n))
diff --git a/graphics/digital_differential_analyzer_line.py b/graphics/digital_differential_analyzer_line.py
new file mode 100644
index 000000000000..f7269ab09856
--- /dev/null
+++ b/graphics/digital_differential_analyzer_line.py
@@ -0,0 +1,52 @@
+import matplotlib.pyplot as plt
+
+
+def digital_differential_analyzer_line(
+ p1: tuple[int, int], p2: tuple[int, int]
+) -> list[tuple[int, int]]:
+ """
+ Draws a line between two points using the DDA algorithm.
+
+ Args:
+ - p1: Coordinates of the starting point.
+ - p2: Coordinates of the ending point.
+ Returns:
+ - List of coordinate points that form the line.
+
+ >>> digital_differential_analyzer_line((1, 1), (4, 4))
+ [(2, 2), (3, 3), (4, 4)]
+ """
+ x1, y1 = p1
+ x2, y2 = p2
+ dx = x2 - x1
+ dy = y2 - y1
+ steps = max(abs(dx), abs(dy))
+ x_increment = dx / float(steps)
+ y_increment = dy / float(steps)
+ coordinates = []
+ x: float = x1
+ y: float = y1
+ for _ in range(steps):
+ x += x_increment
+ y += y_increment
+ coordinates.append((round(x), round(y)))
+ return coordinates
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ x1 = int(input("Enter the x-coordinate of the starting point: "))
+ y1 = int(input("Enter the y-coordinate of the starting point: "))
+ x2 = int(input("Enter the x-coordinate of the ending point: "))
+ y2 = int(input("Enter the y-coordinate of the ending point: "))
+ coordinates = digital_differential_analyzer_line((x1, y1), (x2, y2))
+ x_points, y_points = zip(*coordinates)
+ plt.plot(x_points, y_points, marker="o")
+ plt.title("Digital Differential Analyzer Line Drawing Algorithm")
+ plt.xlabel("X-axis")
+ plt.ylabel("Y-axis")
+ plt.grid()
+ plt.show()
diff --git a/graphics/vector3_for_2d_rendering.py b/graphics/vector3_for_2d_rendering.py
index dfa22262a8d8..a332206e67b6 100644
--- a/graphics/vector3_for_2d_rendering.py
+++ b/graphics/vector3_for_2d_rendering.py
@@ -28,9 +28,8 @@ def convert_to_2d(
TypeError: Input values must either be float or int: ['1', 2, 3, 10, 10]
"""
if not all(isinstance(val, (float, int)) for val in locals().values()):
- raise TypeError(
- "Input values must either be float or int: " f"{list(locals().values())}"
- )
+ msg = f"Input values must either be float or int: {list(locals().values())}"
+ raise TypeError(msg)
projected_x = ((x * distance) / (z + distance)) * scale
projected_y = ((y * distance) / (z + distance)) * scale
return projected_x, projected_y
@@ -71,10 +70,11 @@ def rotate(
input_variables = locals()
del input_variables["axis"]
if not all(isinstance(val, (float, int)) for val in input_variables.values()):
- raise TypeError(
+ msg = (
"Input values except axis must either be float or int: "
f"{list(input_variables.values())}"
)
+ raise TypeError(msg)
angle = (angle % 360) / 450 * 180 / math.pi
if axis == "z":
new_x = x * math.cos(angle) - y * math.sin(angle)
diff --git a/graphs/a_star.py b/graphs/a_star.py
index cb5b2fcd16e8..1d7063ccc55a 100644
--- a/graphs/a_star.py
+++ b/graphs/a_star.py
@@ -1,38 +1,46 @@
-grid = [
- [0, 1, 0, 0, 0, 0],
- [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles
- [0, 1, 0, 0, 0, 0],
- [0, 1, 0, 0, 1, 0],
- [0, 0, 0, 0, 1, 0],
-]
-
-"""
-heuristic = [[9, 8, 7, 6, 5, 4],
- [8, 7, 6, 5, 4, 3],
- [7, 6, 5, 4, 3, 2],
- [6, 5, 4, 3, 2, 1],
- [5, 4, 3, 2, 1, 0]]"""
-
-init = [0, 0]
-goal = [len(grid) - 1, len(grid[0]) - 1] # all coordinates are given in format [y,x]
-cost = 1
-
-# the cost map which pushes the path closer to the goal
-heuristic = [[0 for row in range(len(grid[0]))] for col in range(len(grid))]
-for i in range(len(grid)):
- for j in range(len(grid[0])):
- heuristic[i][j] = abs(i - goal[0]) + abs(j - goal[1])
- if grid[i][j] == 1:
- heuristic[i][j] = 99 # added extra penalty in the heuristic map
-
+from __future__ import annotations
-# the actions we can take
-delta = [[-1, 0], [0, -1], [1, 0], [0, 1]] # go up # go left # go down # go right
+DIRECTIONS = [
+ [-1, 0], # left
+ [0, -1], # down
+ [1, 0], # right
+ [0, 1], # up
+]
# function to search the path
-def search(grid, init, goal, cost, heuristic):
-
+def search(
+ grid: list[list[int]],
+ init: list[int],
+ goal: list[int],
+ cost: int,
+ heuristic: list[list[int]],
+) -> tuple[list[list[int]], list[list[int]]]:
+ """
+ Search for a path on a grid avoiding obstacles.
+ >>> grid = [[0, 1, 0, 0, 0, 0],
+ ... [0, 1, 0, 0, 0, 0],
+ ... [0, 1, 0, 0, 0, 0],
+ ... [0, 1, 0, 0, 1, 0],
+ ... [0, 0, 0, 0, 1, 0]]
+ >>> init = [0, 0]
+ >>> goal = [len(grid) - 1, len(grid[0]) - 1]
+ >>> cost = 1
+ >>> heuristic = [[0] * len(grid[0]) for _ in range(len(grid))]
+ >>> heuristic = [[0 for row in range(len(grid[0]))] for col in range(len(grid))]
+ >>> for i in range(len(grid)):
+ ... for j in range(len(grid[0])):
+ ... heuristic[i][j] = abs(i - goal[0]) + abs(j - goal[1])
+ ... if grid[i][j] == 1:
+ ... heuristic[i][j] = 99
+ >>> path, action = search(grid, init, goal, cost, heuristic)
+ >>> path # doctest: +NORMALIZE_WHITESPACE
+ [[0, 0], [1, 0], [2, 0], [3, 0], [4, 0], [4, 1], [4, 2], [4, 3], [3, 3],
+ [2, 3], [2, 4], [2, 5], [3, 5], [4, 5]]
+ >>> action # doctest: +NORMALIZE_WHITESPACE
+ [[0, 0, 0, 0, 0, 0], [2, 0, 0, 0, 0, 0], [2, 0, 0, 0, 3, 3],
+ [2, 0, 0, 0, 0, 2], [2, 3, 3, 3, 0, 2]]
+ """
closed = [
[0 for col in range(len(grid[0]))] for row in range(len(grid))
] # the reference grid
@@ -44,7 +52,7 @@ def search(grid, init, goal, cost, heuristic):
x = init[0]
y = init[1]
g = 0
- f = g + heuristic[init[0]][init[0]]
+ f = g + heuristic[x][y] # cost from starting cell to destination cell
cell = [[f, g, x, y]]
found = False # flag that is set when search is complete
@@ -52,35 +60,41 @@ def search(grid, init, goal, cost, heuristic):
while not found and not resign:
if len(cell) == 0:
- return "FAIL"
+ raise ValueError("Algorithm is unable to find solution")
else: # to choose the least costliest action so as to move closer to the goal
cell.sort()
cell.reverse()
- next = cell.pop()
- x = next[2]
- y = next[3]
- g = next[1]
+ next_cell = cell.pop()
+ x = next_cell[2]
+ y = next_cell[3]
+ g = next_cell[1]
if x == goal[0] and y == goal[1]:
found = True
else:
- for i in range(len(delta)): # to try out different valid actions
- x2 = x + delta[i][0]
- y2 = y + delta[i][1]
- if x2 >= 0 and x2 < len(grid) and y2 >= 0 and y2 < len(grid[0]):
- if closed[x2][y2] == 0 and grid[x2][y2] == 0:
- g2 = g + cost
- f2 = g2 + heuristic[x2][y2]
- cell.append([f2, g2, x2, y2])
- closed[x2][y2] = 1
- action[x2][y2] = i
+ for i in range(len(DIRECTIONS)): # to try out different valid actions
+ x2 = x + DIRECTIONS[i][0]
+ y2 = y + DIRECTIONS[i][1]
+ if (
+ x2 >= 0
+ and x2 < len(grid)
+ and y2 >= 0
+ and y2 < len(grid[0])
+ and closed[x2][y2] == 0
+ and grid[x2][y2] == 0
+ ):
+ g2 = g + cost
+ f2 = g2 + heuristic[x2][y2]
+ cell.append([f2, g2, x2, y2])
+ closed[x2][y2] = 1
+ action[x2][y2] = i
invpath = []
x = goal[0]
y = goal[1]
invpath.append([x, y]) # we get the reverse path from here
while x != init[0] or y != init[1]:
- x2 = x - delta[action[x][y]][0]
- y2 = y - delta[action[x][y]][1]
+ x2 = x - DIRECTIONS[action[x][y]][0]
+ y2 = y - DIRECTIONS[action[x][y]][1]
x = x2
y = y2
invpath.append([x, y])
@@ -88,13 +102,37 @@ def search(grid, init, goal, cost, heuristic):
path = []
for i in range(len(invpath)):
path.append(invpath[len(invpath) - 1 - i])
+ return path, action
+
+
+if __name__ == "__main__":
+ grid = [
+ [0, 1, 0, 0, 0, 0],
+ [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles
+ [0, 1, 0, 0, 0, 0],
+ [0, 1, 0, 0, 1, 0],
+ [0, 0, 0, 0, 1, 0],
+ ]
+
+ init = [0, 0]
+ # all coordinates are given in format [y,x]
+ goal = [len(grid) - 1, len(grid[0]) - 1]
+ cost = 1
+
+ # the cost map which pushes the path closer to the goal
+ heuristic = [[0 for row in range(len(grid[0]))] for col in range(len(grid))]
+ for i in range(len(grid)):
+ for j in range(len(grid[0])):
+ heuristic[i][j] = abs(i - goal[0]) + abs(j - goal[1])
+ if grid[i][j] == 1:
+ # added extra penalty in the heuristic map
+ heuristic[i][j] = 99
+
+ path, action = search(grid, init, goal, cost, heuristic)
+
print("ACTION MAP")
for i in range(len(action)):
print(action[i])
- return path
-
-
-a = search(grid, init, goal, cost, heuristic)
-for i in range(len(a)):
- print(a[i])
+ for i in range(len(path)):
+ print(path[i])
diff --git a/graphs/ant_colony_optimization_algorithms.py b/graphs/ant_colony_optimization_algorithms.py
new file mode 100644
index 000000000000..753f4c0962c8
--- /dev/null
+++ b/graphs/ant_colony_optimization_algorithms.py
@@ -0,0 +1,224 @@
+"""
+Use an ant colony optimization algorithm to solve the travelling salesman problem (TSP)
+which asks the following question:
+"Given a list of cities and the distances between each pair of cities, what is the
+ shortest possible route that visits each city exactly once and returns to the origin
+ city?"
+
+https://en.wikipedia.org/wiki/Ant_colony_optimization_algorithms
+https://en.wikipedia.org/wiki/Travelling_salesman_problem
+
+Author: Clark
+"""
+
+import copy
+import random
+
+cities = {
+ 0: [0, 0],
+ 1: [0, 5],
+ 2: [3, 8],
+ 3: [8, 10],
+ 4: [12, 8],
+ 5: [12, 4],
+ 6: [8, 0],
+ 7: [6, 2],
+}
+
+
+def main(
+ cities: dict[int, list[int]],
+ ants_num: int,
+ iterations_num: int,
+ pheromone_evaporation: float,
+ alpha: float,
+ beta: float,
+ q: float, # Pheromone system parameters Q, which is a constant
+) -> tuple[list[int], float]:
+ """
+ Ant colony algorithm main function
+ >>> main(cities=cities, ants_num=10, iterations_num=20,
+ ... pheromone_evaporation=0.7, alpha=1.0, beta=5.0, q=10)
+ ([0, 1, 2, 3, 4, 5, 6, 7, 0], 37.909778143828696)
+ >>> main(cities={0: [0, 0], 1: [2, 2]}, ants_num=5, iterations_num=5,
+ ... pheromone_evaporation=0.7, alpha=1.0, beta=5.0, q=10)
+ ([0, 1, 0], 5.656854249492381)
+ >>> main(cities={0: [0, 0], 1: [2, 2], 4: [4, 4]}, ants_num=5, iterations_num=5,
+ ... pheromone_evaporation=0.7, alpha=1.0, beta=5.0, q=10)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> main(cities={}, ants_num=5, iterations_num=5,
+ ... pheromone_evaporation=0.7, alpha=1.0, beta=5.0, q=10)
+ Traceback (most recent call last):
+ ...
+ StopIteration
+ >>> main(cities={0: [0, 0], 1: [2, 2]}, ants_num=0, iterations_num=5,
+ ... pheromone_evaporation=0.7, alpha=1.0, beta=5.0, q=10)
+ ([], inf)
+ >>> main(cities={0: [0, 0], 1: [2, 2]}, ants_num=5, iterations_num=0,
+ ... pheromone_evaporation=0.7, alpha=1.0, beta=5.0, q=10)
+ ([], inf)
+ >>> main(cities={0: [0, 0], 1: [2, 2]}, ants_num=5, iterations_num=5,
+ ... pheromone_evaporation=1, alpha=1.0, beta=5.0, q=10)
+ ([0, 1, 0], 5.656854249492381)
+ >>> main(cities={0: [0, 0], 1: [2, 2]}, ants_num=5, iterations_num=5,
+ ... pheromone_evaporation=0, alpha=1.0, beta=5.0, q=10)
+ ([0, 1, 0], 5.656854249492381)
+ """
+ # Initialize the pheromone matrix
+ cities_num = len(cities)
+ pheromone = [[1.0] * cities_num] * cities_num
+
+ best_path: list[int] = []
+ best_distance = float("inf")
+ for _ in range(iterations_num):
+ ants_route = []
+ for _ in range(ants_num):
+ unvisited_cities = copy.deepcopy(cities)
+ current_city = {next(iter(cities.keys())): next(iter(cities.values()))}
+ del unvisited_cities[next(iter(current_city.keys()))]
+ ant_route = [next(iter(current_city.keys()))]
+ while unvisited_cities:
+ current_city, unvisited_cities = city_select(
+ pheromone, current_city, unvisited_cities, alpha, beta
+ )
+ ant_route.append(next(iter(current_city.keys())))
+ ant_route.append(0)
+ ants_route.append(ant_route)
+
+ pheromone, best_path, best_distance = pheromone_update(
+ pheromone,
+ cities,
+ pheromone_evaporation,
+ ants_route,
+ q,
+ best_path,
+ best_distance,
+ )
+ return best_path, best_distance
+
+
+def distance(city1: list[int], city2: list[int]) -> float:
+ """
+ Calculate the distance between two coordinate points
+ >>> distance([0, 0], [3, 4] )
+ 5.0
+ >>> distance([0, 0], [-3, 4] )
+ 5.0
+ >>> distance([0, 0], [-3, -4] )
+ 5.0
+ """
+ return (((city1[0] - city2[0]) ** 2) + ((city1[1] - city2[1]) ** 2)) ** 0.5
+
+
+def pheromone_update(
+ pheromone: list[list[float]],
+ cities: dict[int, list[int]],
+ pheromone_evaporation: float,
+ ants_route: list[list[int]],
+ q: float, # Pheromone system parameters Q, which is a constant
+ best_path: list[int],
+ best_distance: float,
+) -> tuple[list[list[float]], list[int], float]:
+ """
+ Update pheromones on the route and update the best route
+ >>>
+ >>> pheromone_update(pheromone=[[1.0, 1.0], [1.0, 1.0]],
+ ... cities={0: [0,0], 1: [2,2]}, pheromone_evaporation=0.7,
+ ... ants_route=[[0, 1, 0]], q=10, best_path=[],
+ ... best_distance=float("inf"))
+ ([[0.7, 4.235533905932737], [4.235533905932737, 0.7]], [0, 1, 0], 5.656854249492381)
+ >>> pheromone_update(pheromone=[],
+ ... cities={0: [0,0], 1: [2,2]}, pheromone_evaporation=0.7,
+ ... ants_route=[[0, 1, 0]], q=10, best_path=[],
+ ... best_distance=float("inf"))
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> pheromone_update(pheromone=[[1.0, 1.0], [1.0, 1.0]],
+ ... cities={}, pheromone_evaporation=0.7,
+ ... ants_route=[[0, 1, 0]], q=10, best_path=[],
+ ... best_distance=float("inf"))
+ Traceback (most recent call last):
+ ...
+ KeyError: 0
+ """
+ for a in range(len(cities)): # Update the volatilization of pheromone on all routes
+ for b in range(len(cities)):
+ pheromone[a][b] *= pheromone_evaporation
+ for ant_route in ants_route:
+ total_distance = 0.0
+ for i in range(len(ant_route) - 1): # Calculate total distance
+ total_distance += distance(cities[ant_route[i]], cities[ant_route[i + 1]])
+ delta_pheromone = q / total_distance
+ for i in range(len(ant_route) - 1): # Update pheromones
+ pheromone[ant_route[i]][ant_route[i + 1]] += delta_pheromone
+ pheromone[ant_route[i + 1]][ant_route[i]] = pheromone[ant_route[i]][
+ ant_route[i + 1]
+ ]
+
+ if total_distance < best_distance:
+ best_path = ant_route
+ best_distance = total_distance
+
+ return pheromone, best_path, best_distance
+
+
+def city_select(
+ pheromone: list[list[float]],
+ current_city: dict[int, list[int]],
+ unvisited_cities: dict[int, list[int]],
+ alpha: float,
+ beta: float,
+) -> tuple[dict[int, list[int]], dict[int, list[int]]]:
+ """
+ Choose the next city for ants
+ >>> city_select(pheromone=[[1.0, 1.0], [1.0, 1.0]], current_city={0: [0, 0]},
+ ... unvisited_cities={1: [2, 2]}, alpha=1.0, beta=5.0)
+ ({1: [2, 2]}, {})
+ >>> city_select(pheromone=[], current_city={0: [0,0]},
+ ... unvisited_cities={1: [2, 2]}, alpha=1.0, beta=5.0)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> city_select(pheromone=[[1.0, 1.0], [1.0, 1.0]], current_city={},
+ ... unvisited_cities={1: [2, 2]}, alpha=1.0, beta=5.0)
+ Traceback (most recent call last):
+ ...
+ StopIteration
+ >>> city_select(pheromone=[[1.0, 1.0], [1.0, 1.0]], current_city={0: [0, 0]},
+ ... unvisited_cities={}, alpha=1.0, beta=5.0)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ """
+ probabilities = []
+ for city, value in unvisited_cities.items():
+ city_distance = distance(value, next(iter(current_city.values())))
+ probability = (pheromone[city][next(iter(current_city.keys()))] ** alpha) * (
+ (1 / city_distance) ** beta
+ )
+ probabilities.append(probability)
+
+ chosen_city_i = random.choices(
+ list(unvisited_cities.keys()), weights=probabilities
+ )[0]
+ chosen_city = {chosen_city_i: unvisited_cities[chosen_city_i]}
+ del unvisited_cities[next(iter(chosen_city.keys()))]
+ return chosen_city, unvisited_cities
+
+
+if __name__ == "__main__":
+ best_path, best_distance = main(
+ cities=cities,
+ ants_num=10,
+ iterations_num=20,
+ pheromone_evaporation=0.7,
+ alpha=1.0,
+ beta=5.0,
+ q=10,
+ )
+
+ print(f"{best_path = }")
+ print(f"{best_distance = }")
diff --git a/graphs/articulation_points.py b/graphs/articulation_points.py
index 7197369de090..0bf16e55bc04 100644
--- a/graphs/articulation_points.py
+++ b/graphs/articulation_points.py
@@ -1,47 +1,47 @@
# Finding Articulation Points in Undirected Graph
-def computeAP(l): # noqa: E741
- n = len(l)
- outEdgeCount = 0
+def compute_ap(graph):
+ n = len(graph)
+ out_edge_count = 0
low = [0] * n
visited = [False] * n
- isArt = [False] * n
+ is_art = [False] * n
- def dfs(root, at, parent, outEdgeCount):
+ def dfs(root, at, parent, out_edge_count):
if parent == root:
- outEdgeCount += 1
+ out_edge_count += 1
visited[at] = True
low[at] = at
- for to in l[at]:
+ for to in graph[at]:
if to == parent:
pass
elif not visited[to]:
- outEdgeCount = dfs(root, to, at, outEdgeCount)
+ out_edge_count = dfs(root, to, at, out_edge_count)
low[at] = min(low[at], low[to])
# AP found via bridge
if at < low[to]:
- isArt[at] = True
+ is_art[at] = True
# AP found via cycle
if at == low[to]:
- isArt[at] = True
+ is_art[at] = True
else:
low[at] = min(low[at], to)
- return outEdgeCount
+ return out_edge_count
for i in range(n):
if not visited[i]:
- outEdgeCount = 0
- outEdgeCount = dfs(i, i, -1, outEdgeCount)
- isArt[i] = outEdgeCount > 1
+ out_edge_count = 0
+ out_edge_count = dfs(i, i, -1, out_edge_count)
+ is_art[i] = out_edge_count > 1
- for x in range(len(isArt)):
- if isArt[x] is True:
+ for x in range(len(is_art)):
+ if is_art[x] is True:
print(x)
# Adjacency list of graph
-data = {
+graph = {
0: [1, 2],
1: [0, 2],
2: [0, 1, 3, 5],
@@ -52,4 +52,4 @@ def dfs(root, at, parent, outEdgeCount):
7: [6, 8],
8: [5, 7],
}
-computeAP(data)
+compute_ap(graph)
diff --git a/graphs/basic_graphs.py b/graphs/basic_graphs.py
index 0f73d8d07b2a..286e9b195796 100644
--- a/graphs/basic_graphs.py
+++ b/graphs/basic_graphs.py
@@ -1,42 +1,69 @@
from collections import deque
-if __name__ == "__main__":
- # Accept No. of Nodes and edges
- n, m = map(int, input().split(" "))
- # Initialising Dictionary of edges
- g = {}
- for i in range(n):
- g[i + 1] = []
+def _input(message):
+ return input(message).strip().split(" ")
- """
- ----------------------------------------------------------------------------
- Accepting edges of Unweighted Directed Graphs
- ----------------------------------------------------------------------------
- """
- for _ in range(m):
- x, y = map(int, input().strip().split(" "))
- g[x].append(y)
- """
- ----------------------------------------------------------------------------
- Accepting edges of Unweighted Undirected Graphs
- ----------------------------------------------------------------------------
- """
- for _ in range(m):
- x, y = map(int, input().strip().split(" "))
- g[x].append(y)
- g[y].append(x)
+def initialize_unweighted_directed_graph(
+ node_count: int, edge_count: int
+) -> dict[int, list[int]]:
+ graph: dict[int, list[int]] = {}
+ for i in range(node_count):
+ graph[i + 1] = []
+
+ for e in range(edge_count):
+ x, y = (int(i) for i in _input(f"Edge {e + 1}: "))
+ graph[x].append(y)
+ return graph
+
+
+def initialize_unweighted_undirected_graph(
+ node_count: int, edge_count: int
+) -> dict[int, list[int]]:
+ graph: dict[int, list[int]] = {}
+ for i in range(node_count):
+ graph[i + 1] = []
+
+ for e in range(edge_count):
+ x, y = (int(i) for i in _input(f"Edge {e + 1}: "))
+ graph[x].append(y)
+ graph[y].append(x)
+ return graph
+
+
+def initialize_weighted_undirected_graph(
+ node_count: int, edge_count: int
+) -> dict[int, list[tuple[int, int]]]:
+ graph: dict[int, list[tuple[int, int]]] = {}
+ for i in range(node_count):
+ graph[i + 1] = []
+
+ for e in range(edge_count):
+ x, y, w = (int(i) for i in _input(f"Edge {e + 1}: "))
+ graph[x].append((y, w))
+ graph[y].append((x, w))
+ return graph
+
+
+if __name__ == "__main__":
+ n, m = (int(i) for i in _input("Number of nodes and edges: "))
+
+ graph_choice = int(
+ _input(
+ "Press 1 or 2 or 3 \n"
+ "1. Unweighted directed \n"
+ "2. Unweighted undirected \n"
+ "3. Weighted undirected \n"
+ )[0]
+ )
+
+ g = {
+ 1: initialize_unweighted_directed_graph,
+ 2: initialize_unweighted_undirected_graph,
+ 3: initialize_weighted_undirected_graph,
+ }[graph_choice](n, m)
- """
- ----------------------------------------------------------------------------
- Accepting edges of Weighted Undirected Graphs
- ----------------------------------------------------------------------------
- """
- for _ in range(m):
- x, y, r = map(int, input().strip().split(" "))
- g[x].append([y, r])
- g[y].append([x, r])
"""
--------------------------------------------------------------------------------
@@ -49,20 +76,28 @@
"""
-def dfs(G, s):
- vis, S = {s}, [s]
+def dfs(g, s):
+ """
+ >>> dfs({1: [2, 3], 2: [4, 5], 3: [], 4: [], 5: []}, 1)
+ 1
+ 2
+ 4
+ 5
+ 3
+ """
+ vis, _s = {s}, [s]
print(s)
- while S:
+ while _s:
flag = 0
- for i in G[S[-1]]:
+ for i in g[_s[-1]]:
if i not in vis:
- S.append(i)
+ _s.append(i)
vis.add(i)
flag = 1
print(i)
break
if not flag:
- S.pop()
+ _s.pop()
"""
@@ -76,15 +111,26 @@ def dfs(G, s):
"""
-def bfs(G, s):
- vis, Q = {s}, deque([s])
+def bfs(g, s):
+ """
+ >>> bfs({1: [2, 3], 2: [4, 5], 3: [6, 7], 4: [], 5: [8], 6: [], 7: [], 8: []}, 1)
+ 1
+ 2
+ 3
+ 4
+ 5
+ 6
+ 7
+ 8
+ """
+ vis, q = {s}, deque([s])
print(s)
- while Q:
- u = Q.popleft()
- for v in G[u]:
+ while q:
+ u = q.popleft()
+ for v in g[u]:
if v not in vis:
vis.add(v)
- Q.append(v)
+ q.append(v)
print(v)
@@ -100,25 +146,37 @@ def bfs(G, s):
"""
-def dijk(G, s):
+def dijk(g, s):
+ """
+ dijk({1: [(2, 7), (3, 9), (6, 14)],
+ 2: [(1, 7), (3, 10), (4, 15)],
+ 3: [(1, 9), (2, 10), (4, 11), (6, 2)],
+ 4: [(2, 15), (3, 11), (5, 6)],
+ 5: [(4, 6), (6, 9)],
+ 6: [(1, 14), (3, 2), (5, 9)]}, 1)
+ 7
+ 9
+ 11
+ 20
+ 20
+ """
dist, known, path = {s: 0}, set(), {s: 0}
while True:
- if len(known) == len(G) - 1:
+ if len(known) == len(g) - 1:
break
mini = 100000
- for i in dist:
- if i not in known and dist[i] < mini:
- mini = dist[i]
- u = i
+ for key, value in dist:
+ if key not in known and value < mini:
+ mini = value
+ u = key
known.add(u)
- for v in G[u]:
- if v[0] not in known:
- if dist[u] + v[1] < dist.get(v[0], 100000):
- dist[v[0]] = dist[u] + v[1]
- path[v[0]] = u
- for i in dist:
- if i != s:
- print(dist[i])
+ for v in g[u]:
+ if v[0] not in known and dist[u] + v[1] < dist.get(v[0], 100000):
+ dist[v[0]] = dist[u] + v[1]
+ path[v[0]] = u
+ for key, value in dist.items():
+ if key != s:
+ print(value)
"""
@@ -128,27 +186,27 @@ def dijk(G, s):
"""
-def topo(G, ind=None, Q=None):
- if Q is None:
- Q = [1]
+def topo(g, ind=None, q=None):
+ if q is None:
+ q = [1]
if ind is None:
- ind = [0] * (len(G) + 1) # SInce oth Index is ignored
- for u in G:
- for v in G[u]:
+ ind = [0] * (len(g) + 1) # SInce oth Index is ignored
+ for u in g:
+ for v in g[u]:
ind[v] += 1
- Q = deque()
- for i in G:
+ q = deque()
+ for i in g:
if ind[i] == 0:
- Q.append(i)
- if len(Q) == 0:
+ q.append(i)
+ if len(q) == 0:
return
- v = Q.popleft()
+ v = q.popleft()
print(v)
- for w in G[v]:
+ for w in g[v]:
ind[w] -= 1
if ind[w] == 0:
- Q.append(w)
- topo(G, ind, Q)
+ q.append(w)
+ topo(g, ind, q)
"""
@@ -159,10 +217,29 @@ def topo(G, ind=None, Q=None):
def adjm():
- n = input().strip()
+ r"""
+ Reading an Adjacency matrix
+
+ Parameters:
+ None
+
+ Returns:
+ tuple: A tuple containing a list of edges and number of edges
+
+ Example:
+ >>> # Simulate user input for 3 nodes
+ >>> input_data = "4\n0 1 0 1\n1 0 1 0\n0 1 0 1\n1 0 1 0\n"
+ >>> import sys,io
+ >>> original_input = sys.stdin
+ >>> sys.stdin = io.StringIO(input_data) # Redirect stdin for testing
+ >>> adjm()
+ ([(0, 1, 0, 1), (1, 0, 1, 0), (0, 1, 0, 1), (1, 0, 1, 0)], 4)
+ >>> sys.stdin = original_input # Restore original stdin
+ """
+ n = int(input().strip())
a = []
- for i in range(n):
- a.append(map(int, input().strip().split()))
+ for _ in range(n):
+ a.append(tuple(map(int, input().strip().split())))
return a, n
@@ -179,9 +256,9 @@ def adjm():
"""
-def floy(A_and_n):
- (A, n) = A_and_n
- dist = list(A)
+def floy(a_and_n):
+ (a, n) = a_and_n
+ dist = list(a)
path = [[0] * n for i in range(n)]
for k in range(n):
for i in range(n):
@@ -204,22 +281,21 @@ def floy(A_and_n):
"""
-def prim(G, s):
+def prim(g, s):
dist, known, path = {s: 0}, set(), {s: 0}
while True:
- if len(known) == len(G) - 1:
+ if len(known) == len(g) - 1:
break
mini = 100000
- for i in dist:
- if i not in known and dist[i] < mini:
- mini = dist[i]
- u = i
+ for key, value in dist.items():
+ if key not in known and value < mini:
+ mini = value
+ u = key
known.add(u)
- for v in G[u]:
- if v[0] not in known:
- if v[1] < dist.get(v[0], 100000):
- dist[v[0]] = v[1]
- path[v[0]] = u
+ for v in g[u]:
+ if v[0] not in known and v[1] < dist.get(v[0], 100000):
+ dist[v[0]] = v[1]
+ path[v[0]] = u
return dist
@@ -235,10 +311,29 @@ def prim(G, s):
def edglist():
- n, m = map(int, input().split(" "))
+ r"""
+ Get the edges and number of edges from the user
+
+ Parameters:
+ None
+
+ Returns:
+ tuple: A tuple containing a list of edges and number of edges
+
+ Example:
+ >>> # Simulate user input for 3 edges and 4 vertices: (1, 2), (2, 3), (3, 4)
+ >>> input_data = "4 3\n1 2\n2 3\n3 4\n"
+ >>> import sys,io
+ >>> original_input = sys.stdin
+ >>> sys.stdin = io.StringIO(input_data) # Redirect stdin for testing
+ >>> edglist()
+ ([(1, 2), (2, 3), (3, 4)], 4)
+ >>> sys.stdin = original_input # Restore original stdin
+ """
+ n, m = tuple(map(int, input().split(" ")))
edges = []
- for i in range(m):
- edges.append(map(int, input().split(" ")))
+ for _ in range(m):
+ edges.append(tuple(map(int, input().split(" "))))
return edges, n
@@ -252,16 +347,18 @@ def edglist():
"""
-def krusk(E_and_n):
- # Sort edges on the basis of distance
- (E, n) = E_and_n
- E.sort(reverse=True, key=lambda x: x[2])
+def krusk(e_and_n):
+ """
+ Sort edges on the basis of distance
+ """
+ (e, n) = e_and_n
+ e.sort(reverse=True, key=lambda x: x[2])
s = [{i} for i in range(1, n + 1)]
while True:
if len(s) == 1:
break
print(s)
- x = E.pop()
+ x = e.pop()
for i in range(len(s)):
if x[0] in s[i]:
break
@@ -274,8 +371,37 @@ def krusk(E_and_n):
break
-# find the isolated node in the graph
def find_isolated_nodes(graph):
+ """
+ Find the isolated node in the graph
+
+ Parameters:
+ graph (dict): A dictionary representing a graph.
+
+ Returns:
+ list: A list of isolated nodes.
+
+ Examples:
+ >>> graph1 = {1: [2, 3], 2: [1, 3], 3: [1, 2], 4: []}
+ >>> find_isolated_nodes(graph1)
+ [4]
+
+ >>> graph2 = {'A': ['B', 'C'], 'B': ['A'], 'C': ['A'], 'D': []}
+ >>> find_isolated_nodes(graph2)
+ ['D']
+
+ >>> graph3 = {'X': [], 'Y': [], 'Z': []}
+ >>> find_isolated_nodes(graph3)
+ ['X', 'Y', 'Z']
+
+ >>> graph4 = {1: [2, 3], 2: [1, 3], 3: [1, 2]}
+ >>> find_isolated_nodes(graph4)
+ []
+
+ >>> graph5 = {}
+ >>> find_isolated_nodes(graph5)
+ []
+ """
isolated = []
for node in graph:
if not graph[node]:
diff --git a/graphs/bellman_ford.py b/graphs/bellman_ford.py
index ace7985647bb..9ac8bae85d4f 100644
--- a/graphs/bellman_ford.py
+++ b/graphs/bellman_ford.py
@@ -1,56 +1,73 @@
from __future__ import annotations
-def printDist(dist, V):
- print("Vertex Distance")
- distances = ("INF" if d == float("inf") else d for d in dist)
- print("\t".join(f"{i}\t{d}" for i, d in enumerate(distances)))
+def print_distance(distance: list[float], src):
+ print(f"Vertex\tShortest Distance from vertex {src}")
+ for i, d in enumerate(distance):
+ print(f"{i}\t\t{d}")
-def BellmanFord(graph: list[dict[str, int]], V: int, E: int, src: int) -> int:
+def check_negative_cycle(
+ graph: list[dict[str, int]], distance: list[float], edge_count: int
+):
+ for j in range(edge_count):
+ u, v, w = (graph[j][k] for k in ["src", "dst", "weight"])
+ if distance[u] != float("inf") and distance[u] + w < distance[v]:
+ return True
+ return False
+
+
+def bellman_ford(
+ graph: list[dict[str, int]], vertex_count: int, edge_count: int, src: int
+) -> list[float]:
"""
Returns shortest paths from a vertex src to all
other vertices.
+ >>> edges = [(2, 1, -10), (3, 2, 3), (0, 3, 5), (0, 1, 4)]
+ >>> g = [{"src": s, "dst": d, "weight": w} for s, d, w in edges]
+ >>> bellman_ford(g, 4, 4, 0)
+ [0.0, -2.0, 8.0, 5.0]
+ >>> g = [{"src": s, "dst": d, "weight": w} for s, d, w in edges + [(1, 3, 5)]]
+ >>> bellman_ford(g, 4, 5, 0)
+ Traceback (most recent call last):
+ ...
+ Exception: Negative cycle found
"""
- mdist = [float("inf") for i in range(V)]
- mdist[src] = 0.0
+ distance = [float("inf")] * vertex_count
+ distance[src] = 0.0
- for i in range(V - 1):
- for j in range(E):
- u = graph[j]["src"]
- v = graph[j]["dst"]
- w = graph[j]["weight"]
+ for _ in range(vertex_count - 1):
+ for j in range(edge_count):
+ u, v, w = (graph[j][k] for k in ["src", "dst", "weight"])
- if mdist[u] != float("inf") and mdist[u] + w < mdist[v]:
- mdist[v] = mdist[u] + w
- for j in range(E):
- u = graph[j]["src"]
- v = graph[j]["dst"]
- w = graph[j]["weight"]
+ if distance[u] != float("inf") and distance[u] + w < distance[v]:
+ distance[v] = distance[u] + w
- if mdist[u] != float("inf") and mdist[u] + w < mdist[v]:
- print("Negative cycle found. Solution not possible.")
- return
+ negative_cycle_exists = check_negative_cycle(graph, distance, edge_count)
+ if negative_cycle_exists:
+ raise Exception("Negative cycle found")
- printDist(mdist, V)
- return src
+ return distance
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
V = int(input("Enter number of vertices: ").strip())
E = int(input("Enter number of edges: ").strip())
- graph = [dict() for j in range(E)]
+ graph: list[dict[str, int]] = [{} for _ in range(E)]
for i in range(E):
- graph[i][i] = 0.0
+ print("Edge ", i + 1)
+ src, dest, weight = (
+ int(x)
+ for x in input("Enter source, destination, weight: ").strip().split(" ")
+ )
+ graph[i] = {"src": src, "dst": dest, "weight": weight}
- for i in range(E):
- print("\nEdge ", i + 1)
- src = int(input("Enter source:").strip())
- dst = int(input("Enter destination:").strip())
- weight = float(input("Enter weight:").strip())
- graph[i] = {"src": src, "dst": dst, "weight": weight}
-
- gsrc = int(input("\nEnter shortest path source:").strip())
- BellmanFord(graph, V, E, gsrc)
+ source = int(input("\nEnter shortest path source:").strip())
+ shortest_distance = bellman_ford(graph, V, E, source)
+ print_distance(shortest_distance, 0)
diff --git a/graphs/bi_directional_dijkstra.py b/graphs/bi_directional_dijkstra.py
new file mode 100644
index 000000000000..d2c4030b921b
--- /dev/null
+++ b/graphs/bi_directional_dijkstra.py
@@ -0,0 +1,140 @@
+"""
+Bi-directional Dijkstra's algorithm.
+
+A bi-directional approach is an efficient and
+less time consuming optimization for Dijkstra's
+searching algorithm
+
+Reference: shorturl.at/exHM7
+"""
+
+# Author: Swayam Singh (https://github.com/practice404)
+
+from queue import PriorityQueue
+from typing import Any
+
+import numpy as np
+
+
+def pass_and_relaxation(
+ graph: dict,
+ v: str,
+ visited_forward: set,
+ visited_backward: set,
+ cst_fwd: dict,
+ cst_bwd: dict,
+ queue: PriorityQueue,
+ parent: dict,
+ shortest_distance: float,
+) -> float:
+ for nxt, d in graph[v]:
+ if nxt in visited_forward:
+ continue
+ old_cost_f = cst_fwd.get(nxt, np.inf)
+ new_cost_f = cst_fwd[v] + d
+ if new_cost_f < old_cost_f:
+ queue.put((new_cost_f, nxt))
+ cst_fwd[nxt] = new_cost_f
+ parent[nxt] = v
+ if (
+ nxt in visited_backward
+ and cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance
+ ):
+ shortest_distance = cst_fwd[v] + d + cst_bwd[nxt]
+ return shortest_distance
+
+
+def bidirectional_dij(
+ source: str, destination: str, graph_forward: dict, graph_backward: dict
+) -> int:
+ """
+ Bi-directional Dijkstra's algorithm.
+
+ Returns:
+ shortest_path_distance (int): length of the shortest path.
+
+ Warnings:
+ If the destination is not reachable, function returns -1
+
+ >>> bidirectional_dij("E", "F", graph_fwd, graph_bwd)
+ 3
+ """
+ shortest_path_distance = -1
+
+ visited_forward = set()
+ visited_backward = set()
+ cst_fwd = {source: 0}
+ cst_bwd = {destination: 0}
+ parent_forward = {source: None}
+ parent_backward = {destination: None}
+ queue_forward: PriorityQueue[Any] = PriorityQueue()
+ queue_backward: PriorityQueue[Any] = PriorityQueue()
+
+ shortest_distance = np.inf
+
+ queue_forward.put((0, source))
+ queue_backward.put((0, destination))
+
+ if source == destination:
+ return 0
+
+ while not queue_forward.empty() and not queue_backward.empty():
+ _, v_fwd = queue_forward.get()
+ visited_forward.add(v_fwd)
+
+ _, v_bwd = queue_backward.get()
+ visited_backward.add(v_bwd)
+
+ shortest_distance = pass_and_relaxation(
+ graph_forward,
+ v_fwd,
+ visited_forward,
+ visited_backward,
+ cst_fwd,
+ cst_bwd,
+ queue_forward,
+ parent_forward,
+ shortest_distance,
+ )
+
+ shortest_distance = pass_and_relaxation(
+ graph_backward,
+ v_bwd,
+ visited_backward,
+ visited_forward,
+ cst_bwd,
+ cst_fwd,
+ queue_backward,
+ parent_backward,
+ shortest_distance,
+ )
+
+ if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance:
+ break
+
+ if shortest_distance != np.inf:
+ shortest_path_distance = shortest_distance
+ return shortest_path_distance
+
+
+graph_fwd = {
+ "B": [["C", 1]],
+ "C": [["D", 1]],
+ "D": [["F", 1]],
+ "E": [["B", 1], ["G", 2]],
+ "F": [],
+ "G": [["F", 1]],
+}
+graph_bwd = {
+ "B": [["E", 1]],
+ "C": [["B", 1]],
+ "D": [["C", 1]],
+ "F": [["D", 1], ["G", 1]],
+ "E": [[None, np.inf]],
+ "G": [["E", 2]],
+}
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/bidirectional_a_star.py b/graphs/bidirectional_a_star.py
index 72ff4fa65ff0..00f623de3493 100644
--- a/graphs/bidirectional_a_star.py
+++ b/graphs/bidirectional_a_star.py
@@ -22,6 +22,8 @@
delta = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right
+TPosition = tuple[int, int]
+
class Node:
"""
@@ -39,7 +41,15 @@ class Node:
True
"""
- def __init__(self, pos_x, pos_y, goal_x, goal_y, g_cost, parent):
+ def __init__(
+ self,
+ pos_x: int,
+ pos_y: int,
+ goal_x: int,
+ goal_y: int,
+ g_cost: int,
+ parent: Node | None,
+ ) -> None:
self.pos_x = pos_x
self.pos_y = pos_y
self.pos = (pos_y, pos_x)
@@ -59,9 +69,9 @@ def calculate_heuristic(self) -> float:
if HEURISTIC == 1:
return abs(dx) + abs(dy)
else:
- return sqrt(dy ** 2 + dx ** 2)
+ return sqrt(dy**2 + dx**2)
- def __lt__(self, other) -> bool:
+ def __lt__(self, other: Node) -> bool:
return self.f_cost < other.f_cost
@@ -81,23 +91,22 @@ class AStar:
(4, 3), (4, 4), (5, 4), (5, 5), (6, 5), (6, 6)]
"""
- def __init__(self, start, goal):
+ def __init__(self, start: TPosition, goal: TPosition):
self.start = Node(start[1], start[0], goal[1], goal[0], 0, None)
self.target = Node(goal[1], goal[0], goal[1], goal[0], 99999, None)
self.open_nodes = [self.start]
- self.closed_nodes = []
+ self.closed_nodes: list[Node] = []
self.reached = False
- def search(self) -> list[tuple[int]]:
+ def search(self) -> list[TPosition]:
while self.open_nodes:
# Open Nodes are sorted using __lt__
self.open_nodes.sort()
current_node = self.open_nodes.pop(0)
if current_node.pos == self.target.pos:
- self.reached = True
return self.retrace_path(current_node)
self.closed_nodes.append(current_node)
@@ -118,8 +127,7 @@ def search(self) -> list[tuple[int]]:
else:
self.open_nodes.append(better_node)
- if not (self.reached):
- return [(self.start.pos)]
+ return [self.start.pos]
def get_successors(self, parent: Node) -> list[Node]:
"""
@@ -147,7 +155,7 @@ def get_successors(self, parent: Node) -> list[Node]:
)
return successors
- def retrace_path(self, node: Node) -> list[tuple[int]]:
+ def retrace_path(self, node: Node | None) -> list[TPosition]:
"""
Retrace the path from parents to parents until start node
"""
@@ -173,12 +181,12 @@ class BidirectionalAStar:
(2, 5), (3, 5), (4, 5), (5, 5), (5, 6), (6, 6)]
"""
- def __init__(self, start, goal):
+ def __init__(self, start: TPosition, goal: TPosition) -> None:
self.fwd_astar = AStar(start, goal)
self.bwd_astar = AStar(goal, start)
self.reached = False
- def search(self) -> list[tuple[int]]:
+ def search(self) -> list[TPosition]:
while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes:
self.fwd_astar.open_nodes.sort()
self.bwd_astar.open_nodes.sort()
@@ -186,7 +194,6 @@ def search(self) -> list[tuple[int]]:
current_bwd_node = self.bwd_astar.open_nodes.pop(0)
if current_bwd_node.pos == current_fwd_node.pos:
- self.reached = True
return self.retrace_bidirectional_path(
current_fwd_node, current_bwd_node
)
@@ -220,12 +227,11 @@ def search(self) -> list[tuple[int]]:
else:
astar.open_nodes.append(better_node)
- if not self.reached:
- return [self.fwd_astar.start.pos]
+ return [self.fwd_astar.start.pos]
def retrace_bidirectional_path(
self, fwd_node: Node, bwd_node: Node
- ) -> list[tuple[int]]:
+ ) -> list[TPosition]:
fwd_path = self.fwd_astar.retrace_path(fwd_node)
bwd_path = self.bwd_astar.retrace_path(bwd_node)
bwd_path.pop()
@@ -236,9 +242,6 @@ def retrace_bidirectional_path(
if __name__ == "__main__":
# all coordinates are given in format [y,x]
- import doctest
-
- doctest.testmod()
init = (0, 0)
goal = (len(grid) - 1, len(grid[0]) - 1)
for elem in grid:
@@ -252,6 +255,5 @@ def retrace_bidirectional_path(
bd_start_time = time.time()
bidir_astar = BidirectionalAStar(init, goal)
- path = bidir_astar.search()
bd_end_time = time.time() - bd_start_time
print(f"BidirectionalAStar execution time = {bd_end_time:f} seconds")
diff --git a/graphs/bidirectional_breadth_first_search.py b/graphs/bidirectional_breadth_first_search.py
index 39d8dc7d4187..71c5a9aff08f 100644
--- a/graphs/bidirectional_breadth_first_search.py
+++ b/graphs/bidirectional_breadth_first_search.py
@@ -6,6 +6,8 @@
import time
+Path = list[tuple[int, int]]
+
grid = [
[0, 0, 0, 0, 0, 0, 0],
[0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles
@@ -20,7 +22,9 @@
class Node:
- def __init__(self, pos_x, pos_y, goal_x, goal_y, parent):
+ def __init__(
+ self, pos_x: int, pos_y: int, goal_x: int, goal_y: int, parent: Node | None
+ ):
self.pos_x = pos_x
self.pos_y = pos_y
self.pos = (pos_y, pos_x)
@@ -31,28 +35,31 @@ def __init__(self, pos_x, pos_y, goal_x, goal_y, parent):
class BreadthFirstSearch:
"""
- >>> bfs = BreadthFirstSearch((0, 0), (len(grid) - 1, len(grid[0]) - 1))
- >>> (bfs.start.pos_y + delta[3][0], bfs.start.pos_x + delta[3][1])
+ # Comment out slow pytests...
+ # 9.15s call graphs/bidirectional_breadth_first_search.py:: \
+ # graphs.bidirectional_breadth_first_search.BreadthFirstSearch
+ # >>> bfs = BreadthFirstSearch((0, 0), (len(grid) - 1, len(grid[0]) - 1))
+ # >>> (bfs.start.pos_y + delta[3][0], bfs.start.pos_x + delta[3][1])
(0, 1)
- >>> [x.pos for x in bfs.get_successors(bfs.start)]
+ # >>> [x.pos for x in bfs.get_successors(bfs.start)]
[(1, 0), (0, 1)]
- >>> (bfs.start.pos_y + delta[2][0], bfs.start.pos_x + delta[2][1])
+ # >>> (bfs.start.pos_y + delta[2][0], bfs.start.pos_x + delta[2][1])
(1, 0)
- >>> bfs.retrace_path(bfs.start)
+ # >>> bfs.retrace_path(bfs.start)
[(0, 0)]
- >>> bfs.search() # doctest: +NORMALIZE_WHITESPACE
+ # >>> bfs.search() # doctest: +NORMALIZE_WHITESPACE
[(0, 0), (1, 0), (2, 0), (3, 0), (3, 1), (4, 1),
(5, 1), (5, 2), (5, 3), (5, 4), (5, 5), (6, 5), (6, 6)]
"""
- def __init__(self, start, goal):
+ def __init__(self, start: tuple[int, int], goal: tuple[int, int]):
self.start = Node(start[1], start[0], goal[1], goal[0], None)
self.target = Node(goal[1], goal[0], goal[1], goal[0], None)
self.node_queue = [self.start]
self.reached = False
- def search(self) -> list[tuple[int]]:
+ def search(self) -> Path | None:
while self.node_queue:
current_node = self.node_queue.pop(0)
@@ -65,8 +72,9 @@ def search(self) -> list[tuple[int]]:
for node in successors:
self.node_queue.append(node)
- if not (self.reached):
- return [(self.start.pos)]
+ if not self.reached:
+ return [self.start.pos]
+ return None
def get_successors(self, parent: Node) -> list[Node]:
"""
@@ -87,7 +95,7 @@ def get_successors(self, parent: Node) -> list[Node]:
)
return successors
- def retrace_path(self, node: Node) -> list[tuple[int]]:
+ def retrace_path(self, node: Node | None) -> Path:
"""
Retrace the path from parents to parents until start node
"""
@@ -119,7 +127,7 @@ def __init__(self, start, goal):
self.bwd_bfs = BreadthFirstSearch(goal, start)
self.reached = False
- def search(self) -> list[tuple[int]]:
+ def search(self) -> Path | None:
while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue:
current_fwd_node = self.fwd_bfs.node_queue.pop(0)
current_bwd_node = self.bwd_bfs.node_queue.pop(0)
@@ -144,10 +152,9 @@ def search(self) -> list[tuple[int]]:
if not self.reached:
return [self.fwd_bfs.start.pos]
+ return None
- def retrace_bidirectional_path(
- self, fwd_node: Node, bwd_node: Node
- ) -> list[tuple[int]]:
+ def retrace_bidirectional_path(self, fwd_node: Node, bwd_node: Node) -> Path:
fwd_path = self.fwd_bfs.retrace_path(fwd_node)
bwd_path = self.bwd_bfs.retrace_path(bwd_node)
bwd_path.pop()
diff --git a/graphs/boruvka.py b/graphs/boruvka.py
new file mode 100644
index 000000000000..3dc059ff6a62
--- /dev/null
+++ b/graphs/boruvka.py
@@ -0,0 +1,176 @@
+"""Borůvka's algorithm.
+
+Determines the minimum spanning tree (MST) of a graph using the Borůvka's algorithm.
+Borůvka's algorithm is a greedy algorithm for finding a minimum spanning tree in a
+connected graph, or a minimum spanning forest if a graph that is not connected.
+
+The time complexity of this algorithm is O(ELogV), where E represents the number
+of edges, while V represents the number of nodes.
+O(number_of_edges Log number_of_nodes)
+
+The space complexity of this algorithm is O(V + E), since we have to keep a couple
+of lists whose sizes are equal to the number of nodes, as well as keep all the
+edges of a graph inside of the data structure itself.
+
+Borůvka's algorithm gives us pretty much the same result as other MST Algorithms -
+they all find the minimum spanning tree, and the time complexity is approximately
+the same.
+
+One advantage that Borůvka's algorithm has compared to the alternatives is that it
+doesn't need to presort the edges or maintain a priority queue in order to find the
+minimum spanning tree.
+Even though that doesn't help its complexity, since it still passes the edges logE
+times, it is a bit simpler to code.
+
+Details: https://en.wikipedia.org/wiki/Bor%C5%AFvka%27s_algorithm
+"""
+
+from __future__ import annotations
+
+from typing import Any
+
+
+class Graph:
+ def __init__(self, num_of_nodes: int) -> None:
+ """
+ Arguments:
+ num_of_nodes - the number of nodes in the graph
+ Attributes:
+ m_num_of_nodes - the number of nodes in the graph.
+ m_edges - the list of edges.
+ m_component - the dictionary which stores the index of the component which
+ a node belongs to.
+ """
+
+ self.m_num_of_nodes = num_of_nodes
+ self.m_edges: list[list[int]] = []
+ self.m_component: dict[int, int] = {}
+
+ def add_edge(self, u_node: int, v_node: int, weight: int) -> None:
+ """Adds an edge in the format [first, second, edge weight] to graph."""
+
+ self.m_edges.append([u_node, v_node, weight])
+
+ def find_component(self, u_node: int) -> int:
+ """Propagates a new component throughout a given component."""
+
+ if self.m_component[u_node] == u_node:
+ return u_node
+ return self.find_component(self.m_component[u_node])
+
+ def set_component(self, u_node: int) -> None:
+ """Finds the component index of a given node"""
+
+ if self.m_component[u_node] != u_node:
+ for k in self.m_component:
+ self.m_component[k] = self.find_component(k)
+
+ def union(self, component_size: list[int], u_node: int, v_node: int) -> None:
+ """Union finds the roots of components for two nodes, compares the components
+ in terms of size, and attaches the smaller one to the larger one to form
+ single component"""
+
+ if component_size[u_node] <= component_size[v_node]:
+ self.m_component[u_node] = v_node
+ component_size[v_node] += component_size[u_node]
+ self.set_component(u_node)
+
+ elif component_size[u_node] >= component_size[v_node]:
+ self.m_component[v_node] = self.find_component(u_node)
+ component_size[u_node] += component_size[v_node]
+ self.set_component(v_node)
+
+ def boruvka(self) -> None:
+ """Performs Borůvka's algorithm to find MST."""
+
+ # Initialize additional lists required to algorithm.
+ component_size = []
+ mst_weight = 0
+
+ minimum_weight_edge: list[Any] = [-1] * self.m_num_of_nodes
+
+ # A list of components (initialized to all of the nodes)
+ for node in range(self.m_num_of_nodes):
+ self.m_component.update({node: node})
+ component_size.append(1)
+
+ num_of_components = self.m_num_of_nodes
+
+ while num_of_components > 1:
+ for edge in self.m_edges:
+ u, v, w = edge
+
+ u_component = self.m_component[u]
+ v_component = self.m_component[v]
+
+ if u_component != v_component:
+ """If the current minimum weight edge of component u doesn't
+ exist (is -1), or if it's greater than the edge we're
+ observing right now, we will assign the value of the edge
+ we're observing to it.
+
+ If the current minimum weight edge of component v doesn't
+ exist (is -1), or if it's greater than the edge we're
+ observing right now, we will assign the value of the edge
+ we're observing to it"""
+
+ for component in (u_component, v_component):
+ if (
+ minimum_weight_edge[component] == -1
+ or minimum_weight_edge[component][2] > w
+ ):
+ minimum_weight_edge[component] = [u, v, w]
+
+ for edge in minimum_weight_edge:
+ if isinstance(edge, list):
+ u, v, w = edge
+
+ u_component = self.m_component[u]
+ v_component = self.m_component[v]
+
+ if u_component != v_component:
+ mst_weight += w
+ self.union(component_size, u_component, v_component)
+ print(f"Added edge [{u} - {v}]\nAdded weight: {w}\n")
+ num_of_components -= 1
+
+ minimum_weight_edge = [-1] * self.m_num_of_nodes
+ print(f"The total weight of the minimal spanning tree is: {mst_weight}")
+
+
+def test_vector() -> None:
+ """
+ >>> g = Graph(8)
+ >>> for u_v_w in ((0, 1, 10), (0, 2, 6), (0, 3, 5), (1, 3, 15), (2, 3, 4),
+ ... (3, 4, 8), (4, 5, 10), (4, 6, 6), (4, 7, 5), (5, 7, 15), (6, 7, 4)):
+ ... g.add_edge(*u_v_w)
+ >>> g.boruvka()
+ Added edge [0 - 3]
+ Added weight: 5
+
+ Added edge [0 - 1]
+ Added weight: 10
+
+ Added edge [2 - 3]
+ Added weight: 4
+
+ Added edge [4 - 7]
+ Added weight: 5
+
+ Added edge [4 - 5]
+ Added weight: 10
+
+ Added edge [6 - 7]
+ Added weight: 4
+
+ Added edge [3 - 4]
+ Added weight: 8
+
+ The total weight of the minimal spanning tree is: 46
+ """
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/breadth_first_search.py b/graphs/breadth_first_search.py
index e40ec9d1d06d..cab79be39ed3 100644
--- a/graphs/breadth_first_search.py
+++ b/graphs/breadth_first_search.py
@@ -1,59 +1,93 @@
#!/usr/bin/python
-""" Author: OMKAR PATHAK """
+"""Author: OMKAR PATHAK"""
+
+from __future__ import annotations
+
+from queue import Queue
class Graph:
- def __init__(self):
- self.vertices = {}
+ def __init__(self) -> None:
+ self.vertices: dict[int, list[int]] = {}
- def printGraph(self):
- """prints adjacency list representation of graaph"""
- for i in self.vertices.keys():
+ def print_graph(self) -> None:
+ """
+ prints adjacency list representation of graaph
+ >>> g = Graph()
+ >>> g.print_graph()
+ >>> g.add_edge(0, 1)
+ >>> g.print_graph()
+ 0 : 1
+ """
+ for i in self.vertices:
print(i, " : ", " -> ".join([str(j) for j in self.vertices[i]]))
- def addEdge(self, fromVertex, toVertex):
- """adding the edge between two vertices"""
- if fromVertex in self.vertices.keys():
- self.vertices[fromVertex].append(toVertex)
+ def add_edge(self, from_vertex: int, to_vertex: int) -> None:
+ """
+ adding the edge between two vertices
+ >>> g = Graph()
+ >>> g.print_graph()
+ >>> g.add_edge(0, 1)
+ >>> g.print_graph()
+ 0 : 1
+ """
+ if from_vertex in self.vertices:
+ self.vertices[from_vertex].append(to_vertex)
else:
- self.vertices[fromVertex] = [toVertex]
+ self.vertices[from_vertex] = [to_vertex]
- def BFS(self, startVertex):
+ def bfs(self, start_vertex: int) -> set[int]:
+ """
+ >>> g = Graph()
+ >>> g.add_edge(0, 1)
+ >>> g.add_edge(0, 1)
+ >>> g.add_edge(0, 2)
+ >>> g.add_edge(1, 2)
+ >>> g.add_edge(2, 0)
+ >>> g.add_edge(2, 3)
+ >>> g.add_edge(3, 3)
+ >>> sorted(g.bfs(2))
+ [0, 1, 2, 3]
+ """
# initialize set for storing already visited vertices
visited = set()
# create a first in first out queue to store all the vertices for BFS
- queue = []
+ queue: Queue = Queue()
# mark the source node as visited and enqueue it
- visited.add(startVertex)
- queue.append(startVertex)
+ visited.add(start_vertex)
+ queue.put(start_vertex)
- while queue:
- vertex = queue.pop(0)
+ while not queue.empty():
+ vertex = queue.get()
# loop through all adjacent vertex and enqueue it if not yet visited
for adjacent_vertex in self.vertices[vertex]:
if adjacent_vertex not in visited:
- queue.append(adjacent_vertex)
+ queue.put(adjacent_vertex)
visited.add(adjacent_vertex)
return visited
if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod(verbose=True)
+
g = Graph()
- g.addEdge(0, 1)
- g.addEdge(0, 2)
- g.addEdge(1, 2)
- g.addEdge(2, 0)
- g.addEdge(2, 3)
- g.addEdge(3, 3)
-
- g.printGraph()
+ g.add_edge(0, 1)
+ g.add_edge(0, 2)
+ g.add_edge(1, 2)
+ g.add_edge(2, 0)
+ g.add_edge(2, 3)
+ g.add_edge(3, 3)
+
+ g.print_graph()
# 0 : 1 -> 2
# 1 : 2
# 2 : 0 -> 3
# 3 : 3
- assert sorted(g.BFS(2)) == [0, 1, 2, 3]
+ assert sorted(g.bfs(2)) == [0, 1, 2, 3]
diff --git a/graphs/breadth_first_search_2.py b/graphs/breadth_first_search_2.py
index a90e963a4043..ccadfa346bf1 100644
--- a/graphs/breadth_first_search_2.py
+++ b/graphs/breadth_first_search_2.py
@@ -12,8 +12,13 @@
mark w as explored
add w to Q (at the end)
"""
+
from __future__ import annotations
+from collections import deque
+from queue import Queue
+from timeit import timeit
+
G = {
"A": ["B", "C"],
"B": ["A", "D", "E"],
@@ -24,21 +29,60 @@
}
-def breadth_first_search(graph: dict, start: str) -> set[str]:
+def breadth_first_search(graph: dict, start: str) -> list[str]:
"""
- >>> ''.join(sorted(breadth_first_search(G, 'A')))
+ Implementation of breadth first search using queue.Queue.
+
+ >>> ''.join(breadth_first_search(G, 'A'))
'ABCDEF'
"""
explored = {start}
- queue = [start]
- while queue:
- v = queue.pop(0) # queue.popleft()
+ result = [start]
+ queue: Queue = Queue()
+ queue.put(start)
+ while not queue.empty():
+ v = queue.get()
for w in graph[v]:
if w not in explored:
explored.add(w)
- queue.append(w)
- return explored
+ result.append(w)
+ queue.put(w)
+ return result
+
+
+def breadth_first_search_with_deque(graph: dict, start: str) -> list[str]:
+ """
+ Implementation of breadth first search using collection.queue.
+
+ >>> ''.join(breadth_first_search_with_deque(G, 'A'))
+ 'ABCDEF'
+ """
+ visited = {start}
+ result = [start]
+ queue = deque([start])
+ while queue:
+ v = queue.popleft()
+ for child in graph[v]:
+ if child not in visited:
+ visited.add(child)
+ result.append(child)
+ queue.append(child)
+ return result
+
+
+def benchmark_function(name: str) -> None:
+ setup = f"from __main__ import G, {name}"
+ number = 10000
+ res = timeit(f"{name}(G, 'A')", setup=setup, number=number)
+ print(f"{name:<35} finished {number} runs in {res:.5f} seconds")
if __name__ == "__main__":
- print(breadth_first_search(G, "A"))
+ import doctest
+
+ doctest.testmod()
+
+ benchmark_function("breadth_first_search")
+ benchmark_function("breadth_first_search_with_deque")
+ # breadth_first_search finished 10000 runs in 0.20999 seconds
+ # breadth_first_search_with_deque finished 10000 runs in 0.01421 seconds
diff --git a/graphs/breadth_first_search_shortest_path.py b/graphs/breadth_first_search_shortest_path.py
index b43479d4659c..c06440bccef3 100644
--- a/graphs/breadth_first_search_shortest_path.py
+++ b/graphs/breadth_first_search_shortest_path.py
@@ -1,6 +1,7 @@
"""Breath First Search (BFS) can be used when finding the shortest path
from a given source node to a target node in an unweighted graph.
"""
+
from __future__ import annotations
graph = {
@@ -15,17 +16,19 @@
class Graph:
- def __init__(self, graph: dict[str, str], source_vertex: str) -> None:
- """Graph is implemented as dictionary of adjacency lists. Also,
+ def __init__(self, graph: dict[str, list[str]], source_vertex: str) -> None:
+ """
+ Graph is implemented as dictionary of adjacency lists. Also,
Source vertex have to be defined upon initialization.
"""
self.graph = graph
# mapping node to its parent in resulting breadth first tree
- self.parent = {}
+ self.parent: dict[str, str | None] = {}
self.source_vertex = source_vertex
def breath_first_search(self) -> None:
- """This function is a helper for running breath first search on this graph.
+ """
+ This function is a helper for running breath first search on this graph.
>>> g = Graph(graph, "G")
>>> g.breath_first_search()
>>> g.parent
@@ -44,7 +47,8 @@ def breath_first_search(self) -> None:
queue.append(adjacent_vertex)
def shortest_path(self, target_vertex: str) -> str:
- """This shortest path function returns a string, describing the result:
+ """
+ This shortest path function returns a string, describing the result:
1.) No path is found. The string is a human readable message to indicate this.
2.) The shortest path is found. The string is in the form
`v1(->v2->v3->...->vn)`, where v1 is the source vertex and vn is the target
@@ -55,7 +59,9 @@ def shortest_path(self, target_vertex: str) -> str:
Case 1 - No path is found.
>>> g.shortest_path("Foo")
- 'No path from vertex:G to vertex:Foo'
+ Traceback (most recent call last):
+ ...
+ ValueError: No path from vertex: G to vertex: Foo
Case 2 - The path is found.
>>> g.shortest_path("D")
@@ -64,17 +70,19 @@ def shortest_path(self, target_vertex: str) -> str:
'G'
"""
if target_vertex == self.source_vertex:
- return f"{self.source_vertex}"
- elif not self.parent.get(target_vertex):
- return f"No path from vertex:{self.source_vertex} to vertex:{target_vertex}"
- else:
- return self.shortest_path(self.parent[target_vertex]) + f"->{target_vertex}"
+ return self.source_vertex
+ target_vertex_parent = self.parent.get(target_vertex)
+ if target_vertex_parent is None:
+ msg = (
+ f"No path from vertex: {self.source_vertex} to vertex: {target_vertex}"
+ )
+ raise ValueError(msg)
+
+ return self.shortest_path(target_vertex_parent) + f"->{target_vertex}"
-if __name__ == "__main__":
- import doctest
- doctest.testmod()
+if __name__ == "__main__":
g = Graph(graph, "G")
g.breath_first_search()
print(g.shortest_path("D"))
diff --git a/graphs/bfs_shortest_path.py b/graphs/breadth_first_search_shortest_path_2.py
similarity index 75%
rename from graphs/bfs_shortest_path.py
rename to graphs/breadth_first_search_shortest_path_2.py
index 1655ca64208d..4f9b6e65bdf3 100644
--- a/graphs/bfs_shortest_path.py
+++ b/graphs/breadth_first_search_shortest_path_2.py
@@ -1,12 +1,11 @@
"""Breadth-first search shortest path implementations.
-
- doctest:
- python -m doctest -v bfs_shortest_path.py
-
- Manual test:
- python bfs_shortest_path.py
+doctest:
+python -m doctest -v bfs_shortest_path.py
+Manual test:
+python bfs_shortest_path.py
"""
-graph = {
+
+demo_graph = {
"A": ["B", "C", "E"],
"B": ["A", "D", "E"],
"C": ["A", "F", "G"],
@@ -17,30 +16,31 @@
}
-def bfs_shortest_path(graph: dict, start, goal) -> str:
+def bfs_shortest_path(graph: dict, start, goal) -> list[str]:
"""Find shortest path between `start` and `goal` nodes.
-
Args:
graph (dict): node/list of neighboring nodes key/value pairs.
start: start node.
goal: target node.
-
Returns:
Shortest path between `start` and `goal` nodes as a string of nodes.
'Not found' string if no path found.
-
Example:
- >>> bfs_shortest_path(graph, "G", "D")
+ >>> bfs_shortest_path(demo_graph, "G", "D")
['G', 'C', 'A', 'B', 'D']
+ >>> bfs_shortest_path(demo_graph, "G", "G")
+ ['G']
+ >>> bfs_shortest_path(demo_graph, "G", "Unknown")
+ []
"""
# keep track of explored nodes
- explored = []
+ explored = set()
# keep track of all the paths to be checked
queue = [[start]]
# return path if start is goal
if start == goal:
- return "That was easy! Start = goal"
+ return [start]
# keeps looping until all possible paths have been checked
while queue:
@@ -61,30 +61,27 @@ def bfs_shortest_path(graph: dict, start, goal) -> str:
return new_path
# mark node as explored
- explored.append(node)
+ explored.add(node)
# in case there's no path between the 2 nodes
- return "So sorry, but a connecting path doesn't exist :("
+ return []
def bfs_shortest_path_distance(graph: dict, start, target) -> int:
"""Find shortest path distance between `start` and `target` nodes.
-
Args:
graph: node/list of neighboring nodes key/value pairs.
start: node to start search from.
target: node to search for.
-
Returns:
Number of edges in shortest path between `start` and `target` nodes.
-1 if no path exists.
-
Example:
- >>> bfs_shortest_path_distance(graph, "G", "D")
+ >>> bfs_shortest_path_distance(demo_graph, "G", "D")
4
- >>> bfs_shortest_path_distance(graph, "A", "A")
+ >>> bfs_shortest_path_distance(demo_graph, "A", "A")
0
- >>> bfs_shortest_path_distance(graph, "A", "H")
+ >>> bfs_shortest_path_distance(demo_graph, "A", "Unknown")
-1
"""
if not graph or start not in graph or target not in graph:
@@ -92,7 +89,7 @@ def bfs_shortest_path_distance(graph: dict, start, target) -> int:
if start == target:
return 0
queue = [start]
- visited = [start]
+ visited = set(start)
# Keep tab on distances from `start` node.
dist = {start: 0, target: -1}
while queue:
@@ -103,12 +100,12 @@ def bfs_shortest_path_distance(graph: dict, start, target) -> int:
)
for adjacent in graph[node]:
if adjacent not in visited:
- visited.append(adjacent)
+ visited.add(adjacent)
queue.append(adjacent)
dist[adjacent] = dist[node] + 1
return dist[target]
if __name__ == "__main__":
- print(bfs_shortest_path(graph, "G", "D")) # returns ['G', 'C', 'A', 'B', 'D']
- print(bfs_shortest_path_distance(graph, "G", "D")) # returns 4
+ print(bfs_shortest_path(demo_graph, "G", "D")) # returns ['G', 'C', 'A', 'B', 'D']
+ print(bfs_shortest_path_distance(demo_graph, "G", "D")) # returns 4
diff --git a/graphs/bfs_zero_one_shortest_path.py b/graphs/breadth_first_search_zero_one_shortest_path.py
similarity index 89%
rename from graphs/bfs_zero_one_shortest_path.py
rename to graphs/breadth_first_search_zero_one_shortest_path.py
index a725fae7e48f..d3a255bac1ef 100644
--- a/graphs/bfs_zero_one_shortest_path.py
+++ b/graphs/breadth_first_search_zero_one_shortest_path.py
@@ -1,13 +1,15 @@
-from collections import deque
-from dataclasses import dataclass
-from typing import Iterator, List
-
"""
Finding the shortest path in 0-1-graph in O(E + V) which is faster than dijkstra.
0-1-graph is the weighted graph with the weights equal to 0 or 1.
Link: https://codeforces.com/blog/entry/22276
"""
+from __future__ import annotations
+
+from collections import deque
+from collections.abc import Iterator
+from dataclasses import dataclass
+
@dataclass
class Edge:
@@ -21,7 +23,7 @@ class AdjacencyList:
"""Graph adjacency list."""
def __init__(self, size: int):
- self._graph: List[List[Edge]] = [[] for _ in range(size)]
+ self._graph: list[list[Edge]] = [[] for _ in range(size)]
self._size = size
def __getitem__(self, vertex: int) -> Iterator[Edge]:
@@ -58,7 +60,7 @@ def add_edge(self, from_vertex: int, to_vertex: int, weight: int):
self._graph[from_vertex].append(Edge(to_vertex, weight))
- def get_shortest_path(self, start_vertex: int, finish_vertex: int) -> int:
+ def get_shortest_path(self, start_vertex: int, finish_vertex: int) -> int | None:
"""
Return the shortest distance from start_vertex to finish_vertex in 0-1-graph.
1 1 1
@@ -106,18 +108,21 @@ def get_shortest_path(self, start_vertex: int, finish_vertex: int) -> int:
ValueError: No path from start_vertex to finish_vertex.
"""
queue = deque([start_vertex])
- distances = [None for i in range(self.size)]
+ distances: list[int | None] = [None] * self.size
distances[start_vertex] = 0
while queue:
current_vertex = queue.popleft()
current_distance = distances[current_vertex]
+ if current_distance is None:
+ continue
for edge in self[current_vertex]:
new_distance = current_distance + edge.weight
+ dest_vertex_distance = distances[edge.destination_vertex]
if (
- distances[edge.destination_vertex] is not None
- and new_distance >= distances[edge.destination_vertex]
+ isinstance(dest_vertex_distance, int)
+ and new_distance >= dest_vertex_distance
):
continue
distances[edge.destination_vertex] = new_distance
diff --git a/graphs/check_bipartite_graph_bfs.py b/graphs/check_bipartite_graph_bfs.py
deleted file mode 100644
index 00b771649b5d..000000000000
--- a/graphs/check_bipartite_graph_bfs.py
+++ /dev/null
@@ -1,45 +0,0 @@
-# Check whether Graph is Bipartite or Not using BFS
-
-
-# A Bipartite Graph is a graph whose vertices can be divided into two independent sets,
-# U and V such that every edge (u, v) either connects a vertex from U to V or a vertex
-# from V to U. In other words, for every edge (u, v), either u belongs to U and v to V,
-# or u belongs to V and v to U. We can also say that there is no edge that connects
-# vertices of same set.
-def checkBipartite(graph):
- queue = []
- visited = [False] * len(graph)
- color = [-1] * len(graph)
-
- def bfs():
- while queue:
- u = queue.pop(0)
- visited[u] = True
-
- for neighbour in graph[u]:
-
- if neighbour == u:
- return False
-
- if color[neighbour] == -1:
- color[neighbour] = 1 - color[u]
- queue.append(neighbour)
-
- elif color[neighbour] == color[u]:
- return False
-
- return True
-
- for i in range(len(graph)):
- if not visited[i]:
- queue.append(i)
- color[i] = 0
- if bfs() is False:
- return False
-
- return True
-
-
-if __name__ == "__main__":
- # Adjacency List of graph
- print(checkBipartite({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2]}))
diff --git a/graphs/check_bipartite_graph_dfs.py b/graphs/check_bipartite_graph_dfs.py
deleted file mode 100644
index fd644230449c..000000000000
--- a/graphs/check_bipartite_graph_dfs.py
+++ /dev/null
@@ -1,34 +0,0 @@
-# Check whether Graph is Bipartite or Not using DFS
-
-
-# A Bipartite Graph is a graph whose vertices can be divided into two independent sets,
-# U and V such that every edge (u, v) either connects a vertex from U to V or a vertex
-# from V to U. In other words, for every edge (u, v), either u belongs to U and v to V,
-# or u belongs to V and v to U. We can also say that there is no edge that connects
-# vertices of same set.
-def check_bipartite_dfs(graph):
- visited = [False] * len(graph)
- color = [-1] * len(graph)
-
- def dfs(v, c):
- visited[v] = True
- color[v] = c
- for u in graph[v]:
- if not visited[u]:
- dfs(u, 1 - c)
-
- for i in range(len(graph)):
- if not visited[i]:
- dfs(i, 0)
-
- for i in range(len(graph)):
- for j in graph[i]:
- if color[i] == color[j]:
- return False
-
- return True
-
-
-# Adjacency list of graph
-graph = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []}
-print(check_bipartite_dfs(graph))
diff --git a/graphs/check_bipatrite.py b/graphs/check_bipatrite.py
new file mode 100644
index 000000000000..213f3f9480b5
--- /dev/null
+++ b/graphs/check_bipatrite.py
@@ -0,0 +1,183 @@
+from collections import defaultdict, deque
+
+
+def is_bipartite_dfs(graph: defaultdict[int, list[int]]) -> bool:
+ """
+ Check if a graph is bipartite using depth-first search (DFS).
+
+ Args:
+ `graph`: Adjacency list representing the graph.
+
+ Returns:
+ ``True`` if bipartite, ``False`` otherwise.
+
+ Checks if the graph can be divided into two sets of vertices, such that no two
+ vertices within the same set are connected by an edge.
+
+ Examples:
+
+ >>> # FIXME: This test should pass.
+ >>> is_bipartite_dfs(defaultdict(list, {0: [1, 2], 1: [0, 3], 2: [0, 4]}))
+ Traceback (most recent call last):
+ ...
+ RuntimeError: dictionary changed size during iteration
+ >>> is_bipartite_dfs(defaultdict(list, {0: [1, 2], 1: [0, 3], 2: [0, 1]}))
+ False
+ >>> is_bipartite_dfs({})
+ True
+ >>> is_bipartite_dfs({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2]})
+ True
+ >>> is_bipartite_dfs({0: [1, 2, 3], 1: [0, 2], 2: [0, 1, 3], 3: [0, 2]})
+ False
+ >>> is_bipartite_dfs({0: [4], 1: [], 2: [4], 3: [4], 4: [0, 2, 3]})
+ True
+ >>> is_bipartite_dfs({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: [0]})
+ False
+ >>> is_bipartite_dfs({7: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: [0]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 0
+
+ >>> # FIXME: This test should fails with KeyError: 4.
+ >>> is_bipartite_dfs({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 9: [0]})
+ False
+ >>> is_bipartite_dfs({0: [-1, 3], 1: [0, -2]})
+ Traceback (most recent call last):
+ ...
+ KeyError: -1
+ >>> is_bipartite_dfs({-1: [0, 2], 0: [-1, 1], 1: [0, 2], 2: [-1, 1]})
+ True
+ >>> is_bipartite_dfs({0.9: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 0
+
+ >>> # FIXME: This test should fails with
+ >>> # TypeError: list indices must be integers or...
+ >>> is_bipartite_dfs({0: [1.0, 3.0], 1.0: [0, 2.0], 2.0: [1.0, 3.0], 3.0: [0, 2.0]})
+ True
+ >>> is_bipartite_dfs({"a": [1, 3], "b": [0, 2], "c": [1, 3], "d": [0, 2]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 1
+ >>> is_bipartite_dfs({0: ["b", "d"], 1: ["a", "c"], 2: ["b", "d"], 3: ["a", "c"]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 'b'
+ """
+
+ def depth_first_search(node: int, color: int) -> bool:
+ """
+ Perform Depth-First Search (DFS) on the graph starting from a node.
+
+ Args:
+ node: The current node being visited.
+ color: The color assigned to the current node.
+
+ Returns:
+ True if the graph is bipartite starting from the current node,
+ False otherwise.
+ """
+ if visited[node] == -1:
+ visited[node] = color
+ for neighbor in graph[node]:
+ if not depth_first_search(neighbor, 1 - color):
+ return False
+ return visited[node] == color
+
+ visited: defaultdict[int, int] = defaultdict(lambda: -1)
+ for node in graph:
+ if visited[node] == -1 and not depth_first_search(node, 0):
+ return False
+ return True
+
+
+def is_bipartite_bfs(graph: defaultdict[int, list[int]]) -> bool:
+ """
+ Check if a graph is bipartite using a breadth-first search (BFS).
+
+ Args:
+ `graph`: Adjacency list representing the graph.
+
+ Returns:
+ ``True`` if bipartite, ``False`` otherwise.
+
+ Check if the graph can be divided into two sets of vertices, such that no two
+ vertices within the same set are connected by an edge.
+
+ Examples:
+
+ >>> # FIXME: This test should pass.
+ >>> is_bipartite_bfs(defaultdict(list, {0: [1, 2], 1: [0, 3], 2: [0, 4]}))
+ Traceback (most recent call last):
+ ...
+ RuntimeError: dictionary changed size during iteration
+ >>> is_bipartite_bfs(defaultdict(list, {0: [1, 2], 1: [0, 2], 2: [0, 1]}))
+ False
+ >>> is_bipartite_bfs({})
+ True
+ >>> is_bipartite_bfs({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2]})
+ True
+ >>> is_bipartite_bfs({0: [1, 2, 3], 1: [0, 2], 2: [0, 1, 3], 3: [0, 2]})
+ False
+ >>> is_bipartite_bfs({0: [4], 1: [], 2: [4], 3: [4], 4: [0, 2, 3]})
+ True
+ >>> is_bipartite_bfs({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: [0]})
+ False
+ >>> is_bipartite_bfs({7: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: [0]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 0
+
+ >>> # FIXME: This test should fails with KeyError: 4.
+ >>> is_bipartite_bfs({0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 9: [0]})
+ False
+ >>> is_bipartite_bfs({0: [-1, 3], 1: [0, -2]})
+ Traceback (most recent call last):
+ ...
+ KeyError: -1
+ >>> is_bipartite_bfs({-1: [0, 2], 0: [-1, 1], 1: [0, 2], 2: [-1, 1]})
+ True
+ >>> is_bipartite_bfs({0.9: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 0
+
+ >>> # FIXME: This test should fails with
+ >>> # TypeError: list indices must be integers or...
+ >>> is_bipartite_bfs({0: [1.0, 3.0], 1.0: [0, 2.0], 2.0: [1.0, 3.0], 3.0: [0, 2.0]})
+ True
+ >>> is_bipartite_bfs({"a": [1, 3], "b": [0, 2], "c": [1, 3], "d": [0, 2]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 1
+ >>> is_bipartite_bfs({0: ["b", "d"], 1: ["a", "c"], 2: ["b", "d"], 3: ["a", "c"]})
+ Traceback (most recent call last):
+ ...
+ KeyError: 'b'
+ """
+ visited: defaultdict[int, int] = defaultdict(lambda: -1)
+ for node in graph:
+ if visited[node] == -1:
+ queue: deque[int] = deque()
+ queue.append(node)
+ visited[node] = 0
+ while queue:
+ curr_node = queue.popleft()
+ for neighbor in graph[curr_node]:
+ if visited[neighbor] == -1:
+ visited[neighbor] = 1 - visited[curr_node]
+ queue.append(neighbor)
+ elif visited[neighbor] == visited[curr_node]:
+ return False
+ return True
+
+
+if __name__ == "__main":
+ import doctest
+
+ result = doctest.testmod()
+ if result.failed:
+ print(f"{result.failed} test(s) failed.")
+ else:
+ print("All tests passed!")
diff --git a/graphs/check_cycle.py b/graphs/check_cycle.py
new file mode 100644
index 000000000000..9fd1cd80f116
--- /dev/null
+++ b/graphs/check_cycle.py
@@ -0,0 +1,52 @@
+"""
+Program to check if a cycle is present in a given graph
+"""
+
+
+def check_cycle(graph: dict) -> bool:
+ """
+ Returns True if graph is cyclic else False
+ >>> check_cycle(graph={0:[], 1:[0, 3], 2:[0, 4], 3:[5], 4:[5], 5:[]})
+ False
+ >>> check_cycle(graph={0:[1, 2], 1:[2], 2:[0, 3], 3:[3]})
+ True
+ """
+ # Keep track of visited nodes
+ visited: set[int] = set()
+ # To detect a back edge, keep track of vertices currently in the recursion stack
+ rec_stk: set[int] = set()
+ return any(
+ node not in visited and depth_first_search(graph, node, visited, rec_stk)
+ for node in graph
+ )
+
+
+def depth_first_search(graph: dict, vertex: int, visited: set, rec_stk: set) -> bool:
+ """
+ Recur for all neighbours.
+ If any neighbour is visited and in rec_stk then graph is cyclic.
+ >>> graph = {0:[], 1:[0, 3], 2:[0, 4], 3:[5], 4:[5], 5:[]}
+ >>> vertex, visited, rec_stk = 0, set(), set()
+ >>> depth_first_search(graph, vertex, visited, rec_stk)
+ False
+ """
+ # Mark current node as visited and add to recursion stack
+ visited.add(vertex)
+ rec_stk.add(vertex)
+
+ for node in graph[vertex]:
+ if node not in visited:
+ if depth_first_search(graph, node, visited, rec_stk):
+ return True
+ elif node in rec_stk:
+ return True
+
+ # The node needs to be removed from recursion stack before function ends
+ rec_stk.remove(vertex)
+ return False
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/graphs/connected_components.py b/graphs/connected_components.py
index 4af7803d74a7..15c7633e13e8 100644
--- a/graphs/connected_components.py
+++ b/graphs/connected_components.py
@@ -27,7 +27,7 @@ def dfs(graph: dict, vert: int, visited: list) -> list:
if not visited[neighbour]:
connected_verts += dfs(graph, neighbour, visited)
- return [vert] + connected_verts
+ return [vert, *connected_verts]
def connected_components(graph: dict) -> list:
diff --git a/graphs/deep_clone_graph.py b/graphs/deep_clone_graph.py
new file mode 100644
index 000000000000..18ea99c6a52d
--- /dev/null
+++ b/graphs/deep_clone_graph.py
@@ -0,0 +1,78 @@
+"""
+LeetCode 133. Clone Graph
+https://leetcode.com/problems/clone-graph/
+
+Given a reference of a node in a connected undirected graph.
+
+Return a deep copy (clone) of the graph.
+
+Each node in the graph contains a value (int) and a list (List[Node]) of its
+neighbors.
+"""
+
+from dataclasses import dataclass
+
+
+@dataclass
+class Node:
+ value: int = 0
+ neighbors: list["Node"] | None = None
+
+ def __post_init__(self) -> None:
+ """
+ >>> Node(3).neighbors
+ []
+ """
+ self.neighbors = self.neighbors or []
+
+ def __hash__(self) -> int:
+ """
+ >>> hash(Node(3)) != 0
+ True
+ """
+ return id(self)
+
+
+def clone_graph(node: Node | None) -> Node | None:
+ """
+ This function returns a clone of a connected undirected graph.
+ >>> clone_graph(Node(1))
+ Node(value=1, neighbors=[])
+ >>> clone_graph(Node(1, [Node(2)]))
+ Node(value=1, neighbors=[Node(value=2, neighbors=[])])
+ >>> clone_graph(None) is None
+ True
+ """
+ if not node:
+ return None
+
+ originals_to_clones = {} # map nodes to clones
+
+ stack = [node]
+
+ while stack:
+ original = stack.pop()
+
+ if original in originals_to_clones:
+ continue
+
+ originals_to_clones[original] = Node(original.value)
+
+ stack.extend(original.neighbors or [])
+
+ for original, clone in originals_to_clones.items():
+ for neighbor in original.neighbors or []:
+ cloned_neighbor = originals_to_clones[neighbor]
+
+ if not clone.neighbors:
+ clone.neighbors = []
+
+ clone.neighbors.append(cloned_neighbor)
+
+ return originals_to_clones[node]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/depth_first_search.py b/graphs/depth_first_search.py
index 907cc172f253..a666e74ce607 100644
--- a/graphs/depth_first_search.py
+++ b/graphs/depth_first_search.py
@@ -3,19 +3,18 @@
from __future__ import annotations
-def depth_first_search(graph: dict, start: str) -> set[int]:
+def depth_first_search(graph: dict, start: str) -> set[str]:
"""Depth First Search on Graph
:param graph: directed graph in dictionary format
- :param vertex: starting vertex as a string
+ :param start: starting vertex as a string
:returns: the trace of the search
- >>> G = { "A": ["B", "C", "D"], "B": ["A", "D", "E"],
+ >>> input_G = { "A": ["B", "C", "D"], "B": ["A", "D", "E"],
... "C": ["A", "F"], "D": ["B", "D"], "E": ["B", "F"],
... "F": ["C", "E", "G"], "G": ["F"] }
- >>> start = "A"
>>> output_G = list({'A', 'B', 'C', 'D', 'E', 'F', 'G'})
- >>> all(x in output_G for x in list(depth_first_search(G, "A")))
+ >>> all(x in output_G for x in list(depth_first_search(input_G, "A")))
True
- >>> all(x in output_G for x in list(depth_first_search(G, "G")))
+ >>> all(x in output_G for x in list(depth_first_search(input_G, "G")))
True
"""
explored, stack = set(start), [start]
diff --git a/graphs/depth_first_search_2.py b/graphs/depth_first_search_2.py
index 3072d527c1c7..8fe48b7f2b42 100644
--- a/graphs/depth_first_search_2.py
+++ b/graphs/depth_first_search_2.py
@@ -1,6 +1,6 @@
#!/usr/bin/python
-""" Author: OMKAR PATHAK """
+"""Author: OMKAR PATHAK"""
class Graph:
@@ -9,12 +9,44 @@ def __init__(self):
# for printing the Graph vertices
def print_graph(self) -> None:
+ """
+ Print the graph vertices.
+
+ Example:
+ >>> g = Graph()
+ >>> g.add_edge(0, 1)
+ >>> g.add_edge(0, 2)
+ >>> g.add_edge(1, 2)
+ >>> g.add_edge(2, 0)
+ >>> g.add_edge(2, 3)
+ >>> g.add_edge(3, 3)
+ >>> g.print_graph()
+ {0: [1, 2], 1: [2], 2: [0, 3], 3: [3]}
+ 0 -> 1 -> 2
+ 1 -> 2
+ 2 -> 0 -> 3
+ 3 -> 3
+ """
print(self.vertex)
for i in self.vertex:
print(i, " -> ", " -> ".join([str(j) for j in self.vertex[i]]))
# for adding the edge between two vertices
def add_edge(self, from_vertex: int, to_vertex: int) -> None:
+ """
+ Add an edge between two vertices.
+
+ :param from_vertex: The source vertex.
+ :param to_vertex: The destination vertex.
+
+ Example:
+ >>> g = Graph()
+ >>> g.add_edge(0, 1)
+ >>> g.add_edge(0, 2)
+ >>> g.print_graph()
+ {0: [1, 2]}
+ 0 -> 1 -> 2
+ """
# check if vertex is already present,
if from_vertex in self.vertex:
self.vertex[from_vertex].append(to_vertex)
@@ -23,6 +55,21 @@ def add_edge(self, from_vertex: int, to_vertex: int) -> None:
self.vertex[from_vertex] = [to_vertex]
def dfs(self) -> None:
+ """
+ Perform depth-first search (DFS) traversal on the graph
+ and print the visited vertices.
+
+ Example:
+ >>> g = Graph()
+ >>> g.add_edge(0, 1)
+ >>> g.add_edge(0, 2)
+ >>> g.add_edge(1, 2)
+ >>> g.add_edge(2, 0)
+ >>> g.add_edge(2, 3)
+ >>> g.add_edge(3, 3)
+ >>> g.dfs()
+ 0 1 2 3
+ """
# visited array for storing already visited nodes
visited = [False] * len(self.vertex)
@@ -32,18 +79,41 @@ def dfs(self) -> None:
self.dfs_recursive(i, visited)
def dfs_recursive(self, start_vertex: int, visited: list) -> None:
+ """
+ Perform a recursive depth-first search (DFS) traversal on the graph.
+
+ :param start_vertex: The starting vertex for the traversal.
+ :param visited: A list to track visited vertices.
+
+ Example:
+ >>> g = Graph()
+ >>> g.add_edge(0, 1)
+ >>> g.add_edge(0, 2)
+ >>> g.add_edge(1, 2)
+ >>> g.add_edge(2, 0)
+ >>> g.add_edge(2, 3)
+ >>> g.add_edge(3, 3)
+ >>> visited = [False] * len(g.vertex)
+ >>> g.dfs_recursive(0, visited)
+ 0 1 2 3
+ """
# mark start vertex as visited
visited[start_vertex] = True
- print(start_vertex, end=" ")
+ print(start_vertex, end="")
# Recur for all the vertices that are adjacent to this node
for i in self.vertex:
if not visited[i]:
+ print(" ", end="")
self.dfs_recursive(i, visited)
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
g = Graph()
g.add_edge(0, 1)
g.add_edge(0, 2)
@@ -55,11 +125,3 @@ def dfs_recursive(self, start_vertex: int, visited: list) -> None:
g.print_graph()
print("DFS:")
g.dfs()
-
- # OUTPUT:
- # 0 -> 1 -> 2
- # 1 -> 2
- # 2 -> 0 -> 3
- # 3 -> 3
- # DFS:
- # 0 1 2 3
diff --git a/graphs/dijkstra.py b/graphs/dijkstra.py
index d15fcbbfeef0..87e9d2233bb2 100644
--- a/graphs/dijkstra.py
+++ b/graphs/dijkstra.py
@@ -30,6 +30,7 @@
distance between each vertex that makes up the path from start vertex to target
vertex.
"""
+
import heapq
@@ -56,8 +57,8 @@ def dijkstra(graph, start, end):
for v, c in graph[u]:
if v in visited:
continue
- next = cost + c
- heapq.heappush(heap, (next, v))
+ next_item = cost + c
+ heapq.heappush(heap, (next_item, v))
return -1
@@ -103,14 +104,14 @@ def dijkstra(graph, start, end):
"G": [["F", 1]],
}
-shortDistance = dijkstra(G, "E", "C")
-print(shortDistance) # E -- 3 --> F -- 3 --> C == 6
+short_distance = dijkstra(G, "E", "C")
+print(short_distance) # E -- 3 --> F -- 3 --> C == 6
-shortDistance = dijkstra(G2, "E", "F")
-print(shortDistance) # E -- 3 --> F == 3
+short_distance = dijkstra(G2, "E", "F")
+print(short_distance) # E -- 3 --> F == 3
-shortDistance = dijkstra(G3, "E", "F")
-print(shortDistance) # E -- 2 --> G -- 1 --> F == 3
+short_distance = dijkstra(G3, "E", "F")
+print(short_distance) # E -- 2 --> G -- 1 --> F == 3
if __name__ == "__main__":
import doctest
diff --git a/graphs/dijkstra_2.py b/graphs/dijkstra_2.py
index 762884136e4a..f548463ff7bd 100644
--- a/graphs/dijkstra_2.py
+++ b/graphs/dijkstra_2.py
@@ -1,6 +1,6 @@
-def printDist(dist, V):
+def print_dist(dist, v):
print("\nVertex Distance")
- for i in range(V):
+ for i in range(v):
if dist[i] != float("inf"):
print(i, "\t", int(dist[i]), end="\t")
else:
@@ -8,34 +8,34 @@ def printDist(dist, V):
print()
-def minDist(mdist, vset, V):
- minVal = float("inf")
- minInd = -1
- for i in range(V):
- if (not vset[i]) and mdist[i] < minVal:
- minInd = i
- minVal = mdist[i]
- return minInd
+def min_dist(mdist, vset, v):
+ min_val = float("inf")
+ min_ind = -1
+ for i in range(v):
+ if (not vset[i]) and mdist[i] < min_val:
+ min_ind = i
+ min_val = mdist[i]
+ return min_ind
-def Dijkstra(graph, V, src):
- mdist = [float("inf") for i in range(V)]
- vset = [False for i in range(V)]
+def dijkstra(graph, v, src):
+ mdist = [float("inf") for _ in range(v)]
+ vset = [False for _ in range(v)]
mdist[src] = 0.0
- for i in range(V - 1):
- u = minDist(mdist, vset, V)
+ for _ in range(v - 1):
+ u = min_dist(mdist, vset, v)
vset[u] = True
- for v in range(V):
+ for i in range(v):
if (
- (not vset[v])
- and graph[u][v] != float("inf")
- and mdist[u] + graph[u][v] < mdist[v]
+ (not vset[i])
+ and graph[u][i] != float("inf")
+ and mdist[u] + graph[u][i] < mdist[i]
):
- mdist[v] = mdist[u] + graph[u][v]
+ mdist[i] = mdist[u] + graph[u][i]
- printDist(mdist, V)
+ print_dist(mdist, i)
if __name__ == "__main__":
@@ -55,4 +55,4 @@ def Dijkstra(graph, V, src):
graph[src][dst] = weight
gsrc = int(input("\nEnter shortest path source:").strip())
- Dijkstra(graph, V, gsrc)
+ dijkstra(graph, V, gsrc)
diff --git a/graphs/dijkstra_algorithm.py b/graphs/dijkstra_algorithm.py
index 6b64834acd81..51412b790bac 100644
--- a/graphs/dijkstra_algorithm.py
+++ b/graphs/dijkstra_algorithm.py
@@ -11,35 +11,127 @@
class PriorityQueue:
# Based on Min Heap
def __init__(self):
+ """
+ Priority queue class constructor method.
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.cur_size
+ 0
+ >>> priority_queue_test.array
+ []
+ >>> priority_queue_test.pos
+ {}
+ """
self.cur_size = 0
self.array = []
self.pos = {} # To store the pos of node in array
- def isEmpty(self):
+ def is_empty(self):
+ """
+ Conditional boolean method to determine if the priority queue is empty or not.
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.is_empty()
+ True
+ >>> priority_queue_test.insert((2, 'A'))
+ >>> priority_queue_test.is_empty()
+ False
+ """
return self.cur_size == 0
def min_heapify(self, idx):
+ """
+ Sorts the queue array so that the minimum element is root.
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.cur_size = 3
+ >>> priority_queue_test.pos = {'A': 0, 'B': 1, 'C': 2}
+
+ >>> priority_queue_test.array = [(5, 'A'), (10, 'B'), (15, 'C')]
+ >>> priority_queue_test.min_heapify(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'list' object is not callable
+ >>> priority_queue_test.array
+ [(5, 'A'), (10, 'B'), (15, 'C')]
+
+ >>> priority_queue_test.array = [(10, 'A'), (5, 'B'), (15, 'C')]
+ >>> priority_queue_test.min_heapify(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'list' object is not callable
+ >>> priority_queue_test.array
+ [(10, 'A'), (5, 'B'), (15, 'C')]
+
+ >>> priority_queue_test.array = [(10, 'A'), (15, 'B'), (5, 'C')]
+ >>> priority_queue_test.min_heapify(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'list' object is not callable
+ >>> priority_queue_test.array
+ [(10, 'A'), (15, 'B'), (5, 'C')]
+
+ >>> priority_queue_test.array = [(10, 'A'), (5, 'B')]
+ >>> priority_queue_test.cur_size = len(priority_queue_test.array)
+ >>> priority_queue_test.pos = {'A': 0, 'B': 1}
+ >>> priority_queue_test.min_heapify(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'list' object is not callable
+ >>> priority_queue_test.array
+ [(10, 'A'), (5, 'B')]
+ """
lc = self.left(idx)
rc = self.right(idx)
- if lc < self.cur_size and self.array(lc)[0] < self.array(idx)[0]:
+ if lc < self.cur_size and self.array(lc)[0] < self.array[idx][0]:
smallest = lc
else:
smallest = idx
- if rc < self.cur_size and self.array(rc)[0] < self.array(smallest)[0]:
+ if rc < self.cur_size and self.array(rc)[0] < self.array[smallest][0]:
smallest = rc
if smallest != idx:
self.swap(idx, smallest)
self.min_heapify(smallest)
def insert(self, tup):
- # Inserts a node into the Priority Queue
+ """
+ Inserts a node into the Priority Queue.
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.insert((10, 'A'))
+ >>> priority_queue_test.array
+ [(10, 'A')]
+ >>> priority_queue_test.insert((15, 'B'))
+ >>> priority_queue_test.array
+ [(10, 'A'), (15, 'B')]
+ >>> priority_queue_test.insert((5, 'C'))
+ >>> priority_queue_test.array
+ [(5, 'C'), (10, 'A'), (15, 'B')]
+ """
self.pos[tup[1]] = self.cur_size
self.cur_size += 1
self.array.append((sys.maxsize, tup[1]))
self.decrease_key((sys.maxsize, tup[1]), tup[0])
def extract_min(self):
- # Removes and returns the min element at top of priority queue
+ """
+ Removes and returns the min element at top of priority queue.
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.array = [(10, 'A'), (15, 'B')]
+ >>> priority_queue_test.cur_size = len(priority_queue_test.array)
+ >>> priority_queue_test.pos = {'A': 0, 'B': 1}
+ >>> priority_queue_test.insert((5, 'C'))
+ >>> priority_queue_test.extract_min()
+ 'C'
+ >>> priority_queue_test.array[0]
+ (15, 'B')
+ """
min_node = self.array[0][1]
self.array[0] = self.array[self.cur_size - 1]
self.cur_size -= 1
@@ -48,20 +140,61 @@ def extract_min(self):
return min_node
def left(self, i):
- # returns the index of left child
+ """
+ Returns the index of left child
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.left(0)
+ 1
+ >>> priority_queue_test.left(1)
+ 3
+ """
return 2 * i + 1
def right(self, i):
- # returns the index of right child
+ """
+ Returns the index of right child
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.right(0)
+ 2
+ >>> priority_queue_test.right(1)
+ 4
+ """
return 2 * i + 2
def par(self, i):
- # returns the index of parent
+ """
+ Returns the index of parent
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.par(1)
+ 0
+ >>> priority_queue_test.par(2)
+ 1
+ >>> priority_queue_test.par(4)
+ 2
+ """
return math.floor(i / 2)
def swap(self, i, j):
- # swaps array elements at indices i and j
- # update the pos{}
+ """
+ Swaps array elements at indices i and j, update the pos{}
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.array = [(10, 'A'), (15, 'B')]
+ >>> priority_queue_test.cur_size = len(priority_queue_test.array)
+ >>> priority_queue_test.pos = {'A': 0, 'B': 1}
+ >>> priority_queue_test.swap(0, 1)
+ >>> priority_queue_test.array
+ [(15, 'B'), (10, 'A')]
+ >>> priority_queue_test.pos
+ {'A': 1, 'B': 0}
+ """
self.pos[self.array[i][1]] = j
self.pos[self.array[j][1]] = i
temp = self.array[i]
@@ -69,8 +202,20 @@ def swap(self, i, j):
self.array[j] = temp
def decrease_key(self, tup, new_d):
+ """
+ Decrease the key value for a given tuple, assuming the new_d is at most old_d.
+
+ Examples:
+ >>> priority_queue_test = PriorityQueue()
+ >>> priority_queue_test.array = [(10, 'A'), (15, 'B')]
+ >>> priority_queue_test.cur_size = len(priority_queue_test.array)
+ >>> priority_queue_test.pos = {'A': 0, 'B': 1}
+ >>> priority_queue_test.decrease_key((10, 'A'), 5)
+ >>> priority_queue_test.array
+ [(5, 'A'), (15, 'B')]
+ """
idx = self.pos[tup[1]]
- # assuming the new_d is atmost old_d
+ # assuming the new_d is at most old_d
self.array[idx] = (new_d, tup[1])
while idx > 0 and self.array[self.par(idx)][0] > self.array[idx][0]:
self.swap(idx, self.par(idx))
@@ -79,6 +224,20 @@ def decrease_key(self, tup, new_d):
class Graph:
def __init__(self, num):
+ """
+ Graph class constructor
+
+ Examples:
+ >>> graph_test = Graph(1)
+ >>> graph_test.num_nodes
+ 1
+ >>> graph_test.dist
+ [0]
+ >>> graph_test.par
+ [-1]
+ >>> graph_test.adjList
+ {}
+ """
self.adjList = {} # To store graph: u -> (v,w)
self.num_nodes = num # Number of nodes in graph
# To store the distance from source vertex
@@ -86,48 +245,144 @@ def __init__(self, num):
self.par = [-1] * self.num_nodes # To store the path
def add_edge(self, u, v, w):
- # Edge going from node u to v and v to u with weight w
- # u (w)-> v, v (w) -> u
+ """
+ Add edge going from node u to v and v to u with weight w: u (w)-> v, v (w) -> u
+
+ Examples:
+ >>> graph_test = Graph(1)
+ >>> graph_test.add_edge(1, 2, 1)
+ >>> graph_test.add_edge(2, 3, 2)
+ >>> graph_test.adjList
+ {1: [(2, 1)], 2: [(1, 1), (3, 2)], 3: [(2, 2)]}
+ """
# Check if u already in graph
- if u in self.adjList.keys():
+ if u in self.adjList:
self.adjList[u].append((v, w))
else:
self.adjList[u] = [(v, w)]
# Assuming undirected graph
- if v in self.adjList.keys():
+ if v in self.adjList:
self.adjList[v].append((u, w))
else:
self.adjList[v] = [(u, w)]
def show_graph(self):
- # u -> v(w)
+ """
+ Show the graph: u -> v(w)
+
+ Examples:
+ >>> graph_test = Graph(1)
+ >>> graph_test.add_edge(1, 2, 1)
+ >>> graph_test.show_graph()
+ 1 -> 2(1)
+ 2 -> 1(1)
+ >>> graph_test.add_edge(2, 3, 2)
+ >>> graph_test.show_graph()
+ 1 -> 2(1)
+ 2 -> 1(1) -> 3(2)
+ 3 -> 2(2)
+ """
for u in self.adjList:
print(u, "->", " -> ".join(str(f"{v}({w})") for v, w in self.adjList[u]))
def dijkstra(self, src):
+ """
+ Dijkstra algorithm
+
+ Examples:
+ >>> graph_test = Graph(3)
+ >>> graph_test.add_edge(0, 1, 2)
+ >>> graph_test.add_edge(1, 2, 2)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 2
+ Node 2 has distance: 4
+ >>> graph_test.dist
+ [0, 2, 4]
+
+ >>> graph_test = Graph(2)
+ >>> graph_test.add_edge(0, 1, 2)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 2
+ >>> graph_test.dist
+ [0, 2]
+
+ >>> graph_test = Graph(3)
+ >>> graph_test.add_edge(0, 1, 2)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 2
+ Node 2 has distance: 0
+ >>> graph_test.dist
+ [0, 2, 0]
+
+ >>> graph_test = Graph(3)
+ >>> graph_test.add_edge(0, 1, 2)
+ >>> graph_test.add_edge(1, 2, 2)
+ >>> graph_test.add_edge(0, 2, 1)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 2
+ Node 2 has distance: 1
+ >>> graph_test.dist
+ [0, 2, 1]
+
+ >>> graph_test = Graph(4)
+ >>> graph_test.add_edge(0, 1, 4)
+ >>> graph_test.add_edge(1, 2, 2)
+ >>> graph_test.add_edge(2, 3, 1)
+ >>> graph_test.add_edge(0, 2, 3)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 4
+ Node 2 has distance: 3
+ Node 3 has distance: 4
+ >>> graph_test.dist
+ [0, 4, 3, 4]
+
+ >>> graph_test = Graph(4)
+ >>> graph_test.add_edge(0, 1, 4)
+ >>> graph_test.add_edge(1, 2, 2)
+ >>> graph_test.add_edge(2, 3, 1)
+ >>> graph_test.add_edge(0, 2, 7)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 4
+ Node 2 has distance: 6
+ Node 3 has distance: 7
+ >>> graph_test.dist
+ [0, 4, 6, 7]
+ """
# Flush old junk values in par[]
self.par = [-1] * self.num_nodes
# src is the source node
self.dist[src] = 0
- Q = PriorityQueue()
- Q.insert((0, src)) # (dist from src, node)
- for u in self.adjList.keys():
+ q = PriorityQueue()
+ q.insert((0, src)) # (dist from src, node)
+ for u in self.adjList:
if u != src:
self.dist[u] = sys.maxsize # Infinity
self.par[u] = -1
- while not Q.isEmpty():
- u = Q.extract_min() # Returns node with the min dist from source
+ while not q.is_empty():
+ u = q.extract_min() # Returns node with the min dist from source
# Update the distance of all the neighbours of u and
# if their prev dist was INFINITY then push them in Q
for v, w in self.adjList[u]:
new_dist = self.dist[u] + w
if self.dist[v] > new_dist:
if self.dist[v] == sys.maxsize:
- Q.insert((new_dist, v))
+ q.insert((new_dist, v))
else:
- Q.decrease_key((self.dist[v], v), new_dist)
+ q.decrease_key((self.dist[v], v), new_dist)
self.dist[v] = new_dist
self.par[v] = u
@@ -135,13 +390,40 @@ def dijkstra(self, src):
self.show_distances(src)
def show_distances(self, src):
+ """
+ Show the distances from src to all other nodes in a graph
+
+ Examples:
+ >>> graph_test = Graph(1)
+ >>> graph_test.show_distances(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ """
print(f"Distance from node: {src}")
for u in range(self.num_nodes):
print(f"Node {u} has distance: {self.dist[u]}")
def show_path(self, src, dest):
- # To show the shortest path from src to dest
- # WARNING: Use it *after* calling dijkstra
+ """
+ Shows the shortest path from src to dest.
+ WARNING: Use it *after* calling dijkstra.
+
+ Examples:
+ >>> graph_test = Graph(4)
+ >>> graph_test.add_edge(0, 1, 1)
+ >>> graph_test.add_edge(1, 2, 2)
+ >>> graph_test.add_edge(2, 3, 3)
+ >>> graph_test.dijkstra(0)
+ Distance from node: 0
+ Node 0 has distance: 0
+ Node 1 has distance: 1
+ Node 2 has distance: 3
+ Node 3 has distance: 6
+ >>> graph_test.show_path(0, 3) # doctest: +NORMALIZE_WHITESPACE
+ ----Path to reach 3 from 0----
+ 0 -> 1 -> 2 -> 3
+ Total cost of path: 6
+ """
path = []
cost = 0
temp = dest
@@ -167,6 +449,9 @@ def show_path(self, src, dest):
if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
graph = Graph(9)
graph.add_edge(0, 1, 4)
graph.add_edge(0, 7, 8)
diff --git a/graphs/dijkstra_alternate.py b/graphs/dijkstra_alternate.py
new file mode 100644
index 000000000000..7beef6b04da1
--- /dev/null
+++ b/graphs/dijkstra_alternate.py
@@ -0,0 +1,98 @@
+from __future__ import annotations
+
+
+class Graph:
+ def __init__(self, vertices: int) -> None:
+ """
+ >>> graph = Graph(2)
+ >>> graph.vertices
+ 2
+ >>> len(graph.graph)
+ 2
+ >>> len(graph.graph[0])
+ 2
+ """
+ self.vertices = vertices
+ self.graph = [[0] * vertices for _ in range(vertices)]
+
+ def print_solution(self, distances_from_source: list[int]) -> None:
+ """
+ >>> Graph(0).print_solution([]) # doctest: +NORMALIZE_WHITESPACE
+ Vertex Distance from Source
+ """
+ print("Vertex \t Distance from Source")
+ for vertex in range(self.vertices):
+ print(vertex, "\t\t", distances_from_source[vertex])
+
+ def minimum_distance(
+ self, distances_from_source: list[int], visited: list[bool]
+ ) -> int:
+ """
+ A utility function to find the vertex with minimum distance value, from the set
+ of vertices not yet included in shortest path tree.
+
+ >>> Graph(3).minimum_distance([1, 2, 3], [False, False, True])
+ 0
+ """
+
+ # Initialize minimum distance for next node
+ minimum = 1e7
+ min_index = 0
+
+ # Search not nearest vertex not in the shortest path tree
+ for vertex in range(self.vertices):
+ if distances_from_source[vertex] < minimum and visited[vertex] is False:
+ minimum = distances_from_source[vertex]
+ min_index = vertex
+ return min_index
+
+ def dijkstra(self, source: int) -> None:
+ """
+ Function that implements Dijkstra's single source shortest path algorithm for a
+ graph represented using adjacency matrix representation.
+
+ >>> Graph(4).dijkstra(1) # doctest: +NORMALIZE_WHITESPACE
+ Vertex Distance from Source
+ 0 10000000
+ 1 0
+ 2 10000000
+ 3 10000000
+ """
+
+ distances = [int(1e7)] * self.vertices # distances from the source
+ distances[source] = 0
+ visited = [False] * self.vertices
+
+ for _ in range(self.vertices):
+ u = self.minimum_distance(distances, visited)
+ visited[u] = True
+
+ # Update dist value of the adjacent vertices
+ # of the picked vertex only if the current
+ # distance is greater than new distance and
+ # the vertex in not in the shortest path tree
+ for v in range(self.vertices):
+ if (
+ self.graph[u][v] > 0
+ and visited[v] is False
+ and distances[v] > distances[u] + self.graph[u][v]
+ ):
+ distances[v] = distances[u] + self.graph[u][v]
+
+ self.print_solution(distances)
+
+
+if __name__ == "__main__":
+ graph = Graph(9)
+ graph.graph = [
+ [0, 4, 0, 0, 0, 0, 0, 8, 0],
+ [4, 0, 8, 0, 0, 0, 0, 11, 0],
+ [0, 8, 0, 7, 0, 4, 0, 0, 2],
+ [0, 0, 7, 0, 9, 14, 0, 0, 0],
+ [0, 0, 0, 9, 0, 10, 0, 0, 0],
+ [0, 0, 4, 14, 10, 0, 2, 0, 0],
+ [0, 0, 0, 0, 0, 2, 0, 1, 6],
+ [8, 11, 0, 0, 0, 0, 1, 0, 7],
+ [0, 0, 2, 0, 0, 0, 6, 7, 0],
+ ]
+ graph.dijkstra(0)
diff --git a/graphs/dijkstra_binary_grid.py b/graphs/dijkstra_binary_grid.py
new file mode 100644
index 000000000000..06293a87da2d
--- /dev/null
+++ b/graphs/dijkstra_binary_grid.py
@@ -0,0 +1,89 @@
+"""
+This script implements the Dijkstra algorithm on a binary grid.
+The grid consists of 0s and 1s, where 1 represents
+a walkable node and 0 represents an obstacle.
+The algorithm finds the shortest path from a start node to a destination node.
+Diagonal movement can be allowed or disallowed.
+"""
+
+from heapq import heappop, heappush
+
+import numpy as np
+
+
+def dijkstra(
+ grid: np.ndarray,
+ source: tuple[int, int],
+ destination: tuple[int, int],
+ allow_diagonal: bool,
+) -> tuple[float | int, list[tuple[int, int]]]:
+ """
+ Implements Dijkstra's algorithm on a binary grid.
+
+ Args:
+ grid (np.ndarray): A 2D numpy array representing the grid.
+ 1 represents a walkable node and 0 represents an obstacle.
+ source (Tuple[int, int]): A tuple representing the start node.
+ destination (Tuple[int, int]): A tuple representing the
+ destination node.
+ allow_diagonal (bool): A boolean determining whether
+ diagonal movements are allowed.
+
+ Returns:
+ Tuple[Union[float, int], List[Tuple[int, int]]]:
+ The shortest distance from the start node to the destination node
+ and the shortest path as a list of nodes.
+
+ >>> dijkstra(np.array([[1, 1, 1], [0, 1, 0], [0, 1, 1]]), (0, 0), (2, 2), False)
+ (4.0, [(0, 0), (0, 1), (1, 1), (2, 1), (2, 2)])
+
+ >>> dijkstra(np.array([[1, 1, 1], [0, 1, 0], [0, 1, 1]]), (0, 0), (2, 2), True)
+ (2.0, [(0, 0), (1, 1), (2, 2)])
+
+ >>> dijkstra(np.array([[1, 1, 1], [0, 0, 1], [0, 1, 1]]), (0, 0), (2, 2), False)
+ (4.0, [(0, 0), (0, 1), (0, 2), (1, 2), (2, 2)])
+ """
+ rows, cols = grid.shape
+ dx = [-1, 1, 0, 0]
+ dy = [0, 0, -1, 1]
+ if allow_diagonal:
+ dx += [-1, -1, 1, 1]
+ dy += [-1, 1, -1, 1]
+
+ queue, visited = [(0, source)], set()
+ matrix = np.full((rows, cols), np.inf)
+ matrix[source] = 0
+ predecessors = np.empty((rows, cols), dtype=object)
+ predecessors[source] = None
+
+ while queue:
+ (dist, (x, y)) = heappop(queue)
+ if (x, y) in visited:
+ continue
+ visited.add((x, y))
+
+ if (x, y) == destination:
+ path = []
+ while (x, y) != source:
+ path.append((x, y))
+ x, y = predecessors[x, y]
+ path.append(source) # add the source manually
+ path.reverse()
+ return float(matrix[destination]), path
+
+ for i in range(len(dx)):
+ nx, ny = x + dx[i], y + dy[i]
+ if 0 <= nx < rows and 0 <= ny < cols:
+ next_node = grid[nx][ny]
+ if next_node == 1 and matrix[nx, ny] > dist + 1:
+ heappush(queue, (dist + 1, (nx, ny)))
+ matrix[nx, ny] = dist + 1
+ predecessors[nx, ny] = (x, y)
+
+ return np.inf, []
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/dinic.py b/graphs/dinic.py
index aaf3a119525c..7919e6bc060a 100644
--- a/graphs/dinic.py
+++ b/graphs/dinic.py
@@ -37,7 +37,7 @@ def depth_first_search(self, vertex, sink, flow):
# Here we calculate the flow that reaches the sink
def max_flow(self, source, sink):
flow, self.q[0] = 0, source
- for l in range(31): # noqa: E741 l = 30 maybe faster for random data
+ for l in range(31): # l = 30 maybe faster for random data # noqa: E741
while True:
self.lvl, self.ptr = [0] * len(self.q), [0] * len(self.q)
qi, qe, self.lvl[source] = 0, 1, 1
diff --git a/graphs/directed_and_undirected_(weighted)_graph.py b/graphs/directed_and_undirected_weighted_graph.py
similarity index 93%
rename from graphs/directed_and_undirected_(weighted)_graph.py
rename to graphs/directed_and_undirected_weighted_graph.py
index 5cfa9e13edd9..8ca645fdace8 100644
--- a/graphs/directed_and_undirected_(weighted)_graph.py
+++ b/graphs/directed_and_undirected_weighted_graph.py
@@ -39,7 +39,7 @@ def dfs(self, s=-2, d=-1):
stack = []
visited = []
if s == -2:
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
ss = s
@@ -87,7 +87,7 @@ def bfs(self, s=-2):
d = deque()
visited = []
if s == -2:
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
d.append(s)
visited.append(s)
while d:
@@ -114,7 +114,7 @@ def topological_sort(self, s=-2):
stack = []
visited = []
if s == -2:
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
ss = s
@@ -146,7 +146,7 @@ def topological_sort(self, s=-2):
def cycle_nodes(self):
stack = []
visited = []
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
parent = -2
@@ -167,7 +167,7 @@ def cycle_nodes(self):
and not on_the_way_back
):
len_stack = len(stack) - 1
- while True and len_stack >= 0:
+ while len_stack >= 0:
if stack[len_stack] == node[1]:
anticipating_nodes.add(node[1])
break
@@ -199,7 +199,7 @@ def cycle_nodes(self):
def has_cycle(self):
stack = []
visited = []
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
parent = -2
@@ -220,15 +220,12 @@ def has_cycle(self):
and not on_the_way_back
):
len_stack_minus_one = len(stack) - 1
- while True and len_stack_minus_one >= 0:
+ while len_stack_minus_one >= 0:
if stack[len_stack_minus_one] == node[1]:
anticipating_nodes.add(node[1])
break
else:
return True
- # TODO:The following code is unreachable.
- anticipating_nodes.add(stack[len_stack_minus_one])
- len_stack_minus_one -= 1
if visited.count(node[1]) < 1:
stack.append(node[1])
visited.append(node[1])
@@ -308,7 +305,7 @@ def dfs(self, s=-2, d=-1):
stack = []
visited = []
if s == -2:
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
ss = s
@@ -356,7 +353,7 @@ def bfs(self, s=-2):
d = deque()
visited = []
if s == -2:
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
d.append(s)
visited.append(s)
while d:
@@ -374,7 +371,7 @@ def degree(self, u):
def cycle_nodes(self):
stack = []
visited = []
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
parent = -2
@@ -395,7 +392,7 @@ def cycle_nodes(self):
and not on_the_way_back
):
len_stack = len(stack) - 1
- while True and len_stack >= 0:
+ while len_stack >= 0:
if stack[len_stack] == node[1]:
anticipating_nodes.add(node[1])
break
@@ -427,7 +424,7 @@ def cycle_nodes(self):
def has_cycle(self):
stack = []
visited = []
- s = list(self.graph)[0]
+ s = next(iter(self.graph))
stack.append(s)
visited.append(s)
parent = -2
@@ -448,16 +445,12 @@ def has_cycle(self):
and not on_the_way_back
):
len_stack_minus_one = len(stack) - 1
- while True and len_stack_minus_one >= 0:
+ while len_stack_minus_one >= 0:
if stack[len_stack_minus_one] == node[1]:
anticipating_nodes.add(node[1])
break
else:
return True
- # TODO: the following code is unreachable
- # is this meant to be called in the else ?
- anticipating_nodes.add(stack[len_stack_minus_one])
- len_stack_minus_one -= 1
if visited.count(node[1]) < 1:
stack.append(node[1])
visited.append(node[1])
diff --git a/graphs/edmonds_karp_multiple_source_and_sink.py b/graphs/edmonds_karp_multiple_source_and_sink.py
index 0f359ff1aea3..5c774f4b812b 100644
--- a/graphs/edmonds_karp_multiple_source_and_sink.py
+++ b/graphs/edmonds_karp_multiple_source_and_sink.py
@@ -1,15 +1,15 @@
class FlowNetwork:
def __init__(self, graph, sources, sinks):
- self.sourceIndex = None
- self.sinkIndex = None
+ self.source_index = None
+ self.sink_index = None
self.graph = graph
- self._normalizeGraph(sources, sinks)
- self.verticesCount = len(graph)
- self.maximumFlowAlgorithm = None
+ self._normalize_graph(sources, sinks)
+ self.vertices_count = len(graph)
+ self.maximum_flow_algorithm = None
# make only one source and one sink
- def _normalizeGraph(self, sources, sinks):
+ def _normalize_graph(self, sources, sinks):
if sources is int:
sources = [sources]
if sinks is int:
@@ -18,54 +18,54 @@ def _normalizeGraph(self, sources, sinks):
if len(sources) == 0 or len(sinks) == 0:
return
- self.sourceIndex = sources[0]
- self.sinkIndex = sinks[0]
+ self.source_index = sources[0]
+ self.sink_index = sinks[0]
# make fake vertex if there are more
# than one source or sink
if len(sources) > 1 or len(sinks) > 1:
- maxInputFlow = 0
+ max_input_flow = 0
for i in sources:
- maxInputFlow += sum(self.graph[i])
+ max_input_flow += sum(self.graph[i])
size = len(self.graph) + 1
for room in self.graph:
room.insert(0, 0)
self.graph.insert(0, [0] * size)
for i in sources:
- self.graph[0][i + 1] = maxInputFlow
- self.sourceIndex = 0
+ self.graph[0][i + 1] = max_input_flow
+ self.source_index = 0
size = len(self.graph) + 1
for room in self.graph:
room.append(0)
self.graph.append([0] * size)
for i in sinks:
- self.graph[i + 1][size - 1] = maxInputFlow
- self.sinkIndex = size - 1
+ self.graph[i + 1][size - 1] = max_input_flow
+ self.sink_index = size - 1
- def findMaximumFlow(self):
- if self.maximumFlowAlgorithm is None:
+ def find_maximum_flow(self):
+ if self.maximum_flow_algorithm is None:
raise Exception("You need to set maximum flow algorithm before.")
- if self.sourceIndex is None or self.sinkIndex is None:
+ if self.source_index is None or self.sink_index is None:
return 0
- self.maximumFlowAlgorithm.execute()
- return self.maximumFlowAlgorithm.getMaximumFlow()
+ self.maximum_flow_algorithm.execute()
+ return self.maximum_flow_algorithm.getMaximumFlow()
- def setMaximumFlowAlgorithm(self, Algorithm):
- self.maximumFlowAlgorithm = Algorithm(self)
+ def set_maximum_flow_algorithm(self, algorithm):
+ self.maximum_flow_algorithm = algorithm(self)
class FlowNetworkAlgorithmExecutor:
- def __init__(self, flowNetwork):
- self.flowNetwork = flowNetwork
- self.verticesCount = flowNetwork.verticesCount
- self.sourceIndex = flowNetwork.sourceIndex
- self.sinkIndex = flowNetwork.sinkIndex
+ def __init__(self, flow_network):
+ self.flow_network = flow_network
+ self.verticies_count = flow_network.verticesCount
+ self.source_index = flow_network.sourceIndex
+ self.sink_index = flow_network.sinkIndex
# it's just a reference, so you shouldn't change
# it in your algorithms, use deep copy before doing that
- self.graph = flowNetwork.graph
+ self.graph = flow_network.graph
self.executed = False
def execute(self):
@@ -79,95 +79,95 @@ def _algorithm(self):
class MaximumFlowAlgorithmExecutor(FlowNetworkAlgorithmExecutor):
- def __init__(self, flowNetwork):
- super().__init__(flowNetwork)
+ def __init__(self, flow_network):
+ super().__init__(flow_network)
# use this to save your result
- self.maximumFlow = -1
+ self.maximum_flow = -1
- def getMaximumFlow(self):
+ def get_maximum_flow(self):
if not self.executed:
raise Exception("You should execute algorithm before using its result!")
- return self.maximumFlow
+ return self.maximum_flow
class PushRelabelExecutor(MaximumFlowAlgorithmExecutor):
- def __init__(self, flowNetwork):
- super().__init__(flowNetwork)
+ def __init__(self, flow_network):
+ super().__init__(flow_network)
- self.preflow = [[0] * self.verticesCount for i in range(self.verticesCount)]
+ self.preflow = [[0] * self.verticies_count for i in range(self.verticies_count)]
- self.heights = [0] * self.verticesCount
- self.excesses = [0] * self.verticesCount
+ self.heights = [0] * self.verticies_count
+ self.excesses = [0] * self.verticies_count
def _algorithm(self):
- self.heights[self.sourceIndex] = self.verticesCount
+ self.heights[self.source_index] = self.verticies_count
# push some substance to graph
- for nextVertexIndex, bandwidth in enumerate(self.graph[self.sourceIndex]):
- self.preflow[self.sourceIndex][nextVertexIndex] += bandwidth
- self.preflow[nextVertexIndex][self.sourceIndex] -= bandwidth
- self.excesses[nextVertexIndex] += bandwidth
+ for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index]):
+ self.preflow[self.source_index][nextvertex_index] += bandwidth
+ self.preflow[nextvertex_index][self.source_index] -= bandwidth
+ self.excesses[nextvertex_index] += bandwidth
# Relabel-to-front selection rule
- verticesList = [
+ vertices_list = [
i
- for i in range(self.verticesCount)
- if i != self.sourceIndex and i != self.sinkIndex
+ for i in range(self.verticies_count)
+ if i not in {self.source_index, self.sink_index}
]
# move through list
i = 0
- while i < len(verticesList):
- vertexIndex = verticesList[i]
- previousHeight = self.heights[vertexIndex]
- self.processVertex(vertexIndex)
- if self.heights[vertexIndex] > previousHeight:
+ while i < len(vertices_list):
+ vertex_index = vertices_list[i]
+ previous_height = self.heights[vertex_index]
+ self.process_vertex(vertex_index)
+ if self.heights[vertex_index] > previous_height:
# if it was relabeled, swap elements
# and start from 0 index
- verticesList.insert(0, verticesList.pop(i))
+ vertices_list.insert(0, vertices_list.pop(i))
i = 0
else:
i += 1
- self.maximumFlow = sum(self.preflow[self.sourceIndex])
+ self.maximum_flow = sum(self.preflow[self.source_index])
- def processVertex(self, vertexIndex):
- while self.excesses[vertexIndex] > 0:
- for neighbourIndex in range(self.verticesCount):
+ def process_vertex(self, vertex_index):
+ while self.excesses[vertex_index] > 0:
+ for neighbour_index in range(self.verticies_count):
# if it's neighbour and current vertex is higher
if (
- self.graph[vertexIndex][neighbourIndex]
- - self.preflow[vertexIndex][neighbourIndex]
+ self.graph[vertex_index][neighbour_index]
+ - self.preflow[vertex_index][neighbour_index]
> 0
- and self.heights[vertexIndex] > self.heights[neighbourIndex]
+ and self.heights[vertex_index] > self.heights[neighbour_index]
):
- self.push(vertexIndex, neighbourIndex)
+ self.push(vertex_index, neighbour_index)
- self.relabel(vertexIndex)
+ self.relabel(vertex_index)
- def push(self, fromIndex, toIndex):
- preflowDelta = min(
- self.excesses[fromIndex],
- self.graph[fromIndex][toIndex] - self.preflow[fromIndex][toIndex],
+ def push(self, from_index, to_index):
+ preflow_delta = min(
+ self.excesses[from_index],
+ self.graph[from_index][to_index] - self.preflow[from_index][to_index],
)
- self.preflow[fromIndex][toIndex] += preflowDelta
- self.preflow[toIndex][fromIndex] -= preflowDelta
- self.excesses[fromIndex] -= preflowDelta
- self.excesses[toIndex] += preflowDelta
-
- def relabel(self, vertexIndex):
- minHeight = None
- for toIndex in range(self.verticesCount):
+ self.preflow[from_index][to_index] += preflow_delta
+ self.preflow[to_index][from_index] -= preflow_delta
+ self.excesses[from_index] -= preflow_delta
+ self.excesses[to_index] += preflow_delta
+
+ def relabel(self, vertex_index):
+ min_height = None
+ for to_index in range(self.verticies_count):
if (
- self.graph[vertexIndex][toIndex] - self.preflow[vertexIndex][toIndex]
+ self.graph[vertex_index][to_index]
+ - self.preflow[vertex_index][to_index]
> 0
- ):
- if minHeight is None or self.heights[toIndex] < minHeight:
- minHeight = self.heights[toIndex]
+ ) and (min_height is None or self.heights[to_index] < min_height):
+ min_height = self.heights[to_index]
- if minHeight is not None:
- self.heights[vertexIndex] = minHeight + 1
+ if min_height is not None:
+ self.heights[vertex_index] = min_height + 1
if __name__ == "__main__":
@@ -184,10 +184,10 @@ def relabel(self, vertexIndex):
graph = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]]
# prepare our network
- flowNetwork = FlowNetwork(graph, entrances, exits)
+ flow_network = FlowNetwork(graph, entrances, exits)
# set algorithm
- flowNetwork.setMaximumFlowAlgorithm(PushRelabelExecutor)
+ flow_network.set_maximum_flow_algorithm(PushRelabelExecutor)
# and calculate
- maximumFlow = flowNetwork.findMaximumFlow()
+ maximum_flow = flow_network.find_maximum_flow()
- print(f"maximum flow is {maximumFlow}")
+ print(f"maximum flow is {maximum_flow}")
diff --git a/graphs/eulerian_path_and_circuit_for_undirected_graph.py b/graphs/eulerian_path_and_circuit_for_undirected_graph.py
index 7850933b0201..5b146eaa845b 100644
--- a/graphs/eulerian_path_and_circuit_for_undirected_graph.py
+++ b/graphs/eulerian_path_and_circuit_for_undirected_graph.py
@@ -6,8 +6,8 @@
# using dfs for finding eulerian path traversal
-def dfs(u, graph, visited_edge, path=[]):
- path = path + [u]
+def dfs(u, graph, visited_edge, path=None):
+ path = (path or []) + [u]
for v in graph[u]:
if visited_edge[u][v] is False:
visited_edge[u][v], visited_edge[v][u] = True, True
@@ -20,7 +20,7 @@ def check_circuit_or_path(graph, max_node):
odd_degree_nodes = 0
odd_node = -1
for i in range(max_node):
- if i not in graph.keys():
+ if i not in graph:
continue
if len(graph[i]) % 2 == 1:
odd_degree_nodes += 1
@@ -50,21 +50,21 @@ def check_euler(graph, max_node):
def main():
- G1 = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]}
- G2 = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]}
- G3 = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]}
- G4 = {1: [2, 3], 2: [1, 3], 3: [1, 2]}
- G5 = {
+ g1 = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]}
+ g2 = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]}
+ g3 = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]}
+ g4 = {1: [2, 3], 2: [1, 3], 3: [1, 2]}
+ g5 = {
1: [],
- 2: []
+ 2: [],
# all degree is zero
}
max_node = 10
- check_euler(G1, max_node)
- check_euler(G2, max_node)
- check_euler(G3, max_node)
- check_euler(G4, max_node)
- check_euler(G5, max_node)
+ check_euler(g1, max_node)
+ check_euler(g2, max_node)
+ check_euler(g3, max_node)
+ check_euler(g4, max_node)
+ check_euler(g5, max_node)
if __name__ == "__main__":
diff --git a/graphs/even_tree.py b/graphs/even_tree.py
index c9aef6e7861f..7d47899527a7 100644
--- a/graphs/even_tree.py
+++ b/graphs/even_tree.py
@@ -12,16 +12,17 @@
Note: The tree input will be such that it can always be decomposed into
components containing an even number of nodes.
"""
+
# pylint: disable=invalid-name
from collections import defaultdict
-def dfs(start):
+def dfs(start: int) -> int:
"""DFS traversal"""
# pylint: disable=redefined-outer-name
ret = 1
visited[start] = True
- for v in tree.get(start):
+ for v in tree[start]:
if v not in visited:
ret += dfs(v)
if ret % 2 == 0:
@@ -48,8 +49,8 @@ def even_tree():
if __name__ == "__main__":
n, m = 10, 9
tree = defaultdict(list)
- visited = {}
- cuts = []
+ visited: dict[int, bool] = {}
+ cuts: list[int] = []
count = 0
edges = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)]
for u, v in edges:
diff --git a/graphs/finding_bridges.py b/graphs/finding_bridges.py
index 6555dd7bc29e..c17606745ad8 100644
--- a/graphs/finding_bridges.py
+++ b/graphs/finding_bridges.py
@@ -1,42 +1,106 @@
-# Finding Bridges in Undirected Graph
-def computeBridges(graph):
- id = 0
+"""
+An edge is a bridge if, after removing it count of connected components in graph will
+be increased by one. Bridges represent vulnerabilities in a connected network and are
+useful for designing reliable networks. For example, in a wired computer network, an
+articulation point indicates the critical computers and a bridge indicates the critical
+wires or connections.
+
+For more details, refer this article:
+https://www.geeksforgeeks.org/bridge-in-a-graph/
+"""
+
+
+def __get_demo_graph(index):
+ return [
+ {
+ 0: [1, 2],
+ 1: [0, 2],
+ 2: [0, 1, 3, 5],
+ 3: [2, 4],
+ 4: [3],
+ 5: [2, 6, 8],
+ 6: [5, 7],
+ 7: [6, 8],
+ 8: [5, 7],
+ },
+ {
+ 0: [6],
+ 1: [9],
+ 2: [4, 5],
+ 3: [4],
+ 4: [2, 3],
+ 5: [2],
+ 6: [0, 7],
+ 7: [6],
+ 8: [],
+ 9: [1],
+ },
+ {
+ 0: [4],
+ 1: [6],
+ 2: [],
+ 3: [5, 6, 7],
+ 4: [0, 6],
+ 5: [3, 8, 9],
+ 6: [1, 3, 4, 7],
+ 7: [3, 6, 8, 9],
+ 8: [5, 7],
+ 9: [5, 7],
+ },
+ {
+ 0: [1, 3],
+ 1: [0, 2, 4],
+ 2: [1, 3, 4],
+ 3: [0, 2, 4],
+ 4: [1, 2, 3],
+ },
+ ][index]
+
+
+def compute_bridges(graph: dict[int, list[int]]) -> list[tuple[int, int]]:
+ """
+ Return the list of undirected graph bridges [(a1, b1), ..., (ak, bk)]; ai <= bi
+ >>> compute_bridges(__get_demo_graph(0))
+ [(3, 4), (2, 3), (2, 5)]
+ >>> compute_bridges(__get_demo_graph(1))
+ [(6, 7), (0, 6), (1, 9), (3, 4), (2, 4), (2, 5)]
+ >>> compute_bridges(__get_demo_graph(2))
+ [(1, 6), (4, 6), (0, 4)]
+ >>> compute_bridges(__get_demo_graph(3))
+ []
+ >>> compute_bridges({})
+ []
+ """
+
+ id_ = 0
n = len(graph) # No of vertices in graph
low = [0] * n
visited = [False] * n
- def dfs(at, parent, bridges, id):
+ def dfs(at, parent, bridges, id_):
visited[at] = True
- low[at] = id
- id += 1
+ low[at] = id_
+ id_ += 1
for to in graph[at]:
if to == parent:
pass
elif not visited[to]:
- dfs(to, at, bridges, id)
+ dfs(to, at, bridges, id_)
low[at] = min(low[at], low[to])
- if at < low[to]:
- bridges.append([at, to])
+ if id_ <= low[to]:
+ bridges.append((at, to) if at < to else (to, at))
else:
# This edge is a back edge and cannot be a bridge
- low[at] = min(low[at], to)
+ low[at] = min(low[at], low[to])
- bridges = []
+ bridges: list[tuple[int, int]] = []
for i in range(n):
if not visited[i]:
- dfs(i, -1, bridges, id)
- print(bridges)
-
-
-graph = {
- 0: [1, 2],
- 1: [0, 2],
- 2: [0, 1, 3, 5],
- 3: [2, 4],
- 4: [3],
- 5: [2, 6, 8],
- 6: [5, 7],
- 7: [6, 8],
- 8: [5, 7],
-}
-computeBridges(graph)
+ dfs(i, -1, bridges, id_)
+ return bridges
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/frequent_pattern_graph_miner.py b/graphs/frequent_pattern_graph_miner.py
index ff7063082267..f8da73f3438e 100644
--- a/graphs/frequent_pattern_graph_miner.py
+++ b/graphs/frequent_pattern_graph_miner.py
@@ -8,6 +8,7 @@
URL: https://www.researchgate.net/publication/235255851
"""
+
# fmt: off
edge_array = [
['ab-e1', 'ac-e3', 'ad-e5', 'bc-e4', 'bd-e2', 'be-e6', 'bh-e12', 'cd-e2', 'ce-e4',
@@ -54,7 +55,7 @@ def get_frequency_table(edge_array):
Returns Frequency Table
"""
distinct_edge = get_distinct_edge(edge_array)
- frequency_table = dict()
+ frequency_table = {}
for item in distinct_edge:
bit = get_bitcode(edge_array, item)
@@ -79,7 +80,7 @@ def get_nodes(frequency_table):
{'11111': ['ab', 'ac', 'df', 'bd', 'bc']}
"""
nodes = {}
- for i, item in enumerate(frequency_table):
+ for _, item in enumerate(frequency_table):
nodes.setdefault(item[2], []).append(item[0])
return nodes
@@ -130,11 +131,11 @@ def create_edge(nodes, graph, cluster, c1):
"""
create edge between the nodes
"""
- for i in cluster[c1].keys():
+ for i in cluster[c1]:
count = 0
c2 = c1 + 1
while c2 < max(cluster.keys()):
- for j in cluster[c2].keys():
+ for j in cluster[c2]:
"""
creates edge only if the condition satisfies
"""
@@ -151,16 +152,16 @@ def create_edge(nodes, graph, cluster, c1):
def construct_graph(cluster, nodes):
- X = cluster[max(cluster.keys())]
+ x = cluster[max(cluster.keys())]
cluster[max(cluster.keys()) + 1] = "Header"
graph = {}
- for i in X:
- if tuple(["Header"]) in graph:
- graph[tuple(["Header"])].append(X[i])
+ for i in x:
+ if (["Header"],) in graph:
+ graph[(["Header"],)].append(x[i])
else:
- graph[tuple(["Header"])] = [X[i]]
- for i in X:
- graph[tuple(X[i])] = [["Header"]]
+ graph[(["Header"],)] = [x[i]]
+ for i in x:
+ graph[(x[i],)] = [["Header"]]
i = 1
while i < max(cluster) - 1:
create_edge(nodes, graph, cluster, i)
@@ -168,16 +169,16 @@ def construct_graph(cluster, nodes):
return graph
-def myDFS(graph, start, end, path=[]):
+def my_dfs(graph, start, end, path=None):
"""
find different DFS walk from given node to Header node
"""
- path = path + [start]
+ path = (path or []) + [start]
if start == end:
paths.append(path)
for node in graph[start]:
if tuple(node) not in path:
- myDFS(graph, tuple(node), end, path)
+ my_dfs(graph, tuple(node), end, path)
def find_freq_subgraph_given_support(s, cluster, graph):
@@ -185,24 +186,24 @@ def find_freq_subgraph_given_support(s, cluster, graph):
find edges of multiple frequent subgraphs
"""
k = int(s / 100 * (len(cluster) - 1))
- for i in cluster[k].keys():
- myDFS(graph, tuple(cluster[k][i]), tuple(["Header"]))
+ for i in cluster[k]:
+ my_dfs(graph, tuple(cluster[k][i]), (["Header"],))
def freq_subgraphs_edge_list(paths):
"""
returns Edge list for frequent subgraphs
"""
- freq_sub_EL = []
+ freq_sub_el = []
for edges in paths:
- EL = []
+ el = []
for j in range(len(edges) - 1):
temp = list(edges[j])
for e in temp:
edge = (e[0], e[1])
- EL.append(edge)
- freq_sub_EL.append(EL)
- return freq_sub_EL
+ el.append(edge)
+ freq_sub_el.append(el)
+ return freq_sub_el
def preprocess(edge_array):
@@ -227,6 +228,6 @@ def preprocess(edge_array):
support = get_support(cluster)
graph = construct_graph(cluster, nodes)
find_freq_subgraph_given_support(60, cluster, graph)
- paths = []
+ paths: list = []
freq_subgraph_edge_list = freq_subgraphs_edge_list(paths)
print_all()
diff --git a/graphs/gale_shapley_bigraph.py b/graphs/gale_shapley_bigraph.py
index 59baf8296ea6..f4b3153817c4 100644
--- a/graphs/gale_shapley_bigraph.py
+++ b/graphs/gale_shapley_bigraph.py
@@ -1,7 +1,9 @@
from __future__ import annotations
-def stable_matching(donor_pref: list[int], recipient_pref: list[int]) -> list[int]:
+def stable_matching(
+ donor_pref: list[list[int]], recipient_pref: list[list[int]]
+) -> list[int]:
"""
Finds the stable match in any bipartite graph, i.e a pairing where no 2 objects
prefer each other over their partner. The function accepts the preferences of
@@ -15,15 +17,17 @@ def stable_matching(donor_pref: list[int], recipient_pref: list[int]) -> list[in
>>> donor_pref = [[0, 1, 3, 2], [0, 2, 3, 1], [1, 0, 2, 3], [0, 3, 1, 2]]
>>> recipient_pref = [[3, 1, 2, 0], [3, 1, 0, 2], [0, 3, 1, 2], [1, 0, 3, 2]]
- >>> print(stable_matching(donor_pref, recipient_pref))
+ >>> stable_matching(donor_pref, recipient_pref)
[1, 2, 3, 0]
"""
assert len(donor_pref) == len(recipient_pref)
+
n = len(donor_pref)
unmatched_donors = list(range(n))
donor_record = [-1] * n # who the donor has donated to
rec_record = [-1] * n # who the recipient has received from
num_donations = [0] * n
+
while unmatched_donors:
donor = unmatched_donors[0]
donor_preference = donor_pref[donor]
@@ -31,6 +35,7 @@ def stable_matching(donor_pref: list[int], recipient_pref: list[int]) -> list[in
num_donations[donor] += 1
rec_preference = recipient_pref[recipient]
prev_donor = rec_record[recipient]
+
if prev_donor != -1:
if rec_preference.index(prev_donor) > rec_preference.index(donor):
rec_record[recipient] = donor
diff --git a/graphs/graph_adjacency_list.py b/graphs/graph_adjacency_list.py
new file mode 100644
index 000000000000..abc75311cd60
--- /dev/null
+++ b/graphs/graph_adjacency_list.py
@@ -0,0 +1,588 @@
+#!/usr/bin/env python3
+"""
+Author: Vikram Nithyanandam
+
+Description:
+The following implementation is a robust unweighted Graph data structure
+implemented using an adjacency list. This vertices and edges of this graph can be
+effectively initialized and modified while storing your chosen generic
+value in each vertex.
+
+Adjacency List: https://en.wikipedia.org/wiki/Adjacency_list
+
+Potential Future Ideas:
+- Add a flag to set edge weights on and set edge weights
+- Make edge weights and vertex values customizable to store whatever the client wants
+- Support multigraph functionality if the client wants it
+"""
+
+from __future__ import annotations
+
+import random
+import unittest
+from pprint import pformat
+from typing import Generic, TypeVar
+
+import pytest
+
+T = TypeVar("T")
+
+
+class GraphAdjacencyList(Generic[T]):
+ def __init__(
+ self, vertices: list[T], edges: list[list[T]], directed: bool = True
+ ) -> None:
+ """
+ Parameters:
+ - vertices: (list[T]) The list of vertex names the client wants to
+ pass in. Default is empty.
+ - edges: (list[list[T]]) The list of edges the client wants to
+ pass in. Each edge is a 2-element list. Default is empty.
+ - directed: (bool) Indicates if graph is directed or undirected.
+ Default is True.
+ """
+ self.adj_list: dict[T, list[T]] = {} # dictionary of lists of T
+ self.directed = directed
+
+ # Falsey checks
+ edges = edges or []
+ vertices = vertices or []
+
+ for vertex in vertices:
+ self.add_vertex(vertex)
+
+ for edge in edges:
+ if len(edge) != 2:
+ msg = f"Invalid input: {edge} is the wrong length."
+ raise ValueError(msg)
+ self.add_edge(edge[0], edge[1])
+
+ def add_vertex(self, vertex: T) -> None:
+ """
+ Adds a vertex to the graph. If the given vertex already exists,
+ a ValueError will be thrown.
+ """
+ if self.contains_vertex(vertex):
+ msg = f"Incorrect input: {vertex} is already in the graph."
+ raise ValueError(msg)
+ self.adj_list[vertex] = []
+
+ def add_edge(self, source_vertex: T, destination_vertex: T) -> None:
+ """
+ Creates an edge from source vertex to destination vertex. If any
+ given vertex doesn't exist or the edge already exists, a ValueError
+ will be thrown.
+ """
+ if not (
+ self.contains_vertex(source_vertex)
+ and self.contains_vertex(destination_vertex)
+ ):
+ msg = (
+ f"Incorrect input: Either {source_vertex} or "
+ f"{destination_vertex} does not exist"
+ )
+ raise ValueError(msg)
+ if self.contains_edge(source_vertex, destination_vertex):
+ msg = (
+ "Incorrect input: The edge already exists between "
+ f"{source_vertex} and {destination_vertex}"
+ )
+ raise ValueError(msg)
+
+ # add the destination vertex to the list associated with the source vertex
+ # and vice versa if not directed
+ self.adj_list[source_vertex].append(destination_vertex)
+ if not self.directed:
+ self.adj_list[destination_vertex].append(source_vertex)
+
+ def remove_vertex(self, vertex: T) -> None:
+ """
+ Removes the given vertex from the graph and deletes all incoming and
+ outgoing edges from the given vertex as well. If the given vertex
+ does not exist, a ValueError will be thrown.
+ """
+ if not self.contains_vertex(vertex):
+ msg = f"Incorrect input: {vertex} does not exist in this graph."
+ raise ValueError(msg)
+
+ if not self.directed:
+ # If not directed, find all neighboring vertices and delete all references
+ # of edges connecting to the given vertex
+ for neighbor in self.adj_list[vertex]:
+ self.adj_list[neighbor].remove(vertex)
+ else:
+ # If directed, search all neighbors of all vertices and delete all
+ # references of edges connecting to the given vertex
+ for edge_list in self.adj_list.values():
+ if vertex in edge_list:
+ edge_list.remove(vertex)
+
+ # Finally, delete the given vertex and all of its outgoing edge references
+ self.adj_list.pop(vertex)
+
+ def remove_edge(self, source_vertex: T, destination_vertex: T) -> None:
+ """
+ Removes the edge between the two vertices. If any given vertex
+ doesn't exist or the edge does not exist, a ValueError will be thrown.
+ """
+ if not (
+ self.contains_vertex(source_vertex)
+ and self.contains_vertex(destination_vertex)
+ ):
+ msg = (
+ f"Incorrect input: Either {source_vertex} or "
+ f"{destination_vertex} does not exist"
+ )
+ raise ValueError(msg)
+ if not self.contains_edge(source_vertex, destination_vertex):
+ msg = (
+ "Incorrect input: The edge does NOT exist between "
+ f"{source_vertex} and {destination_vertex}"
+ )
+ raise ValueError(msg)
+
+ # remove the destination vertex from the list associated with the source
+ # vertex and vice versa if not directed
+ self.adj_list[source_vertex].remove(destination_vertex)
+ if not self.directed:
+ self.adj_list[destination_vertex].remove(source_vertex)
+
+ def contains_vertex(self, vertex: T) -> bool:
+ """
+ Returns True if the graph contains the vertex, False otherwise.
+ """
+ return vertex in self.adj_list
+
+ def contains_edge(self, source_vertex: T, destination_vertex: T) -> bool:
+ """
+ Returns True if the graph contains the edge from the source_vertex to the
+ destination_vertex, False otherwise. If any given vertex doesn't exist, a
+ ValueError will be thrown.
+ """
+ if not (
+ self.contains_vertex(source_vertex)
+ and self.contains_vertex(destination_vertex)
+ ):
+ msg = (
+ f"Incorrect input: Either {source_vertex} "
+ f"or {destination_vertex} does not exist."
+ )
+ raise ValueError(msg)
+
+ return destination_vertex in self.adj_list[source_vertex]
+
+ def clear_graph(self) -> None:
+ """
+ Clears all vertices and edges.
+ """
+ self.adj_list = {}
+
+ def __repr__(self) -> str:
+ return pformat(self.adj_list)
+
+
+class TestGraphAdjacencyList(unittest.TestCase):
+ def __assert_graph_edge_exists_check(
+ self,
+ undirected_graph: GraphAdjacencyList,
+ directed_graph: GraphAdjacencyList,
+ edge: list[int],
+ ) -> None:
+ assert undirected_graph.contains_edge(edge[0], edge[1])
+ assert undirected_graph.contains_edge(edge[1], edge[0])
+ assert directed_graph.contains_edge(edge[0], edge[1])
+
+ def __assert_graph_edge_does_not_exist_check(
+ self,
+ undirected_graph: GraphAdjacencyList,
+ directed_graph: GraphAdjacencyList,
+ edge: list[int],
+ ) -> None:
+ assert not undirected_graph.contains_edge(edge[0], edge[1])
+ assert not undirected_graph.contains_edge(edge[1], edge[0])
+ assert not directed_graph.contains_edge(edge[0], edge[1])
+
+ def __assert_graph_vertex_exists_check(
+ self,
+ undirected_graph: GraphAdjacencyList,
+ directed_graph: GraphAdjacencyList,
+ vertex: int,
+ ) -> None:
+ assert undirected_graph.contains_vertex(vertex)
+ assert directed_graph.contains_vertex(vertex)
+
+ def __assert_graph_vertex_does_not_exist_check(
+ self,
+ undirected_graph: GraphAdjacencyList,
+ directed_graph: GraphAdjacencyList,
+ vertex: int,
+ ) -> None:
+ assert not undirected_graph.contains_vertex(vertex)
+ assert not directed_graph.contains_vertex(vertex)
+
+ def __generate_random_edges(
+ self, vertices: list[int], edge_pick_count: int
+ ) -> list[list[int]]:
+ assert edge_pick_count <= len(vertices)
+
+ random_source_vertices: list[int] = random.sample(
+ vertices[0 : int(len(vertices) / 2)], edge_pick_count
+ )
+ random_destination_vertices: list[int] = random.sample(
+ vertices[int(len(vertices) / 2) :], edge_pick_count
+ )
+ random_edges: list[list[int]] = []
+
+ for source in random_source_vertices:
+ for dest in random_destination_vertices:
+ random_edges.append([source, dest])
+
+ return random_edges
+
+ def __generate_graphs(
+ self, vertex_count: int, min_val: int, max_val: int, edge_pick_count: int
+ ) -> tuple[GraphAdjacencyList, GraphAdjacencyList, list[int], list[list[int]]]:
+ if max_val - min_val + 1 < vertex_count:
+ raise ValueError(
+ "Will result in duplicate vertices. Either increase range "
+ "between min_val and max_val or decrease vertex count."
+ )
+
+ # generate graph input
+ random_vertices: list[int] = random.sample(
+ range(min_val, max_val + 1), vertex_count
+ )
+ random_edges: list[list[int]] = self.__generate_random_edges(
+ random_vertices, edge_pick_count
+ )
+
+ # build graphs
+ undirected_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=random_edges, directed=False
+ )
+ directed_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=random_edges, directed=True
+ )
+
+ return undirected_graph, directed_graph, random_vertices, random_edges
+
+ def test_init_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ # test graph initialization with vertices and edges
+ for num in random_vertices:
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, num
+ )
+
+ for edge in random_edges:
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+ assert not undirected_graph.directed
+ assert directed_graph.directed
+
+ def test_contains_vertex(self) -> None:
+ random_vertices: list[int] = random.sample(range(101), 20)
+
+ # Build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=[], directed=True
+ )
+
+ # Test contains_vertex
+ for num in range(101):
+ assert (num in random_vertices) == undirected_graph.contains_vertex(num)
+ assert (num in random_vertices) == directed_graph.contains_vertex(num)
+
+ def test_add_vertices(self) -> None:
+ random_vertices: list[int] = random.sample(range(101), 20)
+
+ # build empty graphs
+ undirected_graph: GraphAdjacencyList = GraphAdjacencyList(
+ vertices=[], edges=[], directed=False
+ )
+ directed_graph: GraphAdjacencyList = GraphAdjacencyList(
+ vertices=[], edges=[], directed=True
+ )
+
+ # run add_vertex
+ for num in random_vertices:
+ undirected_graph.add_vertex(num)
+
+ for num in random_vertices:
+ directed_graph.add_vertex(num)
+
+ # test add_vertex worked
+ for num in random_vertices:
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, num
+ )
+
+ def test_remove_vertices(self) -> None:
+ random_vertices: list[int] = random.sample(range(101), 20)
+
+ # build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=[], directed=True
+ )
+
+ # test remove_vertex worked
+ for num in random_vertices:
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, num
+ )
+
+ undirected_graph.remove_vertex(num)
+ directed_graph.remove_vertex(num)
+
+ self.__assert_graph_vertex_does_not_exist_check(
+ undirected_graph, directed_graph, num
+ )
+
+ def test_add_and_remove_vertices_repeatedly(self) -> None:
+ random_vertices1: list[int] = random.sample(range(51), 20)
+ random_vertices2: list[int] = random.sample(range(51, 101), 20)
+
+ # build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyList(
+ vertices=random_vertices1, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyList(
+ vertices=random_vertices1, edges=[], directed=True
+ )
+
+ # test adding and removing vertices
+ for i, _ in enumerate(random_vertices1):
+ undirected_graph.add_vertex(random_vertices2[i])
+ directed_graph.add_vertex(random_vertices2[i])
+
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, random_vertices2[i]
+ )
+
+ undirected_graph.remove_vertex(random_vertices1[i])
+ directed_graph.remove_vertex(random_vertices1[i])
+
+ self.__assert_graph_vertex_does_not_exist_check(
+ undirected_graph, directed_graph, random_vertices1[i]
+ )
+
+ # remove all vertices
+ for i, _ in enumerate(random_vertices1):
+ undirected_graph.remove_vertex(random_vertices2[i])
+ directed_graph.remove_vertex(random_vertices2[i])
+
+ self.__assert_graph_vertex_does_not_exist_check(
+ undirected_graph, directed_graph, random_vertices2[i]
+ )
+
+ def test_contains_edge(self) -> None:
+ # generate graphs and graph input
+ vertex_count = 20
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(vertex_count, 0, 100, 4)
+
+ # generate all possible edges for testing
+ all_possible_edges: list[list[int]] = []
+ for i in range(vertex_count - 1):
+ for j in range(i + 1, vertex_count):
+ all_possible_edges.append([random_vertices[i], random_vertices[j]])
+ all_possible_edges.append([random_vertices[j], random_vertices[i]])
+
+ # test contains_edge function
+ for edge in all_possible_edges:
+ if edge in random_edges:
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+ elif [edge[1], edge[0]] in random_edges:
+ # since this edge exists for undirected but the reverse
+ # may not exist for directed
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, [edge[1], edge[0]]
+ )
+ else:
+ self.__assert_graph_edge_does_not_exist_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ def test_add_edge(self) -> None:
+ # generate graph input
+ random_vertices: list[int] = random.sample(range(101), 15)
+ random_edges: list[list[int]] = self.__generate_random_edges(random_vertices, 4)
+
+ # build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyList(
+ vertices=random_vertices, edges=[], directed=True
+ )
+
+ # run and test add_edge
+ for edge in random_edges:
+ undirected_graph.add_edge(edge[0], edge[1])
+ directed_graph.add_edge(edge[0], edge[1])
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ def test_remove_edge(self) -> None:
+ # generate graph input and graphs
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ # run and test remove_edge
+ for edge in random_edges:
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+ undirected_graph.remove_edge(edge[0], edge[1])
+ directed_graph.remove_edge(edge[0], edge[1])
+ self.__assert_graph_edge_does_not_exist_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ def test_add_and_remove_edges_repeatedly(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ # make some more edge options!
+ more_random_edges: list[list[int]] = []
+
+ while len(more_random_edges) != len(random_edges):
+ edges: list[list[int]] = self.__generate_random_edges(random_vertices, 4)
+ for edge in edges:
+ if len(more_random_edges) == len(random_edges):
+ break
+ elif edge not in more_random_edges and edge not in random_edges:
+ more_random_edges.append(edge)
+
+ for i, _ in enumerate(random_edges):
+ undirected_graph.add_edge(more_random_edges[i][0], more_random_edges[i][1])
+ directed_graph.add_edge(more_random_edges[i][0], more_random_edges[i][1])
+
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, more_random_edges[i]
+ )
+
+ undirected_graph.remove_edge(random_edges[i][0], random_edges[i][1])
+ directed_graph.remove_edge(random_edges[i][0], random_edges[i][1])
+
+ self.__assert_graph_edge_does_not_exist_check(
+ undirected_graph, directed_graph, random_edges[i]
+ )
+
+ def test_add_vertex_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for vertex in random_vertices:
+ with pytest.raises(ValueError):
+ undirected_graph.add_vertex(vertex)
+ with pytest.raises(ValueError):
+ directed_graph.add_vertex(vertex)
+
+ def test_remove_vertex_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for i in range(101):
+ if i not in random_vertices:
+ with pytest.raises(ValueError):
+ undirected_graph.remove_vertex(i)
+ with pytest.raises(ValueError):
+ directed_graph.remove_vertex(i)
+
+ def test_add_edge_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for edge in random_edges:
+ with pytest.raises(ValueError):
+ undirected_graph.add_edge(edge[0], edge[1])
+ with pytest.raises(ValueError):
+ directed_graph.add_edge(edge[0], edge[1])
+
+ def test_remove_edge_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ more_random_edges: list[list[int]] = []
+
+ while len(more_random_edges) != len(random_edges):
+ edges: list[list[int]] = self.__generate_random_edges(random_vertices, 4)
+ for edge in edges:
+ if len(more_random_edges) == len(random_edges):
+ break
+ elif edge not in more_random_edges and edge not in random_edges:
+ more_random_edges.append(edge)
+
+ for edge in more_random_edges:
+ with pytest.raises(ValueError):
+ undirected_graph.remove_edge(edge[0], edge[1])
+ with pytest.raises(ValueError):
+ directed_graph.remove_edge(edge[0], edge[1])
+
+ def test_contains_edge_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for vertex in random_vertices:
+ with pytest.raises(ValueError):
+ undirected_graph.contains_edge(vertex, 102)
+ with pytest.raises(ValueError):
+ directed_graph.contains_edge(vertex, 102)
+
+ with pytest.raises(ValueError):
+ undirected_graph.contains_edge(103, 102)
+ with pytest.raises(ValueError):
+ directed_graph.contains_edge(103, 102)
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/graphs/graph_adjacency_matrix.py b/graphs/graph_adjacency_matrix.py
new file mode 100644
index 000000000000..568c84166e4b
--- /dev/null
+++ b/graphs/graph_adjacency_matrix.py
@@ -0,0 +1,609 @@
+#!/usr/bin/env python3
+"""
+Author: Vikram Nithyanandam
+
+Description:
+The following implementation is a robust unweighted Graph data structure
+implemented using an adjacency matrix. This vertices and edges of this graph can be
+effectively initialized and modified while storing your chosen generic
+value in each vertex.
+
+Adjacency Matrix: https://mathworld.wolfram.com/AdjacencyMatrix.html
+
+Potential Future Ideas:
+- Add a flag to set edge weights on and set edge weights
+- Make edge weights and vertex values customizable to store whatever the client wants
+- Support multigraph functionality if the client wants it
+"""
+
+from __future__ import annotations
+
+import random
+import unittest
+from pprint import pformat
+from typing import Generic, TypeVar
+
+import pytest
+
+T = TypeVar("T")
+
+
+class GraphAdjacencyMatrix(Generic[T]):
+ def __init__(
+ self, vertices: list[T], edges: list[list[T]], directed: bool = True
+ ) -> None:
+ """
+ Parameters:
+ - vertices: (list[T]) The list of vertex names the client wants to
+ pass in. Default is empty.
+ - edges: (list[list[T]]) The list of edges the client wants to
+ pass in. Each edge is a 2-element list. Default is empty.
+ - directed: (bool) Indicates if graph is directed or undirected.
+ Default is True.
+ """
+ self.directed = directed
+ self.vertex_to_index: dict[T, int] = {}
+ self.adj_matrix: list[list[int]] = []
+
+ # Falsey checks
+ edges = edges or []
+ vertices = vertices or []
+
+ for vertex in vertices:
+ self.add_vertex(vertex)
+
+ for edge in edges:
+ if len(edge) != 2:
+ msg = f"Invalid input: {edge} must have length 2."
+ raise ValueError(msg)
+ self.add_edge(edge[0], edge[1])
+
+ def add_edge(self, source_vertex: T, destination_vertex: T) -> None:
+ """
+ Creates an edge from source vertex to destination vertex. If any
+ given vertex doesn't exist or the edge already exists, a ValueError
+ will be thrown.
+ """
+ if not (
+ self.contains_vertex(source_vertex)
+ and self.contains_vertex(destination_vertex)
+ ):
+ msg = (
+ f"Incorrect input: Either {source_vertex} or "
+ f"{destination_vertex} does not exist"
+ )
+ raise ValueError(msg)
+ if self.contains_edge(source_vertex, destination_vertex):
+ msg = (
+ "Incorrect input: The edge already exists between "
+ f"{source_vertex} and {destination_vertex}"
+ )
+ raise ValueError(msg)
+
+ # Get the indices of the corresponding vertices and set their edge value to 1.
+ u: int = self.vertex_to_index[source_vertex]
+ v: int = self.vertex_to_index[destination_vertex]
+ self.adj_matrix[u][v] = 1
+ if not self.directed:
+ self.adj_matrix[v][u] = 1
+
+ def remove_edge(self, source_vertex: T, destination_vertex: T) -> None:
+ """
+ Removes the edge between the two vertices. If any given vertex
+ doesn't exist or the edge does not exist, a ValueError will be thrown.
+ """
+ if not (
+ self.contains_vertex(source_vertex)
+ and self.contains_vertex(destination_vertex)
+ ):
+ msg = (
+ f"Incorrect input: Either {source_vertex} or "
+ f"{destination_vertex} does not exist"
+ )
+ raise ValueError(msg)
+ if not self.contains_edge(source_vertex, destination_vertex):
+ msg = (
+ "Incorrect input: The edge does NOT exist between "
+ f"{source_vertex} and {destination_vertex}"
+ )
+ raise ValueError(msg)
+
+ # Get the indices of the corresponding vertices and set their edge value to 0.
+ u: int = self.vertex_to_index[source_vertex]
+ v: int = self.vertex_to_index[destination_vertex]
+ self.adj_matrix[u][v] = 0
+ if not self.directed:
+ self.adj_matrix[v][u] = 0
+
+ def add_vertex(self, vertex: T) -> None:
+ """
+ Adds a vertex to the graph. If the given vertex already exists,
+ a ValueError will be thrown.
+ """
+ if self.contains_vertex(vertex):
+ msg = f"Incorrect input: {vertex} already exists in this graph."
+ raise ValueError(msg)
+
+ # build column for vertex
+ for row in self.adj_matrix:
+ row.append(0)
+
+ # build row for vertex and update other data structures
+ self.adj_matrix.append([0] * (len(self.adj_matrix) + 1))
+ self.vertex_to_index[vertex] = len(self.adj_matrix) - 1
+
+ def remove_vertex(self, vertex: T) -> None:
+ """
+ Removes the given vertex from the graph and deletes all incoming and
+ outgoing edges from the given vertex as well. If the given vertex
+ does not exist, a ValueError will be thrown.
+ """
+ if not self.contains_vertex(vertex):
+ msg = f"Incorrect input: {vertex} does not exist in this graph."
+ raise ValueError(msg)
+
+ # first slide up the rows by deleting the row corresponding to
+ # the vertex being deleted.
+ start_index = self.vertex_to_index[vertex]
+ self.adj_matrix.pop(start_index)
+
+ # next, slide the columns to the left by deleting the values in
+ # the column corresponding to the vertex being deleted
+ for lst in self.adj_matrix:
+ lst.pop(start_index)
+
+ # final clean up
+ self.vertex_to_index.pop(vertex)
+
+ # decrement indices for vertices shifted by the deleted vertex in the adj matrix
+ for inner_vertex in self.vertex_to_index:
+ if self.vertex_to_index[inner_vertex] >= start_index:
+ self.vertex_to_index[inner_vertex] = (
+ self.vertex_to_index[inner_vertex] - 1
+ )
+
+ def contains_vertex(self, vertex: T) -> bool:
+ """
+ Returns True if the graph contains the vertex, False otherwise.
+ """
+ return vertex in self.vertex_to_index
+
+ def contains_edge(self, source_vertex: T, destination_vertex: T) -> bool:
+ """
+ Returns True if the graph contains the edge from the source_vertex to the
+ destination_vertex, False otherwise. If any given vertex doesn't exist, a
+ ValueError will be thrown.
+ """
+ if not (
+ self.contains_vertex(source_vertex)
+ and self.contains_vertex(destination_vertex)
+ ):
+ msg = (
+ f"Incorrect input: Either {source_vertex} "
+ f"or {destination_vertex} does not exist."
+ )
+ raise ValueError(msg)
+
+ u = self.vertex_to_index[source_vertex]
+ v = self.vertex_to_index[destination_vertex]
+ return self.adj_matrix[u][v] == 1
+
+ def clear_graph(self) -> None:
+ """
+ Clears all vertices and edges.
+ """
+ self.vertex_to_index = {}
+ self.adj_matrix = []
+
+ def __repr__(self) -> str:
+ first = "Adj Matrix:\n" + pformat(self.adj_matrix)
+ second = "\nVertex to index mapping:\n" + pformat(self.vertex_to_index)
+ return first + second
+
+
+class TestGraphMatrix(unittest.TestCase):
+ def __assert_graph_edge_exists_check(
+ self,
+ undirected_graph: GraphAdjacencyMatrix,
+ directed_graph: GraphAdjacencyMatrix,
+ edge: list[int],
+ ) -> None:
+ assert undirected_graph.contains_edge(edge[0], edge[1])
+ assert undirected_graph.contains_edge(edge[1], edge[0])
+ assert directed_graph.contains_edge(edge[0], edge[1])
+
+ def __assert_graph_edge_does_not_exist_check(
+ self,
+ undirected_graph: GraphAdjacencyMatrix,
+ directed_graph: GraphAdjacencyMatrix,
+ edge: list[int],
+ ) -> None:
+ assert not undirected_graph.contains_edge(edge[0], edge[1])
+ assert not undirected_graph.contains_edge(edge[1], edge[0])
+ assert not directed_graph.contains_edge(edge[0], edge[1])
+
+ def __assert_graph_vertex_exists_check(
+ self,
+ undirected_graph: GraphAdjacencyMatrix,
+ directed_graph: GraphAdjacencyMatrix,
+ vertex: int,
+ ) -> None:
+ assert undirected_graph.contains_vertex(vertex)
+ assert directed_graph.contains_vertex(vertex)
+
+ def __assert_graph_vertex_does_not_exist_check(
+ self,
+ undirected_graph: GraphAdjacencyMatrix,
+ directed_graph: GraphAdjacencyMatrix,
+ vertex: int,
+ ) -> None:
+ assert not undirected_graph.contains_vertex(vertex)
+ assert not directed_graph.contains_vertex(vertex)
+
+ def __generate_random_edges(
+ self, vertices: list[int], edge_pick_count: int
+ ) -> list[list[int]]:
+ assert edge_pick_count <= len(vertices)
+
+ random_source_vertices: list[int] = random.sample(
+ vertices[0 : int(len(vertices) / 2)], edge_pick_count
+ )
+ random_destination_vertices: list[int] = random.sample(
+ vertices[int(len(vertices) / 2) :], edge_pick_count
+ )
+ random_edges: list[list[int]] = []
+
+ for source in random_source_vertices:
+ for dest in random_destination_vertices:
+ random_edges.append([source, dest])
+
+ return random_edges
+
+ def __generate_graphs(
+ self, vertex_count: int, min_val: int, max_val: int, edge_pick_count: int
+ ) -> tuple[GraphAdjacencyMatrix, GraphAdjacencyMatrix, list[int], list[list[int]]]:
+ if max_val - min_val + 1 < vertex_count:
+ raise ValueError(
+ "Will result in duplicate vertices. Either increase "
+ "range between min_val and max_val or decrease vertex count"
+ )
+
+ # generate graph input
+ random_vertices: list[int] = random.sample(
+ range(min_val, max_val + 1), vertex_count
+ )
+ random_edges: list[list[int]] = self.__generate_random_edges(
+ random_vertices, edge_pick_count
+ )
+
+ # build graphs
+ undirected_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=random_edges, directed=False
+ )
+ directed_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=random_edges, directed=True
+ )
+
+ return undirected_graph, directed_graph, random_vertices, random_edges
+
+ def test_init_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ # test graph initialization with vertices and edges
+ for num in random_vertices:
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, num
+ )
+
+ for edge in random_edges:
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ assert not undirected_graph.directed
+ assert directed_graph.directed
+
+ def test_contains_vertex(self) -> None:
+ random_vertices: list[int] = random.sample(range(101), 20)
+
+ # Build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=[], directed=True
+ )
+
+ # Test contains_vertex
+ for num in range(101):
+ assert (num in random_vertices) == undirected_graph.contains_vertex(num)
+ assert (num in random_vertices) == directed_graph.contains_vertex(num)
+
+ def test_add_vertices(self) -> None:
+ random_vertices: list[int] = random.sample(range(101), 20)
+
+ # build empty graphs
+ undirected_graph: GraphAdjacencyMatrix = GraphAdjacencyMatrix(
+ vertices=[], edges=[], directed=False
+ )
+ directed_graph: GraphAdjacencyMatrix = GraphAdjacencyMatrix(
+ vertices=[], edges=[], directed=True
+ )
+
+ # run add_vertex
+ for num in random_vertices:
+ undirected_graph.add_vertex(num)
+
+ for num in random_vertices:
+ directed_graph.add_vertex(num)
+
+ # test add_vertex worked
+ for num in random_vertices:
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, num
+ )
+
+ def test_remove_vertices(self) -> None:
+ random_vertices: list[int] = random.sample(range(101), 20)
+
+ # build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=[], directed=True
+ )
+
+ # test remove_vertex worked
+ for num in random_vertices:
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, num
+ )
+
+ undirected_graph.remove_vertex(num)
+ directed_graph.remove_vertex(num)
+
+ self.__assert_graph_vertex_does_not_exist_check(
+ undirected_graph, directed_graph, num
+ )
+
+ def test_add_and_remove_vertices_repeatedly(self) -> None:
+ random_vertices1: list[int] = random.sample(range(51), 20)
+ random_vertices2: list[int] = random.sample(range(51, 101), 20)
+
+ # build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices1, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices1, edges=[], directed=True
+ )
+
+ # test adding and removing vertices
+ for i, _ in enumerate(random_vertices1):
+ undirected_graph.add_vertex(random_vertices2[i])
+ directed_graph.add_vertex(random_vertices2[i])
+
+ self.__assert_graph_vertex_exists_check(
+ undirected_graph, directed_graph, random_vertices2[i]
+ )
+
+ undirected_graph.remove_vertex(random_vertices1[i])
+ directed_graph.remove_vertex(random_vertices1[i])
+
+ self.__assert_graph_vertex_does_not_exist_check(
+ undirected_graph, directed_graph, random_vertices1[i]
+ )
+
+ # remove all vertices
+ for i, _ in enumerate(random_vertices1):
+ undirected_graph.remove_vertex(random_vertices2[i])
+ directed_graph.remove_vertex(random_vertices2[i])
+
+ self.__assert_graph_vertex_does_not_exist_check(
+ undirected_graph, directed_graph, random_vertices2[i]
+ )
+
+ def test_contains_edge(self) -> None:
+ # generate graphs and graph input
+ vertex_count = 20
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(vertex_count, 0, 100, 4)
+
+ # generate all possible edges for testing
+ all_possible_edges: list[list[int]] = []
+ for i in range(vertex_count - 1):
+ for j in range(i + 1, vertex_count):
+ all_possible_edges.append([random_vertices[i], random_vertices[j]])
+ all_possible_edges.append([random_vertices[j], random_vertices[i]])
+
+ # test contains_edge function
+ for edge in all_possible_edges:
+ if edge in random_edges:
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+ elif [edge[1], edge[0]] in random_edges:
+ # since this edge exists for undirected but the reverse may
+ # not exist for directed
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, [edge[1], edge[0]]
+ )
+ else:
+ self.__assert_graph_edge_does_not_exist_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ def test_add_edge(self) -> None:
+ # generate graph input
+ random_vertices: list[int] = random.sample(range(101), 15)
+ random_edges: list[list[int]] = self.__generate_random_edges(random_vertices, 4)
+
+ # build graphs WITHOUT edges
+ undirected_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=[], directed=False
+ )
+ directed_graph = GraphAdjacencyMatrix(
+ vertices=random_vertices, edges=[], directed=True
+ )
+
+ # run and test add_edge
+ for edge in random_edges:
+ undirected_graph.add_edge(edge[0], edge[1])
+ directed_graph.add_edge(edge[0], edge[1])
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ def test_remove_edge(self) -> None:
+ # generate graph input and graphs
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ # run and test remove_edge
+ for edge in random_edges:
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, edge
+ )
+ undirected_graph.remove_edge(edge[0], edge[1])
+ directed_graph.remove_edge(edge[0], edge[1])
+ self.__assert_graph_edge_does_not_exist_check(
+ undirected_graph, directed_graph, edge
+ )
+
+ def test_add_and_remove_edges_repeatedly(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ # make some more edge options!
+ more_random_edges: list[list[int]] = []
+
+ while len(more_random_edges) != len(random_edges):
+ edges: list[list[int]] = self.__generate_random_edges(random_vertices, 4)
+ for edge in edges:
+ if len(more_random_edges) == len(random_edges):
+ break
+ elif edge not in more_random_edges and edge not in random_edges:
+ more_random_edges.append(edge)
+
+ for i, _ in enumerate(random_edges):
+ undirected_graph.add_edge(more_random_edges[i][0], more_random_edges[i][1])
+ directed_graph.add_edge(more_random_edges[i][0], more_random_edges[i][1])
+
+ self.__assert_graph_edge_exists_check(
+ undirected_graph, directed_graph, more_random_edges[i]
+ )
+
+ undirected_graph.remove_edge(random_edges[i][0], random_edges[i][1])
+ directed_graph.remove_edge(random_edges[i][0], random_edges[i][1])
+
+ self.__assert_graph_edge_does_not_exist_check(
+ undirected_graph, directed_graph, random_edges[i]
+ )
+
+ def test_add_vertex_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for vertex in random_vertices:
+ with pytest.raises(ValueError):
+ undirected_graph.add_vertex(vertex)
+ with pytest.raises(ValueError):
+ directed_graph.add_vertex(vertex)
+
+ def test_remove_vertex_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for i in range(101):
+ if i not in random_vertices:
+ with pytest.raises(ValueError):
+ undirected_graph.remove_vertex(i)
+ with pytest.raises(ValueError):
+ directed_graph.remove_vertex(i)
+
+ def test_add_edge_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for edge in random_edges:
+ with pytest.raises(ValueError):
+ undirected_graph.add_edge(edge[0], edge[1])
+ with pytest.raises(ValueError):
+ directed_graph.add_edge(edge[0], edge[1])
+
+ def test_remove_edge_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ more_random_edges: list[list[int]] = []
+
+ while len(more_random_edges) != len(random_edges):
+ edges: list[list[int]] = self.__generate_random_edges(random_vertices, 4)
+ for edge in edges:
+ if len(more_random_edges) == len(random_edges):
+ break
+ elif edge not in more_random_edges and edge not in random_edges:
+ more_random_edges.append(edge)
+
+ for edge in more_random_edges:
+ with pytest.raises(ValueError):
+ undirected_graph.remove_edge(edge[0], edge[1])
+ with pytest.raises(ValueError):
+ directed_graph.remove_edge(edge[0], edge[1])
+
+ def test_contains_edge_exception_check(self) -> None:
+ (
+ undirected_graph,
+ directed_graph,
+ random_vertices,
+ random_edges,
+ ) = self.__generate_graphs(20, 0, 100, 4)
+
+ for vertex in random_vertices:
+ with pytest.raises(ValueError):
+ undirected_graph.contains_edge(vertex, 102)
+ with pytest.raises(ValueError):
+ directed_graph.contains_edge(vertex, 102)
+
+ with pytest.raises(ValueError):
+ undirected_graph.contains_edge(103, 102)
+ with pytest.raises(ValueError):
+ directed_graph.contains_edge(103, 102)
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/graphs/graph_list.py b/graphs/graph_list.py
index a812fecd961e..6563cbb76132 100644
--- a/graphs/graph_list.py
+++ b/graphs/graph_list.py
@@ -1,44 +1,150 @@
-#!/usr/bin/python
+#!/usr/bin/env python3
-# Author: OMKAR PATHAK
+# Author: OMKAR PATHAK, Nwachukwu Chidiebere
-# We can use Python's dictionary for constructing the graph.
+# Use a Python dictionary to construct the graph.
+from __future__ import annotations
+from pprint import pformat
+from typing import Generic, TypeVar
-class AdjacencyList:
- def __init__(self):
- self.adj_list = {}
+T = TypeVar("T")
- def add_edge(self, from_vertex: int, to_vertex: int) -> None:
- # check if vertex is already present
- if from_vertex in self.adj_list:
- self.adj_list[from_vertex].append(to_vertex)
+
+class GraphAdjacencyList(Generic[T]):
+ """
+ Adjacency List type Graph Data Structure that accounts for directed and undirected
+ Graphs. Initialize graph object indicating whether it's directed or undirected.
+
+ Directed graph example:
+ >>> d_graph = GraphAdjacencyList()
+ >>> print(d_graph)
+ {}
+ >>> d_graph.add_edge(0, 1)
+ {0: [1], 1: []}
+ >>> d_graph.add_edge(1, 2).add_edge(1, 4).add_edge(1, 5)
+ {0: [1], 1: [2, 4, 5], 2: [], 4: [], 5: []}
+ >>> d_graph.add_edge(2, 0).add_edge(2, 6).add_edge(2, 7)
+ {0: [1], 1: [2, 4, 5], 2: [0, 6, 7], 4: [], 5: [], 6: [], 7: []}
+ >>> d_graph
+ {0: [1], 1: [2, 4, 5], 2: [0, 6, 7], 4: [], 5: [], 6: [], 7: []}
+ >>> print(repr(d_graph))
+ {0: [1], 1: [2, 4, 5], 2: [0, 6, 7], 4: [], 5: [], 6: [], 7: []}
+
+ Undirected graph example:
+ >>> u_graph = GraphAdjacencyList(directed=False)
+ >>> u_graph.add_edge(0, 1)
+ {0: [1], 1: [0]}
+ >>> u_graph.add_edge(1, 2).add_edge(1, 4).add_edge(1, 5)
+ {0: [1], 1: [0, 2, 4, 5], 2: [1], 4: [1], 5: [1]}
+ >>> u_graph.add_edge(2, 0).add_edge(2, 6).add_edge(2, 7)
+ {0: [1, 2], 1: [0, 2, 4, 5], 2: [1, 0, 6, 7], 4: [1], 5: [1], 6: [2], 7: [2]}
+ >>> u_graph.add_edge(4, 5)
+ {0: [1, 2],
+ 1: [0, 2, 4, 5],
+ 2: [1, 0, 6, 7],
+ 4: [1, 5],
+ 5: [1, 4],
+ 6: [2],
+ 7: [2]}
+ >>> print(u_graph)
+ {0: [1, 2],
+ 1: [0, 2, 4, 5],
+ 2: [1, 0, 6, 7],
+ 4: [1, 5],
+ 5: [1, 4],
+ 6: [2],
+ 7: [2]}
+ >>> print(repr(u_graph))
+ {0: [1, 2],
+ 1: [0, 2, 4, 5],
+ 2: [1, 0, 6, 7],
+ 4: [1, 5],
+ 5: [1, 4],
+ 6: [2],
+ 7: [2]}
+ >>> char_graph = GraphAdjacencyList(directed=False)
+ >>> char_graph.add_edge('a', 'b')
+ {'a': ['b'], 'b': ['a']}
+ >>> char_graph.add_edge('b', 'c').add_edge('b', 'e').add_edge('b', 'f')
+ {'a': ['b'], 'b': ['a', 'c', 'e', 'f'], 'c': ['b'], 'e': ['b'], 'f': ['b']}
+ >>> char_graph
+ {'a': ['b'], 'b': ['a', 'c', 'e', 'f'], 'c': ['b'], 'e': ['b'], 'f': ['b']}
+ """
+
+ def __init__(self, directed: bool = True) -> None:
+ """
+ Parameters:
+ directed: (bool) Indicates if graph is directed or undirected. Default is True.
+ """
+
+ self.adj_list: dict[T, list[T]] = {} # dictionary of lists
+ self.directed = directed
+
+ def add_edge(
+ self, source_vertex: T, destination_vertex: T
+ ) -> GraphAdjacencyList[T]:
+ """
+ Connects vertices together. Creates and Edge from source vertex to destination
+ vertex.
+ Vertices will be created if not found in graph
+ """
+
+ if not self.directed: # For undirected graphs
+ # if both source vertex and destination vertex are both present in the
+ # adjacency list, add destination vertex to source vertex list of adjacent
+ # vertices and add source vertex to destination vertex list of adjacent
+ # vertices.
+ if source_vertex in self.adj_list and destination_vertex in self.adj_list:
+ self.adj_list[source_vertex].append(destination_vertex)
+ self.adj_list[destination_vertex].append(source_vertex)
+ # if only source vertex is present in adjacency list, add destination vertex
+ # to source vertex list of adjacent vertices, then create a new vertex with
+ # destination vertex as key and assign a list containing the source vertex
+ # as it's first adjacent vertex.
+ elif source_vertex in self.adj_list:
+ self.adj_list[source_vertex].append(destination_vertex)
+ self.adj_list[destination_vertex] = [source_vertex]
+ # if only destination vertex is present in adjacency list, add source vertex
+ # to destination vertex list of adjacent vertices, then create a new vertex
+ # with source vertex as key and assign a list containing the source vertex
+ # as it's first adjacent vertex.
+ elif destination_vertex in self.adj_list:
+ self.adj_list[destination_vertex].append(source_vertex)
+ self.adj_list[source_vertex] = [destination_vertex]
+ # if both source vertex and destination vertex are not present in adjacency
+ # list, create a new vertex with source vertex as key and assign a list
+ # containing the destination vertex as it's first adjacent vertex also
+ # create a new vertex with destination vertex as key and assign a list
+ # containing the source vertex as it's first adjacent vertex.
+ else:
+ self.adj_list[source_vertex] = [destination_vertex]
+ self.adj_list[destination_vertex] = [source_vertex]
+ # For directed graphs
+ # if both source vertex and destination vertex are present in adjacency
+ # list, add destination vertex to source vertex list of adjacent vertices.
+ elif source_vertex in self.adj_list and destination_vertex in self.adj_list:
+ self.adj_list[source_vertex].append(destination_vertex)
+ # if only source vertex is present in adjacency list, add destination
+ # vertex to source vertex list of adjacent vertices and create a new vertex
+ # with destination vertex as key, which has no adjacent vertex
+ elif source_vertex in self.adj_list:
+ self.adj_list[source_vertex].append(destination_vertex)
+ self.adj_list[destination_vertex] = []
+ # if only destination vertex is present in adjacency list, create a new
+ # vertex with source vertex as key and assign a list containing destination
+ # vertex as first adjacent vertex
+ elif destination_vertex in self.adj_list:
+ self.adj_list[source_vertex] = [destination_vertex]
+ # if both source vertex and destination vertex are not present in adjacency
+ # list, create a new vertex with source vertex as key and a list containing
+ # destination vertex as it's first adjacent vertex. Then create a new vertex
+ # with destination vertex as key, which has no adjacent vertex
else:
- self.adj_list[from_vertex] = [to_vertex]
-
- def print_list(self) -> None:
- for i in self.adj_list:
- print((i, "->", " -> ".join([str(j) for j in self.adj_list[i]])))
-
-
-if __name__ == "__main__":
- al = AdjacencyList()
- al.add_edge(0, 1)
- al.add_edge(0, 4)
- al.add_edge(4, 1)
- al.add_edge(4, 3)
- al.add_edge(1, 0)
- al.add_edge(1, 4)
- al.add_edge(1, 3)
- al.add_edge(1, 2)
- al.add_edge(2, 3)
- al.add_edge(3, 4)
-
- al.print_list()
-
- # OUTPUT:
- # 0 -> 1 -> 4
- # 1 -> 0 -> 4 -> 3 -> 2
- # 2 -> 3
- # 3 -> 4
- # 4 -> 1 -> 3
+ self.adj_list[source_vertex] = [destination_vertex]
+ self.adj_list[destination_vertex] = []
+
+ return self
+
+ def __repr__(self) -> str:
+ return pformat(self.adj_list)
diff --git a/graphs/graph_matrix.py b/graphs/graph_matrix.py
deleted file mode 100644
index 987168426ba5..000000000000
--- a/graphs/graph_matrix.py
+++ /dev/null
@@ -1,25 +0,0 @@
-class Graph:
- def __init__(self, vertex):
- self.vertex = vertex
- self.graph = [[0] * vertex for i in range(vertex)]
-
- def add_edge(self, u, v):
- self.graph[u - 1][v - 1] = 1
- self.graph[v - 1][u - 1] = 1
-
- def show(self):
-
- for i in self.graph:
- for j in i:
- print(j, end=" ")
- print(" ")
-
-
-g = Graph(100)
-
-g.add_edge(1, 4)
-g.add_edge(4, 2)
-g.add_edge(4, 5)
-g.add_edge(2, 5)
-g.add_edge(5, 3)
-g.show()
diff --git a/graphs/graphs_floyd_warshall.py b/graphs/graphs_floyd_warshall.py
index 56cf8b9e382b..aaed9ac5df8b 100644
--- a/graphs/graphs_floyd_warshall.py
+++ b/graphs/graphs_floyd_warshall.py
@@ -1,7 +1,7 @@
# floyd_warshall.py
"""
- The problem is to find the shortest distance between all pairs of vertices in a
- weighted directed graph that can have negative edge weights.
+The problem is to find the shortest distance between all pairs of vertices in a
+weighted directed graph that can have negative edge weights.
"""
diff --git a/graphs/greedy_best_first.py b/graphs/greedy_best_first.py
index 4b80a6853d3f..bb3160047e34 100644
--- a/graphs/greedy_best_first.py
+++ b/graphs/greedy_best_first.py
@@ -4,14 +4,34 @@
from __future__ import annotations
-grid = [
- [0, 0, 0, 0, 0, 0, 0],
- [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles
- [0, 0, 0, 0, 0, 0, 0],
- [0, 0, 1, 0, 0, 0, 0],
- [1, 0, 1, 0, 0, 0, 0],
- [0, 0, 0, 0, 0, 0, 0],
- [0, 0, 0, 0, 1, 0, 0],
+Path = list[tuple[int, int]]
+
+# 0's are free path whereas 1's are obstacles
+TEST_GRIDS = [
+ [
+ [0, 0, 0, 0, 0, 0, 0],
+ [0, 1, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0],
+ [0, 0, 1, 0, 0, 0, 0],
+ [1, 0, 1, 0, 0, 0, 0],
+ [0, 0, 0, 0, 0, 0, 0],
+ [0, 0, 0, 0, 1, 0, 0],
+ ],
+ [
+ [0, 0, 0, 1, 1, 0, 0],
+ [0, 0, 0, 0, 1, 0, 1],
+ [0, 0, 0, 1, 1, 0, 0],
+ [0, 1, 0, 0, 1, 0, 0],
+ [1, 0, 0, 1, 1, 0, 1],
+ [0, 0, 0, 0, 0, 0, 0],
+ ],
+ [
+ [0, 0, 1, 0, 0],
+ [0, 1, 0, 0, 0],
+ [0, 0, 1, 0, 1],
+ [1, 0, 0, 1, 1],
+ [0, 0, 0, 0, 0],
+ ],
]
delta = ([-1, 0], [0, -1], [1, 0], [0, 1]) # up, left, down, right
@@ -33,7 +53,15 @@ class Node:
True
"""
- def __init__(self, pos_x, pos_y, goal_x, goal_y, g_cost, parent):
+ def __init__(
+ self,
+ pos_x: int,
+ pos_y: int,
+ goal_x: int,
+ goal_y: int,
+ g_cost: float,
+ parent: Node | None,
+ ):
self.pos_x = pos_x
self.pos_y = pos_y
self.pos = (pos_y, pos_x)
@@ -48,17 +76,21 @@ def calculate_heuristic(self) -> float:
The heuristic here is the Manhattan Distance
Could elaborate to offer more than one choice
"""
- dy = abs(self.pos_x - self.goal_x)
- dx = abs(self.pos_y - self.goal_y)
+ dx = abs(self.pos_x - self.goal_x)
+ dy = abs(self.pos_y - self.goal_y)
return dx + dy
def __lt__(self, other) -> bool:
return self.f_cost < other.f_cost
+ def __eq__(self, other) -> bool:
+ return self.pos == other.pos
+
class GreedyBestFirst:
"""
- >>> gbf = GreedyBestFirst((0, 0), (len(grid) - 1, len(grid[0]) - 1))
+ >>> grid = TEST_GRIDS[2]
+ >>> gbf = GreedyBestFirst(grid, (0, 0), (len(grid) - 1, len(grid[0]) - 1))
>>> [x.pos for x in gbf.get_successors(gbf.start)]
[(1, 0), (0, 1)]
>>> (gbf.start.pos_y + delta[3][0], gbf.start.pos_x + delta[3][1])
@@ -68,20 +100,23 @@ class GreedyBestFirst:
>>> gbf.retrace_path(gbf.start)
[(0, 0)]
>>> gbf.search() # doctest: +NORMALIZE_WHITESPACE
- [(0, 0), (1, 0), (2, 0), (3, 0), (3, 1), (4, 1), (5, 1), (6, 1),
- (6, 2), (6, 3), (5, 3), (5, 4), (5, 5), (6, 5), (6, 6)]
+ [(0, 0), (1, 0), (2, 0), (2, 1), (3, 1), (4, 1), (4, 2), (4, 3),
+ (4, 4)]
"""
- def __init__(self, start, goal):
+ def __init__(
+ self, grid: list[list[int]], start: tuple[int, int], goal: tuple[int, int]
+ ):
+ self.grid = grid
self.start = Node(start[1], start[0], goal[1], goal[0], 0, None)
self.target = Node(goal[1], goal[0], goal[1], goal[0], 99999, None)
self.open_nodes = [self.start]
- self.closed_nodes = []
+ self.closed_nodes: list[Node] = []
self.reached = False
- def search(self) -> list[tuple[int]]:
+ def search(self) -> Path | None:
"""
Search for the path,
if a path is not found, only the starting position is returned
@@ -104,46 +139,33 @@ def search(self) -> list[tuple[int]]:
if child_node not in self.open_nodes:
self.open_nodes.append(child_node)
- else:
- # retrieve the best current path
- better_node = self.open_nodes.pop(self.open_nodes.index(child_node))
-
- if child_node.g_cost < better_node.g_cost:
- self.open_nodes.append(child_node)
- else:
- self.open_nodes.append(better_node)
- if not (self.reached):
+ if not self.reached:
return [self.start.pos]
+ return None
def get_successors(self, parent: Node) -> list[Node]:
"""
Returns a list of successors (both in the grid and free spaces)
"""
- successors = []
- for action in delta:
- pos_x = parent.pos_x + action[1]
- pos_y = parent.pos_y + action[0]
-
- if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(grid) - 1):
- continue
-
- if grid[pos_y][pos_x] != 0:
- continue
-
- successors.append(
- Node(
- pos_x,
- pos_y,
- self.target.pos_y,
- self.target.pos_x,
- parent.g_cost + 1,
- parent,
- )
+ return [
+ Node(
+ pos_x,
+ pos_y,
+ self.target.pos_x,
+ self.target.pos_y,
+ parent.g_cost + 1,
+ parent,
+ )
+ for action in delta
+ if (
+ 0 <= (pos_x := parent.pos_x + action[1]) < len(self.grid[0])
+ and 0 <= (pos_y := parent.pos_y + action[0]) < len(self.grid)
+ and self.grid[pos_y][pos_x] == 0
)
- return successors
+ ]
- def retrace_path(self, node: Node) -> list[tuple[int]]:
+ def retrace_path(self, node: Node | None) -> Path:
"""
Retrace the path from parents to parents until start node
"""
@@ -157,18 +179,21 @@ def retrace_path(self, node: Node) -> list[tuple[int]]:
if __name__ == "__main__":
- init = (0, 0)
- goal = (len(grid) - 1, len(grid[0]) - 1)
- for elem in grid:
- print(elem)
+ for idx, grid in enumerate(TEST_GRIDS):
+ print(f"==grid-{idx + 1}==")
- print("------")
+ init = (0, 0)
+ goal = (len(grid) - 1, len(grid[0]) - 1)
+ for elem in grid:
+ print(elem)
- greedy_bf = GreedyBestFirst(init, goal)
- path = greedy_bf.search()
+ print("------")
- for elem in path:
- grid[elem[0]][elem[1]] = 2
+ greedy_bf = GreedyBestFirst(grid, init, goal)
+ path = greedy_bf.search()
+ if path:
+ for pos_x, pos_y in path:
+ grid[pos_x][pos_y] = 2
- for elem in grid:
- print(elem)
+ for elem in grid:
+ print(elem)
diff --git a/graphs/greedy_min_vertex_cover.py b/graphs/greedy_min_vertex_cover.py
new file mode 100644
index 000000000000..cdef69141bd6
--- /dev/null
+++ b/graphs/greedy_min_vertex_cover.py
@@ -0,0 +1,64 @@
+"""
+* Author: Manuel Di Lullo (https://github.com/manueldilullo)
+* Description: Approximization algorithm for minimum vertex cover problem.
+ Greedy Approach. Uses graphs represented with an adjacency list
+URL: https://mathworld.wolfram.com/MinimumVertexCover.html
+URL: https://cs.stackexchange.com/questions/129017/greedy-algorithm-for-vertex-cover
+"""
+
+import heapq
+
+
+def greedy_min_vertex_cover(graph: dict) -> set[int]:
+ """
+ Greedy APX Algorithm for min Vertex Cover
+ @input: graph (graph stored in an adjacency list where each vertex
+ is represented with an integer)
+ @example:
+ >>> graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]}
+ >>> greedy_min_vertex_cover(graph)
+ {0, 1, 2, 4}
+ """
+ # queue used to store nodes and their rank
+ queue: list[list] = []
+
+ # for each node and his adjacency list add them and the rank of the node to queue
+ # using heapq module the queue will be filled like a Priority Queue
+ # heapq works with a min priority queue, so I used -1*len(v) to build it
+ for key, value in graph.items():
+ # O(log(n))
+ heapq.heappush(queue, [-1 * len(value), (key, value)])
+
+ # chosen_vertices = set of chosen vertices
+ chosen_vertices = set()
+
+ # while queue isn't empty and there are still edges
+ # (queue[0][0] is the rank of the node with max rank)
+ while queue and queue[0][0] != 0:
+ # extract vertex with max rank from queue and add it to chosen_vertices
+ argmax = heapq.heappop(queue)[1][0]
+ chosen_vertices.add(argmax)
+
+ # Remove all arcs adjacent to argmax
+ for elem in queue:
+ # if v haven't adjacent node, skip
+ if elem[0] == 0:
+ continue
+ # if argmax is reachable from elem
+ # remove argmax from elem's adjacent list and update his rank
+ if argmax in elem[1][1]:
+ index = elem[1][1].index(argmax)
+ del elem[1][1][index]
+ elem[0] += 1
+ # re-order the queue
+ heapq.heapify(queue)
+ return chosen_vertices
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]}
+ print(f"Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}")
diff --git a/graphs/kahns_algorithm_long.py b/graphs/kahns_algorithm_long.py
index fed7517a21e2..1f16b90c0745 100644
--- a/graphs/kahns_algorithm_long.py
+++ b/graphs/kahns_algorithm_long.py
@@ -1,10 +1,10 @@
# Finding longest distance in Directed Acyclic Graph using KahnsAlgorithm
-def longestDistance(graph):
+def longest_distance(graph):
indegree = [0] * len(graph)
queue = []
- longDist = [1] * len(graph)
+ long_dist = [1] * len(graph)
- for key, values in graph.items():
+ for values in graph.values():
for i in values:
indegree[i] += 1
@@ -17,15 +17,14 @@ def longestDistance(graph):
for x in graph[vertex]:
indegree[x] -= 1
- if longDist[vertex] + 1 > longDist[x]:
- longDist[x] = longDist[vertex] + 1
+ long_dist[x] = max(long_dist[x], long_dist[vertex] + 1)
if indegree[x] == 0:
queue.append(x)
- print(max(longDist))
+ print(max(long_dist))
# Adjacency list of Graph
graph = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []}
-longestDistance(graph)
+longest_distance(graph)
diff --git a/graphs/kahns_algorithm_topo.py b/graphs/kahns_algorithm_topo.py
index bf9f90299361..c956cf9f48fd 100644
--- a/graphs/kahns_algorithm_topo.py
+++ b/graphs/kahns_algorithm_topo.py
@@ -1,36 +1,61 @@
-def topologicalSort(graph):
+def topological_sort(graph: dict[int, list[int]]) -> list[int] | None:
"""
- Kahn's Algorithm is used to find Topological ordering of Directed Acyclic Graph
- using BFS
+ Perform topological sorting of a Directed Acyclic Graph (DAG)
+ using Kahn's Algorithm via Breadth-First Search (BFS).
+
+ Topological sorting is a linear ordering of vertices in a graph such that for
+ every directed edge u → v, vertex u comes before vertex v in the ordering.
+
+ Parameters:
+ graph: Adjacency list representing the directed graph where keys are
+ vertices, and values are lists of adjacent vertices.
+
+ Returns:
+ The topologically sorted order of vertices if the graph is a DAG.
+ Returns None if the graph contains a cycle.
+
+ Example:
+ >>> graph = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []}
+ >>> topological_sort(graph)
+ [0, 1, 2, 3, 4, 5]
+
+ >>> graph_with_cycle = {0: [1], 1: [2], 2: [0]}
+ >>> topological_sort(graph_with_cycle)
"""
+
indegree = [0] * len(graph)
queue = []
- topo = []
- cnt = 0
+ topo_order = []
+ processed_vertices_count = 0
- for key, values in graph.items():
+ # Calculate the indegree of each vertex
+ for values in graph.values():
for i in values:
indegree[i] += 1
+ # Add all vertices with 0 indegree to the queue
for i in range(len(indegree)):
if indegree[i] == 0:
queue.append(i)
+ # Perform BFS
while queue:
vertex = queue.pop(0)
- cnt += 1
- topo.append(vertex)
- for x in graph[vertex]:
- indegree[x] -= 1
- if indegree[x] == 0:
- queue.append(x)
-
- if cnt != len(graph):
- print("Cycle exists")
- else:
- print(topo)
-
-
-# Adjacency List of Graph
-graph = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []}
-topologicalSort(graph)
+ processed_vertices_count += 1
+ topo_order.append(vertex)
+
+ # Traverse neighbors
+ for neighbor in graph[vertex]:
+ indegree[neighbor] -= 1
+ if indegree[neighbor] == 0:
+ queue.append(neighbor)
+
+ if processed_vertices_count != len(graph):
+ return None # no topological ordering exists due to cycle
+ return topo_order # valid topological ordering
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/karger.py b/graphs/karger.py
index f72128c8178a..3ef65c0d6d32 100644
--- a/graphs/karger.py
+++ b/graphs/karger.py
@@ -47,7 +47,6 @@ def partition_graph(graph: dict[str, list[str]]) -> set[tuple[str, str]]:
graph_copy = {node: graph[node][:] for node in graph}
while len(graph_copy) > 2:
-
# Choose a random edge.
u = random.choice(list(graph_copy.keys()))
v = random.choice(graph_copy[u])
diff --git a/graphs/lanczos_eigenvectors.py b/graphs/lanczos_eigenvectors.py
new file mode 100644
index 000000000000..581a81a1127f
--- /dev/null
+++ b/graphs/lanczos_eigenvectors.py
@@ -0,0 +1,206 @@
+"""
+Lanczos Method for Finding Eigenvalues and Eigenvectors of a Graph.
+
+This module demonstrates the Lanczos method to approximate the largest eigenvalues
+and corresponding eigenvectors of a symmetric matrix represented as a graph's
+adjacency list. The method efficiently handles large, sparse matrices by converting
+the graph to a tridiagonal matrix, whose eigenvalues and eigenvectors are then
+computed.
+
+Key Functions:
+- `find_lanczos_eigenvectors`: Computes the k largest eigenvalues and vectors.
+- `lanczos_iteration`: Constructs the tridiagonal matrix and orthonormal basis vectors.
+- `multiply_matrix_vector`: Multiplies an adjacency list graph with a vector.
+
+Complexity:
+- Time: O(k * n), where k is the number of eigenvalues and n is the matrix size.
+- Space: O(n), due to sparse representation and tridiagonal matrix structure.
+
+Further Reading:
+- Lanczos Algorithm: https://en.wikipedia.org/wiki/Lanczos_algorithm
+- Eigenvector Centrality: https://en.wikipedia.org/wiki/Eigenvector_centrality
+
+Example Usage:
+Given a graph represented by an adjacency list, the `find_lanczos_eigenvectors`
+function returns the largest eigenvalues and eigenvectors. This can be used to
+analyze graph centrality.
+"""
+
+import numpy as np
+
+
+def validate_adjacency_list(graph: list[list[int | None]]) -> None:
+ """Validates the adjacency list format for the graph.
+
+ Args:
+ graph: A list of lists where each sublist contains the neighbors of a node.
+
+ Raises:
+ ValueError: If the graph is not a list of lists, or if any node has
+ invalid neighbors (e.g., out-of-range or non-integer values).
+
+ >>> validate_adjacency_list([[1, 2], [0], [0, 1]])
+ >>> validate_adjacency_list([[]]) # No neighbors, valid case
+ >>> validate_adjacency_list([[1], [2], [-1]]) # Invalid neighbor
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid neighbor -1 in node 2 adjacency list.
+ """
+ if not isinstance(graph, list):
+ raise ValueError("Graph should be a list of lists.")
+
+ for node_index, neighbors in enumerate(graph):
+ if not isinstance(neighbors, list):
+ no_neighbors_message: str = (
+ f"Node {node_index} should have a list of neighbors."
+ )
+ raise ValueError(no_neighbors_message)
+ for neighbor_index in neighbors:
+ if (
+ not isinstance(neighbor_index, int)
+ or neighbor_index < 0
+ or neighbor_index >= len(graph)
+ ):
+ invalid_neighbor_message: str = (
+ f"Invalid neighbor {neighbor_index} in node {node_index} "
+ f"adjacency list."
+ )
+ raise ValueError(invalid_neighbor_message)
+
+
+def lanczos_iteration(
+ graph: list[list[int | None]], num_eigenvectors: int
+) -> tuple[np.ndarray, np.ndarray]:
+ """Constructs the tridiagonal matrix and orthonormal basis vectors using the
+ Lanczos method.
+
+ Args:
+ graph: The graph represented as a list of adjacency lists.
+ num_eigenvectors: The number of largest eigenvalues and eigenvectors
+ to approximate.
+
+ Returns:
+ A tuple containing:
+ - tridiagonal_matrix: A (num_eigenvectors x num_eigenvectors) symmetric
+ matrix.
+ - orthonormal_basis: A (num_nodes x num_eigenvectors) matrix of orthonormal
+ basis vectors.
+
+ Raises:
+ ValueError: If num_eigenvectors is less than 1 or greater than the number of
+ nodes.
+
+ >>> graph = [[1, 2], [0, 2], [0, 1]]
+ >>> T, Q = lanczos_iteration(graph, 2)
+ >>> T.shape == (2, 2) and Q.shape == (3, 2)
+ True
+ """
+ num_nodes: int = len(graph)
+ if not (1 <= num_eigenvectors <= num_nodes):
+ raise ValueError(
+ "Number of eigenvectors must be between 1 and the number of "
+ "nodes in the graph."
+ )
+
+ orthonormal_basis: np.ndarray = np.zeros((num_nodes, num_eigenvectors))
+ tridiagonal_matrix: np.ndarray = np.zeros((num_eigenvectors, num_eigenvectors))
+
+ rng = np.random.default_rng()
+ initial_vector: np.ndarray = rng.random(num_nodes)
+ initial_vector /= np.sqrt(np.dot(initial_vector, initial_vector))
+ orthonormal_basis[:, 0] = initial_vector
+
+ prev_beta: float = 0.0
+ for iter_index in range(num_eigenvectors):
+ result_vector: np.ndarray = multiply_matrix_vector(
+ graph, orthonormal_basis[:, iter_index]
+ )
+ if iter_index > 0:
+ result_vector -= prev_beta * orthonormal_basis[:, iter_index - 1]
+ alpha_value: float = np.dot(orthonormal_basis[:, iter_index], result_vector)
+ result_vector -= alpha_value * orthonormal_basis[:, iter_index]
+
+ prev_beta = np.sqrt(np.dot(result_vector, result_vector))
+ if iter_index < num_eigenvectors - 1 and prev_beta > 1e-10:
+ orthonormal_basis[:, iter_index + 1] = result_vector / prev_beta
+ tridiagonal_matrix[iter_index, iter_index] = alpha_value
+ if iter_index < num_eigenvectors - 1:
+ tridiagonal_matrix[iter_index, iter_index + 1] = prev_beta
+ tridiagonal_matrix[iter_index + 1, iter_index] = prev_beta
+ return tridiagonal_matrix, orthonormal_basis
+
+
+def multiply_matrix_vector(
+ graph: list[list[int | None]], vector: np.ndarray
+) -> np.ndarray:
+ """Performs multiplication of a graph's adjacency list representation with a vector.
+
+ Args:
+ graph: The adjacency list of the graph.
+ vector: A 1D numpy array representing the vector to multiply.
+
+ Returns:
+ A numpy array representing the product of the adjacency list and the vector.
+
+ Raises:
+ ValueError: If the vector's length does not match the number of nodes in the
+ graph.
+
+ >>> multiply_matrix_vector([[1, 2], [0, 2], [0, 1]], np.array([1, 1, 1]))
+ array([2., 2., 2.])
+ >>> multiply_matrix_vector([[1, 2], [0, 2], [0, 1]], np.array([0, 1, 0]))
+ array([1., 0., 1.])
+ """
+ num_nodes: int = len(graph)
+ if vector.shape[0] != num_nodes:
+ raise ValueError("Vector length must match the number of nodes in the graph.")
+
+ result: np.ndarray = np.zeros(num_nodes)
+ for node_index, neighbors in enumerate(graph):
+ for neighbor_index in neighbors:
+ result[node_index] += vector[neighbor_index]
+ return result
+
+
+def find_lanczos_eigenvectors(
+ graph: list[list[int | None]], num_eigenvectors: int
+) -> tuple[np.ndarray, np.ndarray]:
+ """Computes the largest eigenvalues and their corresponding eigenvectors using the
+ Lanczos method.
+
+ Args:
+ graph: The graph as a list of adjacency lists.
+ num_eigenvectors: Number of largest eigenvalues and eigenvectors to compute.
+
+ Returns:
+ A tuple containing:
+ - eigenvalues: 1D array of the largest eigenvalues in descending order.
+ - eigenvectors: 2D array where each column is an eigenvector corresponding
+ to an eigenvalue.
+
+ Raises:
+ ValueError: If the graph format is invalid or num_eigenvectors is out of bounds.
+
+ >>> eigenvalues, eigenvectors = find_lanczos_eigenvectors(
+ ... [[1, 2], [0, 2], [0, 1]], 2
+ ... )
+ >>> len(eigenvalues) == 2 and eigenvectors.shape[1] == 2
+ True
+ """
+ validate_adjacency_list(graph)
+ tridiagonal_matrix, orthonormal_basis = lanczos_iteration(graph, num_eigenvectors)
+ eigenvalues, eigenvectors = np.linalg.eigh(tridiagonal_matrix)
+ return eigenvalues[::-1], np.dot(orthonormal_basis, eigenvectors[:, ::-1])
+
+
+def main() -> None:
+ """
+ Main driver function for testing the implementation with doctests.
+ """
+ import doctest
+
+ doctest.testmod()
+
+
+if __name__ == "__main__":
+ main()
diff --git a/other/markov_chain.py b/graphs/markov_chain.py
similarity index 99%
rename from other/markov_chain.py
rename to graphs/markov_chain.py
index b93c408cd288..0b6659822dc4 100644
--- a/other/markov_chain.py
+++ b/graphs/markov_chain.py
@@ -35,6 +35,7 @@ def transition(self, node: str) -> str:
current_probability += self.connections[node][dest]
if current_probability > random_value:
return dest
+ return ""
def get_transitions(
diff --git a/graphs/matching_min_vertex_cover.py b/graphs/matching_min_vertex_cover.py
new file mode 100644
index 000000000000..5ac944ec1a07
--- /dev/null
+++ b/graphs/matching_min_vertex_cover.py
@@ -0,0 +1,62 @@
+"""
+* Author: Manuel Di Lullo (https://github.com/manueldilullo)
+* Description: Approximization algorithm for minimum vertex cover problem.
+ Matching Approach. Uses graphs represented with an adjacency list
+
+URL: https://mathworld.wolfram.com/MinimumVertexCover.html
+URL: https://www.princeton.edu/~aaa/Public/Teaching/ORF523/ORF523_Lec6.pdf
+"""
+
+
+def matching_min_vertex_cover(graph: dict) -> set:
+ """
+ APX Algorithm for min Vertex Cover using Matching Approach
+ @input: graph (graph stored in an adjacency list where each vertex
+ is represented as an integer)
+ @example:
+ >>> graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]}
+ >>> matching_min_vertex_cover(graph)
+ {0, 1, 2, 4}
+ """
+ # chosen_vertices = set of chosen vertices
+ chosen_vertices = set()
+ # edges = list of graph's edges
+ edges = get_edges(graph)
+
+ # While there are still elements in edges list, take an arbitrary edge
+ # (from_node, to_node) and add his extremity to chosen_vertices and then
+ # remove all arcs adjacent to the from_node and to_node
+ while edges:
+ from_node, to_node = edges.pop()
+ chosen_vertices.add(from_node)
+ chosen_vertices.add(to_node)
+ for edge in edges.copy():
+ if from_node in edge or to_node in edge:
+ edges.discard(edge)
+ return chosen_vertices
+
+
+def get_edges(graph: dict) -> set:
+ """
+ Return a set of couples that represents all of the edges.
+ @input: graph (graph stored in an adjacency list where each vertex is
+ represented as an integer)
+ @example:
+ >>> graph = {0: [1, 3], 1: [0, 3], 2: [0, 3], 3: [0, 1, 2]}
+ >>> get_edges(graph)
+ {(0, 1), (3, 1), (0, 3), (2, 0), (3, 0), (2, 3), (1, 0), (3, 2), (1, 3)}
+ """
+ edges = set()
+ for from_node, to_nodes in graph.items():
+ for to_node in to_nodes:
+ edges.add((from_node, to_node))
+ return edges
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]}
+ # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
diff --git a/graphs/minimum_path_sum.py b/graphs/minimum_path_sum.py
new file mode 100644
index 000000000000..df1e545df3d0
--- /dev/null
+++ b/graphs/minimum_path_sum.py
@@ -0,0 +1,63 @@
+def min_path_sum(grid: list) -> int:
+ """
+ Find the path from top left to bottom right of array of numbers
+ with the lowest possible sum and return the sum along this path.
+ >>> min_path_sum([
+ ... [1, 3, 1],
+ ... [1, 5, 1],
+ ... [4, 2, 1],
+ ... ])
+ 7
+
+ >>> min_path_sum([
+ ... [1, 0, 5, 6, 7],
+ ... [8, 9, 0, 4, 2],
+ ... [4, 4, 4, 5, 1],
+ ... [9, 6, 3, 1, 0],
+ ... [8, 4, 3, 2, 7],
+ ... ])
+ 20
+
+ >>> min_path_sum(None)
+ Traceback (most recent call last):
+ ...
+ TypeError: The grid does not contain the appropriate information
+
+ >>> min_path_sum([[]])
+ Traceback (most recent call last):
+ ...
+ TypeError: The grid does not contain the appropriate information
+ """
+
+ if not grid or not grid[0]:
+ raise TypeError("The grid does not contain the appropriate information")
+
+ for cell_n in range(1, len(grid[0])):
+ grid[0][cell_n] += grid[0][cell_n - 1]
+ row_above = grid[0]
+
+ for row_n in range(1, len(grid)):
+ current_row = grid[row_n]
+ grid[row_n] = fill_row(current_row, row_above)
+ row_above = grid[row_n]
+
+ return grid[-1][-1]
+
+
+def fill_row(current_row: list, row_above: list) -> list:
+ """
+ >>> fill_row([2, 2, 2], [1, 2, 3])
+ [3, 4, 5]
+ """
+
+ current_row[0] += row_above[0]
+ for cell_n in range(1, len(current_row)):
+ current_row[cell_n] += min(current_row[cell_n - 1], row_above[cell_n])
+
+ return current_row
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/minimum_spanning_tree_boruvka.py b/graphs/minimum_spanning_tree_boruvka.py
index 32548b2ecb6c..f234d65ab765 100644
--- a/graphs/minimum_spanning_tree_boruvka.py
+++ b/graphs/minimum_spanning_tree_boruvka.py
@@ -4,7 +4,6 @@ class Graph:
"""
def __init__(self):
-
self.num_vertices = 0
self.num_edges = 0
self.adjacency = {}
@@ -63,7 +62,7 @@ def __str__(self):
for tail in self.adjacency:
for head in self.adjacency[tail]:
weight = self.adjacency[head][tail]
- string += "%d -> %d == %d\n" % (head, tail, weight)
+ string += f"{head} -> {tail} == {weight}\n"
return string.rstrip("\n")
def get_edges(self):
@@ -145,6 +144,7 @@ def union(self, item1, item2):
self.rank[root1] += 1
self.parent[root2] = root1
return root1
+ return None
@staticmethod
def boruvka_mst(graph):
@@ -185,12 +185,12 @@ def boruvka_mst(graph):
if cheap_edge[set2] == -1 or cheap_edge[set2][2] > weight:
cheap_edge[set2] = [head, tail, weight]
- for vertex in cheap_edge:
- if cheap_edge[vertex] != -1:
- head, tail, weight = cheap_edge[vertex]
+ for head_tail_weight in cheap_edge.values():
+ if head_tail_weight != -1:
+ head, tail, weight = head_tail_weight
if union_find.find(head) != union_find.find(tail):
union_find.union(head, tail)
- mst_edges.append(cheap_edge[vertex])
+ mst_edges.append(head_tail_weight)
num_components = num_components - 1
mst = Graph.build(edges=mst_edges)
return mst
diff --git a/graphs/minimum_spanning_tree_kruskal.py b/graphs/minimum_spanning_tree_kruskal.py
index a51f970341f7..85d937010489 100644
--- a/graphs/minimum_spanning_tree_kruskal.py
+++ b/graphs/minimum_spanning_tree_kruskal.py
@@ -1,15 +1,14 @@
-from typing import List, Tuple
-
-
-def kruskal(num_nodes: int, num_edges: int, edges: List[Tuple[int, int, int]]) -> int:
+def kruskal(
+ num_nodes: int, edges: list[tuple[int, int, int]]
+) -> list[tuple[int, int, int]]:
"""
- >>> kruskal(4, 3, [(0, 1, 3), (1, 2, 5), (2, 3, 1)])
+ >>> kruskal(4, [(0, 1, 3), (1, 2, 5), (2, 3, 1)])
[(2, 3, 1), (0, 1, 3), (1, 2, 5)]
- >>> kruskal(4, 5, [(0, 1, 3), (1, 2, 5), (2, 3, 1), (0, 2, 1), (0, 3, 2)])
+ >>> kruskal(4, [(0, 1, 3), (1, 2, 5), (2, 3, 1), (0, 2, 1), (0, 3, 2)])
[(2, 3, 1), (0, 2, 1), (0, 1, 3)]
- >>> kruskal(4, 6, [(0, 1, 3), (1, 2, 5), (2, 3, 1), (0, 2, 1), (0, 3, 2),
+ >>> kruskal(4, [(0, 1, 3), (1, 2, 5), (2, 3, 1), (0, 2, 1), (0, 3, 2),
... (2, 1, 1)])
[(2, 3, 1), (0, 2, 1), (2, 1, 1)]
"""
@@ -41,7 +40,7 @@ def find_parent(i):
edges = []
for _ in range(num_edges):
- node1, node2, cost = [int(x) for x in input().strip().split()]
+ node1, node2, cost = (int(x) for x in input().strip().split())
edges.append((node1, node2, cost))
- kruskal(num_nodes, num_edges, edges)
+ kruskal(num_nodes, edges)
diff --git a/graphs/minimum_spanning_tree_kruskal2.py b/graphs/minimum_spanning_tree_kruskal2.py
index dfb87efeb89a..0ddb43ce8e6e 100644
--- a/graphs/minimum_spanning_tree_kruskal2.py
+++ b/graphs/minimum_spanning_tree_kruskal2.py
@@ -1,78 +1,93 @@
from __future__ import annotations
+from typing import Generic, TypeVar
-class DisjointSetTreeNode:
+T = TypeVar("T")
+
+
+class DisjointSetTreeNode(Generic[T]):
# Disjoint Set Node to store the parent and rank
- def __init__(self, key: int) -> None:
- self.key = key
+ def __init__(self, data: T) -> None:
+ self.data = data
self.parent = self
self.rank = 0
-class DisjointSetTree:
+class DisjointSetTree(Generic[T]):
# Disjoint Set DataStructure
- def __init__(self):
+ def __init__(self) -> None:
# map from node name to the node object
- self.map = {}
+ self.map: dict[T, DisjointSetTreeNode[T]] = {}
- def make_set(self, x: int) -> None:
+ def make_set(self, data: T) -> None:
# create a new set with x as its member
- self.map[x] = DisjointSetTreeNode(x)
+ self.map[data] = DisjointSetTreeNode(data)
- def find_set(self, x: int) -> DisjointSetTreeNode:
+ def find_set(self, data: T) -> DisjointSetTreeNode[T]:
# find the set x belongs to (with path-compression)
- elem_ref = self.map[x]
+ elem_ref = self.map[data]
if elem_ref != elem_ref.parent:
- elem_ref.parent = self.find_set(elem_ref.parent.key)
+ elem_ref.parent = self.find_set(elem_ref.parent.data)
return elem_ref.parent
- def link(self, x: int, y: int) -> None:
+ def link(
+ self, node1: DisjointSetTreeNode[T], node2: DisjointSetTreeNode[T]
+ ) -> None:
# helper function for union operation
- if x.rank > y.rank:
- y.parent = x
+ if node1.rank > node2.rank:
+ node2.parent = node1
else:
- x.parent = y
- if x.rank == y.rank:
- y.rank += 1
+ node1.parent = node2
+ if node1.rank == node2.rank:
+ node2.rank += 1
- def union(self, x: int, y: int) -> None:
+ def union(self, data1: T, data2: T) -> None:
# merge 2 disjoint sets
- self.link(self.find_set(x), self.find_set(y))
+ self.link(self.find_set(data1), self.find_set(data2))
-class GraphUndirectedWeighted:
- def __init__(self):
+class GraphUndirectedWeighted(Generic[T]):
+ def __init__(self) -> None:
# connections: map from the node to the neighbouring nodes (with weights)
- self.connections = {}
+ self.connections: dict[T, dict[T, int]] = {}
- def add_node(self, node: int) -> None:
+ def add_node(self, node: T) -> None:
# add a node ONLY if its not present in the graph
if node not in self.connections:
self.connections[node] = {}
- def add_edge(self, node1: int, node2: int, weight: int) -> None:
+ def add_edge(self, node1: T, node2: T, weight: int) -> None:
# add an edge with the given weight
self.add_node(node1)
self.add_node(node2)
self.connections[node1][node2] = weight
self.connections[node2][node1] = weight
- def kruskal(self) -> GraphUndirectedWeighted:
+ def kruskal(self) -> GraphUndirectedWeighted[T]:
# Kruskal's Algorithm to generate a Minimum Spanning Tree (MST) of a graph
"""
Details: https://en.wikipedia.org/wiki/Kruskal%27s_algorithm
Example:
-
- >>> graph = GraphUndirectedWeighted()
- >>> graph.add_edge(1, 2, 1)
- >>> graph.add_edge(2, 3, 2)
- >>> graph.add_edge(3, 4, 1)
- >>> graph.add_edge(3, 5, 100) # Removed in MST
- >>> graph.add_edge(4, 5, 5)
- >>> assert 5 in graph.connections[3]
- >>> mst = graph.kruskal()
+ >>> g1 = GraphUndirectedWeighted[int]()
+ >>> g1.add_edge(1, 2, 1)
+ >>> g1.add_edge(2, 3, 2)
+ >>> g1.add_edge(3, 4, 1)
+ >>> g1.add_edge(3, 5, 100) # Removed in MST
+ >>> g1.add_edge(4, 5, 5)
+ >>> assert 5 in g1.connections[3]
+ >>> mst = g1.kruskal()
>>> assert 5 not in mst.connections[3]
+
+ >>> g2 = GraphUndirectedWeighted[str]()
+ >>> g2.add_edge('A', 'B', 1)
+ >>> g2.add_edge('B', 'C', 2)
+ >>> g2.add_edge('C', 'D', 1)
+ >>> g2.add_edge('C', 'E', 100) # Removed in MST
+ >>> g2.add_edge('D', 'E', 5)
+ >>> assert 'E' in g2.connections["C"]
+ >>> mst = g2.kruskal()
+ >>> assert 'E' not in mst.connections['C']
"""
# getting the edges in ascending order of weights
@@ -84,26 +99,23 @@ def kruskal(self) -> GraphUndirectedWeighted:
seen.add((end, start))
edges.append((start, end, self.connections[start][end]))
edges.sort(key=lambda x: x[2])
+
# creating the disjoint set
- disjoint_set = DisjointSetTree()
- [disjoint_set.make_set(node) for node in self.connections]
+ disjoint_set = DisjointSetTree[T]()
+ for node in self.connections:
+ disjoint_set.make_set(node)
+
# MST generation
num_edges = 0
index = 0
- graph = GraphUndirectedWeighted()
+ graph = GraphUndirectedWeighted[T]()
while num_edges < len(self.connections) - 1:
u, v, w = edges[index]
index += 1
- parentu = disjoint_set.find_set(u)
- parentv = disjoint_set.find_set(v)
- if parentu != parentv:
+ parent_u = disjoint_set.find_set(u)
+ parent_v = disjoint_set.find_set(v)
+ if parent_u != parent_v:
num_edges += 1
graph.add_edge(u, v, w)
disjoint_set.union(u, v)
return graph
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/graphs/minimum_spanning_tree_prims.py b/graphs/minimum_spanning_tree_prims.py
index 16b4286140ec..d0b45d7ef139 100644
--- a/graphs/minimum_spanning_tree_prims.py
+++ b/graphs/minimum_spanning_tree_prims.py
@@ -2,115 +2,134 @@
from collections import defaultdict
-def PrimsAlgorithm(l): # noqa: E741
+class Heap:
+ def __init__(self):
+ self.node_position = []
- nodePosition = []
+ def get_position(self, vertex):
+ return self.node_position[vertex]
- def get_position(vertex):
- return nodePosition[vertex]
+ def set_position(self, vertex, pos):
+ self.node_position[vertex] = pos
- def set_position(vertex, pos):
- nodePosition[vertex] = pos
-
- def top_to_bottom(heap, start, size, positions):
+ def top_to_bottom(self, heap, start, size, positions):
if start > size // 2 - 1:
return
else:
- if 2 * start + 2 >= size:
- m = 2 * start + 1
+ if 2 * start + 2 >= size: # noqa: SIM114
+ smallest_child = 2 * start + 1
+ elif heap[2 * start + 1] < heap[2 * start + 2]:
+ smallest_child = 2 * start + 1
else:
- if heap[2 * start + 1] < heap[2 * start + 2]:
- m = 2 * start + 1
- else:
- m = 2 * start + 2
- if heap[m] < heap[start]:
- temp, temp1 = heap[m], positions[m]
- heap[m], positions[m] = heap[start], positions[start]
+ smallest_child = 2 * start + 2
+ if heap[smallest_child] < heap[start]:
+ temp, temp1 = heap[smallest_child], positions[smallest_child]
+ heap[smallest_child], positions[smallest_child] = (
+ heap[start],
+ positions[start],
+ )
heap[start], positions[start] = temp, temp1
- temp = get_position(positions[m])
- set_position(positions[m], get_position(positions[start]))
- set_position(positions[start], temp)
+ temp = self.get_position(positions[smallest_child])
+ self.set_position(
+ positions[smallest_child], self.get_position(positions[start])
+ )
+ self.set_position(positions[start], temp)
- top_to_bottom(heap, m, size, positions)
+ self.top_to_bottom(heap, smallest_child, size, positions)
# Update function if value of any node in min-heap decreases
- def bottom_to_top(val, index, heap, position):
+ def bottom_to_top(self, val, index, heap, position):
temp = position[index]
while index != 0:
- if index % 2 == 0:
- parent = int((index - 2) / 2)
- else:
- parent = int((index - 1) / 2)
+ parent = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2)
if val < heap[parent]:
heap[index] = heap[parent]
position[index] = position[parent]
- set_position(position[parent], index)
+ self.set_position(position[parent], index)
else:
heap[index] = val
position[index] = temp
- set_position(temp, index)
+ self.set_position(temp, index)
break
index = parent
else:
heap[0] = val
position[0] = temp
- set_position(temp, 0)
+ self.set_position(temp, 0)
- def heapify(heap, positions):
+ def heapify(self, heap, positions):
start = len(heap) // 2 - 1
for i in range(start, -1, -1):
- top_to_bottom(heap, i, len(heap), positions)
+ self.top_to_bottom(heap, i, len(heap), positions)
- def deleteMinimum(heap, positions):
+ def delete_minimum(self, heap, positions):
temp = positions[0]
heap[0] = sys.maxsize
- top_to_bottom(heap, 0, len(heap), positions)
+ self.top_to_bottom(heap, 0, len(heap), positions)
return temp
- visited = [0 for i in range(len(l))]
- Nbr_TV = [-1 for i in range(len(l))] # Neighboring Tree Vertex of selected vertex
+
+def prisms_algorithm(adjacency_list):
+ """
+ >>> adjacency_list = {0: [[1, 1], [3, 3]],
+ ... 1: [[0, 1], [2, 6], [3, 5], [4, 1]],
+ ... 2: [[1, 6], [4, 5], [5, 2]],
+ ... 3: [[0, 3], [1, 5], [4, 1]],
+ ... 4: [[1, 1], [2, 5], [3, 1], [5, 4]],
+ ... 5: [[2, 2], [4, 4]]}
+ >>> prisms_algorithm(adjacency_list)
+ [(0, 1), (1, 4), (4, 3), (4, 5), (5, 2)]
+ """
+
+ heap = Heap()
+
+ visited = [0] * len(adjacency_list)
+ nbr_tv = [-1] * len(adjacency_list) # Neighboring Tree Vertex of selected vertex
# Minimum Distance of explored vertex with neighboring vertex of partial tree
# formed in graph
- Distance_TV = [] # Heap of Distance of vertices from their neighboring vertex
- Positions = []
+ distance_tv = [] # Heap of Distance of vertices from their neighboring vertex
+ positions = []
- for x in range(len(l)):
- p = sys.maxsize
- Distance_TV.append(p)
- Positions.append(x)
- nodePosition.append(x)
+ for vertex in range(len(adjacency_list)):
+ distance_tv.append(sys.maxsize)
+ positions.append(vertex)
+ heap.node_position.append(vertex)
- TreeEdges = []
+ tree_edges = []
visited[0] = 1
- Distance_TV[0] = sys.maxsize
- for x in l[0]:
- Nbr_TV[x[0]] = 0
- Distance_TV[x[0]] = x[1]
- heapify(Distance_TV, Positions)
-
- for i in range(1, len(l)):
- vertex = deleteMinimum(Distance_TV, Positions)
+ distance_tv[0] = sys.maxsize
+ for neighbor, distance in adjacency_list[0]:
+ nbr_tv[neighbor] = 0
+ distance_tv[neighbor] = distance
+ heap.heapify(distance_tv, positions)
+
+ for _ in range(1, len(adjacency_list)):
+ vertex = heap.delete_minimum(distance_tv, positions)
if visited[vertex] == 0:
- TreeEdges.append((Nbr_TV[vertex], vertex))
+ tree_edges.append((nbr_tv[vertex], vertex))
visited[vertex] = 1
- for v in l[vertex]:
- if visited[v[0]] == 0 and v[1] < Distance_TV[get_position(v[0])]:
- Distance_TV[get_position(v[0])] = v[1]
- bottom_to_top(v[1], get_position(v[0]), Distance_TV, Positions)
- Nbr_TV[v[0]] = vertex
- return TreeEdges
+ for neighbor, distance in adjacency_list[vertex]:
+ if (
+ visited[neighbor] == 0
+ and distance < distance_tv[heap.get_position(neighbor)]
+ ):
+ distance_tv[heap.get_position(neighbor)] = distance
+ heap.bottom_to_top(
+ distance, heap.get_position(neighbor), distance_tv, positions
+ )
+ nbr_tv[neighbor] = vertex
+ return tree_edges
if __name__ == "__main__": # pragma: no cover
# < --------- Prims Algorithm --------- >
- n = int(input("Enter number of vertices: ").strip())
- e = int(input("Enter number of edges: ").strip())
- adjlist = defaultdict(list)
- for x in range(e):
- l = [int(x) for x in input().strip().split()] # noqa: E741
- adjlist[l[0]].append([l[1], l[2]])
- adjlist[l[1]].append([l[0], l[2]])
- print(PrimsAlgorithm(adjlist))
+ edges_number = int(input("Enter number of edges: ").strip())
+ adjacency_list = defaultdict(list)
+ for _ in range(edges_number):
+ edge = [int(x) for x in input().strip().split()]
+ adjacency_list[edge[0]].append([edge[1], edge[2]])
+ adjacency_list[edge[1]].append([edge[0], edge[2]])
+ print(prisms_algorithm(adjacency_list))
diff --git a/graphs/minimum_spanning_tree_prims2.py b/graphs/minimum_spanning_tree_prims2.py
index 10ed736c9d17..6870cc80f844 100644
--- a/graphs/minimum_spanning_tree_prims2.py
+++ b/graphs/minimum_spanning_tree_prims2.py
@@ -7,8 +7,12 @@
connection from the tree to another vertex.
"""
+from __future__ import annotations
+
from sys import maxsize
-from typing import Dict, Optional, Tuple, Union
+from typing import Generic, TypeVar
+
+T = TypeVar("T")
def get_parent_position(position: int) -> int:
@@ -43,7 +47,7 @@ def get_child_right_position(position: int) -> int:
return (2 * position) + 2
-class MinPriorityQueue:
+class MinPriorityQueue(Generic[T]):
"""
Minimum Priority Queue Class
@@ -66,23 +70,23 @@ class MinPriorityQueue:
>>> queue.push(3, 4000)
>>> queue.push(4, 3000)
- >>> print(queue.extract_min())
+ >>> queue.extract_min()
2
>>> queue.update_key(4, 50)
- >>> print(queue.extract_min())
+ >>> queue.extract_min()
4
- >>> print(queue.extract_min())
+ >>> queue.extract_min()
1
- >>> print(queue.extract_min())
+ >>> queue.extract_min()
3
"""
def __init__(self) -> None:
- self.heap = []
- self.position_map = {}
- self.elements = 0
+ self.heap: list[tuple[T, int]] = []
+ self.position_map: dict[T, int] = {}
+ self.elements: int = 0
def __len__(self) -> int:
return self.elements
@@ -94,14 +98,14 @@ def is_empty(self) -> bool:
# Check if the priority queue is empty
return self.elements == 0
- def push(self, elem: Union[int, str], weight: int) -> None:
+ def push(self, elem: T, weight: int) -> None:
# Add an element with given priority to the queue
self.heap.append((elem, weight))
self.position_map[elem] = self.elements
self.elements += 1
self._bubble_up(elem)
- def extract_min(self) -> Union[int, str]:
+ def extract_min(self) -> T:
# Remove and return the element with lowest weight (highest priority)
if self.elements > 1:
self._swap_nodes(0, self.elements - 1)
@@ -113,7 +117,7 @@ def extract_min(self) -> Union[int, str]:
self._bubble_down(bubble_down_elem)
return elem
- def update_key(self, elem: Union[int, str], weight: int) -> None:
+ def update_key(self, elem: T, weight: int) -> None:
# Update the weight of the given key
position = self.position_map[elem]
self.heap[position] = (elem, weight)
@@ -127,21 +131,21 @@ def update_key(self, elem: Union[int, str], weight: int) -> None:
else:
self._bubble_down(elem)
- def _bubble_up(self, elem: Union[int, str]) -> None:
+ def _bubble_up(self, elem: T) -> None:
# Place a node at the proper position (upward movement) [to be used internally
# only]
curr_pos = self.position_map[elem]
if curr_pos == 0:
- return
+ return None
parent_position = get_parent_position(curr_pos)
_, weight = self.heap[curr_pos]
_, parent_weight = self.heap[parent_position]
if parent_weight > weight:
self._swap_nodes(parent_position, curr_pos)
return self._bubble_up(elem)
- return
+ return None
- def _bubble_down(self, elem: Union[int, str]) -> None:
+ def _bubble_down(self, elem: T) -> None:
# Place a node at the proper position (downward movement) [to be used
# internally only]
curr_pos = self.position_map[elem]
@@ -151,24 +155,22 @@ def _bubble_down(self, elem: Union[int, str]) -> None:
if child_left_position < self.elements and child_right_position < self.elements:
_, child_left_weight = self.heap[child_left_position]
_, child_right_weight = self.heap[child_right_position]
- if child_right_weight < child_left_weight:
- if child_right_weight < weight:
- self._swap_nodes(child_right_position, curr_pos)
- return self._bubble_down(elem)
+ if child_right_weight < child_left_weight and child_right_weight < weight:
+ self._swap_nodes(child_right_position, curr_pos)
+ return self._bubble_down(elem)
if child_left_position < self.elements:
_, child_left_weight = self.heap[child_left_position]
if child_left_weight < weight:
self._swap_nodes(child_left_position, curr_pos)
return self._bubble_down(elem)
else:
- return
+ return None
if child_right_position < self.elements:
_, child_right_weight = self.heap[child_right_position]
if child_right_weight < weight:
self._swap_nodes(child_right_position, curr_pos)
return self._bubble_down(elem)
- else:
- return
+ return None
def _swap_nodes(self, node1_pos: int, node2_pos: int) -> None:
# Swap the nodes at the given positions
@@ -182,7 +184,7 @@ def _swap_nodes(self, node1_pos: int, node2_pos: int) -> None:
self.position_map[node2_elem] = node1_pos
-class GraphUndirectedWeighted:
+class GraphUndirectedWeighted(Generic[T]):
"""
Graph Undirected Weighted Class
@@ -192,8 +194,8 @@ class GraphUndirectedWeighted:
"""
def __init__(self) -> None:
- self.connections = {}
- self.nodes = 0
+ self.connections: dict[T, dict[T, int]] = {}
+ self.nodes: int = 0
def __repr__(self) -> str:
return str(self.connections)
@@ -201,15 +203,13 @@ def __repr__(self) -> str:
def __len__(self) -> int:
return self.nodes
- def add_node(self, node: Union[int, str]) -> None:
+ def add_node(self, node: T) -> None:
# Add a node in the graph if it is not in the graph
if node not in self.connections:
self.connections[node] = {}
self.nodes += 1
- def add_edge(
- self, node1: Union[int, str], node2: Union[int, str], weight: int
- ) -> None:
+ def add_edge(self, node1: T, node2: T, weight: int) -> None:
# Add an edge between 2 nodes in the graph
self.add_node(node1)
self.add_node(node2)
@@ -218,8 +218,8 @@ def add_edge(
def prims_algo(
- graph: GraphUndirectedWeighted,
-) -> Tuple[Dict[str, int], Dict[str, Optional[str]]]:
+ graph: GraphUndirectedWeighted[T],
+) -> tuple[dict[T, int], dict[T, T | None]]:
"""
>>> graph = GraphUndirectedWeighted()
@@ -239,10 +239,13 @@ def prims_algo(
13
"""
# prim's algorithm for minimum spanning tree
- dist = {node: maxsize for node in graph.connections}
- parent = {node: None for node in graph.connections}
- priority_queue = MinPriorityQueue()
- [priority_queue.push(node, weight) for node, weight in dist.items()]
+ dist: dict[T, int] = dict.fromkeys(graph.connections, maxsize)
+ parent: dict[T, T | None] = dict.fromkeys(graph.connections)
+
+ priority_queue: MinPriorityQueue[T] = MinPriorityQueue()
+ for node, weight in dist.items():
+ priority_queue.push(node, weight)
+
if priority_queue.is_empty():
return dist, parent
@@ -254,6 +257,7 @@ def prims_algo(
dist[neighbour] = dist[node] + graph.connections[node][neighbour]
priority_queue.update_key(neighbour, dist[neighbour])
parent[neighbour] = node
+
# running prim's algorithm
while not priority_queue.is_empty():
node = priority_queue.extract_min()
@@ -263,9 +267,3 @@ def prims_algo(
priority_queue.update_key(neighbour, dist[neighbour])
parent[neighbour] = node
return dist, parent
-
-
-if __name__ == "__main__":
- from doctest import testmod
-
- testmod()
diff --git a/graphs/multi_heuristic_astar.py b/graphs/multi_heuristic_astar.py
index 77ca5760d5f0..38b07e1ca675 100644
--- a/graphs/multi_heuristic_astar.py
+++ b/graphs/multi_heuristic_astar.py
@@ -1,7 +1,10 @@
import heapq
+import sys
import numpy as np
+TPos = tuple[int, int]
+
class PriorityQueue:
def __init__(self):
@@ -30,7 +33,7 @@ def put(self, item, priority):
temp.append((pri, x))
(pri, x) = heapq.heappop(self.elements)
temp.append((priority, item))
- for (pro, xxx) in temp:
+ for pro, xxx in temp:
heapq.heappush(self.elements, (pro, xxx))
def remove_element(self, item):
@@ -41,7 +44,7 @@ def remove_element(self, item):
while x != item:
temp.append((pro, x))
(pro, x) = heapq.heappop(self.elements)
- for (prito, yyy) in temp:
+ for prito, yyy in temp:
heapq.heappush(self.elements, (prito, yyy))
def top_show(self):
@@ -53,30 +56,30 @@ def get(self):
return (priority, item)
-def consistent_heuristic(P, goal):
+def consistent_heuristic(p: TPos, goal: TPos):
# euclidean distance
- a = np.array(P)
+ a = np.array(p)
b = np.array(goal)
return np.linalg.norm(a - b)
-def heuristic_2(P, goal):
+def heuristic_2(p: TPos, goal: TPos):
# integer division by time variable
- return consistent_heuristic(P, goal) // t
+ return consistent_heuristic(p, goal) // t
-def heuristic_1(P, goal):
+def heuristic_1(p: TPos, goal: TPos):
# manhattan distance
- return abs(P[0] - goal[0]) + abs(P[1] - goal[1])
+ return abs(p[0] - goal[0]) + abs(p[1] - goal[1])
-def key(start, i, goal, g_function):
+def key(start: TPos, i: int, goal: TPos, g_function: dict[TPos, float]):
ans = g_function[start] + W1 * heuristics[i](start, goal)
return ans
def do_something(back_pointer, goal, start):
- grid = np.chararray((n, n))
+ grid = np.char.chararray((n, n))
for i in range(n):
for j in range(n):
grid[i][j] = "*"
@@ -114,15 +117,13 @@ def do_something(back_pointer, goal, start):
print(x, end=" ")
x = back_pointer[x]
print(x)
- quit()
+ sys.exit()
-def valid(p):
+def valid(p: TPos):
if p[0] < 0 or p[0] > n - 1:
return False
- if p[1] < 0 or p[1] > n - 1:
- return False
- return True
+ return not (p[1] < 0 or p[1] > n - 1)
def expand_state(
@@ -215,7 +216,6 @@ def make_common_ground():
(18, 1),
(19, 1),
]
-blocks_no = []
blocks_all = make_common_ground()
@@ -233,7 +233,7 @@ def make_common_ground():
t = 1
-def multi_a_star(start, goal, n_heuristic):
+def multi_a_star(start: TPos, goal: TPos, n_heuristic: int):
g_function = {start: 0, goal: float("inf")}
back_pointer = {start: -1, goal: -1}
open_list = []
@@ -243,8 +243,8 @@ def multi_a_star(start, goal, n_heuristic):
open_list.append(PriorityQueue())
open_list[i].put(start, key(start, i, goal, g_function))
- close_list_anchor = []
- close_list_inad = []
+ close_list_anchor: list[int] = []
+ close_list_inad: list[int] = []
while open_list[0].minkey() < float("inf"):
for i in range(1, n_heuristic):
# print(open_list[0].minkey(), open_list[i].minkey())
@@ -268,24 +268,23 @@ def multi_a_star(start, goal, n_heuristic):
back_pointer,
)
close_list_inad.append(get_s)
+ elif g_function[goal] <= open_list[0].minkey():
+ if g_function[goal] < float("inf"):
+ do_something(back_pointer, goal, start)
else:
- if g_function[goal] <= open_list[0].minkey():
- if g_function[goal] < float("inf"):
- do_something(back_pointer, goal, start)
- else:
- get_s = open_list[0].top_show()
- visited.add(get_s)
- expand_state(
- get_s,
- 0,
- visited,
- g_function,
- close_list_anchor,
- close_list_inad,
- open_list,
- back_pointer,
- )
- close_list_anchor.append(get_s)
+ get_s = open_list[0].top_show()
+ visited.add(get_s)
+ expand_state(
+ get_s,
+ 0,
+ visited,
+ g_function,
+ close_list_anchor,
+ close_list_inad,
+ open_list,
+ back_pointer,
+ )
+ close_list_anchor.append(get_s)
print("No path found to goal")
print()
for i in range(n - 1, -1, -1):
diff --git a/graphs/page_rank.py b/graphs/page_rank.py
index 0f5129146ddf..c0ce3a94c76b 100644
--- a/graphs/page_rank.py
+++ b/graphs/page_rank.py
@@ -1,6 +1,7 @@
"""
Author: https://github.com/bhushan-borole
"""
+
"""
The input graph for the algorithm is:
@@ -27,7 +28,7 @@ def add_outbound(self, node):
self.outbound.append(node)
def __repr__(self):
- return f"Node {self.name}: Inbound: {self.inbound} ; Outbound: {self.outbound}"
+ return f""
def page_rank(nodes, limit=3, d=0.85):
@@ -41,9 +42,9 @@ def page_rank(nodes, limit=3, d=0.85):
for i in range(limit):
print(f"======= Iteration {i + 1} =======")
- for j, node in enumerate(nodes):
+ for _, node in enumerate(nodes):
ranks[node.name] = (1 - d) + d * sum(
- [ranks[ib] / outbounds[ib] for ib in node.inbound]
+ ranks[ib] / outbounds[ib] for ib in node.inbound
)
print(ranks)
diff --git a/graphs/prim.py b/graphs/prim.py
index 70329da7e8e2..5b3ce04441ec 100644
--- a/graphs/prim.py
+++ b/graphs/prim.py
@@ -1,19 +1,19 @@
"""Prim's Algorithm.
- Determines the minimum spanning tree(MST) of a graph using the Prim's Algorithm.
+Determines the minimum spanning tree(MST) of a graph using the Prim's Algorithm.
- Details: https://en.wikipedia.org/wiki/Prim%27s_algorithm
+Details: https://en.wikipedia.org/wiki/Prim%27s_algorithm
"""
import heapq as hq
import math
-from typing import Iterator
+from collections.abc import Iterator
class Vertex:
"""Class Vertex."""
- def __init__(self, id):
+ def __init__(self, id_):
"""
Arguments:
id - input an id to identify the vertex
@@ -21,7 +21,7 @@ def __init__(self, id):
neighbors - a list of the vertices it is linked to
edges - a dict to store the edges's weight
"""
- self.id = str(id)
+ self.id = str(id_)
self.key = None
self.pi = None
self.neighbors = []
diff --git a/graphs/random_graph_generator.py b/graphs/random_graph_generator.py
new file mode 100644
index 000000000000..0e7e18bc8fd9
--- /dev/null
+++ b/graphs/random_graph_generator.py
@@ -0,0 +1,67 @@
+"""
+* Author: Manuel Di Lullo (https://github.com/manueldilullo)
+* Description: Random graphs generator.
+ Uses graphs represented with an adjacency list.
+
+URL: https://en.wikipedia.org/wiki/Random_graph
+"""
+
+import random
+
+
+def random_graph(
+ vertices_number: int, probability: float, directed: bool = False
+) -> dict:
+ """
+ Generate a random graph
+ @input: vertices_number (number of vertices),
+ probability (probability that a generic edge (u,v) exists),
+ directed (if True: graph will be a directed graph,
+ otherwise it will be an undirected graph)
+ @examples:
+ >>> random.seed(1)
+ >>> random_graph(4, 0.5)
+ {0: [1], 1: [0, 2, 3], 2: [1, 3], 3: [1, 2]}
+ >>> random.seed(1)
+ >>> random_graph(4, 0.5, True)
+ {0: [1], 1: [2, 3], 2: [3], 3: []}
+ """
+ graph: dict = {i: [] for i in range(vertices_number)}
+
+ # if probability is greater or equal than 1, then generate a complete graph
+ if probability >= 1:
+ return complete_graph(vertices_number)
+ # if probability is lower or equal than 0, then return a graph without edges
+ if probability <= 0:
+ return graph
+
+ # for each couple of nodes, add an edge from u to v
+ # if the number randomly generated is greater than probability probability
+ for i in range(vertices_number):
+ for j in range(i + 1, vertices_number):
+ if random.random() < probability:
+ graph[i].append(j)
+ if not directed:
+ # if the graph is undirected, add an edge in from j to i, either
+ graph[j].append(i)
+ return graph
+
+
+def complete_graph(vertices_number: int) -> dict:
+ """
+ Generate a complete graph with vertices_number vertices.
+ @input: vertices_number (number of vertices),
+ directed (False if the graph is undirected, True otherwise)
+ @example:
+ >>> complete_graph(3)
+ {0: [1, 2], 1: [0, 2], 2: [0, 1]}
+ """
+ return {
+ i: [j for j in range(vertices_number) if i != j] for i in range(vertices_number)
+ }
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/graphs/scc_kosaraju.py b/graphs/scc_kosaraju.py
index 573c1bf5e363..39211c64b687 100644
--- a/graphs/scc_kosaraju.py
+++ b/graphs/scc_kosaraju.py
@@ -1,25 +1,28 @@
+from __future__ import annotations
+
+
def dfs(u):
- global g, r, scc, component, visit, stack
+ global graph, reversed_graph, scc, component, visit, stack
if visit[u]:
return
visit[u] = True
- for v in g[u]:
+ for v in graph[u]:
dfs(v)
stack.append(u)
def dfs2(u):
- global g, r, scc, component, visit, stack
+ global graph, reversed_graph, scc, component, visit, stack
if visit[u]:
return
visit[u] = True
component.append(u)
- for v in r[u]:
+ for v in reversed_graph[u]:
dfs2(v)
def kosaraju():
- global g, r, scc, component, visit, stack
+ global graph, reversed_graph, scc, component, visit, stack
for i in range(n):
dfs(i)
visit = [False] * n
@@ -36,16 +39,16 @@ def kosaraju():
# n - no of nodes, m - no of edges
n, m = list(map(int, input().strip().split()))
- g = [[] for i in range(n)] # graph
- r = [[] for i in range(n)] # reversed graph
+ graph: list[list[int]] = [[] for _ in range(n)] # graph
+ reversed_graph: list[list[int]] = [[] for i in range(n)] # reversed graph
# input graph data (edges)
- for i in range(m):
+ for _ in range(m):
u, v = list(map(int, input().strip().split()))
- g[u].append(v)
- r[v].append(u)
+ graph[u].append(v)
+ reversed_graph[v].append(u)
- stack = []
- visit = [False] * n
- scc = []
- component = []
+ stack: list[int] = []
+ visit: list[bool] = [False] * n
+ scc: list[int] = []
+ component: list[int] = []
print(kosaraju())
diff --git a/graphs/strongly_connected_components.py b/graphs/strongly_connected_components.py
index d469df0c625b..4d4cf88035b5 100644
--- a/graphs/strongly_connected_components.py
+++ b/graphs/strongly_connected_components.py
@@ -10,7 +10,9 @@
test_graph_2 = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]}
-def topology_sort(graph: dict, vert: int, visited: list) -> list:
+def topology_sort(
+ graph: dict[int, list[int]], vert: int, visited: list[bool]
+) -> list[int]:
"""
Use depth first search to sort graph
At this time graph is the same as input
@@ -32,9 +34,11 @@ def topology_sort(graph: dict, vert: int, visited: list) -> list:
return order
-def find_components(reversed_graph: dict, vert: int, visited: list) -> list:
+def find_components(
+ reversed_graph: dict[int, list[int]], vert: int, visited: list[bool]
+) -> list[int]:
"""
- Use depth first search to find strongliy connected
+ Use depth first search to find strongly connected
vertices. Now graph is reversed
>>> find_components({0: [1], 1: [2], 2: [0]}, 0, 5 * [False])
[0, 1, 2]
@@ -52,7 +56,7 @@ def find_components(reversed_graph: dict, vert: int, visited: list) -> list:
return component
-def strongly_connected_components(graph: dict) -> list:
+def strongly_connected_components(graph: dict[int, list[int]]) -> list[list[int]]:
"""
This function takes graph as a parameter
and then returns the list of strongly connected components
@@ -63,7 +67,7 @@ def strongly_connected_components(graph: dict) -> list:
"""
visited = len(graph) * [False]
- reversed_graph = {vert: [] for vert in range(len(graph))}
+ reversed_graph: dict[int, list[int]] = {vert: [] for vert in range(len(graph))}
for vert, neighbours in graph.items():
for neighbour in neighbours:
@@ -84,9 +88,3 @@ def strongly_connected_components(graph: dict) -> list:
components_list.append(component)
return components_list
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/graphs/tarjans_scc.py b/graphs/tarjans_scc.py
index 30f8ca8a204f..b4a3bd5c4c35 100644
--- a/graphs/tarjans_scc.py
+++ b/graphs/tarjans_scc.py
@@ -1,7 +1,7 @@
from collections import deque
-def tarjan(g):
+def tarjan(g: list[list[int]]) -> list[list[int]]:
"""
Tarjan's algo for finding strongly connected components in a directed graph
@@ -19,15 +19,30 @@ def tarjan(g):
Complexity: strong_connect() is called at most once for each node and has a
complexity of O(|E|) as it is DFS.
Therefore this has complexity O(|V| + |E|) for a graph G = (V, E)
+
+ >>> tarjan([[2, 3, 4], [2, 3, 4], [0, 1, 3], [0, 1, 2], [1]])
+ [[4, 3, 1, 2, 0]]
+ >>> tarjan([[], [], [], []])
+ [[0], [1], [2], [3]]
+ >>> a = [0, 1, 2, 3, 4, 5, 4]
+ >>> b = [1, 0, 3, 2, 5, 4, 0]
+ >>> n = 7
+ >>> sorted(tarjan(create_graph(n, list(zip(a, b))))) == sorted(
+ ... tarjan(create_graph(n, list(zip(a[::-1], b[::-1])))))
+ True
+ >>> a = [0, 1, 2, 3, 4, 5, 6]
+ >>> b = [0, 1, 2, 3, 4, 5, 6]
+ >>> sorted(tarjan(create_graph(n, list(zip(a, b)))))
+ [[0], [1], [2], [3], [4], [5], [6]]
"""
n = len(g)
- stack = deque()
+ stack: deque[int] = deque()
on_stack = [False for _ in range(n)]
index_of = [-1 for _ in range(n)]
lowlink_of = index_of[:]
- def strong_connect(v, index, components):
+ def strong_connect(v: int, index: int, components: list[list[int]]) -> int:
index_of[v] = index # the number when this node is seen
lowlink_of[v] = index # lowest rank node reachable from here
index += 1
@@ -57,7 +72,7 @@ def strong_connect(v, index, components):
components.append(component)
return index
- components = []
+ components: list[list[int]] = []
for v in range(n):
if index_of[v] == -1:
strong_connect(v, 0, components)
@@ -65,8 +80,16 @@ def strong_connect(v, index, components):
return components
-def create_graph(n, edges):
- g = [[] for _ in range(n)]
+def create_graph(n: int, edges: list[tuple[int, int]]) -> list[list[int]]:
+ """
+ >>> n = 7
+ >>> source = [0, 0, 1, 2, 3, 3, 4, 4, 6]
+ >>> target = [1, 3, 2, 0, 1, 4, 5, 6, 5]
+ >>> edges = list(zip(source, target))
+ >>> create_graph(n, edges)
+ [[1, 3], [2], [0], [1, 4], [5, 6], [], [5]]
+ """
+ g: list[list[int]] = [[] for _ in range(n)]
for u, v in edges:
g[u].append(v)
return g
@@ -77,7 +100,7 @@ def create_graph(n, edges):
n_vertices = 7
source = [0, 0, 1, 2, 3, 3, 4, 4, 6]
target = [1, 3, 2, 0, 1, 4, 5, 6, 5]
- edges = [(u, v) for u, v in zip(source, target)]
+ edges = list(zip(source, target))
g = create_graph(n_vertices, edges)
- assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
+ assert tarjan(g) == [[5], [6], [4], [3, 2, 1, 0]]
diff --git a/graphs/tests/__init__.py b/graphs/tests/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/graphs/tests/test_min_spanning_tree_kruskal.py b/graphs/tests/test_min_spanning_tree_kruskal.py
index 3a527aef384f..d6df242ec6d1 100644
--- a/graphs/tests/test_min_spanning_tree_kruskal.py
+++ b/graphs/tests/test_min_spanning_tree_kruskal.py
@@ -2,7 +2,7 @@
def test_kruskal_successful_result():
- num_nodes, num_edges = 9, 14
+ num_nodes = 9
edges = [
[0, 1, 4],
[0, 7, 8],
@@ -20,7 +20,7 @@ def test_kruskal_successful_result():
[1, 7, 11],
]
- result = kruskal(num_nodes, num_edges, edges)
+ result = kruskal(num_nodes, edges)
expected = [
[7, 6, 1],
diff --git a/graphs/tests/test_min_spanning_tree_prim.py b/graphs/tests/test_min_spanning_tree_prim.py
index 048fbf595fa6..66e5706dadb1 100644
--- a/graphs/tests/test_min_spanning_tree_prim.py
+++ b/graphs/tests/test_min_spanning_tree_prim.py
@@ -1,6 +1,6 @@
from collections import defaultdict
-from graphs.minimum_spanning_tree_prims import PrimsAlgorithm as mst
+from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst
def test_prim_successful_result():
@@ -22,12 +22,12 @@ def test_prim_successful_result():
[1, 7, 11],
]
- adjancency = defaultdict(list)
+ adjacency = defaultdict(list)
for node1, node2, cost in edges:
- adjancency[node1].append([node2, cost])
- adjancency[node2].append([node1, cost])
+ adjacency[node1].append([node2, cost])
+ adjacency[node2].append([node1, cost])
- result = mst(adjancency)
+ result = mst(adjacency)
expected = [
[7, 6, 1],
diff --git a/greedy_methods/__init__.py b/greedy_methods/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/greedy_methods/best_time_to_buy_and_sell_stock.py b/greedy_methods/best_time_to_buy_and_sell_stock.py
new file mode 100644
index 000000000000..4aea19172ece
--- /dev/null
+++ b/greedy_methods/best_time_to_buy_and_sell_stock.py
@@ -0,0 +1,42 @@
+"""
+Given a list of stock prices calculate the maximum profit that can be made from a
+single buy and sell of one share of stock. We only allowed to complete one buy
+transaction and one sell transaction but must buy before we sell.
+
+Example : prices = [7, 1, 5, 3, 6, 4]
+max_profit will return 5 - which is by buying at price 1 and selling at price 6.
+
+This problem can be solved using the concept of "GREEDY ALGORITHM".
+
+We iterate over the price array once, keeping track of the lowest price point
+(buy) and the maximum profit we can get at each point. The greedy choice at each point
+is to either buy at the current price if it's less than our current buying price, or
+sell at the current price if the profit is more than our current maximum profit.
+"""
+
+
+def max_profit(prices: list[int]) -> int:
+ """
+ >>> max_profit([7, 1, 5, 3, 6, 4])
+ 5
+ >>> max_profit([7, 6, 4, 3, 1])
+ 0
+ """
+ if not prices:
+ return 0
+
+ min_price = prices[0]
+ max_profit: int = 0
+
+ for price in prices:
+ min_price = min(price, min_price)
+ max_profit = max(price - min_price, max_profit)
+
+ return max_profit
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(max_profit([7, 1, 5, 3, 6, 4]))
diff --git a/greedy_methods/fractional_cover_problem.py b/greedy_methods/fractional_cover_problem.py
new file mode 100644
index 000000000000..e37c363f1db9
--- /dev/null
+++ b/greedy_methods/fractional_cover_problem.py
@@ -0,0 +1,102 @@
+# https://en.wikipedia.org/wiki/Set_cover_problem
+
+from dataclasses import dataclass
+from operator import attrgetter
+
+
+@dataclass
+class Item:
+ weight: int
+ value: int
+
+ @property
+ def ratio(self) -> float:
+ """
+ Return the value-to-weight ratio for the item.
+
+ Returns:
+ float: The value-to-weight ratio for the item.
+
+ Examples:
+ >>> Item(10, 65).ratio
+ 6.5
+
+ >>> Item(20, 100).ratio
+ 5.0
+
+ >>> Item(30, 120).ratio
+ 4.0
+ """
+ return self.value / self.weight
+
+
+def fractional_cover(items: list[Item], capacity: int) -> float:
+ """
+ Solve the Fractional Cover Problem.
+
+ Args:
+ items: A list of items, where each item has weight and value attributes.
+ capacity: The maximum weight capacity of the knapsack.
+
+ Returns:
+ The maximum value that can be obtained by selecting fractions of items to cover
+ the knapsack's capacity.
+
+ Raises:
+ ValueError: If capacity is negative.
+
+ Examples:
+ >>> fractional_cover((Item(10, 60), Item(20, 100), Item(30, 120)), capacity=50)
+ 240.0
+
+ >>> fractional_cover([Item(20, 100), Item(30, 120), Item(10, 60)], capacity=25)
+ 135.0
+
+ >>> fractional_cover([Item(10, 60), Item(20, 100), Item(30, 120)], capacity=60)
+ 280.0
+
+ >>> fractional_cover(items=[Item(5, 30), Item(10, 60), Item(15, 90)], capacity=30)
+ 180.0
+
+ >>> fractional_cover(items=[], capacity=50)
+ 0.0
+
+ >>> fractional_cover(items=[Item(10, 60)], capacity=5)
+ 30.0
+
+ >>> fractional_cover(items=[Item(10, 60)], capacity=1)
+ 6.0
+
+ >>> fractional_cover(items=[Item(10, 60)], capacity=0)
+ 0.0
+
+ >>> fractional_cover(items=[Item(10, 60)], capacity=-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Capacity cannot be negative
+ """
+ if capacity < 0:
+ raise ValueError("Capacity cannot be negative")
+
+ total_value = 0.0
+ remaining_capacity = capacity
+
+ # Sort the items by their value-to-weight ratio in descending order
+ for item in sorted(items, key=attrgetter("ratio"), reverse=True):
+ if remaining_capacity == 0:
+ break
+
+ weight_taken = min(item.weight, remaining_capacity)
+ total_value += weight_taken * item.ratio
+ remaining_capacity -= weight_taken
+
+ return total_value
+
+
+if __name__ == "__main__":
+ import doctest
+
+ if result := doctest.testmod().failed:
+ print(f"{result} test(s) failed")
+ else:
+ print("All tests passed")
diff --git a/greedy_methods/fractional_knapsack.py b/greedy_methods/fractional_knapsack.py
new file mode 100644
index 000000000000..d52b56f23569
--- /dev/null
+++ b/greedy_methods/fractional_knapsack.py
@@ -0,0 +1,51 @@
+from bisect import bisect
+from itertools import accumulate
+
+
+def frac_knapsack(vl, wt, w, n):
+ """
+ >>> frac_knapsack([60, 100, 120], [10, 20, 30], 50, 3)
+ 240.0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 10, 4)
+ 105.0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 8, 4)
+ 95.0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6], 8, 4)
+ 60.0
+ >>> frac_knapsack([10, 40, 30], [5, 4, 6, 3], 8, 4)
+ 60.0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 0, 4)
+ 0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 8, 0)
+ 95.0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], -8, 4)
+ 0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 8, -4)
+ 95.0
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 800, 4)
+ 130
+ >>> frac_knapsack([10, 40, 30, 50], [5, 4, 6, 3], 8, 400)
+ 95.0
+ >>> frac_knapsack("ABCD", [5, 4, 6, 3], 8, 400)
+ Traceback (most recent call last):
+ ...
+ TypeError: unsupported operand type(s) for /: 'str' and 'int'
+ """
+
+ r = sorted(zip(vl, wt), key=lambda x: x[0] / x[1], reverse=True)
+ vl, wt = [i[0] for i in r], [i[1] for i in r]
+ acc = list(accumulate(wt))
+ k = bisect(acc, w)
+ return (
+ 0
+ if k == 0
+ else sum(vl[:k]) + (w - acc[k - 1]) * (vl[k]) / (wt[k])
+ if k != n
+ else sum(vl[:k])
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/dynamic_programming/fractional_knapsack_2.py b/greedy_methods/fractional_knapsack_2.py
similarity index 68%
rename from dynamic_programming/fractional_knapsack_2.py
rename to greedy_methods/fractional_knapsack_2.py
index cae57738311b..6d9ed2ec3b6b 100644
--- a/dynamic_programming/fractional_knapsack_2.py
+++ b/greedy_methods/fractional_knapsack_2.py
@@ -1,60 +1,53 @@
-# https://en.wikipedia.org/wiki/Continuous_knapsack_problem
-# https://www.guru99.com/fractional-knapsack-problem-greedy.html
-# https://medium.com/walkinthecode/greedy-algorithm-fractional-knapsack-problem-9aba1daecc93
-
-from __future__ import annotations
-
-
-def fractional_knapsack(
- value: list[int], weight: list[int], capacity: int
-) -> tuple[int, list[int]]:
- """
- >>> value = [1, 3, 5, 7, 9]
- >>> weight = [0.9, 0.7, 0.5, 0.3, 0.1]
- >>> fractional_knapsack(value, weight, 5)
- (25, [1, 1, 1, 1, 1])
- >>> fractional_knapsack(value, weight, 15)
- (25, [1, 1, 1, 1, 1])
- >>> fractional_knapsack(value, weight, 25)
- (25, [1, 1, 1, 1, 1])
- >>> fractional_knapsack(value, weight, 26)
- (25, [1, 1, 1, 1, 1])
- >>> fractional_knapsack(value, weight, -1)
- (-90.0, [0, 0, 0, 0, -10.0])
- >>> fractional_knapsack([1, 3, 5, 7], weight, 30)
- (16, [1, 1, 1, 1])
- >>> fractional_knapsack(value, [0.9, 0.7, 0.5, 0.3, 0.1], 30)
- (25, [1, 1, 1, 1, 1])
- >>> fractional_knapsack([], [], 30)
- (0, [])
- """
- index = list(range(len(value)))
- ratio = [v / w for v, w in zip(value, weight)]
- index.sort(key=lambda i: ratio[i], reverse=True)
-
- max_value = 0
- fractions = [0] * len(value)
- for i in index:
- if weight[i] <= capacity:
- fractions[i] = 1
- max_value += value[i]
- capacity -= weight[i]
- else:
- fractions[i] = capacity / weight[i]
- max_value += value[i] * capacity / weight[i]
- break
-
- return max_value, fractions
-
-
-if __name__ == "__main__":
- n = int(input("Enter number of items: "))
- value = input(f"Enter the values of the {n} item(s) in order: ").split()
- value = [int(v) for v in value]
- weight = input(f"Enter the positive weights of the {n} item(s) in order: ".split())
- weight = [int(w) for w in weight]
- capacity = int(input("Enter maximum weight: "))
-
- max_value, fractions = fractional_knapsack(value, weight, capacity)
- print("The maximum value of items that can be carried:", max_value)
- print("The fractions in which the items should be taken:", fractions)
+# https://en.wikipedia.org/wiki/Continuous_knapsack_problem
+# https://www.guru99.com/fractional-knapsack-problem-greedy.html
+# https://medium.com/walkinthecode/greedy-algorithm-fractional-knapsack-problem-9aba1daecc93
+
+from __future__ import annotations
+
+
+def fractional_knapsack(
+ value: list[int], weight: list[int], capacity: int
+) -> tuple[float, list[float]]:
+ """
+ >>> value = [1, 3, 5, 7, 9]
+ >>> weight = [0.9, 0.7, 0.5, 0.3, 0.1]
+ >>> fractional_knapsack(value, weight, 5)
+ (25, [1, 1, 1, 1, 1])
+ >>> fractional_knapsack(value, weight, 15)
+ (25, [1, 1, 1, 1, 1])
+ >>> fractional_knapsack(value, weight, 25)
+ (25, [1, 1, 1, 1, 1])
+ >>> fractional_knapsack(value, weight, 26)
+ (25, [1, 1, 1, 1, 1])
+ >>> fractional_knapsack(value, weight, -1)
+ (-90.0, [0, 0, 0, 0, -10.0])
+ >>> fractional_knapsack([1, 3, 5, 7], weight, 30)
+ (16, [1, 1, 1, 1])
+ >>> fractional_knapsack(value, [0.9, 0.7, 0.5, 0.3, 0.1], 30)
+ (25, [1, 1, 1, 1, 1])
+ >>> fractional_knapsack([], [], 30)
+ (0, [])
+ """
+ index = list(range(len(value)))
+ ratio = [v / w for v, w in zip(value, weight)]
+ index.sort(key=lambda i: ratio[i], reverse=True)
+
+ max_value: float = 0
+ fractions: list[float] = [0] * len(value)
+ for i in index:
+ if weight[i] <= capacity:
+ fractions[i] = 1
+ max_value += value[i]
+ capacity -= weight[i]
+ else:
+ fractions[i] = capacity / weight[i]
+ max_value += value[i] * capacity / weight[i]
+ break
+
+ return max_value, fractions
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/greedy_methods/gas_station.py b/greedy_methods/gas_station.py
new file mode 100644
index 000000000000..6391ce379329
--- /dev/null
+++ b/greedy_methods/gas_station.py
@@ -0,0 +1,98 @@
+"""
+Task:
+There are n gas stations along a circular route, where the amount of gas
+at the ith station is gas_quantities[i].
+
+You have a car with an unlimited gas tank and it costs costs[i] of gas
+to travel from the ith station to its next (i + 1)th station.
+You begin the journey with an empty tank at one of the gas stations.
+
+Given two integer arrays gas_quantities and costs, return the starting
+gas station's index if you can travel around the circuit once
+in the clockwise direction otherwise, return -1.
+If there exists a solution, it is guaranteed to be unique
+
+Reference: https://leetcode.com/problems/gas-station/description
+
+Implementation notes:
+First, check whether the total gas is enough to complete the journey. If not, return -1.
+However, if there is enough gas, it is guaranteed that there is a valid
+starting index to reach the end of the journey.
+Greedily calculate the net gain (gas_quantity - cost) at each station.
+If the net gain ever goes below 0 while iterating through the stations,
+start checking from the next station.
+
+"""
+
+from dataclasses import dataclass
+
+
+@dataclass
+class GasStation:
+ gas_quantity: int
+ cost: int
+
+
+def get_gas_stations(
+ gas_quantities: list[int], costs: list[int]
+) -> tuple[GasStation, ...]:
+ """
+ This function returns a tuple of gas stations.
+
+ Args:
+ gas_quantities: Amount of gas available at each station
+ costs: The cost of gas required to move from one station to the next
+
+ Returns:
+ A tuple of gas stations
+
+ >>> gas_stations = get_gas_stations([1, 2, 3, 4, 5], [3, 4, 5, 1, 2])
+ >>> len(gas_stations)
+ 5
+ >>> gas_stations[0]
+ GasStation(gas_quantity=1, cost=3)
+ >>> gas_stations[-1]
+ GasStation(gas_quantity=5, cost=2)
+ """
+ return tuple(
+ GasStation(quantity, cost) for quantity, cost in zip(gas_quantities, costs)
+ )
+
+
+def can_complete_journey(gas_stations: tuple[GasStation, ...]) -> int:
+ """
+ This function returns the index from which to start the journey
+ in order to reach the end.
+
+ Args:
+ gas_quantities [list]: Amount of gas available at each station
+ cost [list]: The cost of gas required to move from one station to the next
+
+ Returns:
+ start [int]: start index needed to complete the journey
+
+ Examples:
+ >>> can_complete_journey(get_gas_stations([1, 2, 3, 4, 5], [3, 4, 5, 1, 2]))
+ 3
+ >>> can_complete_journey(get_gas_stations([2, 3, 4], [3, 4, 3]))
+ -1
+ """
+ total_gas = sum(gas_station.gas_quantity for gas_station in gas_stations)
+ total_cost = sum(gas_station.cost for gas_station in gas_stations)
+ if total_gas < total_cost:
+ return -1
+
+ start = 0
+ net = 0
+ for i, gas_station in enumerate(gas_stations):
+ net += gas_station.gas_quantity - gas_station.cost
+ if net < 0:
+ start = i + 1
+ net = 0
+ return start
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/greedy_methods/minimum_coin_change.py b/greedy_methods/minimum_coin_change.py
new file mode 100644
index 000000000000..db2c381bc84a
--- /dev/null
+++ b/greedy_methods/minimum_coin_change.py
@@ -0,0 +1,100 @@
+"""
+Test cases:
+Do you want to enter your denominations ? (Y/N) :N
+Enter the change you want to make in Indian Currency: 987
+Following is minimal change for 987 :
+500 100 100 100 100 50 20 10 5 2
+
+Do you want to enter your denominations ? (Y/N) :Y
+Enter number of denomination:10
+1
+5
+10
+20
+50
+100
+200
+500
+1000
+2000
+Enter the change you want to make: 18745
+Following is minimal change for 18745 :
+2000 2000 2000 2000 2000 2000 2000 2000 2000 500 200 20 20 5
+
+Do you want to enter your denominations ? (Y/N) :N
+Enter the change you want to make: 0
+The total value cannot be zero or negative.
+Do you want to enter your denominations ? (Y/N) :N
+Enter the change you want to make: -98
+The total value cannot be zero or negative.
+
+Do you want to enter your denominations ? (Y/N) :Y
+Enter number of denomination:5
+1
+5
+100
+500
+1000
+Enter the change you want to make: 456
+Following is minimal change for 456 :
+100 100 100 100 5 5 5 5 5 5 5 5 5 5 5 1
+"""
+
+
+def find_minimum_change(denominations: list[int], value: str) -> list[int]:
+ """
+ Find the minimum change from the given denominations and value
+ >>> find_minimum_change([1, 5, 10, 20, 50, 100, 200, 500, 1000,2000], 18745)
+ [2000, 2000, 2000, 2000, 2000, 2000, 2000, 2000, 2000, 500, 200, 20, 20, 5]
+ >>> find_minimum_change([1, 2, 5, 10, 20, 50, 100, 500, 2000], 987)
+ [500, 100, 100, 100, 100, 50, 20, 10, 5, 2]
+ >>> find_minimum_change([1, 2, 5, 10, 20, 50, 100, 500, 2000], 0)
+ []
+ >>> find_minimum_change([1, 2, 5, 10, 20, 50, 100, 500, 2000], -98)
+ []
+ >>> find_minimum_change([1, 5, 100, 500, 1000], 456)
+ [100, 100, 100, 100, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 1]
+ """
+ total_value = int(value)
+
+ # Initialize Result
+ answer = []
+
+ # Traverse through all denomination
+ for denomination in reversed(denominations):
+ # Find denominations
+ while int(total_value) >= int(denomination):
+ total_value -= int(denomination)
+ answer.append(denomination) # Append the "answers" array
+
+ return answer
+
+
+# Driver Code
+if __name__ == "__main__":
+ denominations = []
+ value = "0"
+
+ if (
+ input("Do you want to enter your denominations ? (yY/n): ").strip().lower()
+ == "y"
+ ):
+ n = int(input("Enter the number of denominations you want to add: ").strip())
+
+ for i in range(n):
+ denominations.append(int(input(f"Denomination {i}: ").strip()))
+ value = input("Enter the change you want to make in Indian Currency: ").strip()
+ else:
+ # All denominations of Indian Currency if user does not enter
+ denominations = [1, 2, 5, 10, 20, 50, 100, 500, 2000]
+ value = input("Enter the change you want to make: ").strip()
+
+ if int(value) == 0 or int(value) < 0:
+ print("The total value cannot be zero or negative.")
+
+ else:
+ print(f"Following is minimal change for {value}: ")
+ answer = find_minimum_change(denominations, value)
+ # Print result
+ for i in range(len(answer)):
+ print(answer[i], end=" ")
diff --git a/greedy_methods/minimum_waiting_time.py b/greedy_methods/minimum_waiting_time.py
new file mode 100644
index 000000000000..aaae8cf8f720
--- /dev/null
+++ b/greedy_methods/minimum_waiting_time.py
@@ -0,0 +1,48 @@
+"""
+Calculate the minimum waiting time using a greedy algorithm.
+reference: https://www.youtube.com/watch?v=Sf3eiO12eJs
+
+For doctests run following command:
+python -m doctest -v minimum_waiting_time.py
+
+The minimum_waiting_time function uses a greedy algorithm to calculate the minimum
+time for queries to complete. It sorts the list in non-decreasing order, calculates
+the waiting time for each query by multiplying its position in the list with the
+sum of all remaining query times, and returns the total waiting time. A doctest
+ensures that the function produces the correct output.
+"""
+
+
+def minimum_waiting_time(queries: list[int]) -> int:
+ """
+ This function takes a list of query times and returns the minimum waiting time
+ for all queries to be completed.
+
+ Args:
+ queries: A list of queries measured in picoseconds
+
+ Returns:
+ total_waiting_time: Minimum waiting time measured in picoseconds
+
+ Examples:
+ >>> minimum_waiting_time([3, 2, 1, 2, 6])
+ 17
+ >>> minimum_waiting_time([3, 2, 1])
+ 4
+ >>> minimum_waiting_time([1, 2, 3, 4])
+ 10
+ >>> minimum_waiting_time([5, 5, 5, 5])
+ 30
+ >>> minimum_waiting_time([])
+ 0
+ """
+ n = len(queries)
+ if n in (0, 1):
+ return 0
+ return sum(query * (n - i - 1) for i, query in enumerate(sorted(queries)))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/greedy_methods/optimal_merge_pattern.py b/greedy_methods/optimal_merge_pattern.py
new file mode 100644
index 000000000000..a1c934f84498
--- /dev/null
+++ b/greedy_methods/optimal_merge_pattern.py
@@ -0,0 +1,56 @@
+"""
+This is a pure Python implementation of the greedy-merge-sort algorithm
+reference: https://www.geeksforgeeks.org/optimal-file-merge-patterns/
+
+For doctests run following command:
+python3 -m doctest -v greedy_merge_sort.py
+
+Objective
+Merge a set of sorted files of different length into a single sorted file.
+We need to find an optimal solution, where the resultant file
+will be generated in minimum time.
+
+Approach
+If the number of sorted files are given, there are many ways
+to merge them into a single sorted file.
+This merge can be performed pair wise.
+To merge a m-record file and a n-record file requires possibly m+n record moves
+the optimal choice being,
+merge the two smallest files together at each step (greedy approach).
+"""
+
+
+def optimal_merge_pattern(files: list) -> float:
+ """Function to merge all the files with optimum cost
+
+ Args:
+ files [list]: A list of sizes of different files to be merged
+
+ Returns:
+ optimal_merge_cost [int]: Optimal cost to merge all those files
+
+ Examples:
+ >>> optimal_merge_pattern([2, 3, 4])
+ 14
+ >>> optimal_merge_pattern([5, 10, 20, 30, 30])
+ 205
+ >>> optimal_merge_pattern([8, 8, 8, 8, 8])
+ 96
+ """
+ optimal_merge_cost = 0
+ while len(files) > 1:
+ temp = 0
+ # Consider two files with minimum cost to be merged
+ for _ in range(2):
+ min_index = files.index(min(files))
+ temp += files[min_index]
+ files.pop(min_index)
+ files.append(temp)
+ optimal_merge_cost += temp
+ return optimal_merge_cost
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/greedy_methods/smallest_range.py b/greedy_methods/smallest_range.py
new file mode 100644
index 000000000000..9adb12bf9029
--- /dev/null
+++ b/greedy_methods/smallest_range.py
@@ -0,0 +1,72 @@
+"""
+smallest_range function takes a list of sorted integer lists and finds the smallest
+range that includes at least one number from each list, using a min heap for efficiency.
+"""
+
+from heapq import heappop, heappush
+from sys import maxsize
+
+
+def smallest_range(nums: list[list[int]]) -> list[int]:
+ """
+ Find the smallest range from each list in nums.
+
+ Uses min heap for efficiency. The range includes at least one number from each list.
+
+ Args:
+ `nums`: List of k sorted integer lists.
+
+ Returns:
+ list: Smallest range as a two-element list.
+
+ Examples:
+
+ >>> smallest_range([[4, 10, 15, 24, 26], [0, 9, 12, 20], [5, 18, 22, 30]])
+ [20, 24]
+ >>> smallest_range([[1, 2, 3], [1, 2, 3], [1, 2, 3]])
+ [1, 1]
+ >>> smallest_range(((1, 2, 3), (1, 2, 3), (1, 2, 3)))
+ [1, 1]
+ >>> smallest_range(((-3, -2, -1), (0, 0, 0), (1, 2, 3)))
+ [-1, 1]
+ >>> smallest_range([[1, 2, 3], [4, 5, 6], [7, 8, 9]])
+ [3, 7]
+ >>> smallest_range([[0, 0, 0], [0, 0, 0], [0, 0, 0]])
+ [0, 0]
+ >>> smallest_range([[], [], []])
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ """
+
+ min_heap: list[tuple[int, int, int]] = []
+ current_max = -maxsize - 1
+
+ for i, items in enumerate(nums):
+ heappush(min_heap, (items[0], i, 0))
+ current_max = max(current_max, items[0])
+
+ # Initialize smallest_range with large integer values
+ smallest_range = [-maxsize - 1, maxsize]
+
+ while min_heap:
+ current_min, list_index, element_index = heappop(min_heap)
+
+ if current_max - current_min < smallest_range[1] - smallest_range[0]:
+ smallest_range = [current_min, current_max]
+
+ if element_index == len(nums[list_index]) - 1:
+ break
+
+ next_element = nums[list_index][element_index + 1]
+ heappush(min_heap, (next_element, list_index, element_index + 1))
+ current_max = max(current_max, next_element)
+
+ return smallest_range
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ print(f"{smallest_range([[1, 2, 3], [1, 2, 3], [1, 2, 3]])}") # Output: [1, 1]
diff --git a/hashes/README.md b/hashes/README.md
new file mode 100644
index 000000000000..0237260eaa67
--- /dev/null
+++ b/hashes/README.md
@@ -0,0 +1,17 @@
+# Hashes
+Hashing is the process of mapping any amount of data to a specified size using an algorithm. This is known as a hash value (or, if you're feeling fancy, a hash code, hash sums, or even a hash digest). Hashing is a one-way function, whereas encryption is a two-way function. While it is functionally conceivable to reverse-hash stuff, the required computing power makes it impractical. Hashing is a one-way street.
+Unlike encryption, which is intended to protect data in transit, hashing is intended to authenticate that a file or piece of data has not been altered—that it is authentic. In other words, it functions as a checksum.
+
+## Common hashing algorithms
+### MD5
+This is one of the first algorithms that has gained widespread acceptance. MD5 is hashing algorithm made by Ray Rivest that is known to suffer vulnerabilities. It was created in 1992 as the successor to MD4. Currently MD6 is in the works, but as of 2009 Rivest had removed it from NIST consideration for SHA-3.
+
+### SHA
+SHA stands for Security Hashing Algorithm and it’s probably best known as the hashing algorithm used in most SSL/TLS cipher suites. A cipher suite is a collection of ciphers and algorithms that are used for SSL/TLS connections. SHA handles the hashing aspects. SHA-1, as we mentioned earlier, is now deprecated. SHA-2 is now mandatory. SHA-2 is sometimes known as SHA-256, though variants with longer bit lengths are also available.
+
+### SHA256
+SHA 256 is a member of the SHA 2 algorithm family, under which SHA stands for Secure Hash Algorithm. It was a collaborative effort between both the NSA and NIST to implement a successor to the SHA 1 family, which was beginning to lose potency against brute force attacks. It was published in 2001.
+The importance of the 256 in the name refers to the final hash digest value, i.e. the hash value will remain 256 bits regardless of the size of the plaintext/cleartext. Other algorithms in the SHA family are similar to SHA 256 in some ways.
+
+### Luhn
+The Luhn algorithm, also renowned as the modulus 10 or mod 10 algorithm, is a straightforward checksum formula used to validate a wide range of identification numbers, including credit card numbers, IMEI numbers, and Canadian Social Insurance Numbers. A community of mathematicians developed the LUHN formula in the late 1960s. Companies offering credit cards quickly followed suit. Since the algorithm is in the public interest, anyone can use it. The algorithm is used by most credit cards and many government identification numbers as a simple method of differentiating valid figures from mistyped or otherwise incorrect numbers. It was created to guard against unintentional errors, not malicious attacks.
diff --git a/hashes/adler32.py b/hashes/adler32.py
index fad747abe3c3..38d76ab12aa0 100644
--- a/hashes/adler32.py
+++ b/hashes/adler32.py
@@ -1,18 +1,20 @@
"""
- Adler-32 is a checksum algorithm which was invented by Mark Adler in 1995.
- Compared to a cyclic redundancy check of the same length, it trades reliability for
- speed (preferring the latter).
- Adler-32 is more reliable than Fletcher-16, and slightly less reliable than
- Fletcher-32.[2]
+Adler-32 is a checksum algorithm which was invented by Mark Adler in 1995.
+Compared to a cyclic redundancy check of the same length, it trades reliability for
+speed (preferring the latter).
+Adler-32 is more reliable than Fletcher-16, and slightly less reliable than
+Fletcher-32.[2]
- source: https://en.wikipedia.org/wiki/Adler-32
+source: https://en.wikipedia.org/wiki/Adler-32
"""
+MOD_ADLER = 65521
-def adler32(plain_text: str) -> str:
+
+def adler32(plain_text: str) -> int:
"""
Function implements adler-32 hash.
- Itterates and evaluates new value for each character
+ Iterates and evaluates a new value for each character
>>> adler32('Algorithms')
363791387
@@ -20,7 +22,6 @@ def adler32(plain_text: str) -> str:
>>> adler32('go adler em all')
708642122
"""
- MOD_ADLER = 65521
a = 1
b = 0
for plain_chr in plain_text:
diff --git a/hashes/chaos_machine.py b/hashes/chaos_machine.py
index 1bdf984b68de..d2fde2f5e371 100644
--- a/hashes/chaos_machine.py
+++ b/hashes/chaos_machine.py
@@ -6,7 +6,8 @@
m = 5
# Buffer Space (with Parameters Space)
-buffer_space, params_space = [], []
+buffer_space: list[float] = []
+params_space: list[float] = []
# Machine Time
machine_time = 0
@@ -42,17 +43,17 @@ def pull():
global buffer_space, params_space, machine_time, K, m, t
# PRNG (Xorshift by George Marsaglia)
- def xorshift(X, Y):
- X ^= Y >> 13
- Y ^= X << 17
- X ^= Y >> 5
- return X
+ def xorshift(x, y):
+ x ^= y >> 13
+ y ^= x << 17
+ x ^= y >> 5
+ return x
# Choosing Dynamical Systems (Increment)
key = machine_time % m
# Evolution (Time Length)
- for i in range(0, t):
+ for _ in range(t):
# Variables (Position + Parameters)
r = params_space[key]
value = buffer_space[key]
@@ -62,13 +63,13 @@ def xorshift(X, Y):
params_space[key] = (machine_time * 0.01 + r * 1.01) % 1 + 3
# Choosing Chaotic Data
- X = int(buffer_space[(key + 2) % m] * (10 ** 10))
- Y = int(buffer_space[(key - 2) % m] * (10 ** 10))
+ x = int(buffer_space[(key + 2) % m] * (10**10))
+ y = int(buffer_space[(key - 2) % m] * (10**10))
# Machine Time
machine_time += 1
- return xorshift(X, Y) % 0xFFFFFFFF
+ return xorshift(x, y) % 0xFFFFFFFF
def reset():
@@ -95,7 +96,7 @@ def reset():
# Pulling Data (Output)
while inp in ("e", "E"):
- print("%s" % format(pull(), "#04x"))
+ print(f"{format(pull(), '#04x')}")
print(buffer_space)
print(params_space)
inp = input("(e)exit? ").strip()
diff --git a/hashes/djb2.py b/hashes/djb2.py
index 2d1c9aabb1fb..4c84635098f2 100644
--- a/hashes/djb2.py
+++ b/hashes/djb2.py
@@ -29,7 +29,7 @@ def djb2(s: str) -> int:
>>> djb2('scramble bits')
1609059040
"""
- hash = 5381
+ hash_value = 5381
for x in s:
- hash = ((hash << 5) + hash) + ord(x)
- return hash & 0xFFFFFFFF
+ hash_value = ((hash_value << 5) + hash_value) + ord(x)
+ return hash_value & 0xFFFFFFFF
diff --git a/hashes/elf.py b/hashes/elf.py
new file mode 100644
index 000000000000..e4bfcec22c22
--- /dev/null
+++ b/hashes/elf.py
@@ -0,0 +1,21 @@
+def elf_hash(data: str) -> int:
+ """
+ Implementation of ElfHash Algorithm, a variant of PJW hash function.
+
+ >>> elf_hash('lorem ipsum')
+ 253956621
+ """
+ hash_ = x = 0
+ for letter in data:
+ hash_ = (hash_ << 4) + ord(letter)
+ x = hash_ & 0xF0000000
+ if x != 0:
+ hash_ ^= x >> 24
+ hash_ &= ~x
+ return hash_
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/hashes/enigma_machine.py b/hashes/enigma_machine.py
index 5420bacc1409..0da8e4113de9 100644
--- a/hashes/enigma_machine.py
+++ b/hashes/enigma_machine.py
@@ -1,8 +1,8 @@
alphabets = [chr(i) for i in range(32, 126)]
-gear_one = [i for i in range(len(alphabets))]
-gear_two = [i for i in range(len(alphabets))]
-gear_three = [i for i in range(len(alphabets))]
-reflector = [i for i in reversed(range(len(alphabets)))]
+gear_one = list(range(len(alphabets)))
+gear_two = list(range(len(alphabets)))
+gear_three = list(range(len(alphabets)))
+reflector = list(reversed(range(len(alphabets))))
code = []
gear_one_pos = gear_two_pos = gear_three_pos = 0
@@ -15,12 +15,12 @@ def rotator():
gear_one.append(i)
del gear_one[0]
gear_one_pos += 1
- if gear_one_pos % int(len(alphabets)) == 0:
+ if gear_one_pos % len(alphabets) == 0:
i = gear_two[0]
gear_two.append(i)
del gear_two[0]
gear_two_pos += 1
- if gear_two_pos % int(len(alphabets)) == 0:
+ if gear_two_pos % len(alphabets) == 0:
i = gear_three[0]
gear_three.append(i)
del gear_three[0]
@@ -41,20 +41,19 @@ def engine(input_character):
if __name__ == "__main__":
- decode = input("Type your message:\n")
- decode = list(decode)
+ decode = list(input("Type your message:\n"))
while True:
try:
token = int(input("Please set token:(must be only digits)\n"))
break
except Exception as error:
print(error)
- for i in range(token):
+ for _ in range(token):
rotator()
- for i in decode:
- engine(i)
+ for j in decode:
+ engine(j)
print("\n" + "".join(code))
print(
f"\nYour Token is {token} please write it down.\nIf you want to decode "
- f"this message again you should input same digits as token!"
+ "this message again you should input same digits as token!"
)
diff --git a/hashes/fletcher16.py b/hashes/fletcher16.py
new file mode 100644
index 000000000000..add8e185bc06
--- /dev/null
+++ b/hashes/fletcher16.py
@@ -0,0 +1,36 @@
+"""
+The Fletcher checksum is an algorithm for computing a position-dependent
+checksum devised by John G. Fletcher (1934-2012) at Lawrence Livermore Labs
+in the late 1970s.[1] The objective of the Fletcher checksum was to
+provide error-detection properties approaching those of a cyclic
+redundancy check but with the lower computational effort associated
+with summation techniques.
+
+Source: https://en.wikipedia.org/wiki/Fletcher%27s_checksum
+"""
+
+
+def fletcher16(text: str) -> int:
+ """
+ Loop through every character in the data and add to two sums.
+
+ >>> fletcher16('hello world')
+ 6752
+ >>> fletcher16('onethousandfourhundredthirtyfour')
+ 28347
+ >>> fletcher16('The quick brown fox jumps over the lazy dog.')
+ 5655
+ """
+ data = bytes(text, "ascii")
+ sum1 = 0
+ sum2 = 0
+ for character in data:
+ sum1 = (sum1 + character) % 255
+ sum2 = (sum1 + sum2) % 255
+ return (sum2 << 8) | sum1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/hashes/hamming_code.py b/hashes/hamming_code.py
index 4a32bae1a51c..b3095852ac51 100644
--- a/hashes/hamming_code.py
+++ b/hashes/hamming_code.py
@@ -4,44 +4,44 @@
# Black: True
"""
- * This code implement the Hamming code:
- https://en.wikipedia.org/wiki/Hamming_code - In telecommunication,
- Hamming codes are a family of linear error-correcting codes. Hamming
- codes can detect up to two-bit errors or correct one-bit errors
- without detection of uncorrected errors. By contrast, the simple
- parity code cannot correct errors, and can detect only an odd number
- of bits in error. Hamming codes are perfect codes, that is, they
- achieve the highest possible rate for codes with their block length
- and minimum distance of three.
-
- * the implemented code consists of:
- * a function responsible for encoding the message (emitterConverter)
- * return the encoded message
- * a function responsible for decoding the message (receptorConverter)
- * return the decoded message and a ack of data integrity
-
- * how to use:
- to be used you must declare how many parity bits (sizePari)
- you want to include in the message.
- it is desired (for test purposes) to select a bit to be set
- as an error. This serves to check whether the code is working correctly.
- Lastly, the variable of the message/word that must be desired to be
- encoded (text).
-
- * how this work:
- declaration of variables (sizePari, be, text)
-
- converts the message/word (text) to binary using the
- text_to_bits function
- encodes the message using the rules of hamming encoding
- decodes the message using the rules of hamming encoding
- print the original message, the encoded message and the
- decoded message
-
- forces an error in the coded text variable
- decodes the message that was forced the error
- print the original message, the encoded message, the bit changed
- message and the decoded message
+* This code implement the Hamming code:
+ https://en.wikipedia.org/wiki/Hamming_code - In telecommunication,
+Hamming codes are a family of linear error-correcting codes. Hamming
+codes can detect up to two-bit errors or correct one-bit errors
+without detection of uncorrected errors. By contrast, the simple
+parity code cannot correct errors, and can detect only an odd number
+of bits in error. Hamming codes are perfect codes, that is, they
+achieve the highest possible rate for codes with their block length
+and minimum distance of three.
+
+* the implemented code consists of:
+ * a function responsible for encoding the message (emitterConverter)
+ * return the encoded message
+ * a function responsible for decoding the message (receptorConverter)
+ * return the decoded message and a ack of data integrity
+
+* how to use:
+ to be used you must declare how many parity bits (sizePari)
+ you want to include in the message.
+ it is desired (for test purposes) to select a bit to be set
+ as an error. This serves to check whether the code is working correctly.
+ Lastly, the variable of the message/word that must be desired to be
+ encoded (text).
+
+* how this work:
+ declaration of variables (sizePari, be, text)
+
+ converts the message/word (text) to binary using the
+ text_to_bits function
+ encodes the message using the rules of hamming encoding
+ decodes the message using the rules of hamming encoding
+ print the original message, the encoded message and the
+ decoded message
+
+ forces an error in the coded text variable
+ decodes the message that was forced the error
+ print the original message, the encoded message, the bit changed
+ message and the decoded message
"""
# Imports
@@ -68,177 +68,173 @@ def text_from_bits(bits, encoding="utf-8", errors="surrogatepass"):
# Functions of hamming code-------------------------------------------
-def emitterConverter(sizePar, data):
+def emitter_converter(size_par, data):
"""
- :param sizePar: how many parity bits the message must have
+ :param size_par: how many parity bits the message must have
:param data: information bits
:return: message to be transmitted by unreliable medium
- bits of information merged with parity bits
- >>> emitterConverter(4, "101010111111")
+ >>> emitter_converter(4, "101010111111")
['1', '1', '1', '1', '0', '1', '0', '0', '1', '0', '1', '1', '1', '1', '1', '1']
+ >>> emitter_converter(5, "101010111111")
+ Traceback (most recent call last):
+ ...
+ ValueError: size of parity don't match with size of data
"""
- if sizePar + len(data) <= 2 ** sizePar - (len(data) - 1):
- print("ERROR - size of parity don't match with size of data")
- exit(0)
+ if size_par + len(data) <= 2**size_par - (len(data) - 1):
+ raise ValueError("size of parity don't match with size of data")
- dataOut = []
+ data_out = []
parity = []
- binPos = [bin(x)[2:] for x in range(1, sizePar + len(data) + 1)]
+ bin_pos = [bin(x)[2:] for x in range(1, size_par + len(data) + 1)]
# sorted information data for the size of the output data
- dataOrd = []
+ data_ord = []
# data position template + parity
- dataOutGab = []
+ data_out_gab = []
# parity bit counter
- qtdBP = 0
+ qtd_bp = 0
# counter position of data bits
- contData = 0
+ cont_data = 0
- for x in range(1, sizePar + len(data) + 1):
+ for x in range(1, size_par + len(data) + 1):
# Performs a template of bit positions - who should be given,
# and who should be parity
- if qtdBP < sizePar:
+ if qtd_bp < size_par:
if (np.log(x) / np.log(2)).is_integer():
- dataOutGab.append("P")
- qtdBP = qtdBP + 1
+ data_out_gab.append("P")
+ qtd_bp = qtd_bp + 1
else:
- dataOutGab.append("D")
+ data_out_gab.append("D")
else:
- dataOutGab.append("D")
+ data_out_gab.append("D")
# Sorts the data to the new output size
- if dataOutGab[-1] == "D":
- dataOrd.append(data[contData])
- contData += 1
+ if data_out_gab[-1] == "D":
+ data_ord.append(data[cont_data])
+ cont_data += 1
else:
- dataOrd.append(None)
+ data_ord.append(None)
# Calculates parity
- qtdBP = 0 # parity bit counter
- for bp in range(1, sizePar + 1):
+ qtd_bp = 0 # parity bit counter
+ for bp in range(1, size_par + 1):
# Bit counter one for a given parity
- contBO = 0
+ cont_bo = 0
# counter to control the loop reading
- contLoop = 0
- for x in dataOrd:
+ for cont_loop, x in enumerate(data_ord):
if x is not None:
try:
- aux = (binPos[contLoop])[-1 * (bp)]
+ aux = (bin_pos[cont_loop])[-1 * (bp)]
except IndexError:
aux = "0"
- if aux == "1":
- if x == "1":
- contBO += 1
- contLoop += 1
- parity.append(contBO % 2)
+ if aux == "1" and x == "1":
+ cont_bo += 1
+ parity.append(cont_bo % 2)
- qtdBP += 1
+ qtd_bp += 1
# Mount the message
- ContBP = 0 # parity bit counter
- for x in range(0, sizePar + len(data)):
- if dataOrd[x] is None:
- dataOut.append(str(parity[ContBP]))
- ContBP += 1
+ cont_bp = 0 # parity bit counter
+ for x in range(size_par + len(data)):
+ if data_ord[x] is None:
+ data_out.append(str(parity[cont_bp]))
+ cont_bp += 1
else:
- dataOut.append(dataOrd[x])
+ data_out.append(data_ord[x])
- return dataOut
+ return data_out
-def receptorConverter(sizePar, data):
+def receptor_converter(size_par, data):
"""
- >>> receptorConverter(4, "1111010010111111")
+ >>> receptor_converter(4, "1111010010111111")
(['1', '0', '1', '0', '1', '0', '1', '1', '1', '1', '1', '1'], True)
"""
# data position template + parity
- dataOutGab = []
+ data_out_gab = []
# Parity bit counter
- qtdBP = 0
+ qtd_bp = 0
# Counter p data bit reading
- contData = 0
+ cont_data = 0
# list of parity received
- parityReceived = []
- dataOutput = []
+ parity_received = []
+ data_output = []
- for x in range(1, len(data) + 1):
+ for i, item in enumerate(data, 1):
# Performs a template of bit positions - who should be given,
# and who should be parity
- if qtdBP < sizePar and (np.log(x) / np.log(2)).is_integer():
- dataOutGab.append("P")
- qtdBP = qtdBP + 1
+ if qtd_bp < size_par and (np.log(i) / np.log(2)).is_integer():
+ data_out_gab.append("P")
+ qtd_bp = qtd_bp + 1
else:
- dataOutGab.append("D")
+ data_out_gab.append("D")
# Sorts the data to the new output size
- if dataOutGab[-1] == "D":
- dataOutput.append(data[contData])
+ if data_out_gab[-1] == "D":
+ data_output.append(item)
else:
- parityReceived.append(data[contData])
- contData += 1
+ parity_received.append(item)
# -----------calculates the parity with the data
- dataOut = []
+ data_out = []
parity = []
- binPos = [bin(x)[2:] for x in range(1, sizePar + len(dataOutput) + 1)]
+ bin_pos = [bin(x)[2:] for x in range(1, size_par + len(data_output) + 1)]
# sorted information data for the size of the output data
- dataOrd = []
+ data_ord = []
# Data position feedback + parity
- dataOutGab = []
+ data_out_gab = []
# Parity bit counter
- qtdBP = 0
+ qtd_bp = 0
# Counter p data bit reading
- contData = 0
+ cont_data = 0
- for x in range(1, sizePar + len(dataOutput) + 1):
+ for x in range(1, size_par + len(data_output) + 1):
# Performs a template position of bits - who should be given,
# and who should be parity
- if qtdBP < sizePar and (np.log(x) / np.log(2)).is_integer():
- dataOutGab.append("P")
- qtdBP = qtdBP + 1
+ if qtd_bp < size_par and (np.log(x) / np.log(2)).is_integer():
+ data_out_gab.append("P")
+ qtd_bp = qtd_bp + 1
else:
- dataOutGab.append("D")
+ data_out_gab.append("D")
# Sorts the data to the new output size
- if dataOutGab[-1] == "D":
- dataOrd.append(dataOutput[contData])
- contData += 1
+ if data_out_gab[-1] == "D":
+ data_ord.append(data_output[cont_data])
+ cont_data += 1
else:
- dataOrd.append(None)
+ data_ord.append(None)
# Calculates parity
- qtdBP = 0 # parity bit counter
- for bp in range(1, sizePar + 1):
+ qtd_bp = 0 # parity bit counter
+ for bp in range(1, size_par + 1):
# Bit counter one for a certain parity
- contBO = 0
- # Counter to control loop reading
- contLoop = 0
- for x in dataOrd:
+ cont_bo = 0
+ for cont_loop, x in enumerate(data_ord):
if x is not None:
try:
- aux = (binPos[contLoop])[-1 * (bp)]
+ aux = (bin_pos[cont_loop])[-1 * (bp)]
except IndexError:
aux = "0"
if aux == "1" and x == "1":
- contBO += 1
- contLoop += 1
- parity.append(str(contBO % 2))
+ cont_bo += 1
+ parity.append(str(cont_bo % 2))
- qtdBP += 1
+ qtd_bp += 1
# Mount the message
- ContBP = 0 # Parity bit counter
- for x in range(0, sizePar + len(dataOutput)):
- if dataOrd[x] is None:
- dataOut.append(str(parity[ContBP]))
- ContBP += 1
+ cont_bp = 0 # Parity bit counter
+ for x in range(size_par + len(data_output)):
+ if data_ord[x] is None:
+ data_out.append(str(parity[cont_bp]))
+ cont_bp += 1
else:
- dataOut.append(dataOrd[x])
+ data_out.append(data_ord[x])
- ack = parityReceived == parity
- return dataOutput, ack
+ ack = parity_received == parity
+ return data_output, ack
# ---------------------------------------------------------------------
diff --git a/hashes/luhn.py b/hashes/luhn.py
new file mode 100644
index 000000000000..a29bf39e3d82
--- /dev/null
+++ b/hashes/luhn.py
@@ -0,0 +1,43 @@
+"""Luhn Algorithm"""
+
+from __future__ import annotations
+
+
+def is_luhn(string: str) -> bool:
+ """
+ Perform Luhn validation on an input string
+ Algorithm:
+ * Double every other digit starting from 2nd last digit.
+ * Subtract 9 if number is greater than 9.
+ * Sum the numbers
+ *
+ >>> test_cases = (79927398710, 79927398711, 79927398712, 79927398713,
+ ... 79927398714, 79927398715, 79927398716, 79927398717, 79927398718,
+ ... 79927398719)
+ >>> [is_luhn(str(test_case)) for test_case in test_cases]
+ [False, False, False, True, False, False, False, False, False, False]
+ """
+ check_digit: int
+ _vector: list[str] = list(string)
+ __vector, check_digit = _vector[:-1], int(_vector[-1])
+ vector: list[int] = [int(digit) for digit in __vector]
+
+ vector.reverse()
+ for i, digit in enumerate(vector):
+ if i & 1 == 0:
+ doubled: int = digit * 2
+ if doubled > 9:
+ doubled -= 9
+ check_digit += doubled
+ else:
+ check_digit += digit
+
+ return check_digit % 10 == 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ assert is_luhn("79927398713")
+ assert not is_luhn("79927398714")
diff --git a/hashes/md5.py b/hashes/md5.py
index b7888fb610ac..f9d802ff0308 100644
--- a/hashes/md5.py
+++ b/hashes/md5.py
@@ -1,91 +1,223 @@
-import math
+"""
+The MD5 algorithm is a hash function that's commonly used as a checksum to
+detect data corruption. The algorithm works by processing a given message in
+blocks of 512 bits, padding the message as needed. It uses the blocks to operate
+a 128-bit state and performs a total of 64 such operations. Note that all values
+are little-endian, so inputs are converted as needed.
+Although MD5 was used as a cryptographic hash function in the past, it's since
+been cracked, so it shouldn't be used for security purposes.
-def rearrange(bitString32):
- """[summary]
- Regroups the given binary string.
+For more info, see https://en.wikipedia.org/wiki/MD5
+"""
+
+from collections.abc import Generator
+from math import sin
+
+
+def to_little_endian(string_32: bytes) -> bytes:
+ """
+ Converts the given string to little-endian in groups of 8 chars.
Arguments:
- bitString32 {[string]} -- [32 bit binary]
+ string_32 {[string]} -- [32-char string]
Raises:
- ValueError -- [if the given string not are 32 bit binary string]
+ ValueError -- [input is not 32 char]
Returns:
- [string] -- [32 bit binary string]
- >>> rearrange('1234567890abcdfghijklmnopqrstuvw')
- 'pqrstuvwhijklmno90abcdfg12345678'
+ 32-char little-endian string
+ >>> to_little_endian(b'1234567890abcdfghijklmnopqrstuvw')
+ b'pqrstuvwhijklmno90abcdfg12345678'
+ >>> to_little_endian(b'1234567890')
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be of length 32
"""
+ if len(string_32) != 32:
+ raise ValueError("Input must be of length 32")
- if len(bitString32) != 32:
- raise ValueError("Need length 32")
- newString = ""
+ little_endian = b""
for i in [3, 2, 1, 0]:
- newString += bitString32[8 * i : 8 * i + 8]
- return newString
+ little_endian += string_32[8 * i : 8 * i + 8]
+ return little_endian
+
+
+def reformat_hex(i: int) -> bytes:
+ """
+ Converts the given non-negative integer to hex string.
+ Example: Suppose the input is the following:
+ i = 1234
-def reformatHex(i):
- """[summary]
- Converts the given integer into 8-digit hex number.
+ The input is 0x000004d2 in hex, so the little-endian hex string is
+ "d2040000".
Arguments:
- i {[int]} -- [integer]
- >>> reformatHex(666)
- '9a020000'
+ i {[int]} -- [integer]
+
+ Raises:
+ ValueError -- [input is negative]
+
+ Returns:
+ 8-char little-endian hex string
+
+ >>> reformat_hex(1234)
+ b'd2040000'
+ >>> reformat_hex(666)
+ b'9a020000'
+ >>> reformat_hex(0)
+ b'00000000'
+ >>> reformat_hex(1234567890)
+ b'd2029649'
+ >>> reformat_hex(1234567890987654321)
+ b'b11c6cb1'
+ >>> reformat_hex(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be non-negative
"""
+ if i < 0:
+ raise ValueError("Input must be non-negative")
- hexrep = format(i, "08x")
- thing = ""
- for i in [3, 2, 1, 0]:
- thing += hexrep[2 * i : 2 * i + 2]
- return thing
+ hex_rep = format(i, "08x")[-8:]
+ little_endian_hex = b""
+ for j in [3, 2, 1, 0]:
+ little_endian_hex += hex_rep[2 * j : 2 * j + 2].encode("utf-8")
+ return little_endian_hex
-def pad(bitString):
- """[summary]
- Fills up the binary string to a 512 bit binary string
+def preprocess(message: bytes) -> bytes:
+ """
+ Preprocesses the message string:
+ - Convert message to bit string
+ - Pad bit string to a multiple of 512 chars:
+ - Append a 1
+ - Append 0's until length = 448 (mod 512)
+ - Append length of original message (64 chars)
+
+ Example: Suppose the input is the following:
+ message = "a"
+
+ The message bit string is "01100001", which is 8 bits long. Thus, the
+ bit string needs 439 bits of padding so that
+ (bit_string + "1" + padding) = 448 (mod 512).
+ The message length is "000010000...0" in 64-bit little-endian binary.
+ The combined bit string is then 512 bits long.
Arguments:
- bitString {[string]} -- [binary string]
+ message {[string]} -- [message string]
Returns:
- [string] -- [binary string]
+ processed bit string padded to a multiple of 512 chars
+
+ >>> preprocess(b"a") == (b"01100001" + b"1" +
+ ... (b"0" * 439) + b"00001000" + (b"0" * 56))
+ True
+ >>> preprocess(b"") == b"1" + (b"0" * 447) + (b"0" * 64)
+ True
"""
- startLength = len(bitString)
- bitString += "1"
- while len(bitString) % 512 != 448:
- bitString += "0"
- lastPart = format(startLength, "064b")
- bitString += rearrange(lastPart[32:]) + rearrange(lastPart[:32])
- return bitString
+ bit_string = b""
+ for char in message:
+ bit_string += format(char, "08b").encode("utf-8")
+ start_len = format(len(bit_string), "064b").encode("utf-8")
+
+ # Pad bit_string to a multiple of 512 chars
+ bit_string += b"1"
+ while len(bit_string) % 512 != 448:
+ bit_string += b"0"
+ bit_string += to_little_endian(start_len[32:]) + to_little_endian(start_len[:32])
+
+ return bit_string
+
+
+def get_block_words(bit_string: bytes) -> Generator[list[int]]:
+ """
+ Splits bit string into blocks of 512 chars and yields each block as a list
+ of 32-bit words
+
+ Example: Suppose the input is the following:
+ bit_string =
+ "000000000...0" + # 0x00 (32 bits, padded to the right)
+ "000000010...0" + # 0x01 (32 bits, padded to the right)
+ "000000100...0" + # 0x02 (32 bits, padded to the right)
+ "000000110...0" + # 0x03 (32 bits, padded to the right)
+ ...
+ "000011110...0" # 0x0a (32 bits, padded to the right)
+ Then len(bit_string) == 512, so there'll be 1 block. The block is split
+ into 32-bit words, and each word is converted to little endian. The
+ first word is interpreted as 0 in decimal, the second word is
+ interpreted as 1 in decimal, etc.
-def getBlock(bitString):
- """[summary]
- Iterator:
- Returns by each call a list of length 16 with the 32 bit
- integer blocks.
+ Thus, block_words == [[0, 1, 2, 3, ..., 15]].
Arguments:
- bitString {[string]} -- [binary string >= 512]
+ bit_string {[string]} -- [bit string with multiple of 512 as length]
+
+ Raises:
+ ValueError -- [length of bit string isn't multiple of 512]
+
+ Yields:
+ a list of 16 32-bit words
+
+ >>> test_string = ("".join(format(n << 24, "032b") for n in range(16))
+ ... .encode("utf-8"))
+ >>> list(get_block_words(test_string))
+ [[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15]]
+ >>> list(get_block_words(test_string * 4)) == [list(range(16))] * 4
+ True
+ >>> list(get_block_words(b"1" * 512)) == [[4294967295] * 16]
+ True
+ >>> list(get_block_words(b""))
+ []
+ >>> list(get_block_words(b"1111"))
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must have length that's a multiple of 512
"""
+ if len(bit_string) % 512 != 0:
+ raise ValueError("Input must have length that's a multiple of 512")
- currPos = 0
- while currPos < len(bitString):
- currPart = bitString[currPos : currPos + 512]
- mySplits = []
- for i in range(16):
- mySplits.append(int(rearrange(currPart[32 * i : 32 * i + 32]), 2))
- yield mySplits
- currPos += 512
+ for pos in range(0, len(bit_string), 512):
+ block = bit_string[pos : pos + 512]
+ block_words = []
+ for i in range(0, 512, 32):
+ block_words.append(int(to_little_endian(block[i : i + 32]), 2))
+ yield block_words
-def not32(i):
+def not_32(i: int) -> int:
"""
- >>> not32(34)
+ Perform bitwise NOT on given int.
+
+ Arguments:
+ i {[int]} -- [given int]
+
+ Raises:
+ ValueError -- [input is negative]
+
+ Returns:
+ Result of bitwise NOT on i
+
+ >>> not_32(34)
4294967261
+ >>> not_32(1234)
+ 4294966061
+ >>> not_32(4294966061)
+ 1234
+ >>> not_32(0)
+ 4294967295
+ >>> not_32(1)
+ 4294967294
+ >>> not_32(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be non-negative
"""
+ if i < 0:
+ raise ValueError("Input must be non-negative")
+
i_str = format(i, "032b")
new_str = ""
for c in i_str:
@@ -93,36 +225,114 @@ def not32(i):
return int(new_str, 2)
-def sum32(a, b):
- """"""
- return (a + b) % 2 ** 32
+def sum_32(a: int, b: int) -> int:
+ """
+ Add two numbers as 32-bit ints.
+ Arguments:
+ a {[int]} -- [first given int]
+ b {[int]} -- [second given int]
-def leftrot32(i, s):
- return (i << s) ^ (i >> (32 - s))
+ Returns:
+ (a + b) as an unsigned 32-bit int
+
+ >>> sum_32(1, 1)
+ 2
+ >>> sum_32(2, 3)
+ 5
+ >>> sum_32(0, 0)
+ 0
+ >>> sum_32(-1, -1)
+ 4294967294
+ >>> sum_32(4294967295, 1)
+ 0
+ """
+ return (a + b) % 2**32
-def md5me(testString):
- """[summary]
- Returns a 32-bit hash code of the string 'testString'
+def left_rotate_32(i: int, shift: int) -> int:
+ """
+ Rotate the bits of a given int left by a given amount.
Arguments:
- testString {[string]} -- [message]
+ i {[int]} -- [given int]
+ shift {[int]} -- [shift amount]
+
+ Raises:
+ ValueError -- [either given int or shift is negative]
+
+ Returns:
+ `i` rotated to the left by `shift` bits
+
+ >>> left_rotate_32(1234, 1)
+ 2468
+ >>> left_rotate_32(1111, 4)
+ 17776
+ >>> left_rotate_32(2147483648, 1)
+ 1
+ >>> left_rotate_32(2147483648, 3)
+ 4
+ >>> left_rotate_32(4294967295, 4)
+ 4294967295
+ >>> left_rotate_32(1234, 0)
+ 1234
+ >>> left_rotate_32(0, 0)
+ 0
+ >>> left_rotate_32(-1, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be non-negative
+ >>> left_rotate_32(0, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Shift must be non-negative
"""
+ if i < 0:
+ raise ValueError("Input must be non-negative")
+ if shift < 0:
+ raise ValueError("Shift must be non-negative")
+ return ((i << shift) ^ (i >> (32 - shift))) % 2**32
- bs = ""
- for i in testString:
- bs += format(ord(i), "08b")
- bs = pad(bs)
- tvals = [int(2 ** 32 * abs(math.sin(i + 1))) for i in range(64)]
+def md5_me(message: bytes) -> bytes:
+ """
+ Returns the 32-char MD5 hash of a given message.
+
+ Reference: https://en.wikipedia.org/wiki/MD5#Algorithm
+
+ Arguments:
+ message {[string]} -- [message]
+
+ Returns:
+ 32-char MD5 hash string
+
+ >>> md5_me(b"")
+ b'd41d8cd98f00b204e9800998ecf8427e'
+ >>> md5_me(b"The quick brown fox jumps over the lazy dog")
+ b'9e107d9d372bb6826bd81d3542a419d6'
+ >>> md5_me(b"The quick brown fox jumps over the lazy dog.")
+ b'e4d909c290d0fb1ca068ffaddf22cbd0'
+
+ >>> import hashlib
+ >>> from string import ascii_letters
+ >>> msgs = [b"", ascii_letters.encode("utf-8"), "Üñîçø∂é".encode("utf-8"),
+ ... b"The quick brown fox jumps over the lazy dog."]
+ >>> all(md5_me(msg) == hashlib.md5(msg).hexdigest().encode("utf-8") for msg in msgs)
+ True
+ """
+
+ # Convert to bit string, add padding and append message length
+ bit_string = preprocess(message)
+
+ added_consts = [int(2**32 * abs(sin(i + 1))) for i in range(64)]
+ # Starting states
a0 = 0x67452301
b0 = 0xEFCDAB89
c0 = 0x98BADCFE
d0 = 0x10325476
- s = [
+ shift_amounts = [
7,
12,
17,
@@ -189,51 +399,46 @@ def md5me(testString):
21,
]
- for m in getBlock(bs):
- A = a0
- B = b0
- C = c0
- D = d0
+ # Process bit string in chunks, each with 16 32-char words
+ for block_words in get_block_words(bit_string):
+ a = a0
+ b = b0
+ c = c0
+ d = d0
+
+ # Hash current chunk
for i in range(64):
if i <= 15:
- # f = (B & C) | (not32(B) & D)
- f = D ^ (B & (C ^ D))
+ # f = (b & c) | (not_32(b) & d) # Alternate definition for f
+ f = d ^ (b & (c ^ d))
g = i
elif i <= 31:
- # f = (D & B) | (not32(D) & C)
- f = C ^ (D & (B ^ C))
+ # f = (d & b) | (not_32(d) & c) # Alternate definition for f
+ f = c ^ (d & (b ^ c))
g = (5 * i + 1) % 16
elif i <= 47:
- f = B ^ C ^ D
+ f = b ^ c ^ d
g = (3 * i + 5) % 16
else:
- f = C ^ (B | not32(D))
+ f = c ^ (b | not_32(d))
g = (7 * i) % 16
- dtemp = D
- D = C
- C = B
- B = sum32(B, leftrot32((A + f + tvals[i] + m[g]) % 2 ** 32, s[i]))
- A = dtemp
- a0 = sum32(a0, A)
- b0 = sum32(b0, B)
- c0 = sum32(c0, C)
- d0 = sum32(d0, D)
-
- digest = reformatHex(a0) + reformatHex(b0) + reformatHex(c0) + reformatHex(d0)
+ f = (f + a + added_consts[i] + block_words[g]) % 2**32
+ a = d
+ d = c
+ c = b
+ b = sum_32(b, left_rotate_32(f, shift_amounts[i]))
+
+ # Add hashed chunk to running total
+ a0 = sum_32(a0, a)
+ b0 = sum_32(b0, b)
+ c0 = sum_32(c0, c)
+ d0 = sum_32(d0, d)
+
+ digest = reformat_hex(a0) + reformat_hex(b0) + reformat_hex(c0) + reformat_hex(d0)
return digest
-def test():
- assert md5me("") == "d41d8cd98f00b204e9800998ecf8427e"
- assert (
- md5me("The quick brown fox jumps over the lazy dog")
- == "9e107d9d372bb6826bd81d3542a419d6"
- )
- print("Success.")
-
-
if __name__ == "__main__":
- test()
import doctest
doctest.testmod()
diff --git a/hashes/sdbm.py b/hashes/sdbm.py
index 86d47a1d9967..a5abc6f3185b 100644
--- a/hashes/sdbm.py
+++ b/hashes/sdbm.py
@@ -1,25 +1,25 @@
"""
- This algorithm was created for sdbm (a public-domain reimplementation of ndbm)
- database library.
- It was found to do well in scrambling bits, causing better distribution of the keys
- and fewer splits.
- It also happens to be a good general hashing function with good distribution.
- The actual function (pseudo code) is:
- for i in i..len(str):
- hash(i) = hash(i - 1) * 65599 + str[i];
+This algorithm was created for sdbm (a public-domain reimplementation of ndbm)
+database library.
+It was found to do well in scrambling bits, causing better distribution of the keys
+and fewer splits.
+It also happens to be a good general hashing function with good distribution.
+The actual function (pseudo code) is:
+ for i in i..len(str):
+ hash(i) = hash(i - 1) * 65599 + str[i];
- What is included below is the faster version used in gawk. [there is even a faster,
- duff-device version]
- The magic constant 65599 was picked out of thin air while experimenting with
- different constants.
- It turns out to be a prime.
- This is one of the algorithms used in berkeley db (see sleepycat) and elsewhere.
+What is included below is the faster version used in gawk. [there is even a faster,
+duff-device version]
+The magic constant 65599 was picked out of thin air while experimenting with
+different constants.
+It turns out to be a prime.
+This is one of the algorithms used in berkeley db (see sleepycat) and elsewhere.
- source: http://www.cse.yorku.ca/~oz/hash.html
+source: http://www.cse.yorku.ca/~oz/hash.html
"""
-def sdbm(plain_text: str) -> str:
+def sdbm(plain_text: str) -> int:
"""
Function implements sdbm hash, easy to use, great for bits scrambling.
iterates over each character in the given string and applies function to each of
@@ -31,7 +31,9 @@ def sdbm(plain_text: str) -> str:
>>> sdbm('scramble bits')
730247649148944819640658295400555317318720608290373040936089
"""
- hash = 0
+ hash_value = 0
for plain_chr in plain_text:
- hash = ord(plain_chr) + (hash << 6) + (hash << 16) - hash
- return hash
+ hash_value = (
+ ord(plain_chr) + (hash_value << 6) + (hash_value << 16) - hash_value
+ )
+ return hash_value
diff --git a/hashes/sha1.py b/hashes/sha1.py
index cca38b7c3fdc..75a1423e9b5f 100644
--- a/hashes/sha1.py
+++ b/hashes/sha1.py
@@ -1,48 +1,50 @@
"""
-Demonstrates implementation of SHA1 Hash function in a Python class and gives utilities
-to find hash of string or hash of text from a file.
+Implementation of the SHA1 hash function and gives utilities to find hash of string or
+hash of text from a file. Also contains a Test class to verify that the generated hash
+matches what is returned by the hashlib library
+
Usage: python sha1.py --string "Hello World!!"
python sha1.py --file "hello_world.txt"
When run without any arguments, it prints the hash of the string "Hello World!!
Welcome to Cryptography"
-Also contains a Test class to verify that the generated Hash is same as that
-returned by the hashlib library
-SHA1 hash or SHA1 sum of a string is a cryptographic function which means it is easy
+SHA1 hash or SHA1 sum of a string is a cryptographic function, which means it is easy
to calculate forwards but extremely difficult to calculate backwards. What this means
-is, you can easily calculate the hash of a string, but it is extremely difficult to
-know the original string if you have its hash. This property is useful to communicate
-securely, send encrypted messages and is very useful in payment systems, blockchain
-and cryptocurrency etc.
-The Algorithm as described in the reference:
+is you can easily calculate the hash of a string, but it is extremely difficult to know
+the original string if you have its hash. This property is useful for communicating
+securely, send encrypted messages and is very useful in payment systems, blockchain and
+cryptocurrency etc.
+
+The algorithm as described in the reference:
First we start with a message. The message is padded and the length of the message
is added to the end. It is then split into blocks of 512 bits or 64 bytes. The blocks
are then processed one at a time. Each block must be expanded and compressed.
-The value after each compression is added to a 160bit buffer called the current hash
-state. After the last block is processed the current hash state is returned as
+The value after each compression is added to a 160-bit buffer called the current hash
+state. After the last block is processed, the current hash state is returned as
the final hash.
+
Reference: https://deadhacker.com/2006/02/21/sha-1-illustrated/
"""
+
import argparse
import hashlib # hashlib is only used inside the Test class
import struct
-import unittest
class SHA1Hash:
"""
- Class to contain the entire pipeline for SHA1 Hashing Algorithm
+ Class to contain the entire pipeline for SHA1 hashing algorithm
>>> SHA1Hash(bytes('Allan', 'utf-8')).final_hash()
'872af2d8ac3d8695387e7c804bf0e02c18df9e6e'
"""
def __init__(self, data):
"""
- Inititates the variables data and h. h is a list of 5 8-digit Hexadecimal
+ Initiates the variables data and h. h is a list of 5 8-digit hexadecimal
numbers corresponding to
(1732584193, 4023233417, 2562383102, 271733878, 3285377520)
respectively. We will start with this as a message digest. 0x is how you write
- Hexadecimal numbers in Python
+ hexadecimal numbers in Python
"""
self.data = data
self.h = [0x67452301, 0xEFCDAB89, 0x98BADCFE, 0x10325476, 0xC3D2E1F0]
@@ -91,14 +93,14 @@ def final_hash(self):
For each block, the variable h that was initialized is copied to a,b,c,d,e
and these 5 variables a,b,c,d,e undergo several changes. After all the blocks
are processed, these 5 variables are pairwise added to h ie a to h[0], b to h[1]
- and so on. This h becomes our final hash which is returned.
+ and so on. This h becomes our final hash which is returned.
"""
self.padded_data = self.padding()
self.blocks = self.split_blocks()
for block in self.blocks:
expanded_block = self.expand_block(block)
a, b, c, d, e = self.h
- for i in range(0, 80):
+ for i in range(80):
if 0 <= i < 20:
f = (b & c) | ((~b) & d)
k = 0x5A827999
@@ -118,30 +120,25 @@ def final_hash(self):
c,
d,
)
- self.h = (
- self.h[0] + a & 0xFFFFFFFF,
- self.h[1] + b & 0xFFFFFFFF,
- self.h[2] + c & 0xFFFFFFFF,
- self.h[3] + d & 0xFFFFFFFF,
- self.h[4] + e & 0xFFFFFFFF,
- )
- return "%08x%08x%08x%08x%08x" % tuple(self.h)
+ self.h = (
+ self.h[0] + a & 0xFFFFFFFF,
+ self.h[1] + b & 0xFFFFFFFF,
+ self.h[2] + c & 0xFFFFFFFF,
+ self.h[3] + d & 0xFFFFFFFF,
+ self.h[4] + e & 0xFFFFFFFF,
+ )
+ return ("{:08x}" * 5).format(*self.h)
-class SHA1HashTest(unittest.TestCase):
- """
- Test class for the SHA1Hash class. Inherits the TestCase class from unittest
- """
-
- def testMatchHashes(self):
- msg = bytes("Test String", "utf-8")
- self.assertEqual(SHA1Hash(msg).final_hash(), hashlib.sha1(msg).hexdigest())
+def test_sha1_hash():
+ msg = b"Test String"
+ assert SHA1Hash(msg).final_hash() == hashlib.sha1(msg).hexdigest() # noqa: S324
def main():
"""
Provides option 'string' or 'file' to take input and prints the calculated SHA1
- hash. unittest.main() has been commented because we probably don't want to run
+ hash. unittest.main() has been commented out because we probably don't want to run
the test each time.
"""
# unittest.main()
diff --git a/hashes/sha256.py b/hashes/sha256.py
new file mode 100644
index 000000000000..bcc83edca480
--- /dev/null
+++ b/hashes/sha256.py
@@ -0,0 +1,248 @@
+# Author: M. Yathurshan
+# Black Formatter: True
+
+"""
+Implementation of SHA256 Hash function in a Python class and provides utilities
+to find hash of string or hash of text from a file.
+
+Usage: python sha256.py --string "Hello World!!"
+ python sha256.py --file "hello_world.txt"
+ When run without any arguments,
+ it prints the hash of the string "Hello World!! Welcome to Cryptography"
+
+References:
+https://qvault.io/cryptography/how-sha-2-works-step-by-step-sha-256/
+https://en.wikipedia.org/wiki/SHA-2
+"""
+
+import argparse
+import struct
+import unittest
+
+
+class SHA256:
+ """
+ Class to contain the entire pipeline for SHA1 Hashing Algorithm
+
+ >>> SHA256(b'Python').hash
+ '18885f27b5af9012df19e496460f9294d5ab76128824c6f993787004f6d9a7db'
+
+ >>> SHA256(b'hello world').hash
+ 'b94d27b9934d3e08a52e52d7da7dabfac484efe37a5380ee9088f7ace2efcde9'
+ """
+
+ def __init__(self, data: bytes) -> None:
+ self.data = data
+
+ # Initialize hash values
+ self.hashes = [
+ 0x6A09E667,
+ 0xBB67AE85,
+ 0x3C6EF372,
+ 0xA54FF53A,
+ 0x510E527F,
+ 0x9B05688C,
+ 0x1F83D9AB,
+ 0x5BE0CD19,
+ ]
+
+ # Initialize round constants
+ self.round_constants = [
+ 0x428A2F98,
+ 0x71374491,
+ 0xB5C0FBCF,
+ 0xE9B5DBA5,
+ 0x3956C25B,
+ 0x59F111F1,
+ 0x923F82A4,
+ 0xAB1C5ED5,
+ 0xD807AA98,
+ 0x12835B01,
+ 0x243185BE,
+ 0x550C7DC3,
+ 0x72BE5D74,
+ 0x80DEB1FE,
+ 0x9BDC06A7,
+ 0xC19BF174,
+ 0xE49B69C1,
+ 0xEFBE4786,
+ 0x0FC19DC6,
+ 0x240CA1CC,
+ 0x2DE92C6F,
+ 0x4A7484AA,
+ 0x5CB0A9DC,
+ 0x76F988DA,
+ 0x983E5152,
+ 0xA831C66D,
+ 0xB00327C8,
+ 0xBF597FC7,
+ 0xC6E00BF3,
+ 0xD5A79147,
+ 0x06CA6351,
+ 0x14292967,
+ 0x27B70A85,
+ 0x2E1B2138,
+ 0x4D2C6DFC,
+ 0x53380D13,
+ 0x650A7354,
+ 0x766A0ABB,
+ 0x81C2C92E,
+ 0x92722C85,
+ 0xA2BFE8A1,
+ 0xA81A664B,
+ 0xC24B8B70,
+ 0xC76C51A3,
+ 0xD192E819,
+ 0xD6990624,
+ 0xF40E3585,
+ 0x106AA070,
+ 0x19A4C116,
+ 0x1E376C08,
+ 0x2748774C,
+ 0x34B0BCB5,
+ 0x391C0CB3,
+ 0x4ED8AA4A,
+ 0x5B9CCA4F,
+ 0x682E6FF3,
+ 0x748F82EE,
+ 0x78A5636F,
+ 0x84C87814,
+ 0x8CC70208,
+ 0x90BEFFFA,
+ 0xA4506CEB,
+ 0xBEF9A3F7,
+ 0xC67178F2,
+ ]
+
+ self.preprocessed_data = self.preprocessing(self.data)
+ self.final_hash()
+
+ @staticmethod
+ def preprocessing(data: bytes) -> bytes:
+ padding = b"\x80" + (b"\x00" * (63 - (len(data) + 8) % 64))
+ big_endian_integer = struct.pack(">Q", (len(data) * 8))
+ return data + padding + big_endian_integer
+
+ def final_hash(self) -> None:
+ # Convert into blocks of 64 bytes
+ self.blocks = [
+ self.preprocessed_data[x : x + 64]
+ for x in range(0, len(self.preprocessed_data), 64)
+ ]
+
+ for block in self.blocks:
+ # Convert the given block into a list of 4 byte integers
+ words = list(struct.unpack(">16L", block))
+ # add 48 0-ed integers
+ words += [0] * 48
+
+ a, b, c, d, e, f, g, h = self.hashes
+
+ for index in range(64):
+ if index > 15:
+ # modify the zero-ed indexes at the end of the array
+ s0 = (
+ self.ror(words[index - 15], 7)
+ ^ self.ror(words[index - 15], 18)
+ ^ (words[index - 15] >> 3)
+ )
+ s1 = (
+ self.ror(words[index - 2], 17)
+ ^ self.ror(words[index - 2], 19)
+ ^ (words[index - 2] >> 10)
+ )
+
+ words[index] = (
+ words[index - 16] + s0 + words[index - 7] + s1
+ ) % 0x100000000
+
+ # Compression
+ s1 = self.ror(e, 6) ^ self.ror(e, 11) ^ self.ror(e, 25)
+ ch = (e & f) ^ ((~e & (0xFFFFFFFF)) & g)
+ temp1 = (
+ h + s1 + ch + self.round_constants[index] + words[index]
+ ) % 0x100000000
+ s0 = self.ror(a, 2) ^ self.ror(a, 13) ^ self.ror(a, 22)
+ maj = (a & b) ^ (a & c) ^ (b & c)
+ temp2 = (s0 + maj) % 0x100000000
+
+ h, g, f, e, d, c, b, a = (
+ g,
+ f,
+ e,
+ ((d + temp1) % 0x100000000),
+ c,
+ b,
+ a,
+ ((temp1 + temp2) % 0x100000000),
+ )
+
+ mutated_hash_values = [a, b, c, d, e, f, g, h]
+
+ # Modify final values
+ self.hashes = [
+ ((element + mutated_hash_values[index]) % 0x100000000)
+ for index, element in enumerate(self.hashes)
+ ]
+
+ self.hash = "".join([hex(value)[2:].zfill(8) for value in self.hashes])
+
+ def ror(self, value: int, rotations: int) -> int:
+ """
+ Right rotate a given unsigned number by a certain amount of rotations
+ """
+ return 0xFFFFFFFF & (value << (32 - rotations)) | (value >> rotations)
+
+
+class SHA256HashTest(unittest.TestCase):
+ """
+ Test class for the SHA256 class. Inherits the TestCase class from unittest
+ """
+
+ def test_match_hashes(self) -> None:
+ import hashlib
+
+ msg = bytes("Test String", "utf-8")
+ assert SHA256(msg).hash == hashlib.sha256(msg).hexdigest()
+
+
+def main() -> None:
+ """
+ Provides option 'string' or 'file' to take input
+ and prints the calculated SHA-256 hash
+ """
+
+ # unittest.main()
+
+ import doctest
+
+ doctest.testmod()
+
+ parser = argparse.ArgumentParser()
+ parser.add_argument(
+ "-s",
+ "--string",
+ dest="input_string",
+ default="Hello World!! Welcome to Cryptography",
+ help="Hash the string",
+ )
+ parser.add_argument(
+ "-f", "--file", dest="input_file", help="Hash contents of a file"
+ )
+
+ args = parser.parse_args()
+
+ input_string = args.input_string
+
+ # hash input should be a bytestring
+ if args.input_file:
+ with open(args.input_file, "rb") as f:
+ hash_input = f.read()
+ else:
+ hash_input = bytes(input_string, "utf-8")
+
+ print(SHA256(hash_input).hash)
+
+
+if __name__ == "__main__":
+ main()
diff --git a/images/Travis_CI_fail_1.png b/images/Travis_CI_fail_1.png
deleted file mode 100644
index 451e54e4844a..000000000000
Binary files a/images/Travis_CI_fail_1.png and /dev/null differ
diff --git a/images/Travis_CI_fail_2.png b/images/Travis_CI_fail_2.png
deleted file mode 100644
index caa406099da1..000000000000
Binary files a/images/Travis_CI_fail_2.png and /dev/null differ
diff --git a/index.md b/index.md
new file mode 100644
index 000000000000..134520cb94aa
--- /dev/null
+++ b/index.md
@@ -0,0 +1,10 @@
+# TheAlgorithms/Python
+```{toctree}
+:maxdepth: 2
+:caption: index.md
+
+
+CONTRIBUTING.md
+README.md
+LICENSE.md
+```
diff --git a/knapsack/README.md b/knapsack/README.md
index 6041c1e48eb8..f31e5f591412 100644
--- a/knapsack/README.md
+++ b/knapsack/README.md
@@ -17,7 +17,7 @@ The knapsack problem has been studied for more than a century, with early works
## Documentation
This module uses docstrings to enable the use of Python's in-built `help(...)` function.
-For instance, try `help(Vector)`, `help(unitBasisVector)`, and `help(CLASSNAME.METHODNAME)`.
+For instance, try `help(Vector)`, `help(unit_basis_vector)`, and `help(CLASSNAME.METHODNAME)`.
---
diff --git a/greedy_method/greedy_knapsack.py b/knapsack/greedy_knapsack.py
similarity index 100%
rename from greedy_method/greedy_knapsack.py
rename to knapsack/greedy_knapsack.py
diff --git a/knapsack/knapsack.py b/knapsack/knapsack.py
index 756443ea6163..bb507be1ba3c 100644
--- a/knapsack/knapsack.py
+++ b/knapsack/knapsack.py
@@ -1,11 +1,11 @@
-from typing import List
-
-""" A naive recursive implementation of 0-1 Knapsack Problem
- https://en.wikipedia.org/wiki/Knapsack_problem
+"""A naive recursive implementation of 0-1 Knapsack Problem
+https://en.wikipedia.org/wiki/Knapsack_problem
"""
+from __future__ import annotations
+
-def knapsack(capacity: int, weights: List[int], values: List[int], counter: int) -> int:
+def knapsack(capacity: int, weights: list[int], values: list[int], counter: int) -> int:
"""
Returns the maximum value that can be put in a knapsack of a capacity cap,
whereby each weight w has a specific value val.
diff --git a/knapsack/recursive_approach_knapsack.py b/knapsack/recursive_approach_knapsack.py
new file mode 100644
index 000000000000..9a8ed1886a5b
--- /dev/null
+++ b/knapsack/recursive_approach_knapsack.py
@@ -0,0 +1,51 @@
+# To get an insight into naive recursive way to solve the Knapsack problem
+
+
+"""
+A shopkeeper has bags of wheat that each have different weights and different profits.
+eg.
+no_of_items 4
+profit 5 4 8 6
+weight 1 2 4 5
+max_weight 5
+Constraints:
+max_weight > 0
+profit[i] >= 0
+weight[i] >= 0
+Calculate the maximum profit that the shopkeeper can make given maxmum weight that can
+be carried.
+"""
+
+
+def knapsack(
+ weights: list, values: list, number_of_items: int, max_weight: int, index: int
+) -> int:
+ """
+ Function description is as follows-
+ :param weights: Take a list of weights
+ :param values: Take a list of profits corresponding to the weights
+ :param number_of_items: number of items available to pick from
+ :param max_weight: Maximum weight that could be carried
+ :param index: the element we are looking at
+ :return: Maximum expected gain
+ >>> knapsack([1, 2, 4, 5], [5, 4, 8, 6], 4, 5, 0)
+ 13
+ >>> knapsack([3 ,4 , 5], [10, 9 , 8], 3, 25, 0)
+ 27
+ """
+ if index == number_of_items:
+ return 0
+ ans1 = 0
+ ans2 = 0
+ ans1 = knapsack(weights, values, number_of_items, max_weight, index + 1)
+ if weights[index] <= max_weight:
+ ans2 = values[index] + knapsack(
+ weights, values, number_of_items, max_weight - weights[index], index + 1
+ )
+ return max(ans1, ans2)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/knapsack/tests/__init__.py b/knapsack/tests/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/greedy_method/test_knapsack.py b/knapsack/tests/test_greedy_knapsack.py
similarity index 77%
rename from greedy_method/test_knapsack.py
rename to knapsack/tests/test_greedy_knapsack.py
index 5e277a92114e..e6a40084109e 100644
--- a/greedy_method/test_knapsack.py
+++ b/knapsack/tests/test_greedy_knapsack.py
@@ -1,6 +1,8 @@
import unittest
-from . import greedy_knapsack as kp
+import pytest
+
+from knapsack import greedy_knapsack as kp
class TestClass(unittest.TestCase):
@@ -16,7 +18,7 @@ def test_sorted(self):
profit = [10, 20, 30, 40, 50, 60]
weight = [2, 4, 6, 8, 10, 12]
max_weight = 100
- self.assertEqual(kp.calc_profit(profit, weight, max_weight), 210)
+ assert kp.calc_profit(profit, weight, max_weight) == 210
def test_negative_max_weight(self):
"""
@@ -26,7 +28,7 @@ def test_negative_max_weight(self):
# profit = [10, 20, 30, 40, 50, 60]
# weight = [2, 4, 6, 8, 10, 12]
# max_weight = -15
- self.assertRaisesRegex(ValueError, "max_weight must greater than zero.")
+ pytest.raises(ValueError, match="max_weight must greater than zero.")
def test_negative_profit_value(self):
"""
@@ -36,7 +38,7 @@ def test_negative_profit_value(self):
# profit = [10, -20, 30, 40, 50, 60]
# weight = [2, 4, 6, 8, 10, 12]
# max_weight = 15
- self.assertRaisesRegex(ValueError, "Weight can not be negative.")
+ pytest.raises(ValueError, match="Weight can not be negative.")
def test_negative_weight_value(self):
"""
@@ -46,7 +48,7 @@ def test_negative_weight_value(self):
# profit = [10, 20, 30, 40, 50, 60]
# weight = [2, -4, 6, -8, 10, 12]
# max_weight = 15
- self.assertRaisesRegex(ValueError, "Profit can not be negative.")
+ pytest.raises(ValueError, match="Profit can not be negative.")
def test_null_max_weight(self):
"""
@@ -56,7 +58,7 @@ def test_null_max_weight(self):
# profit = [10, 20, 30, 40, 50, 60]
# weight = [2, 4, 6, 8, 10, 12]
# max_weight = null
- self.assertRaisesRegex(ValueError, "max_weight must greater than zero.")
+ pytest.raises(ValueError, match="max_weight must greater than zero.")
def test_unequal_list_length(self):
"""
@@ -66,9 +68,7 @@ def test_unequal_list_length(self):
# profit = [10, 20, 30, 40, 50]
# weight = [2, 4, 6, 8, 10, 12]
# max_weight = 100
- self.assertRaisesRegex(
- IndexError, "The length of profit and weight must be same."
- )
+ pytest.raises(IndexError, match="The length of profit and weight must be same.")
if __name__ == "__main__":
diff --git a/knapsack/test_knapsack.py b/knapsack/tests/test_knapsack.py
similarity index 79%
rename from knapsack/test_knapsack.py
rename to knapsack/tests/test_knapsack.py
index 248855fbce53..7bfb8780627b 100644
--- a/knapsack/test_knapsack.py
+++ b/knapsack/tests/test_knapsack.py
@@ -6,6 +6,7 @@
This file contains the test-suite for the knapsack problem.
"""
+
import unittest
from knapsack import knapsack as k
@@ -20,12 +21,12 @@ def test_base_case(self):
val = [0]
w = [0]
c = len(val)
- self.assertEqual(k.knapsack(cap, w, val, c), 0)
+ assert k.knapsack(cap, w, val, c) == 0
val = [60]
w = [10]
c = len(val)
- self.assertEqual(k.knapsack(cap, w, val, c), 0)
+ assert k.knapsack(cap, w, val, c) == 0
def test_easy_case(self):
"""
@@ -35,7 +36,7 @@ def test_easy_case(self):
val = [1, 2, 3]
w = [3, 2, 1]
c = len(val)
- self.assertEqual(k.knapsack(cap, w, val, c), 5)
+ assert k.knapsack(cap, w, val, c) == 5
def test_knapsack(self):
"""
@@ -45,7 +46,7 @@ def test_knapsack(self):
val = [60, 100, 120]
w = [10, 20, 30]
c = len(val)
- self.assertEqual(k.knapsack(cap, w, val, c), 220)
+ assert k.knapsack(cap, w, val, c) == 220
if __name__ == "__main__":
diff --git a/linear_algebra/README.md b/linear_algebra/README.md
index dc6085090d02..35b50b5e0f0a 100644
--- a/linear_algebra/README.md
+++ b/linear_algebra/README.md
@@ -10,56 +10,56 @@ This module contains classes and functions for doing linear algebra.
-
- This class represents a vector of arbitrary size and related operations.
- **Overview about the methods:**
+ **Overview of the methods:**
- - constructor(components : list) : init the vector
- - set(components : list) : changes the vector components.
+ - constructor(components) : init the vector
+ - set(components) : changes the vector components.
- \_\_str\_\_() : toString method
- - component(i : int): gets the i-th component (start by 0)
+ - component(i): gets the i-th component (0-indexed)
- \_\_len\_\_() : gets the size / length of the vector (number of components)
- - euclidLength() : returns the eulidean length of the vector.
+ - euclidean_length() : returns the eulidean length of the vector
- operator + : vector addition
- operator - : vector subtraction
- operator * : scalar multiplication and dot product
- - copy() : copies this vector and returns it.
- - changeComponent(pos,value) : changes the specified component.
+ - copy() : copies this vector and returns it
+ - change_component(pos,value) : changes the specified component
-- function zeroVector(dimension)
+- function zero_vector(dimension)
- returns a zero vector of 'dimension'
-- function unitBasisVector(dimension,pos)
- - returns a unit basis vector with a One at index 'pos' (indexing at 0)
-- function axpy(scalar,vector1,vector2)
+- function unit_basis_vector(dimension, pos)
+ - returns a unit basis vector with a one at index 'pos' (0-indexed)
+- function axpy(scalar, vector1, vector2)
- computes the axpy operation
-- function randomVector(N,a,b)
- - returns a random vector of size N, with random integer components between 'a' and 'b'.
+- function random_vector(N, a, b)
+ - returns a random vector of size N, with random integer components between 'a' and 'b' inclusive
### class Matrix
-
- This class represents a matrix of arbitrary size and operations on it.
- **Overview about the methods:**
+ **Overview of the methods:**
- \_\_str\_\_() : returns a string representation
- operator * : implements the matrix vector multiplication
implements the matrix-scalar multiplication.
- - changeComponent(x,y,value) : changes the specified component.
- - component(x,y) : returns the specified component.
+ - change_component(x, y, value) : changes the specified component.
+ - component(x, y) : returns the specified component.
- width() : returns the width of the matrix
- height() : returns the height of the matrix
- - determinate() : returns the determinate of the matrix if it is square
+ - determinant() : returns the determinant of the matrix if it is square
- operator + : implements the matrix-addition.
- - operator - _ implements the matrix-subtraction
+ - operator - : implements the matrix-subtraction
-- function squareZeroMatrix(N)
+- function square_zero_matrix(N)
- returns a square zero-matrix of dimension NxN
-- function randomMatrix(W,H,a,b)
- - returns a random matrix WxH with integer components between 'a' and 'b'
+- function random_matrix(W, H, a, b)
+ - returns a random matrix WxH with integer components between 'a' and 'b' inclusive
---
## Documentation
This module uses docstrings to enable the use of Python's in-built `help(...)` function.
-For instance, try `help(Vector)`, `help(unitBasisVector)`, and `help(CLASSNAME.METHODNAME)`.
+For instance, try `help(Vector)`, `help(unit_basis_vector)`, and `help(CLASSNAME.METHODNAME)`.
---
diff --git a/arithmetic_analysis/gaussian_elimination.py b/linear_algebra/gaussian_elimination.py
similarity index 56%
rename from arithmetic_analysis/gaussian_elimination.py
rename to linear_algebra/gaussian_elimination.py
index 51207686c12a..6f4075b710fd 100644
--- a/arithmetic_analysis/gaussian_elimination.py
+++ b/linear_algebra/gaussian_elimination.py
@@ -1,21 +1,29 @@
"""
-Gaussian elimination method for solving a system of linear equations.
-Gaussian elimination - https://en.wikipedia.org/wiki/Gaussian_elimination
+| Gaussian elimination method for solving a system of linear equations.
+| Gaussian elimination - https://en.wikipedia.org/wiki/Gaussian_elimination
"""
-
import numpy as np
+from numpy import float64
+from numpy.typing import NDArray
-def retroactive_resolution(coefficients: np.matrix, vector: np.array) -> np.array:
+def retroactive_resolution(
+ coefficients: NDArray[float64], vector: NDArray[float64]
+) -> NDArray[float64]:
"""
This function performs a retroactive linear system resolution
- for triangular matrix
+ for triangular matrix
Examples:
- 2x1 + 2x2 - 1x3 = 5 2x1 + 2x2 = -1
- 0x1 - 2x2 - 1x3 = -7 0x1 - 2x2 = -1
- 0x1 + 0x2 + 5x3 = 15
+ 1.
+ * 2x1 + 2x2 - 1x3 = 5
+ * 0x1 - 2x2 - 1x3 = -7
+ * 0x1 + 0x2 + 5x3 = 15
+ 2.
+ * 2x1 + 2x2 = -1
+ * 0x1 - 2x2 = -1
+
>>> gaussian_elimination([[2, 2, -1], [0, -2, -1], [0, 0, 5]], [[5], [-7], [15]])
array([[2.],
[2.],
@@ -27,25 +35,29 @@ def retroactive_resolution(coefficients: np.matrix, vector: np.array) -> np.arra
rows, columns = np.shape(coefficients)
- x = np.zeros((rows, 1), dtype=float)
+ x: NDArray[float64] = np.zeros((rows, 1), dtype=float)
for row in reversed(range(rows)):
- sum = 0
- for col in range(row + 1, columns):
- sum += coefficients[row, col] * x[col]
-
- x[row, 0] = (vector[row] - sum) / coefficients[row, row]
+ total = np.dot(coefficients[row, row + 1 :], x[row + 1 :])
+ x[row, 0] = (vector[row][0] - total[0]) / coefficients[row, row]
return x
-def gaussian_elimination(coefficients: np.matrix, vector: np.array) -> np.array:
+def gaussian_elimination(
+ coefficients: NDArray[float64], vector: NDArray[float64]
+) -> NDArray[float64]:
"""
This function performs Gaussian elimination method
Examples:
- 1x1 - 4x2 - 2x3 = -2 1x1 + 2x2 = 5
- 5x1 + 2x2 - 2x3 = -3 5x1 + 2x2 = 5
- 1x1 - 1x2 + 0x3 = 4
+ 1.
+ * 1x1 - 4x2 - 2x3 = -2
+ * 5x1 + 2x2 - 2x3 = -3
+ * 1x1 - 1x2 + 0x3 = 4
+ 2.
+ * 1x1 + 2x2 = 5
+ * 5x1 + 2x2 = 5
+
>>> gaussian_elimination([[1, -4, -2], [5, 2, -2], [1, -1, 0]], [[-2], [-3], [4]])
array([[ 2.3 ],
[-1.7 ],
@@ -57,10 +69,10 @@ def gaussian_elimination(coefficients: np.matrix, vector: np.array) -> np.array:
# coefficients must to be a square matrix so we need to check first
rows, columns = np.shape(coefficients)
if rows != columns:
- return []
+ return np.array((), dtype=float)
# augmented matrix
- augmented_mat = np.concatenate((coefficients, vector), axis=1)
+ augmented_mat: NDArray[float64] = np.concatenate((coefficients, vector), axis=1)
augmented_mat = augmented_mat.astype("float64")
# scale the matrix leaving it triangular
diff --git a/linear_algebra/jacobi_iteration_method.py b/linear_algebra/jacobi_iteration_method.py
new file mode 100644
index 000000000000..2cc9c103018b
--- /dev/null
+++ b/linear_algebra/jacobi_iteration_method.py
@@ -0,0 +1,204 @@
+"""
+Jacobi Iteration Method - https://en.wikipedia.org/wiki/Jacobi_method
+"""
+
+from __future__ import annotations
+
+import numpy as np
+from numpy import float64
+from numpy.typing import NDArray
+
+
+# Method to find solution of system of linear equations
+def jacobi_iteration_method(
+ coefficient_matrix: NDArray[float64],
+ constant_matrix: NDArray[float64],
+ init_val: list[float],
+ iterations: int,
+) -> list[float]:
+ """
+ Jacobi Iteration Method:
+ An iterative algorithm to determine the solutions of strictly diagonally dominant
+ system of linear equations
+
+ 4x1 + x2 + x3 = 2
+ x1 + 5x2 + 2x3 = -6
+ x1 + 2x2 + 4x3 = -4
+
+ x_init = [0.5, -0.5 , -0.5]
+
+ Examples:
+
+ >>> coefficient = np.array([[4, 1, 1], [1, 5, 2], [1, 2, 4]])
+ >>> constant = np.array([[2], [-6], [-4]])
+ >>> init_val = [0.5, -0.5, -0.5]
+ >>> iterations = 3
+ >>> jacobi_iteration_method(coefficient, constant, init_val, iterations)
+ [0.909375, -1.14375, -0.7484375]
+
+
+ >>> coefficient = np.array([[4, 1, 1], [1, 5, 2]])
+ >>> constant = np.array([[2], [-6], [-4]])
+ >>> init_val = [0.5, -0.5, -0.5]
+ >>> iterations = 3
+ >>> jacobi_iteration_method(coefficient, constant, init_val, iterations)
+ Traceback (most recent call last):
+ ...
+ ValueError: Coefficient matrix dimensions must be nxn but received 2x3
+
+ >>> coefficient = np.array([[4, 1, 1], [1, 5, 2], [1, 2, 4]])
+ >>> constant = np.array([[2], [-6]])
+ >>> init_val = [0.5, -0.5, -0.5]
+ >>> iterations = 3
+ >>> jacobi_iteration_method(
+ ... coefficient, constant, init_val, iterations
+ ... ) # doctest: +NORMALIZE_WHITESPACE
+ Traceback (most recent call last):
+ ...
+ ValueError: Coefficient and constant matrices dimensions must be nxn and nx1 but
+ received 3x3 and 2x1
+
+ >>> coefficient = np.array([[4, 1, 1], [1, 5, 2], [1, 2, 4]])
+ >>> constant = np.array([[2], [-6], [-4]])
+ >>> init_val = [0.5, -0.5]
+ >>> iterations = 3
+ >>> jacobi_iteration_method(
+ ... coefficient, constant, init_val, iterations
+ ... ) # doctest: +NORMALIZE_WHITESPACE
+ Traceback (most recent call last):
+ ...
+ ValueError: Number of initial values must be equal to number of rows in coefficient
+ matrix but received 2 and 3
+
+ >>> coefficient = np.array([[4, 1, 1], [1, 5, 2], [1, 2, 4]])
+ >>> constant = np.array([[2], [-6], [-4]])
+ >>> init_val = [0.5, -0.5, -0.5]
+ >>> iterations = 0
+ >>> jacobi_iteration_method(coefficient, constant, init_val, iterations)
+ Traceback (most recent call last):
+ ...
+ ValueError: Iterations must be at least 1
+ """
+
+ rows1, cols1 = coefficient_matrix.shape
+ rows2, cols2 = constant_matrix.shape
+
+ if rows1 != cols1:
+ msg = f"Coefficient matrix dimensions must be nxn but received {rows1}x{cols1}"
+ raise ValueError(msg)
+
+ if cols2 != 1:
+ msg = f"Constant matrix must be nx1 but received {rows2}x{cols2}"
+ raise ValueError(msg)
+
+ if rows1 != rows2:
+ msg = (
+ "Coefficient and constant matrices dimensions must be nxn and nx1 but "
+ f"received {rows1}x{cols1} and {rows2}x{cols2}"
+ )
+ raise ValueError(msg)
+
+ if len(init_val) != rows1:
+ msg = (
+ "Number of initial values must be equal to number of rows in coefficient "
+ f"matrix but received {len(init_val)} and {rows1}"
+ )
+ raise ValueError(msg)
+
+ if iterations <= 0:
+ raise ValueError("Iterations must be at least 1")
+
+ table: NDArray[float64] = np.concatenate(
+ (coefficient_matrix, constant_matrix), axis=1
+ )
+
+ rows, cols = table.shape
+
+ strictly_diagonally_dominant(table)
+
+ """
+ # Iterates the whole matrix for given number of times
+ for _ in range(iterations):
+ new_val = []
+ for row in range(rows):
+ temp = 0
+ for col in range(cols):
+ if col == row:
+ denom = table[row][col]
+ elif col == cols - 1:
+ val = table[row][col]
+ else:
+ temp += (-1) * table[row][col] * init_val[col]
+ temp = (temp + val) / denom
+ new_val.append(temp)
+ init_val = new_val
+ """
+
+ # denominator - a list of values along the diagonal
+ denominator = np.diag(coefficient_matrix)
+
+ # val_last - values of the last column of the table array
+ val_last = table[:, -1]
+
+ # masks - boolean mask of all strings without diagonal
+ # elements array coefficient_matrix
+ masks = ~np.eye(coefficient_matrix.shape[0], dtype=bool)
+
+ # no_diagonals - coefficient_matrix array values without diagonal elements
+ no_diagonals = coefficient_matrix[masks].reshape(-1, rows - 1)
+
+ # Here we get 'i_col' - these are the column numbers, for each row
+ # without diagonal elements, except for the last column.
+ i_row, i_col = np.where(masks)
+ ind = i_col.reshape(-1, rows - 1)
+
+ #'i_col' is converted to a two-dimensional list 'ind', which will be
+ # used to make selections from 'init_val' ('arr' array see below).
+
+ # Iterates the whole matrix for given number of times
+ for _ in range(iterations):
+ arr = np.take(init_val, ind)
+ sum_product_rows = np.sum((-1) * no_diagonals * arr, axis=1)
+ new_val = (sum_product_rows + val_last) / denominator
+ init_val = new_val
+
+ return new_val.tolist()
+
+
+# Checks if the given matrix is strictly diagonally dominant
+def strictly_diagonally_dominant(table: NDArray[float64]) -> bool:
+ """
+ >>> table = np.array([[4, 1, 1, 2], [1, 5, 2, -6], [1, 2, 4, -4]])
+ >>> strictly_diagonally_dominant(table)
+ True
+
+ >>> table = np.array([[4, 1, 1, 2], [1, 5, 2, -6], [1, 2, 3, -4]])
+ >>> strictly_diagonally_dominant(table)
+ Traceback (most recent call last):
+ ...
+ ValueError: Coefficient matrix is not strictly diagonally dominant
+ """
+
+ rows, cols = table.shape
+
+ is_diagonally_dominant = True
+
+ for i in range(rows):
+ total = 0
+ for j in range(cols - 1):
+ if i == j:
+ continue
+ else:
+ total += table[i][j]
+
+ if table[i][i] <= total:
+ raise ValueError("Coefficient matrix is not strictly diagonally dominant")
+
+ return is_diagonally_dominant
+
+
+# Test Cases
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/linear_algebra/lu_decomposition.py b/linear_algebra/lu_decomposition.py
new file mode 100644
index 000000000000..3d89b53a48fb
--- /dev/null
+++ b/linear_algebra/lu_decomposition.py
@@ -0,0 +1,114 @@
+"""
+Lower-upper (LU) decomposition factors a matrix as a product of a lower
+triangular matrix and an upper triangular matrix. A square matrix has an LU
+decomposition under the following conditions:
+
+ - If the matrix is invertible, then it has an LU decomposition if and only
+ if all of its leading principal minors are non-zero (see
+ https://en.wikipedia.org/wiki/Minor_(linear_algebra) for an explanation of
+ leading principal minors of a matrix).
+ - If the matrix is singular (i.e., not invertible) and it has a rank of k
+ (i.e., it has k linearly independent columns), then it has an LU
+ decomposition if its first k leading principal minors are non-zero.
+
+This algorithm will simply attempt to perform LU decomposition on any square
+matrix and raise an error if no such decomposition exists.
+
+Reference: https://en.wikipedia.org/wiki/LU_decomposition
+"""
+
+from __future__ import annotations
+
+import numpy as np
+
+
+def lower_upper_decomposition(table: np.ndarray) -> tuple[np.ndarray, np.ndarray]:
+ """
+ Perform LU decomposition on a given matrix and raises an error if the matrix
+ isn't square or if no such decomposition exists
+
+ >>> matrix = np.array([[2, -2, 1], [0, 1, 2], [5, 3, 1]])
+ >>> lower_mat, upper_mat = lower_upper_decomposition(matrix)
+ >>> lower_mat
+ array([[1. , 0. , 0. ],
+ [0. , 1. , 0. ],
+ [2.5, 8. , 1. ]])
+ >>> upper_mat
+ array([[ 2. , -2. , 1. ],
+ [ 0. , 1. , 2. ],
+ [ 0. , 0. , -17.5]])
+
+ >>> matrix = np.array([[4, 3], [6, 3]])
+ >>> lower_mat, upper_mat = lower_upper_decomposition(matrix)
+ >>> lower_mat
+ array([[1. , 0. ],
+ [1.5, 1. ]])
+ >>> upper_mat
+ array([[ 4. , 3. ],
+ [ 0. , -1.5]])
+
+ >>> # Matrix is not square
+ >>> matrix = np.array([[2, -2, 1], [0, 1, 2]])
+ >>> lower_mat, upper_mat = lower_upper_decomposition(matrix)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'table' has to be of square shaped array but got a 2x3 array:
+ [[ 2 -2 1]
+ [ 0 1 2]]
+
+ >>> # Matrix is invertible, but its first leading principal minor is 0
+ >>> matrix = np.array([[0, 1], [1, 0]])
+ >>> lower_mat, upper_mat = lower_upper_decomposition(matrix)
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: No LU decomposition exists
+
+ >>> # Matrix is singular, but its first leading principal minor is 1
+ >>> matrix = np.array([[1, 0], [1, 0]])
+ >>> lower_mat, upper_mat = lower_upper_decomposition(matrix)
+ >>> lower_mat
+ array([[1., 0.],
+ [1., 1.]])
+ >>> upper_mat
+ array([[1., 0.],
+ [0., 0.]])
+
+ >>> # Matrix is singular, but its first leading principal minor is 0
+ >>> matrix = np.array([[0, 1], [0, 1]])
+ >>> lower_mat, upper_mat = lower_upper_decomposition(matrix)
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: No LU decomposition exists
+ """
+ # Ensure that table is a square array
+ rows, columns = np.shape(table)
+ if rows != columns:
+ msg = (
+ "'table' has to be of square shaped array but got a "
+ f"{rows}x{columns} array:\n{table}"
+ )
+ raise ValueError(msg)
+
+ lower = np.zeros((rows, columns))
+ upper = np.zeros((rows, columns))
+
+ # in 'total', the necessary data is extracted through slices
+ # and the sum of the products is obtained.
+
+ for i in range(columns):
+ for j in range(i):
+ total = np.sum(lower[i, :i] * upper[:i, j])
+ if upper[j][j] == 0:
+ raise ArithmeticError("No LU decomposition exists")
+ lower[i][j] = (table[i][j] - total) / upper[j][j]
+ lower[i][i] = 1
+ for j in range(i, columns):
+ total = np.sum(lower[i, :i] * upper[:i, j])
+ upper[i][j] = table[i][j] - total
+ return lower, upper
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/linear_algebra/src/conjugate_gradient.py b/linear_algebra/src/conjugate_gradient.py
new file mode 100644
index 000000000000..45da35813978
--- /dev/null
+++ b/linear_algebra/src/conjugate_gradient.py
@@ -0,0 +1,180 @@
+"""
+Resources:
+- https://en.wikipedia.org/wiki/Conjugate_gradient_method
+- https://en.wikipedia.org/wiki/Definite_symmetric_matrix
+"""
+
+from typing import Any
+
+import numpy as np
+
+
+def _is_matrix_spd(matrix: np.ndarray) -> bool:
+ """
+ Returns True if input matrix is symmetric positive definite.
+ Returns False otherwise.
+
+ For a matrix to be SPD, all eigenvalues must be positive.
+
+ >>> import numpy as np
+ >>> matrix = np.array([
+ ... [4.12401784, -5.01453636, -0.63865857],
+ ... [-5.01453636, 12.33347422, -3.40493586],
+ ... [-0.63865857, -3.40493586, 5.78591885]])
+ >>> _is_matrix_spd(matrix)
+ True
+ >>> matrix = np.array([
+ ... [0.34634879, 1.96165514, 2.18277744],
+ ... [0.74074469, -1.19648894, -1.34223498],
+ ... [-0.7687067 , 0.06018373, -1.16315631]])
+ >>> _is_matrix_spd(matrix)
+ False
+ """
+ # Ensure matrix is square.
+ assert np.shape(matrix)[0] == np.shape(matrix)[1]
+
+ # If matrix not symmetric, exit right away.
+ if np.allclose(matrix, matrix.T) is False:
+ return False
+
+ # Get eigenvalues and eignevectors for a symmetric matrix.
+ eigen_values, _ = np.linalg.eigh(matrix)
+
+ # Check sign of all eigenvalues.
+ # np.all returns a value of type np.bool_
+ return bool(np.all(eigen_values > 0))
+
+
+def _create_spd_matrix(dimension: int) -> Any:
+ """
+ Returns a symmetric positive definite matrix given a dimension.
+
+ Input:
+ dimension gives the square matrix dimension.
+
+ Output:
+ spd_matrix is an diminesion x dimensions symmetric positive definite (SPD) matrix.
+
+ >>> import numpy as np
+ >>> dimension = 3
+ >>> spd_matrix = _create_spd_matrix(dimension)
+ >>> _is_matrix_spd(spd_matrix)
+ True
+ """
+ rng = np.random.default_rng()
+ random_matrix = rng.normal(size=(dimension, dimension))
+ spd_matrix = np.dot(random_matrix, random_matrix.T)
+ assert _is_matrix_spd(spd_matrix)
+ return spd_matrix
+
+
+def conjugate_gradient(
+ spd_matrix: np.ndarray,
+ load_vector: np.ndarray,
+ max_iterations: int = 1000,
+ tol: float = 1e-8,
+) -> Any:
+ """
+ Returns solution to the linear system np.dot(spd_matrix, x) = b.
+
+ Input:
+ spd_matrix is an NxN Symmetric Positive Definite (SPD) matrix.
+ load_vector is an Nx1 vector.
+
+ Output:
+ x is an Nx1 vector that is the solution vector.
+
+ >>> import numpy as np
+ >>> spd_matrix = np.array([
+ ... [8.73256573, -5.02034289, -2.68709226],
+ ... [-5.02034289, 3.78188322, 0.91980451],
+ ... [-2.68709226, 0.91980451, 1.94746467]])
+ >>> b = np.array([
+ ... [-5.80872761],
+ ... [ 3.23807431],
+ ... [ 1.95381422]])
+ >>> conjugate_gradient(spd_matrix, b)
+ array([[-0.63114139],
+ [-0.01561498],
+ [ 0.13979294]])
+ """
+ # Ensure proper dimensionality.
+ assert np.shape(spd_matrix)[0] == np.shape(spd_matrix)[1]
+ assert np.shape(load_vector)[0] == np.shape(spd_matrix)[0]
+ assert _is_matrix_spd(spd_matrix)
+
+ # Initialize solution guess, residual, search direction.
+ x0 = np.zeros((np.shape(load_vector)[0], 1))
+ r0 = np.copy(load_vector)
+ p0 = np.copy(r0)
+
+ # Set initial errors in solution guess and residual.
+ error_residual = 1e9
+ error_x_solution = 1e9
+ error = 1e9
+
+ # Set iteration counter to threshold number of iterations.
+ iterations = 0
+
+ while error > tol:
+ # Save this value so we only calculate the matrix-vector product once.
+ w = np.dot(spd_matrix, p0)
+
+ # The main algorithm.
+
+ # Update search direction magnitude.
+ alpha = np.dot(r0.T, r0) / np.dot(p0.T, w)
+ # Update solution guess.
+ x = x0 + alpha * p0
+ # Calculate new residual.
+ r = r0 - alpha * w
+ # Calculate new Krylov subspace scale.
+ beta = np.dot(r.T, r) / np.dot(r0.T, r0)
+ # Calculate new A conjuage search direction.
+ p = r + beta * p0
+
+ # Calculate errors.
+ error_residual = np.linalg.norm(r - r0)
+ error_x_solution = np.linalg.norm(x - x0)
+ error = np.maximum(error_residual, error_x_solution)
+
+ # Update variables.
+ x0 = np.copy(x)
+ r0 = np.copy(r)
+ p0 = np.copy(p)
+
+ # Update number of iterations.
+ iterations += 1
+ if iterations > max_iterations:
+ break
+
+ return x
+
+
+def test_conjugate_gradient() -> None:
+ """
+ >>> test_conjugate_gradient() # self running tests
+ """
+ # Create linear system with SPD matrix and known solution x_true.
+ dimension = 3
+ spd_matrix = _create_spd_matrix(dimension)
+ rng = np.random.default_rng()
+ x_true = rng.normal(size=(dimension, 1))
+ b = np.dot(spd_matrix, x_true)
+
+ # Numpy solution.
+ x_numpy = np.linalg.solve(spd_matrix, b)
+
+ # Our implementation.
+ x_conjugate_gradient = conjugate_gradient(spd_matrix, b)
+
+ # Ensure both solutions are close to x_true (and therefore one another).
+ assert np.linalg.norm(x_numpy - x_true) <= 1e-6
+ assert np.linalg.norm(x_conjugate_gradient - x_true) <= 1e-6
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ test_conjugate_gradient()
diff --git a/linear_algebra/src/gaussian_elimination_pivoting.py b/linear_algebra/src/gaussian_elimination_pivoting.py
new file mode 100644
index 000000000000..540f57b0cff6
--- /dev/null
+++ b/linear_algebra/src/gaussian_elimination_pivoting.py
@@ -0,0 +1,88 @@
+import numpy as np
+
+
+def solve_linear_system(matrix: np.ndarray) -> np.ndarray:
+ """
+ Solve a linear system of equations using Gaussian elimination with partial pivoting
+
+ Args:
+ - `matrix`: Coefficient matrix with the last column representing the constants.
+
+ Returns:
+ - Solution vector.
+
+ Raises:
+ - ``ValueError``: If the matrix is not correct (i.e., singular).
+
+ https://courses.engr.illinois.edu/cs357/su2013/lect.htm Lecture 7
+
+ Example:
+
+ >>> A = np.array([[2, 1, -1], [-3, -1, 2], [-2, 1, 2]], dtype=float)
+ >>> B = np.array([8, -11, -3], dtype=float)
+ >>> solution = solve_linear_system(np.column_stack((A, B)))
+ >>> np.allclose(solution, np.array([2., 3., -1.]))
+ True
+ >>> solve_linear_system(np.array([[0, 0, 0]], dtype=float))
+ Traceback (most recent call last):
+ ...
+ ValueError: Matrix is not square
+ >>> solve_linear_system(np.array([[0, 0, 0], [0, 0, 0]], dtype=float))
+ Traceback (most recent call last):
+ ...
+ ValueError: Matrix is singular
+ """
+ ab = np.copy(matrix)
+ num_of_rows = ab.shape[0]
+ num_of_columns = ab.shape[1] - 1
+ x_lst: list[float] = []
+
+ if num_of_rows != num_of_columns:
+ raise ValueError("Matrix is not square")
+
+ for column_num in range(num_of_rows):
+ # Lead element search
+ for i in range(column_num, num_of_columns):
+ if abs(ab[i][column_num]) > abs(ab[column_num][column_num]):
+ ab[[column_num, i]] = ab[[i, column_num]]
+
+ # Upper triangular matrix
+ if abs(ab[column_num, column_num]) < 1e-8:
+ raise ValueError("Matrix is singular")
+
+ if column_num != 0:
+ for i in range(column_num, num_of_rows):
+ ab[i, :] -= (
+ ab[i, column_num - 1]
+ / ab[column_num - 1, column_num - 1]
+ * ab[column_num - 1, :]
+ )
+
+ # Find x vector (Back Substitution)
+ for column_num in range(num_of_rows - 1, -1, -1):
+ x = ab[column_num, -1] / ab[column_num, column_num]
+ x_lst.insert(0, x)
+ for i in range(column_num - 1, -1, -1):
+ ab[i, -1] -= ab[i, column_num] * x
+
+ # Return the solution vector
+ return np.asarray(x_lst)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+
+ example_matrix = np.array(
+ [
+ [5.0, -5.0, -3.0, 4.0, -11.0],
+ [1.0, -4.0, 6.0, -4.0, -10.0],
+ [-2.0, -5.0, 4.0, -5.0, -12.0],
+ [-3.0, -3.0, 5.0, -5.0, 8.0],
+ ],
+ dtype=float,
+ )
+
+ print(f"Matrix:\n{example_matrix}")
+ print(f"{solve_linear_system(example_matrix) = }")
diff --git a/linear_algebra/src/lib.py b/linear_algebra/src/lib.py
index 353c8334093b..0d6a348475cd 100644
--- a/linear_algebra/src/lib.py
+++ b/linear_algebra/src/lib.py
@@ -10,18 +10,21 @@
Overview:
- class Vector
-- function zeroVector(dimension)
-- function unitBasisVector(dimension,pos)
-- function axpy(scalar,vector1,vector2)
-- function randomVector(N,a,b)
+- function zero_vector(dimension)
+- function unit_basis_vector(dimension, pos)
+- function axpy(scalar, vector1, vector2)
+- function random_vector(N, a, b)
- class Matrix
-- function squareZeroMatrix(N)
-- function randomMatrix(W,H,a,b)
+- function square_zero_matrix(N)
+- function random_matrix(W, H, a, b)
"""
+from __future__ import annotations
import math
import random
+from collections.abc import Collection
+from typing import overload
class Vector:
@@ -29,23 +32,23 @@ class Vector:
This class represents a vector of arbitrary size.
You need to give the vector components.
- Overview about the methods:
-
- constructor(components : list) : init the vector
- set(components : list) : changes the vector components.
- __str__() : toString method
- component(i : int): gets the i-th component (start by 0)
- __len__() : gets the size of the vector (number of components)
- euclidLength() : returns the euclidean length of the vector.
- operator + : vector addition
- operator - : vector subtraction
- operator * : scalar multiplication and dot product
- copy() : copies this vector and returns it.
- changeComponent(pos,value) : changes the specified component.
- TODO: compare-operator
+ Overview of the methods:
+
+ __init__(components: Collection[float] | None): init the vector
+ __len__(): gets the size of the vector (number of components)
+ __str__(): returns a string representation
+ __add__(other: Vector): vector addition
+ __sub__(other: Vector): vector subtraction
+ __mul__(other: float): scalar multiplication
+ __mul__(other: Vector): dot product
+ copy(): copies this vector and returns it
+ component(i): gets the i-th component (0-indexed)
+ change_component(pos: int, value: float): changes specified component
+ euclidean_length(): returns the euclidean length of the vector
+ angle(other: Vector, deg: bool): returns the angle between two vectors
"""
- def __init__(self, components=None):
+ def __init__(self, components: Collection[float] | None = None) -> None:
"""
input: components or nothing
simple constructor for init the vector
@@ -54,49 +57,19 @@ def __init__(self, components=None):
components = []
self.__components = list(components)
- def set(self, components):
- """
- input: new components
- changes the components of the vector.
- replace the components with newer one.
- """
- if len(components) > 0:
- self.__components = list(components)
- else:
- raise Exception("please give any vector")
-
- def __str__(self):
- """
- returns a string representation of the vector
- """
- return "(" + ",".join(map(str, self.__components)) + ")"
-
- def component(self, i):
- """
- input: index (start at 0)
- output: the i-th component of the vector.
- """
- if type(i) is int and -len(self.__components) <= i < len(self.__components):
- return self.__components[i]
- else:
- raise Exception("index out of range")
-
- def __len__(self):
+ def __len__(self) -> int:
"""
returns the size of the vector
"""
return len(self.__components)
- def euclidLength(self):
+ def __str__(self) -> str:
"""
- returns the euclidean length of the vector
+ returns a string representation of the vector
"""
- summe = 0
- for c in self.__components:
- summe += c ** 2
- return math.sqrt(summe)
+ return "(" + ",".join(map(str, self.__components)) + ")"
- def __add__(self, other):
+ def __add__(self, other: Vector) -> Vector:
"""
input: other vector
assumes: other vector has the same size
@@ -109,7 +82,7 @@ def __add__(self, other):
else:
raise Exception("must have the same size")
- def __sub__(self, other):
+ def __sub__(self, other: Vector) -> Vector:
"""
input: other vector
assumes: other vector has the same size
@@ -122,30 +95,54 @@ def __sub__(self, other):
else: # error case
raise Exception("must have the same size")
- def __mul__(self, other):
+ def __eq__(self, other: object) -> bool:
+ """
+ performs the comparison between two vectors
+ """
+ if not isinstance(other, Vector):
+ return NotImplemented
+ if len(self) != len(other):
+ return False
+ return all(self.component(i) == other.component(i) for i in range(len(self)))
+
+ @overload
+ def __mul__(self, other: float) -> Vector: ...
+
+ @overload
+ def __mul__(self, other: Vector) -> float: ...
+
+ def __mul__(self, other: float | Vector) -> float | Vector:
"""
mul implements the scalar multiplication
and the dot-product
"""
- if isinstance(other, float) or isinstance(other, int):
+ if isinstance(other, (float, int)):
ans = [c * other for c in self.__components]
return Vector(ans)
- elif isinstance(other, Vector) and (len(self) == len(other)):
+ elif isinstance(other, Vector) and len(self) == len(other):
size = len(self)
- summe = 0
- for i in range(size):
- summe += self.__components[i] * other.component(i)
- return summe
+ prods = [self.__components[i] * other.component(i) for i in range(size)]
+ return sum(prods)
else: # error case
raise Exception("invalid operand!")
- def copy(self):
+ def copy(self) -> Vector:
"""
copies this vector and returns it.
"""
return Vector(self.__components)
- def changeComponent(self, pos, value):
+ def component(self, i: int) -> float:
+ """
+ input: index (0-indexed)
+ output: the i-th component of the vector.
+ """
+ if isinstance(i, int) and -len(self.__components) <= i < len(self.__components):
+ return self.__components[i]
+ else:
+ raise Exception("index out of range")
+
+ def change_component(self, pos: int, value: float) -> None:
"""
input: an index (pos) and a value
changes the specified component (pos) with the
@@ -155,8 +152,48 @@ def changeComponent(self, pos, value):
assert -len(self.__components) <= pos < len(self.__components)
self.__components[pos] = value
+ def euclidean_length(self) -> float:
+ """
+ returns the euclidean length of the vector
-def zeroVector(dimension):
+ >>> Vector([2, 3, 4]).euclidean_length()
+ 5.385164807134504
+ >>> Vector([1]).euclidean_length()
+ 1.0
+ >>> Vector([0, -1, -2, -3, 4, 5, 6]).euclidean_length()
+ 9.539392014169456
+ >>> Vector([]).euclidean_length()
+ Traceback (most recent call last):
+ ...
+ Exception: Vector is empty
+ """
+ if len(self.__components) == 0:
+ raise Exception("Vector is empty")
+ squares = [c**2 for c in self.__components]
+ return math.sqrt(sum(squares))
+
+ def angle(self, other: Vector, deg: bool = False) -> float:
+ """
+ find angle between two Vector (self, Vector)
+
+ >>> Vector([3, 4, -1]).angle(Vector([2, -1, 1]))
+ 1.4906464636572374
+ >>> Vector([3, 4, -1]).angle(Vector([2, -1, 1]), deg = True)
+ 85.40775111366095
+ >>> Vector([3, 4, -1]).angle(Vector([2, -1]))
+ Traceback (most recent call last):
+ ...
+ Exception: invalid operand!
+ """
+ num = self * other
+ den = self.euclidean_length() * other.euclidean_length()
+ if deg:
+ return math.degrees(math.acos(num / den))
+ else:
+ return math.acos(num / den)
+
+
+def zero_vector(dimension: int) -> Vector:
"""
returns a zero-vector of size 'dimension'
"""
@@ -165,34 +202,33 @@ def zeroVector(dimension):
return Vector([0] * dimension)
-def unitBasisVector(dimension, pos):
+def unit_basis_vector(dimension: int, pos: int) -> Vector:
"""
returns a unit basis vector with a One
at index 'pos' (indexing at 0)
"""
# precondition
- assert isinstance(dimension, int) and (isinstance(pos, int))
+ assert isinstance(dimension, int)
+ assert isinstance(pos, int)
ans = [0] * dimension
ans[pos] = 1
return Vector(ans)
-def axpy(scalar, x, y):
+def axpy(scalar: float, x: Vector, y: Vector) -> Vector:
"""
input: a 'scalar' and two vectors 'x' and 'y'
output: a vector
computes the axpy operation
"""
# precondition
- assert (
- isinstance(x, Vector)
- and (isinstance(y, Vector))
- and (isinstance(scalar, int) or isinstance(scalar, float))
- )
+ assert isinstance(x, Vector)
+ assert isinstance(y, Vector)
+ assert isinstance(scalar, (int, float))
return x * scalar + y
-def randomVector(N, a, b):
+def random_vector(n: int, a: int, b: int) -> Vector:
"""
input: size (N) of the vector.
random range (a,b)
@@ -200,41 +236,43 @@ def randomVector(N, a, b):
random integer components between 'a' and 'b'.
"""
random.seed(None)
- ans = [random.randint(a, b) for i in range(N)]
+ ans = [random.randint(a, b) for _ in range(n)]
return Vector(ans)
class Matrix:
"""
class: Matrix
- This class represents a arbitrary matrix.
-
- Overview about the methods:
-
- __str__() : returns a string representation
- operator * : implements the matrix vector multiplication
- implements the matrix-scalar multiplication.
- changeComponent(x,y,value) : changes the specified component.
- component(x,y) : returns the specified component.
- width() : returns the width of the matrix
- height() : returns the height of the matrix
- operator + : implements the matrix-addition.
- operator - _ implements the matrix-subtraction
+ This class represents an arbitrary matrix.
+
+ Overview of the methods:
+
+ __init__():
+ __str__(): returns a string representation
+ __add__(other: Matrix): matrix addition
+ __sub__(other: Matrix): matrix subtraction
+ __mul__(other: float): scalar multiplication
+ __mul__(other: Vector): vector multiplication
+ height() : returns height
+ width() : returns width
+ component(x: int, y: int): returns specified component
+ change_component(x: int, y: int, value: float): changes specified component
+ minor(x: int, y: int): returns minor along (x, y)
+ cofactor(x: int, y: int): returns cofactor along (x, y)
+ determinant() : returns determinant
"""
- def __init__(self, matrix, w, h):
+ def __init__(self, matrix: list[list[float]], w: int, h: int) -> None:
"""
- simple constructor for initializing
- the matrix with components.
+ simple constructor for initializing the matrix with components.
"""
self.__matrix = matrix
self.__width = w
self.__height = h
- def __str__(self):
+ def __str__(self) -> str:
"""
- returns a string representation of this
- matrix.
+ returns a string representation of this matrix.
"""
ans = ""
for i in range(self.__height):
@@ -246,134 +284,161 @@ def __str__(self):
ans += str(self.__matrix[i][j]) + "|\n"
return ans
- def changeComponent(self, x, y, value):
+ def __add__(self, other: Matrix) -> Matrix:
"""
- changes the x-y component of this matrix
+ implements matrix addition.
"""
- if 0 <= x < self.__height and 0 <= y < self.__width:
- self.__matrix[x][y] = value
+ if self.__width == other.width() and self.__height == other.height():
+ matrix = []
+ for i in range(self.__height):
+ row = [
+ self.__matrix[i][j] + other.component(i, j)
+ for j in range(self.__width)
+ ]
+ matrix.append(row)
+ return Matrix(matrix, self.__width, self.__height)
else:
- raise Exception("changeComponent: indices out of bounds")
+ raise Exception("matrix must have the same dimension!")
- def component(self, x, y):
+ def __sub__(self, other: Matrix) -> Matrix:
"""
- returns the specified (x,y) component
+ implements matrix subtraction.
"""
- if 0 <= x < self.__height and 0 <= y < self.__width:
- return self.__matrix[x][y]
+ if self.__width == other.width() and self.__height == other.height():
+ matrix = []
+ for i in range(self.__height):
+ row = [
+ self.__matrix[i][j] - other.component(i, j)
+ for j in range(self.__width)
+ ]
+ matrix.append(row)
+ return Matrix(matrix, self.__width, self.__height)
else:
- raise Exception("changeComponent: indices out of bounds")
+ raise Exception("matrices must have the same dimension!")
- def width(self):
- """
- getter for the width
- """
- return self.__width
+ @overload
+ def __mul__(self, other: float) -> Matrix: ...
- def height(self):
- """
- getter for the height
- """
- return self.__height
-
- def determinate(self) -> float:
- """
- returns the determinate of an nxn matrix using Laplace expansion
- """
- if self.__height == self.__width and self.__width >= 2:
- total = 0
- if self.__width > 2:
- for x in range(0, self.__width):
- for y in range(0, self.__height):
- total += (
- self.__matrix[x][y]
- * (-1) ** (x + y)
- * Matrix(
- self.__matrix[0:x] + self.__matrix[x + 1 :],
- self.__width - 1,
- self.__height - 1,
- ).determinate()
- )
- else:
- return (
- self.__matrix[0][0] * self.__matrix[1][1]
- - self.__matrix[0][1] * self.__matrix[1][0]
- )
- return total
- else:
- raise Exception("matrix is not square")
+ @overload
+ def __mul__(self, other: Vector) -> Vector: ...
- def __mul__(self, other):
+ def __mul__(self, other: float | Vector) -> Vector | Matrix:
"""
implements the matrix-vector multiplication.
implements the matrix-scalar multiplication
"""
- if isinstance(other, Vector): # vector-matrix
+ if isinstance(other, Vector): # matrix-vector
if len(other) == self.__width:
- ans = zeroVector(self.__height)
+ ans = zero_vector(self.__height)
for i in range(self.__height):
- summe = 0
- for j in range(self.__width):
- summe += other.component(j) * self.__matrix[i][j]
- ans.changeComponent(i, summe)
- summe = 0
+ prods = [
+ self.__matrix[i][j] * other.component(j)
+ for j in range(self.__width)
+ ]
+ ans.change_component(i, sum(prods))
return ans
else:
raise Exception(
"vector must have the same size as the "
- + "number of columns of the matrix!"
+ "number of columns of the matrix!"
)
- elif isinstance(other, int) or isinstance(other, float): # matrix-scalar
+ elif isinstance(other, (int, float)): # matrix-scalar
matrix = [
[self.__matrix[i][j] * other for j in range(self.__width)]
for i in range(self.__height)
]
return Matrix(matrix, self.__width, self.__height)
+ return None
- def __add__(self, other):
+ def height(self) -> int:
"""
- implements the matrix-addition.
+ getter for the height
"""
- if self.__width == other.width() and self.__height == other.height():
- matrix = []
- for i in range(self.__height):
- row = []
- for j in range(self.__width):
- row.append(self.__matrix[i][j] + other.component(i, j))
- matrix.append(row)
- return Matrix(matrix, self.__width, self.__height)
+ return self.__height
+
+ def width(self) -> int:
+ """
+ getter for the width
+ """
+ return self.__width
+
+ def component(self, x: int, y: int) -> float:
+ """
+ returns the specified (x,y) component
+ """
+ if 0 <= x < self.__height and 0 <= y < self.__width:
+ return self.__matrix[x][y]
else:
- raise Exception("matrix must have the same dimension!")
+ raise Exception("change_component: indices out of bounds")
- def __sub__(self, other):
+ def change_component(self, x: int, y: int, value: float) -> None:
"""
- implements the matrix-subtraction.
+ changes the x-y component of this matrix
"""
- if self.__width == other.width() and self.__height == other.height():
- matrix = []
- for i in range(self.__height):
- row = []
- for j in range(self.__width):
- row.append(self.__matrix[i][j] - other.component(i, j))
- matrix.append(row)
- return Matrix(matrix, self.__width, self.__height)
+ if 0 <= x < self.__height and 0 <= y < self.__width:
+ self.__matrix[x][y] = value
else:
- raise Exception("matrix must have the same dimension!")
+ raise Exception("change_component: indices out of bounds")
+
+ def minor(self, x: int, y: int) -> float:
+ """
+ returns the minor along (x, y)
+ """
+ if self.__height != self.__width:
+ raise Exception("Matrix is not square")
+ minor = self.__matrix[:x] + self.__matrix[x + 1 :]
+ for i in range(len(minor)):
+ minor[i] = minor[i][:y] + minor[i][y + 1 :]
+ return Matrix(minor, self.__width - 1, self.__height - 1).determinant()
+
+ def cofactor(self, x: int, y: int) -> float:
+ """
+ returns the cofactor (signed minor) along (x, y)
+ """
+ if self.__height != self.__width:
+ raise Exception("Matrix is not square")
+ if 0 <= x < self.__height and 0 <= y < self.__width:
+ return (-1) ** (x + y) * self.minor(x, y)
+ else:
+ raise Exception("Indices out of bounds")
+
+ def determinant(self) -> float:
+ """
+ returns the determinant of an nxn matrix using Laplace expansion
+ """
+ if self.__height != self.__width:
+ raise Exception("Matrix is not square")
+ if self.__height < 1:
+ raise Exception("Matrix has no element")
+ elif self.__height == 1:
+ return self.__matrix[0][0]
+ elif self.__height == 2:
+ return (
+ self.__matrix[0][0] * self.__matrix[1][1]
+ - self.__matrix[0][1] * self.__matrix[1][0]
+ )
+ else:
+ cofactor_prods = [
+ self.__matrix[0][y] * self.cofactor(0, y) for y in range(self.__width)
+ ]
+ return sum(cofactor_prods)
-def squareZeroMatrix(N):
+def square_zero_matrix(n: int) -> Matrix:
"""
returns a square zero-matrix of dimension NxN
"""
- ans = [[0] * N for i in range(N)]
- return Matrix(ans, N, N)
+ ans: list[list[float]] = [[0] * n for _ in range(n)]
+ return Matrix(ans, n, n)
-def randomMatrix(W, H, a, b):
+def random_matrix(width: int, height: int, a: int, b: int) -> Matrix:
"""
returns a random matrix WxH with integer components
between 'a' and 'b'
"""
random.seed(None)
- matrix = [[random.randint(a, b) for j in range(W)] for i in range(H)]
- return Matrix(matrix, W, H)
+ matrix: list[list[float]] = [
+ [random.randint(a, b) for _ in range(width)] for _ in range(height)
+ ]
+ return Matrix(matrix, width, height)
diff --git a/linear_algebra/src/polynom_for_points.py b/linear_algebra/src/polynom_for_points.py
index 7a363723d9d2..452f3edd4aee 100644
--- a/linear_algebra/src/polynom_for_points.py
+++ b/linear_algebra/src/polynom_for_points.py
@@ -1,122 +1,93 @@
-from __future__ import annotations
-
-
def points_to_polynomial(coordinates: list[list[int]]) -> str:
"""
coordinates is a two dimensional matrix: [[x, y], [x, y], ...]
number of points you want to use
- >>> print(points_to_polynomial([]))
- The program cannot work out a fitting polynomial.
- >>> print(points_to_polynomial([[]]))
- The program cannot work out a fitting polynomial.
- >>> print(points_to_polynomial([[1, 0], [2, 0], [3, 0]]))
- f(x)=x^2*0.0+x^1*-0.0+x^0*0.0
- >>> print(points_to_polynomial([[1, 1], [2, 1], [3, 1]]))
- f(x)=x^2*0.0+x^1*-0.0+x^0*1.0
- >>> print(points_to_polynomial([[1, 3], [2, 3], [3, 3]]))
- f(x)=x^2*0.0+x^1*-0.0+x^0*3.0
- >>> print(points_to_polynomial([[1, 1], [2, 2], [3, 3]]))
- f(x)=x^2*0.0+x^1*1.0+x^0*0.0
- >>> print(points_to_polynomial([[1, 1], [2, 4], [3, 9]]))
- f(x)=x^2*1.0+x^1*-0.0+x^0*0.0
- >>> print(points_to_polynomial([[1, 3], [2, 6], [3, 11]]))
- f(x)=x^2*1.0+x^1*-0.0+x^0*2.0
- >>> print(points_to_polynomial([[1, -3], [2, -6], [3, -11]]))
- f(x)=x^2*-1.0+x^1*-0.0+x^0*-2.0
- >>> print(points_to_polynomial([[1, 5], [2, 2], [3, 9]]))
- f(x)=x^2*5.0+x^1*-18.0+x^0*18.0
+ >>> points_to_polynomial([])
+ Traceback (most recent call last):
+ ...
+ ValueError: The program cannot work out a fitting polynomial.
+ >>> points_to_polynomial([[]])
+ Traceback (most recent call last):
+ ...
+ ValueError: The program cannot work out a fitting polynomial.
+ >>> points_to_polynomial([[1, 0], [2, 0], [3, 0]])
+ 'f(x)=x^2*0.0+x^1*-0.0+x^0*0.0'
+ >>> points_to_polynomial([[1, 1], [2, 1], [3, 1]])
+ 'f(x)=x^2*0.0+x^1*-0.0+x^0*1.0'
+ >>> points_to_polynomial([[1, 3], [2, 3], [3, 3]])
+ 'f(x)=x^2*0.0+x^1*-0.0+x^0*3.0'
+ >>> points_to_polynomial([[1, 1], [2, 2], [3, 3]])
+ 'f(x)=x^2*0.0+x^1*1.0+x^0*0.0'
+ >>> points_to_polynomial([[1, 1], [2, 4], [3, 9]])
+ 'f(x)=x^2*1.0+x^1*-0.0+x^0*0.0'
+ >>> points_to_polynomial([[1, 3], [2, 6], [3, 11]])
+ 'f(x)=x^2*1.0+x^1*-0.0+x^0*2.0'
+ >>> points_to_polynomial([[1, -3], [2, -6], [3, -11]])
+ 'f(x)=x^2*-1.0+x^1*-0.0+x^0*-2.0'
+ >>> points_to_polynomial([[1, 5], [2, 2], [3, 9]])
+ 'f(x)=x^2*5.0+x^1*-18.0+x^0*18.0'
+ >>> points_to_polynomial([[1, 1], [1, 2], [1, 3]])
+ 'x=1'
+ >>> points_to_polynomial([[1, 1], [2, 2], [2, 2]])
+ Traceback (most recent call last):
+ ...
+ ValueError: The program cannot work out a fitting polynomial.
"""
- try:
- check = 1
- more_check = 0
- d = coordinates[0][0]
- for j in range(len(coordinates)):
- if j == 0:
- continue
- if d == coordinates[j][0]:
- more_check += 1
- solved = "x=" + str(coordinates[j][0])
- if more_check == len(coordinates) - 1:
- check = 2
- break
- elif more_check > 0 and more_check != len(coordinates) - 1:
- check = 3
- else:
- check = 1
-
- if len(coordinates) == 1 and coordinates[0][0] == 0:
- check = 2
- solved = "x=0"
- except Exception:
- check = 3
-
- x = len(coordinates)
+ if len(coordinates) == 0 or not all(len(pair) == 2 for pair in coordinates):
+ raise ValueError("The program cannot work out a fitting polynomial.")
- if check == 1:
- count_of_line = 0
- matrix = []
- # put the x and x to the power values in a matrix
- while count_of_line < x:
- count_in_line = 0
- a = coordinates[count_of_line][0]
- count_line: list[int] = []
- while count_in_line < x:
- count_line.append(a ** (x - (count_in_line + 1)))
- count_in_line += 1
- matrix.append(count_line)
- count_of_line += 1
+ if len({tuple(pair) for pair in coordinates}) != len(coordinates):
+ raise ValueError("The program cannot work out a fitting polynomial.")
- count_of_line = 0
- # put the y values into a vector
- vector: list[int] = []
- while count_of_line < x:
- vector.append(coordinates[count_of_line][1])
- count_of_line += 1
+ set_x = {x for x, _ in coordinates}
+ if len(set_x) == 1:
+ return f"x={coordinates[0][0]}"
- count = 0
+ if len(set_x) != len(coordinates):
+ raise ValueError("The program cannot work out a fitting polynomial.")
- while count < x:
- zahlen = 0
- while zahlen < x:
- if count == zahlen:
- zahlen += 1
- if zahlen == x:
- break
- bruch = matrix[zahlen][count] / matrix[count][count]
- for counting_columns, item in enumerate(matrix[count]):
- # manipulating all the values in the matrix
- matrix[zahlen][counting_columns] -= item * bruch
- # manipulating the values in the vector
- vector[zahlen] -= vector[count] * bruch
- zahlen += 1
- count += 1
-
- count = 0
- # make solutions
- solution: list[str] = []
- while count < x:
- solution.append(vector[count] / matrix[count][count])
- count += 1
-
- count = 0
- solved = "f(x)="
-
- while count < x:
- remove_e: list[str] = str(solution[count]).split("E")
- if len(remove_e) > 1:
- solution[count] = remove_e[0] + "*10^" + remove_e[1]
- solved += "x^" + str(x - (count + 1)) + "*" + str(solution[count])
- if count + 1 != x:
- solved += "+"
- count += 1
-
- return solved
+ x = len(coordinates)
- elif check == 2:
- return solved
- else:
- return "The program cannot work out a fitting polynomial."
+ # put the x and x to the power values in a matrix
+ matrix: list[list[float]] = [
+ [
+ coordinates[count_of_line][0] ** (x - (count_in_line + 1))
+ for count_in_line in range(x)
+ ]
+ for count_of_line in range(x)
+ ]
+
+ # put the y values into a vector
+ vector: list[float] = [coordinates[count_of_line][1] for count_of_line in range(x)]
+
+ for count in range(x):
+ for number in range(x):
+ if count == number:
+ continue
+ fraction = matrix[number][count] / matrix[count][count]
+ for counting_columns, item in enumerate(matrix[count]):
+ # manipulating all the values in the matrix
+ matrix[number][counting_columns] -= item * fraction
+ # manipulating the values in the vector
+ vector[number] -= vector[count] * fraction
+
+ # make solutions
+ solution: list[str] = [
+ str(vector[count] / matrix[count][count]) for count in range(x)
+ ]
+
+ solved = "f(x)="
+
+ for count in range(x):
+ remove_e: list[str] = solution[count].split("E")
+ if len(remove_e) > 1:
+ solution[count] = f"{remove_e[0]}*10^{remove_e[1]}"
+ solved += f"x^{x - (count + 1)}*{solution[count]}"
+ if count + 1 != x:
+ solved += "+"
+
+ return solved
if __name__ == "__main__":
diff --git a/linear_algebra/src/power_iteration.py b/linear_algebra/src/power_iteration.py
index 476361e0d433..83c2ce48c3a0 100644
--- a/linear_algebra/src/power_iteration.py
+++ b/linear_algebra/src/power_iteration.py
@@ -2,14 +2,17 @@
def power_iteration(
- input_matrix: np.array, vector: np.array, error_tol=1e-12, max_iterations=100
-) -> [float, np.array]:
+ input_matrix: np.ndarray,
+ vector: np.ndarray,
+ error_tol: float = 1e-12,
+ max_iterations: int = 100,
+) -> tuple[float, np.ndarray]:
"""
Power Iteration.
- Find the largest eignevalue and corresponding eigenvector
+ Find the largest eigenvalue and corresponding eigenvector
of matrix input_matrix given a random vector in the same space.
Will work so long as vector has component of largest eigenvector.
- input_matrix must be symmetric.
+ input_matrix must be either real or Hermitian.
Input
input_matrix: input matrix whose largest eigenvalue we will find.
@@ -38,12 +41,18 @@ def power_iteration(
assert np.shape(input_matrix)[0] == np.shape(input_matrix)[1]
# Ensure proper dimensionality.
assert np.shape(input_matrix)[0] == np.shape(vector)[0]
+ # Ensure inputs are either both complex or both real
+ assert np.iscomplexobj(input_matrix) == np.iscomplexobj(vector)
+ is_complex = np.iscomplexobj(input_matrix)
+ if is_complex:
+ # Ensure complex input_matrix is Hermitian
+ assert np.array_equal(input_matrix, input_matrix.conj().T)
# Set convergence to False. Will define convergence when we exceed max_iterations
# or when we have small changes from one iteration to next.
convergence = False
- lamda_previous = 0
+ lambda_previous = 0
iterations = 0
error = 1e12
@@ -54,44 +63,62 @@ def power_iteration(
vector = w / np.linalg.norm(w)
# Find rayleigh quotient
# (faster than usual b/c we know vector is normalized already)
- lamda = np.dot(vector.T, np.dot(input_matrix, vector))
+ vector_h = vector.conj().T if is_complex else vector.T
+ lambda_ = np.dot(vector_h, np.dot(input_matrix, vector))
# Check convergence.
- error = np.abs(lamda - lamda_previous) / lamda
+ error = np.abs(lambda_ - lambda_previous) / lambda_
iterations += 1
if error <= error_tol or iterations >= max_iterations:
convergence = True
- lamda_previous = lamda
+ lambda_previous = lambda_
- return lamda, vector
+ if is_complex:
+ lambda_ = np.real(lambda_)
+
+ return float(lambda_), vector
def test_power_iteration() -> None:
"""
>>> test_power_iteration() # self running tests
"""
- # Our implementation.
- input_matrix = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]])
- vector = np.array([41, 4, 20])
- eigen_value, eigen_vector = power_iteration(input_matrix, vector)
-
- # Numpy implementation.
-
- # Get eigen values and eigen vectors using built in numpy
- # eigh (eigh used for symmetric or hermetian matrices).
- eigen_values, eigen_vectors = np.linalg.eigh(input_matrix)
- # Last eigen value is the maximum one.
- eigen_value_max = eigen_values[-1]
- # Last column in this matrix is eigen vector corresponding to largest eigen value.
- eigen_vector_max = eigen_vectors[:, -1]
-
- # Check our implementation and numpy gives close answers.
- assert np.abs(eigen_value - eigen_value_max) <= 1e-6
- # Take absolute values element wise of each eigenvector.
- # as they are only unique to a minus sign.
- assert np.linalg.norm(np.abs(eigen_vector) - np.abs(eigen_vector_max)) <= 1e-6
+ real_input_matrix = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]])
+ real_vector = np.array([41, 4, 20])
+ complex_input_matrix = real_input_matrix.astype(np.complex128)
+ imag_matrix = np.triu(1j * complex_input_matrix, 1)
+ complex_input_matrix += imag_matrix
+ complex_input_matrix += -1 * imag_matrix.T
+ complex_vector = np.array([41, 4, 20]).astype(np.complex128)
+
+ for problem_type in ["real", "complex"]:
+ if problem_type == "real":
+ input_matrix = real_input_matrix
+ vector = real_vector
+ elif problem_type == "complex":
+ input_matrix = complex_input_matrix
+ vector = complex_vector
+
+ # Our implementation.
+ eigen_value, eigen_vector = power_iteration(input_matrix, vector)
+
+ # Numpy implementation.
+
+ # Get eigenvalues and eigenvectors using built-in numpy
+ # eigh (eigh used for symmetric or hermetian matrices).
+ eigen_values, eigen_vectors = np.linalg.eigh(input_matrix)
+ # Last eigenvalue is the maximum one.
+ eigen_value_max = eigen_values[-1]
+ # Last column in this matrix is eigenvector corresponding to largest eigenvalue.
+ eigen_vector_max = eigen_vectors[:, -1]
+
+ # Check our implementation and numpy gives close answers.
+ assert np.abs(eigen_value - eigen_value_max) <= 1e-6
+ # Take absolute values element wise of each eigenvector.
+ # as they are only unique to a minus sign.
+ assert np.linalg.norm(np.abs(eigen_vector) - np.abs(eigen_vector_max)) <= 1e-6
if __name__ == "__main__":
diff --git a/linear_algebra/src/rank_of_matrix.py b/linear_algebra/src/rank_of_matrix.py
new file mode 100644
index 000000000000..2c4fe2a8d1da
--- /dev/null
+++ b/linear_algebra/src/rank_of_matrix.py
@@ -0,0 +1,93 @@
+"""
+Calculate the rank of a matrix.
+
+See: https://en.wikipedia.org/wiki/Rank_(linear_algebra)
+"""
+
+
+def rank_of_matrix(matrix: list[list[int | float]]) -> int:
+ """
+ Finds the rank of a matrix.
+
+ Args:
+ `matrix`: The matrix as a list of lists.
+
+ Returns:
+ The rank of the matrix.
+
+ Example:
+
+ >>> matrix1 = [[1, 2, 3],
+ ... [4, 5, 6],
+ ... [7, 8, 9]]
+ >>> rank_of_matrix(matrix1)
+ 2
+ >>> matrix2 = [[1, 0, 0],
+ ... [0, 1, 0],
+ ... [0, 0, 0]]
+ >>> rank_of_matrix(matrix2)
+ 2
+ >>> matrix3 = [[1, 2, 3, 4],
+ ... [5, 6, 7, 8],
+ ... [9, 10, 11, 12]]
+ >>> rank_of_matrix(matrix3)
+ 2
+ >>> rank_of_matrix([[2,3,-1,-1],
+ ... [1,-1,-2,4],
+ ... [3,1,3,-2],
+ ... [6,3,0,-7]])
+ 4
+ >>> rank_of_matrix([[2,1,-3,-6],
+ ... [3,-3,1,2],
+ ... [1,1,1,2]])
+ 3
+ >>> rank_of_matrix([[2,-1,0],
+ ... [1,3,4],
+ ... [4,1,-3]])
+ 3
+ >>> rank_of_matrix([[3,2,1],
+ ... [-6,-4,-2]])
+ 1
+ >>> rank_of_matrix([[],[]])
+ 0
+ >>> rank_of_matrix([[1]])
+ 1
+ >>> rank_of_matrix([[]])
+ 0
+ """
+
+ rows = len(matrix)
+ columns = len(matrix[0])
+ rank = min(rows, columns)
+
+ for row in range(rank):
+ # Check if diagonal element is not zero
+ if matrix[row][row] != 0:
+ # Eliminate all the elements below the diagonal
+ for col in range(row + 1, rows):
+ multiplier = matrix[col][row] / matrix[row][row]
+ for i in range(row, columns):
+ matrix[col][i] -= multiplier * matrix[row][i]
+ else:
+ # Find a non-zero diagonal element to swap rows
+ reduce = True
+ for i in range(row + 1, rows):
+ if matrix[i][row] != 0:
+ matrix[row], matrix[i] = matrix[i], matrix[row]
+ reduce = False
+ break
+ if reduce:
+ rank -= 1
+ for i in range(rows):
+ matrix[i][row] = matrix[i][rank]
+
+ # Reduce the row pointer by one to stay on the same row
+ row -= 1
+
+ return rank
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/linear_algebra/src/rayleigh_quotient.py b/linear_algebra/src/rayleigh_quotient.py
index 69bbbac119e8..46bf1671d2b1 100644
--- a/linear_algebra/src/rayleigh_quotient.py
+++ b/linear_algebra/src/rayleigh_quotient.py
@@ -1,10 +1,13 @@
"""
https://en.wikipedia.org/wiki/Rayleigh_quotient
"""
+
+from typing import Any
+
import numpy as np
-def is_hermitian(matrix: np.array) -> bool:
+def is_hermitian(matrix: np.ndarray) -> bool:
"""
Checks if a matrix is Hermitian.
>>> import numpy as np
@@ -24,7 +27,7 @@ def is_hermitian(matrix: np.array) -> bool:
return np.array_equal(matrix, matrix.conjugate().T)
-def rayleigh_quotient(A: np.array, v: np.array) -> float:
+def rayleigh_quotient(a: np.ndarray, v: np.ndarray) -> Any:
"""
Returns the Rayleigh quotient of a Hermitian matrix A and
vector v.
@@ -43,18 +46,20 @@ def rayleigh_quotient(A: np.array, v: np.array) -> float:
array([[3.]])
"""
v_star = v.conjugate().T
- return (v_star.dot(A).dot(v)) / (v_star.dot(v))
+ v_star_dot = v_star.dot(a)
+ assert isinstance(v_star_dot, np.ndarray)
+ return (v_star_dot.dot(v)) / (v_star.dot(v))
def tests() -> None:
- A = np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]])
+ a = np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]])
v = np.array([[1], [2], [3]])
- assert is_hermitian(A), f"{A} is not hermitian."
- print(rayleigh_quotient(A, v))
+ assert is_hermitian(a), f"{a} is not hermitian."
+ print(rayleigh_quotient(a, v))
- A = np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]])
- assert is_hermitian(A), f"{A} is not hermitian."
- assert rayleigh_quotient(A, v) == float(3)
+ a = np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]])
+ assert is_hermitian(a), f"{a} is not hermitian."
+ assert rayleigh_quotient(a, v) == float(3)
if __name__ == "__main__":
diff --git a/linear_algebra/src/schur_complement.py b/linear_algebra/src/schur_complement.py
new file mode 100644
index 000000000000..74ac75e3fce2
--- /dev/null
+++ b/linear_algebra/src/schur_complement.py
@@ -0,0 +1,98 @@
+import unittest
+
+import numpy as np
+import pytest
+
+
+def schur_complement(
+ mat_a: np.ndarray,
+ mat_b: np.ndarray,
+ mat_c: np.ndarray,
+ pseudo_inv: np.ndarray | None = None,
+) -> np.ndarray:
+ """
+ Schur complement of a symmetric matrix X given as a 2x2 block matrix
+ consisting of matrices `A`, `B` and `C`.
+ Matrix `A` must be quadratic and non-singular.
+ In case `A` is singular, a pseudo-inverse may be provided using
+ the `pseudo_inv` argument.
+
+ | Link to Wiki: https://en.wikipedia.org/wiki/Schur_complement
+ | See also Convex Optimization - Boyd and Vandenberghe, A.5.5
+
+ >>> import numpy as np
+ >>> a = np.array([[1, 2], [2, 1]])
+ >>> b = np.array([[0, 3], [3, 0]])
+ >>> c = np.array([[2, 1], [6, 3]])
+ >>> schur_complement(a, b, c)
+ array([[ 5., -5.],
+ [ 0., 6.]])
+ """
+ shape_a = np.shape(mat_a)
+ shape_b = np.shape(mat_b)
+ shape_c = np.shape(mat_c)
+
+ if shape_a[0] != shape_b[0]:
+ msg = (
+ "Expected the same number of rows for A and B. "
+ f"Instead found A of size {shape_a} and B of size {shape_b}"
+ )
+ raise ValueError(msg)
+
+ if shape_b[1] != shape_c[1]:
+ msg = (
+ "Expected the same number of columns for B and C. "
+ f"Instead found B of size {shape_b} and C of size {shape_c}"
+ )
+ raise ValueError(msg)
+
+ a_inv = pseudo_inv
+ if a_inv is None:
+ try:
+ a_inv = np.linalg.inv(mat_a)
+ except np.linalg.LinAlgError:
+ raise ValueError(
+ "Input matrix A is not invertible. Cannot compute Schur complement."
+ )
+
+ return mat_c - mat_b.T @ a_inv @ mat_b
+
+
+class TestSchurComplement(unittest.TestCase):
+ def test_schur_complement(self) -> None:
+ a = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]])
+ b = np.array([[0, 3], [3, 0], [2, 3]])
+ c = np.array([[2, 1], [6, 3]])
+
+ s = schur_complement(a, b, c)
+
+ input_matrix = np.block([[a, b], [b.T, c]])
+
+ det_x = np.linalg.det(input_matrix)
+ det_a = np.linalg.det(a)
+ det_s = np.linalg.det(s)
+
+ assert np.is_close(det_x, det_a * det_s)
+
+ def test_improper_a_b_dimensions(self) -> None:
+ a = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]])
+ b = np.array([[0, 3], [3, 0], [2, 3]])
+ c = np.array([[2, 1], [6, 3]])
+
+ with pytest.raises(ValueError):
+ schur_complement(a, b, c)
+
+ def test_improper_b_c_dimensions(self) -> None:
+ a = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]])
+ b = np.array([[0, 3], [3, 0], [2, 3]])
+ c = np.array([[2, 1, 3], [6, 3, 5]])
+
+ with pytest.raises(ValueError):
+ schur_complement(a, b, c)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ unittest.main()
diff --git a/linear_algebra/src/test_linear_algebra.py b/linear_algebra/src/test_linear_algebra.py
index 6eba3a1638bd..5209c152013e 100644
--- a/linear_algebra/src/test_linear_algebra.py
+++ b/linear_algebra/src/test_linear_algebra.py
@@ -6,149 +6,205 @@
This file contains the test-suite for the linear algebra library.
"""
+
import unittest
-from .lib import Matrix, Vector, axpy, squareZeroMatrix, unitBasisVector, zeroVector
+import pytest
+
+from .lib import (
+ Matrix,
+ Vector,
+ axpy,
+ square_zero_matrix,
+ unit_basis_vector,
+ zero_vector,
+)
class Test(unittest.TestCase):
- def test_component(self):
+ def test_component(self) -> None:
"""
- test for method component
+ test for method component()
"""
x = Vector([1, 2, 3])
- self.assertEqual(x.component(0), 1)
- self.assertEqual(x.component(2), 3)
+ assert x.component(0) == 1
+ assert x.component(2) == 3
_ = Vector()
- def test_str(self):
+ def test_str(self) -> None:
"""
- test for toString() method
+ test for method toString()
"""
x = Vector([0, 0, 0, 0, 0, 1])
- self.assertEqual(str(x), "(0,0,0,0,0,1)")
+ assert str(x) == "(0,0,0,0,0,1)"
- def test_size(self):
+ def test_size(self) -> None:
"""
- test for size()-method
+ test for method size()
"""
x = Vector([1, 2, 3, 4])
- self.assertEqual(len(x), 4)
+ assert len(x) == 4
- def test_euclidLength(self):
+ def test_euclidean_length(self) -> None:
"""
- test for the eulidean length
+ test for method euclidean_length()
"""
x = Vector([1, 2])
- self.assertAlmostEqual(x.euclidLength(), 2.236, 3)
+ y = Vector([1, 2, 3, 4, 5])
+ z = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0])
+ w = Vector([1, -1, 1, -1, 2, -3, 4, -5])
+ assert x.euclidean_length() == pytest.approx(2.236, abs=1e-3)
+ assert y.euclidean_length() == pytest.approx(7.416, abs=1e-3)
+ assert z.euclidean_length() == 0
+ assert w.euclidean_length() == pytest.approx(7.616, abs=1e-3)
- def test_add(self):
+ def test_add(self) -> None:
"""
test for + operator
"""
x = Vector([1, 2, 3])
y = Vector([1, 1, 1])
- self.assertEqual((x + y).component(0), 2)
- self.assertEqual((x + y).component(1), 3)
- self.assertEqual((x + y).component(2), 4)
+ assert (x + y).component(0) == 2
+ assert (x + y).component(1) == 3
+ assert (x + y).component(2) == 4
- def test_sub(self):
+ def test_sub(self) -> None:
"""
test for - operator
"""
x = Vector([1, 2, 3])
y = Vector([1, 1, 1])
- self.assertEqual((x - y).component(0), 0)
- self.assertEqual((x - y).component(1), 1)
- self.assertEqual((x - y).component(2), 2)
+ assert (x - y).component(0) == 0
+ assert (x - y).component(1) == 1
+ assert (x - y).component(2) == 2
- def test_mul(self):
+ def test_mul(self) -> None:
"""
test for * operator
"""
x = Vector([1, 2, 3])
- a = Vector([2, -1, 4]) # for test of dot-product
+ a = Vector([2, -1, 4]) # for test of dot product
b = Vector([1, -2, -1])
- self.assertEqual(str(x * 3.0), "(3.0,6.0,9.0)")
- self.assertEqual((a * b), 0)
+ assert str(x * 3.0) == "(3.0,6.0,9.0)"
+ assert a * b == 0
- def test_zeroVector(self):
+ def test_zero_vector(self) -> None:
"""
- test for the global function zeroVector(...)
+ test for global function zero_vector()
"""
- self.assertTrue(str(zeroVector(10)).count("0") == 10)
+ assert str(zero_vector(10)).count("0") == 10
- def test_unitBasisVector(self):
+ def test_unit_basis_vector(self) -> None:
"""
- test for the global function unitBasisVector(...)
+ test for global function unit_basis_vector()
"""
- self.assertEqual(str(unitBasisVector(3, 1)), "(0,1,0)")
+ assert str(unit_basis_vector(3, 1)) == "(0,1,0)"
- def test_axpy(self):
+ def test_axpy(self) -> None:
"""
- test for the global function axpy(...) (operation)
+ test for global function axpy() (operation)
"""
x = Vector([1, 2, 3])
y = Vector([1, 0, 1])
- self.assertEqual(str(axpy(2, x, y)), "(3,4,7)")
+ assert str(axpy(2, x, y)) == "(3,4,7)"
- def test_copy(self):
+ def test_copy(self) -> None:
"""
- test for the copy()-method
+ test for method copy()
"""
x = Vector([1, 0, 0, 0, 0, 0])
y = x.copy()
- self.assertEqual(str(x), str(y))
+ assert str(x) == str(y)
- def test_changeComponent(self):
+ def test_change_component(self) -> None:
"""
- test for the changeComponent(...)-method
+ test for method change_component()
"""
x = Vector([1, 0, 0])
- x.changeComponent(0, 0)
- x.changeComponent(1, 1)
- self.assertEqual(str(x), "(0,1,0)")
+ x.change_component(0, 0)
+ x.change_component(1, 1)
+ assert str(x) == "(0,1,0)"
+
+ def test_str_matrix(self) -> None:
+ """
+ test for Matrix method str()
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ assert str(a) == "|1,2,3|\n|2,4,5|\n|6,7,8|\n"
+
+ def test_minor(self) -> None:
+ """
+ test for Matrix method minor()
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ minors = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]]
+ for x in range(a.height()):
+ for y in range(a.width()):
+ assert minors[x][y] == a.minor(x, y)
- def test_str_matrix(self):
- A = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
- self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n", str(A))
+ def test_cofactor(self) -> None:
+ """
+ test for Matrix method cofactor()
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ cofactors = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]]
+ for x in range(a.height()):
+ for y in range(a.width()):
+ assert cofactors[x][y] == a.cofactor(x, y)
- def test_determinate(self):
+ def test_determinant(self) -> None:
"""
- test for determinate()
+ test for Matrix method determinant()
"""
- A = Matrix([[1, 1, 4, 5], [3, 3, 3, 2], [5, 1, 9, 0], [9, 7, 7, 9]], 4, 4)
- self.assertEqual(-376, A.determinate())
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ assert a.determinant() == -5
- def test__mul__matrix(self):
- A = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]], 3, 3)
+ def test__mul__matrix(self) -> None:
+ """
+ test for Matrix * operator
+ """
+ a = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]], 3, 3)
x = Vector([1, 2, 3])
- self.assertEqual("(14,32,50)", str(A * x))
- self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n", str(A * 2))
-
- def test_changeComponent_matrix(self):
- A = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
- A.changeComponent(0, 2, 5)
- self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n", str(A))
-
- def test_component_matrix(self):
- A = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
- self.assertEqual(7, A.component(2, 1), 0.01)
-
- def test__add__matrix(self):
- A = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
- B = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]], 3, 3)
- self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n", str(A + B))
-
- def test__sub__matrix(self):
- A = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
- B = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]], 3, 3)
- self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n", str(A - B))
-
- def test_squareZeroMatrix(self):
- self.assertEqual(
- "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|" + "\n|0,0,0,0,0|\n",
- str(squareZeroMatrix(5)),
+ assert str(a * x) == "(14,32,50)"
+ assert str(a * 2) == "|2,4,6|\n|8,10,12|\n|14,16,18|\n"
+
+ def test_change_component_matrix(self) -> None:
+ """
+ test for Matrix method change_component()
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ a.change_component(0, 2, 5)
+ assert str(a) == "|1,2,5|\n|2,4,5|\n|6,7,8|\n"
+
+ def test_component_matrix(self) -> None:
+ """
+ test for Matrix method component()
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ assert a.component(2, 1) == 7, "0.01"
+
+ def test__add__matrix(self) -> None:
+ """
+ test for Matrix + operator
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ b = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]], 3, 3)
+ assert str(a + b) == "|2,4,10|\n|4,8,10|\n|12,14,18|\n"
+
+ def test__sub__matrix(self) -> None:
+ """
+ test for Matrix - operator
+ """
+ a = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3)
+ b = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]], 3, 3)
+ assert str(a - b) == "|0,0,-4|\n|0,0,0|\n|0,0,-2|\n"
+
+ def test_square_zero_matrix(self) -> None:
+ """
+ test for global function square_zero_matrix()
+ """
+ assert str(square_zero_matrix(5)) == (
+ "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n"
)
diff --git a/linear_algebra/src/transformations_2d.py b/linear_algebra/src/transformations_2d.py
index 6a15189c5676..5dee59024752 100644
--- a/linear_algebra/src/transformations_2d.py
+++ b/linear_algebra/src/transformations_2d.py
@@ -3,15 +3,16 @@
I have added the codes for reflection, projection, scaling and rotation 2D matrices.
+.. code-block:: python
+
scaling(5) = [[5.0, 0.0], [0.0, 5.0]]
- rotation(45) = [[0.5253219888177297, -0.8509035245341184],
- [0.8509035245341184, 0.5253219888177297]]
-projection(45) = [[0.27596319193541496, 0.446998331800279],
- [0.446998331800279, 0.7240368080645851]]
-reflection(45) = [[0.05064397763545947, 0.893996663600558],
- [0.893996663600558, 0.7018070490682369]]
+ rotation(45) = [[0.5253219888177297, -0.8509035245341184],
+ [0.8509035245341184, 0.5253219888177297]]
+ projection(45) = [[0.27596319193541496, 0.446998331800279],
+ [0.446998331800279, 0.7240368080645851]]
+ reflection(45) = [[0.05064397763545947, 0.893996663600558],
+ [0.893996663600558, 0.7018070490682369]]
"""
-from __future__ import annotations
from math import cos, sin
diff --git a/linear_programming/__init__.py b/linear_programming/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/linear_programming/simplex.py b/linear_programming/simplex.py
new file mode 100644
index 000000000000..a8affe1b72d2
--- /dev/null
+++ b/linear_programming/simplex.py
@@ -0,0 +1,339 @@
+"""
+Python implementation of the simplex algorithm for solving linear programs in
+tabular form with
+- `>=`, `<=`, and `=` constraints and
+- each variable `x1, x2, ...>= 0`.
+
+See https://gist.github.com/imengus/f9619a568f7da5bc74eaf20169a24d98 for how to
+convert linear programs to simplex tableaus, and the steps taken in the simplex
+algorithm.
+
+Resources:
+https://en.wikipedia.org/wiki/Simplex_algorithm
+https://tinyurl.com/simplex4beginners
+"""
+
+from typing import Any
+
+import numpy as np
+
+
+class Tableau:
+ """Operate on simplex tableaus
+
+ >>> Tableau(np.array([[-1,-1,0,0,1],[1,3,1,0,4],[3,1,0,1,4]]), 2, 2)
+ Traceback (most recent call last):
+ ...
+ TypeError: Tableau must have type float64
+
+ >>> Tableau(np.array([[-1,-1,0,0,-1],[1,3,1,0,4],[3,1,0,1,4.]]), 2, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: RHS must be > 0
+
+ >>> Tableau(np.array([[-1,-1,0,0,1],[1,3,1,0,4],[3,1,0,1,4.]]), -2, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: number of (artificial) variables must be a natural number
+ """
+
+ # Max iteration number to prevent cycling
+ maxiter = 100
+
+ def __init__(
+ self, tableau: np.ndarray, n_vars: int, n_artificial_vars: int
+ ) -> None:
+ if tableau.dtype != "float64":
+ raise TypeError("Tableau must have type float64")
+
+ # Check if RHS is negative
+ if not (tableau[:, -1] >= 0).all():
+ raise ValueError("RHS must be > 0")
+
+ if n_vars < 2 or n_artificial_vars < 0:
+ raise ValueError(
+ "number of (artificial) variables must be a natural number"
+ )
+
+ self.tableau = tableau
+ self.n_rows, n_cols = tableau.shape
+
+ # Number of decision variables x1, x2, x3...
+ self.n_vars, self.n_artificial_vars = n_vars, n_artificial_vars
+
+ # 2 if there are >= or == constraints (nonstandard), 1 otherwise (std)
+ self.n_stages = (self.n_artificial_vars > 0) + 1
+
+ # Number of slack variables added to make inequalities into equalities
+ self.n_slack = n_cols - self.n_vars - self.n_artificial_vars - 1
+
+ # Objectives for each stage
+ self.objectives = ["max"]
+
+ # In two stage simplex, first minimise then maximise
+ if self.n_artificial_vars:
+ self.objectives.append("min")
+
+ self.col_titles = self.generate_col_titles()
+
+ # Index of current pivot row and column
+ self.row_idx = None
+ self.col_idx = None
+
+ # Does objective row only contain (non)-negative values?
+ self.stop_iter = False
+
+ def generate_col_titles(self) -> list[str]:
+ """Generate column titles for tableau of specific dimensions
+
+ >>> Tableau(np.array([[-1,-1,0,0,1],[1,3,1,0,4],[3,1,0,1,4.]]),
+ ... 2, 0).generate_col_titles()
+ ['x1', 'x2', 's1', 's2', 'RHS']
+
+ >>> Tableau(np.array([[-1,-1,0,0,1],[1,3,1,0,4],[3,1,0,1,4.]]),
+ ... 2, 2).generate_col_titles()
+ ['x1', 'x2', 'RHS']
+ """
+ args = (self.n_vars, self.n_slack)
+
+ # decision | slack
+ string_starts = ["x", "s"]
+ titles = []
+ for i in range(2):
+ for j in range(args[i]):
+ titles.append(string_starts[i] + str(j + 1))
+ titles.append("RHS")
+ return titles
+
+ def find_pivot(self) -> tuple[Any, Any]:
+ """Finds the pivot row and column.
+ >>> tuple(int(x) for x in Tableau(np.array([[-2,1,0,0,0], [3,1,1,0,6],
+ ... [1,2,0,1,7.]]), 2, 0).find_pivot())
+ (1, 0)
+ """
+ objective = self.objectives[-1]
+
+ # Find entries of highest magnitude in objective rows
+ sign = (objective == "min") - (objective == "max")
+ col_idx = np.argmax(sign * self.tableau[0, :-1])
+
+ # Choice is only valid if below 0 for maximise, and above for minimise
+ if sign * self.tableau[0, col_idx] <= 0:
+ self.stop_iter = True
+ return 0, 0
+
+ # Pivot row is chosen as having the lowest quotient when elements of
+ # the pivot column divide the right-hand side
+
+ # Slice excluding the objective rows
+ s = slice(self.n_stages, self.n_rows)
+
+ # RHS
+ dividend = self.tableau[s, -1]
+
+ # Elements of pivot column within slice
+ divisor = self.tableau[s, col_idx]
+
+ # Array filled with nans
+ nans = np.full(self.n_rows - self.n_stages, np.nan)
+
+ # If element in pivot column is greater than zero, return
+ # quotient or nan otherwise
+ quotients = np.divide(dividend, divisor, out=nans, where=divisor > 0)
+
+ # Arg of minimum quotient excluding the nan values. n_stages is added
+ # to compensate for earlier exclusion of objective columns
+ row_idx = np.nanargmin(quotients) + self.n_stages
+ return row_idx, col_idx
+
+ def pivot(self, row_idx: int, col_idx: int) -> np.ndarray:
+ """Pivots on value on the intersection of pivot row and column.
+
+ >>> Tableau(np.array([[-2,-3,0,0,0],[1,3,1,0,4],[3,1,0,1,4.]]),
+ ... 2, 2).pivot(1, 0).tolist()
+ ... # doctest: +NORMALIZE_WHITESPACE
+ [[0.0, 3.0, 2.0, 0.0, 8.0],
+ [1.0, 3.0, 1.0, 0.0, 4.0],
+ [0.0, -8.0, -3.0, 1.0, -8.0]]
+ """
+ # Avoid changes to original tableau
+ piv_row = self.tableau[row_idx].copy()
+
+ piv_val = piv_row[col_idx]
+
+ # Entry becomes 1
+ piv_row *= 1 / piv_val
+
+ # Variable in pivot column becomes basic, ie the only non-zero entry
+ for idx, coeff in enumerate(self.tableau[:, col_idx]):
+ self.tableau[idx] += -coeff * piv_row
+ self.tableau[row_idx] = piv_row
+ return self.tableau
+
+ def change_stage(self) -> np.ndarray:
+ """Exits first phase of the two-stage method by deleting artificial
+ rows and columns, or completes the algorithm if exiting the standard
+ case.
+
+ >>> Tableau(np.array([
+ ... [3, 3, -1, -1, 0, 0, 4],
+ ... [2, 1, 0, 0, 0, 0, 0.],
+ ... [1, 2, -1, 0, 1, 0, 2],
+ ... [2, 1, 0, -1, 0, 1, 2]
+ ... ]), 2, 2).change_stage().tolist()
+ ... # doctest: +NORMALIZE_WHITESPACE
+ [[2.0, 1.0, 0.0, 0.0, 0.0],
+ [1.0, 2.0, -1.0, 0.0, 2.0],
+ [2.0, 1.0, 0.0, -1.0, 2.0]]
+ """
+ # Objective of original objective row remains
+ self.objectives.pop()
+
+ if not self.objectives:
+ return self.tableau
+
+ # Slice containing ids for artificial columns
+ s = slice(-self.n_artificial_vars - 1, -1)
+
+ # Delete the artificial variable columns
+ self.tableau = np.delete(self.tableau, s, axis=1)
+
+ # Delete the objective row of the first stage
+ self.tableau = np.delete(self.tableau, 0, axis=0)
+
+ self.n_stages = 1
+ self.n_rows -= 1
+ self.n_artificial_vars = 0
+ self.stop_iter = False
+ return self.tableau
+
+ def run_simplex(self) -> dict[Any, Any]:
+ """Operate on tableau until objective function cannot be
+ improved further.
+
+ # Standard linear program:
+ Max: x1 + x2
+ ST: x1 + 3x2 <= 4
+ 3x1 + x2 <= 4
+ >>> {key: float(value) for key, value in Tableau(np.array([[-1,-1,0,0,0],
+ ... [1,3,1,0,4],[3,1,0,1,4.]]), 2, 0).run_simplex().items()}
+ {'P': 2.0, 'x1': 1.0, 'x2': 1.0}
+
+ # Standard linear program with 3 variables:
+ Max: 3x1 + x2 + 3x3
+ ST: 2x1 + x2 + x3 ≤ 2
+ x1 + 2x2 + 3x3 ≤ 5
+ 2x1 + 2x2 + x3 ≤ 6
+ >>> {key: float(value) for key, value in Tableau(np.array([
+ ... [-3,-1,-3,0,0,0,0],
+ ... [2,1,1,1,0,0,2],
+ ... [1,2,3,0,1,0,5],
+ ... [2,2,1,0,0,1,6.]
+ ... ]),3,0).run_simplex().items()} # doctest: +ELLIPSIS
+ {'P': 5.4, 'x1': 0.199..., 'x3': 1.6}
+
+
+ # Optimal tableau input:
+ >>> {key: float(value) for key, value in Tableau(np.array([
+ ... [0, 0, 0.25, 0.25, 2],
+ ... [0, 1, 0.375, -0.125, 1],
+ ... [1, 0, -0.125, 0.375, 1]
+ ... ]), 2, 0).run_simplex().items()}
+ {'P': 2.0, 'x1': 1.0, 'x2': 1.0}
+
+ # Non-standard: >= constraints
+ Max: 2x1 + 3x2 + x3
+ ST: x1 + x2 + x3 <= 40
+ 2x1 + x2 - x3 >= 10
+ - x2 + x3 >= 10
+ >>> {key: float(value) for key, value in Tableau(np.array([
+ ... [2, 0, 0, 0, -1, -1, 0, 0, 20],
+ ... [-2, -3, -1, 0, 0, 0, 0, 0, 0],
+ ... [1, 1, 1, 1, 0, 0, 0, 0, 40],
+ ... [2, 1, -1, 0, -1, 0, 1, 0, 10],
+ ... [0, -1, 1, 0, 0, -1, 0, 1, 10.]
+ ... ]), 3, 2).run_simplex().items()}
+ {'P': 70.0, 'x1': 10.0, 'x2': 10.0, 'x3': 20.0}
+
+ # Non standard: minimisation and equalities
+ Min: x1 + x2
+ ST: 2x1 + x2 = 12
+ 6x1 + 5x2 = 40
+ >>> {key: float(value) for key, value in Tableau(np.array([
+ ... [8, 6, 0, 0, 52],
+ ... [1, 1, 0, 0, 0],
+ ... [2, 1, 1, 0, 12],
+ ... [6, 5, 0, 1, 40.],
+ ... ]), 2, 2).run_simplex().items()}
+ {'P': 7.0, 'x1': 5.0, 'x2': 2.0}
+
+
+ # Pivot on slack variables
+ Max: 8x1 + 6x2
+ ST: x1 + 3x2 <= 33
+ 4x1 + 2x2 <= 48
+ 2x1 + 4x2 <= 48
+ x1 + x2 >= 10
+ x1 >= 2
+ >>> {key: float(value) for key, value in Tableau(np.array([
+ ... [2, 1, 0, 0, 0, -1, -1, 0, 0, 12.0],
+ ... [-8, -6, 0, 0, 0, 0, 0, 0, 0, 0.0],
+ ... [1, 3, 1, 0, 0, 0, 0, 0, 0, 33.0],
+ ... [4, 2, 0, 1, 0, 0, 0, 0, 0, 60.0],
+ ... [2, 4, 0, 0, 1, 0, 0, 0, 0, 48.0],
+ ... [1, 1, 0, 0, 0, -1, 0, 1, 0, 10.0],
+ ... [1, 0, 0, 0, 0, 0, -1, 0, 1, 2.0]
+ ... ]), 2, 2).run_simplex().items()} # doctest: +ELLIPSIS
+ {'P': 132.0, 'x1': 12.000... 'x2': 5.999...}
+ """
+ # Stop simplex algorithm from cycling.
+ for _ in range(Tableau.maxiter):
+ # Completion of each stage removes an objective. If both stages
+ # are complete, then no objectives are left
+ if not self.objectives:
+ # Find the values of each variable at optimal solution
+ return self.interpret_tableau()
+
+ row_idx, col_idx = self.find_pivot()
+
+ # If there are no more negative values in objective row
+ if self.stop_iter:
+ # Delete artificial variable columns and rows. Update attributes
+ self.tableau = self.change_stage()
+ else:
+ self.tableau = self.pivot(row_idx, col_idx)
+ return {}
+
+ def interpret_tableau(self) -> dict[str, float]:
+ """Given the final tableau, add the corresponding values of the basic
+ decision variables to the `output_dict`
+ >>> {key: float(value) for key, value in Tableau(np.array([
+ ... [0,0,0.875,0.375,5],
+ ... [0,1,0.375,-0.125,1],
+ ... [1,0,-0.125,0.375,1]
+ ... ]),2, 0).interpret_tableau().items()}
+ {'P': 5.0, 'x1': 1.0, 'x2': 1.0}
+ """
+ # P = RHS of final tableau
+ output_dict = {"P": abs(self.tableau[0, -1])}
+
+ for i in range(self.n_vars):
+ # Gives indices of nonzero entries in the ith column
+ nonzero = np.nonzero(self.tableau[:, i])
+ n_nonzero = len(nonzero[0])
+
+ # First entry in the nonzero indices
+ nonzero_rowidx = nonzero[0][0]
+ nonzero_val = self.tableau[nonzero_rowidx, i]
+
+ # If there is only one nonzero value in column, which is one
+ if n_nonzero == 1 and nonzero_val == 1:
+ rhs_val = self.tableau[nonzero_rowidx, -1]
+ output_dict[self.col_titles[i]] = rhs_val
+ return output_dict
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/machine_learning/apriori_algorithm.py b/machine_learning/apriori_algorithm.py
new file mode 100644
index 000000000000..09a89ac236bd
--- /dev/null
+++ b/machine_learning/apriori_algorithm.py
@@ -0,0 +1,113 @@
+"""
+Apriori Algorithm is a Association rule mining technique, also known as market basket
+analysis, aims to discover interesting relationships or associations among a set of
+items in a transactional or relational database.
+
+For example, Apriori Algorithm states: "If a customer buys item A and item B, then they
+are likely to buy item C." This rule suggests a relationship between items A, B, and C,
+indicating that customers who purchased A and B are more likely to also purchase item C.
+
+WIKI: https://en.wikipedia.org/wiki/Apriori_algorithm
+Examples: https://www.kaggle.com/code/earthian/apriori-association-rules-mining
+"""
+
+from itertools import combinations
+
+
+def load_data() -> list[list[str]]:
+ """
+ Returns a sample transaction dataset.
+
+ >>> load_data()
+ [['milk'], ['milk', 'butter'], ['milk', 'bread'], ['milk', 'bread', 'chips']]
+ """
+ return [["milk"], ["milk", "butter"], ["milk", "bread"], ["milk", "bread", "chips"]]
+
+
+def prune(itemset: list, candidates: list, length: int) -> list:
+ """
+ Prune candidate itemsets that are not frequent.
+ The goal of pruning is to filter out candidate itemsets that are not frequent. This
+ is done by checking if all the (k-1) subsets of a candidate itemset are present in
+ the frequent itemsets of the previous iteration (valid subsequences of the frequent
+ itemsets from the previous iteration).
+
+ Prunes candidate itemsets that are not frequent.
+
+ >>> itemset = ['X', 'Y', 'Z']
+ >>> candidates = [['X', 'Y'], ['X', 'Z'], ['Y', 'Z']]
+ >>> prune(itemset, candidates, 2)
+ [['X', 'Y'], ['X', 'Z'], ['Y', 'Z']]
+
+ >>> itemset = ['1', '2', '3', '4']
+ >>> candidates = ['1', '2', '4']
+ >>> prune(itemset, candidates, 3)
+ []
+ """
+ pruned = []
+ for candidate in candidates:
+ is_subsequence = True
+ for item in candidate:
+ if item not in itemset or itemset.count(item) < length - 1:
+ is_subsequence = False
+ break
+ if is_subsequence:
+ pruned.append(candidate)
+ return pruned
+
+
+def apriori(data: list[list[str]], min_support: int) -> list[tuple[list[str], int]]:
+ """
+ Returns a list of frequent itemsets and their support counts.
+
+ >>> data = [['A', 'B', 'C'], ['A', 'B'], ['A', 'C'], ['A', 'D'], ['B', 'C']]
+ >>> apriori(data, 2)
+ [(['A', 'B'], 1), (['A', 'C'], 2), (['B', 'C'], 2)]
+
+ >>> data = [['1', '2', '3'], ['1', '2'], ['1', '3'], ['1', '4'], ['2', '3']]
+ >>> apriori(data, 3)
+ []
+ """
+ itemset = [list(transaction) for transaction in data]
+ frequent_itemsets = []
+ length = 1
+
+ while itemset:
+ # Count itemset support
+ counts = [0] * len(itemset)
+ for transaction in data:
+ for j, candidate in enumerate(itemset):
+ if all(item in transaction for item in candidate):
+ counts[j] += 1
+
+ # Prune infrequent itemsets
+ itemset = [item for i, item in enumerate(itemset) if counts[i] >= min_support]
+
+ # Append frequent itemsets (as a list to maintain order)
+ for i, item in enumerate(itemset):
+ frequent_itemsets.append((sorted(item), counts[i]))
+
+ length += 1
+ itemset = prune(itemset, list(combinations(itemset, length)), length)
+
+ return frequent_itemsets
+
+
+if __name__ == "__main__":
+ """
+ Apriori algorithm for finding frequent itemsets.
+
+ Args:
+ data: A list of transactions, where each transaction is a list of items.
+ min_support: The minimum support threshold for frequent itemsets.
+
+ Returns:
+ A list of frequent itemsets along with their support counts.
+ """
+ import doctest
+
+ doctest.testmod()
+
+ # user-defined threshold or minimum support level
+ frequent_itemsets = apriori(data=load_data(), min_support=2)
+ print("\n".join(f"{itemset}: {support}" for itemset, support in frequent_itemsets))
diff --git a/machine_learning/astar.py b/machine_learning/astar.py
index ee3fcff0b7bf..a5859e51fe70 100644
--- a/machine_learning/astar.py
+++ b/machine_learning/astar.py
@@ -1,41 +1,39 @@
"""
-The A* algorithm combines features of uniform-cost search and pure
-heuristic search to efficiently compute optimal solutions.
-A* algorithm is a best-first search algorithm in which the cost
-associated with a node is f(n) = g(n) + h(n),
-where g(n) is the cost of the path from the initial state to node n and
-h(n) is the heuristic estimate or the cost or a path
-from node n to a goal.A* algorithm introduces a heuristic into a
-regular graph-searching algorithm,
-essentially planning ahead at each step so a more optimal decision
-is made.A* also known as the algorithm with brains
+The A* algorithm combines features of uniform-cost search and pure heuristic search to
+efficiently compute optimal solutions.
+
+The A* algorithm is a best-first search algorithm in which the cost associated with a
+node is f(n) = g(n) + h(n), where g(n) is the cost of the path from the initial state to
+node n and h(n) is the heuristic estimate or the cost or a path from node n to a goal.
+
+The A* algorithm introduces a heuristic into a regular graph-searching algorithm,
+essentially planning ahead at each step so a more optimal decision is made. For this
+reason, A* is known as an algorithm with brains.
+
+https://en.wikipedia.org/wiki/A*_search_algorithm
"""
+
import numpy as np
class Cell:
"""
- Class cell represents a cell in the world which have the property
- position : The position of the represented by tupleof x and y
- coordinates initially set to (0,0)
- parent : This contains the parent cell object which we visited
- before arrinving this cell
- g,h,f : The parameters for constructing the heuristic function
- which can be any function. for simplicity used line
- distance
+ Class cell represents a cell in the world which have the properties:
+ position: represented by tuple of x and y coordinates initially set to (0,0).
+ parent: Contains the parent cell object visited before we arrived at this cell.
+ g, h, f: Parameters used when calling our heuristic function.
"""
def __init__(self):
self.position = (0, 0)
self.parent = None
-
self.g = 0
self.h = 0
self.f = 0
"""
- overrides equals method because otherwise cell assign will give
- wrong results
+ Overrides equals method because otherwise cell assign will give
+ wrong results.
"""
def __eq__(self, cell):
@@ -48,8 +46,8 @@ def showcell(self):
class Gridworld:
"""
Gridworld class represents the external world here a grid M*M
- matrix
- world_size: create a numpy array with the given world_size default is 5
+ matrix.
+ world_size: create a numpy array with the given world_size default is 5.
"""
def __init__(self, world_size=(5, 5)):
@@ -60,7 +58,7 @@ def __init__(self, world_size=(5, 5)):
def show(self):
print(self.w)
- def get_neigbours(self, cell):
+ def get_neighbours(self, cell):
"""
Return the neighbours of cell
"""
@@ -90,10 +88,10 @@ def get_neigbours(self, cell):
def astar(world, start, goal):
"""
- Implementation of a start algorithm
- world : Object of the world object
- start : Object of the cell as start position
- stop : Object of the cell as goal position
+ Implementation of a start algorithm.
+ world : Object of the world object.
+ start : Object of the cell as start position.
+ stop : Object of the cell as goal position.
>>> p = Gridworld()
>>> start = Cell()
@@ -113,7 +111,7 @@ def astar(world, start, goal):
_closed.append(_open.pop(min_f))
if current == goal:
break
- for n in world.get_neigbours(current):
+ for n in world.get_neighbours(current):
for c in _closed:
if c == n:
continue
@@ -137,14 +135,14 @@ def astar(world, start, goal):
if __name__ == "__main__":
world = Gridworld()
- # stat position and Goal
+ # Start position and goal
start = Cell()
start.position = (0, 0)
goal = Cell()
goal.position = (4, 4)
print(f"path from {start.position} to {goal.position}")
s = astar(world, start, goal)
- # Just for visual reasons
+ # Just for visual reasons.
for i in s:
world.w[i] = 1
print(world.w)
diff --git a/machine_learning/automatic_differentiation.py b/machine_learning/automatic_differentiation.py
new file mode 100644
index 000000000000..5c2708247c21
--- /dev/null
+++ b/machine_learning/automatic_differentiation.py
@@ -0,0 +1,328 @@
+"""
+Demonstration of the Automatic Differentiation (Reverse mode).
+
+Reference: https://en.wikipedia.org/wiki/Automatic_differentiation
+
+Author: Poojan Smart
+Email: smrtpoojan@gmail.com
+"""
+
+from __future__ import annotations
+
+from collections import defaultdict
+from enum import Enum
+from types import TracebackType
+from typing import Any
+
+import numpy as np
+from typing_extensions import Self # noqa: UP035
+
+
+class OpType(Enum):
+ """
+ Class represents list of supported operations on Variable for gradient calculation.
+ """
+
+ ADD = 0
+ SUB = 1
+ MUL = 2
+ DIV = 3
+ MATMUL = 4
+ POWER = 5
+ NOOP = 6
+
+
+class Variable:
+ """
+ Class represents n-dimensional object which is used to wrap numpy array on which
+ operations will be performed and the gradient will be calculated.
+
+ Examples:
+ >>> Variable(5.0)
+ Variable(5.0)
+ >>> Variable([5.0, 2.9])
+ Variable([5. 2.9])
+ >>> Variable([5.0, 2.9]) + Variable([1.0, 5.5])
+ Variable([6. 8.4])
+ >>> Variable([[8.0, 10.0]])
+ Variable([[ 8. 10.]])
+ """
+
+ def __init__(self, value: Any) -> None:
+ self.value = np.array(value)
+
+ # pointers to the operations to which the Variable is input
+ self.param_to: list[Operation] = []
+ # pointer to the operation of which the Variable is output of
+ self.result_of: Operation = Operation(OpType.NOOP)
+
+ def __repr__(self) -> str:
+ return f"Variable({self.value})"
+
+ def to_ndarray(self) -> np.ndarray:
+ return self.value
+
+ def __add__(self, other: Variable) -> Variable:
+ result = Variable(self.value + other.value)
+
+ with GradientTracker() as tracker:
+ # if tracker is enabled, computation graph will be updated
+ if tracker.enabled:
+ tracker.append(OpType.ADD, params=[self, other], output=result)
+ return result
+
+ def __sub__(self, other: Variable) -> Variable:
+ result = Variable(self.value - other.value)
+
+ with GradientTracker() as tracker:
+ # if tracker is enabled, computation graph will be updated
+ if tracker.enabled:
+ tracker.append(OpType.SUB, params=[self, other], output=result)
+ return result
+
+ def __mul__(self, other: Variable) -> Variable:
+ result = Variable(self.value * other.value)
+
+ with GradientTracker() as tracker:
+ # if tracker is enabled, computation graph will be updated
+ if tracker.enabled:
+ tracker.append(OpType.MUL, params=[self, other], output=result)
+ return result
+
+ def __truediv__(self, other: Variable) -> Variable:
+ result = Variable(self.value / other.value)
+
+ with GradientTracker() as tracker:
+ # if tracker is enabled, computation graph will be updated
+ if tracker.enabled:
+ tracker.append(OpType.DIV, params=[self, other], output=result)
+ return result
+
+ def __matmul__(self, other: Variable) -> Variable:
+ result = Variable(self.value @ other.value)
+
+ with GradientTracker() as tracker:
+ # if tracker is enabled, computation graph will be updated
+ if tracker.enabled:
+ tracker.append(OpType.MATMUL, params=[self, other], output=result)
+ return result
+
+ def __pow__(self, power: int) -> Variable:
+ result = Variable(self.value**power)
+
+ with GradientTracker() as tracker:
+ # if tracker is enabled, computation graph will be updated
+ if tracker.enabled:
+ tracker.append(
+ OpType.POWER,
+ params=[self],
+ output=result,
+ other_params={"power": power},
+ )
+ return result
+
+ def add_param_to(self, param_to: Operation) -> None:
+ self.param_to.append(param_to)
+
+ def add_result_of(self, result_of: Operation) -> None:
+ self.result_of = result_of
+
+
+class Operation:
+ """
+ Class represents operation between single or two Variable objects.
+ Operation objects contains type of operation, pointers to input Variable
+ objects and pointer to resulting Variable from the operation.
+ """
+
+ def __init__(
+ self,
+ op_type: OpType,
+ other_params: dict | None = None,
+ ) -> None:
+ self.op_type = op_type
+ self.other_params = {} if other_params is None else other_params
+
+ def add_params(self, params: list[Variable]) -> None:
+ self.params = params
+
+ def add_output(self, output: Variable) -> None:
+ self.output = output
+
+ def __eq__(self, value) -> bool:
+ return self.op_type == value if isinstance(value, OpType) else False
+
+
+class GradientTracker:
+ """
+ Class contains methods to compute partial derivatives of Variable
+ based on the computation graph.
+
+ Examples:
+
+ >>> with GradientTracker() as tracker:
+ ... a = Variable([2.0, 5.0])
+ ... b = Variable([1.0, 2.0])
+ ... m = Variable([1.0, 2.0])
+ ... c = a + b
+ ... d = a * b
+ ... e = c / d
+ >>> tracker.gradient(e, a)
+ array([-0.25, -0.04])
+ >>> tracker.gradient(e, b)
+ array([-1. , -0.25])
+ >>> tracker.gradient(e, m) is None
+ True
+
+ >>> with GradientTracker() as tracker:
+ ... a = Variable([[2.0, 5.0]])
+ ... b = Variable([[1.0], [2.0]])
+ ... c = a @ b
+ >>> tracker.gradient(c, a)
+ array([[1., 2.]])
+ >>> tracker.gradient(c, b)
+ array([[2.],
+ [5.]])
+
+ >>> with GradientTracker() as tracker:
+ ... a = Variable([[2.0, 5.0]])
+ ... b = a ** 3
+ >>> tracker.gradient(b, a)
+ array([[12., 75.]])
+ """
+
+ instance = None
+
+ def __new__(cls) -> Self:
+ """
+ Executes at the creation of class object and returns if
+ object is already created. This class follows singleton
+ design pattern.
+ """
+ if cls.instance is None:
+ cls.instance = super().__new__(cls)
+ return cls.instance
+
+ def __init__(self) -> None:
+ self.enabled = False
+
+ def __enter__(self) -> Self:
+ self.enabled = True
+ return self
+
+ def __exit__(
+ self,
+ exc_type: type[BaseException] | None,
+ exc: BaseException | None,
+ traceback: TracebackType | None,
+ ) -> None:
+ self.enabled = False
+
+ def append(
+ self,
+ op_type: OpType,
+ params: list[Variable],
+ output: Variable,
+ other_params: dict | None = None,
+ ) -> None:
+ """
+ Adds Operation object to the related Variable objects for
+ creating computational graph for calculating gradients.
+
+ Args:
+ op_type: Operation type
+ params: Input parameters to the operation
+ output: Output variable of the operation
+ """
+ operation = Operation(op_type, other_params=other_params)
+ param_nodes = []
+ for param in params:
+ param.add_param_to(operation)
+ param_nodes.append(param)
+ output.add_result_of(operation)
+
+ operation.add_params(param_nodes)
+ operation.add_output(output)
+
+ def gradient(self, target: Variable, source: Variable) -> np.ndarray | None:
+ """
+ Reverse accumulation of partial derivatives to calculate gradients
+ of target variable with respect to source variable.
+
+ Args:
+ target: target variable for which gradients are calculated.
+ source: source variable with respect to which the gradients are
+ calculated.
+
+ Returns:
+ Gradient of the source variable with respect to the target variable
+ """
+
+ # partial derivatives with respect to target
+ partial_deriv = defaultdict(lambda: 0)
+ partial_deriv[target] = np.ones_like(target.to_ndarray())
+
+ # iterating through each operations in the computation graph
+ operation_queue = [target.result_of]
+ while len(operation_queue) > 0:
+ operation = operation_queue.pop()
+ for param in operation.params:
+ # as per the chain rule, multiplying partial derivatives
+ # of variables with respect to the target
+ dparam_doutput = self.derivative(param, operation)
+ dparam_dtarget = dparam_doutput * partial_deriv[operation.output]
+ partial_deriv[param] += dparam_dtarget
+
+ if param.result_of and param.result_of != OpType.NOOP:
+ operation_queue.append(param.result_of)
+
+ return partial_deriv.get(source)
+
+ def derivative(self, param: Variable, operation: Operation) -> np.ndarray:
+ """
+ Compute the derivative of given operation/function
+
+ Args:
+ param: variable to be differentiated
+ operation: function performed on the input variable
+
+ Returns:
+ Derivative of input variable with respect to the output of
+ the operation
+ """
+ params = operation.params
+
+ if operation == OpType.ADD:
+ return np.ones_like(params[0].to_ndarray(), dtype=np.float64)
+ if operation == OpType.SUB:
+ if params[0] == param:
+ return np.ones_like(params[0].to_ndarray(), dtype=np.float64)
+ return -np.ones_like(params[1].to_ndarray(), dtype=np.float64)
+ if operation == OpType.MUL:
+ return (
+ params[1].to_ndarray().T
+ if params[0] == param
+ else params[0].to_ndarray().T
+ )
+ if operation == OpType.DIV:
+ if params[0] == param:
+ return 1 / params[1].to_ndarray()
+ return -params[0].to_ndarray() / (params[1].to_ndarray() ** 2)
+ if operation == OpType.MATMUL:
+ return (
+ params[1].to_ndarray().T
+ if params[0] == param
+ else params[0].to_ndarray().T
+ )
+ if operation == OpType.POWER:
+ power = operation.other_params["power"]
+ return power * (params[0].to_ndarray() ** (power - 1))
+
+ err_msg = f"invalid operation type: {operation.op_type}"
+ raise ValueError(err_msg)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/machine_learning/data_transformations.py b/machine_learning/data_transformations.py
index 9e0d747e93fa..a1c28d514fd5 100644
--- a/machine_learning/data_transformations.py
+++ b/machine_learning/data_transformations.py
@@ -1,5 +1,7 @@
"""
-Normalization Wikipedia: https://en.wikipedia.org/wiki/Normalization
+Normalization.
+
+Wikipedia: https://en.wikipedia.org/wiki/Normalization
Normalization is the process of converting numerical data to a standard range of values.
This range is typically between [0, 1] or [-1, 1]. The equation for normalization is
x_norm = (x - x_min)/(x_max - x_min) where x_norm is the normalized value, x is the
@@ -23,12 +25,14 @@
2. non-gaussian (non-normal) distributions work better with normalization
3. If a column or list of values has extreme values / outliers, use standardization
"""
+
from statistics import mean, stdev
def normalization(data: list, ndigits: int = 3) -> list:
"""
- Returns a normalized list of values
+ Return a normalized list of values.
+
@params: data, a list of values to normalize
@returns: a list of normalized values (rounded to ndigits decimal places)
@examples:
@@ -46,7 +50,8 @@ def normalization(data: list, ndigits: int = 3) -> list:
def standardization(data: list, ndigits: int = 3) -> list:
"""
- Returns a standardized list of values
+ Return a standardized list of values.
+
@params: data, a list of values to standardize
@returns: a list of standardized values (rounded to ndigits decimal places)
@examples:
diff --git a/machine_learning/decision_tree.py b/machine_learning/decision_tree.py
index ace6fb0fa883..72970431c3fc 100644
--- a/machine_learning/decision_tree.py
+++ b/machine_learning/decision_tree.py
@@ -3,10 +3,11 @@
Input data set: The input data set must be 1-dimensional with continuous labels.
Output: The decision tree maps a real number input to a real number output.
"""
+
import numpy as np
-class Decision_Tree:
+class DecisionTree:
def __init__(self, depth=5, min_leaf_size=5):
self.depth = depth
self.decision_boundary = 0
@@ -18,22 +19,22 @@ def __init__(self, depth=5, min_leaf_size=5):
def mean_squared_error(self, labels, prediction):
"""
mean_squared_error:
- @param labels: a one dimensional numpy array
+ @param labels: a one-dimensional numpy array
@param prediction: a floating point value
return value: mean_squared_error calculates the error if prediction is used to
estimate the labels
- >>> tester = Decision_Tree()
+ >>> tester = DecisionTree()
>>> test_labels = np.array([1,2,3,4,5,6,7,8,9,10])
- >>> test_prediction = np.float(6)
- >>> tester.mean_squared_error(test_labels, test_prediction) == (
- ... Test_Decision_Tree.helper_mean_squared_error_test(test_labels,
- ... test_prediction))
+ >>> test_prediction = float(6)
+ >>> bool(tester.mean_squared_error(test_labels, test_prediction) == (
+ ... TestDecisionTree.helper_mean_squared_error_test(test_labels,
+ ... test_prediction)))
True
>>> test_labels = np.array([1,2,3])
- >>> test_prediction = np.float(2)
- >>> tester.mean_squared_error(test_labels, test_prediction) == (
- ... Test_Decision_Tree.helper_mean_squared_error_test(test_labels,
- ... test_prediction))
+ >>> test_prediction = float(2)
+ >>> bool(tester.mean_squared_error(test_labels, test_prediction) == (
+ ... TestDecisionTree.helper_mean_squared_error_test(test_labels,
+ ... test_prediction)))
True
"""
if labels.ndim != 1:
@@ -41,31 +42,52 @@ def mean_squared_error(self, labels, prediction):
return np.mean((labels - prediction) ** 2)
- def train(self, X, y):
+ def train(self, x, y):
"""
train:
- @param X: a one dimensional numpy array
- @param y: a one dimensional numpy array.
+ @param x: a one-dimensional numpy array
+ @param y: a one-dimensional numpy array.
The contents of y are the labels for the corresponding X values
- train does not have a return value
- """
-
- """
- this section is to check that the inputs conform to our dimensionality
+ train() does not have a return value
+
+ Examples:
+ 1. Try to train when x & y are of same length & 1 dimensions (No errors)
+ >>> dt = DecisionTree()
+ >>> dt.train(np.array([10,20,30,40,50]),np.array([0,0,0,1,1]))
+
+ 2. Try to train when x is 2 dimensions
+ >>> dt = DecisionTree()
+ >>> dt.train(np.array([[1,2,3,4,5],[1,2,3,4,5]]),np.array([0,0,0,1,1]))
+ Traceback (most recent call last):
+ ...
+ ValueError: Input data set must be one-dimensional
+
+ 3. Try to train when x and y are not of the same length
+ >>> dt = DecisionTree()
+ >>> dt.train(np.array([1,2,3,4,5]),np.array([[0,0,0,1,1],[0,0,0,1,1]]))
+ Traceback (most recent call last):
+ ...
+ ValueError: x and y have different lengths
+
+ 4. Try to train when x & y are of the same length but different dimensions
+ >>> dt = DecisionTree()
+ >>> dt.train(np.array([1,2,3,4,5]),np.array([[1],[2],[3],[4],[5]]))
+ Traceback (most recent call last):
+ ...
+ ValueError: Data set labels must be one-dimensional
+
+ This section is to check that the inputs conform to our dimensionality
constraints
"""
- if X.ndim != 1:
- print("Error: Input data set must be one dimensional")
- return
- if len(X) != len(y):
- print("Error: X and y have different lengths")
- return
+ if x.ndim != 1:
+ raise ValueError("Input data set must be one-dimensional")
+ if len(x) != len(y):
+ raise ValueError("x and y have different lengths")
if y.ndim != 1:
- print("Error: Data set labels must be one dimensional")
- return
+ raise ValueError("Data set labels must be one-dimensional")
- if len(X) < 2 * self.min_leaf_size:
+ if len(x) < 2 * self.min_leaf_size:
self.prediction = np.mean(y)
return
@@ -74,7 +96,7 @@ def train(self, X, y):
return
best_split = 0
- min_error = self.mean_squared_error(X, np.mean(y)) * 2
+ min_error = self.mean_squared_error(x, np.mean(y)) * 2
"""
loop over all possible splits for the decision tree. find the best split.
@@ -82,34 +104,34 @@ def train(self, X, y):
then the data set is not split and the average for the entire array is used as
the predictor
"""
- for i in range(len(X)):
- if len(X[:i]) < self.min_leaf_size:
+ for i in range(len(x)):
+ if len(x[:i]) < self.min_leaf_size: # noqa: SIM114
continue
- elif len(X[i:]) < self.min_leaf_size:
+ elif len(x[i:]) < self.min_leaf_size:
continue
else:
- error_left = self.mean_squared_error(X[:i], np.mean(y[:i]))
- error_right = self.mean_squared_error(X[i:], np.mean(y[i:]))
+ error_left = self.mean_squared_error(x[:i], np.mean(y[:i]))
+ error_right = self.mean_squared_error(x[i:], np.mean(y[i:]))
error = error_left + error_right
if error < min_error:
best_split = i
min_error = error
if best_split != 0:
- left_X = X[:best_split]
+ left_x = x[:best_split]
left_y = y[:best_split]
- right_X = X[best_split:]
+ right_x = x[best_split:]
right_y = y[best_split:]
- self.decision_boundary = X[best_split]
- self.left = Decision_Tree(
+ self.decision_boundary = x[best_split]
+ self.left = DecisionTree(
depth=self.depth - 1, min_leaf_size=self.min_leaf_size
)
- self.right = Decision_Tree(
+ self.right = DecisionTree(
depth=self.depth - 1, min_leaf_size=self.min_leaf_size
)
- self.left.train(left_X, left_y)
- self.right.train(right_X, right_y)
+ self.left.train(left_x, left_y)
+ self.right.train(right_x, right_y)
else:
self.prediction = np.mean(y)
@@ -134,7 +156,7 @@ def predict(self, x):
return None
-class Test_Decision_Tree:
+class TestDecisionTree:
"""Decision Tres test class"""
@staticmethod
@@ -145,11 +167,11 @@ def helper_mean_squared_error_test(labels, prediction):
@param prediction: a floating point value
return value: helper_mean_squared_error_test calculates the mean squared error
"""
- squared_error_sum = np.float(0)
+ squared_error_sum = float(0)
for label in labels:
squared_error_sum += (label - prediction) ** 2
- return np.float(squared_error_sum / labels.size)
+ return float(squared_error_sum / labels.size)
def main():
@@ -159,13 +181,14 @@ def main():
predict the label of 10 different test values. Then the mean squared error over
this test is displayed.
"""
- X = np.arange(-1.0, 1.0, 0.005)
- y = np.sin(X)
+ x = np.arange(-1.0, 1.0, 0.005)
+ y = np.sin(x)
- tree = Decision_Tree(depth=10, min_leaf_size=10)
- tree.train(X, y)
+ tree = DecisionTree(depth=10, min_leaf_size=10)
+ tree.train(x, y)
- test_cases = (np.random.rand(10) * 2) - 1
+ rng = np.random.default_rng()
+ test_cases = (rng.random(10) * 2) - 1
predictions = np.array([tree.predict(x) for x in test_cases])
avg_error = np.mean((predictions - test_cases) ** 2)
diff --git a/machine_learning/dimensionality_reduction.py b/machine_learning/dimensionality_reduction.py
new file mode 100644
index 000000000000..50d442ecc3de
--- /dev/null
+++ b/machine_learning/dimensionality_reduction.py
@@ -0,0 +1,198 @@
+# Copyright (c) 2023 Diego Gasco (diego.gasco99@gmail.com), Diegomangasco on GitHub
+
+"""
+Requirements:
+ - numpy version 1.21
+ - scipy version 1.3.3
+Notes:
+ - Each column of the features matrix corresponds to a class item
+"""
+
+import logging
+
+import numpy as np
+import pytest
+from scipy.linalg import eigh
+
+logging.basicConfig(level=logging.INFO, format="%(message)s")
+
+
+def column_reshape(input_array: np.ndarray) -> np.ndarray:
+ """Function to reshape a row Numpy array into a column Numpy array
+ >>> input_array = np.array([1, 2, 3])
+ >>> column_reshape(input_array)
+ array([[1],
+ [2],
+ [3]])
+ """
+
+ return input_array.reshape((input_array.size, 1))
+
+
+def covariance_within_classes(
+ features: np.ndarray, labels: np.ndarray, classes: int
+) -> np.ndarray:
+ """Function to compute the covariance matrix inside each class.
+ >>> features = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]])
+ >>> labels = np.array([0, 1, 0])
+ >>> covariance_within_classes(features, labels, 2)
+ array([[0.66666667, 0.66666667, 0.66666667],
+ [0.66666667, 0.66666667, 0.66666667],
+ [0.66666667, 0.66666667, 0.66666667]])
+ """
+
+ covariance_sum = np.nan
+ for i in range(classes):
+ data = features[:, labels == i]
+ data_mean = data.mean(1)
+ # Centralize the data of class i
+ centered_data = data - column_reshape(data_mean)
+ if i > 0:
+ # If covariance_sum is not None
+ covariance_sum += np.dot(centered_data, centered_data.T)
+ else:
+ # If covariance_sum is np.nan (i.e. first loop)
+ covariance_sum = np.dot(centered_data, centered_data.T)
+
+ return covariance_sum / features.shape[1]
+
+
+def covariance_between_classes(
+ features: np.ndarray, labels: np.ndarray, classes: int
+) -> np.ndarray:
+ """Function to compute the covariance matrix between multiple classes
+ >>> features = np.array([[9, 2, 3], [4, 3, 6], [1, 8, 9]])
+ >>> labels = np.array([0, 1, 0])
+ >>> covariance_between_classes(features, labels, 2)
+ array([[ 3.55555556, 1.77777778, -2.66666667],
+ [ 1.77777778, 0.88888889, -1.33333333],
+ [-2.66666667, -1.33333333, 2. ]])
+ """
+
+ general_data_mean = features.mean(1)
+ covariance_sum = np.nan
+ for i in range(classes):
+ data = features[:, labels == i]
+ device_data = data.shape[1]
+ data_mean = data.mean(1)
+ if i > 0:
+ # If covariance_sum is not None
+ covariance_sum += device_data * np.dot(
+ column_reshape(data_mean) - column_reshape(general_data_mean),
+ (column_reshape(data_mean) - column_reshape(general_data_mean)).T,
+ )
+ else:
+ # If covariance_sum is np.nan (i.e. first loop)
+ covariance_sum = device_data * np.dot(
+ column_reshape(data_mean) - column_reshape(general_data_mean),
+ (column_reshape(data_mean) - column_reshape(general_data_mean)).T,
+ )
+
+ return covariance_sum / features.shape[1]
+
+
+def principal_component_analysis(features: np.ndarray, dimensions: int) -> np.ndarray:
+ """
+ Principal Component Analysis.
+
+ For more details, see: https://en.wikipedia.org/wiki/Principal_component_analysis.
+ Parameters:
+ * features: the features extracted from the dataset
+ * dimensions: to filter the projected data for the desired dimension
+
+ >>> test_principal_component_analysis()
+ """
+
+ # Check if the features have been loaded
+ if features.any():
+ data_mean = features.mean(1)
+ # Center the dataset
+ centered_data = features - np.reshape(data_mean, (data_mean.size, 1))
+ covariance_matrix = np.dot(centered_data, centered_data.T) / features.shape[1]
+ _, eigenvectors = np.linalg.eigh(covariance_matrix)
+ # Take all the columns in the reverse order (-1), and then takes only the first
+ filtered_eigenvectors = eigenvectors[:, ::-1][:, 0:dimensions]
+ # Project the database on the new space
+ projected_data = np.dot(filtered_eigenvectors.T, features)
+ logging.info("Principal Component Analysis computed")
+
+ return projected_data
+ else:
+ logging.basicConfig(level=logging.ERROR, format="%(message)s", force=True)
+ logging.error("Dataset empty")
+ raise AssertionError
+
+
+def linear_discriminant_analysis(
+ features: np.ndarray, labels: np.ndarray, classes: int, dimensions: int
+) -> np.ndarray:
+ """
+ Linear Discriminant Analysis.
+
+ For more details, see: https://en.wikipedia.org/wiki/Linear_discriminant_analysis.
+ Parameters:
+ * features: the features extracted from the dataset
+ * labels: the class labels of the features
+ * classes: the number of classes present in the dataset
+ * dimensions: to filter the projected data for the desired dimension
+
+ >>> test_linear_discriminant_analysis()
+ """
+
+ # Check if the dimension desired is less than the number of classes
+ assert classes > dimensions
+
+ # Check if features have been already loaded
+ if features.any:
+ _, eigenvectors = eigh(
+ covariance_between_classes(features, labels, classes),
+ covariance_within_classes(features, labels, classes),
+ )
+ filtered_eigenvectors = eigenvectors[:, ::-1][:, :dimensions]
+ svd_matrix, _, _ = np.linalg.svd(filtered_eigenvectors)
+ filtered_svd_matrix = svd_matrix[:, 0:dimensions]
+ projected_data = np.dot(filtered_svd_matrix.T, features)
+ logging.info("Linear Discriminant Analysis computed")
+
+ return projected_data
+ else:
+ logging.basicConfig(level=logging.ERROR, format="%(message)s", force=True)
+ logging.error("Dataset empty")
+ raise AssertionError
+
+
+def test_linear_discriminant_analysis() -> None:
+ # Create dummy dataset with 2 classes and 3 features
+ features = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]])
+ labels = np.array([0, 0, 0, 1, 1])
+ classes = 2
+ dimensions = 2
+
+ # Assert that the function raises an AssertionError if dimensions > classes
+ with pytest.raises(AssertionError) as error_info: # noqa: PT012
+ projected_data = linear_discriminant_analysis(
+ features, labels, classes, dimensions
+ )
+ if isinstance(projected_data, np.ndarray):
+ raise AssertionError(
+ "Did not raise AssertionError for dimensions > classes"
+ )
+ assert error_info.type is AssertionError
+
+
+def test_principal_component_analysis() -> None:
+ features = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]])
+ dimensions = 2
+ expected_output = np.array([[6.92820323, 8.66025404, 10.39230485], [3.0, 3.0, 3.0]])
+
+ with pytest.raises(AssertionError) as error_info: # noqa: PT012
+ output = principal_component_analysis(features, dimensions)
+ if not np.allclose(expected_output, output):
+ raise AssertionError
+ assert error_info.type is AssertionError
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/machine_learning/forecasting/ex_data.csv b/machine_learning/forecasting/ex_data.csv
index 1c429e649755..e6e73c4a1ca4 100644
--- a/machine_learning/forecasting/ex_data.csv
+++ b/machine_learning/forecasting/ex_data.csv
@@ -1,4 +1,4 @@
-total_user,total_events,days
+total_users,total_events,days
18231,0.0,1
22621,1.0,2
15675,0.0,3
diff --git a/machine_learning/forecasting/run.py b/machine_learning/forecasting/run.py
index 0e11f958825f..9d81b03cd09e 100644
--- a/machine_learning/forecasting/run.py
+++ b/machine_learning/forecasting/run.py
@@ -1,7 +1,7 @@
"""
this is code for forecasting
-but i modified it and used it for safety checker of data
-for ex: you have a online shop and for some reason some data are
+but I modified it and used it for safety checker of data
+for ex: you have an online shop and for some reason some data are
missing (the amount of data that u expected are not supposed to be)
then we can use it
*ps : 1. ofc we can use normal statistic method but in this case
@@ -11,6 +11,8 @@
u can just adjust it for ur own purpose
"""
+from warnings import simplefilter
+
import numpy as np
import pandas as pd
from sklearn.preprocessing import Normalizer
@@ -26,11 +28,10 @@ def linear_regression_prediction(
input : training data (date, total_user, total_event) in list of float
output : list of total user prediction in float
>>> n = linear_regression_prediction([2,3,4,5], [5,3,4,6], [3,1,2,4], [2,1], [2,2])
- >>> abs(n - 5.0) < 1e-6 # Checking precision because of floating point errors
+ >>> bool(abs(n - 5.0) < 1e-6) # Checking precision because of floating point errors
True
"""
- x = [[1, item, train_mtch[i]] for i, item in enumerate(train_dt)]
- x = np.array(x)
+ x = np.array([[1, item, train_mtch[i]] for i, item in enumerate(train_dt)])
y = np.array(train_usr)
beta = np.dot(np.dot(np.linalg.inv(np.dot(x.transpose(), x)), x.transpose()), y)
return abs(beta[0] + test_dt[0] * beta[1] + test_mtch[0] + beta[2])
@@ -46,14 +47,16 @@ def sarimax_predictor(train_user: list, train_match: list, test_match: list) ->
>>> sarimax_predictor([4,2,6,8], [3,1,2,4], [2])
6.6666671111109626
"""
+ # Suppress the User Warning raised by SARIMAX due to insufficient observations
+ simplefilter("ignore", UserWarning)
order = (1, 2, 1)
- seasonal_order = (1, 1, 0, 7)
+ seasonal_order = (1, 1, 1, 7)
model = SARIMAX(
train_user, exog=train_match, order=order, seasonal_order=seasonal_order
)
model_fit = model.fit(disp=False, maxiter=600, method="nm")
result = model_fit.predict(1, len(test_match), exog=[test_match])
- return result[0]
+ return float(result[0])
def support_vector_regressor(x_train: list, x_test: list, train_user: list) -> float:
@@ -72,7 +75,7 @@ def support_vector_regressor(x_train: list, x_test: list, train_user: list) -> f
regressor = SVR(kernel="rbf", C=1, gamma=0.1, epsilon=0.1)
regressor.fit(x_train, train_user)
y_pred = regressor.predict(x_test)
- return y_pred[0]
+ return float(y_pred[0])
def interquartile_range_checker(train_user: list) -> float:
@@ -89,69 +92,71 @@ def interquartile_range_checker(train_user: list) -> float:
q3 = np.percentile(train_user, 75)
iqr = q3 - q1
low_lim = q1 - (iqr * 0.1)
- return low_lim
+ return float(low_lim)
-def data_safety_checker(list_vote: list, actual_result: float) -> None:
+def data_safety_checker(list_vote: list, actual_result: float) -> bool:
"""
Used to review all the votes (list result prediction)
and compare it to the actual result.
input : list of predictions
output : print whether it's safe or not
- >>> data_safety_checker([2,3,4],5.0)
- Today's data is not safe.
+ >>> data_safety_checker([2, 3, 4], 5.0)
+ False
"""
safe = 0
not_safe = 0
+
+ if not isinstance(actual_result, float):
+ raise TypeError("Actual result should be float. Value passed is a list")
+
for i in list_vote:
if i > actual_result:
safe = not_safe + 1
+ elif abs(abs(i) - abs(actual_result)) <= 0.1:
+ safe += 1
else:
- if abs(abs(i) - abs(actual_result)) <= 0.1:
- safe = safe + 1
- else:
- not_safe = not_safe + 1
- print(f"Today's data is {'not ' if safe <= not_safe else ''}safe.")
+ not_safe += 1
+ return safe > not_safe
-# data_input_df = pd.read_csv("ex_data.csv", header=None)
-data_input = [[18231, 0.0, 1], [22621, 1.0, 2], [15675, 0.0, 3], [23583, 1.0, 4]]
-data_input_df = pd.DataFrame(data_input, columns=["total_user", "total_even", "days"])
-
-"""
-data column = total user in a day, how much online event held in one day,
-what day is that(sunday-saturday)
-"""
-
-# start normalization
-normalize_df = Normalizer().fit_transform(data_input_df.values)
-# split data
-total_date = normalize_df[:, 2].tolist()
-total_user = normalize_df[:, 0].tolist()
-total_match = normalize_df[:, 1].tolist()
-
-# for svr (input variable = total date and total match)
-x = normalize_df[:, [1, 2]].tolist()
-x_train = x[: len(x) - 1]
-x_test = x[len(x) - 1 :]
-
-# for linear reression & sarimax
-trn_date = total_date[: len(total_date) - 1]
-trn_user = total_user[: len(total_user) - 1]
-trn_match = total_match[: len(total_match) - 1]
-
-tst_date = total_date[len(total_date) - 1 :]
-tst_user = total_user[len(total_user) - 1 :]
-tst_match = total_match[len(total_match) - 1 :]
-
-
-# voting system with forecasting
-res_vote = []
-res_vote.append(
- linear_regression_prediction(trn_date, trn_user, trn_match, tst_date, tst_match)
-)
-res_vote.append(sarimax_predictor(trn_user, trn_match, tst_match))
-res_vote.append(support_vector_regressor(x_train, x_test, trn_user))
-
-# check the safety of todays'data^^
-data_safety_checker(res_vote, tst_user)
+if __name__ == "__main__":
+ """
+ data column = total user in a day, how much online event held in one day,
+ what day is that(sunday-saturday)
+ """
+ data_input_df = pd.read_csv("ex_data.csv")
+
+ # start normalization
+ normalize_df = Normalizer().fit_transform(data_input_df.values)
+ # split data
+ total_date = normalize_df[:, 2].tolist()
+ total_user = normalize_df[:, 0].tolist()
+ total_match = normalize_df[:, 1].tolist()
+
+ # for svr (input variable = total date and total match)
+ x = normalize_df[:, [1, 2]].tolist()
+ x_train = x[: len(x) - 1]
+ x_test = x[len(x) - 1 :]
+
+ # for linear regression & sarimax
+ train_date = total_date[: len(total_date) - 1]
+ train_user = total_user[: len(total_user) - 1]
+ train_match = total_match[: len(total_match) - 1]
+
+ test_date = total_date[len(total_date) - 1 :]
+ test_user = total_user[len(total_user) - 1 :]
+ test_match = total_match[len(total_match) - 1 :]
+
+ # voting system with forecasting
+ res_vote = [
+ linear_regression_prediction(
+ train_date, train_user, train_match, test_date, test_match
+ ),
+ sarimax_predictor(train_user, train_match, test_match),
+ support_vector_regressor(x_train, x_test, train_user),
+ ]
+
+ # check the safety of today's data
+ not_str = "" if data_safety_checker(res_vote, test_user[0]) else "not "
+ print(f"Today's data is {not_str}safe.")
diff --git a/machine_learning/frequent_pattern_growth.py b/machine_learning/frequent_pattern_growth.py
new file mode 100644
index 000000000000..fae2df16efb1
--- /dev/null
+++ b/machine_learning/frequent_pattern_growth.py
@@ -0,0 +1,350 @@
+"""
+The Frequent Pattern Growth algorithm (FP-Growth) is a widely used data mining
+technique for discovering frequent itemsets in large transaction databases.
+
+It overcomes some of the limitations of traditional methods such as Apriori by
+efficiently constructing the FP-Tree
+
+WIKI: https://athena.ecs.csus.edu/~mei/associationcw/FpGrowth.html
+
+Examples: https://www.javatpoint.com/fp-growth-algorithm-in-data-mining
+"""
+
+from __future__ import annotations
+
+from dataclasses import dataclass, field
+
+
+@dataclass
+class TreeNode:
+ """
+ A node in a Frequent Pattern tree.
+
+ Args:
+ name: The name of this node.
+ num_occur: The number of occurrences of the node.
+ parent_node: The parent node.
+
+ Example:
+ >>> parent = TreeNode("Parent", 1, None)
+ >>> child = TreeNode("Child", 2, parent)
+ >>> child.name
+ 'Child'
+ >>> child.count
+ 2
+ """
+
+ name: str
+ count: int
+ parent: TreeNode | None = None
+ children: dict[str, TreeNode] = field(default_factory=dict)
+ node_link: TreeNode | None = None
+
+ def __repr__(self) -> str:
+ return f"TreeNode({self.name!r}, {self.count!r}, {self.parent!r})"
+
+ def inc(self, num_occur: int) -> None:
+ self.count += num_occur
+
+ def disp(self, ind: int = 1) -> None:
+ print(f"{' ' * ind} {self.name} {self.count}")
+ for child in self.children.values():
+ child.disp(ind + 1)
+
+
+def create_tree(data_set: list, min_sup: int = 1) -> tuple[TreeNode, dict]:
+ """
+ Create Frequent Pattern tree
+
+ Args:
+ data_set: A list of transactions, where each transaction is a list of items.
+ min_sup: The minimum support threshold.
+ Items with support less than this will be pruned. Default is 1.
+
+ Returns:
+ The root of the FP-Tree.
+ header_table: The header table dictionary with item information.
+
+ Example:
+ >>> data_set = [
+ ... ['A', 'B', 'C'],
+ ... ['A', 'C'],
+ ... ['A', 'B', 'E'],
+ ... ['A', 'B', 'C', 'E'],
+ ... ['B', 'E']
+ ... ]
+ >>> min_sup = 2
+ >>> fp_tree, header_table = create_tree(data_set, min_sup)
+ >>> fp_tree
+ TreeNode('Null Set', 1, None)
+ >>> len(header_table)
+ 4
+ >>> header_table["A"]
+ [[4, None], TreeNode('A', 4, TreeNode('Null Set', 1, None))]
+ >>> header_table["E"][1] # doctest: +NORMALIZE_WHITESPACE
+ TreeNode('E', 1, TreeNode('B', 3, TreeNode('A', 4, TreeNode('Null Set', 1, None))))
+ >>> sorted(header_table)
+ ['A', 'B', 'C', 'E']
+ >>> fp_tree.name
+ 'Null Set'
+ >>> sorted(fp_tree.children)
+ ['A', 'B']
+ >>> fp_tree.children['A'].name
+ 'A'
+ >>> sorted(fp_tree.children['A'].children)
+ ['B', 'C']
+ """
+ header_table: dict = {}
+ for trans in data_set:
+ for item in trans:
+ header_table[item] = header_table.get(item, [0, None])
+ header_table[item][0] += 1
+
+ for k in list(header_table):
+ if header_table[k][0] < min_sup:
+ del header_table[k]
+
+ if not (freq_item_set := set(header_table)):
+ return TreeNode("Null Set", 1, None), {}
+
+ for key, value in header_table.items():
+ header_table[key] = [value, None]
+
+ fp_tree = TreeNode("Null Set", 1, None) # Parent is None for the root node
+ for tran_set in data_set:
+ local_d = {
+ item: header_table[item][0] for item in tran_set if item in freq_item_set
+ }
+ if local_d:
+ sorted_items = sorted(
+ local_d.items(), key=lambda item_info: item_info[1], reverse=True
+ )
+ ordered_items = [item[0] for item in sorted_items]
+ update_tree(ordered_items, fp_tree, header_table, 1)
+
+ return fp_tree, header_table
+
+
+def update_tree(items: list, in_tree: TreeNode, header_table: dict, count: int) -> None:
+ """
+ Update the FP-Tree with a transaction.
+
+ Args:
+ items: List of items in the transaction.
+ in_tree: The current node in the FP-Tree.
+ header_table: The header table dictionary with item information.
+ count: The count of the transaction.
+
+ Example:
+ >>> data_set = [
+ ... ['A', 'B', 'C'],
+ ... ['A', 'C'],
+ ... ['A', 'B', 'E'],
+ ... ['A', 'B', 'C', 'E'],
+ ... ['B', 'E']
+ ... ]
+ >>> min_sup = 2
+ >>> fp_tree, header_table = create_tree(data_set, min_sup)
+ >>> fp_tree
+ TreeNode('Null Set', 1, None)
+ >>> transaction = ['A', 'B', 'E']
+ >>> update_tree(transaction, fp_tree, header_table, 1)
+ >>> fp_tree
+ TreeNode('Null Set', 1, None)
+ >>> fp_tree.children['A'].children['B'].children['E'].children
+ {}
+ >>> fp_tree.children['A'].children['B'].children['E'].count
+ 2
+ >>> header_table['E'][1].name
+ 'E'
+ """
+ if items[0] in in_tree.children:
+ in_tree.children[items[0]].inc(count)
+ else:
+ in_tree.children[items[0]] = TreeNode(items[0], count, in_tree)
+ if header_table[items[0]][1] is None:
+ header_table[items[0]][1] = in_tree.children[items[0]]
+ else:
+ update_header(header_table[items[0]][1], in_tree.children[items[0]])
+ if len(items) > 1:
+ update_tree(items[1:], in_tree.children[items[0]], header_table, count)
+
+
+def update_header(node_to_test: TreeNode, target_node: TreeNode) -> TreeNode:
+ """
+ Update the header table with a node link.
+
+ Args:
+ node_to_test: The node to be updated in the header table.
+ target_node: The node to link to.
+
+ Example:
+ >>> data_set = [
+ ... ['A', 'B', 'C'],
+ ... ['A', 'C'],
+ ... ['A', 'B', 'E'],
+ ... ['A', 'B', 'C', 'E'],
+ ... ['B', 'E']
+ ... ]
+ >>> min_sup = 2
+ >>> fp_tree, header_table = create_tree(data_set, min_sup)
+ >>> fp_tree
+ TreeNode('Null Set', 1, None)
+ >>> node1 = TreeNode("A", 3, None)
+ >>> node2 = TreeNode("B", 4, None)
+ >>> node1
+ TreeNode('A', 3, None)
+ >>> node1 = update_header(node1, node2)
+ >>> node1
+ TreeNode('A', 3, None)
+ >>> node1.node_link
+ TreeNode('B', 4, None)
+ >>> node2.node_link is None
+ True
+ """
+ while node_to_test.node_link is not None:
+ node_to_test = node_to_test.node_link
+ if node_to_test.node_link is None:
+ node_to_test.node_link = target_node
+ # Return the updated node
+ return node_to_test
+
+
+def ascend_tree(leaf_node: TreeNode, prefix_path: list[str]) -> None:
+ """
+ Ascend the FP-Tree from a leaf node to its root, adding item names to the prefix
+ path.
+
+ Args:
+ leaf_node: The leaf node to start ascending from.
+ prefix_path: A list to store the item as they are ascended.
+
+ Example:
+ >>> data_set = [
+ ... ['A', 'B', 'C'],
+ ... ['A', 'C'],
+ ... ['A', 'B', 'E'],
+ ... ['A', 'B', 'C', 'E'],
+ ... ['B', 'E']
+ ... ]
+ >>> min_sup = 2
+ >>> fp_tree, header_table = create_tree(data_set, min_sup)
+
+ >>> path = []
+ >>> ascend_tree(fp_tree.children['A'], path)
+ >>> path # ascending from a leaf node 'A'
+ ['A']
+ """
+ if leaf_node.parent is not None:
+ prefix_path.append(leaf_node.name)
+ ascend_tree(leaf_node.parent, prefix_path)
+
+
+def find_prefix_path(base_pat: frozenset, tree_node: TreeNode | None) -> dict: # noqa: ARG001
+ """
+ Find the conditional pattern base for a given base pattern.
+
+ Args:
+ base_pat: The base pattern for which to find the conditional pattern base.
+ tree_node: The node in the FP-Tree.
+
+ Example:
+ >>> data_set = [
+ ... ['A', 'B', 'C'],
+ ... ['A', 'C'],
+ ... ['A', 'B', 'E'],
+ ... ['A', 'B', 'C', 'E'],
+ ... ['B', 'E']
+ ... ]
+ >>> min_sup = 2
+ >>> fp_tree, header_table = create_tree(data_set, min_sup)
+ >>> fp_tree
+ TreeNode('Null Set', 1, None)
+ >>> len(header_table)
+ 4
+ >>> base_pattern = frozenset(['A'])
+ >>> sorted(find_prefix_path(base_pattern, fp_tree.children['A']))
+ []
+ """
+ cond_pats: dict = {}
+ while tree_node is not None:
+ prefix_path: list = []
+ ascend_tree(tree_node, prefix_path)
+ if len(prefix_path) > 1:
+ cond_pats[frozenset(prefix_path[1:])] = tree_node.count
+ tree_node = tree_node.node_link
+ return cond_pats
+
+
+def mine_tree(
+ in_tree: TreeNode, # noqa: ARG001
+ header_table: dict,
+ min_sup: int,
+ pre_fix: set,
+ freq_item_list: list,
+) -> None:
+ """
+ Mine the FP-Tree recursively to discover frequent itemsets.
+
+ Args:
+ in_tree: The FP-Tree to mine.
+ header_table: The header table dictionary with item information.
+ min_sup: The minimum support threshold.
+ pre_fix: A set of items as a prefix for the itemsets being mined.
+ freq_item_list: A list to store the frequent itemsets.
+
+ Example:
+ >>> data_set = [
+ ... ['A', 'B', 'C'],
+ ... ['A', 'C'],
+ ... ['A', 'B', 'E'],
+ ... ['A', 'B', 'C', 'E'],
+ ... ['B', 'E']
+ ... ]
+ >>> min_sup = 2
+ >>> fp_tree, header_table = create_tree(data_set, min_sup)
+ >>> fp_tree
+ TreeNode('Null Set', 1, None)
+ >>> frequent_itemsets = []
+ >>> mine_tree(fp_tree, header_table, min_sup, set([]), frequent_itemsets)
+ >>> expe_itm = [{'C'}, {'C', 'A'}, {'E'}, {'A', 'E'}, {'E', 'B'}, {'A'}, {'B'}]
+ >>> all(expected in frequent_itemsets for expected in expe_itm)
+ True
+ """
+ sorted_items = sorted(header_table.items(), key=lambda item_info: item_info[1][0])
+ big_l = [item[0] for item in sorted_items]
+ for base_pat in big_l:
+ new_freq_set = pre_fix.copy()
+ new_freq_set.add(base_pat)
+ freq_item_list.append(new_freq_set)
+ cond_patt_bases = find_prefix_path(base_pat, header_table[base_pat][1])
+ my_cond_tree, my_head = create_tree(list(cond_patt_bases), min_sup)
+ if my_head is not None:
+ # Pass header_table[base_pat][1] as node_to_test to update_header
+ header_table[base_pat][1] = update_header(
+ header_table[base_pat][1], my_cond_tree
+ )
+ mine_tree(my_cond_tree, my_head, min_sup, new_freq_set, freq_item_list)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ data_set: list[frozenset] = [
+ frozenset(["bread", "milk", "cheese"]),
+ frozenset(["bread", "milk"]),
+ frozenset(["bread", "diapers"]),
+ frozenset(["bread", "milk", "diapers"]),
+ frozenset(["milk", "diapers"]),
+ frozenset(["milk", "cheese"]),
+ frozenset(["diapers", "cheese"]),
+ frozenset(["bread", "milk", "cheese", "diapers"]),
+ ]
+ print(f"{len(data_set) = }")
+ fp_tree, header_table = create_tree(data_set, min_sup=3)
+ print(f"{fp_tree = }")
+ print(f"{len(header_table) = }")
+ freq_items: list = []
+ mine_tree(fp_tree, header_table, 3, set(), freq_items)
+ print(f"{freq_items = }")
diff --git a/machine_learning/gaussian_naive_bayes.py b/machine_learning/gaussian_naive_bayes.py.broken.txt
similarity index 57%
rename from machine_learning/gaussian_naive_bayes.py
rename to machine_learning/gaussian_naive_bayes.py.broken.txt
index c200aa5a4d2d..7e9a8d7f6dcf 100644
--- a/machine_learning/gaussian_naive_bayes.py
+++ b/machine_learning/gaussian_naive_bayes.py.broken.txt
@@ -1,7 +1,9 @@
# Gaussian Naive Bayes Example
+import time
+
from matplotlib import pyplot as plt
from sklearn.datasets import load_iris
-from sklearn.metrics import plot_confusion_matrix
+from sklearn.metrics import accuracy_score, plot_confusion_matrix
from sklearn.model_selection import train_test_split
from sklearn.naive_bayes import GaussianNB
@@ -17,28 +19,34 @@ def main():
iris = load_iris()
# Split dataset into train and test data
- X = iris["data"] # features
- Y = iris["target"]
+ x = iris["data"] # features
+ y = iris["target"]
x_train, x_test, y_train, y_test = train_test_split(
- X, Y, test_size=0.3, random_state=1
+ x, y, test_size=0.3, random_state=1
)
# Gaussian Naive Bayes
- NB_model = GaussianNB()
- NB_model.fit(x_train, y_train)
+ nb_model = GaussianNB()
+ time.sleep(2.9)
+ model_fit = nb_model.fit(x_train, y_train)
+ y_pred = model_fit.predict(x_test) # Predictions on the test set
# Display Confusion Matrix
plot_confusion_matrix(
- NB_model,
+ nb_model,
x_test,
y_test,
display_labels=iris["target_names"],
- cmap="Blues",
+ cmap="Blues", # although, Greys_r has a better contrast...
normalize="true",
)
plt.title("Normalized Confusion Matrix - IRIS Dataset")
plt.show()
+ time.sleep(1.8)
+ final_accuracy = 100 * accuracy_score(y_true=y_test, y_pred=y_pred)
+ print(f"The overall accuracy of the model is: {round(final_accuracy, 2)}%")
+
if __name__ == "__main__":
main()
diff --git a/machine_learning/gradient_boosting_classifier.py b/machine_learning/gradient_boosting_classifier.py
new file mode 100644
index 000000000000..2902394d8226
--- /dev/null
+++ b/machine_learning/gradient_boosting_classifier.py
@@ -0,0 +1,118 @@
+import numpy as np
+from sklearn.datasets import load_iris
+from sklearn.metrics import accuracy_score
+from sklearn.model_selection import train_test_split
+from sklearn.tree import DecisionTreeRegressor
+
+
+class GradientBoostingClassifier:
+ def __init__(self, n_estimators: int = 100, learning_rate: float = 0.1) -> None:
+ """
+ Initialize a GradientBoostingClassifier.
+
+ Parameters:
+ - n_estimators (int): The number of weak learners to train.
+ - learning_rate (float): The learning rate for updating the model.
+
+ Attributes:
+ - n_estimators (int): The number of weak learners.
+ - learning_rate (float): The learning rate.
+ - models (list): A list to store the trained weak learners.
+ """
+ self.n_estimators = n_estimators
+ self.learning_rate = learning_rate
+ self.models: list[tuple[DecisionTreeRegressor, float]] = []
+
+ def fit(self, features: np.ndarray, target: np.ndarray) -> None:
+ """
+ Fit the GradientBoostingClassifier to the training data.
+
+ Parameters:
+ - features (np.ndarray): The training features.
+ - target (np.ndarray): The target values.
+
+ Returns:
+ None
+
+ >>> import numpy as np
+ >>> from sklearn.datasets import load_iris
+ >>> clf = GradientBoostingClassifier(n_estimators=100, learning_rate=0.1)
+ >>> iris = load_iris()
+ >>> X, y = iris.data, iris.target
+ >>> clf.fit(X, y)
+ >>> # Check if the model is trained
+ >>> len(clf.models) == 100
+ True
+ """
+ for _ in range(self.n_estimators):
+ # Calculate the pseudo-residuals
+ residuals = -self.gradient(target, self.predict(features))
+ # Fit a weak learner (e.g., decision tree) to the residuals
+ model = DecisionTreeRegressor(max_depth=1)
+ model.fit(features, residuals)
+ # Update the model by adding the weak learner with a learning rate
+ self.models.append((model, self.learning_rate))
+
+ def predict(self, features: np.ndarray) -> np.ndarray:
+ """
+ Make predictions on input data.
+
+ Parameters:
+ - features (np.ndarray): The input data for making predictions.
+
+ Returns:
+ - np.ndarray: An array of binary predictions (-1 or 1).
+
+ >>> import numpy as np
+ >>> from sklearn.datasets import load_iris
+ >>> clf = GradientBoostingClassifier(n_estimators=100, learning_rate=0.1)
+ >>> iris = load_iris()
+ >>> X, y = iris.data, iris.target
+ >>> clf.fit(X, y)
+ >>> y_pred = clf.predict(X)
+ >>> # Check if the predictions have the correct shape
+ >>> y_pred.shape == y.shape
+ True
+ """
+ # Initialize predictions with zeros
+ predictions = np.zeros(features.shape[0])
+ for model, learning_rate in self.models:
+ predictions += learning_rate * model.predict(features)
+ return np.sign(predictions) # Convert to binary predictions (-1 or 1)
+
+ def gradient(self, target: np.ndarray, y_pred: np.ndarray) -> np.ndarray:
+ """
+ Calculate the negative gradient (pseudo-residuals) for logistic loss.
+
+ Parameters:
+ - target (np.ndarray): The target values.
+ - y_pred (np.ndarray): The predicted values.
+
+ Returns:
+ - np.ndarray: An array of pseudo-residuals.
+
+ >>> import numpy as np
+ >>> clf = GradientBoostingClassifier(n_estimators=100, learning_rate=0.1)
+ >>> target = np.array([0, 1, 0, 1])
+ >>> y_pred = np.array([0.2, 0.8, 0.3, 0.7])
+ >>> residuals = clf.gradient(target, y_pred)
+ >>> # Check if residuals have the correct shape
+ >>> residuals.shape == target.shape
+ True
+ """
+ return -target / (1 + np.exp(target * y_pred))
+
+
+if __name__ == "__main__":
+ iris = load_iris()
+ X, y = iris.data, iris.target
+ X_train, X_test, y_train, y_test = train_test_split(
+ X, y, test_size=0.2, random_state=42
+ )
+
+ clf = GradientBoostingClassifier(n_estimators=100, learning_rate=0.1)
+ clf.fit(X_train, y_train)
+
+ y_pred = clf.predict(X_test)
+ accuracy = accuracy_score(y_test, y_pred)
+ print(f"Accuracy: {accuracy:.2f}")
diff --git a/machine_learning/gradient_boosting_regressor.py b/machine_learning/gradient_boosting_regressor.py.broken.txt
similarity index 80%
rename from machine_learning/gradient_boosting_regressor.py
rename to machine_learning/gradient_boosting_regressor.py.broken.txt
index 0aa0e7a10ac5..c082f3cafe10 100644
--- a/machine_learning/gradient_boosting_regressor.py
+++ b/machine_learning/gradient_boosting_regressor.py.broken.txt
@@ -26,30 +26,30 @@ def main():
print(df_boston.describe().T)
# Feature selection
- X = df_boston.iloc[:, :-1]
+ x = df_boston.iloc[:, :-1]
y = df_boston.iloc[:, -1] # target variable
# split the data with 75% train and 25% test sets.
- X_train, X_test, y_train, y_test = train_test_split(
- X, y, random_state=0, test_size=0.25
+ x_train, x_test, y_train, y_test = train_test_split(
+ x, y, random_state=0, test_size=0.25
)
model = GradientBoostingRegressor(
n_estimators=500, max_depth=5, min_samples_split=4, learning_rate=0.01
)
# training the model
- model.fit(X_train, y_train)
+ model.fit(x_train, y_train)
# to see how good the model fit the data
- training_score = model.score(X_train, y_train).round(3)
- test_score = model.score(X_test, y_test).round(3)
+ training_score = model.score(x_train, y_train).round(3)
+ test_score = model.score(x_test, y_test).round(3)
print("Training score of GradientBoosting is :", training_score)
print("The test score of GradientBoosting is :", test_score)
# Let us evaluation the model by finding the errors
- y_pred = model.predict(X_test)
+ y_pred = model.predict(x_test)
# The mean squared error
- print("Mean squared error: %.2f" % mean_squared_error(y_test, y_pred))
+ print(f"Mean squared error: {mean_squared_error(y_test, y_pred):.2f}")
# Explained variance score: 1 is perfect prediction
- print("Test Variance score: %.2f" % r2_score(y_test, y_pred))
+ print(f"Test Variance score: {r2_score(y_test, y_pred):.2f}")
# So let's run the model against the test data
fig, ax = plt.subplots()
diff --git a/machine_learning/gradient_descent.py b/machine_learning/gradient_descent.py
index 9fa460a07562..95463faf5635 100644
--- a/machine_learning/gradient_descent.py
+++ b/machine_learning/gradient_descent.py
@@ -2,7 +2,8 @@
Implementation of gradient descent algorithm for minimizing cost of a linear hypothesis
function.
"""
-import numpy
+
+import numpy as np
# List of input, output pairs
train_data = (
@@ -55,6 +56,7 @@ def output(example_no, data_set):
return train_data[example_no][1]
elif data_set == "test":
return test_data[example_no][1]
+ return None
def calculate_hypothesis_value(example_no, data_set):
@@ -68,6 +70,7 @@ def calculate_hypothesis_value(example_no, data_set):
return _hypothesis_value(train_data[example_no][0])
elif data_set == "test":
return _hypothesis_value(test_data[example_no][0])
+ return None
def summation_of_cost_derivative(index, end=m):
@@ -108,12 +111,12 @@ def run_gradient_descent():
while True:
j += 1
temp_parameter_vector = [0, 0, 0, 0]
- for i in range(0, len(parameter_vector)):
+ for i in range(len(parameter_vector)):
cost_derivative = get_cost_derivative(i - 1)
temp_parameter_vector[i] = (
parameter_vector[i] - LEARNING_RATE * cost_derivative
)
- if numpy.allclose(
+ if np.allclose(
parameter_vector,
temp_parameter_vector,
atol=absolute_error_limit,
diff --git a/machine_learning/k_means_clust.py b/machine_learning/k_means_clust.py
index f155d4845f41..a926362fc18b 100644
--- a/machine_learning/k_means_clust.py
+++ b/machine_learning/k_means_clust.py
@@ -11,10 +11,10 @@
- initial_centroids , initial centroid values generated by utility function(mentioned
in usage).
- maxiter , maximum number of iterations to process.
- - heterogeneity , empty list that will be filled with hetrogeneity values if passed
+ - heterogeneity , empty list that will be filled with heterogeneity values if passed
to kmeans func.
Usage:
- 1. define 'k' value, 'X' features array and 'hetrogeneity' empty list
+ 1. define 'k' value, 'X' features array and 'heterogeneity' empty list
2. create initial_centroids,
initial_centroids = get_initial_centroids(
X,
@@ -31,8 +31,8 @@
record_heterogeneity=heterogeneity,
verbose=True # whether to print logs in console or not.(default=False)
)
- 4. Plot the loss function, hetrogeneity values for every iteration saved in
- hetrogeneity list.
+ 4. Plot the loss function and heterogeneity values for every iteration saved in
+ heterogeneity list.
plot_heterogeneity(
heterogeneity,
k
@@ -40,6 +40,7 @@
5. Transfers Dataframe into excel format it must have feature called
'Clust' with k means clustering numbers in it.
"""
+
import warnings
import numpy as np
@@ -54,12 +55,12 @@
def get_initial_centroids(data, k, seed=None):
"""Randomly choose k data points as initial centroids"""
- if seed is not None: # useful for obtaining consistent results
- np.random.seed(seed)
+ # useful for obtaining consistent results
+ rng = np.random.default_rng(seed)
n = data.shape[0] # number of data points
# Pick K indices from range [0, N).
- rand_indices = np.random.randint(0, n, k)
+ rand_indices = rng.integers(0, n, k)
# Keep centroids as dense format, as many entries will be nonzero due to averaging.
# As long as at least one document in a cluster contains a word,
@@ -69,12 +70,11 @@ def get_initial_centroids(data, k, seed=None):
return centroids
-def centroid_pairwise_dist(X, centroids):
- return pairwise_distances(X, centroids, metric="euclidean")
+def centroid_pairwise_dist(x, centroids):
+ return pairwise_distances(x, centroids, metric="euclidean")
def assign_clusters(data, centroids):
-
# Compute distances between each data point and the set of centroids:
# Fill in the blank (RHS only)
distances_from_centroids = centroid_pairwise_dist(data, centroids)
@@ -100,10 +100,8 @@ def revise_centroids(data, k, cluster_assignment):
def compute_heterogeneity(data, k, centroids, cluster_assignment):
-
heterogeneity = 0.0
for i in range(k):
-
# Select all data points that belong to cluster i. Fill in the blank (RHS only)
member_data_points = data[cluster_assignment == i, :]
@@ -112,7 +110,7 @@ def compute_heterogeneity(data, k, centroids, cluster_assignment):
distances = pairwise_distances(
member_data_points, [centroids[i]], metric="euclidean"
)
- squared_distances = distances ** 2
+ squared_distances = distances**2
heterogeneity += np.sum(squared_distances)
return heterogeneity
@@ -131,7 +129,7 @@ def plot_heterogeneity(heterogeneity, k):
def kmeans(
data, k, initial_centroids, maxiter=500, record_heterogeneity=None, verbose=False
):
- """This function runs k-means on given data and initial set of centroids.
+ """Runs k-means on given data and initial set of centroids.
maxiter: maximum number of iterations to run.(default=500)
record_heterogeneity: (optional) a list, to store the history of heterogeneity
as function of iterations
@@ -164,9 +162,7 @@ def kmeans(
num_changed = np.sum(prev_cluster_assignment != cluster_assignment)
if verbose:
print(
- " {:5d} elements changed their cluster assignment.".format(
- num_changed
- )
+ f" {num_changed:5d} elements changed their cluster assignment."
)
# Record heterogeneity convergence metric
@@ -199,24 +195,21 @@ def kmeans(
plot_heterogeneity(heterogeneity, k)
-def ReportGenerator(
- df: pd.DataFrame, ClusteringVariables: np.array, FillMissingReport=None
+def report_generator(
+ predicted: pd.DataFrame, clustering_variables: np.ndarray, fill_missing_report=None
) -> pd.DataFrame:
"""
- Function generates easy-erading clustering report. It takes 2 arguments as an input:
- DataFrame - dataframe with predicted cluester column;
- FillMissingReport - dictionary of rules how we are going to fill missing
- values of for final report generate (not included in modeling);
- in order to run the function following libraries must be imported:
- import pandas as pd
- import numpy as np
- >>> data = pd.DataFrame()
- >>> data['numbers'] = [1, 2, 3]
- >>> data['col1'] = [0.5, 2.5, 4.5]
- >>> data['col2'] = [100, 200, 300]
- >>> data['col3'] = [10, 20, 30]
- >>> data['Cluster'] = [1, 1, 2]
- >>> ReportGenerator(data, ['col1', 'col2'], 0)
+ Generate a clustering report given these two arguments:
+ predicted - dataframe with predicted cluster column
+ fill_missing_report - dictionary of rules on how we are going to fill in missing
+ values for final generated report (not included in modelling);
+ >>> predicted = pd.DataFrame()
+ >>> predicted['numbers'] = [1, 2, 3]
+ >>> predicted['col1'] = [0.5, 2.5, 4.5]
+ >>> predicted['col2'] = [100, 200, 300]
+ >>> predicted['col3'] = [10, 20, 30]
+ >>> predicted['Cluster'] = [1, 1, 2]
+ >>> report_generator(predicted, ['col1', 'col2'], 0)
Features Type Mark 1 2
0 # of Customers ClusterSize False 2.000000 1.000000
1 % of Customers ClusterProportion False 0.666667 0.333333
@@ -233,19 +226,19 @@ def ReportGenerator(
[104 rows x 5 columns]
"""
# Fill missing values with given rules
- if FillMissingReport:
- df.fillna(value=FillMissingReport, inplace=True)
- df["dummy"] = 1
- numeric_cols = df.select_dtypes(np.number).columns
+ if fill_missing_report:
+ predicted = predicted.fillna(value=fill_missing_report)
+ predicted["dummy"] = 1
+ numeric_cols = predicted.select_dtypes(np.number).columns
report = (
- df.groupby(["Cluster"])[ # construct report dataframe
+ predicted.groupby(["Cluster"])[ # construct report dataframe
numeric_cols
] # group by cluster number
.agg(
[
- ("sum", np.sum),
+ ("sum", "sum"),
("mean_with_zeros", lambda x: np.mean(np.nan_to_num(x))),
- ("mean_without_zeros", lambda x: x.replace(0, np.NaN).mean()),
+ ("mean_without_zeros", lambda x: x.replace(0, np.nan).mean()),
(
"mean_25-75",
lambda x: np.mean(
@@ -256,7 +249,7 @@ def ReportGenerator(
)
),
),
- ("mean_with_na", np.mean),
+ ("mean_with_na", "mean"),
("min", lambda x: x.min()),
("5%", lambda x: x.quantile(0.05)),
("25%", lambda x: x.quantile(0.25)),
@@ -275,47 +268,44 @@ def ReportGenerator(
.rename(index=str, columns={"level_0": "Features", "level_1": "Type"})
) # rename columns
# calculate the size of cluster(count of clientID's)
+ # avoid SettingWithCopyWarning
clustersize = report[
(report["Features"] == "dummy") & (report["Type"] == "count")
- ].copy() # avoid SettingWithCopyWarning
- clustersize.Type = (
- "ClusterSize" # rename created cluster df to match report column names
- )
+ ].copy()
+ # rename created predicted cluster to match report column names
+ clustersize.Type = "ClusterSize"
clustersize.Features = "# of Customers"
+ # calculating the proportion of cluster
clusterproportion = pd.DataFrame(
- clustersize.iloc[:, 2:].values
- / clustersize.iloc[:, 2:].values.sum() # calculating the proportion of cluster
+ clustersize.iloc[:, 2:].to_numpy() / clustersize.iloc[:, 2:].to_numpy().sum()
)
- clusterproportion[
- "Type"
- ] = "% of Customers" # rename created cluster df to match report column names
+ # rename created predicted cluster to match report column names
+ clusterproportion["Type"] = "% of Customers"
clusterproportion["Features"] = "ClusterProportion"
cols = clusterproportion.columns.tolist()
cols = cols[-2:] + cols[:-2]
clusterproportion = clusterproportion[cols] # rearrange columns to match report
clusterproportion.columns = report.columns
+ # generating dataframe with count of nan values
a = pd.DataFrame(
abs(
- report[report["Type"] == "count"].iloc[:, 2:].values
- - clustersize.iloc[:, 2:].values
+ report[report["Type"] == "count"].iloc[:, 2:].to_numpy()
+ - clustersize.iloc[:, 2:].to_numpy()
)
- ) # generating df with count of nan values
+ )
a["Features"] = 0
a["Type"] = "# of nan"
- a.Features = report[
- report["Type"] == "count"
- ].Features.tolist() # filling values in order to match report
+ # filling values in order to match report
+ a.Features = report[report["Type"] == "count"].Features.tolist()
cols = a.columns.tolist()
cols = cols[-2:] + cols[:-2]
a = a[cols] # rearrange columns to match report
a.columns = report.columns # rename columns to match report
- report = report.drop(
- report[report.Type == "count"].index
- ) # drop count values except cluster size
- report = pd.concat(
- [report, a, clustersize, clusterproportion], axis=0
- ) # concat report with clustert size and nan values
- report["Mark"] = report["Features"].isin(ClusteringVariables)
+ # drop count values except for cluster size
+ report = report.drop(report[report.Type == "count"].index)
+ # concat report with cluster size and nan values
+ report = pd.concat([report, a, clustersize, clusterproportion], axis=0)
+ report["Mark"] = report["Features"].isin(clustering_variables)
cols = report.columns.tolist()
cols = cols[0:2] + cols[-1:] + cols[2:-1]
report = report[cols]
@@ -343,7 +333,7 @@ def ReportGenerator(
)
report.columns.name = ""
report = report.reset_index()
- report.drop(columns=["index"], inplace=True)
+ report = report.drop(columns=["index"])
return report
diff --git a/machine_learning/k_nearest_neighbours.py b/machine_learning/k_nearest_neighbours.py
index e90ea09a58c1..fbc1b8bd227e 100644
--- a/machine_learning/k_nearest_neighbours.py
+++ b/machine_learning/k_nearest_neighbours.py
@@ -1,58 +1,88 @@
+"""
+k-Nearest Neighbours (kNN) is a simple non-parametric supervised learning
+algorithm used for classification. Given some labelled training data, a given
+point is classified using its k nearest neighbours according to some distance
+metric. The most commonly occurring label among the neighbours becomes the label
+of the given point. In effect, the label of the given point is decided by a
+majority vote.
+
+This implementation uses the commonly used Euclidean distance metric, but other
+distance metrics can also be used.
+
+Reference: https://en.wikipedia.org/wiki/K-nearest_neighbors_algorithm
+"""
+
from collections import Counter
+from heapq import nsmallest
import numpy as np
from sklearn import datasets
from sklearn.model_selection import train_test_split
-data = datasets.load_iris()
-
-X = np.array(data["data"])
-y = np.array(data["target"])
-classes = data["target_names"]
-
-X_train, X_test, y_train, y_test = train_test_split(X, y)
-
-
-def euclidean_distance(a, b):
- """
- Gives the euclidean distance between two points
- >>> euclidean_distance([0, 0], [3, 4])
- 5.0
- >>> euclidean_distance([1, 2, 3], [1, 8, 11])
- 10.0
- """
- return np.linalg.norm(np.array(a) - np.array(b))
-
-
-def classifier(train_data, train_target, classes, point, k=5):
- """
- Classifies the point using the KNN algorithm
- k closest points are found (ranked in ascending order of euclidean distance)
- Params:
- :train_data: Set of points that are classified into two or more classes
- :train_target: List of classes in the order of train_data points
- :classes: Labels of the classes
- :point: The data point that needs to be classifed
-
- >>> X_train = [[0, 0], [1, 0], [0, 1], [0.5, 0.5], [3, 3], [2, 3], [3, 2]]
- >>> y_train = [0, 0, 0, 0, 1, 1, 1]
- >>> classes = ['A','B']; point = [1.2,1.2]
- >>> classifier(X_train, y_train, classes,point)
- 'A'
- """
- data = zip(train_data, train_target)
- # List of distances of all points from the point to be classified
- distances = []
- for data_point in data:
- distance = euclidean_distance(data_point[0], point)
- distances.append((distance, data_point[1]))
- # Choosing 'k' points with the least distances.
- votes = [i[1] for i in sorted(distances)[:k]]
- # Most commonly occurring class among them
- # is the class into which the point is classified
- result = Counter(votes).most_common(1)[0][0]
- return classes[result]
+
+class KNN:
+ def __init__(
+ self,
+ train_data: np.ndarray[float],
+ train_target: np.ndarray[int],
+ class_labels: list[str],
+ ) -> None:
+ """
+ Create a kNN classifier using the given training data and class labels
+ """
+ self.data = zip(train_data, train_target)
+ self.labels = class_labels
+
+ @staticmethod
+ def _euclidean_distance(a: np.ndarray[float], b: np.ndarray[float]) -> float:
+ """
+ Calculate the Euclidean distance between two points
+ >>> KNN._euclidean_distance(np.array([0, 0]), np.array([3, 4]))
+ 5.0
+ >>> KNN._euclidean_distance(np.array([1, 2, 3]), np.array([1, 8, 11]))
+ 10.0
+ """
+ return float(np.linalg.norm(a - b))
+
+ def classify(self, pred_point: np.ndarray[float], k: int = 5) -> str:
+ """
+ Classify a given point using the kNN algorithm
+ >>> train_X = np.array(
+ ... [[0, 0], [1, 0], [0, 1], [0.5, 0.5], [3, 3], [2, 3], [3, 2]]
+ ... )
+ >>> train_y = np.array([0, 0, 0, 0, 1, 1, 1])
+ >>> classes = ['A', 'B']
+ >>> knn = KNN(train_X, train_y, classes)
+ >>> point = np.array([1.2, 1.2])
+ >>> knn.classify(point)
+ 'A'
+ """
+ # Distances of all points from the point to be classified
+ distances = (
+ (self._euclidean_distance(data_point[0], pred_point), data_point[1])
+ for data_point in self.data
+ )
+
+ # Choosing k points with the shortest distances
+ votes = (i[1] for i in nsmallest(k, distances))
+
+ # Most commonly occurring class is the one into which the point is classified
+ result = Counter(votes).most_common(1)[0][0]
+ return self.labels[result]
if __name__ == "__main__":
- print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
+ import doctest
+
+ doctest.testmod()
+
+ iris = datasets.load_iris()
+
+ X = np.array(iris["data"])
+ y = np.array(iris["target"])
+ iris_classes = iris["target_names"]
+
+ X_train, X_test, y_train, y_test = train_test_split(X, y, random_state=0)
+ iris_point = np.array([4.4, 3.1, 1.3, 1.4])
+ classifier = KNN(X_train, y_train, iris_classes)
+ print(classifier.classify(iris_point, k=3))
diff --git a/machine_learning/knn_sklearn.py b/machine_learning/knn_sklearn.py
deleted file mode 100644
index 9a9114102ff3..000000000000
--- a/machine_learning/knn_sklearn.py
+++ /dev/null
@@ -1,31 +0,0 @@
-from sklearn.datasets import load_iris
-from sklearn.model_selection import train_test_split
-from sklearn.neighbors import KNeighborsClassifier
-
-# Load iris file
-iris = load_iris()
-iris.keys()
-
-
-print(f"Target names: \n {iris.target_names} ")
-print(f"\n Features: \n {iris.feature_names}")
-
-# Train set e Test set
-X_train, X_test, y_train, y_test = train_test_split(
- iris["data"], iris["target"], random_state=4
-)
-
-# KNN
-
-knn = KNeighborsClassifier(n_neighbors=1)
-knn.fit(X_train, y_train)
-
-# new array to test
-X_new = [[1, 2, 1, 4], [2, 3, 4, 5]]
-
-prediction = knn.predict(X_new)
-
-print(
- "\nNew array: \n {}"
- "\n\nTarget Names Prediction: \n {}".format(X_new, iris["target_names"][prediction])
-)
diff --git a/machine_learning/linear_discriminant_analysis.py b/machine_learning/linear_discriminant_analysis.py
index 0d19e970e973..8528ccbbae51 100644
--- a/machine_learning/linear_discriminant_analysis.py
+++ b/machine_learning/linear_discriminant_analysis.py
@@ -1,51 +1,53 @@
"""
- Linear Discriminant Analysis
+Linear Discriminant Analysis
- Assumptions About Data :
- 1. The input variables has a gaussian distribution.
- 2. The variance calculated for each input variables by class grouping is the
- same.
- 3. The mix of classes in your training set is representative of the problem.
+Assumptions About Data :
+ 1. The input variables has a gaussian distribution.
+ 2. The variance calculated for each input variables by class grouping is the
+ same.
+ 3. The mix of classes in your training set is representative of the problem.
- Learning The Model :
- The LDA model requires the estimation of statistics from the training data :
- 1. Mean of each input value for each class.
- 2. Probability of an instance belong to each class.
- 3. Covariance for the input data for each class
+Learning The Model :
+ The LDA model requires the estimation of statistics from the training data :
+ 1. Mean of each input value for each class.
+ 2. Probability of an instance belong to each class.
+ 3. Covariance for the input data for each class
- Calculate the class means :
- mean(x) = 1/n ( for i = 1 to i = n --> sum(xi))
+ Calculate the class means :
+ mean(x) = 1/n ( for i = 1 to i = n --> sum(xi))
- Calculate the class probabilities :
- P(y = 0) = count(y = 0) / (count(y = 0) + count(y = 1))
- P(y = 1) = count(y = 1) / (count(y = 0) + count(y = 1))
+ Calculate the class probabilities :
+ P(y = 0) = count(y = 0) / (count(y = 0) + count(y = 1))
+ P(y = 1) = count(y = 1) / (count(y = 0) + count(y = 1))
- Calculate the variance :
- We can calculate the variance for dataset in two steps :
- 1. Calculate the squared difference for each input variable from the
- group mean.
- 2. Calculate the mean of the squared difference.
- ------------------------------------------------
- Squared_Difference = (x - mean(k)) ** 2
- Variance = (1 / (count(x) - count(classes))) *
- (for i = 1 to i = n --> sum(Squared_Difference(xi)))
+ Calculate the variance :
+ We can calculate the variance for dataset in two steps :
+ 1. Calculate the squared difference for each input variable from the
+ group mean.
+ 2. Calculate the mean of the squared difference.
+ ------------------------------------------------
+ Squared_Difference = (x - mean(k)) ** 2
+ Variance = (1 / (count(x) - count(classes))) *
+ (for i = 1 to i = n --> sum(Squared_Difference(xi)))
- Making Predictions :
- discriminant(x) = x * (mean / variance) -
- ((mean ** 2) / (2 * variance)) + Ln(probability)
- ---------------------------------------------------------------------------
- After calculating the discriminant value for each class, the class with the
- largest discriminant value is taken as the prediction.
+Making Predictions :
+ discriminant(x) = x * (mean / variance) -
+ ((mean ** 2) / (2 * variance)) + Ln(probability)
+ ---------------------------------------------------------------------------
+ After calculating the discriminant value for each class, the class with the
+ largest discriminant value is taken as the prediction.
- Author: @EverLookNeverSee
+Author: @EverLookNeverSee
"""
+
+from collections.abc import Callable
from math import log
from os import name, system
from random import gauss, seed
-from typing import Callable, TypeVar
+from typing import TypeVar
# Make a training dataset drawn from a gaussian distribution
@@ -254,8 +256,8 @@ def valid_input(
input_type: Callable[[object], num], # Usually float or int
input_msg: str,
err_msg: str,
- condition: Callable[[num], bool] = lambda x: True,
- default: str = None,
+ condition: Callable[[num], bool] = lambda _: True,
+ default: str | None = None,
) -> num:
"""
Ask for user value and validate that it fulfill a condition.
@@ -283,7 +285,7 @@ def valid_input(
# Main Function
def main():
- """ This function starts execution phase """
+ """This function starts execution phase"""
while True:
print(" Linear Discriminant Analysis ".center(50, "*"))
print("*" * 50, "\n")
@@ -320,7 +322,7 @@ def main():
user_count = valid_input(
input_type=int,
condition=lambda x: x > 0,
- input_msg=(f"Enter The number of instances for class_{i+1}: "),
+ input_msg=(f"Enter The number of instances for class_{i + 1}: "),
err_msg="Number of instances should be positive!",
)
counts.append(user_count)
@@ -331,7 +333,7 @@ def main():
for a in range(n_classes):
user_mean = valid_input(
input_type=float,
- input_msg=(f"Enter the value of mean for class_{a+1}: "),
+ input_msg=(f"Enter the value of mean for class_{a + 1}: "),
err_msg="This is an invalid value.",
)
user_means.append(user_mean)
@@ -398,7 +400,7 @@ def main():
if input("Press any key to restart or 'q' for quit: ").strip().lower() == "q":
print("\n" + "GoodBye!".center(100, "-") + "\n")
break
- system("cls" if name == "nt" else "clear")
+ system("cls" if name == "nt" else "clear") # noqa: S605
if __name__ == "__main__":
diff --git a/machine_learning/linear_regression.py b/machine_learning/linear_regression.py
index a726629efe00..1d11e5a9cc2b 100644
--- a/machine_learning/linear_regression.py
+++ b/machine_learning/linear_regression.py
@@ -7,6 +7,7 @@
fit our dataset. In this particular code, I had used a CSGO dataset (ADR vs
Rating). We try to best fit a line through dataset and estimate the parameters.
"""
+
import numpy as np
import requests
@@ -17,9 +18,9 @@ def collect_dataset():
:return : dataset obtained from the link, as matrix
"""
response = requests.get(
- "https://raw.githubusercontent.com/yashLadha/"
- + "The_Math_of_Intelligence/master/Week1/ADRvs"
- + "Rating.csv"
+ "https://raw.githubusercontent.com/yashLadha/The_Math_of_Intelligence/"
+ "master/Week1/ADRvsRating.csv",
+ timeout=10,
)
lines = response.text.splitlines()
data = []
@@ -40,6 +41,14 @@ def run_steep_gradient_descent(data_x, data_y, len_data, alpha, theta):
:param theta : Feature vector (weight's for our model)
;param return : Updated Feature's, using
curr_features - alpha_ * gradient(w.r.t. feature)
+ >>> import numpy as np
+ >>> data_x = np.array([[1, 2], [3, 4]])
+ >>> data_y = np.array([5, 6])
+ >>> len_data = len(data_x)
+ >>> alpha = 0.01
+ >>> theta = np.array([0.1, 0.2])
+ >>> run_steep_gradient_descent(data_x, data_y, len_data, alpha, theta)
+ array([0.196, 0.343])
"""
n = len_data
@@ -57,6 +66,12 @@ def sum_of_square_error(data_x, data_y, len_data, theta):
:param len_data : len of the dataset
:param theta : contains the feature vector
:return : sum of square error computed from given feature's
+
+ Example:
+ >>> vc_x = np.array([[1.1], [2.1], [3.1]])
+ >>> vc_y = np.array([1.2, 2.2, 3.2])
+ >>> round(sum_of_square_error(vc_x, vc_y, 3, np.array([1])),3)
+ np.float64(0.005)
"""
prod = np.dot(theta, data_x.transpose())
prod -= data_y.transpose()
@@ -79,16 +94,31 @@ def run_linear_regression(data_x, data_y):
theta = np.zeros((1, no_features))
- for i in range(0, iterations):
+ for i in range(iterations):
theta = run_steep_gradient_descent(data_x, data_y, len_data, alpha, theta)
error = sum_of_square_error(data_x, data_y, len_data, theta)
- print("At Iteration %d - Error is %.5f " % (i + 1, error))
+ print(f"At Iteration {i + 1} - Error is {error:.5f}")
return theta
+def mean_absolute_error(predicted_y, original_y):
+ """Return sum of square error for error calculation
+ :param predicted_y : contains the output of prediction (result vector)
+ :param original_y : contains values of expected outcome
+ :return : mean absolute error computed from given feature's
+
+ >>> predicted_y = [3, -0.5, 2, 7]
+ >>> original_y = [2.5, 0.0, 2, 8]
+ >>> mean_absolute_error(predicted_y, original_y)
+ 0.5
+ """
+ total = sum(abs(y - predicted_y[i]) for i, y in enumerate(original_y))
+ return total / len(original_y)
+
+
def main():
- """ Driver function """
+ """Driver function"""
data = collect_dataset()
len_data = data.shape[0]
@@ -98,9 +128,12 @@ def main():
theta = run_linear_regression(data_x, data_y)
len_result = theta.shape[1]
print("Resultant Feature vector : ")
- for i in range(0, len_result):
- print("%.5f" % (theta[0, i]))
+ for i in range(len_result):
+ print(f"{theta[0, i]:.5f}")
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
main()
diff --git a/machine_learning/local_weighted_learning/README.md b/machine_learning/local_weighted_learning/README.md
new file mode 100644
index 000000000000..ef4dbc958600
--- /dev/null
+++ b/machine_learning/local_weighted_learning/README.md
@@ -0,0 +1,66 @@
+# Locally Weighted Linear Regression
+It is a non-parametric ML algorithm that does not learn on a fixed set of parameters such as **linear regression**. \
+So, here comes a question of what is *linear regression*? \
+**Linear regression** is a supervised learning algorithm used for computing linear relationships between input (X) and output (Y). \
+
+### Terminology Involved
+
+number_of_features(i) = Number of features involved. \
+number_of_training_examples(m) = Number of training examples. \
+output_sequence(y) = Output Sequence. \
+$\theta$ $^T$ x = predicted point. \
+J($\theta$) = COst function of point.
+
+The steps involved in ordinary linear regression are:
+
+Training phase: Compute \theta to minimize the cost. \
+J($\theta$) = $\sum_{i=1}^m$ (($\theta$)$^T$ $x^i$ - $y^i$)$^2$
+
+Predict output: for given query point x, \
+ return: ($\theta$)$^T$ x
+
+
+
+This training phase is possible when data points are linear, but there again comes a question can we predict non-linear relationship between x and y ? as shown below
+
+
+
+
+So, here comes the role of non-parametric algorithm which doesn't compute predictions based on fixed set of params. Rather parameters $\theta$ are computed individually for each query point/data point x.
+
+
+While Computing $\theta$ , a higher preference is given to points in the vicinity of x than points farther from x.
+
+Cost Function J($\theta$) = $\sum_{i=1}^m$ $w^i$ (($\theta$)$^T$ $x^i$ - $y^i$)$^2$
+
+$w^i$ is non-negative weight associated to training point $x^i$. \
+$w^i$ is large fr $x^i$'s lying closer to query point $x_i$. \
+$w^i$ is small for $x^i$'s lying farther to query point $x_i$.
+
+A Typical weight can be computed using \
+
+$w^i$ = $\exp$(-$\frac{(x^i-x)(x^i-x)^T}{2\tau^2}$)
+
+Where $\tau$ is the bandwidth parameter that controls $w^i$ distance from x.
+
+Let's look at a example :
+
+Suppose, we had a query point x=5.0 and training points $x^1$=4.9 and $x^2$=5.0 than we can calculate weights as :
+
+$w^i$ = $\exp$(-$\frac{(x^i-x)(x^i-x)^T}{2\tau^2}$) with $\tau$=0.5
+
+$w^1$ = $\exp$(-$\frac{(4.9-5)^2}{2(0.5)^2}$) = 0.9802
+
+$w^2$ = $\exp$(-$\frac{(3-5)^2}{2(0.5)^2}$) = 0.000335
+
+So, J($\theta$) = 0.9802*($\theta$ $^T$ $x^1$ - $y^1$) + 0.000335*($\theta$ $^T$ $x^2$ - $y^2$)
+
+So, here by we can conclude that the weight fall exponentially as the distance between x & $x^i$ increases and So, does the contribution of error in prediction for $x^i$ to the cost.
+
+Steps involved in LWL are : \
+Compute \theta to minimize the cost.
+J($\theta$) = $\sum_{i=1}^m$ $w^i$ (($\theta$)$^T$ $x^i$ - $y^i$)$^2$ \
+Predict Output: for given query point x, \
+return : $\theta$ $^T$ x
+
+
diff --git a/machine_learning/local_weighted_learning/__init__.py b/machine_learning/local_weighted_learning/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/machine_learning/local_weighted_learning/local_weighted_learning.py b/machine_learning/local_weighted_learning/local_weighted_learning.py
new file mode 100644
index 000000000000..f3056da40e24
--- /dev/null
+++ b/machine_learning/local_weighted_learning/local_weighted_learning.py
@@ -0,0 +1,185 @@
+"""
+Locally weighted linear regression, also called local regression, is a type of
+non-parametric linear regression that prioritizes data closest to a given
+prediction point. The algorithm estimates the vector of model coefficients β
+using weighted least squares regression:
+
+β = (XᵀWX)⁻¹(XᵀWy),
+
+where X is the design matrix, y is the response vector, and W is the diagonal
+weight matrix.
+
+This implementation calculates wᵢ, the weight of the ith training sample, using
+the Gaussian weight:
+
+wᵢ = exp(-‖xᵢ - x‖²/(2τ²)),
+
+where xᵢ is the ith training sample, x is the prediction point, τ is the
+"bandwidth", and ‖x‖ is the Euclidean norm (also called the 2-norm or the L²
+norm). The bandwidth τ controls how quickly the weight of a training sample
+decreases as its distance from the prediction point increases. One can think of
+the Gaussian weight as a bell curve centered around the prediction point: a
+training sample is weighted lower if it's farther from the center, and τ
+controls the spread of the bell curve.
+
+Other types of locally weighted regression such as locally estimated scatterplot
+smoothing (LOESS) typically use different weight functions.
+
+References:
+ - https://en.wikipedia.org/wiki/Local_regression
+ - https://en.wikipedia.org/wiki/Weighted_least_squares
+ - https://cs229.stanford.edu/notes2022fall/main_notes.pdf
+"""
+
+import matplotlib.pyplot as plt
+import numpy as np
+
+
+def weight_matrix(point: np.ndarray, x_train: np.ndarray, tau: float) -> np.ndarray:
+ """
+ Calculate the weight of every point in the training data around a given
+ prediction point
+
+ Args:
+ point: x-value at which the prediction is being made
+ x_train: ndarray of x-values for training
+ tau: bandwidth value, controls how quickly the weight of training values
+ decreases as the distance from the prediction point increases
+
+ Returns:
+ m x m weight matrix around the prediction point, where m is the size of
+ the training set
+ >>> weight_matrix(
+ ... np.array([1., 1.]),
+ ... np.array([[16.99, 10.34], [21.01,23.68], [24.59,25.69]]),
+ ... 0.6
+ ... )
+ array([[1.43807972e-207, 0.00000000e+000, 0.00000000e+000],
+ [0.00000000e+000, 0.00000000e+000, 0.00000000e+000],
+ [0.00000000e+000, 0.00000000e+000, 0.00000000e+000]])
+ """
+ m = len(x_train) # Number of training samples
+ weights = np.eye(m) # Initialize weights as identity matrix
+ for j in range(m):
+ diff = point - x_train[j]
+ weights[j, j] = np.exp(diff @ diff.T / (-2.0 * tau**2))
+
+ return weights
+
+
+def local_weight(
+ point: np.ndarray, x_train: np.ndarray, y_train: np.ndarray, tau: float
+) -> np.ndarray:
+ """
+ Calculate the local weights at a given prediction point using the weight
+ matrix for that point
+
+ Args:
+ point: x-value at which the prediction is being made
+ x_train: ndarray of x-values for training
+ y_train: ndarray of y-values for training
+ tau: bandwidth value, controls how quickly the weight of training values
+ decreases as the distance from the prediction point increases
+ Returns:
+ ndarray of local weights
+ >>> local_weight(
+ ... np.array([1., 1.]),
+ ... np.array([[16.99, 10.34], [21.01,23.68], [24.59,25.69]]),
+ ... np.array([[1.01, 1.66, 3.5]]),
+ ... 0.6
+ ... )
+ array([[0.00873174],
+ [0.08272556]])
+ """
+ weight_mat = weight_matrix(point, x_train, tau)
+ weight = np.linalg.inv(x_train.T @ weight_mat @ x_train) @ (
+ x_train.T @ weight_mat @ y_train.T
+ )
+
+ return weight
+
+
+def local_weight_regression(
+ x_train: np.ndarray, y_train: np.ndarray, tau: float
+) -> np.ndarray:
+ """
+ Calculate predictions for each point in the training data
+
+ Args:
+ x_train: ndarray of x-values for training
+ y_train: ndarray of y-values for training
+ tau: bandwidth value, controls how quickly the weight of training values
+ decreases as the distance from the prediction point increases
+
+ Returns:
+ ndarray of predictions
+ >>> local_weight_regression(
+ ... np.array([[16.99, 10.34], [21.01, 23.68], [24.59, 25.69]]),
+ ... np.array([[1.01, 1.66, 3.5]]),
+ ... 0.6
+ ... )
+ array([1.07173261, 1.65970737, 3.50160179])
+ """
+ y_pred = np.zeros(len(x_train)) # Initialize array of predictions
+ for i, item in enumerate(x_train):
+ y_pred[i] = np.dot(item, local_weight(item, x_train, y_train, tau)).item()
+
+ return y_pred
+
+
+def load_data(
+ dataset_name: str, x_name: str, y_name: str
+) -> tuple[np.ndarray, np.ndarray, np.ndarray]:
+ """
+ Load data from seaborn and split it into x and y points
+ >>> pass # No doctests, function is for demo purposes only
+ """
+ import seaborn as sns
+
+ data = sns.load_dataset(dataset_name)
+ x_data = np.array(data[x_name])
+ y_data = np.array(data[y_name])
+
+ one = np.ones(len(y_data))
+
+ # pairing elements of one and x_data
+ x_train = np.column_stack((one, x_data))
+
+ return x_train, x_data, y_data
+
+
+def plot_preds(
+ x_train: np.ndarray,
+ preds: np.ndarray,
+ x_data: np.ndarray,
+ y_data: np.ndarray,
+ x_name: str,
+ y_name: str,
+) -> None:
+ """
+ Plot predictions and display the graph
+ >>> pass # No doctests, function is for demo purposes only
+ """
+ x_train_sorted = np.sort(x_train, axis=0)
+ plt.scatter(x_data, y_data, color="blue")
+ plt.plot(
+ x_train_sorted[:, 1],
+ preds[x_train[:, 1].argsort(0)],
+ color="yellow",
+ linewidth=5,
+ )
+ plt.title("Local Weighted Regression")
+ plt.xlabel(x_name)
+ plt.ylabel(y_name)
+ plt.show()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ # Demo with a dataset from the seaborn module
+ training_data_x, total_bill, tip = load_data("tips", "total_bill", "tip")
+ predictions = local_weight_regression(training_data_x, tip, 5)
+ plot_preds(training_data_x, predictions, total_bill, tip, "total_bill", "tip")
diff --git a/machine_learning/logistic_regression.py b/machine_learning/logistic_regression.py
index 48d88ef61185..496026631fbe 100644
--- a/machine_learning/logistic_regression.py
+++ b/machine_learning/logistic_regression.py
@@ -14,6 +14,7 @@
Coursera ML course
https://medium.com/@martinpella/logistic-regression-from-scratch-in-python-124c5636b8ac
"""
+
import numpy as np
from matplotlib import pyplot as plt
from sklearn import datasets
@@ -27,57 +28,128 @@
# classification problems
-def sigmoid_function(z):
+def sigmoid_function(z: float | np.ndarray) -> float | np.ndarray:
+ """
+ Also known as Logistic Function.
+
+ 1
+ f(x) = -------
+ 1 + e⁻ˣ
+
+ The sigmoid function approaches a value of 1 as its input 'x' becomes
+ increasing positive. Opposite for negative values.
+
+ Reference: https://en.wikipedia.org/wiki/Sigmoid_function
+
+ @param z: input to the function
+ @returns: returns value in the range 0 to 1
+
+ Examples:
+ >>> float(sigmoid_function(4))
+ 0.9820137900379085
+ >>> sigmoid_function(np.array([-3, 3]))
+ array([0.04742587, 0.95257413])
+ >>> sigmoid_function(np.array([-3, 3, 1]))
+ array([0.04742587, 0.95257413, 0.73105858])
+ >>> sigmoid_function(np.array([-0.01, -2, -1.9]))
+ array([0.49750002, 0.11920292, 0.13010847])
+ >>> sigmoid_function(np.array([-1.3, 5.3, 12]))
+ array([0.21416502, 0.9950332 , 0.99999386])
+ >>> sigmoid_function(np.array([0.01, 0.02, 4.1]))
+ array([0.50249998, 0.50499983, 0.9836975 ])
+ >>> sigmoid_function(np.array([0.8]))
+ array([0.68997448])
+ """
return 1 / (1 + np.exp(-z))
-def cost_function(h, y):
- return (-y * np.log(h) - (1 - y) * np.log(1 - h)).mean()
+def cost_function(h: np.ndarray, y: np.ndarray) -> float:
+ """
+ Cost function quantifies the error between predicted and expected values.
+ The cost function used in Logistic Regression is called Log Loss
+ or Cross Entropy Function.
+
+ J(θ) = (1/m) * Σ [ -y * log(hθ(x)) - (1 - y) * log(1 - hθ(x)) ]
+
+ Where:
+ - J(θ) is the cost that we want to minimize during training
+ - m is the number of training examples
+ - Σ represents the summation over all training examples
+ - y is the actual binary label (0 or 1) for a given example
+ - hθ(x) is the predicted probability that x belongs to the positive class
+
+ @param h: the output of sigmoid function. It is the estimated probability
+ that the input example 'x' belongs to the positive class
+
+ @param y: the actual binary label associated with input example 'x'
+ Examples:
+ >>> estimations = sigmoid_function(np.array([0.3, -4.3, 8.1]))
+ >>> cost_function(h=estimations,y=np.array([1, 0, 1]))
+ 0.18937868932131605
+ >>> estimations = sigmoid_function(np.array([4, 3, 1]))
+ >>> cost_function(h=estimations,y=np.array([1, 0, 0]))
+ 1.459999655669926
+ >>> estimations = sigmoid_function(np.array([4, -3, -1]))
+ >>> cost_function(h=estimations,y=np.array([1,0,0]))
+ 0.1266663223365915
+ >>> estimations = sigmoid_function(0)
+ >>> cost_function(h=estimations,y=np.array([1]))
+ 0.6931471805599453
-def log_likelihood(X, Y, weights):
- scores = np.dot(X, weights)
- return np.sum(Y * scores - np.log(1 + np.exp(scores)))
+ References:
+ - https://en.wikipedia.org/wiki/Logistic_regression
+ """
+ return float((-y * np.log(h) - (1 - y) * np.log(1 - h)).mean())
+
+
+def log_likelihood(x, y, weights):
+ scores = np.dot(x, weights)
+ return np.sum(y * scores - np.log(1 + np.exp(scores)))
# here alpha is the learning rate, X is the feature matrix,y is the target matrix
-def logistic_reg(alpha, X, y, max_iterations=70000):
- theta = np.zeros(X.shape[1])
+def logistic_reg(alpha, x, y, max_iterations=70000):
+ theta = np.zeros(x.shape[1])
for iterations in range(max_iterations):
- z = np.dot(X, theta)
+ z = np.dot(x, theta)
h = sigmoid_function(z)
- gradient = np.dot(X.T, h - y) / y.size
+ gradient = np.dot(x.T, h - y) / y.size
theta = theta - alpha * gradient # updating the weights
- z = np.dot(X, theta)
+ z = np.dot(x, theta)
h = sigmoid_function(z)
- J = cost_function(h, y)
+ j = cost_function(h, y)
if iterations % 100 == 0:
- print(f"loss: {J} \t") # printing the loss after every 100 iterations
+ print(f"loss: {j} \t") # printing the loss after every 100 iterations
return theta
# In[68]:
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
iris = datasets.load_iris()
- X = iris.data[:, :2]
+ x = iris.data[:, :2]
y = (iris.target != 0) * 1
alpha = 0.1
- theta = logistic_reg(alpha, X, y, max_iterations=70000)
+ theta = logistic_reg(alpha, x, y, max_iterations=70000)
print("theta: ", theta) # printing the theta i.e our weights vector
- def predict_prob(X):
+ def predict_prob(x):
return sigmoid_function(
- np.dot(X, theta)
+ np.dot(x, theta)
) # predicting the value of probability from the logistic regression algorithm
plt.figure(figsize=(10, 6))
- plt.scatter(X[y == 0][:, 0], X[y == 0][:, 1], color="b", label="0")
- plt.scatter(X[y == 1][:, 0], X[y == 1][:, 1], color="r", label="1")
- (x1_min, x1_max) = (X[:, 0].min(), X[:, 0].max())
- (x2_min, x2_max) = (X[:, 1].min(), X[:, 1].max())
+ plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color="b", label="0")
+ plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color="r", label="1")
+ (x1_min, x1_max) = (x[:, 0].min(), x[:, 0].max())
+ (x2_min, x2_max) = (x[:, 1].min(), x[:, 1].max())
(xx1, xx2) = np.meshgrid(np.linspace(x1_min, x1_max), np.linspace(x2_min, x2_max))
grid = np.c_[xx1.ravel(), xx2.ravel()]
probs = predict_prob(grid).reshape(xx1.shape)
diff --git a/machine_learning/loss_functions.py b/machine_learning/loss_functions.py
new file mode 100644
index 000000000000..0bd9aa8b5401
--- /dev/null
+++ b/machine_learning/loss_functions.py
@@ -0,0 +1,669 @@
+import numpy as np
+
+
+def binary_cross_entropy(
+ y_true: np.ndarray, y_pred: np.ndarray, epsilon: float = 1e-15
+) -> float:
+ """
+ Calculate the mean binary cross-entropy (BCE) loss between true labels and predicted
+ probabilities.
+
+ BCE loss quantifies dissimilarity between true labels (0 or 1) and predicted
+ probabilities. It's widely used in binary classification tasks.
+
+ BCE = -Σ(y_true * ln(y_pred) + (1 - y_true) * ln(1 - y_pred))
+
+ Reference: https://en.wikipedia.org/wiki/Cross_entropy
+
+ Parameters:
+ - y_true: True binary labels (0 or 1)
+ - y_pred: Predicted probabilities for class 1
+ - epsilon: Small constant to avoid numerical instability
+
+ >>> true_labels = np.array([0, 1, 1, 0, 1])
+ >>> predicted_probs = np.array([0.2, 0.7, 0.9, 0.3, 0.8])
+ >>> float(binary_cross_entropy(true_labels, predicted_probs))
+ 0.2529995012327421
+ >>> true_labels = np.array([0, 1, 1, 0, 1])
+ >>> predicted_probs = np.array([0.3, 0.8, 0.9, 0.2])
+ >>> binary_cross_entropy(true_labels, predicted_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+
+ y_pred = np.clip(y_pred, epsilon, 1 - epsilon) # Clip predictions to avoid log(0)
+ bce_loss = -(y_true * np.log(y_pred) + (1 - y_true) * np.log(1 - y_pred))
+ return np.mean(bce_loss)
+
+
+def binary_focal_cross_entropy(
+ y_true: np.ndarray,
+ y_pred: np.ndarray,
+ gamma: float = 2.0,
+ alpha: float = 0.25,
+ epsilon: float = 1e-15,
+) -> float:
+ """
+ Calculate the mean binary focal cross-entropy (BFCE) loss between true labels
+ and predicted probabilities.
+
+ BFCE loss quantifies dissimilarity between true labels (0 or 1) and predicted
+ probabilities. It's a variation of binary cross-entropy that addresses class
+ imbalance by focusing on hard examples.
+
+ BCFE = -Σ(alpha * (1 - y_pred)**gamma * y_true * log(y_pred)
+ + (1 - alpha) * y_pred**gamma * (1 - y_true) * log(1 - y_pred))
+
+ Reference: [Lin et al., 2018](https://arxiv.org/pdf/1708.02002.pdf)
+
+ Parameters:
+ - y_true: True binary labels (0 or 1).
+ - y_pred: Predicted probabilities for class 1.
+ - gamma: Focusing parameter for modulating the loss (default: 2.0).
+ - alpha: Weighting factor for class 1 (default: 0.25).
+ - epsilon: Small constant to avoid numerical instability.
+
+ >>> true_labels = np.array([0, 1, 1, 0, 1])
+ >>> predicted_probs = np.array([0.2, 0.7, 0.9, 0.3, 0.8])
+ >>> float(binary_focal_cross_entropy(true_labels, predicted_probs))
+ 0.008257977659239775
+ >>> true_labels = np.array([0, 1, 1, 0, 1])
+ >>> predicted_probs = np.array([0.3, 0.8, 0.9, 0.2])
+ >>> binary_focal_cross_entropy(true_labels, predicted_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+ # Clip predicted probabilities to avoid log(0)
+ y_pred = np.clip(y_pred, epsilon, 1 - epsilon)
+
+ bcfe_loss = -(
+ alpha * (1 - y_pred) ** gamma * y_true * np.log(y_pred)
+ + (1 - alpha) * y_pred**gamma * (1 - y_true) * np.log(1 - y_pred)
+ )
+
+ return np.mean(bcfe_loss)
+
+
+def categorical_cross_entropy(
+ y_true: np.ndarray, y_pred: np.ndarray, epsilon: float = 1e-15
+) -> float:
+ """
+ Calculate categorical cross-entropy (CCE) loss between true class labels and
+ predicted class probabilities.
+
+ CCE = -Σ(y_true * ln(y_pred))
+
+ Reference: https://en.wikipedia.org/wiki/Cross_entropy
+
+ Parameters:
+ - y_true: True class labels (one-hot encoded)
+ - y_pred: Predicted class probabilities
+ - epsilon: Small constant to avoid numerical instability
+
+ >>> true_labels = np.array([[1, 0, 0], [0, 1, 0], [0, 0, 1]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1], [0.0, 0.1, 0.9]])
+ >>> float(categorical_cross_entropy(true_labels, pred_probs))
+ 0.567395975254385
+ >>> true_labels = np.array([[1, 0], [0, 1]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1]])
+ >>> categorical_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same shape.
+ >>> true_labels = np.array([[2, 0, 1], [1, 0, 0]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1]])
+ >>> categorical_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: y_true must be one-hot encoded.
+ >>> true_labels = np.array([[1, 0, 1], [1, 0, 0]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1]])
+ >>> categorical_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: y_true must be one-hot encoded.
+ >>> true_labels = np.array([[1, 0, 0], [0, 1, 0]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.1], [0.2, 0.7, 0.1]])
+ >>> categorical_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Predicted probabilities must sum to approximately 1.
+ """
+ if y_true.shape != y_pred.shape:
+ raise ValueError("Input arrays must have the same shape.")
+
+ if np.any((y_true != 0) & (y_true != 1)) or np.any(y_true.sum(axis=1) != 1):
+ raise ValueError("y_true must be one-hot encoded.")
+
+ if not np.all(np.isclose(np.sum(y_pred, axis=1), 1, rtol=epsilon, atol=epsilon)):
+ raise ValueError("Predicted probabilities must sum to approximately 1.")
+
+ y_pred = np.clip(y_pred, epsilon, 1) # Clip predictions to avoid log(0)
+ return -np.sum(y_true * np.log(y_pred))
+
+
+def categorical_focal_cross_entropy(
+ y_true: np.ndarray,
+ y_pred: np.ndarray,
+ alpha: np.ndarray = None,
+ gamma: float = 2.0,
+ epsilon: float = 1e-15,
+) -> float:
+ """
+ Calculate the mean categorical focal cross-entropy (CFCE) loss between true
+ labels and predicted probabilities for multi-class classification.
+
+ CFCE loss is a generalization of binary focal cross-entropy for multi-class
+ classification. It addresses class imbalance by focusing on hard examples.
+
+ CFCE = -Σ alpha * (1 - y_pred)**gamma * y_true * log(y_pred)
+
+ Reference: [Lin et al., 2018](https://arxiv.org/pdf/1708.02002.pdf)
+
+ Parameters:
+ - y_true: True labels in one-hot encoded form.
+ - y_pred: Predicted probabilities for each class.
+ - alpha: Array of weighting factors for each class.
+ - gamma: Focusing parameter for modulating the loss (default: 2.0).
+ - epsilon: Small constant to avoid numerical instability.
+
+ Returns:
+ - The mean categorical focal cross-entropy loss.
+
+ >>> true_labels = np.array([[1, 0, 0], [0, 1, 0], [0, 0, 1]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1], [0.0, 0.1, 0.9]])
+ >>> alpha = np.array([0.6, 0.2, 0.7])
+ >>> float(categorical_focal_cross_entropy(true_labels, pred_probs, alpha))
+ 0.0025966118981496423
+
+ >>> true_labels = np.array([[0, 1, 0], [0, 0, 1]])
+ >>> pred_probs = np.array([[0.05, 0.95, 0], [0.1, 0.8, 0.1]])
+ >>> alpha = np.array([0.25, 0.25, 0.25])
+ >>> float(categorical_focal_cross_entropy(true_labels, pred_probs, alpha))
+ 0.23315276982014324
+
+ >>> true_labels = np.array([[1, 0], [0, 1]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1]])
+ >>> categorical_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same shape.
+
+ >>> true_labels = np.array([[2, 0, 1], [1, 0, 0]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1]])
+ >>> categorical_focal_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: y_true must be one-hot encoded.
+
+ >>> true_labels = np.array([[1, 0, 1], [1, 0, 0]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1]])
+ >>> categorical_focal_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: y_true must be one-hot encoded.
+
+ >>> true_labels = np.array([[1, 0, 0], [0, 1, 0]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.1], [0.2, 0.7, 0.1]])
+ >>> categorical_focal_cross_entropy(true_labels, pred_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Predicted probabilities must sum to approximately 1.
+
+ >>> true_labels = np.array([[1, 0, 0], [0, 1, 0], [0, 0, 1]])
+ >>> pred_probs = np.array([[0.9, 0.1, 0.0], [0.2, 0.7, 0.1], [0.0, 0.1, 0.9]])
+ >>> alpha = np.array([0.6, 0.2])
+ >>> categorical_focal_cross_entropy(true_labels, pred_probs, alpha)
+ Traceback (most recent call last):
+ ...
+ ValueError: Length of alpha must match the number of classes.
+ """
+ if y_true.shape != y_pred.shape:
+ raise ValueError("Shape of y_true and y_pred must be the same.")
+
+ if alpha is None:
+ alpha = np.ones(y_true.shape[1])
+
+ if np.any((y_true != 0) & (y_true != 1)) or np.any(y_true.sum(axis=1) != 1):
+ raise ValueError("y_true must be one-hot encoded.")
+
+ if len(alpha) != y_true.shape[1]:
+ raise ValueError("Length of alpha must match the number of classes.")
+
+ if not np.all(np.isclose(np.sum(y_pred, axis=1), 1, rtol=epsilon, atol=epsilon)):
+ raise ValueError("Predicted probabilities must sum to approximately 1.")
+
+ # Clip predicted probabilities to avoid log(0)
+ y_pred = np.clip(y_pred, epsilon, 1 - epsilon)
+
+ # Calculate loss for each class and sum across classes
+ cfce_loss = -np.sum(
+ alpha * np.power(1 - y_pred, gamma) * y_true * np.log(y_pred), axis=1
+ )
+
+ return np.mean(cfce_loss)
+
+
+def hinge_loss(y_true: np.ndarray, y_pred: np.ndarray) -> float:
+ """
+ Calculate the mean hinge loss for between true labels and predicted probabilities
+ for training support vector machines (SVMs).
+
+ Hinge loss = max(0, 1 - true * pred)
+
+ Reference: https://en.wikipedia.org/wiki/Hinge_loss
+
+ Args:
+ - y_true: actual values (ground truth) encoded as -1 or 1
+ - y_pred: predicted values
+
+ >>> true_labels = np.array([-1, 1, 1, -1, 1])
+ >>> pred = np.array([-4, -0.3, 0.7, 5, 10])
+ >>> float(hinge_loss(true_labels, pred))
+ 1.52
+ >>> true_labels = np.array([-1, 1, 1, -1, 1, 1])
+ >>> pred = np.array([-4, -0.3, 0.7, 5, 10])
+ >>> hinge_loss(true_labels, pred)
+ Traceback (most recent call last):
+ ...
+ ValueError: Length of predicted and actual array must be same.
+ >>> true_labels = np.array([-1, 1, 10, -1, 1])
+ >>> pred = np.array([-4, -0.3, 0.7, 5, 10])
+ >>> hinge_loss(true_labels, pred)
+ Traceback (most recent call last):
+ ...
+ ValueError: y_true can have values -1 or 1 only.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Length of predicted and actual array must be same.")
+
+ if np.any((y_true != -1) & (y_true != 1)):
+ raise ValueError("y_true can have values -1 or 1 only.")
+
+ hinge_losses = np.maximum(0, 1.0 - (y_true * y_pred))
+ return np.mean(hinge_losses)
+
+
+def huber_loss(y_true: np.ndarray, y_pred: np.ndarray, delta: float) -> float:
+ """
+ Calculate the mean Huber loss between the given ground truth and predicted values.
+
+ The Huber loss describes the penalty incurred by an estimation procedure, and it
+ serves as a measure of accuracy for regression models.
+
+ Huber loss =
+ 0.5 * (y_true - y_pred)^2 if |y_true - y_pred| <= delta
+ delta * |y_true - y_pred| - 0.5 * delta^2 otherwise
+
+ Reference: https://en.wikipedia.org/wiki/Huber_loss
+
+ Parameters:
+ - y_true: The true values (ground truth)
+ - y_pred: The predicted values
+
+ >>> true_values = np.array([0.9, 10.0, 2.0, 1.0, 5.2])
+ >>> predicted_values = np.array([0.8, 2.1, 2.9, 4.2, 5.2])
+ >>> bool(np.isclose(huber_loss(true_values, predicted_values, 1.0), 2.102))
+ True
+ >>> true_labels = np.array([11.0, 21.0, 3.32, 4.0, 5.0])
+ >>> predicted_probs = np.array([8.3, 20.8, 2.9, 11.2, 5.0])
+ >>> bool(np.isclose(huber_loss(true_labels, predicted_probs, 1.0), 1.80164))
+ True
+ >>> true_labels = np.array([11.0, 21.0, 3.32, 4.0])
+ >>> predicted_probs = np.array([8.3, 20.8, 2.9, 11.2, 5.0])
+ >>> huber_loss(true_labels, predicted_probs, 1.0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+
+ huber_mse = 0.5 * (y_true - y_pred) ** 2
+ huber_mae = delta * (np.abs(y_true - y_pred) - 0.5 * delta)
+ return np.where(np.abs(y_true - y_pred) <= delta, huber_mse, huber_mae).mean()
+
+
+def mean_squared_error(y_true: np.ndarray, y_pred: np.ndarray) -> float:
+ """
+ Calculate the mean squared error (MSE) between ground truth and predicted values.
+
+ MSE measures the squared difference between true values and predicted values, and it
+ serves as a measure of accuracy for regression models.
+
+ MSE = (1/n) * Σ(y_true - y_pred)^2
+
+ Reference: https://en.wikipedia.org/wiki/Mean_squared_error
+
+ Parameters:
+ - y_true: The true values (ground truth)
+ - y_pred: The predicted values
+
+ >>> true_values = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_values = np.array([0.8, 2.1, 2.9, 4.2, 5.2])
+ >>> bool(np.isclose(mean_squared_error(true_values, predicted_values), 0.028))
+ True
+ >>> true_labels = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_probs = np.array([0.3, 0.8, 0.9, 0.2])
+ >>> mean_squared_error(true_labels, predicted_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+
+ squared_errors = (y_true - y_pred) ** 2
+ return np.mean(squared_errors)
+
+
+def mean_absolute_error(y_true: np.ndarray, y_pred: np.ndarray) -> float:
+ """
+ Calculates the Mean Absolute Error (MAE) between ground truth (observed)
+ and predicted values.
+
+ MAE measures the absolute difference between true values and predicted values.
+
+ Equation:
+ MAE = (1/n) * Σ(abs(y_true - y_pred))
+
+ Reference: https://en.wikipedia.org/wiki/Mean_absolute_error
+
+ Parameters:
+ - y_true: The true values (ground truth)
+ - y_pred: The predicted values
+
+ >>> true_values = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_values = np.array([0.8, 2.1, 2.9, 4.2, 5.2])
+ >>> bool(np.isclose(mean_absolute_error(true_values, predicted_values), 0.16))
+ True
+ >>> true_values = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_values = np.array([0.8, 2.1, 2.9, 4.2, 5.2])
+ >>> bool(np.isclose(mean_absolute_error(true_values, predicted_values), 2.16))
+ False
+ >>> true_labels = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_probs = np.array([0.3, 0.8, 0.9, 5.2])
+ >>> mean_absolute_error(true_labels, predicted_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+
+ return np.mean(abs(y_true - y_pred))
+
+
+def mean_squared_logarithmic_error(y_true: np.ndarray, y_pred: np.ndarray) -> float:
+ """
+ Calculate the mean squared logarithmic error (MSLE) between ground truth and
+ predicted values.
+
+ MSLE measures the squared logarithmic difference between true values and predicted
+ values for regression models. It's particularly useful for dealing with skewed or
+ large-value data, and it's often used when the relative differences between
+ predicted and true values are more important than absolute differences.
+
+ MSLE = (1/n) * Σ(log(1 + y_true) - log(1 + y_pred))^2
+
+ Reference: https://insideaiml.com/blog/MeanSquared-Logarithmic-Error-Loss-1035
+
+ Parameters:
+ - y_true: The true values (ground truth)
+ - y_pred: The predicted values
+
+ >>> true_values = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_values = np.array([0.8, 2.1, 2.9, 4.2, 5.2])
+ >>> float(mean_squared_logarithmic_error(true_values, predicted_values))
+ 0.0030860877925181344
+ >>> true_labels = np.array([1.0, 2.0, 3.0, 4.0, 5.0])
+ >>> predicted_probs = np.array([0.3, 0.8, 0.9, 0.2])
+ >>> mean_squared_logarithmic_error(true_labels, predicted_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+
+ squared_logarithmic_errors = (np.log1p(y_true) - np.log1p(y_pred)) ** 2
+ return np.mean(squared_logarithmic_errors)
+
+
+def mean_absolute_percentage_error(
+ y_true: np.ndarray, y_pred: np.ndarray, epsilon: float = 1e-15
+) -> float:
+ """
+ Calculate the Mean Absolute Percentage Error between y_true and y_pred.
+
+ Mean Absolute Percentage Error calculates the average of the absolute
+ percentage differences between the predicted and true values.
+
+ Formula = (Σ|y_true[i]-Y_pred[i]/y_true[i]|)/n
+
+ Source: https://stephenallwright.com/good-mape-score/
+
+ Parameters:
+ y_true (np.ndarray): Numpy array containing true/target values.
+ y_pred (np.ndarray): Numpy array containing predicted values.
+
+ Returns:
+ float: The Mean Absolute Percentage error between y_true and y_pred.
+
+ Examples:
+ >>> y_true = np.array([10, 20, 30, 40])
+ >>> y_pred = np.array([12, 18, 33, 45])
+ >>> float(mean_absolute_percentage_error(y_true, y_pred))
+ 0.13125
+
+ >>> y_true = np.array([1, 2, 3, 4])
+ >>> y_pred = np.array([2, 3, 4, 5])
+ >>> float(mean_absolute_percentage_error(y_true, y_pred))
+ 0.5208333333333333
+
+ >>> y_true = np.array([34, 37, 44, 47, 48, 48, 46, 43, 32, 27, 26, 24])
+ >>> y_pred = np.array([37, 40, 46, 44, 46, 50, 45, 44, 34, 30, 22, 23])
+ >>> float(mean_absolute_percentage_error(y_true, y_pred))
+ 0.064671076436071
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("The length of the two arrays should be the same.")
+
+ y_true = np.where(y_true == 0, epsilon, y_true)
+ absolute_percentage_diff = np.abs((y_true - y_pred) / y_true)
+
+ return np.mean(absolute_percentage_diff)
+
+
+def perplexity_loss(
+ y_true: np.ndarray, y_pred: np.ndarray, epsilon: float = 1e-7
+) -> float:
+ """
+ Calculate the perplexity for the y_true and y_pred.
+
+ Compute the Perplexity which useful in predicting language model
+ accuracy in Natural Language Processing (NLP.)
+ Perplexity is measure of how certain the model in its predictions.
+
+ Perplexity Loss = exp(-1/N (Σ ln(p(x)))
+
+ Reference:
+ https://en.wikipedia.org/wiki/Perplexity
+
+ Args:
+ y_true: Actual label encoded sentences of shape (batch_size, sentence_length)
+ y_pred: Predicted sentences of shape (batch_size, sentence_length, vocab_size)
+ epsilon: Small floating point number to avoid getting inf for log(0)
+
+ Returns:
+ Perplexity loss between y_true and y_pred.
+
+ >>> y_true = np.array([[1, 4], [2, 3]])
+ >>> y_pred = np.array(
+ ... [[[0.28, 0.19, 0.21 , 0.15, 0.15],
+ ... [0.24, 0.19, 0.09, 0.18, 0.27]],
+ ... [[0.03, 0.26, 0.21, 0.18, 0.30],
+ ... [0.28, 0.10, 0.33, 0.15, 0.12]]]
+ ... )
+ >>> float(perplexity_loss(y_true, y_pred))
+ 5.0247347775367945
+ >>> y_true = np.array([[1, 4], [2, 3]])
+ >>> y_pred = np.array(
+ ... [[[0.28, 0.19, 0.21 , 0.15, 0.15],
+ ... [0.24, 0.19, 0.09, 0.18, 0.27],
+ ... [0.30, 0.10, 0.20, 0.15, 0.25]],
+ ... [[0.03, 0.26, 0.21, 0.18, 0.30],
+ ... [0.28, 0.10, 0.33, 0.15, 0.12],
+ ... [0.30, 0.10, 0.20, 0.15, 0.25]],]
+ ... )
+ >>> perplexity_loss(y_true, y_pred)
+ Traceback (most recent call last):
+ ...
+ ValueError: Sentence length of y_true and y_pred must be equal.
+ >>> y_true = np.array([[1, 4], [2, 11]])
+ >>> y_pred = np.array(
+ ... [[[0.28, 0.19, 0.21 , 0.15, 0.15],
+ ... [0.24, 0.19, 0.09, 0.18, 0.27]],
+ ... [[0.03, 0.26, 0.21, 0.18, 0.30],
+ ... [0.28, 0.10, 0.33, 0.15, 0.12]]]
+ ... )
+ >>> perplexity_loss(y_true, y_pred)
+ Traceback (most recent call last):
+ ...
+ ValueError: Label value must not be greater than vocabulary size.
+ >>> y_true = np.array([[1, 4]])
+ >>> y_pred = np.array(
+ ... [[[0.28, 0.19, 0.21 , 0.15, 0.15],
+ ... [0.24, 0.19, 0.09, 0.18, 0.27]],
+ ... [[0.03, 0.26, 0.21, 0.18, 0.30],
+ ... [0.28, 0.10, 0.33, 0.15, 0.12]]]
+ ... )
+ >>> perplexity_loss(y_true, y_pred)
+ Traceback (most recent call last):
+ ...
+ ValueError: Batch size of y_true and y_pred must be equal.
+ """
+
+ vocab_size = y_pred.shape[2]
+
+ if y_true.shape[0] != y_pred.shape[0]:
+ raise ValueError("Batch size of y_true and y_pred must be equal.")
+ if y_true.shape[1] != y_pred.shape[1]:
+ raise ValueError("Sentence length of y_true and y_pred must be equal.")
+ if np.max(y_true) > vocab_size:
+ raise ValueError("Label value must not be greater than vocabulary size.")
+
+ # Matrix to select prediction value only for true class
+ filter_matrix = np.array(
+ [[list(np.eye(vocab_size)[word]) for word in sentence] for sentence in y_true]
+ )
+
+ # Getting the matrix containing prediction for only true class
+ true_class_pred = np.sum(y_pred * filter_matrix, axis=2).clip(epsilon, 1)
+
+ # Calculating perplexity for each sentence
+ perp_losses = np.exp(np.negative(np.mean(np.log(true_class_pred), axis=1)))
+
+ return np.mean(perp_losses)
+
+
+def smooth_l1_loss(y_true: np.ndarray, y_pred: np.ndarray, beta: float = 1.0) -> float:
+ """
+ Calculate the Smooth L1 Loss between y_true and y_pred.
+
+ The Smooth L1 Loss is less sensitive to outliers than the L2 Loss and is often used
+ in regression problems, such as object detection.
+
+ Smooth L1 Loss =
+ 0.5 * (x - y)^2 / beta, if |x - y| < beta
+ |x - y| - 0.5 * beta, otherwise
+
+ Reference:
+ https://pytorch.org/docs/stable/generated/torch.nn.SmoothL1Loss.html
+
+ Args:
+ y_true: Array of true values.
+ y_pred: Array of predicted values.
+ beta: Specifies the threshold at which to change between L1 and L2 loss.
+
+ Returns:
+ The calculated Smooth L1 Loss between y_true and y_pred.
+
+ Raises:
+ ValueError: If the length of the two arrays is not the same.
+
+ >>> y_true = np.array([3, 5, 2, 7])
+ >>> y_pred = np.array([2.9, 4.8, 2.1, 7.2])
+ >>> float(smooth_l1_loss(y_true, y_pred, 1.0))
+ 0.012500000000000022
+
+ >>> y_true = np.array([2, 4, 6])
+ >>> y_pred = np.array([1, 5, 7])
+ >>> float(smooth_l1_loss(y_true, y_pred, 1.0))
+ 0.5
+
+ >>> y_true = np.array([1, 3, 5, 7])
+ >>> y_pred = np.array([1, 3, 5, 7])
+ >>> float(smooth_l1_loss(y_true, y_pred, 1.0))
+ 0.0
+
+ >>> y_true = np.array([1, 3, 5])
+ >>> y_pred = np.array([1, 3, 5, 7])
+ >>> smooth_l1_loss(y_true, y_pred, 1.0)
+ Traceback (most recent call last):
+ ...
+ ValueError: The length of the two arrays should be the same.
+ """
+
+ if len(y_true) != len(y_pred):
+ raise ValueError("The length of the two arrays should be the same.")
+
+ diff = np.abs(y_true - y_pred)
+ loss = np.where(diff < beta, 0.5 * diff**2 / beta, diff - 0.5 * beta)
+ return np.mean(loss)
+
+
+def kullback_leibler_divergence(y_true: np.ndarray, y_pred: np.ndarray) -> float:
+ """
+ Calculate the Kullback-Leibler divergence (KL divergence) loss between true labels
+ and predicted probabilities.
+
+ KL divergence loss quantifies dissimilarity between true labels and predicted
+ probabilities. It's often used in training generative models.
+
+ KL = Σ(y_true * ln(y_true / y_pred))
+
+ Reference: https://en.wikipedia.org/wiki/Kullback%E2%80%93Leibler_divergence
+
+ Parameters:
+ - y_true: True class probabilities
+ - y_pred: Predicted class probabilities
+
+ >>> true_labels = np.array([0.2, 0.3, 0.5])
+ >>> predicted_probs = np.array([0.3, 0.3, 0.4])
+ >>> float(kullback_leibler_divergence(true_labels, predicted_probs))
+ 0.030478754035472025
+ >>> true_labels = np.array([0.2, 0.3, 0.5])
+ >>> predicted_probs = np.array([0.3, 0.3, 0.4, 0.5])
+ >>> kullback_leibler_divergence(true_labels, predicted_probs)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input arrays must have the same length.
+ """
+ if len(y_true) != len(y_pred):
+ raise ValueError("Input arrays must have the same length.")
+
+ kl_loss = y_true * np.log(y_true / y_pred)
+ return np.sum(kl_loss)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/machine_learning/lstm/lstm_prediction.py b/machine_learning/lstm/lstm_prediction.py
index 5452f0443f62..81ac5f01d3d6 100644
--- a/machine_learning/lstm/lstm_prediction.py
+++ b/machine_learning/lstm/lstm_prediction.py
@@ -1,9 +1,10 @@
"""
- Create a Long Short Term Memory (LSTM) network model
- An LSTM is a type of Recurrent Neural Network (RNN) as discussed at:
- * http://colah.github.io/posts/2015-08-Understanding-LSTMs
- * https://en.wikipedia.org/wiki/Long_short-term_memory
+Create a Long Short Term Memory (LSTM) network model
+An LSTM is a type of Recurrent Neural Network (RNN) as discussed at:
+* https://colah.github.io/posts/2015-08-Understanding-LSTMs
+* https://en.wikipedia.org/wiki/Long_short-term_memory
"""
+
import numpy as np
import pandas as pd
from keras.layers import LSTM, Dense
@@ -17,11 +18,11 @@
make sure you set the price column on line number 21. Here we
use a dataset which have the price on 3rd column.
"""
- df = pd.read_csv("sample_data.csv", header=None)
- len_data = df.shape[:1][0]
+ sample_data = pd.read_csv("sample_data.csv", header=None)
+ len_data = sample_data.shape[:1][0]
# If you're using some other dataset input the target column
- actual_data = df.iloc[:, 1:2]
- actual_data = actual_data.values.reshape(len_data, 1)
+ actual_data = sample_data.iloc[:, 1:2]
+ actual_data = actual_data.to_numpy().reshape(len_data, 1)
actual_data = MinMaxScaler().fit_transform(actual_data)
look_back = 10
forward_days = 5
@@ -32,10 +33,10 @@
train_x, train_y = [], []
test_x, test_y = [], []
- for i in range(0, len(train_data) - forward_days - look_back + 1):
+ for i in range(len(train_data) - forward_days - look_back + 1):
train_x.append(train_data[i : i + look_back])
train_y.append(train_data[i + look_back : i + look_back + forward_days])
- for i in range(0, len(test_data) - forward_days - look_back + 1):
+ for i in range(len(test_data) - forward_days - look_back + 1):
test_x.append(test_data[i : i + look_back])
test_y.append(test_data[i + look_back : i + look_back + forward_days])
x_train = np.array(train_x)
diff --git a/machine_learning/mfcc.py b/machine_learning/mfcc.py
new file mode 100644
index 000000000000..dcc3151d5a1a
--- /dev/null
+++ b/machine_learning/mfcc.py
@@ -0,0 +1,479 @@
+"""
+Mel Frequency Cepstral Coefficients (MFCC) Calculation
+
+MFCC is an algorithm widely used in audio and speech processing to represent the
+short-term power spectrum of a sound signal in a more compact and
+discriminative way. It is particularly popular in speech and audio processing
+tasks such as speech recognition and speaker identification.
+
+How Mel Frequency Cepstral Coefficients are Calculated:
+1. Preprocessing:
+ - Load an audio signal and normalize it to ensure that the values fall
+ within a specific range (e.g., between -1 and 1).
+ - Frame the audio signal into overlapping, fixed-length segments, typically
+ using a technique like windowing to reduce spectral leakage.
+
+2. Fourier Transform:
+ - Apply a Fast Fourier Transform (FFT) to each audio frame to convert it
+ from the time domain to the frequency domain. This results in a
+ representation of the audio frame as a sequence of frequency components.
+
+3. Power Spectrum:
+ - Calculate the power spectrum by taking the squared magnitude of each
+ frequency component obtained from the FFT. This step measures the energy
+ distribution across different frequency bands.
+
+4. Mel Filterbank:
+ - Apply a set of triangular filterbanks spaced in the Mel frequency scale
+ to the power spectrum. These filters mimic the human auditory system's
+ frequency response. Each filterbank sums the power spectrum values within
+ its band.
+
+5. Logarithmic Compression:
+ - Take the logarithm (typically base 10) of the filterbank values to
+ compress the dynamic range. This step mimics the logarithmic response of
+ the human ear to sound intensity.
+
+6. Discrete Cosine Transform (DCT):
+ - Apply the Discrete Cosine Transform to the log filterbank energies to
+ obtain the MFCC coefficients. This transformation helps decorrelate the
+ filterbank energies and captures the most important features of the audio
+ signal.
+
+7. Feature Extraction:
+ - Select a subset of the DCT coefficients to form the feature vector.
+ Often, the first few coefficients (e.g., 12-13) are used for most
+ applications.
+
+References:
+- Mel-Frequency Cepstral Coefficients (MFCCs):
+ https://en.wikipedia.org/wiki/Mel-frequency_cepstrum
+- Speech and Language Processing by Daniel Jurafsky & James H. Martin:
+ https://web.stanford.edu/~jurafsky/slp3/
+- Mel Frequency Cepstral Coefficient (MFCC) tutorial
+ http://practicalcryptography.com/miscellaneous/machine-learning
+ /guide-mel-frequency-cepstral-coefficients-mfccs/
+
+Author: Amir Lavasani
+"""
+
+import logging
+
+import numpy as np
+import scipy.fftpack as fft
+from scipy.signal import get_window
+
+logging.basicConfig(filename=f"{__file__}.log", level=logging.INFO)
+
+
+def mfcc(
+ audio: np.ndarray,
+ sample_rate: int,
+ ftt_size: int = 1024,
+ hop_length: int = 20,
+ mel_filter_num: int = 10,
+ dct_filter_num: int = 40,
+) -> np.ndarray:
+ """
+ Calculate Mel Frequency Cepstral Coefficients (MFCCs) from an audio signal.
+
+ Args:
+ audio: The input audio signal.
+ sample_rate: The sample rate of the audio signal (in Hz).
+ ftt_size: The size of the FFT window (default is 1024).
+ hop_length: The hop length for frame creation (default is 20ms).
+ mel_filter_num: The number of Mel filters (default is 10).
+ dct_filter_num: The number of DCT filters (default is 40).
+
+ Returns:
+ A matrix of MFCCs for the input audio.
+
+ Raises:
+ ValueError: If the input audio is empty.
+
+ Example:
+ >>> sample_rate = 44100 # Sample rate of 44.1 kHz
+ >>> duration = 2.0 # Duration of 1 second
+ >>> t = np.linspace(0, duration, int(sample_rate * duration), endpoint=False)
+ >>> audio = 0.5 * np.sin(2 * np.pi * 440.0 * t) # Generate a 440 Hz sine wave
+ >>> mfccs = mfcc(audio, sample_rate)
+ >>> mfccs.shape
+ (40, 101)
+ """
+ logging.info(f"Sample rate: {sample_rate}Hz")
+ logging.info(f"Audio duration: {len(audio) / sample_rate}s")
+ logging.info(f"Audio min: {np.min(audio)}")
+ logging.info(f"Audio max: {np.max(audio)}")
+
+ # normalize audio
+ audio_normalized = normalize(audio)
+
+ logging.info(f"Normalized audio min: {np.min(audio_normalized)}")
+ logging.info(f"Normalized audio max: {np.max(audio_normalized)}")
+
+ # frame audio into
+ audio_framed = audio_frames(
+ audio_normalized, sample_rate, ftt_size=ftt_size, hop_length=hop_length
+ )
+
+ logging.info(f"Framed audio shape: {audio_framed.shape}")
+ logging.info(f"First frame: {audio_framed[0]}")
+
+ # convert to frequency domain
+ # For simplicity we will choose the Hanning window.
+ window = get_window("hann", ftt_size, fftbins=True)
+ audio_windowed = audio_framed * window
+
+ logging.info(f"Windowed audio shape: {audio_windowed.shape}")
+ logging.info(f"First frame: {audio_windowed[0]}")
+
+ audio_fft = calculate_fft(audio_windowed, ftt_size)
+ logging.info(f"fft audio shape: {audio_fft.shape}")
+ logging.info(f"First frame: {audio_fft[0]}")
+
+ audio_power = calculate_signal_power(audio_fft)
+ logging.info(f"power audio shape: {audio_power.shape}")
+ logging.info(f"First frame: {audio_power[0]}")
+
+ filters = mel_spaced_filterbank(sample_rate, mel_filter_num, ftt_size)
+ logging.info(f"filters shape: {filters.shape}")
+
+ audio_filtered = np.dot(filters, np.transpose(audio_power))
+ audio_log = 10.0 * np.log10(audio_filtered)
+ logging.info(f"audio_log shape: {audio_log.shape}")
+
+ dct_filters = discrete_cosine_transform(dct_filter_num, mel_filter_num)
+ cepstral_coefficents = np.dot(dct_filters, audio_log)
+
+ logging.info(f"cepstral_coefficents shape: {cepstral_coefficents.shape}")
+ return cepstral_coefficents
+
+
+def normalize(audio: np.ndarray) -> np.ndarray:
+ """
+ Normalize an audio signal by scaling it to have values between -1 and 1.
+
+ Args:
+ audio: The input audio signal.
+
+ Returns:
+ The normalized audio signal.
+
+ Examples:
+ >>> audio = np.array([1, 2, 3, 4, 5])
+ >>> normalized_audio = normalize(audio)
+ >>> float(np.max(normalized_audio))
+ 1.0
+ >>> float(np.min(normalized_audio))
+ 0.2
+ """
+ # Divide the entire audio signal by the maximum absolute value
+ return audio / np.max(np.abs(audio))
+
+
+def audio_frames(
+ audio: np.ndarray,
+ sample_rate: int,
+ hop_length: int = 20,
+ ftt_size: int = 1024,
+) -> np.ndarray:
+ """
+ Split an audio signal into overlapping frames.
+
+ Args:
+ audio: The input audio signal.
+ sample_rate: The sample rate of the audio signal.
+ hop_length: The length of the hopping (default is 20ms).
+ ftt_size: The size of the FFT window (default is 1024).
+
+ Returns:
+ An array of overlapping frames.
+
+ Examples:
+ >>> audio = np.array([1, 2, 3, 4, 5, 6, 7, 8, 9, 10]*1000)
+ >>> sample_rate = 8000
+ >>> frames = audio_frames(audio, sample_rate, hop_length=10, ftt_size=512)
+ >>> frames.shape
+ (126, 512)
+ """
+
+ hop_size = np.round(sample_rate * hop_length / 1000).astype(int)
+
+ # Pad the audio signal to handle edge cases
+ audio = np.pad(audio, int(ftt_size / 2), mode="reflect")
+
+ # Calculate the number of frames
+ frame_count = int((len(audio) - ftt_size) / hop_size) + 1
+
+ # Initialize an array to store the frames
+ frames = np.zeros((frame_count, ftt_size))
+
+ # Split the audio signal into frames
+ for n in range(frame_count):
+ frames[n] = audio[n * hop_size : n * hop_size + ftt_size]
+
+ return frames
+
+
+def calculate_fft(audio_windowed: np.ndarray, ftt_size: int = 1024) -> np.ndarray:
+ """
+ Calculate the Fast Fourier Transform (FFT) of windowed audio data.
+
+ Args:
+ audio_windowed: The windowed audio signal.
+ ftt_size: The size of the FFT (default is 1024).
+
+ Returns:
+ The FFT of the audio data.
+
+ Examples:
+ >>> audio_windowed = np.array([[1.0, 2.0, 3.0], [4.0, 5.0, 6.0]])
+ >>> audio_fft = calculate_fft(audio_windowed, ftt_size=4)
+ >>> bool(np.allclose(audio_fft[0], np.array([6.0+0.j, -1.5+0.8660254j,
+ ... -1.5-0.8660254j])))
+ True
+ """
+ # Transpose the audio data to have time in rows and channels in columns
+ audio_transposed = np.transpose(audio_windowed)
+
+ # Initialize an array to store the FFT results
+ audio_fft = np.empty(
+ (int(1 + ftt_size // 2), audio_transposed.shape[1]),
+ dtype=np.complex64,
+ order="F",
+ )
+
+ # Compute FFT for each channel
+ for n in range(audio_fft.shape[1]):
+ audio_fft[:, n] = fft.fft(audio_transposed[:, n], axis=0)[: audio_fft.shape[0]]
+
+ # Transpose the FFT results back to the original shape
+ return np.transpose(audio_fft)
+
+
+def calculate_signal_power(audio_fft: np.ndarray) -> np.ndarray:
+ """
+ Calculate the power of the audio signal from its FFT.
+
+ Args:
+ audio_fft: The FFT of the audio signal.
+
+ Returns:
+ The power of the audio signal.
+
+ Examples:
+ >>> audio_fft = np.array([1+2j, 2+3j, 3+4j, 4+5j])
+ >>> power = calculate_signal_power(audio_fft)
+ >>> np.allclose(power, np.array([5, 13, 25, 41]))
+ True
+ """
+ # Calculate the power by squaring the absolute values of the FFT coefficients
+ return np.square(np.abs(audio_fft))
+
+
+def freq_to_mel(freq: float) -> float:
+ """
+ Convert a frequency in Hertz to the mel scale.
+
+ Args:
+ freq: The frequency in Hertz.
+
+ Returns:
+ The frequency in mel scale.
+
+ Examples:
+ >>> float(round(freq_to_mel(1000), 2))
+ 999.99
+ """
+ # Use the formula to convert frequency to the mel scale
+ return 2595.0 * np.log10(1.0 + freq / 700.0)
+
+
+def mel_to_freq(mels: float) -> float:
+ """
+ Convert a frequency in the mel scale to Hertz.
+
+ Args:
+ mels: The frequency in mel scale.
+
+ Returns:
+ The frequency in Hertz.
+
+ Examples:
+ >>> round(mel_to_freq(999.99), 2)
+ 1000.01
+ """
+ # Use the formula to convert mel scale to frequency
+ return 700.0 * (10.0 ** (mels / 2595.0) - 1.0)
+
+
+def mel_spaced_filterbank(
+ sample_rate: int, mel_filter_num: int = 10, ftt_size: int = 1024
+) -> np.ndarray:
+ """
+ Create a Mel-spaced filter bank for audio processing.
+
+ Args:
+ sample_rate: The sample rate of the audio.
+ mel_filter_num: The number of mel filters (default is 10).
+ ftt_size: The size of the FFT (default is 1024).
+
+ Returns:
+ Mel-spaced filter bank.
+
+ Examples:
+ >>> float(round(mel_spaced_filterbank(8000, 10, 1024)[0][1], 10))
+ 0.0004603981
+ """
+ freq_min = 0
+ freq_high = sample_rate // 2
+
+ logging.info(f"Minimum frequency: {freq_min}")
+ logging.info(f"Maximum frequency: {freq_high}")
+
+ # Calculate filter points and mel frequencies
+ filter_points, mel_freqs = get_filter_points(
+ sample_rate,
+ freq_min,
+ freq_high,
+ mel_filter_num,
+ ftt_size,
+ )
+
+ filters = get_filters(filter_points, ftt_size)
+
+ # normalize filters
+ # taken from the librosa library
+ enorm = 2.0 / (mel_freqs[2 : mel_filter_num + 2] - mel_freqs[:mel_filter_num])
+ return filters * enorm[:, np.newaxis]
+
+
+def get_filters(filter_points: np.ndarray, ftt_size: int) -> np.ndarray:
+ """
+ Generate filters for audio processing.
+
+ Args:
+ filter_points: A list of filter points.
+ ftt_size: The size of the FFT.
+
+ Returns:
+ A matrix of filters.
+
+ Examples:
+ >>> get_filters(np.array([0, 20, 51, 95, 161, 256], dtype=int), 512).shape
+ (4, 257)
+ """
+ num_filters = len(filter_points) - 2
+ filters = np.zeros((num_filters, int(ftt_size / 2) + 1))
+
+ for n in range(num_filters):
+ start = filter_points[n]
+ mid = filter_points[n + 1]
+ end = filter_points[n + 2]
+
+ # Linearly increase values from 0 to 1
+ filters[n, start:mid] = np.linspace(0, 1, mid - start)
+
+ # Linearly decrease values from 1 to 0
+ filters[n, mid:end] = np.linspace(1, 0, end - mid)
+
+ return filters
+
+
+def get_filter_points(
+ sample_rate: int,
+ freq_min: int,
+ freq_high: int,
+ mel_filter_num: int = 10,
+ ftt_size: int = 1024,
+) -> tuple[np.ndarray, np.ndarray]:
+ """
+ Calculate the filter points and frequencies for mel frequency filters.
+
+ Args:
+ sample_rate: The sample rate of the audio.
+ freq_min: The minimum frequency in Hertz.
+ freq_high: The maximum frequency in Hertz.
+ mel_filter_num: The number of mel filters (default is 10).
+ ftt_size: The size of the FFT (default is 1024).
+
+ Returns:
+ Filter points and corresponding frequencies.
+
+ Examples:
+ >>> filter_points = get_filter_points(8000, 0, 4000, mel_filter_num=4, ftt_size=512)
+ >>> filter_points[0]
+ array([ 0, 20, 51, 95, 161, 256])
+ >>> filter_points[1]
+ array([ 0. , 324.46707094, 799.33254207, 1494.30973963,
+ 2511.42581671, 4000. ])
+ """
+ # Convert minimum and maximum frequencies to mel scale
+ fmin_mel = freq_to_mel(freq_min)
+ fmax_mel = freq_to_mel(freq_high)
+
+ logging.info(f"MEL min: {fmin_mel}")
+ logging.info(f"MEL max: {fmax_mel}")
+
+ # Generate equally spaced mel frequencies
+ mels = np.linspace(fmin_mel, fmax_mel, num=mel_filter_num + 2)
+
+ # Convert mel frequencies back to Hertz
+ freqs = mel_to_freq(mels)
+
+ # Calculate filter points as integer values
+ filter_points = np.floor((ftt_size + 1) / sample_rate * freqs).astype(int)
+
+ return filter_points, freqs
+
+
+def discrete_cosine_transform(dct_filter_num: int, filter_num: int) -> np.ndarray:
+ """
+ Compute the Discrete Cosine Transform (DCT) basis matrix.
+
+ Args:
+ dct_filter_num: The number of DCT filters to generate.
+ filter_num: The number of the fbank filters.
+
+ Returns:
+ The DCT basis matrix.
+
+ Examples:
+ >>> float(round(discrete_cosine_transform(3, 5)[0][0], 5))
+ 0.44721
+ """
+ basis = np.empty((dct_filter_num, filter_num))
+ basis[0, :] = 1.0 / np.sqrt(filter_num)
+
+ samples = np.arange(1, 2 * filter_num, 2) * np.pi / (2.0 * filter_num)
+
+ for i in range(1, dct_filter_num):
+ basis[i, :] = np.cos(i * samples) * np.sqrt(2.0 / filter_num)
+
+ return basis
+
+
+def example(wav_file_path: str = "./path-to-file/sample.wav") -> np.ndarray:
+ """
+ Example function to calculate Mel Frequency Cepstral Coefficients
+ (MFCCs) from an audio file.
+
+ Args:
+ wav_file_path: The path to the WAV audio file.
+
+ Returns:
+ np.ndarray: The computed MFCCs for the audio.
+ """
+ from scipy.io import wavfile
+
+ # Load the audio from the WAV file
+ sample_rate, audio = wavfile.read(wav_file_path)
+
+ # Calculate MFCCs
+ return mfcc(audio, sample_rate)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/machine_learning/multilayer_perceptron_classifier.py b/machine_learning/multilayer_perceptron_classifier.py
index 604185cef677..40f998c7dfa2 100644
--- a/machine_learning/multilayer_perceptron_classifier.py
+++ b/machine_learning/multilayer_perceptron_classifier.py
@@ -15,12 +15,12 @@
Y = clf.predict(test)
-def wrapper(Y):
+def wrapper(y):
"""
- >>> wrapper(Y)
+ >>> [int(x) for x in wrapper(Y)]
[0, 0, 1]
"""
- return list(Y)
+ return list(y)
if __name__ == "__main__":
diff --git a/machine_learning/polymonial_regression.py b/machine_learning/polymonial_regression.py
deleted file mode 100644
index 374c35f7f905..000000000000
--- a/machine_learning/polymonial_regression.py
+++ /dev/null
@@ -1,45 +0,0 @@
-import pandas as pd
-from matplotlib import pyplot as plt
-from sklearn.linear_model import LinearRegression
-
-# Splitting the dataset into the Training set and Test set
-from sklearn.model_selection import train_test_split
-
-# Fitting Polynomial Regression to the dataset
-from sklearn.preprocessing import PolynomialFeatures
-
-# Importing the dataset
-dataset = pd.read_csv(
- "https://s3.us-west-2.amazonaws.com/public.gamelab.fun/dataset/"
- "position_salaries.csv"
-)
-X = dataset.iloc[:, 1:2].values
-y = dataset.iloc[:, 2].values
-
-
-X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=0)
-
-
-poly_reg = PolynomialFeatures(degree=4)
-X_poly = poly_reg.fit_transform(X)
-pol_reg = LinearRegression()
-pol_reg.fit(X_poly, y)
-
-
-# Visualizing the Polymonial Regression results
-def viz_polymonial():
- plt.scatter(X, y, color="red")
- plt.plot(X, pol_reg.predict(poly_reg.fit_transform(X)), color="blue")
- plt.title("Truth or Bluff (Linear Regression)")
- plt.xlabel("Position level")
- plt.ylabel("Salary")
- plt.show()
- return
-
-
-if __name__ == "__main__":
- viz_polymonial()
-
- # Predicting a new result with Polymonial Regression
- pol_reg.predict(poly_reg.fit_transform([[5.5]]))
- # output should be 132148.43750003
diff --git a/machine_learning/polynomial_regression.py b/machine_learning/polynomial_regression.py
new file mode 100644
index 000000000000..212f40bea197
--- /dev/null
+++ b/machine_learning/polynomial_regression.py
@@ -0,0 +1,213 @@
+"""
+Polynomial regression is a type of regression analysis that models the relationship
+between a predictor x and the response y as an mth-degree polynomial:
+
+y = β₀ + β₁x + β₂x² + ... + βₘxᵐ + ε
+
+By treating x, x², ..., xᵐ as distinct variables, we see that polynomial regression is a
+special case of multiple linear regression. Therefore, we can use ordinary least squares
+(OLS) estimation to estimate the vector of model parameters β = (β₀, β₁, β₂, ..., βₘ)
+for polynomial regression:
+
+β = (XᵀX)⁻¹Xᵀy = X⁺y
+
+where X is the design matrix, y is the response vector, and X⁺ denotes the Moore-Penrose
+pseudoinverse of X. In the case of polynomial regression, the design matrix is
+
+ |1 x₁ x₁² ⋯ x₁ᵐ|
+X = |1 x₂ x₂² ⋯ x₂ᵐ|
+ |⋮ ⋮ ⋮ ⋱ ⋮ |
+ |1 xₙ xₙ² ⋯ xₙᵐ|
+
+In OLS estimation, inverting XᵀX to compute X⁺ can be very numerically unstable. This
+implementation sidesteps this need to invert XᵀX by computing X⁺ using singular value
+decomposition (SVD):
+
+β = VΣ⁺Uᵀy
+
+where UΣVᵀ is an SVD of X.
+
+References:
+ - https://en.wikipedia.org/wiki/Polynomial_regression
+ - https://en.wikipedia.org/wiki/Moore%E2%80%93Penrose_inverse
+ - https://en.wikipedia.org/wiki/Numerical_methods_for_linear_least_squares
+ - https://en.wikipedia.org/wiki/Singular_value_decomposition
+"""
+
+import matplotlib.pyplot as plt
+import numpy as np
+
+
+class PolynomialRegression:
+ __slots__ = "degree", "params"
+
+ def __init__(self, degree: int) -> None:
+ """
+ @raises ValueError: if the polynomial degree is negative
+ """
+ if degree < 0:
+ raise ValueError("Polynomial degree must be non-negative")
+
+ self.degree = degree
+ self.params = None
+
+ @staticmethod
+ def _design_matrix(data: np.ndarray, degree: int) -> np.ndarray:
+ """
+ Constructs a polynomial regression design matrix for the given input data. For
+ input data x = (x₁, x₂, ..., xₙ) and polynomial degree m, the design matrix is
+ the Vandermonde matrix
+
+ |1 x₁ x₁² ⋯ x₁ᵐ|
+ X = |1 x₂ x₂² ⋯ x₂ᵐ|
+ |⋮ ⋮ ⋮ ⋱ ⋮ |
+ |1 xₙ xₙ² ⋯ xₙᵐ|
+
+ Reference: https://en.wikipedia.org/wiki/Vandermonde_matrix
+
+ @param data: the input predictor values x, either for model fitting or for
+ prediction
+ @param degree: the polynomial degree m
+ @returns: the Vandermonde matrix X (see above)
+ @raises ValueError: if input data is not N x 1
+
+ >>> x = np.array([0, 1, 2])
+ >>> PolynomialRegression._design_matrix(x, degree=0)
+ array([[1],
+ [1],
+ [1]])
+ >>> PolynomialRegression._design_matrix(x, degree=1)
+ array([[1, 0],
+ [1, 1],
+ [1, 2]])
+ >>> PolynomialRegression._design_matrix(x, degree=2)
+ array([[1, 0, 0],
+ [1, 1, 1],
+ [1, 2, 4]])
+ >>> PolynomialRegression._design_matrix(x, degree=3)
+ array([[1, 0, 0, 0],
+ [1, 1, 1, 1],
+ [1, 2, 4, 8]])
+ >>> PolynomialRegression._design_matrix(np.array([[0, 0], [0 , 0]]), degree=3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Data must have dimensions N x 1
+ """
+ rows, *remaining = data.shape
+ if remaining:
+ raise ValueError("Data must have dimensions N x 1")
+
+ return np.vander(data, N=degree + 1, increasing=True)
+
+ def fit(self, x_train: np.ndarray, y_train: np.ndarray) -> None:
+ """
+ Computes the polynomial regression model parameters using ordinary least squares
+ (OLS) estimation:
+
+ β = (XᵀX)⁻¹Xᵀy = X⁺y
+
+ where X⁺ denotes the Moore-Penrose pseudoinverse of the design matrix X. This
+ function computes X⁺ using singular value decomposition (SVD).
+
+ References:
+ - https://en.wikipedia.org/wiki/Moore%E2%80%93Penrose_inverse
+ - https://en.wikipedia.org/wiki/Singular_value_decomposition
+ - https://en.wikipedia.org/wiki/Multicollinearity
+
+ @param x_train: the predictor values x for model fitting
+ @param y_train: the response values y for model fitting
+ @raises ArithmeticError: if X isn't full rank, then XᵀX is singular and β
+ doesn't exist
+
+ >>> x = np.array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10])
+ >>> y = x**3 - 2 * x**2 + 3 * x - 5
+ >>> poly_reg = PolynomialRegression(degree=3)
+ >>> poly_reg.fit(x, y)
+ >>> poly_reg.params
+ array([-5., 3., -2., 1.])
+ >>> poly_reg = PolynomialRegression(degree=20)
+ >>> poly_reg.fit(x, y)
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: Design matrix is not full rank, can't compute coefficients
+
+ Make sure errors don't grow too large:
+ >>> coefs = np.array([-250, 50, -2, 36, 20, -12, 10, 2, -1, -15, 1])
+ >>> y = PolynomialRegression._design_matrix(x, len(coefs) - 1) @ coefs
+ >>> poly_reg = PolynomialRegression(degree=len(coefs) - 1)
+ >>> poly_reg.fit(x, y)
+ >>> np.allclose(poly_reg.params, coefs, atol=10e-3)
+ True
+ """
+ X = PolynomialRegression._design_matrix(x_train, self.degree) # noqa: N806
+ _, cols = X.shape
+ if np.linalg.matrix_rank(X) < cols:
+ raise ArithmeticError(
+ "Design matrix is not full rank, can't compute coefficients"
+ )
+
+ # np.linalg.pinv() computes the Moore-Penrose pseudoinverse using SVD
+ self.params = np.linalg.pinv(X) @ y_train
+
+ def predict(self, data: np.ndarray) -> np.ndarray:
+ """
+ Computes the predicted response values y for the given input data by
+ constructing the design matrix X and evaluating y = Xβ.
+
+ @param data: the predictor values x for prediction
+ @returns: the predicted response values y = Xβ
+ @raises ArithmeticError: if this function is called before the model
+ parameters are fit
+
+ >>> x = np.array([0, 1, 2, 3, 4])
+ >>> y = x**3 - 2 * x**2 + 3 * x - 5
+ >>> poly_reg = PolynomialRegression(degree=3)
+ >>> poly_reg.fit(x, y)
+ >>> poly_reg.predict(np.array([-1]))
+ array([-11.])
+ >>> poly_reg.predict(np.array([-2]))
+ array([-27.])
+ >>> poly_reg.predict(np.array([6]))
+ array([157.])
+ >>> PolynomialRegression(degree=3).predict(x)
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: Predictor hasn't been fit yet
+ """
+ if self.params is None:
+ raise ArithmeticError("Predictor hasn't been fit yet")
+
+ return PolynomialRegression._design_matrix(data, self.degree) @ self.params
+
+
+def main() -> None:
+ """
+ Fit a polynomial regression model to predict fuel efficiency using seaborn's mpg
+ dataset
+
+ >>> pass # Placeholder, function is only for demo purposes
+ """
+ import seaborn as sns
+
+ mpg_data = sns.load_dataset("mpg")
+
+ poly_reg = PolynomialRegression(degree=2)
+ poly_reg.fit(mpg_data.weight, mpg_data.mpg)
+
+ weight_sorted = np.sort(mpg_data.weight)
+ predictions = poly_reg.predict(weight_sorted)
+
+ plt.scatter(mpg_data.weight, mpg_data.mpg, color="gray", alpha=0.5)
+ plt.plot(weight_sorted, predictions, color="red", linewidth=3)
+ plt.title("Predicting Fuel Efficiency Using Polynomial Regression")
+ plt.xlabel("Weight (lbs)")
+ plt.ylabel("Fuel Efficiency (mpg)")
+ plt.show()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ main()
diff --git a/machine_learning/principle_component_analysis.py b/machine_learning/principle_component_analysis.py
new file mode 100644
index 000000000000..46ccdb968494
--- /dev/null
+++ b/machine_learning/principle_component_analysis.py
@@ -0,0 +1,85 @@
+"""
+Principal Component Analysis (PCA) is a dimensionality reduction technique
+used in machine learning. It transforms high-dimensional data into a lower-dimensional
+representation while retaining as much variance as possible.
+
+This implementation follows best practices, including:
+- Standardizing the dataset.
+- Computing principal components using Singular Value Decomposition (SVD).
+- Returning transformed data and explained variance ratio.
+"""
+
+import doctest
+
+import numpy as np
+from sklearn.datasets import load_iris
+from sklearn.decomposition import PCA
+from sklearn.preprocessing import StandardScaler
+
+
+def collect_dataset() -> tuple[np.ndarray, np.ndarray]:
+ """
+ Collects the dataset (Iris dataset) and returns feature matrix and target values.
+
+ :return: Tuple containing feature matrix (X) and target labels (y)
+
+ Example:
+ >>> X, y = collect_dataset()
+ >>> X.shape
+ (150, 4)
+ >>> y.shape
+ (150,)
+ """
+ data = load_iris()
+ return np.array(data.data), np.array(data.target)
+
+
+def apply_pca(data_x: np.ndarray, n_components: int) -> tuple[np.ndarray, np.ndarray]:
+ """
+ Applies Principal Component Analysis (PCA) to reduce dimensionality.
+
+ :param data_x: Original dataset (features)
+ :param n_components: Number of principal components to retain
+ :return: Tuple containing transformed dataset and explained variance ratio
+
+ Example:
+ >>> X, _ = collect_dataset()
+ >>> transformed_X, variance = apply_pca(X, 2)
+ >>> transformed_X.shape
+ (150, 2)
+ >>> len(variance) == 2
+ True
+ """
+ # Standardizing the dataset
+ scaler = StandardScaler()
+ data_x_scaled = scaler.fit_transform(data_x)
+
+ # Applying PCA
+ pca = PCA(n_components=n_components)
+ principal_components = pca.fit_transform(data_x_scaled)
+
+ return principal_components, pca.explained_variance_ratio_
+
+
+def main() -> None:
+ """
+ Driver function to execute PCA and display results.
+ """
+ data_x, data_y = collect_dataset()
+
+ # Number of principal components to retain
+ n_components = 2
+
+ # Apply PCA
+ transformed_data, variance_ratio = apply_pca(data_x, n_components)
+
+ print("Transformed Dataset (First 5 rows):")
+ print(transformed_data[:5])
+
+ print("\nExplained Variance Ratio:")
+ print(variance_ratio)
+
+
+if __name__ == "__main__":
+ doctest.testmod()
+ main()
diff --git a/machine_learning/random_forest_classifier.py b/machine_learning/random_forest_classifier.py.broken.txt
similarity index 91%
rename from machine_learning/random_forest_classifier.py
rename to machine_learning/random_forest_classifier.py.broken.txt
index 6370254090f7..3267fa209660 100644
--- a/machine_learning/random_forest_classifier.py
+++ b/machine_learning/random_forest_classifier.py.broken.txt
@@ -17,10 +17,10 @@ def main():
iris = load_iris()
# Split dataset into train and test data
- X = iris["data"] # features
- Y = iris["target"]
+ x = iris["data"] # features
+ y = iris["target"]
x_train, x_test, y_train, y_test = train_test_split(
- X, Y, test_size=0.3, random_state=1
+ x, y, test_size=0.3, random_state=1
)
# Random Forest Classifier
diff --git a/machine_learning/random_forest_regressor.py b/machine_learning/random_forest_regressor.py.broken.txt
similarity index 91%
rename from machine_learning/random_forest_regressor.py
rename to machine_learning/random_forest_regressor.py.broken.txt
index 0aade626b038..1001931a109d 100644
--- a/machine_learning/random_forest_regressor.py
+++ b/machine_learning/random_forest_regressor.py.broken.txt
@@ -17,10 +17,10 @@ def main():
print(boston.keys())
# Split dataset into train and test data
- X = boston["data"] # features
- Y = boston["target"]
+ x = boston["data"] # features
+ y = boston["target"]
x_train, x_test, y_train, y_test = train_test_split(
- X, Y, test_size=0.3, random_state=1
+ x, y, test_size=0.3, random_state=1
)
# Random Forest Regressor
diff --git a/machine_learning/scoring_functions.py b/machine_learning/scoring_functions.py
index 08b969a95c3b..f6b685f4f98a 100644
--- a/machine_learning/scoring_functions.py
+++ b/machine_learning/scoring_functions.py
@@ -20,11 +20,11 @@ def mae(predict, actual):
"""
Examples(rounded for precision):
>>> actual = [1,2,3];predict = [1,4,3]
- >>> np.around(mae(predict,actual),decimals = 2)
+ >>> float(np.around(mae(predict,actual),decimals = 2))
0.67
>>> actual = [1,1,1];predict = [1,1,1]
- >>> mae(predict,actual)
+ >>> float(mae(predict,actual))
0.0
"""
predict = np.array(predict)
@@ -41,11 +41,11 @@ def mse(predict, actual):
"""
Examples(rounded for precision):
>>> actual = [1,2,3];predict = [1,4,3]
- >>> np.around(mse(predict,actual),decimals = 2)
+ >>> float(np.around(mse(predict,actual),decimals = 2))
1.33
>>> actual = [1,1,1];predict = [1,1,1]
- >>> mse(predict,actual)
+ >>> float(mse(predict,actual))
0.0
"""
predict = np.array(predict)
@@ -63,11 +63,11 @@ def rmse(predict, actual):
"""
Examples(rounded for precision):
>>> actual = [1,2,3];predict = [1,4,3]
- >>> np.around(rmse(predict,actual),decimals = 2)
+ >>> float(np.around(rmse(predict,actual),decimals = 2))
1.15
>>> actual = [1,1,1];predict = [1,1,1]
- >>> rmse(predict,actual)
+ >>> float(rmse(predict,actual))
0.0
"""
predict = np.array(predict)
@@ -84,12 +84,10 @@ def rmse(predict, actual):
def rmsle(predict, actual):
"""
Examples(rounded for precision):
- >>> actual = [10,10,30];predict = [10,2,30]
- >>> np.around(rmsle(predict,actual),decimals = 2)
+ >>> float(np.around(rmsle(predict=[10, 2, 30], actual=[10, 10, 30]), decimals=2))
0.75
- >>> actual = [1,1,1];predict = [1,1,1]
- >>> rmsle(predict,actual)
+ >>> float(rmsle(predict=[1, 1, 1], actual=[1, 1, 1]))
0.0
"""
predict = np.array(predict)
@@ -117,12 +115,12 @@ def mbd(predict, actual):
Here the model overpredicts
>>> actual = [1,2,3];predict = [2,3,4]
- >>> np.around(mbd(predict,actual),decimals = 2)
+ >>> float(np.around(mbd(predict,actual),decimals = 2))
50.0
Here the model underpredicts
>>> actual = [1,2,3];predict = [0,1,1]
- >>> np.around(mbd(predict,actual),decimals = 2)
+ >>> float(np.around(mbd(predict,actual),decimals = 2))
-66.67
"""
predict = np.array(predict)
diff --git a/machine_learning/self_organizing_map.py b/machine_learning/self_organizing_map.py
new file mode 100644
index 000000000000..fb9d0074e791
--- /dev/null
+++ b/machine_learning/self_organizing_map.py
@@ -0,0 +1,73 @@
+"""
+https://en.wikipedia.org/wiki/Self-organizing_map
+"""
+
+import math
+
+
+class SelfOrganizingMap:
+ def get_winner(self, weights: list[list[float]], sample: list[int]) -> int:
+ """
+ Compute the winning vector by Euclidean distance
+
+ >>> SelfOrganizingMap().get_winner([[1, 2, 3], [4, 5, 6]], [1, 2, 3])
+ 1
+ """
+ d0 = 0.0
+ d1 = 0.0
+ for i in range(len(sample)):
+ d0 += math.pow((sample[i] - weights[0][i]), 2)
+ d1 += math.pow((sample[i] - weights[1][i]), 2)
+ return 0 if d0 > d1 else 1
+ return 0
+
+ def update(
+ self, weights: list[list[int | float]], sample: list[int], j: int, alpha: float
+ ) -> list[list[int | float]]:
+ """
+ Update the winning vector.
+
+ >>> SelfOrganizingMap().update([[1, 2, 3], [4, 5, 6]], [1, 2, 3], 1, 0.1)
+ [[1, 2, 3], [3.7, 4.7, 6]]
+ """
+ for i in range(len(weights)):
+ weights[j][i] += alpha * (sample[i] - weights[j][i])
+ return weights
+
+
+# Driver code
+def main() -> None:
+ # Training Examples ( m, n )
+ training_samples = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]]
+
+ # weight initialization ( n, C )
+ weights = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]]
+
+ # training
+ self_organizing_map = SelfOrganizingMap()
+ epochs = 3
+ alpha = 0.5
+
+ for _ in range(epochs):
+ for j in range(len(training_samples)):
+ # training sample
+ sample = training_samples[j]
+
+ # Compute the winning vector
+ winner = self_organizing_map.get_winner(weights, sample)
+
+ # Update the winning vector
+ weights = self_organizing_map.update(weights, sample, winner, alpha)
+
+ # classify test sample
+ sample = [0, 0, 0, 1]
+ winner = self_organizing_map.get_winner(weights, sample)
+
+ # results
+ print(f"Clusters that the test sample belongs to : {winner}")
+ print(f"Weights that have been trained : {weights}")
+
+
+# running the main() function
+if __name__ == "__main__":
+ main()
diff --git a/machine_learning/sequential_minimum_optimization.py b/machine_learning/sequential_minimum_optimization.py
index 98ce05c46cff..625fc28fe60c 100644
--- a/machine_learning/sequential_minimum_optimization.py
+++ b/machine_learning/sequential_minimum_optimization.py
@@ -1,632 +1,622 @@
-"""
- Implementation of sequential minimal optimization (SMO) for support vector machines
- (SVM).
-
- Sequential minimal optimization (SMO) is an algorithm for solving the quadratic
- programming (QP) problem that arises during the training of support vector
- machines.
- It was invented by John Platt in 1998.
-
-Input:
- 0: type: numpy.ndarray.
- 1: first column of ndarray must be tags of samples, must be 1 or -1.
- 2: rows of ndarray represent samples.
-
-Usage:
- Command:
- python3 sequential_minimum_optimization.py
- Code:
- from sequential_minimum_optimization import SmoSVM, Kernel
-
- kernel = Kernel(kernel='poly', degree=3., coef0=1., gamma=0.5)
- init_alphas = np.zeros(train.shape[0])
- SVM = SmoSVM(train=train, alpha_list=init_alphas, kernel_func=kernel, cost=0.4,
- b=0.0, tolerance=0.001)
- SVM.fit()
- predict = SVM.predict(test_samples)
-
-Reference:
- https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/smo-book.pdf
- https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/tr-98-14.pdf
- http://web.cs.iastate.edu/~honavar/smo-svm.pdf
-"""
-
-
-import os
-import sys
-import urllib.request
-
-import numpy as np
-import pandas as pd
-from matplotlib import pyplot as plt
-from sklearn.datasets import make_blobs, make_circles
-from sklearn.preprocessing import StandardScaler
-
-CANCER_DATASET_URL = (
- "http://archive.ics.uci.edu/ml/machine-learning-databases/"
- "breast-cancer-wisconsin/wdbc.data"
-)
-
-
-class SmoSVM:
- def __init__(
- self,
- train,
- kernel_func,
- alpha_list=None,
- cost=0.4,
- b=0.0,
- tolerance=0.001,
- auto_norm=True,
- ):
- self._init = True
- self._auto_norm = auto_norm
- self._c = np.float64(cost)
- self._b = np.float64(b)
- self._tol = np.float64(tolerance) if tolerance > 0.0001 else np.float64(0.001)
-
- self.tags = train[:, 0]
- self.samples = self._norm(train[:, 1:]) if self._auto_norm else train[:, 1:]
- self.alphas = alpha_list if alpha_list is not None else np.zeros(train.shape[0])
- self.Kernel = kernel_func
-
- self._eps = 0.001
- self._all_samples = list(range(self.length))
- self._K_matrix = self._calculate_k_matrix()
- self._error = np.zeros(self.length)
- self._unbound = []
-
- self.choose_alpha = self._choose_alphas()
-
- # Calculate alphas using SMO algorithm
- def fit(self):
- K = self._k
- state = None
- while True:
-
- # 1: Find alpha1, alpha2
- try:
- i1, i2 = self.choose_alpha.send(state)
- state = None
- except StopIteration:
- print("Optimization done!\nEvery sample satisfy the KKT condition!")
- break
-
- # 2: calculate new alpha2 and new alpha1
- y1, y2 = self.tags[i1], self.tags[i2]
- a1, a2 = self.alphas[i1].copy(), self.alphas[i2].copy()
- e1, e2 = self._e(i1), self._e(i2)
- args = (i1, i2, a1, a2, e1, e2, y1, y2)
- a1_new, a2_new = self._get_new_alpha(*args)
- if not a1_new and not a2_new:
- state = False
- continue
- self.alphas[i1], self.alphas[i2] = a1_new, a2_new
-
- # 3: update threshold(b)
- b1_new = np.float64(
- -e1
- - y1 * K(i1, i1) * (a1_new - a1)
- - y2 * K(i2, i1) * (a2_new - a2)
- + self._b
- )
- b2_new = np.float64(
- -e2
- - y2 * K(i2, i2) * (a2_new - a2)
- - y1 * K(i1, i2) * (a1_new - a1)
- + self._b
- )
- if 0.0 < a1_new < self._c:
- b = b1_new
- if 0.0 < a2_new < self._c:
- b = b2_new
- if not (np.float64(0) < a2_new < self._c) and not (
- np.float64(0) < a1_new < self._c
- ):
- b = (b1_new + b2_new) / 2.0
- b_old = self._b
- self._b = b
-
- # 4: update error value,here we only calculate those non-bound samples'
- # error
- self._unbound = [i for i in self._all_samples if self._is_unbound(i)]
- for s in self.unbound:
- if s == i1 or s == i2:
- continue
- self._error[s] += (
- y1 * (a1_new - a1) * K(i1, s)
- + y2 * (a2_new - a2) * K(i2, s)
- + (self._b - b_old)
- )
-
- # if i1 or i2 is non-bound,update there error value to zero
- if self._is_unbound(i1):
- self._error[i1] = 0
- if self._is_unbound(i2):
- self._error[i2] = 0
-
- # Predict test samles
- def predict(self, test_samples, classify=True):
-
- if test_samples.shape[1] > self.samples.shape[1]:
- raise ValueError(
- "Test samples' feature length does not equal to that of train samples"
- )
-
- if self._auto_norm:
- test_samples = self._norm(test_samples)
-
- results = []
- for test_sample in test_samples:
- result = self._predict(test_sample)
- if classify:
- results.append(1 if result > 0 else -1)
- else:
- results.append(result)
- return np.array(results)
-
- # Check if alpha violate KKT condition
- def _check_obey_kkt(self, index):
- alphas = self.alphas
- tol = self._tol
- r = self._e(index) * self.tags[index]
- c = self._c
-
- return (r < -tol and alphas[index] < c) or (r > tol and alphas[index] > 0.0)
-
- # Get value calculated from kernel function
- def _k(self, i1, i2):
- # for test samples,use Kernel function
- if isinstance(i2, np.ndarray):
- return self.Kernel(self.samples[i1], i2)
- # for train samples,Kernel values have been saved in matrix
- else:
- return self._K_matrix[i1, i2]
-
- # Get sample's error
- def _e(self, index):
- """
- Two cases:
- 1:Sample[index] is non-bound,Fetch error from list: _error
- 2:sample[index] is bound,Use predicted value deduct true value: g(xi) - yi
-
- """
- # get from error data
- if self._is_unbound(index):
- return self._error[index]
- # get by g(xi) - yi
- else:
- gx = np.dot(self.alphas * self.tags, self._K_matrix[:, index]) + self._b
- yi = self.tags[index]
- return gx - yi
-
- # Calculate Kernel matrix of all possible i1,i2 ,saving time
- def _calculate_k_matrix(self):
- k_matrix = np.zeros([self.length, self.length])
- for i in self._all_samples:
- for j in self._all_samples:
- k_matrix[i, j] = np.float64(
- self.Kernel(self.samples[i, :], self.samples[j, :])
- )
- return k_matrix
-
- # Predict test sample's tag
- def _predict(self, sample):
- k = self._k
- predicted_value = (
- np.sum(
- [
- self.alphas[i1] * self.tags[i1] * k(i1, sample)
- for i1 in self._all_samples
- ]
- )
- + self._b
- )
- return predicted_value
-
- # Choose alpha1 and alpha2
- def _choose_alphas(self):
- locis = yield from self._choose_a1()
- if not locis:
- return
- return locis
-
- def _choose_a1(self):
- """
- Choose first alpha ;steps:
- 1:First loop over all sample
- 2:Second loop over all non-bound samples till all non-bound samples does not
- voilate kkt condition.
- 3:Repeat this two process endlessly,till all samples does not voilate kkt
- condition samples after first loop.
- """
- while True:
- all_not_obey = True
- # all sample
- print("scanning all sample!")
- for i1 in [i for i in self._all_samples if self._check_obey_kkt(i)]:
- all_not_obey = False
- yield from self._choose_a2(i1)
-
- # non-bound sample
- print("scanning non-bound sample!")
- while True:
- not_obey = True
- for i1 in [
- i
- for i in self._all_samples
- if self._check_obey_kkt(i) and self._is_unbound(i)
- ]:
- not_obey = False
- yield from self._choose_a2(i1)
- if not_obey:
- print("all non-bound samples fit the KKT condition!")
- break
- if all_not_obey:
- print("all samples fit the KKT condition! Optimization done!")
- break
- return False
-
- def _choose_a2(self, i1):
- """
- Choose the second alpha by using heuristic algorithm ;steps:
- 1: Choose alpha2 which gets the maximum step size (|E1 - E2|).
- 2: Start in a random point,loop over all non-bound samples till alpha1 and
- alpha2 are optimized.
- 3: Start in a random point,loop over all samples till alpha1 and alpha2 are
- optimized.
- """
- self._unbound = [i for i in self._all_samples if self._is_unbound(i)]
-
- if len(self.unbound) > 0:
- tmp_error = self._error.copy().tolist()
- tmp_error_dict = {
- index: value
- for index, value in enumerate(tmp_error)
- if self._is_unbound(index)
- }
- if self._e(i1) >= 0:
- i2 = min(tmp_error_dict, key=lambda index: tmp_error_dict[index])
- else:
- i2 = max(tmp_error_dict, key=lambda index: tmp_error_dict[index])
- cmd = yield i1, i2
- if cmd is None:
- return
-
- for i2 in np.roll(self.unbound, np.random.choice(self.length)):
- cmd = yield i1, i2
- if cmd is None:
- return
-
- for i2 in np.roll(self._all_samples, np.random.choice(self.length)):
- cmd = yield i1, i2
- if cmd is None:
- return
-
- # Get the new alpha2 and new alpha1
- def _get_new_alpha(self, i1, i2, a1, a2, e1, e2, y1, y2):
- K = self._k
- if i1 == i2:
- return None, None
-
- # calculate L and H which bound the new alpha2
- s = y1 * y2
- if s == -1:
- L, H = max(0.0, a2 - a1), min(self._c, self._c + a2 - a1)
- else:
- L, H = max(0.0, a2 + a1 - self._c), min(self._c, a2 + a1)
- if L == H:
- return None, None
-
- # calculate eta
- k11 = K(i1, i1)
- k22 = K(i2, i2)
- k12 = K(i1, i2)
- eta = k11 + k22 - 2.0 * k12
-
- # select the new alpha2 which could get the minimal objectives
- if eta > 0.0:
- a2_new_unc = a2 + (y2 * (e1 - e2)) / eta
- # a2_new has a boundary
- if a2_new_unc >= H:
- a2_new = H
- elif a2_new_unc <= L:
- a2_new = L
- else:
- a2_new = a2_new_unc
- else:
- b = self._b
- l1 = a1 + s * (a2 - L)
- h1 = a1 + s * (a2 - H)
-
- # way 1
- f1 = y1 * (e1 + b) - a1 * K(i1, i1) - s * a2 * K(i1, i2)
- f2 = y2 * (e2 + b) - a2 * K(i2, i2) - s * a1 * K(i1, i2)
- ol = (
- l1 * f1
- + L * f2
- + 1 / 2 * l1 ** 2 * K(i1, i1)
- + 1 / 2 * L ** 2 * K(i2, i2)
- + s * L * l1 * K(i1, i2)
- )
- oh = (
- h1 * f1
- + H * f2
- + 1 / 2 * h1 ** 2 * K(i1, i1)
- + 1 / 2 * H ** 2 * K(i2, i2)
- + s * H * h1 * K(i1, i2)
- )
- """
- # way 2
- Use objective function check which alpha2 new could get the minimal
- objectives
- """
- if ol < (oh - self._eps):
- a2_new = L
- elif ol > oh + self._eps:
- a2_new = H
- else:
- a2_new = a2
-
- # a1_new has a boundary too
- a1_new = a1 + s * (a2 - a2_new)
- if a1_new < 0:
- a2_new += s * a1_new
- a1_new = 0
- if a1_new > self._c:
- a2_new += s * (a1_new - self._c)
- a1_new = self._c
-
- return a1_new, a2_new
-
- # Normalise data using min_max way
- def _norm(self, data):
- if self._init:
- self._min = np.min(data, axis=0)
- self._max = np.max(data, axis=0)
- self._init = False
- return (data - self._min) / (self._max - self._min)
- else:
- return (data - self._min) / (self._max - self._min)
-
- def _is_unbound(self, index):
- if 0.0 < self.alphas[index] < self._c:
- return True
- else:
- return False
-
- def _is_support(self, index):
- if self.alphas[index] > 0:
- return True
- else:
- return False
-
- @property
- def unbound(self):
- return self._unbound
-
- @property
- def support(self):
- return [i for i in range(self.length) if self._is_support(i)]
-
- @property
- def length(self):
- return self.samples.shape[0]
-
-
-class Kernel:
- def __init__(self, kernel, degree=1.0, coef0=0.0, gamma=1.0):
- self.degree = np.float64(degree)
- self.coef0 = np.float64(coef0)
- self.gamma = np.float64(gamma)
- self._kernel_name = kernel
- self._kernel = self._get_kernel(kernel_name=kernel)
- self._check()
-
- def _polynomial(self, v1, v2):
- return (self.gamma * np.inner(v1, v2) + self.coef0) ** self.degree
-
- def _linear(self, v1, v2):
- return np.inner(v1, v2) + self.coef0
-
- def _rbf(self, v1, v2):
- return np.exp(-1 * (self.gamma * np.linalg.norm(v1 - v2) ** 2))
-
- def _check(self):
- if self._kernel == self._rbf:
- if self.gamma < 0:
- raise ValueError("gamma value must greater than 0")
-
- def _get_kernel(self, kernel_name):
- maps = {"linear": self._linear, "poly": self._polynomial, "rbf": self._rbf}
- return maps[kernel_name]
-
- def __call__(self, v1, v2):
- return self._kernel(v1, v2)
-
- def __repr__(self):
- return self._kernel_name
-
-
-def count_time(func):
- def call_func(*args, **kwargs):
- import time
-
- start_time = time.time()
- func(*args, **kwargs)
- end_time = time.time()
- print(f"smo algorithm cost {end_time - start_time} seconds")
-
- return call_func
-
-
-@count_time
-def test_cancel_data():
- print("Hello!\nStart test svm by smo algorithm!")
- # 0: download dataset and load into pandas' dataframe
- if not os.path.exists(r"cancel_data.csv"):
- request = urllib.request.Request(
- CANCER_DATASET_URL,
- headers={"User-Agent": "Mozilla/4.0 (compatible; MSIE 5.5; Windows NT)"},
- )
- response = urllib.request.urlopen(request)
- content = response.read().decode("utf-8")
- with open(r"cancel_data.csv", "w") as f:
- f.write(content)
-
- data = pd.read_csv(r"cancel_data.csv", header=None)
-
- # 1: pre-processing data
- del data[data.columns.tolist()[0]]
- data = data.dropna(axis=0)
- data = data.replace({"M": np.float64(1), "B": np.float64(-1)})
- samples = np.array(data)[:, :]
-
- # 2: dividing data into train_data data and test_data data
- train_data, test_data = samples[:328, :], samples[328:, :]
- test_tags, test_samples = test_data[:, 0], test_data[:, 1:]
-
- # 3: choose kernel function,and set initial alphas to zero(optional)
- mykernel = Kernel(kernel="rbf", degree=5, coef0=1, gamma=0.5)
- al = np.zeros(train_data.shape[0])
-
- # 4: calculating best alphas using SMO algorithm and predict test_data samples
- mysvm = SmoSVM(
- train=train_data,
- kernel_func=mykernel,
- alpha_list=al,
- cost=0.4,
- b=0.0,
- tolerance=0.001,
- )
- mysvm.fit()
- predict = mysvm.predict(test_samples)
-
- # 5: check accuracy
- score = 0
- test_num = test_tags.shape[0]
- for i in range(test_tags.shape[0]):
- if test_tags[i] == predict[i]:
- score += 1
- print(f"\nall: {test_num}\nright: {score}\nfalse: {test_num - score}")
- print(f"Rough Accuracy: {score / test_tags.shape[0]}")
-
-
-def test_demonstration():
- # change stdout
- print("\nStart plot,please wait!!!")
- sys.stdout = open(os.devnull, "w")
-
- ax1 = plt.subplot2grid((2, 2), (0, 0))
- ax2 = plt.subplot2grid((2, 2), (0, 1))
- ax3 = plt.subplot2grid((2, 2), (1, 0))
- ax4 = plt.subplot2grid((2, 2), (1, 1))
- ax1.set_title("linear svm,cost:0.1")
- test_linear_kernel(ax1, cost=0.1)
- ax2.set_title("linear svm,cost:500")
- test_linear_kernel(ax2, cost=500)
- ax3.set_title("rbf kernel svm,cost:0.1")
- test_rbf_kernel(ax3, cost=0.1)
- ax4.set_title("rbf kernel svm,cost:500")
- test_rbf_kernel(ax4, cost=500)
-
- sys.stdout = sys.__stdout__
- print("Plot done!!!")
-
-
-def test_linear_kernel(ax, cost):
- train_x, train_y = make_blobs(
- n_samples=500, centers=2, n_features=2, random_state=1
- )
- train_y[train_y == 0] = -1
- scaler = StandardScaler()
- train_x_scaled = scaler.fit_transform(train_x, train_y)
- train_data = np.hstack((train_y.reshape(500, 1), train_x_scaled))
- mykernel = Kernel(kernel="linear", degree=5, coef0=1, gamma=0.5)
- mysvm = SmoSVM(
- train=train_data,
- kernel_func=mykernel,
- cost=cost,
- tolerance=0.001,
- auto_norm=False,
- )
- mysvm.fit()
- plot_partition_boundary(mysvm, train_data, ax=ax)
-
-
-def test_rbf_kernel(ax, cost):
- train_x, train_y = make_circles(
- n_samples=500, noise=0.1, factor=0.1, random_state=1
- )
- train_y[train_y == 0] = -1
- scaler = StandardScaler()
- train_x_scaled = scaler.fit_transform(train_x, train_y)
- train_data = np.hstack((train_y.reshape(500, 1), train_x_scaled))
- mykernel = Kernel(kernel="rbf", degree=5, coef0=1, gamma=0.5)
- mysvm = SmoSVM(
- train=train_data,
- kernel_func=mykernel,
- cost=cost,
- tolerance=0.001,
- auto_norm=False,
- )
- mysvm.fit()
- plot_partition_boundary(mysvm, train_data, ax=ax)
-
-
-def plot_partition_boundary(
- model, train_data, ax, resolution=100, colors=("b", "k", "r")
-):
- """
- We can not get the optimum w of our kernel svm model which is different from linear
- svm. For this reason, we generate randomly distributed points with high desity and
- prediced values of these points are calculated by using our tained model. Then we
- could use this prediced values to draw contour map.
- And this contour map can represent svm's partition boundary.
- """
- train_data_x = train_data[:, 1]
- train_data_y = train_data[:, 2]
- train_data_tags = train_data[:, 0]
- xrange = np.linspace(train_data_x.min(), train_data_x.max(), resolution)
- yrange = np.linspace(train_data_y.min(), train_data_y.max(), resolution)
- test_samples = np.array([(x, y) for x in xrange for y in yrange]).reshape(
- resolution * resolution, 2
- )
-
- test_tags = model.predict(test_samples, classify=False)
- grid = test_tags.reshape((len(xrange), len(yrange)))
-
- # Plot contour map which represents the partition boundary
- ax.contour(
- xrange,
- yrange,
- np.mat(grid).T,
- levels=(-1, 0, 1),
- linestyles=("--", "-", "--"),
- linewidths=(1, 1, 1),
- colors=colors,
- )
- # Plot all train samples
- ax.scatter(
- train_data_x,
- train_data_y,
- c=train_data_tags,
- cmap=plt.cm.Dark2,
- lw=0,
- alpha=0.5,
- )
-
- # Plot support vectors
- support = model.support
- ax.scatter(
- train_data_x[support],
- train_data_y[support],
- c=train_data_tags[support],
- cmap=plt.cm.Dark2,
- )
-
-
-if __name__ == "__main__":
- test_cancel_data()
- test_demonstration()
- plt.show()
+"""
+Sequential minimal optimization (SMO) for support vector machines (SVM)
+
+Sequential minimal optimization (SMO) is an algorithm for solving the quadratic
+programming (QP) problem that arises during the training of SVMs. It was invented by
+John Platt in 1998.
+
+Input:
+ 0: type: numpy.ndarray.
+ 1: first column of ndarray must be tags of samples, must be 1 or -1.
+ 2: rows of ndarray represent samples.
+
+Usage:
+ Command:
+ python3 sequential_minimum_optimization.py
+ Code:
+ from sequential_minimum_optimization import SmoSVM, Kernel
+
+ kernel = Kernel(kernel='poly', degree=3., coef0=1., gamma=0.5)
+ init_alphas = np.zeros(train.shape[0])
+ SVM = SmoSVM(train=train, alpha_list=init_alphas, kernel_func=kernel, cost=0.4,
+ b=0.0, tolerance=0.001)
+ SVM.fit()
+ predict = SVM.predict(test_samples)
+
+Reference:
+ https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/smo-book.pdf
+ https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/tr-98-14.pdf
+"""
+
+import os
+import sys
+import urllib.request
+
+import numpy as np
+import pandas as pd
+from matplotlib import pyplot as plt
+from sklearn.datasets import make_blobs, make_circles
+from sklearn.preprocessing import StandardScaler
+
+CANCER_DATASET_URL = (
+ "https://archive.ics.uci.edu/ml/machine-learning-databases/"
+ "breast-cancer-wisconsin/wdbc.data"
+)
+
+
+class SmoSVM:
+ def __init__(
+ self,
+ train,
+ kernel_func,
+ alpha_list=None,
+ cost=0.4,
+ b=0.0,
+ tolerance=0.001,
+ auto_norm=True,
+ ):
+ self._init = True
+ self._auto_norm = auto_norm
+ self._c = np.float64(cost)
+ self._b = np.float64(b)
+ self._tol = np.float64(tolerance) if tolerance > 0.0001 else np.float64(0.001)
+
+ self.tags = train[:, 0]
+ self.samples = self._norm(train[:, 1:]) if self._auto_norm else train[:, 1:]
+ self.alphas = alpha_list if alpha_list is not None else np.zeros(train.shape[0])
+ self.Kernel = kernel_func
+
+ self._eps = 0.001
+ self._all_samples = list(range(self.length))
+ self._K_matrix = self._calculate_k_matrix()
+ self._error = np.zeros(self.length)
+ self._unbound = []
+
+ self.choose_alpha = self._choose_alphas()
+
+ # Calculate alphas using SMO algorithm
+ def fit(self):
+ k = self._k
+ state = None
+ while True:
+ # 1: Find alpha1, alpha2
+ try:
+ i1, i2 = self.choose_alpha.send(state)
+ state = None
+ except StopIteration:
+ print("Optimization done!\nEvery sample satisfy the KKT condition!")
+ break
+
+ # 2: calculate new alpha2 and new alpha1
+ y1, y2 = self.tags[i1], self.tags[i2]
+ a1, a2 = self.alphas[i1].copy(), self.alphas[i2].copy()
+ e1, e2 = self._e(i1), self._e(i2)
+ args = (i1, i2, a1, a2, e1, e2, y1, y2)
+ a1_new, a2_new = self._get_new_alpha(*args)
+ if not a1_new and not a2_new:
+ state = False
+ continue
+ self.alphas[i1], self.alphas[i2] = a1_new, a2_new
+
+ # 3: update threshold(b)
+ b1_new = np.float64(
+ -e1
+ - y1 * k(i1, i1) * (a1_new - a1)
+ - y2 * k(i2, i1) * (a2_new - a2)
+ + self._b
+ )
+ b2_new = np.float64(
+ -e2
+ - y2 * k(i2, i2) * (a2_new - a2)
+ - y1 * k(i1, i2) * (a1_new - a1)
+ + self._b
+ )
+ if 0.0 < a1_new < self._c:
+ b = b1_new
+ if 0.0 < a2_new < self._c:
+ b = b2_new
+ if not (np.float64(0) < a2_new < self._c) and not (
+ np.float64(0) < a1_new < self._c
+ ):
+ b = (b1_new + b2_new) / 2.0
+ b_old = self._b
+ self._b = b
+
+ # 4: update error, here we only calculate the error for non-bound samples
+ self._unbound = [i for i in self._all_samples if self._is_unbound(i)]
+ for s in self.unbound:
+ if s in (i1, i2):
+ continue
+ self._error[s] += (
+ y1 * (a1_new - a1) * k(i1, s)
+ + y2 * (a2_new - a2) * k(i2, s)
+ + (self._b - b_old)
+ )
+
+ # if i1 or i2 is non-bound, update their error value to zero
+ if self._is_unbound(i1):
+ self._error[i1] = 0
+ if self._is_unbound(i2):
+ self._error[i2] = 0
+
+ # Predict test samples
+ def predict(self, test_samples, classify=True):
+ if test_samples.shape[1] > self.samples.shape[1]:
+ raise ValueError(
+ "Test samples' feature length does not equal to that of train samples"
+ )
+
+ if self._auto_norm:
+ test_samples = self._norm(test_samples)
+
+ results = []
+ for test_sample in test_samples:
+ result = self._predict(test_sample)
+ if classify:
+ results.append(1 if result > 0 else -1)
+ else:
+ results.append(result)
+ return np.array(results)
+
+ # Check if alpha violates the KKT condition
+ def _check_obey_kkt(self, index):
+ alphas = self.alphas
+ tol = self._tol
+ r = self._e(index) * self.tags[index]
+ c = self._c
+
+ return (r < -tol and alphas[index] < c) or (r > tol and alphas[index] > 0.0)
+
+ # Get value calculated from kernel function
+ def _k(self, i1, i2):
+ # for test samples, use kernel function
+ if isinstance(i2, np.ndarray):
+ return self.Kernel(self.samples[i1], i2)
+ # for training samples, kernel values have been saved in matrix
+ else:
+ return self._K_matrix[i1, i2]
+
+ # Get error for sample
+ def _e(self, index):
+ """
+ Two cases:
+ 1: Sample[index] is non-bound, fetch error from list: _error
+ 2: sample[index] is bound, use predicted value minus true value: g(xi) - yi
+ """
+ # get from error data
+ if self._is_unbound(index):
+ return self._error[index]
+ # get by g(xi) - yi
+ else:
+ gx = np.dot(self.alphas * self.tags, self._K_matrix[:, index]) + self._b
+ yi = self.tags[index]
+ return gx - yi
+
+ # Calculate kernel matrix of all possible i1, i2, saving time
+ def _calculate_k_matrix(self):
+ k_matrix = np.zeros([self.length, self.length])
+ for i in self._all_samples:
+ for j in self._all_samples:
+ k_matrix[i, j] = np.float64(
+ self.Kernel(self.samples[i, :], self.samples[j, :])
+ )
+ return k_matrix
+
+ # Predict tag for test sample
+ def _predict(self, sample):
+ k = self._k
+ predicted_value = (
+ np.sum(
+ [
+ self.alphas[i1] * self.tags[i1] * k(i1, sample)
+ for i1 in self._all_samples
+ ]
+ )
+ + self._b
+ )
+ return predicted_value
+
+ # Choose alpha1 and alpha2
+ def _choose_alphas(self):
+ loci = yield from self._choose_a1()
+ if not loci:
+ return None
+ return loci
+
+ def _choose_a1(self):
+ """
+ Choose first alpha
+ Steps:
+ 1: First loop over all samples
+ 2: Second loop over all non-bound samples until no non-bound samples violate
+ the KKT condition.
+ 3: Repeat these two processes until no samples violate the KKT condition
+ after the first loop.
+ """
+ while True:
+ all_not_obey = True
+ # all sample
+ print("Scanning all samples!")
+ for i1 in [i for i in self._all_samples if self._check_obey_kkt(i)]:
+ all_not_obey = False
+ yield from self._choose_a2(i1)
+
+ # non-bound sample
+ print("Scanning non-bound samples!")
+ while True:
+ not_obey = True
+ for i1 in [
+ i
+ for i in self._all_samples
+ if self._check_obey_kkt(i) and self._is_unbound(i)
+ ]:
+ not_obey = False
+ yield from self._choose_a2(i1)
+ if not_obey:
+ print("All non-bound samples satisfy the KKT condition!")
+ break
+ if all_not_obey:
+ print("All samples satisfy the KKT condition!")
+ break
+ return False
+
+ def _choose_a2(self, i1):
+ """
+ Choose the second alpha using a heuristic algorithm
+ Steps:
+ 1: Choose alpha2 that maximizes the step size (|E1 - E2|).
+ 2: Start in a random point, loop over all non-bound samples till alpha1 and
+ alpha2 are optimized.
+ 3: Start in a random point, loop over all samples till alpha1 and alpha2 are
+ optimized.
+ """
+ self._unbound = [i for i in self._all_samples if self._is_unbound(i)]
+
+ if len(self.unbound) > 0:
+ tmp_error = self._error.copy().tolist()
+ tmp_error_dict = {
+ index: value
+ for index, value in enumerate(tmp_error)
+ if self._is_unbound(index)
+ }
+ if self._e(i1) >= 0:
+ i2 = min(tmp_error_dict, key=lambda index: tmp_error_dict[index])
+ else:
+ i2 = max(tmp_error_dict, key=lambda index: tmp_error_dict[index])
+ cmd = yield i1, i2
+ if cmd is None:
+ return
+
+ rng = np.random.default_rng()
+ for i2 in np.roll(self.unbound, rng.choice(self.length)):
+ cmd = yield i1, i2
+ if cmd is None:
+ return
+
+ for i2 in np.roll(self._all_samples, rng.choice(self.length)):
+ cmd = yield i1, i2
+ if cmd is None:
+ return
+
+ # Get the new alpha2 and new alpha1
+ def _get_new_alpha(self, i1, i2, a1, a2, e1, e2, y1, y2):
+ k = self._k
+ if i1 == i2:
+ return None, None
+
+ # calculate L and H which bound the new alpha2
+ s = y1 * y2
+ if s == -1:
+ l, h = max(0.0, a2 - a1), min(self._c, self._c + a2 - a1) # noqa: E741
+ else:
+ l, h = max(0.0, a2 + a1 - self._c), min(self._c, a2 + a1) # noqa: E741
+ if l == h:
+ return None, None
+
+ # calculate eta
+ k11 = k(i1, i1)
+ k22 = k(i2, i2)
+ k12 = k(i1, i2)
+
+ # select the new alpha2 which could achieve the minimal objectives
+ if (eta := k11 + k22 - 2.0 * k12) > 0.0:
+ a2_new_unc = a2 + (y2 * (e1 - e2)) / eta
+ # a2_new has a boundary
+ if a2_new_unc >= h:
+ a2_new = h
+ elif a2_new_unc <= l:
+ a2_new = l
+ else:
+ a2_new = a2_new_unc
+ else:
+ b = self._b
+ l1 = a1 + s * (a2 - l)
+ h1 = a1 + s * (a2 - h)
+
+ # Method 1
+ f1 = y1 * (e1 + b) - a1 * k(i1, i1) - s * a2 * k(i1, i2)
+ f2 = y2 * (e2 + b) - a2 * k(i2, i2) - s * a1 * k(i1, i2)
+ ol = (
+ l1 * f1
+ + l * f2
+ + 1 / 2 * l1**2 * k(i1, i1)
+ + 1 / 2 * l**2 * k(i2, i2)
+ + s * l * l1 * k(i1, i2)
+ )
+ oh = (
+ h1 * f1
+ + h * f2
+ + 1 / 2 * h1**2 * k(i1, i1)
+ + 1 / 2 * h**2 * k(i2, i2)
+ + s * h * h1 * k(i1, i2)
+ )
+ """
+ Method 2: Use objective function to check which alpha2_new could achieve the
+ minimal objectives
+ """
+ if ol < (oh - self._eps):
+ a2_new = l
+ elif ol > oh + self._eps:
+ a2_new = h
+ else:
+ a2_new = a2
+
+ # a1_new has a boundary too
+ a1_new = a1 + s * (a2 - a2_new)
+ if a1_new < 0:
+ a2_new += s * a1_new
+ a1_new = 0
+ if a1_new > self._c:
+ a2_new += s * (a1_new - self._c)
+ a1_new = self._c
+
+ return a1_new, a2_new
+
+ # Normalize data using min-max method
+ def _norm(self, data):
+ if self._init:
+ self._min = np.min(data, axis=0)
+ self._max = np.max(data, axis=0)
+ self._init = False
+ return (data - self._min) / (self._max - self._min)
+ else:
+ return (data - self._min) / (self._max - self._min)
+
+ def _is_unbound(self, index):
+ return bool(0.0 < self.alphas[index] < self._c)
+
+ def _is_support(self, index):
+ return bool(self.alphas[index] > 0)
+
+ @property
+ def unbound(self):
+ return self._unbound
+
+ @property
+ def support(self):
+ return [i for i in range(self.length) if self._is_support(i)]
+
+ @property
+ def length(self):
+ return self.samples.shape[0]
+
+
+class Kernel:
+ def __init__(self, kernel, degree=1.0, coef0=0.0, gamma=1.0):
+ self.degree = np.float64(degree)
+ self.coef0 = np.float64(coef0)
+ self.gamma = np.float64(gamma)
+ self._kernel_name = kernel
+ self._kernel = self._get_kernel(kernel_name=kernel)
+ self._check()
+
+ def _polynomial(self, v1, v2):
+ return (self.gamma * np.inner(v1, v2) + self.coef0) ** self.degree
+
+ def _linear(self, v1, v2):
+ return np.inner(v1, v2) + self.coef0
+
+ def _rbf(self, v1, v2):
+ return np.exp(-1 * (self.gamma * np.linalg.norm(v1 - v2) ** 2))
+
+ def _check(self):
+ if self._kernel == self._rbf and self.gamma < 0:
+ raise ValueError("gamma value must be non-negative")
+
+ def _get_kernel(self, kernel_name):
+ maps = {"linear": self._linear, "poly": self._polynomial, "rbf": self._rbf}
+ return maps[kernel_name]
+
+ def __call__(self, v1, v2):
+ return self._kernel(v1, v2)
+
+ def __repr__(self):
+ return self._kernel_name
+
+
+def count_time(func):
+ def call_func(*args, **kwargs):
+ import time
+
+ start_time = time.time()
+ func(*args, **kwargs)
+ end_time = time.time()
+ print(f"SMO algorithm cost {end_time - start_time} seconds")
+
+ return call_func
+
+
+@count_time
+def test_cancer_data():
+ print("Hello!\nStart test SVM using the SMO algorithm!")
+ # 0: download dataset and load into pandas' dataframe
+ if not os.path.exists(r"cancer_data.csv"):
+ request = urllib.request.Request( # noqa: S310
+ CANCER_DATASET_URL,
+ headers={"User-Agent": "Mozilla/4.0 (compatible; MSIE 5.5; Windows NT)"},
+ )
+ response = urllib.request.urlopen(request) # noqa: S310
+ content = response.read().decode("utf-8")
+ with open(r"cancer_data.csv", "w") as f:
+ f.write(content)
+
+ data = pd.read_csv(
+ "cancer_data.csv",
+ header=None,
+ dtype={0: str}, # Assuming the first column contains string data
+ )
+
+ # 1: pre-processing data
+ del data[data.columns.tolist()[0]]
+ data = data.dropna(axis=0)
+ data = data.replace({"M": np.float64(1), "B": np.float64(-1)})
+ samples = np.array(data)[:, :]
+
+ # 2: dividing data into train_data data and test_data data
+ train_data, test_data = samples[:328, :], samples[328:, :]
+ test_tags, test_samples = test_data[:, 0], test_data[:, 1:]
+
+ # 3: choose kernel function, and set initial alphas to zero (optional)
+ my_kernel = Kernel(kernel="rbf", degree=5, coef0=1, gamma=0.5)
+ al = np.zeros(train_data.shape[0])
+
+ # 4: calculating best alphas using SMO algorithm and predict test_data samples
+ mysvm = SmoSVM(
+ train=train_data,
+ kernel_func=my_kernel,
+ alpha_list=al,
+ cost=0.4,
+ b=0.0,
+ tolerance=0.001,
+ )
+ mysvm.fit()
+ predict = mysvm.predict(test_samples)
+
+ # 5: check accuracy
+ score = 0
+ test_num = test_tags.shape[0]
+ for i in range(test_tags.shape[0]):
+ if test_tags[i] == predict[i]:
+ score += 1
+ print(f"\nAll: {test_num}\nCorrect: {score}\nIncorrect: {test_num - score}")
+ print(f"Rough Accuracy: {score / test_tags.shape[0]}")
+
+
+def test_demonstration():
+ # change stdout
+ print("\nStarting plot, please wait!")
+ sys.stdout = open(os.devnull, "w")
+
+ ax1 = plt.subplot2grid((2, 2), (0, 0))
+ ax2 = plt.subplot2grid((2, 2), (0, 1))
+ ax3 = plt.subplot2grid((2, 2), (1, 0))
+ ax4 = plt.subplot2grid((2, 2), (1, 1))
+ ax1.set_title("Linear SVM, cost = 0.1")
+ test_linear_kernel(ax1, cost=0.1)
+ ax2.set_title("Linear SVM, cost = 500")
+ test_linear_kernel(ax2, cost=500)
+ ax3.set_title("RBF kernel SVM, cost = 0.1")
+ test_rbf_kernel(ax3, cost=0.1)
+ ax4.set_title("RBF kernel SVM, cost = 500")
+ test_rbf_kernel(ax4, cost=500)
+
+ sys.stdout = sys.__stdout__
+ print("Plot done!")
+
+
+def test_linear_kernel(ax, cost):
+ train_x, train_y = make_blobs(
+ n_samples=500, centers=2, n_features=2, random_state=1
+ )
+ train_y[train_y == 0] = -1
+ scaler = StandardScaler()
+ train_x_scaled = scaler.fit_transform(train_x, train_y)
+ train_data = np.hstack((train_y.reshape(500, 1), train_x_scaled))
+ my_kernel = Kernel(kernel="linear", degree=5, coef0=1, gamma=0.5)
+ mysvm = SmoSVM(
+ train=train_data,
+ kernel_func=my_kernel,
+ cost=cost,
+ tolerance=0.001,
+ auto_norm=False,
+ )
+ mysvm.fit()
+ plot_partition_boundary(mysvm, train_data, ax=ax)
+
+
+def test_rbf_kernel(ax, cost):
+ train_x, train_y = make_circles(
+ n_samples=500, noise=0.1, factor=0.1, random_state=1
+ )
+ train_y[train_y == 0] = -1
+ scaler = StandardScaler()
+ train_x_scaled = scaler.fit_transform(train_x, train_y)
+ train_data = np.hstack((train_y.reshape(500, 1), train_x_scaled))
+ my_kernel = Kernel(kernel="rbf", degree=5, coef0=1, gamma=0.5)
+ mysvm = SmoSVM(
+ train=train_data,
+ kernel_func=my_kernel,
+ cost=cost,
+ tolerance=0.001,
+ auto_norm=False,
+ )
+ mysvm.fit()
+ plot_partition_boundary(mysvm, train_data, ax=ax)
+
+
+def plot_partition_boundary(
+ model, train_data, ax, resolution=100, colors=("b", "k", "r")
+):
+ """
+ We cannot get the optimal w of our kernel SVM model, which is different from a
+ linear SVM. For this reason, we generate randomly distributed points with high
+ density, and predicted values of these points are calculated using our trained
+ model. Then we could use this predicted values to draw contour map, and this contour
+ map represents the SVM's partition boundary.
+ """
+ train_data_x = train_data[:, 1]
+ train_data_y = train_data[:, 2]
+ train_data_tags = train_data[:, 0]
+ xrange = np.linspace(train_data_x.min(), train_data_x.max(), resolution)
+ yrange = np.linspace(train_data_y.min(), train_data_y.max(), resolution)
+ test_samples = np.array([(x, y) for x in xrange for y in yrange]).reshape(
+ resolution * resolution, 2
+ )
+
+ test_tags = model.predict(test_samples, classify=False)
+ grid = test_tags.reshape((len(xrange), len(yrange)))
+
+ # Plot contour map which represents the partition boundary
+ ax.contour(
+ xrange,
+ yrange,
+ np.asmatrix(grid).T,
+ levels=(-1, 0, 1),
+ linestyles=("--", "-", "--"),
+ linewidths=(1, 1, 1),
+ colors=colors,
+ )
+ # Plot all train samples
+ ax.scatter(
+ train_data_x,
+ train_data_y,
+ c=train_data_tags,
+ cmap=plt.cm.Dark2,
+ lw=0,
+ alpha=0.5,
+ )
+
+ # Plot support vectors
+ support = model.support
+ ax.scatter(
+ train_data_x[support],
+ train_data_y[support],
+ c=train_data_tags[support],
+ cmap=plt.cm.Dark2,
+ )
+
+
+if __name__ == "__main__":
+ test_cancer_data()
+ test_demonstration()
+ plt.show()
diff --git a/machine_learning/similarity_search.py b/machine_learning/similarity_search.py
index 6bfb12ed88cb..c8a573796882 100644
--- a/machine_learning/similarity_search.py
+++ b/machine_learning/similarity_search.py
@@ -7,9 +7,13 @@
1. the nearest vector
2. distance between the vector and the nearest vector (float)
"""
+
+from __future__ import annotations
+
import math
import numpy as np
+from numpy.linalg import norm
def euclidean(input_a: np.ndarray, input_b: np.ndarray) -> float:
@@ -30,7 +34,9 @@ def euclidean(input_a: np.ndarray, input_b: np.ndarray) -> float:
return math.sqrt(sum(pow(a - b, 2) for a, b in zip(input_a, input_b)))
-def similarity_search(dataset: np.ndarray, value_array: np.ndarray) -> list:
+def similarity_search(
+ dataset: np.ndarray, value_array: np.ndarray
+) -> list[list[list[float] | float]]:
"""
:param dataset: Set containing the vectors. Should be ndarray.
:param value_array: vector/vectors we want to know the nearest vector from dataset.
@@ -66,7 +72,7 @@ def similarity_search(dataset: np.ndarray, value_array: np.ndarray) -> list:
>>> value_array = np.array([1])
>>> similarity_search(dataset, value_array)
Traceback (most recent call last):
- ...
+ ...
ValueError: Wrong input data's dimensions... dataset : 2, value_array : 1
2. If data's shapes are different.
@@ -76,7 +82,7 @@ def similarity_search(dataset: np.ndarray, value_array: np.ndarray) -> list:
>>> value_array = np.array([[0, 0, 0], [0, 0, 1]])
>>> similarity_search(dataset, value_array)
Traceback (most recent call last):
- ...
+ ...
ValueError: Wrong input data's shape... dataset : 2, value_array : 3
3. If data types are different.
@@ -86,32 +92,35 @@ def similarity_search(dataset: np.ndarray, value_array: np.ndarray) -> list:
>>> value_array = np.array([[0, 0], [0, 1]], dtype=np.int32)
>>> similarity_search(dataset, value_array) # doctest: +NORMALIZE_WHITESPACE
Traceback (most recent call last):
- ...
+ ...
TypeError: Input data have different datatype...
dataset : float32, value_array : int32
"""
if dataset.ndim != value_array.ndim:
- raise ValueError(
- f"Wrong input data's dimensions... dataset : {dataset.ndim}, "
- f"value_array : {value_array.ndim}"
+ msg = (
+ "Wrong input data's dimensions... "
+ f"dataset : {dataset.ndim}, value_array : {value_array.ndim}"
)
+ raise ValueError(msg)
try:
if dataset.shape[1] != value_array.shape[1]:
- raise ValueError(
- f"Wrong input data's shape... dataset : {dataset.shape[1]}, "
- f"value_array : {value_array.shape[1]}"
+ msg = (
+ "Wrong input data's shape... "
+ f"dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}"
)
+ raise ValueError(msg)
except IndexError:
if dataset.ndim != value_array.ndim:
raise TypeError("Wrong shape")
if dataset.dtype != value_array.dtype:
- raise TypeError(
- f"Input data have different datatype... dataset : {dataset.dtype}, "
- f"value_array : {value_array.dtype}"
+ msg = (
+ "Input data have different datatype... "
+ f"dataset : {dataset.dtype}, value_array : {value_array.dtype}"
)
+ raise TypeError(msg)
answer = []
@@ -131,6 +140,22 @@ def similarity_search(dataset: np.ndarray, value_array: np.ndarray) -> list:
return answer
+def cosine_similarity(input_a: np.ndarray, input_b: np.ndarray) -> float:
+ """
+ Calculates cosine similarity between two data.
+ :param input_a: ndarray of first vector.
+ :param input_b: ndarray of second vector.
+ :return: Cosine similarity of input_a and input_b. By using math.sqrt(),
+ result will be float.
+
+ >>> cosine_similarity(np.array([1]), np.array([1]))
+ 1.0
+ >>> cosine_similarity(np.array([1, 2]), np.array([6, 32]))
+ 0.9615239476408232
+ """
+ return float(np.dot(input_a, input_b) / (norm(input_a) * norm(input_b)))
+
+
if __name__ == "__main__":
import doctest
diff --git a/machine_learning/support_vector_machines.py b/machine_learning/support_vector_machines.py
index c5e5085d8748..d17c9044a3e9 100644
--- a/machine_learning/support_vector_machines.py
+++ b/machine_learning/support_vector_machines.py
@@ -1,55 +1,203 @@
-from sklearn import svm
-from sklearn.datasets import load_iris
-from sklearn.model_selection import train_test_split
+import numpy as np
+from numpy import ndarray
+from scipy.optimize import Bounds, LinearConstraint, minimize
-# different functions implementing different types of SVM's
-def NuSVC(train_x, train_y):
- svc_NuSVC = svm.NuSVC()
- svc_NuSVC.fit(train_x, train_y)
- return svc_NuSVC
-
+def norm_squared(vector: ndarray) -> float:
+ """
+ Return the squared second norm of vector
+ norm_squared(v) = sum(x * x for x in v)
-def Linearsvc(train_x, train_y):
- svc_linear = svm.LinearSVC(tol=10e-2)
- svc_linear.fit(train_x, train_y)
- return svc_linear
+ Args:
+ vector (ndarray): input vector
+ Returns:
+ float: squared second norm of vector
-def SVC(train_x, train_y):
- # svm.SVC(C=1.0, kernel='rbf', degree=3, gamma=0.0, coef0=0.0, shrinking=True,
- # probability=False,tol=0.001, cache_size=200, class_weight=None, verbose=False,
- # max_iter=1000, random_state=None)
- # various parameters like "kernel","gamma","C" can effectively tuned for a given
- # machine learning model.
- SVC = svm.SVC(gamma="auto")
- SVC.fit(train_x, train_y)
- return SVC
+ >>> int(norm_squared([1, 2]))
+ 5
+ >>> int(norm_squared(np.asarray([1, 2])))
+ 5
+ >>> int(norm_squared([0, 0]))
+ 0
+ """
+ return np.dot(vector, vector)
-def test(X_new):
+class SVC:
"""
- 3 test cases to be passed
- an array containing the sepal length (cm), sepal width (cm), petal length (cm),
- petal width (cm) based on which the target name will be predicted
- >>> test([1,2,1,4])
- 'virginica'
- >>> test([5, 2, 4, 1])
- 'versicolor'
- >>> test([6,3,4,1])
- 'versicolor'
+ Support Vector Classifier
+
+ Args:
+ kernel (str): kernel to use. Default: linear
+ Possible choices:
+ - linear
+ regularization: constraint for soft margin (data not linearly separable)
+ Default: unbound
+
+ >>> SVC(kernel="asdf")
+ Traceback (most recent call last):
+ ...
+ ValueError: Unknown kernel: asdf
+
+ >>> SVC(kernel="rbf")
+ Traceback (most recent call last):
+ ...
+ ValueError: rbf kernel requires gamma
+
+ >>> SVC(kernel="rbf", gamma=-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: gamma must be > 0
"""
- iris = load_iris()
- # splitting the dataset to test and train
- train_x, test_x, train_y, test_y = train_test_split(
- iris["data"], iris["target"], random_state=4
- )
- # any of the 3 types of SVM can be used
- # current_model=SVC(train_x, train_y)
- # current_model=NuSVC(train_x, train_y)
- current_model = Linearsvc(train_x, train_y)
- prediction = current_model.predict([X_new])
- return iris["target_names"][prediction][0]
+
+ def __init__(
+ self,
+ *,
+ regularization: float = np.inf,
+ kernel: str = "linear",
+ gamma: float = 0.0,
+ ) -> None:
+ self.regularization = regularization
+ self.gamma = gamma
+ if kernel == "linear":
+ self.kernel = self.__linear
+ elif kernel == "rbf":
+ if self.gamma == 0:
+ raise ValueError("rbf kernel requires gamma")
+ if not isinstance(self.gamma, (float, int)):
+ raise ValueError("gamma must be float or int")
+ if not self.gamma > 0:
+ raise ValueError("gamma must be > 0")
+ self.kernel = self.__rbf
+ # in the future, there could be a default value like in sklearn
+ # sklear: def_gamma = 1/(n_features * X.var()) (wiki)
+ # previously it was 1/(n_features)
+ else:
+ msg = f"Unknown kernel: {kernel}"
+ raise ValueError(msg)
+
+ # kernels
+ def __linear(self, vector1: ndarray, vector2: ndarray) -> float:
+ """Linear kernel (as if no kernel used at all)"""
+ return np.dot(vector1, vector2)
+
+ def __rbf(self, vector1: ndarray, vector2: ndarray) -> float:
+ """
+ RBF: Radial Basis Function Kernel
+
+ Note: for more information see:
+ https://en.wikipedia.org/wiki/Radial_basis_function_kernel
+
+ Args:
+ vector1 (ndarray): first vector
+ vector2 (ndarray): second vector)
+
+ Returns:
+ float: exp(-(gamma * norm_squared(vector1 - vector2)))
+ """
+ return np.exp(-(self.gamma * norm_squared(vector1 - vector2)))
+
+ def fit(self, observations: list[ndarray], classes: ndarray) -> None:
+ """
+ Fits the SVC with a set of observations.
+
+ Args:
+ observations (list[ndarray]): list of observations
+ classes (ndarray): classification of each observation (in {1, -1})
+ """
+
+ self.observations = observations
+ self.classes = classes
+
+ # using Wolfe's Dual to calculate w.
+ # Primal problem: minimize 1/2*norm_squared(w)
+ # constraint: yn(w . xn + b) >= 1
+ #
+ # With l a vector
+ # Dual problem: maximize sum_n(ln) -
+ # 1/2 * sum_n(sum_m(ln*lm*yn*ym*xn . xm))
+ # constraint: self.C >= ln >= 0
+ # and sum_n(ln*yn) = 0
+ # Then we get w using w = sum_n(ln*yn*xn)
+ # At the end we can get b ~= mean(yn - w . xn)
+ #
+ # Since we use kernels, we only need l_star to calculate b
+ # and to classify observations
+
+ (n,) = np.shape(classes)
+
+ def to_minimize(candidate: ndarray) -> float:
+ """
+ Opposite of the function to maximize
+
+ Args:
+ candidate (ndarray): candidate array to test
+
+ Return:
+ float: Wolfe's Dual result to minimize
+ """
+ s = 0
+ (n,) = np.shape(candidate)
+ for i in range(n):
+ for j in range(n):
+ s += (
+ candidate[i]
+ * candidate[j]
+ * classes[i]
+ * classes[j]
+ * self.kernel(observations[i], observations[j])
+ )
+ return 1 / 2 * s - sum(candidate)
+
+ ly_contraint = LinearConstraint(classes, 0, 0)
+ l_bounds = Bounds(0, self.regularization)
+
+ l_star = minimize(
+ to_minimize, np.ones(n), bounds=l_bounds, constraints=[ly_contraint]
+ ).x
+ self.optimum = l_star
+
+ # calculating mean offset of separation plane to points
+ s = 0
+ for i in range(n):
+ for j in range(n):
+ s += classes[i] - classes[i] * self.optimum[i] * self.kernel(
+ observations[i], observations[j]
+ )
+ self.offset = s / n
+
+ def predict(self, observation: ndarray) -> int:
+ """
+ Get the expected class of an observation
+
+ Args:
+ observation (Vector): observation
+
+ Returns:
+ int {1, -1}: expected class
+
+ >>> xs = [
+ ... np.asarray([0, 1]), np.asarray([0, 2]),
+ ... np.asarray([1, 1]), np.asarray([1, 2])
+ ... ]
+ >>> y = np.asarray([1, 1, -1, -1])
+ >>> s = SVC()
+ >>> s.fit(xs, y)
+ >>> s.predict(np.asarray([0, 1]))
+ 1
+ >>> s.predict(np.asarray([1, 1]))
+ -1
+ >>> s.predict(np.asarray([2, 2]))
+ -1
+ """
+ s = sum(
+ self.optimum[n]
+ * self.classes[n]
+ * self.kernel(self.observations[n], observation)
+ for n in range(len(self.classes))
+ )
+ return 1 if s + self.offset >= 0 else -1
if __name__ == "__main__":
diff --git a/machine_learning/word_frequency_functions.py b/machine_learning/word_frequency_functions.py
index 9cf7b694c6be..8fd2741f611c 100644
--- a/machine_learning/word_frequency_functions.py
+++ b/machine_learning/word_frequency_functions.py
@@ -61,7 +61,7 @@ def term_frequency(term: str, document: str) -> int:
return len([word for word in tokenize_document if word.lower() == term.lower()])
-def document_frequency(term: str, corpus: str) -> int:
+def document_frequency(term: str, corpus: str) -> tuple[int, int]:
"""
Calculate the number of documents in a corpus that contain a
given term
@@ -83,7 +83,7 @@ def document_frequency(term: str, corpus: str) -> int:
return (len([doc for doc in docs if term in doc]), len(docs))
-def inverse_document_frequency(df: int, N: int, smoothing=False) -> float:
+def inverse_document_frequency(df: int, n: int, smoothing=False) -> float:
"""
Return an integer denoting the importance
of a word. This measure of importance is
@@ -109,15 +109,15 @@ def inverse_document_frequency(df: int, N: int, smoothing=False) -> float:
1.477
"""
if smoothing:
- if N == 0:
+ if n == 0:
raise ValueError("log10(0) is undefined.")
- return round(1 + log10(N / (1 + df)), 3)
+ return round(1 + log10(n / (1 + df)), 3)
if df == 0:
raise ZeroDivisionError("df must be > 0")
- elif N == 0:
+ elif n == 0:
raise ValueError("log10(0) is undefined.")
- return round(log10(N / df), 3)
+ return round(log10(n / df), 3)
def tf_idf(tf: int, idf: int) -> float:
diff --git a/machine_learning/xgboost_classifier.py b/machine_learning/xgboost_classifier.py
new file mode 100644
index 000000000000..1da933cf690f
--- /dev/null
+++ b/machine_learning/xgboost_classifier.py
@@ -0,0 +1,81 @@
+# XGBoost Classifier Example
+import numpy as np
+from matplotlib import pyplot as plt
+from sklearn.datasets import load_iris
+from sklearn.metrics import ConfusionMatrixDisplay
+from sklearn.model_selection import train_test_split
+from xgboost import XGBClassifier
+
+
+def data_handling(data: dict) -> tuple:
+ # Split dataset into features and target
+ # data is features
+ """
+ >>> data_handling(({'data':'[5.1, 3.5, 1.4, 0.2]','target':([0])}))
+ ('[5.1, 3.5, 1.4, 0.2]', [0])
+ >>> data_handling(
+ ... {'data': '[4.9, 3.0, 1.4, 0.2], [4.7, 3.2, 1.3, 0.2]', 'target': ([0, 0])}
+ ... )
+ ('[4.9, 3.0, 1.4, 0.2], [4.7, 3.2, 1.3, 0.2]', [0, 0])
+ """
+ return (data["data"], data["target"])
+
+
+def xgboost(features: np.ndarray, target: np.ndarray) -> XGBClassifier:
+ """
+ # THIS TEST IS BROKEN!! >>> xgboost(np.array([[5.1, 3.6, 1.4, 0.2]]), np.array([0]))
+ XGBClassifier(base_score=0.5, booster='gbtree', callbacks=None,
+ colsample_bylevel=1, colsample_bynode=1, colsample_bytree=1,
+ early_stopping_rounds=None, enable_categorical=False,
+ eval_metric=None, gamma=0, gpu_id=-1, grow_policy='depthwise',
+ importance_type=None, interaction_constraints='',
+ learning_rate=0.300000012, max_bin=256, max_cat_to_onehot=4,
+ max_delta_step=0, max_depth=6, max_leaves=0, min_child_weight=1,
+ missing=nan, monotone_constraints='()', n_estimators=100,
+ n_jobs=0, num_parallel_tree=1, predictor='auto', random_state=0,
+ reg_alpha=0, reg_lambda=1, ...)
+ """
+ classifier = XGBClassifier()
+ classifier.fit(features, target)
+ return classifier
+
+
+def main() -> None:
+ """
+ >>> main()
+
+ Url for the algorithm:
+ https://xgboost.readthedocs.io/en/stable/
+ Iris type dataset is used to demonstrate algorithm.
+ """
+
+ # Load Iris dataset
+ iris = load_iris()
+ features, targets = data_handling(iris)
+ x_train, x_test, y_train, y_test = train_test_split(
+ features, targets, test_size=0.25
+ )
+
+ names = iris["target_names"]
+
+ # Create an XGBoost Classifier from the training data
+ xgboost_classifier = xgboost(x_train, y_train)
+
+ # Display the confusion matrix of the classifier with both training and test sets
+ ConfusionMatrixDisplay.from_estimator(
+ xgboost_classifier,
+ x_test,
+ y_test,
+ display_labels=names,
+ cmap="Blues",
+ normalize="true",
+ )
+ plt.title("Normalized Confusion Matrix - IRIS Dataset")
+ plt.show()
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod(verbose=True)
+ main()
diff --git a/machine_learning/xgboost_regressor.py b/machine_learning/xgboost_regressor.py
new file mode 100644
index 000000000000..52e041c55ea2
--- /dev/null
+++ b/machine_learning/xgboost_regressor.py
@@ -0,0 +1,66 @@
+# XGBoost Regressor Example
+import numpy as np
+from sklearn.datasets import fetch_california_housing
+from sklearn.metrics import mean_absolute_error, mean_squared_error
+from sklearn.model_selection import train_test_split
+from xgboost import XGBRegressor
+
+
+def data_handling(data: dict) -> tuple:
+ # Split dataset into features and target. Data is features.
+ """
+ >>> data_handling((
+ ... {'data':'[ 8.3252 41. 6.9841269 1.02380952 322. 2.55555556 37.88 -122.23 ]'
+ ... ,'target':([4.526])}))
+ ('[ 8.3252 41. 6.9841269 1.02380952 322. 2.55555556 37.88 -122.23 ]', [4.526])
+ """
+ return (data["data"], data["target"])
+
+
+def xgboost(
+ features: np.ndarray, target: np.ndarray, test_features: np.ndarray
+) -> np.ndarray:
+ """
+ >>> xgboost(np.array([[ 2.3571 , 52. , 6.00813008, 1.06775068,
+ ... 907. , 2.45799458, 40.58 , -124.26]]),np.array([1.114]),
+ ... np.array([[1.97840000e+00, 3.70000000e+01, 4.98858447e+00, 1.03881279e+00,
+ ... 1.14300000e+03, 2.60958904e+00, 3.67800000e+01, -1.19780000e+02]]))
+ array([[1.1139996]], dtype=float32)
+ """
+ xgb = XGBRegressor(
+ verbosity=0, random_state=42, tree_method="exact", base_score=0.5
+ )
+ xgb.fit(features, target)
+ # Predict target for test data
+ predictions = xgb.predict(test_features)
+ predictions = predictions.reshape(len(predictions), 1)
+ return predictions
+
+
+def main() -> None:
+ """
+ The URL for this algorithm
+ https://xgboost.readthedocs.io/en/stable/
+ California house price dataset is used to demonstrate the algorithm.
+
+ Expected error values:
+ Mean Absolute Error: 0.30957163379906033
+ Mean Square Error: 0.22611560196662744
+ """
+ # Load California house price dataset
+ california = fetch_california_housing()
+ data, target = data_handling(california)
+ x_train, x_test, y_train, y_test = train_test_split(
+ data, target, test_size=0.25, random_state=1
+ )
+ predictions = xgboost(x_train, y_train, x_test)
+ # Error printing
+ print(f"Mean Absolute Error: {mean_absolute_error(y_test, predictions)}")
+ print(f"Mean Square Error: {mean_squared_error(y_test, predictions)}")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod(verbose=True)
+ main()
diff --git a/maths/3n_plus_1.py b/maths/3n_plus_1.py
deleted file mode 100644
index 28c9fd7b426f..000000000000
--- a/maths/3n_plus_1.py
+++ /dev/null
@@ -1,149 +0,0 @@
-from __future__ import annotations
-
-
-def n31(a: int) -> tuple[list[int], int]:
- """
- Returns the Collatz sequence and its length of any positive integer.
- >>> n31(4)
- ([4, 2, 1], 3)
- """
-
- if not isinstance(a, int):
- raise TypeError("Must be int, not {}".format(type(a).__name__))
- if a < 1:
- raise ValueError(f"Given integer must be greater than 1, not {a}")
-
- path = [a]
- while a != 1:
- if a % 2 == 0:
- a = a // 2
- else:
- a = 3 * a + 1
- path += [a]
- return path, len(path)
-
-
-def test_n31():
- """
- >>> test_n31()
- """
- assert n31(4) == ([4, 2, 1], 3)
- assert n31(11) == ([11, 34, 17, 52, 26, 13, 40, 20, 10, 5, 16, 8, 4, 2, 1], 15)
- assert n31(31) == (
- [
- 31,
- 94,
- 47,
- 142,
- 71,
- 214,
- 107,
- 322,
- 161,
- 484,
- 242,
- 121,
- 364,
- 182,
- 91,
- 274,
- 137,
- 412,
- 206,
- 103,
- 310,
- 155,
- 466,
- 233,
- 700,
- 350,
- 175,
- 526,
- 263,
- 790,
- 395,
- 1186,
- 593,
- 1780,
- 890,
- 445,
- 1336,
- 668,
- 334,
- 167,
- 502,
- 251,
- 754,
- 377,
- 1132,
- 566,
- 283,
- 850,
- 425,
- 1276,
- 638,
- 319,
- 958,
- 479,
- 1438,
- 719,
- 2158,
- 1079,
- 3238,
- 1619,
- 4858,
- 2429,
- 7288,
- 3644,
- 1822,
- 911,
- 2734,
- 1367,
- 4102,
- 2051,
- 6154,
- 3077,
- 9232,
- 4616,
- 2308,
- 1154,
- 577,
- 1732,
- 866,
- 433,
- 1300,
- 650,
- 325,
- 976,
- 488,
- 244,
- 122,
- 61,
- 184,
- 92,
- 46,
- 23,
- 70,
- 35,
- 106,
- 53,
- 160,
- 80,
- 40,
- 20,
- 10,
- 5,
- 16,
- 8,
- 4,
- 2,
- 1,
- ],
- 107,
- )
-
-
-if __name__ == "__main__":
- num = 4
- path, length = n31(num)
- print(f"The Collatz sequence of {num} took {length} steps. \nPath: {path}")
diff --git a/maths/abs.py b/maths/abs.py
index 68c99a1d51d8..b357e98d8680 100644
--- a/maths/abs.py
+++ b/maths/abs.py
@@ -1,7 +1,7 @@
"""Absolute Value."""
-def abs_val(num):
+def abs_val(num: float) -> float:
"""
Find the absolute value of a number.
@@ -15,14 +15,80 @@ def abs_val(num):
return -num if num < 0 else num
+def abs_min(x: list[int]) -> int:
+ """
+ >>> abs_min([0,5,1,11])
+ 0
+ >>> abs_min([3,-10,-2])
+ -2
+ >>> abs_min([])
+ Traceback (most recent call last):
+ ...
+ ValueError: abs_min() arg is an empty sequence
+ """
+ if len(x) == 0:
+ raise ValueError("abs_min() arg is an empty sequence")
+ j = x[0]
+ for i in x:
+ if abs_val(i) < abs_val(j):
+ j = i
+ return j
+
+
+def abs_max(x: list[int]) -> int:
+ """
+ >>> abs_max([0,5,1,11])
+ 11
+ >>> abs_max([3,-10,-2])
+ -10
+ >>> abs_max([])
+ Traceback (most recent call last):
+ ...
+ ValueError: abs_max() arg is an empty sequence
+ """
+ if len(x) == 0:
+ raise ValueError("abs_max() arg is an empty sequence")
+ j = x[0]
+ for i in x:
+ if abs(i) > abs(j):
+ j = i
+ return j
+
+
+def abs_max_sort(x: list[int]) -> int:
+ """
+ >>> abs_max_sort([0,5,1,11])
+ 11
+ >>> abs_max_sort([3,-10,-2])
+ -10
+ >>> abs_max_sort([])
+ Traceback (most recent call last):
+ ...
+ ValueError: abs_max_sort() arg is an empty sequence
+ """
+ if len(x) == 0:
+ raise ValueError("abs_max_sort() arg is an empty sequence")
+ return sorted(x, key=abs)[-1]
+
+
def test_abs_val():
"""
>>> test_abs_val()
"""
- assert 0 == abs_val(0)
- assert 34 == abs_val(34)
- assert 100000000000 == abs_val(-100000000000)
+ assert abs_val(0) == 0
+ assert abs_val(34) == 34
+ assert abs_val(-100000000000) == 100000000000
+
+ a = [-3, -1, 2, -11]
+ assert abs_max(a) == -11
+ assert abs_max_sort(a) == -11
+ assert abs_min(a) == -1
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ test_abs_val()
print(abs_val(-34)) # --> 34
diff --git a/maths/abs_max.py b/maths/abs_max.py
deleted file mode 100644
index e5a8219657ac..000000000000
--- a/maths/abs_max.py
+++ /dev/null
@@ -1,35 +0,0 @@
-from __future__ import annotations
-
-
-def abs_max(x: list[int]) -> int:
- """
- >>> abs_max([0,5,1,11])
- 11
- >>> abs_max([3,-10,-2])
- -10
- """
- j = x[0]
- for i in x:
- if abs(i) > abs(j):
- j = i
- return j
-
-
-def abs_max_sort(x):
- """
- >>> abs_max_sort([0,5,1,11])
- 11
- >>> abs_max_sort([3,-10,-2])
- -10
- """
- return sorted(x, key=abs)[-1]
-
-
-def main():
- a = [1, 2, -11]
- assert abs_max(a) == -11
- assert abs_max_sort(a) == -11
-
-
-if __name__ == "__main__":
- main()
diff --git a/maths/abs_min.py b/maths/abs_min.py
deleted file mode 100644
index eb84de37ce23..000000000000
--- a/maths/abs_min.py
+++ /dev/null
@@ -1,24 +0,0 @@
-from .abs import abs_val
-
-
-def absMin(x):
- """
- >>> absMin([0,5,1,11])
- 0
- >>> absMin([3,-10,-2])
- -2
- """
- j = x[0]
- for i in x:
- if abs_val(i) < abs_val(j):
- j = i
- return j
-
-
-def main():
- a = [-3, -1, 2, -11]
- print(absMin(a)) # = -1
-
-
-if __name__ == "__main__":
- main()
diff --git a/maths/add.py b/maths/add.py
deleted file mode 100644
index 0bc7da9697d3..000000000000
--- a/maths/add.py
+++ /dev/null
@@ -1,19 +0,0 @@
-"""
-Just to check
-"""
-
-
-def add(a, b):
- """
- >>> add(2, 2)
- 4
- >>> add(2, -2)
- 0
- """
- return a + b
-
-
-if __name__ == "__main__":
- a = 5
- b = 6
- print(f"The sum of {a} + {b} is {add(a, b)}")
diff --git a/maths/addition_without_arithmetic.py b/maths/addition_without_arithmetic.py
new file mode 100644
index 000000000000..409604e4c08a
--- /dev/null
+++ b/maths/addition_without_arithmetic.py
@@ -0,0 +1,39 @@
+"""
+Illustrate how to add the integer without arithmetic operation
+Author: suraj Kumar
+Time Complexity: 1
+https://en.wikipedia.org/wiki/Bitwise_operation
+"""
+
+
+def add(first: int, second: int) -> int:
+ """
+ Implementation of addition of integer
+
+ Examples:
+ >>> add(3, 5)
+ 8
+ >>> add(13, 5)
+ 18
+ >>> add(-7, 2)
+ -5
+ >>> add(0, -7)
+ -7
+ >>> add(-321, 0)
+ -321
+ """
+ while second != 0:
+ c = first & second
+ first ^= second
+ second = c << 1
+ return first
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ first = int(input("Enter the first number: ").strip())
+ second = int(input("Enter the second number: ").strip())
+ print(f"{add(first, second) = }")
diff --git a/maths/allocation_number.py b/maths/allocation_number.py
index d419e74d01ff..52f1ac4bdb23 100644
--- a/maths/allocation_number.py
+++ b/maths/allocation_number.py
@@ -5,6 +5,7 @@
for i in allocation_list:
requests.get(url,headers={'Range':f'bytes={i}'})
"""
+
from __future__ import annotations
diff --git a/maths/arc_length.py b/maths/arc_length.py
new file mode 100644
index 000000000000..4c518f321dc7
--- /dev/null
+++ b/maths/arc_length.py
@@ -0,0 +1,17 @@
+from math import pi
+
+
+def arc_length(angle: int, radius: int) -> float:
+ """
+ >>> arc_length(45, 5)
+ 3.9269908169872414
+ >>> arc_length(120, 15)
+ 31.415926535897928
+ >>> arc_length(90, 10)
+ 15.707963267948966
+ """
+ return 2 * pi * radius * (angle / 360)
+
+
+if __name__ == "__main__":
+ print(arc_length(90, 10))
diff --git a/maths/area.py b/maths/area.py
index 24216e223ebf..31a654206977 100644
--- a/maths/area.py
+++ b/maths/area.py
@@ -1,7 +1,9 @@
"""
Find the area of various geometric shapes
+Wikipedia reference: https://en.wikipedia.org/wiki/Area
"""
-from math import pi
+
+from math import pi, sqrt, tan
def surface_area_cube(side_length: float) -> float:
@@ -10,6 +12,10 @@ def surface_area_cube(side_length: float) -> float:
>>> surface_area_cube(1)
6
+ >>> surface_area_cube(1.6)
+ 15.360000000000003
+ >>> surface_area_cube(0)
+ 0
>>> surface_area_cube(3)
54
>>> surface_area_cube(-1)
@@ -19,19 +25,51 @@ def surface_area_cube(side_length: float) -> float:
"""
if side_length < 0:
raise ValueError("surface_area_cube() only accepts non-negative values")
- return 6 * side_length ** 2
+ return 6 * side_length**2
+
+
+def surface_area_cuboid(length: float, breadth: float, height: float) -> float:
+ """
+ Calculate the Surface Area of a Cuboid.
+
+ >>> surface_area_cuboid(1, 2, 3)
+ 22
+ >>> surface_area_cuboid(0, 0, 0)
+ 0
+ >>> surface_area_cuboid(1.6, 2.6, 3.6)
+ 38.56
+ >>> surface_area_cuboid(-1, 2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cuboid() only accepts non-negative values
+ >>> surface_area_cuboid(1, -2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cuboid() only accepts non-negative values
+ >>> surface_area_cuboid(1, 2, -3)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cuboid() only accepts non-negative values
+ """
+ if length < 0 or breadth < 0 or height < 0:
+ raise ValueError("surface_area_cuboid() only accepts non-negative values")
+ return 2 * ((length * breadth) + (breadth * height) + (length * height))
def surface_area_sphere(radius: float) -> float:
"""
Calculate the Surface Area of a Sphere.
Wikipedia reference: https://en.wikipedia.org/wiki/Sphere
- :return 4 * pi * r^2
+ Formula: 4 * pi * r^2
>>> surface_area_sphere(5)
314.1592653589793
>>> surface_area_sphere(1)
12.566370614359172
+ >>> surface_area_sphere(1.6)
+ 32.169908772759484
+ >>> surface_area_sphere(0)
+ 0.0
>>> surface_area_sphere(-1)
Traceback (most recent call last):
...
@@ -39,15 +77,175 @@ def surface_area_sphere(radius: float) -> float:
"""
if radius < 0:
raise ValueError("surface_area_sphere() only accepts non-negative values")
- return 4 * pi * radius ** 2
+ return 4 * pi * radius**2
+
+
+def surface_area_hemisphere(radius: float) -> float:
+ """
+ Calculate the Surface Area of a Hemisphere.
+ Formula: 3 * pi * r^2
+
+ >>> surface_area_hemisphere(5)
+ 235.61944901923448
+ >>> surface_area_hemisphere(1)
+ 9.42477796076938
+ >>> surface_area_hemisphere(0)
+ 0.0
+ >>> surface_area_hemisphere(1.1)
+ 11.40398133253095
+ >>> surface_area_hemisphere(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_hemisphere() only accepts non-negative values
+ """
+ if radius < 0:
+ raise ValueError("surface_area_hemisphere() only accepts non-negative values")
+ return 3 * pi * radius**2
+
+
+def surface_area_cone(radius: float, height: float) -> float:
+ """
+ Calculate the Surface Area of a Cone.
+ Wikipedia reference: https://en.wikipedia.org/wiki/Cone
+ Formula: pi * r * (r + (h ** 2 + r ** 2) ** 0.5)
+
+ >>> surface_area_cone(10, 24)
+ 1130.9733552923256
+ >>> surface_area_cone(6, 8)
+ 301.59289474462014
+ >>> surface_area_cone(1.6, 2.6)
+ 23.387862992395807
+ >>> surface_area_cone(0, 0)
+ 0.0
+ >>> surface_area_cone(-1, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cone() only accepts non-negative values
+ >>> surface_area_cone(1, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cone() only accepts non-negative values
+ >>> surface_area_cone(-1, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cone() only accepts non-negative values
+ """
+ if radius < 0 or height < 0:
+ raise ValueError("surface_area_cone() only accepts non-negative values")
+ return pi * radius * (radius + (height**2 + radius**2) ** 0.5)
+
+
+def surface_area_conical_frustum(
+ radius_1: float, radius_2: float, height: float
+) -> float:
+ """
+ Calculate the Surface Area of a Conical Frustum.
+
+ >>> surface_area_conical_frustum(1, 2, 3)
+ 45.511728065337266
+ >>> surface_area_conical_frustum(4, 5, 6)
+ 300.7913575056268
+ >>> surface_area_conical_frustum(0, 0, 0)
+ 0.0
+ >>> surface_area_conical_frustum(1.6, 2.6, 3.6)
+ 78.57907060751548
+ >>> surface_area_conical_frustum(-1, 2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_conical_frustum() only accepts non-negative values
+ >>> surface_area_conical_frustum(1, -2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_conical_frustum() only accepts non-negative values
+ >>> surface_area_conical_frustum(1, 2, -3)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_conical_frustum() only accepts non-negative values
+ """
+ if radius_1 < 0 or radius_2 < 0 or height < 0:
+ raise ValueError(
+ "surface_area_conical_frustum() only accepts non-negative values"
+ )
+ slant_height = (height**2 + (radius_1 - radius_2) ** 2) ** 0.5
+ return pi * ((slant_height * (radius_1 + radius_2)) + radius_1**2 + radius_2**2)
+
+
+def surface_area_cylinder(radius: float, height: float) -> float:
+ """
+ Calculate the Surface Area of a Cylinder.
+ Wikipedia reference: https://en.wikipedia.org/wiki/Cylinder
+ Formula: 2 * pi * r * (h + r)
+
+ >>> surface_area_cylinder(7, 10)
+ 747.6990515543707
+ >>> surface_area_cylinder(1.6, 2.6)
+ 42.22300526424682
+ >>> surface_area_cylinder(0, 0)
+ 0.0
+ >>> surface_area_cylinder(6, 8)
+ 527.7875658030853
+ >>> surface_area_cylinder(-1, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cylinder() only accepts non-negative values
+ >>> surface_area_cylinder(1, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cylinder() only accepts non-negative values
+ >>> surface_area_cylinder(-1, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_cylinder() only accepts non-negative values
+ """
+ if radius < 0 or height < 0:
+ raise ValueError("surface_area_cylinder() only accepts non-negative values")
+ return 2 * pi * radius * (height + radius)
+
+
+def surface_area_torus(torus_radius: float, tube_radius: float) -> float:
+ """Calculate the Area of a Torus.
+ Wikipedia reference: https://en.wikipedia.org/wiki/Torus
+ :return 4pi^2 * torus_radius * tube_radius
+ >>> surface_area_torus(1, 1)
+ 39.47841760435743
+ >>> surface_area_torus(4, 3)
+ 473.7410112522892
+ >>> surface_area_torus(3, 4)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_torus() does not support spindle or self intersecting tori
+ >>> surface_area_torus(1.6, 1.6)
+ 101.06474906715503
+ >>> surface_area_torus(0, 0)
+ 0.0
+ >>> surface_area_torus(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_torus() only accepts non-negative values
+ >>> surface_area_torus(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: surface_area_torus() only accepts non-negative values
+ """
+ if torus_radius < 0 or tube_radius < 0:
+ raise ValueError("surface_area_torus() only accepts non-negative values")
+ if torus_radius < tube_radius:
+ raise ValueError(
+ "surface_area_torus() does not support spindle or self intersecting tori"
+ )
+ return 4 * pow(pi, 2) * torus_radius * tube_radius
def area_rectangle(length: float, width: float) -> float:
"""
- Calculate the area of a rectangle
+ Calculate the area of a rectangle.
>>> area_rectangle(10, 20)
200
+ >>> area_rectangle(1.6, 2.6)
+ 4.16
+ >>> area_rectangle(0, 0)
+ 0
>>> area_rectangle(-1, -2)
Traceback (most recent call last):
...
@@ -68,10 +266,14 @@ def area_rectangle(length: float, width: float) -> float:
def area_square(side_length: float) -> float:
"""
- Calculate the area of a square
+ Calculate the area of a square.
>>> area_square(10)
100
+ >>> area_square(0)
+ 0
+ >>> area_square(1.6)
+ 2.5600000000000005
>>> area_square(-1)
Traceback (most recent call last):
...
@@ -79,15 +281,19 @@ def area_square(side_length: float) -> float:
"""
if side_length < 0:
raise ValueError("area_square() only accepts non-negative values")
- return side_length ** 2
+ return side_length**2
def area_triangle(base: float, height: float) -> float:
"""
- Calculate the area of a triangle
+ Calculate the area of a triangle given the base and height.
>>> area_triangle(10, 10)
50.0
+ >>> area_triangle(1.6, 2.6)
+ 2.08
+ >>> area_triangle(0, 0)
+ 0.0
>>> area_triangle(-1, -2)
Traceback (most recent call last):
...
@@ -106,12 +312,64 @@ def area_triangle(base: float, height: float) -> float:
return (base * height) / 2
+def area_triangle_three_sides(side1: float, side2: float, side3: float) -> float:
+ """
+ Calculate area of triangle when the length of 3 sides are known.
+ This function uses Heron's formula: https://en.wikipedia.org/wiki/Heron%27s_formula
+
+ >>> area_triangle_three_sides(5, 12, 13)
+ 30.0
+ >>> area_triangle_three_sides(10, 11, 12)
+ 51.521233486786784
+ >>> area_triangle_three_sides(0, 0, 0)
+ 0.0
+ >>> area_triangle_three_sides(1.6, 2.6, 3.6)
+ 1.8703742940919619
+ >>> area_triangle_three_sides(-1, -2, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: area_triangle_three_sides() only accepts non-negative values
+ >>> area_triangle_three_sides(1, -2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: area_triangle_three_sides() only accepts non-negative values
+ >>> area_triangle_three_sides(2, 4, 7)
+ Traceback (most recent call last):
+ ...
+ ValueError: Given three sides do not form a triangle
+ >>> area_triangle_three_sides(2, 7, 4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Given three sides do not form a triangle
+ >>> area_triangle_three_sides(7, 2, 4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Given three sides do not form a triangle
+ """
+ if side1 < 0 or side2 < 0 or side3 < 0:
+ raise ValueError("area_triangle_three_sides() only accepts non-negative values")
+ elif side1 + side2 < side3 or side1 + side3 < side2 or side2 + side3 < side1:
+ raise ValueError("Given three sides do not form a triangle")
+ semi_perimeter = (side1 + side2 + side3) / 2
+ area = sqrt(
+ semi_perimeter
+ * (semi_perimeter - side1)
+ * (semi_perimeter - side2)
+ * (semi_perimeter - side3)
+ )
+ return area
+
+
def area_parallelogram(base: float, height: float) -> float:
"""
- Calculate the area of a parallelogram
+ Calculate the area of a parallelogram.
>>> area_parallelogram(10, 20)
200
+ >>> area_parallelogram(1.6, 2.6)
+ 4.16
+ >>> area_parallelogram(0, 0)
+ 0
>>> area_parallelogram(-1, -2)
Traceback (most recent call last):
...
@@ -132,10 +390,14 @@ def area_parallelogram(base: float, height: float) -> float:
def area_trapezium(base1: float, base2: float, height: float) -> float:
"""
- Calculate the area of a trapezium
+ Calculate the area of a trapezium.
>>> area_trapezium(10, 20, 30)
450.0
+ >>> area_trapezium(1.6, 2.6, 3.6)
+ 7.5600000000000005
+ >>> area_trapezium(0, 0, 0)
+ 0.0
>>> area_trapezium(-1, -2, -3)
Traceback (most recent call last):
...
@@ -172,10 +434,14 @@ def area_trapezium(base1: float, base2: float, height: float) -> float:
def area_circle(radius: float) -> float:
"""
- Calculate the area of a circle
+ Calculate the area of a circle.
>>> area_circle(20)
1256.6370614359173
+ >>> area_circle(1.6)
+ 8.042477193189871
+ >>> area_circle(0)
+ 0.0
>>> area_circle(-1)
Traceback (most recent call last):
...
@@ -183,17 +449,21 @@ def area_circle(radius: float) -> float:
"""
if radius < 0:
raise ValueError("area_circle() only accepts non-negative values")
- return pi * radius ** 2
+ return pi * radius**2
def area_ellipse(radius_x: float, radius_y: float) -> float:
"""
- Calculate the area of a ellipse
+ Calculate the area of a ellipse.
>>> area_ellipse(10, 10)
314.1592653589793
>>> area_ellipse(10, 20)
628.3185307179587
+ >>> area_ellipse(0, 0)
+ 0.0
+ >>> area_ellipse(1.6, 2.6)
+ 13.06902543893354
>>> area_ellipse(-10, 20)
Traceback (most recent call last):
...
@@ -214,10 +484,14 @@ def area_ellipse(radius_x: float, radius_y: float) -> float:
def area_rhombus(diagonal_1: float, diagonal_2: float) -> float:
"""
- Calculate the area of a rhombus
+ Calculate the area of a rhombus.
>>> area_rhombus(10, 20)
100.0
+ >>> area_rhombus(1.6, 2.6)
+ 2.08
+ >>> area_rhombus(0, 0)
+ 0.0
>>> area_rhombus(-1, -2)
Traceback (most recent call last):
...
@@ -236,24 +510,75 @@ def area_rhombus(diagonal_1: float, diagonal_2: float) -> float:
return 1 / 2 * diagonal_1 * diagonal_2
-def main():
- print("Areas of various geometric shapes: \n")
- print(f"Rectangle: {area_rectangle(10, 20)}")
- print(f"Square: {area_square(10)}")
- print(f"Triangle: {area_triangle(10, 10)}")
- print(f"Parallelogram: {area_parallelogram(10, 20)}")
- print(f"Trapezium: {area_trapezium(10, 20, 30)}")
- print(f"Circle: {area_circle(20)}")
- print("\nSurface Areas of various geometric shapes: \n")
- print(f"Cube: {surface_area_cube(20)}")
- print(f"Sphere: {surface_area_sphere(20)}")
- print(f"Rhombus: {area_rhombus(10, 20)}")
+def area_reg_polygon(sides: int, length: float) -> float:
+ """
+ Calculate the area of a regular polygon.
+ Wikipedia reference: https://en.wikipedia.org/wiki/Polygon#Regular_polygons
+ Formula: (n*s^2*cot(pi/n))/4
+
+ >>> area_reg_polygon(3, 10)
+ 43.301270189221945
+ >>> area_reg_polygon(4, 10)
+ 100.00000000000001
+ >>> area_reg_polygon(0, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: area_reg_polygon() only accepts integers greater than or equal to \
+three as number of sides
+ >>> area_reg_polygon(-1, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: area_reg_polygon() only accepts integers greater than or equal to \
+three as number of sides
+ >>> area_reg_polygon(5, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: area_reg_polygon() only accepts non-negative values as \
+length of a side
+ >>> area_reg_polygon(-1, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: area_reg_polygon() only accepts integers greater than or equal to \
+three as number of sides
+ """
+ if not isinstance(sides, int) or sides < 3:
+ raise ValueError(
+ "area_reg_polygon() only accepts integers greater than or \
+equal to three as number of sides"
+ )
+ elif length < 0:
+ raise ValueError(
+ "area_reg_polygon() only accepts non-negative values as \
+length of a side"
+ )
+ return (sides * length**2) / (4 * tan(pi / sides))
+ return (sides * length**2) / (4 * tan(pi / sides))
if __name__ == "__main__":
-
import doctest
doctest.testmod(verbose=True) # verbose so we can see methods missing tests
- main()
+ print("[DEMO] Areas of various geometric shapes: \n")
+ print(f"Rectangle: {area_rectangle(10, 20) = }")
+ print(f"Square: {area_square(10) = }")
+ print(f"Triangle: {area_triangle(10, 10) = }")
+ print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }")
+ print(f"Parallelogram: {area_parallelogram(10, 20) = }")
+ print(f"Rhombus: {area_rhombus(10, 20) = }")
+ print(f"Trapezium: {area_trapezium(10, 20, 30) = }")
+ print(f"Circle: {area_circle(20) = }")
+ print(f"Ellipse: {area_ellipse(10, 20) = }")
+ print("\nSurface Areas of various geometric shapes: \n")
+ print(f"Cube: {surface_area_cube(20) = }")
+ print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }")
+ print(f"Sphere: {surface_area_sphere(20) = }")
+ print(f"Hemisphere: {surface_area_hemisphere(20) = }")
+ print(f"Cone: {surface_area_cone(10, 20) = }")
+ print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }")
+ print(f"Cylinder: {surface_area_cylinder(10, 20) = }")
+ print(f"Torus: {surface_area_torus(20, 10) = }")
+ print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }")
+ print(f"Square: {area_reg_polygon(4, 10) = }")
+ print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
diff --git a/maths/area_under_curve.py b/maths/area_under_curve.py
index 2d01e414b63b..10aec768fa09 100644
--- a/maths/area_under_curve.py
+++ b/maths/area_under_curve.py
@@ -2,13 +2,15 @@
Approximates the area under the curve using the trapezoidal rule
"""
-from typing import Callable, Union
+from __future__ import annotations
+
+from collections.abc import Callable
def trapezoidal_area(
- fnc: Callable[[Union[int, float]], Union[int, float]],
- x_start: Union[int, float],
- x_end: Union[int, float],
+ fnc: Callable[[float], float],
+ x_start: float,
+ x_end: float,
steps: int = 100,
) -> float:
"""
@@ -34,7 +36,7 @@ def trapezoidal_area(
x1 = x_start
fx1 = fnc(x_start)
area = 0.0
- for i in range(steps):
+ for _ in range(steps):
# Approximates small segments of curve as linear and solve
# for trapezoidal area
x2 = (x_end - x_start) / steps + x1
@@ -49,7 +51,7 @@ def trapezoidal_area(
if __name__ == "__main__":
def f(x):
- return x ** 3 + x ** 2
+ return x**3 + x**2
print("f(x) = x^3 + x^2")
print("The area between the curve, x = -5, x = 5 and the x axis is:")
diff --git a/maths/armstrong_numbers.py b/maths/armstrong_numbers.py
deleted file mode 100644
index d30ed2e430a0..000000000000
--- a/maths/armstrong_numbers.py
+++ /dev/null
@@ -1,69 +0,0 @@
-"""
-An Armstrong number is equal to the sum of the cubes of its digits.
-For example, 370 is an Armstrong number because 3*3*3 + 7*7*7 + 0*0*0 = 370.
-An Armstrong number is often called Narcissistic number.
-"""
-
-
-def armstrong_number(n: int) -> bool:
- """
- Return True if n is an Armstrong number or False if it is not.
-
- >>> armstrong_number(153)
- True
- >>> armstrong_number(200)
- False
- >>> armstrong_number(1634)
- True
- >>> armstrong_number(0)
- False
- >>> armstrong_number(-1)
- False
- >>> armstrong_number(1.2)
- False
- """
- if not isinstance(n, int) or n < 1:
- return False
-
- # Initialization of sum and number of digits.
- sum = 0
- number_of_digits = 0
- temp = n
- # Calculation of digits of the number
- while temp > 0:
- number_of_digits += 1
- temp //= 10
- # Dividing number into separate digits and find Armstrong number
- temp = n
- while temp > 0:
- rem = temp % 10
- sum += rem ** number_of_digits
- temp //= 10
- return n == sum
-
-
-def narcissistic_number(n: int) -> bool:
- """Return True if n is a narcissistic number or False if it is not"""
-
- expo = len(str(n)) # power, all number will be raised to
- # each digit will be multiplied expo times
- temp = [(int(i) ** expo) for i in str(n)]
-
- # check if sum of cube of each digit is equal to number
- return n == sum(temp)
-
-
-def main():
- """
- Request that user input an integer and tell them if it is Armstrong number.
- """
- num = int(input("Enter an integer to see if it is an Armstrong number: ").strip())
- print(f"{num} is {'' if armstrong_number(num) else 'not '}an Armstrong number.")
- print(f"{num} is {'' if narcissistic_number(num) else 'not '}an Armstrong number.")
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
- main()
diff --git a/maths/average_absolute_deviation.py b/maths/average_absolute_deviation.py
new file mode 100644
index 000000000000..193d94a2f265
--- /dev/null
+++ b/maths/average_absolute_deviation.py
@@ -0,0 +1,29 @@
+def average_absolute_deviation(nums: list[int]) -> float:
+ """
+ Return the average absolute deviation of a list of numbers.
+ Wiki: https://en.wikipedia.org/wiki/Average_absolute_deviation
+
+ >>> average_absolute_deviation([0])
+ 0.0
+ >>> average_absolute_deviation([4, 1, 3, 2])
+ 1.0
+ >>> average_absolute_deviation([2, 70, 6, 50, 20, 8, 4, 0])
+ 20.0
+ >>> average_absolute_deviation([-20, 0, 30, 15])
+ 16.25
+ >>> average_absolute_deviation([])
+ Traceback (most recent call last):
+ ...
+ ValueError: List is empty
+ """
+ if not nums: # Makes sure that the list is not empty
+ raise ValueError("List is empty")
+
+ average = sum(nums) / len(nums) # Calculate the average
+ return sum(abs(x - average) for x in nums) / len(nums)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/average_mean.py b/maths/average_mean.py
index 4beca1f741a0..274c434ab885 100644
--- a/maths/average_mean.py
+++ b/maths/average_mean.py
@@ -1,20 +1,28 @@
-"""Find mean of a list of numbers."""
+from __future__ import annotations
-def average(nums):
- """Find mean of a list of numbers."""
- return sum(nums) / len(nums)
-
-
-def test_average():
+def mean(nums: list) -> float:
"""
- >>> test_average()
+ Find mean of a list of numbers.
+ Wiki: https://en.wikipedia.org/wiki/Mean
+
+ >>> mean([3, 6, 9, 12, 15, 18, 21])
+ 12.0
+ >>> mean([5, 10, 15, 20, 25, 30, 35])
+ 20.0
+ >>> mean([1, 2, 3, 4, 5, 6, 7, 8])
+ 4.5
+ >>> mean([])
+ Traceback (most recent call last):
+ ...
+ ValueError: List is empty
"""
- assert 12.0 == average([3, 6, 9, 12, 15, 18, 21])
- assert 20 == average([5, 10, 15, 20, 25, 30, 35])
- assert 4.5 == average([1, 2, 3, 4, 5, 6, 7, 8])
+ if not nums:
+ raise ValueError("List is empty")
+ return sum(nums) / len(nums)
if __name__ == "__main__":
- """Call average module to find mean of a specific list of numbers."""
- print(average([2, 4, 6, 8, 20, 50, 70]))
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/average_median.py b/maths/average_median.py
index 0257e3f76f1a..f24e525736b3 100644
--- a/maths/average_median.py
+++ b/maths/average_median.py
@@ -1,10 +1,14 @@
-def median(nums):
+from __future__ import annotations
+
+
+def median(nums: list) -> int | float:
"""
Find median of a list of numbers.
+ Wiki: https://en.wikipedia.org/wiki/Median
>>> median([0])
0
- >>> median([4,1,3,2])
+ >>> median([4, 1, 3, 2])
2.5
>>> median([2, 70, 6, 50, 20, 8, 4])
8
@@ -15,7 +19,9 @@ def median(nums):
Returns:
Median.
"""
- sorted_list = sorted(nums)
+ # The sorted function returns list[SupportsRichComparisonT@sorted]
+ # which does not support `+`
+ sorted_list: list[int] = sorted(nums)
length = len(sorted_list)
mid_index = length >> 1
return (
diff --git a/maths/average_mode.py b/maths/average_mode.py
index d472dc04d4bf..40f88f41f8ca 100644
--- a/maths/average_mode.py
+++ b/maths/average_mode.py
@@ -1,31 +1,32 @@
-import statistics
+from typing import Any
-def mode(input_list): # Defining function "mode."
+def mode(input_list: list) -> list[Any]:
"""This function returns the mode(Mode as in the measures of
central tendency) of the input data.
The input list may contain any Datastructure or any Datatype.
- >>> input_list = [2, 3, 4, 5, 3, 4, 2, 5, 2, 2, 4, 2, 2, 2]
- >>> mode(input_list)
- 2
- >>> input_list = [2, 3, 4, 5, 3, 4, 2, 5, 2, 2, 4, 2, 2, 2]
- >>> mode(input_list) == statistics.mode(input_list)
- True
+ >>> mode([2, 3, 4, 5, 3, 4, 2, 5, 2, 2, 4, 2, 2, 2])
+ [2]
+ >>> mode([3, 4, 5, 3, 4, 2, 5, 2, 2, 4, 4, 2, 2, 2])
+ [2]
+ >>> mode([3, 4, 5, 3, 4, 2, 5, 2, 2, 4, 4, 4, 2, 2, 4, 2])
+ [2, 4]
+ >>> mode(["x", "y", "y", "z"])
+ ['y']
+ >>> mode(["x", "x" , "y", "y", "z"])
+ ['x', 'y']
"""
- # Copying input_list to check with the index number later.
- check_list = input_list.copy()
- result = list() # Empty list to store the counts of elements in input_list
- for x in input_list:
- result.append(input_list.count(x))
- input_list.remove(x)
- y = max(result) # Gets the maximum value in the result list.
- # Returns the value with the maximum number of repetitions.
- return check_list[result.index(y)]
+ if not input_list:
+ return []
+ result = [input_list.count(value) for value in input_list]
+ y = max(result) # Gets the maximum count in the input list.
+ # Gets values of modes
+ return sorted({input_list[i] for i, value in enumerate(result) if value == y})
if __name__ == "__main__":
- data = [2, 3, 4, 5, 3, 4, 2, 5, 2, 2, 4, 2, 2, 2]
- print(mode(data))
- print(statistics.mode(data))
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/bailey_borwein_plouffe.py b/maths/bailey_borwein_plouffe.py
index febf7e975516..389b1566e9de 100644
--- a/maths/bailey_borwein_plouffe.py
+++ b/maths/bailey_borwein_plouffe.py
@@ -67,10 +67,9 @@ def _subsum(
@param precision: same as precision in main function
@return: floating-point number whose integer part is not important
"""
- sum = 0.0
+ total = 0.0
for sum_index in range(digit_pos_to_extract + precision):
denominator = 8 * sum_index + denominator_addend
- exponential_term = 0.0
if sum_index < digit_pos_to_extract:
# if the exponential term is an integer and we mod it by the denominator
# before dividing, only the integer part of the sum will change;
@@ -80,8 +79,8 @@ def _subsum(
)
else:
exponential_term = pow(16, digit_pos_to_extract - 1 - sum_index)
- sum += exponential_term / denominator
- return sum
+ total += exponential_term / denominator
+ return total
if __name__ == "__main__":
diff --git a/maths/base_neg2_conversion.py b/maths/base_neg2_conversion.py
new file mode 100644
index 000000000000..81d40d37e79d
--- /dev/null
+++ b/maths/base_neg2_conversion.py
@@ -0,0 +1,37 @@
+def decimal_to_negative_base_2(num: int) -> int:
+ """
+ This function returns the number negative base 2
+ of the decimal number of the input data.
+
+ Args:
+ int: The decimal number to convert.
+
+ Returns:
+ int: The negative base 2 number.
+
+ Examples:
+ >>> decimal_to_negative_base_2(0)
+ 0
+ >>> decimal_to_negative_base_2(-19)
+ 111101
+ >>> decimal_to_negative_base_2(4)
+ 100
+ >>> decimal_to_negative_base_2(7)
+ 11011
+ """
+ if num == 0:
+ return 0
+ ans = ""
+ while num != 0:
+ num, rem = divmod(num, -2)
+ if rem < 0:
+ rem += 2
+ num += 1
+ ans = str(rem) + ans
+ return int(ans)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/basic_maths.py b/maths/basic_maths.py
index 07ee3b3df296..833f31c18b9e 100644
--- a/maths/basic_maths.py
+++ b/maths/basic_maths.py
@@ -1,4 +1,5 @@
"""Implementation of Basic Math in Python."""
+
import math
@@ -6,7 +7,17 @@ def prime_factors(n: int) -> list:
"""Find Prime Factors.
>>> prime_factors(100)
[2, 2, 5, 5]
+ >>> prime_factors(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive integers have prime factors
+ >>> prime_factors(-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive integers have prime factors
"""
+ if n <= 0:
+ raise ValueError("Only positive integers have prime factors")
pf = []
while n % 2 == 0:
pf.append(2)
@@ -24,7 +35,17 @@ def number_of_divisors(n: int) -> int:
"""Calculate Number of Divisors of an Integer.
>>> number_of_divisors(100)
9
+ >>> number_of_divisors(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive numbers are accepted
+ >>> number_of_divisors(-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive numbers are accepted
"""
+ if n <= 0:
+ raise ValueError("Only positive numbers are accepted")
div = 1
temp = 1
while n % 2 == 0:
@@ -37,6 +58,8 @@ def number_of_divisors(n: int) -> int:
temp += 1
n = int(n / i)
div *= temp
+ if n > 1:
+ div *= 2
return div
@@ -44,21 +67,31 @@ def sum_of_divisors(n: int) -> int:
"""Calculate Sum of Divisors.
>>> sum_of_divisors(100)
217
+ >>> sum_of_divisors(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive numbers are accepted
+ >>> sum_of_divisors(-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive numbers are accepted
"""
+ if n <= 0:
+ raise ValueError("Only positive numbers are accepted")
s = 1
temp = 1
while n % 2 == 0:
temp += 1
n = int(n / 2)
if temp > 1:
- s *= (2 ** temp - 1) / (2 - 1)
+ s *= (2**temp - 1) / (2 - 1)
for i in range(3, int(math.sqrt(n)) + 1, 2):
temp = 1
while n % i == 0:
temp += 1
n = int(n / i)
if temp > 1:
- s *= (i ** temp - 1) / (i - 1)
+ s *= (i**temp - 1) / (i - 1)
return int(s)
@@ -66,7 +99,17 @@ def euler_phi(n: int) -> int:
"""Calculate Euler's Phi Function.
>>> euler_phi(100)
40
+ >>> euler_phi(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive numbers are accepted
+ >>> euler_phi(-10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Only positive numbers are accepted
"""
+ if n <= 0:
+ raise ValueError("Only positive numbers are accepted")
s = n
for x in set(prime_factors(n)):
s *= (x - 1) / x
@@ -74,7 +117,6 @@ def euler_phi(n: int) -> int:
if __name__ == "__main__":
- print(prime_factors(100))
- print(number_of_divisors(100))
- print(sum_of_divisors(100))
- print(euler_phi(100))
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/binary_exp_mod.py b/maths/binary_exp_mod.py
deleted file mode 100644
index 67dd1e728b18..000000000000
--- a/maths/binary_exp_mod.py
+++ /dev/null
@@ -1,28 +0,0 @@
-def bin_exp_mod(a, n, b):
- """
- >>> bin_exp_mod(3, 4, 5)
- 1
- >>> bin_exp_mod(7, 13, 10)
- 7
- """
- # mod b
- assert not (b == 0), "This cannot accept modulo that is == 0"
- if n == 0:
- return 1
-
- if n % 2 == 1:
- return (bin_exp_mod(a, n - 1, b) * a) % b
-
- r = bin_exp_mod(a, n / 2, b)
- return (r * r) % b
-
-
-if __name__ == "__main__":
- try:
- BASE = int(input("Enter Base : ").strip())
- POWER = int(input("Enter Power : ").strip())
- MODULO = int(input("Enter Modulo : ").strip())
- except ValueError:
- print("Invalid literal for integer")
-
- print(bin_exp_mod(BASE, POWER, MODULO))
diff --git a/maths/binary_exponentiation.py b/maths/binary_exponentiation.py
index 8dda5245cf44..51ce86d26c41 100644
--- a/maths/binary_exponentiation.py
+++ b/maths/binary_exponentiation.py
@@ -1,28 +1,196 @@
-"""Binary Exponentiation."""
+"""
+Binary Exponentiation
-# Author : Junth Basnet
-# Time Complexity : O(logn)
+This is a method to find a^b in O(log b) time complexity and is one of the most commonly
+used methods of exponentiation. The method is also useful for modular exponentiation,
+when the solution to (a^b) % c is required.
+To calculate a^b:
+- If b is even, then a^b = (a * a)^(b / 2)
+- If b is odd, then a^b = a * a^(b - 1)
+Repeat until b = 1 or b = 0
-def binary_exponentiation(a, n):
+For modular exponentiation, we use the fact that (a * b) % c = ((a % c) * (b % c)) % c
+"""
- if n == 0:
+
+def binary_exp_recursive(base: float, exponent: int) -> float:
+ """
+ Computes a^b recursively, where a is the base and b is the exponent
+
+ >>> binary_exp_recursive(3, 5)
+ 243
+ >>> binary_exp_recursive(11, 13)
+ 34522712143931
+ >>> binary_exp_recursive(-1, 3)
+ -1
+ >>> binary_exp_recursive(0, 5)
+ 0
+ >>> binary_exp_recursive(3, 1)
+ 3
+ >>> binary_exp_recursive(3, 0)
+ 1
+ >>> binary_exp_recursive(1.5, 4)
+ 5.0625
+ >>> binary_exp_recursive(3, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Exponent must be a non-negative integer
+ """
+ if exponent < 0:
+ raise ValueError("Exponent must be a non-negative integer")
+
+ if exponent == 0:
+ return 1
+
+ if exponent % 2 == 1:
+ return binary_exp_recursive(base, exponent - 1) * base
+
+ b = binary_exp_recursive(base, exponent // 2)
+ return b * b
+
+
+def binary_exp_iterative(base: float, exponent: int) -> float:
+ """
+ Computes a^b iteratively, where a is the base and b is the exponent
+
+ >>> binary_exp_iterative(3, 5)
+ 243
+ >>> binary_exp_iterative(11, 13)
+ 34522712143931
+ >>> binary_exp_iterative(-1, 3)
+ -1
+ >>> binary_exp_iterative(0, 5)
+ 0
+ >>> binary_exp_iterative(3, 1)
+ 3
+ >>> binary_exp_iterative(3, 0)
+ 1
+ >>> binary_exp_iterative(1.5, 4)
+ 5.0625
+ >>> binary_exp_iterative(3, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Exponent must be a non-negative integer
+ """
+ if exponent < 0:
+ raise ValueError("Exponent must be a non-negative integer")
+
+ res: int | float = 1
+ while exponent > 0:
+ if exponent & 1:
+ res *= base
+
+ base *= base
+ exponent >>= 1
+
+ return res
+
+
+def binary_exp_mod_recursive(base: float, exponent: int, modulus: int) -> float:
+ """
+ Computes a^b % c recursively, where a is the base, b is the exponent, and c is the
+ modulus
+
+ >>> binary_exp_mod_recursive(3, 4, 5)
+ 1
+ >>> binary_exp_mod_recursive(11, 13, 7)
+ 4
+ >>> binary_exp_mod_recursive(1.5, 4, 3)
+ 2.0625
+ >>> binary_exp_mod_recursive(7, -1, 10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Exponent must be a non-negative integer
+ >>> binary_exp_mod_recursive(7, 13, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Modulus must be a positive integer
+ """
+ if exponent < 0:
+ raise ValueError("Exponent must be a non-negative integer")
+ if modulus <= 0:
+ raise ValueError("Modulus must be a positive integer")
+
+ if exponent == 0:
return 1
- elif n % 2 == 1:
- return binary_exponentiation(a, n - 1) * a
+ if exponent % 2 == 1:
+ return (binary_exp_mod_recursive(base, exponent - 1, modulus) * base) % modulus
+
+ r = binary_exp_mod_recursive(base, exponent // 2, modulus)
+ return (r * r) % modulus
+
- else:
- b = binary_exponentiation(a, n / 2)
- return b * b
+def binary_exp_mod_iterative(base: float, exponent: int, modulus: int) -> float:
+ """
+ Computes a^b % c iteratively, where a is the base, b is the exponent, and c is the
+ modulus
+
+ >>> binary_exp_mod_iterative(3, 4, 5)
+ 1
+ >>> binary_exp_mod_iterative(11, 13, 7)
+ 4
+ >>> binary_exp_mod_iterative(1.5, 4, 3)
+ 2.0625
+ >>> binary_exp_mod_iterative(7, -1, 10)
+ Traceback (most recent call last):
+ ...
+ ValueError: Exponent must be a non-negative integer
+ >>> binary_exp_mod_iterative(7, 13, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Modulus must be a positive integer
+ """
+ if exponent < 0:
+ raise ValueError("Exponent must be a non-negative integer")
+ if modulus <= 0:
+ raise ValueError("Modulus must be a positive integer")
+
+ res: int | float = 1
+ while exponent > 0:
+ if exponent & 1:
+ res = ((res % modulus) * (base % modulus)) % modulus
+
+ base *= base
+ exponent >>= 1
+
+ return res
if __name__ == "__main__":
- try:
- BASE = int(input("Enter Base : ").strip())
- POWER = int(input("Enter Power : ").strip())
- except ValueError:
- print("Invalid literal for integer")
-
- RESULT = binary_exponentiation(BASE, POWER)
- print(f"{BASE}^({POWER}) : {RESULT}")
+ from timeit import timeit
+
+ a = 1269380576
+ b = 374
+ c = 34
+
+ runs = 100_000
+ print(
+ timeit(
+ f"binary_exp_recursive({a}, {b})",
+ setup="from __main__ import binary_exp_recursive",
+ number=runs,
+ )
+ )
+ print(
+ timeit(
+ f"binary_exp_iterative({a}, {b})",
+ setup="from __main__ import binary_exp_iterative",
+ number=runs,
+ )
+ )
+ print(
+ timeit(
+ f"binary_exp_mod_recursive({a}, {b}, {c})",
+ setup="from __main__ import binary_exp_mod_recursive",
+ number=runs,
+ )
+ )
+ print(
+ timeit(
+ f"binary_exp_mod_iterative({a}, {b}, {c})",
+ setup="from __main__ import binary_exp_mod_iterative",
+ number=runs,
+ )
+ )
diff --git a/maths/binary_multiplication.py b/maths/binary_multiplication.py
new file mode 100644
index 000000000000..0cc5a575f445
--- /dev/null
+++ b/maths/binary_multiplication.py
@@ -0,0 +1,101 @@
+"""
+Binary Multiplication
+This is a method to find a*b in a time complexity of O(log b)
+This is one of the most commonly used methods of finding result of multiplication.
+Also useful in cases where solution to (a*b)%c is required,
+where a,b,c can be numbers over the computers calculation limits.
+Done using iteration, can also be done using recursion
+
+Let's say you need to calculate a * b
+RULE 1 : a * b = (a+a) * (b/2) ---- example : 4 * 4 = (4+4) * (4/2) = 8 * 2
+RULE 2 : IF b is odd, then ---- a * b = a + (a * (b - 1)), where (b - 1) is even.
+Once b is even, repeat the process to get a * b
+Repeat the process until b = 1 or b = 0, because a*1 = a and a*0 = 0
+
+As far as the modulo is concerned,
+the fact : (a+b) % c = ((a%c) + (b%c)) % c
+Now apply RULE 1 or 2, whichever is required.
+
+@author chinmoy159
+"""
+
+
+def binary_multiply(a: int, b: int) -> int:
+ """
+ Multiply 'a' and 'b' using bitwise multiplication.
+
+ Parameters:
+ a (int): The first number.
+ b (int): The second number.
+
+ Returns:
+ int: a * b
+
+ Examples:
+ >>> binary_multiply(2, 3)
+ 6
+ >>> binary_multiply(5, 0)
+ 0
+ >>> binary_multiply(3, 4)
+ 12
+ >>> binary_multiply(10, 5)
+ 50
+ >>> binary_multiply(0, 5)
+ 0
+ >>> binary_multiply(2, 1)
+ 2
+ >>> binary_multiply(1, 10)
+ 10
+ """
+ res = 0
+ while b > 0:
+ if b & 1:
+ res += a
+
+ a += a
+ b >>= 1
+
+ return res
+
+
+def binary_mod_multiply(a: int, b: int, modulus: int) -> int:
+ """
+ Calculate (a * b) % c using binary multiplication and modular arithmetic.
+
+ Parameters:
+ a (int): The first number.
+ b (int): The second number.
+ modulus (int): The modulus.
+
+ Returns:
+ int: (a * b) % modulus.
+
+ Examples:
+ >>> binary_mod_multiply(2, 3, 5)
+ 1
+ >>> binary_mod_multiply(5, 0, 7)
+ 0
+ >>> binary_mod_multiply(3, 4, 6)
+ 0
+ >>> binary_mod_multiply(10, 5, 13)
+ 11
+ >>> binary_mod_multiply(2, 1, 5)
+ 2
+ >>> binary_mod_multiply(1, 10, 3)
+ 1
+ """
+ res = 0
+ while b > 0:
+ if b & 1:
+ res = ((res % modulus) + (a % modulus)) % modulus
+
+ a += a
+ b >>= 1
+
+ return res
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/binomial_coefficient.py b/maths/binomial_coefficient.py
index 4def041492f3..24c54326e305 100644
--- a/maths/binomial_coefficient.py
+++ b/maths/binomial_coefficient.py
@@ -1,20 +1,62 @@
-def binomial_coefficient(n, r):
+def binomial_coefficient(n: int, r: int) -> int:
"""
- Find binomial coefficient using pascals triangle.
+ Find binomial coefficient using Pascal's triangle.
+
+ Calculate C(n, r) using Pascal's triangle.
+
+ :param n: The total number of items.
+ :param r: The number of items to choose.
+ :return: The binomial coefficient C(n, r).
>>> binomial_coefficient(10, 5)
252
+ >>> binomial_coefficient(10, 0)
+ 1
+ >>> binomial_coefficient(0, 10)
+ 1
+ >>> binomial_coefficient(10, 10)
+ 1
+ >>> binomial_coefficient(5, 2)
+ 10
+ >>> binomial_coefficient(5, 6)
+ 0
+ >>> binomial_coefficient(3, 5)
+ 0
+ >>> binomial_coefficient(-2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: n and r must be non-negative integers
+ >>> binomial_coefficient(5, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n and r must be non-negative integers
+ >>> binomial_coefficient(10.1, 5)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'float' object cannot be interpreted as an integer
+ >>> binomial_coefficient(10, 5.1)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'float' object cannot be interpreted as an integer
"""
- C = [0 for i in range(r + 1)]
+ if n < 0 or r < 0:
+ raise ValueError("n and r must be non-negative integers")
+ if 0 in (n, r):
+ return 1
+ c = [0 for i in range(r + 1)]
# nc0 = 1
- C[0] = 1
+ c[0] = 1
for i in range(1, n + 1):
# to compute current row from previous row.
j = min(i, r)
while j > 0:
- C[j] += C[j - 1]
+ c[j] += c[j - 1]
j -= 1
- return C[r]
+ return c[r]
+
+if __name__ == "__main__":
+ from doctest import testmod
-print(binomial_coefficient(n=10, r=5))
+ testmod()
+ print(binomial_coefficient(n=10, r=5))
diff --git a/maths/binomial_distribution.py b/maths/binomial_distribution.py
index a74a5a7ed994..eabcaea0d1b2 100644
--- a/maths/binomial_distribution.py
+++ b/maths/binomial_distribution.py
@@ -1,5 +1,6 @@
"""For more information about the Binomial Distribution -
- https://en.wikipedia.org/wiki/Binomial_distribution"""
+https://en.wikipedia.org/wiki/Binomial_distribution"""
+
from math import factorial
@@ -24,7 +25,7 @@ def binomial_distribution(successes: int, trials: int, prob: float) -> float:
raise ValueError("the function is defined for non-negative integers")
if not 0 < prob < 1:
raise ValueError("prob has to be in range of 1 - 0")
- probability = (prob ** successes) * ((1 - prob) ** (trials - successes))
+ probability = (prob**successes) * ((1 - prob) ** (trials - successes))
# Calculate the binomial coefficient: n! / k!(n-k)!
coefficient = float(factorial(trials))
coefficient /= factorial(successes) * factorial(trials - successes)
diff --git a/maths/ceil.py b/maths/ceil.py
index 97578265c1a9..909e02b3f780 100644
--- a/maths/ceil.py
+++ b/maths/ceil.py
@@ -3,7 +3,7 @@
"""
-def ceil(x) -> int:
+def ceil(x: float) -> int:
"""
Return the ceiling of x as an Integral.
diff --git a/maths/chebyshev_distance.py b/maths/chebyshev_distance.py
new file mode 100644
index 000000000000..4801d391621f
--- /dev/null
+++ b/maths/chebyshev_distance.py
@@ -0,0 +1,20 @@
+def chebyshev_distance(point_a: list[float], point_b: list[float]) -> float:
+ """
+ This function calculates the Chebyshev distance (also known as the
+ Chessboard distance) between two n-dimensional points represented as lists.
+
+ https://en.wikipedia.org/wiki/Chebyshev_distance
+
+ >>> chebyshev_distance([1.0, 1.0], [2.0, 2.0])
+ 1.0
+ >>> chebyshev_distance([1.0, 1.0, 9.0], [2.0, 2.0, -5.2])
+ 14.2
+ >>> chebyshev_distance([1.0], [2.0, 2.0])
+ Traceback (most recent call last):
+ ...
+ ValueError: Both points must have the same dimension.
+ """
+ if len(point_a) != len(point_b):
+ raise ValueError("Both points must have the same dimension.")
+
+ return max(abs(a - b) for a, b in zip(point_a, point_b))
diff --git a/maths/check_polygon.py b/maths/check_polygon.py
new file mode 100644
index 000000000000..1e8dce7183ad
--- /dev/null
+++ b/maths/check_polygon.py
@@ -0,0 +1,44 @@
+from __future__ import annotations
+
+
+def check_polygon(nums: list[float]) -> bool:
+ """
+ Takes list of possible side lengths and determines whether a
+ two-dimensional polygon with such side lengths can exist.
+
+ Returns a boolean value for the < comparison
+ of the largest side length with sum of the rest.
+ Wiki: https://en.wikipedia.org/wiki/Triangle_inequality
+
+ >>> check_polygon([6, 10, 5])
+ True
+ >>> check_polygon([3, 7, 13, 2])
+ False
+ >>> check_polygon([1, 4.3, 5.2, 12.2])
+ False
+ >>> nums = [3, 7, 13, 2]
+ >>> _ = check_polygon(nums) # Run function, do not show answer in output
+ >>> nums # Check numbers are not reordered
+ [3, 7, 13, 2]
+ >>> check_polygon([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Monogons and Digons are not polygons in the Euclidean space
+ >>> check_polygon([-2, 5, 6])
+ Traceback (most recent call last):
+ ...
+ ValueError: All values must be greater than 0
+ """
+ if len(nums) < 2:
+ raise ValueError("Monogons and Digons are not polygons in the Euclidean space")
+ if any(i <= 0 for i in nums):
+ raise ValueError("All values must be greater than 0")
+ copy_nums = nums.copy()
+ copy_nums.sort()
+ return copy_nums[-1] < sum(copy_nums[:-1])
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/blockchain/chinese_remainder_theorem.py b/maths/chinese_remainder_theorem.py
similarity index 69%
rename from blockchain/chinese_remainder_theorem.py
rename to maths/chinese_remainder_theorem.py
index b6a486f0b1ed..18af63d106e8 100644
--- a/blockchain/chinese_remainder_theorem.py
+++ b/maths/chinese_remainder_theorem.py
@@ -1,18 +1,22 @@
-# Chinese Remainder Theorem:
-# GCD ( Greatest Common Divisor ) or HCF ( Highest Common Factor )
+"""
+Chinese Remainder Theorem:
+GCD ( Greatest Common Divisor ) or HCF ( Highest Common Factor )
-# If GCD(a,b) = 1, then for any remainder ra modulo a and any remainder rb modulo b
-# there exists integer n, such that n = ra (mod a) and n = ra(mod b). If n1 and n2 are
-# two such integers, then n1=n2(mod ab)
+If GCD(a,b) = 1, then for any remainder ra modulo a and any remainder rb modulo b
+there exists integer n, such that n = ra (mod a) and n = ra(mod b). If n1 and n2 are
+two such integers, then n1=n2(mod ab)
-# Algorithm :
+Algorithm :
-# 1. Use extended euclid algorithm to find x,y such that a*x + b*y = 1
-# 2. Take n = ra*by + rb*ax
+1. Use extended euclid algorithm to find x,y such that a*x + b*y = 1
+2. Take n = ra*by + rb*ax
+"""
+
+from __future__ import annotations
# Extended Euclid
-def extended_euclid(a, b):
+def extended_euclid(a: int, b: int) -> tuple[int, int]:
"""
>>> extended_euclid(10, 6)
(-1, 2)
@@ -29,7 +33,7 @@ def extended_euclid(a, b):
# Uses ExtendedEuclid to find inverses
-def chinese_remainder_theorem(n1, r1, n2, r2):
+def chinese_remainder_theorem(n1: int, r1: int, n2: int, r2: int) -> int:
"""
>>> chinese_remainder_theorem(5,1,7,3)
31
@@ -50,8 +54,9 @@ def chinese_remainder_theorem(n1, r1, n2, r2):
# ----------SAME SOLUTION USING InvertModulo instead ExtendedEuclid----------------
+
# This function find the inverses of a i.e., a^(-1)
-def invert_modulo(a, n):
+def invert_modulo(a: int, n: int) -> int:
"""
>>> invert_modulo(2, 5)
3
@@ -67,7 +72,7 @@ def invert_modulo(a, n):
# Same a above using InvertingModulo
-def chinese_remainder_theorem2(n1, r1, n2, r2):
+def chinese_remainder_theorem2(n1: int, r1: int, n2: int, r2: int) -> int:
"""
>>> chinese_remainder_theorem2(5,1,7,3)
31
diff --git a/maths/chudnovsky_algorithm.py b/maths/chudnovsky_algorithm.py
index aaee7462822e..d122bf0756f7 100644
--- a/maths/chudnovsky_algorithm.py
+++ b/maths/chudnovsky_algorithm.py
@@ -5,7 +5,7 @@
def pi(precision: int) -> str:
"""
The Chudnovsky algorithm is a fast method for calculating the digits of PI,
- based on Ramanujan’s PI formulae.
+ based on Ramanujan's PI formulae.
https://en.wikipedia.org/wiki/Chudnovsky_algorithm
diff --git a/maths/collatz_sequence.py b/maths/collatz_sequence.py
index 7b3636de69f4..b00dca8d70b7 100644
--- a/maths/collatz_sequence.py
+++ b/maths/collatz_sequence.py
@@ -1,43 +1,66 @@
+"""
+The Collatz conjecture is a famous unsolved problem in mathematics. Given a starting
+positive integer, define the following sequence:
+- If the current term n is even, then the next term is n/2.
+- If the current term n is odd, then the next term is 3n + 1.
+The conjecture claims that this sequence will always reach 1 for any starting number.
+
+Other names for this problem include the 3n + 1 problem, the Ulam conjecture, Kakutani's
+problem, the Thwaites conjecture, Hasse's algorithm, the Syracuse problem, and the
+hailstone sequence.
+
+Reference: https://en.wikipedia.org/wiki/Collatz_conjecture
+"""
+
from __future__ import annotations
+from collections.abc import Generator
-def collatz_sequence(n: int) -> list[int]:
+
+def collatz_sequence(n: int) -> Generator[int]:
"""
- Collatz conjecture: start with any positive integer n. The next term is
- obtained as follows:
- If n term is even, the next term is: n / 2 .
- If n is odd, the next term is: 3 * n + 1.
-
- The conjecture states the sequence will always reach 1 for any starting value n.
- Example:
- >>> collatz_sequence(2.1)
+ Generate the Collatz sequence starting at n.
+ >>> tuple(collatz_sequence(2.1))
Traceback (most recent call last):
...
- Exception: Sequence only defined for natural numbers
- >>> collatz_sequence(0)
+ Exception: Sequence only defined for positive integers
+ >>> tuple(collatz_sequence(0))
Traceback (most recent call last):
...
- Exception: Sequence only defined for natural numbers
- >>> collatz_sequence(43) # doctest: +NORMALIZE_WHITESPACE
- [43, 130, 65, 196, 98, 49, 148, 74, 37, 112, 56, 28, 14, 7,
- 22, 11, 34, 17, 52, 26, 13, 40, 20, 10, 5, 16, 8, 4, 2, 1]
+ Exception: Sequence only defined for positive integers
+ >>> tuple(collatz_sequence(4))
+ (4, 2, 1)
+ >>> tuple(collatz_sequence(11))
+ (11, 34, 17, 52, 26, 13, 40, 20, 10, 5, 16, 8, 4, 2, 1)
+ >>> tuple(collatz_sequence(31)) # doctest: +NORMALIZE_WHITESPACE
+ (31, 94, 47, 142, 71, 214, 107, 322, 161, 484, 242, 121, 364, 182, 91, 274, 137,
+ 412, 206, 103, 310, 155, 466, 233, 700, 350, 175, 526, 263, 790, 395, 1186, 593,
+ 1780, 890, 445, 1336, 668, 334, 167, 502, 251, 754, 377, 1132, 566, 283, 850, 425,
+ 1276, 638, 319, 958, 479, 1438, 719, 2158, 1079, 3238, 1619, 4858, 2429, 7288, 3644,
+ 1822, 911, 2734, 1367, 4102, 2051, 6154, 3077, 9232, 4616, 2308, 1154, 577, 1732,
+ 866, 433, 1300, 650, 325, 976, 488, 244, 122, 61, 184, 92, 46, 23, 70, 35, 106, 53,
+ 160, 80, 40, 20, 10, 5, 16, 8, 4, 2, 1)
+ >>> tuple(collatz_sequence(43)) # doctest: +NORMALIZE_WHITESPACE
+ (43, 130, 65, 196, 98, 49, 148, 74, 37, 112, 56, 28, 14, 7, 22, 11, 34, 17, 52, 26,
+ 13, 40, 20, 10, 5, 16, 8, 4, 2, 1)
"""
-
if not isinstance(n, int) or n < 1:
- raise Exception("Sequence only defined for natural numbers")
+ raise Exception("Sequence only defined for positive integers")
- sequence = [n]
+ yield n
while n != 1:
- n = 3 * n + 1 if n & 1 else n // 2
- sequence.append(n)
- return sequence
+ if n % 2 == 0:
+ n //= 2
+ else:
+ n = 3 * n + 1
+ yield n
def main():
- n = 43
- sequence = collatz_sequence(n)
+ n = int(input("Your number: "))
+ sequence = tuple(collatz_sequence(n))
print(sequence)
- print(f"collatz sequence from {n} took {len(sequence)} steps.")
+ print(f"Collatz sequence from {n} took {len(sequence)} steps.")
if __name__ == "__main__":
diff --git a/maths/combinations.py b/maths/combinations.py
index 40f4f7a9f850..6e9e1a807067 100644
--- a/maths/combinations.py
+++ b/maths/combinations.py
@@ -1,7 +1,6 @@
"""
https://en.wikipedia.org/wiki/Combination
"""
-from math import factorial
def combinations(n: int, k: int) -> int:
@@ -35,18 +34,21 @@ def combinations(n: int, k: int) -> int:
# to calculate a factorial of a negative number, which is not possible
if n < k or k < 0:
raise ValueError("Please enter positive integers for n and k where n >= k")
- return int(factorial(n) / ((factorial(k)) * (factorial(n - k))))
+ res = 1
+ for i in range(k):
+ res *= n - i
+ res //= i + 1
+ return res
if __name__ == "__main__":
-
print(
- "\nThe number of five-card hands possible from a standard",
- f"fifty-two card deck is: {combinations(52, 5)}",
+ "The number of five-card hands possible from a standard",
+ f"fifty-two card deck is: {combinations(52, 5)}\n",
)
print(
- "\nIf a class of 40 students must be arranged into groups of",
+ "If a class of 40 students must be arranged into groups of",
f"4 for group projects, there are {combinations(40, 4)} ways",
"to arrange them.\n",
)
@@ -54,5 +56,5 @@ def combinations(n: int, k: int) -> int:
print(
"If 10 teams are competing in a Formula One race, there",
f"are {combinations(10, 3)} ways that first, second and",
- "third place can be awarded.\n",
+ "third place can be awarded.",
)
diff --git a/maths/continued_fraction.py b/maths/continued_fraction.py
new file mode 100644
index 000000000000..2c38bf88b1e9
--- /dev/null
+++ b/maths/continued_fraction.py
@@ -0,0 +1,57 @@
+"""
+Finding the continuous fraction for a rational number using python
+
+https://en.wikipedia.org/wiki/Continued_fraction
+"""
+
+from fractions import Fraction
+from math import floor
+
+
+def continued_fraction(num: Fraction) -> list[int]:
+ """
+ :param num:
+ Fraction of the number whose continued fractions to be found.
+ Use Fraction(str(number)) for more accurate results due to
+ float inaccuracies.
+
+ :return:
+ The continued fraction of rational number.
+ It is the all commas in the (n + 1)-tuple notation.
+
+ >>> continued_fraction(Fraction(2))
+ [2]
+ >>> continued_fraction(Fraction("3.245"))
+ [3, 4, 12, 4]
+ >>> continued_fraction(Fraction("2.25"))
+ [2, 4]
+ >>> continued_fraction(1/Fraction("2.25"))
+ [0, 2, 4]
+ >>> continued_fraction(Fraction("415/93"))
+ [4, 2, 6, 7]
+ >>> continued_fraction(Fraction(0))
+ [0]
+ >>> continued_fraction(Fraction(0.75))
+ [0, 1, 3]
+ >>> continued_fraction(Fraction("-2.25")) # -2.25 = -3 + 0.75
+ [-3, 1, 3]
+ """
+ numerator, denominator = num.as_integer_ratio()
+ continued_fraction_list: list[int] = []
+ while True:
+ integer_part = floor(numerator / denominator)
+ continued_fraction_list.append(integer_part)
+ numerator -= integer_part * denominator
+ if numerator == 0:
+ break
+ numerator, denominator = denominator, numerator
+
+ return continued_fraction_list
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ print("Continued Fraction of 0.84375 is: ", continued_fraction(Fraction("0.84375")))
diff --git a/maths/decimal_isolate.py b/maths/decimal_isolate.py
index 0e3967a4671d..058ed1bb90d1 100644
--- a/maths/decimal_isolate.py
+++ b/maths/decimal_isolate.py
@@ -4,8 +4,7 @@
"""
-def decimal_isolate(number, digitAmount):
-
+def decimal_isolate(number: float, digit_amount: int) -> float:
"""
Isolates the decimal part of a number.
If digitAmount > 0 round to that decimal place, else print the entire decimal.
@@ -28,8 +27,8 @@ def decimal_isolate(number, digitAmount):
>>> decimal_isolate(-14.123, 3)
-0.123
"""
- if digitAmount > 0:
- return round(number - int(number), digitAmount)
+ if digit_amount > 0:
+ return round(number - int(number), digit_amount)
return number - int(number)
diff --git a/maths/decimal_to_fraction.py b/maths/decimal_to_fraction.py
new file mode 100644
index 000000000000..be42b9fb3b5a
--- /dev/null
+++ b/maths/decimal_to_fraction.py
@@ -0,0 +1,62 @@
+def decimal_to_fraction(decimal: float | str) -> tuple[int, int]:
+ """
+ Return a decimal number in its simplest fraction form
+ >>> decimal_to_fraction(2)
+ (2, 1)
+ >>> decimal_to_fraction(89.)
+ (89, 1)
+ >>> decimal_to_fraction("67")
+ (67, 1)
+ >>> decimal_to_fraction("45.0")
+ (45, 1)
+ >>> decimal_to_fraction(1.5)
+ (3, 2)
+ >>> decimal_to_fraction("6.25")
+ (25, 4)
+ >>> decimal_to_fraction("78td")
+ Traceback (most recent call last):
+ ValueError: Please enter a valid number
+ >>> decimal_to_fraction(0)
+ (0, 1)
+ >>> decimal_to_fraction(-2.5)
+ (-5, 2)
+ >>> decimal_to_fraction(0.125)
+ (1, 8)
+ >>> decimal_to_fraction(1000000.25)
+ (4000001, 4)
+ >>> decimal_to_fraction(1.3333)
+ (13333, 10000)
+ >>> decimal_to_fraction("1.23e2")
+ (123, 1)
+ >>> decimal_to_fraction("0.500")
+ (1, 2)
+ """
+ try:
+ decimal = float(decimal)
+ except ValueError:
+ raise ValueError("Please enter a valid number")
+ fractional_part = decimal - int(decimal)
+ if fractional_part == 0:
+ return int(decimal), 1
+ else:
+ number_of_frac_digits = len(str(decimal).split(".")[1])
+ numerator = int(decimal * (10**number_of_frac_digits))
+ denominator = 10**number_of_frac_digits
+ divisor, dividend = denominator, numerator
+ while True:
+ remainder = dividend % divisor
+ if remainder == 0:
+ break
+ dividend, divisor = divisor, remainder
+ numerator, denominator = numerator // divisor, denominator // divisor
+ return numerator, denominator
+
+
+if __name__ == "__main__":
+ print(f"{decimal_to_fraction(2) = }")
+ print(f"{decimal_to_fraction(89.0) = }")
+ print(f"{decimal_to_fraction('67') = }")
+ print(f"{decimal_to_fraction('45.0') = }")
+ print(f"{decimal_to_fraction(1.5) = }")
+ print(f"{decimal_to_fraction('6.25') = }")
+ print(f"{decimal_to_fraction('78td') = }")
diff --git a/maths/dodecahedron.py b/maths/dodecahedron.py
new file mode 100644
index 000000000000..856245f4a868
--- /dev/null
+++ b/maths/dodecahedron.py
@@ -0,0 +1,73 @@
+# dodecahedron.py
+
+"""
+A regular dodecahedron is a three-dimensional figure made up of
+12 pentagon faces having the same equal size.
+"""
+
+
+def dodecahedron_surface_area(edge: float) -> float:
+ """
+ Calculates the surface area of a regular dodecahedron
+ a = 3 * ((25 + 10 * (5** (1 / 2))) ** (1 / 2 )) * (e**2)
+ where:
+ a --> is the area of the dodecahedron
+ e --> is the length of the edge
+ reference-->"Dodecahedron" Study.com
+
+
+ :param edge: length of the edge of the dodecahedron
+ :type edge: float
+ :return: the surface area of the dodecahedron as a float
+
+
+ Tests:
+ >>> dodecahedron_surface_area(5)
+ 516.1432201766901
+ >>> dodecahedron_surface_area(10)
+ 2064.5728807067603
+ >>> dodecahedron_surface_area(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Length must be a positive.
+ """
+
+ if edge <= 0 or not isinstance(edge, int):
+ raise ValueError("Length must be a positive.")
+ return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2)
+
+
+def dodecahedron_volume(edge: float) -> float:
+ """
+ Calculates the volume of a regular dodecahedron
+ v = ((15 + (7 * (5** (1 / 2)))) / 4) * (e**3)
+ where:
+ v --> is the volume of the dodecahedron
+ e --> is the length of the edge
+ reference-->"Dodecahedron" Study.com
+
+
+ :param edge: length of the edge of the dodecahedron
+ :type edge: float
+ :return: the volume of the dodecahedron as a float
+
+ Tests:
+ >>> dodecahedron_volume(5)
+ 957.8898700780791
+ >>> dodecahedron_volume(10)
+ 7663.118960624633
+ >>> dodecahedron_volume(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Length must be a positive.
+ """
+
+ if edge <= 0 or not isinstance(edge, int):
+ raise ValueError("Length must be a positive.")
+ return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/double_factorial.py b/maths/double_factorial.py
new file mode 100644
index 000000000000..3c3a28304e95
--- /dev/null
+++ b/maths/double_factorial.py
@@ -0,0 +1,60 @@
+def double_factorial_recursive(n: int) -> int:
+ """
+ Compute double factorial using recursive method.
+ Recursion can be costly for large numbers.
+
+ To learn about the theory behind this algorithm:
+ https://en.wikipedia.org/wiki/Double_factorial
+
+ >>> from math import prod
+ >>> all(double_factorial_recursive(i) == prod(range(i, 0, -2)) for i in range(20))
+ True
+ >>> double_factorial_recursive(0.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: double_factorial_recursive() only accepts integral values
+ >>> double_factorial_recursive(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: double_factorial_recursive() not defined for negative values
+ """
+ if not isinstance(n, int):
+ raise ValueError("double_factorial_recursive() only accepts integral values")
+ if n < 0:
+ raise ValueError("double_factorial_recursive() not defined for negative values")
+ return 1 if n <= 1 else n * double_factorial_recursive(n - 2)
+
+
+def double_factorial_iterative(num: int) -> int:
+ """
+ Compute double factorial using iterative method.
+
+ To learn about the theory behind this algorithm:
+ https://en.wikipedia.org/wiki/Double_factorial
+
+ >>> from math import prod
+ >>> all(double_factorial_iterative(i) == prod(range(i, 0, -2)) for i in range(20))
+ True
+ >>> double_factorial_iterative(0.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: double_factorial_iterative() only accepts integral values
+ >>> double_factorial_iterative(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: double_factorial_iterative() not defined for negative values
+ """
+ if not isinstance(num, int):
+ raise ValueError("double_factorial_iterative() only accepts integral values")
+ if num < 0:
+ raise ValueError("double_factorial_iterative() not defined for negative values")
+ value = 1
+ for i in range(num, 0, -2):
+ value *= i
+ return value
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/dual_number_automatic_differentiation.py b/maths/dual_number_automatic_differentiation.py
new file mode 100644
index 000000000000..09aeb17a4aea
--- /dev/null
+++ b/maths/dual_number_automatic_differentiation.py
@@ -0,0 +1,139 @@
+from math import factorial
+
+"""
+https://en.wikipedia.org/wiki/Automatic_differentiation#Automatic_differentiation_using_dual_numbers
+https://blog.jliszka.org/2013/10/24/exact-numeric-nth-derivatives.html
+
+Note this only works for basic functions, f(x) where the power of x is positive.
+"""
+
+
+class Dual:
+ def __init__(self, real, rank):
+ self.real = real
+ if isinstance(rank, int):
+ self.duals = [1] * rank
+ else:
+ self.duals = rank
+
+ def __repr__(self):
+ s = "+".join(f"{dual}E{n}" for n, dual in enumerate(self.duals, 1))
+ return f"{self.real}+{s}"
+
+ def reduce(self):
+ cur = self.duals.copy()
+ while cur[-1] == 0:
+ cur.pop(-1)
+ return Dual(self.real, cur)
+
+ def __add__(self, other):
+ if not isinstance(other, Dual):
+ return Dual(self.real + other, self.duals)
+ s_dual = self.duals.copy()
+ o_dual = other.duals.copy()
+ if len(s_dual) > len(o_dual):
+ o_dual.extend([1] * (len(s_dual) - len(o_dual)))
+ elif len(s_dual) < len(o_dual):
+ s_dual.extend([1] * (len(o_dual) - len(s_dual)))
+ new_duals = []
+ for i in range(len(s_dual)):
+ new_duals.append(s_dual[i] + o_dual[i])
+ return Dual(self.real + other.real, new_duals)
+
+ __radd__ = __add__
+
+ def __sub__(self, other):
+ return self + other * -1
+
+ def __mul__(self, other):
+ if not isinstance(other, Dual):
+ new_duals = []
+ for i in self.duals:
+ new_duals.append(i * other)
+ return Dual(self.real * other, new_duals)
+ new_duals = [0] * (len(self.duals) + len(other.duals) + 1)
+ for i, item in enumerate(self.duals):
+ for j, jtem in enumerate(other.duals):
+ new_duals[i + j + 1] += item * jtem
+ for k in range(len(self.duals)):
+ new_duals[k] += self.duals[k] * other.real
+ for index in range(len(other.duals)):
+ new_duals[index] += other.duals[index] * self.real
+ return Dual(self.real * other.real, new_duals)
+
+ __rmul__ = __mul__
+
+ def __truediv__(self, other):
+ if not isinstance(other, Dual):
+ new_duals = []
+ for i in self.duals:
+ new_duals.append(i / other)
+ return Dual(self.real / other, new_duals)
+ raise ValueError
+
+ def __floordiv__(self, other):
+ if not isinstance(other, Dual):
+ new_duals = []
+ for i in self.duals:
+ new_duals.append(i // other)
+ return Dual(self.real // other, new_duals)
+ raise ValueError
+
+ def __pow__(self, n):
+ if n < 0 or isinstance(n, float):
+ raise ValueError("power must be a positive integer")
+ if n == 0:
+ return 1
+ if n == 1:
+ return self
+ x = self
+ for _ in range(n - 1):
+ x *= self
+ return x
+
+
+def differentiate(func, position, order):
+ """
+ >>> differentiate(lambda x: x**2, 2, 2)
+ 2
+ >>> differentiate(lambda x: x**2 * x**4, 9, 2)
+ 196830
+ >>> differentiate(lambda y: 0.5 * (y + 3) ** 6, 3.5, 4)
+ 7605.0
+ >>> differentiate(lambda y: y ** 2, 4, 3)
+ 0
+ >>> differentiate(8, 8, 8)
+ Traceback (most recent call last):
+ ...
+ ValueError: differentiate() requires a function as input for func
+ >>> differentiate(lambda x: x **2, "", 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: differentiate() requires a float as input for position
+ >>> differentiate(lambda x: x**2, 3, "")
+ Traceback (most recent call last):
+ ...
+ ValueError: differentiate() requires an int as input for order
+ """
+ if not callable(func):
+ raise ValueError("differentiate() requires a function as input for func")
+ if not isinstance(position, (float, int)):
+ raise ValueError("differentiate() requires a float as input for position")
+ if not isinstance(order, int):
+ raise ValueError("differentiate() requires an int as input for order")
+ d = Dual(position, 1)
+ result = func(d)
+ if order == 0:
+ return result.real
+ return result.duals[order - 1] * factorial(order)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ def f(y):
+ return y**2 * y**4
+
+ print(differentiate(f, 9, 2))
diff --git a/maths/entropy.py b/maths/entropy.py
index 43bb3860fc12..b816f1d193f7 100644
--- a/maths/entropy.py
+++ b/maths/entropy.py
@@ -4,6 +4,7 @@
Implementation of entropy of information
https://en.wikipedia.org/wiki/Entropy_(information_theory)
"""
+
from __future__ import annotations
import math
@@ -20,10 +21,10 @@ def calculate_prob(text: str) -> None:
:return: Prints
1) Entropy of information based on 1 alphabet
2) Entropy of information based on couples of 2 alphabet
- 3) print Entropy of H(X n∣Xn−1)
+ 3) print Entropy of H(X n|Xn-1)
Text from random books. Also, random quotes.
- >>> text = ("Behind Winston’s back the voice "
+ >>> text = ("Behind Winston's back the voice "
... "from the telescreen was still "
... "babbling and the overfulfilment")
>>> calculate_prob(text)
@@ -68,7 +69,7 @@ def calculate_prob(text: str) -> None:
my_fir_sum += prob * math.log2(prob) # entropy formula.
# print entropy
- print("{:.1f}".format(round(-1 * my_fir_sum)))
+ print(f"{round(-1 * my_fir_sum):.1f}")
# two len string
all_sum = sum(two_char_strings.values())
@@ -83,10 +84,10 @@ def calculate_prob(text: str) -> None:
my_sec_sum += prob * math.log2(prob)
# print second entropy
- print("{:.1f}".format(round(-1 * my_sec_sum)))
+ print(f"{round(-1 * my_sec_sum):.1f}")
# print the difference between them
- print("{:.1f}".format(round((-1 * my_sec_sum) - (-1 * my_fir_sum))))
+ print(f"{round((-1 * my_sec_sum) - (-1 * my_fir_sum)):.1f}")
def analyze_text(text: str) -> tuple[dict, dict]:
@@ -95,13 +96,13 @@ def analyze_text(text: str) -> tuple[dict, dict]:
The first dictionary stores the frequency of single character strings.
The second dictionary stores the frequency of two character strings.
"""
- single_char_strings = Counter() # type: ignore
- two_char_strings = Counter() # type: ignore
+ single_char_strings = Counter() # type: ignore[var-annotated]
+ two_char_strings = Counter() # type: ignore[var-annotated]
single_char_strings[text[-1]] += 1
# first case when we have space at start.
two_char_strings[" " + text[0]] += 1
- for i in range(0, len(text) - 1):
+ for i in range(len(text) - 1):
single_char_strings[text[i]] += 1
two_char_strings[text[i : i + 2]] += 1
return single_char_strings, two_char_strings
diff --git a/maths/euclidean_distance.py b/maths/euclidean_distance.py
new file mode 100644
index 000000000000..aa7f3efc7684
--- /dev/null
+++ b/maths/euclidean_distance.py
@@ -0,0 +1,65 @@
+from __future__ import annotations
+
+import typing
+from collections.abc import Iterable
+
+import numpy as np
+
+Vector = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007
+VectorOut = typing.Union[np.float64, int, float] # noqa: UP007
+
+
+def euclidean_distance(vector_1: Vector, vector_2: Vector) -> VectorOut:
+ """
+ Calculate the distance between the two endpoints of two vectors.
+ A vector is defined as a list, tuple, or numpy 1D array.
+ >>> float(euclidean_distance((0, 0), (2, 2)))
+ 2.8284271247461903
+ >>> float(euclidean_distance(np.array([0, 0, 0]), np.array([2, 2, 2])))
+ 3.4641016151377544
+ >>> float(euclidean_distance(np.array([1, 2, 3, 4]), np.array([5, 6, 7, 8])))
+ 8.0
+ >>> float(euclidean_distance([1, 2, 3, 4], [5, 6, 7, 8]))
+ 8.0
+ """
+ return np.sqrt(np.sum((np.asarray(vector_1) - np.asarray(vector_2)) ** 2))
+
+
+def euclidean_distance_no_np(vector_1: Vector, vector_2: Vector) -> VectorOut:
+ """
+ Calculate the distance between the two endpoints of two vectors without numpy.
+ A vector is defined as a list, tuple, or numpy 1D array.
+ >>> euclidean_distance_no_np((0, 0), (2, 2))
+ 2.8284271247461903
+ >>> euclidean_distance_no_np([1, 2, 3, 4], [5, 6, 7, 8])
+ 8.0
+ """
+ return sum((v1 - v2) ** 2 for v1, v2 in zip(vector_1, vector_2)) ** (1 / 2)
+
+
+if __name__ == "__main__":
+
+ def benchmark() -> None:
+ """
+ Benchmarks
+ """
+ from timeit import timeit
+
+ print("Without Numpy")
+ print(
+ timeit(
+ "euclidean_distance_no_np([1, 2, 3], [4, 5, 6])",
+ number=10000,
+ globals=globals(),
+ )
+ )
+ print("With Numpy")
+ print(
+ timeit(
+ "euclidean_distance([1, 2, 3], [4, 5, 6])",
+ number=10000,
+ globals=globals(),
+ )
+ )
+
+ benchmark()
diff --git a/maths/euler_method.py b/maths/euler_method.py
new file mode 100644
index 000000000000..c6adb07e2d3d
--- /dev/null
+++ b/maths/euler_method.py
@@ -0,0 +1,47 @@
+from collections.abc import Callable
+
+import numpy as np
+
+
+def explicit_euler(
+ ode_func: Callable, y0: float, x0: float, step_size: float, x_end: float
+) -> np.ndarray:
+ """Calculate numeric solution at each step to an ODE using Euler's Method
+
+ For reference to Euler's method refer to https://en.wikipedia.org/wiki/Euler_method.
+
+ Args:
+ ode_func (Callable): The ordinary differential equation
+ as a function of x and y.
+ y0 (float): The initial value for y.
+ x0 (float): The initial value for x.
+ step_size (float): The increment value for x.
+ x_end (float): The final value of x to be calculated.
+
+ Returns:
+ np.ndarray: Solution of y for every step in x.
+
+ >>> # the exact solution is math.exp(x)
+ >>> def f(x, y):
+ ... return y
+ >>> y0 = 1
+ >>> y = explicit_euler(f, y0, 0.0, 0.01, 5)
+ >>> float(y[-1])
+ 144.77277243257308
+ """
+ n = int(np.ceil((x_end - x0) / step_size))
+ y = np.zeros((n + 1,))
+ y[0] = y0
+ x = x0
+
+ for k in range(n):
+ y[k + 1] = y[k] + step_size * ode_func(x, y[k])
+ x += step_size
+
+ return y
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/euler_modified.py b/maths/euler_modified.py
new file mode 100644
index 000000000000..bb282e9f0ab9
--- /dev/null
+++ b/maths/euler_modified.py
@@ -0,0 +1,54 @@
+from collections.abc import Callable
+
+import numpy as np
+
+
+def euler_modified(
+ ode_func: Callable, y0: float, x0: float, step_size: float, x_end: float
+) -> np.ndarray:
+ """
+ Calculate solution at each step to an ODE using Euler's Modified Method
+ The Euler Method is straightforward to implement, but can't give accurate solutions.
+ So, some changes were proposed to improve accuracy.
+
+ https://en.wikipedia.org/wiki/Euler_method
+
+ Arguments:
+ ode_func -- The ode as a function of x and y
+ y0 -- the initial value for y
+ x0 -- the initial value for x
+ stepsize -- the increment value for x
+ x_end -- the end value for x
+
+ >>> # the exact solution is math.exp(x)
+ >>> def f1(x, y):
+ ... return -2*x*(y**2)
+ >>> y = euler_modified(f1, 1.0, 0.0, 0.2, 1.0)
+ >>> float(y[-1])
+ 0.503338255442106
+ >>> import math
+ >>> def f2(x, y):
+ ... return -2*y + (x**3)*math.exp(-2*x)
+ >>> y = euler_modified(f2, 1.0, 0.0, 0.1, 0.3)
+ >>> float(y[-1])
+ 0.5525976431951775
+ """
+ n = int(np.ceil((x_end - x0) / step_size))
+ y = np.zeros((n + 1,))
+ y[0] = y0
+ x = x0
+
+ for k in range(n):
+ y_get = y[k] + step_size * ode_func(x, y[k])
+ y[k + 1] = y[k] + (
+ (step_size / 2) * (ode_func(x, y[k]) + ode_func(x + step_size, y_get))
+ )
+ x += step_size
+
+ return y
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/eulers_totient.py b/maths/eulers_totient.py
index 6a35e69bde0b..00f0254c215a 100644
--- a/maths/eulers_totient.py
+++ b/maths/eulers_totient.py
@@ -1,12 +1,27 @@
# Eulers Totient function finds the number of relative primes of a number n from 1 to n
def totient(n: int) -> list:
+ """
+ >>> n = 10
+ >>> totient_calculation = totient(n)
+ >>> for i in range(1, n):
+ ... print(f"{i} has {totient_calculation[i]} relative primes.")
+ 1 has 0 relative primes.
+ 2 has 1 relative primes.
+ 3 has 2 relative primes.
+ 4 has 2 relative primes.
+ 5 has 4 relative primes.
+ 6 has 2 relative primes.
+ 7 has 6 relative primes.
+ 8 has 4 relative primes.
+ 9 has 6 relative primes.
+ """
is_prime = [True for i in range(n + 1)]
totients = [i - 1 for i in range(n + 1)]
primes = []
for i in range(2, n + 1):
if is_prime[i]:
primes.append(i)
- for j in range(0, len(primes)):
+ for j in range(len(primes)):
if i * primes[j] >= n:
break
is_prime[i * primes[j]] = False
@@ -20,25 +35,6 @@ def totient(n: int) -> list:
return totients
-def test_totient() -> None:
- """
- >>> n = 10
- >>> totient_calculation = totient(n)
- >>> for i in range(1, n):
- ... print(f"{i} has {totient_calculation[i]} relative primes.")
- 1 has 0 relative primes.
- 2 has 1 relative primes.
- 3 has 2 relative primes.
- 4 has 2 relative primes.
- 5 has 4 relative primes.
- 6 has 2 relative primes.
- 7 has 6 relative primes.
- 8 has 4 relative primes.
- 9 has 6 relative primes.
- """
- pass
-
-
if __name__ == "__main__":
import doctest
diff --git a/maths/explicit_euler.py b/maths/explicit_euler.py
deleted file mode 100644
index 7c780198602b..000000000000
--- a/maths/explicit_euler.py
+++ /dev/null
@@ -1,40 +0,0 @@
-import numpy as np
-
-
-def explicit_euler(ode_func, y0, x0, step_size, x_end):
- """
- Calculate numeric solution at each step to an ODE using Euler's Method
-
- https://en.wikipedia.org/wiki/Euler_method
-
- Arguments:
- ode_func -- The ode as a function of x and y
- y0 -- the initial value for y
- x0 -- the initial value for x
- stepsize -- the increment value for x
- x_end -- the end value for x
-
- >>> # the exact solution is math.exp(x)
- >>> def f(x, y):
- ... return y
- >>> y0 = 1
- >>> y = explicit_euler(f, y0, 0.0, 0.01, 5)
- >>> y[-1]
- 144.77277243257308
- """
- N = int(np.ceil((x_end - x0) / step_size))
- y = np.zeros((N + 1,))
- y[0] = y0
- x = x0
-
- for k in range(N):
- y[k + 1] = y[k] + step_size * ode_func(x, y[k])
- x += step_size
-
- return y
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/maths/extended_euclidean_algorithm.py b/maths/extended_euclidean_algorithm.py
index e7087636ce09..c54909e19101 100644
--- a/maths/extended_euclidean_algorithm.py
+++ b/maths/extended_euclidean_algorithm.py
@@ -12,12 +12,12 @@
# @Email: silentcat@protonmail.com
# @Last modified by: pikulet
# @Last modified time: 2020-10-02
+from __future__ import annotations
import sys
-from typing import Tuple
-def extended_euclidean_algorithm(a: int, b: int) -> Tuple[int, int]:
+def extended_euclidean_algorithm(a: int, b: int) -> tuple[int, int]:
"""
Extended Euclidean Algorithm.
@@ -75,11 +75,12 @@ def main():
"""Call Extended Euclidean Algorithm."""
if len(sys.argv) < 3:
print("2 integer arguments required")
- exit(1)
+ return 1
a = int(sys.argv[1])
b = int(sys.argv[2])
print(extended_euclidean_algorithm(a, b))
+ return 0
if __name__ == "__main__":
- main()
+ raise SystemExit(main())
diff --git a/maths/factorial.py b/maths/factorial.py
new file mode 100644
index 000000000000..aaf90f384bb9
--- /dev/null
+++ b/maths/factorial.py
@@ -0,0 +1,68 @@
+"""
+Factorial of a positive integer -- https://en.wikipedia.org/wiki/Factorial
+"""
+
+
+def factorial(number: int) -> int:
+ """
+ Calculate the factorial of specified number (n!).
+
+ >>> import math
+ >>> all(factorial(i) == math.factorial(i) for i in range(20))
+ True
+ >>> factorial(0.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: factorial() only accepts integral values
+ >>> factorial(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: factorial() not defined for negative values
+ >>> factorial(1)
+ 1
+ >>> factorial(6)
+ 720
+ >>> factorial(0)
+ 1
+ """
+ if number != int(number):
+ raise ValueError("factorial() only accepts integral values")
+ if number < 0:
+ raise ValueError("factorial() not defined for negative values")
+ value = 1
+ for i in range(1, number + 1):
+ value *= i
+ return value
+
+
+def factorial_recursive(n: int) -> int:
+ """
+ Calculate the factorial of a positive integer
+ https://en.wikipedia.org/wiki/Factorial
+
+ >>> import math
+ >>> all(factorial(i) == math.factorial(i) for i in range(20))
+ True
+ >>> factorial(0.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: factorial() only accepts integral values
+ >>> factorial(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: factorial() not defined for negative values
+ """
+ if not isinstance(n, int):
+ raise ValueError("factorial() only accepts integral values")
+ if n < 0:
+ raise ValueError("factorial() not defined for negative values")
+ return 1 if n in {0, 1} else n * factorial(n - 1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ n = int(input("Enter a positive integer: ").strip() or 0)
+ print(f"factorial{n} is {factorial(n)}")
diff --git a/maths/factorial_iterative.py b/maths/factorial_iterative.py
deleted file mode 100644
index 64314790c11c..000000000000
--- a/maths/factorial_iterative.py
+++ /dev/null
@@ -1,30 +0,0 @@
-# factorial of a positive integer -- https://en.wikipedia.org/wiki/Factorial
-
-
-def factorial(n: int) -> int:
- """
- >>> import math
- >>> all(factorial(i) == math.factorial(i) for i in range(20))
- True
- >>> factorial(0.1)
- Traceback (most recent call last):
- ...
- ValueError: factorial() only accepts integral values
- >>> factorial(-1)
- Traceback (most recent call last):
- ...
- ValueError: factorial() not defined for negative values
- """
- if n != int(n):
- raise ValueError("factorial() only accepts integral values")
- if n < 0:
- raise ValueError("factorial() not defined for negative values")
- value = 1
- for i in range(1, n + 1):
- value *= i
- return value
-
-
-if __name__ == "__main__":
- n = int(input("Enter a positive integer: ").strip() or 0)
- print(f"factorial{n} is {factorial(n)}")
diff --git a/maths/factorial_python.py b/maths/factorial_python.py
deleted file mode 100644
index 46688261af56..000000000000
--- a/maths/factorial_python.py
+++ /dev/null
@@ -1,34 +0,0 @@
-def factorial(input_number: int) -> int:
- """
- Calculate the factorial of specified number
-
- >>> factorial(1)
- 1
- >>> factorial(6)
- 720
- >>> factorial(0)
- 1
- >>> factorial(-1)
- Traceback (most recent call last):
- ...
- ValueError: factorial() not defined for negative values
- >>> factorial(0.1)
- Traceback (most recent call last):
- ...
- ValueError: factorial() only accepts integral values
- """
-
- if input_number < 0:
- raise ValueError("factorial() not defined for negative values")
- if not isinstance(input_number, int):
- raise ValueError("factorial() only accepts integral values")
- result = 1
- for i in range(1, input_number):
- result = result * (i + 1)
- return result
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/maths/factorial_recursive.py b/maths/factorial_recursive.py
deleted file mode 100644
index 137112738905..000000000000
--- a/maths/factorial_recursive.py
+++ /dev/null
@@ -1,28 +0,0 @@
-def factorial(n: int) -> int:
- """
- Calculate the factorial of a positive integer
- https://en.wikipedia.org/wiki/Factorial
-
- >>> import math
- >>> all(factorial(i) == math.factorial(i) for i in range(20))
- True
- >>> factorial(0.1)
- Traceback (most recent call last):
- ...
- ValueError: factorial() only accepts integral values
- >>> factorial(-1)
- Traceback (most recent call last):
- ...
- ValueError: factorial() not defined for negative values
- """
- if not isinstance(n, int):
- raise ValueError("factorial() only accepts integral values")
- if n < 0:
- raise ValueError("factorial() not defined for negative values")
- return 1 if n == 0 or n == 1 else n * factorial(n - 1)
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
diff --git a/maths/factors.py b/maths/factors.py
index e2fdc4063a13..ae2e5316cf65 100644
--- a/maths/factors.py
+++ b/maths/factors.py
@@ -1,3 +1,7 @@
+from doctest import testmod
+from math import sqrt
+
+
def factors_of_a_number(num: int) -> list:
"""
>>> factors_of_a_number(1)
@@ -9,10 +13,22 @@ def factors_of_a_number(num: int) -> list:
>>> factors_of_a_number(-24)
[]
"""
- return [i for i in range(1, num + 1) if num % i == 0]
+ facs: list[int] = []
+ if num < 1:
+ return facs
+ facs.append(1)
+ if num == 1:
+ return facs
+ facs.append(num)
+ for i in range(2, int(sqrt(num)) + 1):
+ if num % i == 0: # If i is a factor of num
+ facs.append(i)
+ d = num // i # num//i is the other factor of num
+ if d != i: # If d and i are distinct
+ facs.append(d) # we have found another factor
+ facs.sort()
+ return facs
if __name__ == "__main__":
- num = int(input("Enter a number to find its factors: "))
- factors = factors_of_a_number(num)
- print(f"{num} has {len(factors)} factors: {', '.join(str(f) for f in factors)}")
+ testmod(name="factors_of_a_number", verbose=True)
diff --git a/maths/fast_inverse_sqrt.py b/maths/fast_inverse_sqrt.py
new file mode 100644
index 000000000000..79385bb84877
--- /dev/null
+++ b/maths/fast_inverse_sqrt.py
@@ -0,0 +1,54 @@
+"""
+Fast inverse square root (1/sqrt(x)) using the Quake III algorithm.
+Reference: https://en.wikipedia.org/wiki/Fast_inverse_square_root
+Accuracy: https://en.wikipedia.org/wiki/Fast_inverse_square_root#Accuracy
+"""
+
+import struct
+
+
+def fast_inverse_sqrt(number: float) -> float:
+ """
+ Compute the fast inverse square root of a floating-point number using the famous
+ Quake III algorithm.
+
+ :param float number: Input number for which to calculate the inverse square root.
+ :return float: The fast inverse square root of the input number.
+
+ Example:
+ >>> fast_inverse_sqrt(10)
+ 0.3156857923527257
+ >>> fast_inverse_sqrt(4)
+ 0.49915357479239103
+ >>> fast_inverse_sqrt(4.1)
+ 0.4932849504615651
+ >>> fast_inverse_sqrt(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive number.
+ >>> fast_inverse_sqrt(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive number.
+ >>> from math import isclose, sqrt
+ >>> all(isclose(fast_inverse_sqrt(i), 1 / sqrt(i), rel_tol=0.00132)
+ ... for i in range(50, 60))
+ True
+ """
+ if number <= 0:
+ raise ValueError("Input must be a positive number.")
+ i = struct.unpack(">i", struct.pack(">f", number))[0]
+ i = 0x5F3759DF - (i >> 1)
+ y = struct.unpack(">f", struct.pack(">i", i))[0]
+ return y * (1.5 - 0.5 * number * y * y)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ # https://en.wikipedia.org/wiki/Fast_inverse_square_root#Accuracy
+ from math import sqrt
+
+ for i in range(5, 101, 5):
+ print(f"{i:>3}: {(1 / sqrt(i)) - fast_inverse_sqrt(i):.5f}")
diff --git a/maths/fermat_little_theorem.py b/maths/fermat_little_theorem.py
index 73af3e28c618..4a3ecd05ce91 100644
--- a/maths/fermat_little_theorem.py
+++ b/maths/fermat_little_theorem.py
@@ -5,8 +5,7 @@
# Wikipedia reference: https://en.wikipedia.org/wiki/Fermat%27s_little_theorem
-def binary_exponentiation(a, n, mod):
-
+def binary_exponentiation(a: int, n: float, mod: int) -> int:
if n == 0:
return 1
diff --git a/maths/fibonacci.py b/maths/fibonacci.py
index e6519035401e..24b2d7ae449e 100644
--- a/maths/fibonacci.py
+++ b/maths/fibonacci.py
@@ -1,130 +1,332 @@
-# fibonacci.py
-"""
-1. Calculates the iterative fibonacci sequence
-
-2. Calculates the fibonacci sequence with a formula
- an = [ Phin - (phi)n ]/Sqrt[5]
- reference-->Su, Francis E., et al. "Fibonacci Number Formula." Math Fun Facts.
-
-"""
-import functools
-import math
-import time
-from decimal import Decimal, getcontext
-
-getcontext().prec = 100
-
-
-def timer_decorator(func):
- @functools.wraps(func)
- def timer_wrapper(*args, **kwargs):
- start = time.time()
- func(*args, **kwargs)
- end = time.time()
- if int(end - start) > 0:
- print(f"Run time for {func.__name__}: {(end - start):0.2f}s")
- else:
- print(f"Run time for {func.__name__}: {(end - start)*1000:0.2f}ms")
- return func(*args, **kwargs)
-
- return timer_wrapper
-
-
-# define Python user-defined exceptions
-class Error(Exception):
- """Base class for other exceptions"""
-
-
-class ValueTooLargeError(Error):
- """Raised when the input value is too large"""
-
-
-class ValueTooSmallError(Error):
- """Raised when the input value is not greater than one"""
-
-
-class ValueLessThanZero(Error):
- """Raised when the input value is less than zero"""
-
-
-def _check_number_input(n, min_thresh, max_thresh=None):
- """
- :param n: single integer
- :type n: int
- :param min_thresh: min threshold, single integer
- :type min_thresh: int
- :param max_thresh: max threshold, single integer
- :type max_thresh: int
- :return: boolean
- """
- try:
- if n >= min_thresh and max_thresh is None:
- return True
- elif min_thresh <= n <= max_thresh:
- return True
- elif n < 0:
- raise ValueLessThanZero
- elif n < min_thresh:
- raise ValueTooSmallError
- elif n > max_thresh:
- raise ValueTooLargeError
- except ValueLessThanZero:
- print("Incorrect Input: number must not be less than 0")
- except ValueTooSmallError:
- print(
- f"Incorrect Input: input number must be > {min_thresh} for the recursive "
- "calculation"
- )
- except ValueTooLargeError:
- print(
- f"Incorrect Input: input number must be < {max_thresh} for the recursive "
- "calculation"
- )
- return False
-
-
-@timer_decorator
-def fib_iterative(n):
- """
- :param n: calculate Fibonacci to the nth integer
- :type n:int
- :return: Fibonacci sequence as a list
- """
- n = int(n)
- if _check_number_input(n, 2):
- seq_out = [0, 1]
- a, b = 0, 1
- for _ in range(n - len(seq_out)):
- a, b = b, a + b
- seq_out.append(b)
- return seq_out
-
-
-@timer_decorator
-def fib_formula(n):
- """
- :param n: calculate Fibonacci to the nth integer
- :type n:int
- :return: Fibonacci sequence as a list
- """
- seq_out = [0, 1]
- n = int(n)
- if _check_number_input(n, 2, 1000000):
- sqrt = Decimal(math.sqrt(5))
- phi_1 = Decimal(1 + sqrt) / Decimal(2)
- phi_2 = Decimal(1 - sqrt) / Decimal(2)
- for i in range(2, n):
- temp_out = ((phi_1 ** Decimal(i)) - (phi_2 ** Decimal(i))) * (
- Decimal(sqrt) ** Decimal(-1)
- )
- seq_out.append(int(temp_out))
- return seq_out
-
-
-if __name__ == "__main__":
- num = 20
- # print(f'{fib_recursive(num)}\n')
- # print(f'{fib_iterative(num)}\n')
- # print(f'{fib_formula(num)}\n')
- fib_iterative(num)
- fib_formula(num)
+"""
+Calculates the Fibonacci sequence using iteration, recursion, memoization,
+and a simplified form of Binet's formula
+
+NOTE 1: the iterative, recursive, memoization functions are more accurate than
+the Binet's formula function because the Binet formula function uses floats
+
+NOTE 2: the Binet's formula function is much more limited in the size of inputs
+that it can handle due to the size limitations of Python floats
+NOTE 3: the matrix function is the fastest and most memory efficient for large n
+
+
+See benchmark numbers in __main__ for performance comparisons/
+https://en.wikipedia.org/wiki/Fibonacci_number for more information
+"""
+
+import functools
+from collections.abc import Iterator
+from math import sqrt
+from time import time
+
+import numpy as np
+from numpy import ndarray
+
+
+def time_func(func, *args, **kwargs):
+ """
+ Times the execution of a function with parameters
+ """
+ start = time()
+ output = func(*args, **kwargs)
+ end = time()
+ if int(end - start) > 0:
+ print(f"{func.__name__} runtime: {(end - start):0.4f} s")
+ else:
+ print(f"{func.__name__} runtime: {(end - start) * 1000:0.4f} ms")
+ return output
+
+
+def fib_iterative_yield(n: int) -> Iterator[int]:
+ """
+ Calculates the first n (1-indexed) Fibonacci numbers using iteration with yield
+ >>> list(fib_iterative_yield(0))
+ [0]
+ >>> tuple(fib_iterative_yield(1))
+ (0, 1)
+ >>> tuple(fib_iterative_yield(5))
+ (0, 1, 1, 2, 3, 5)
+ >>> tuple(fib_iterative_yield(10))
+ (0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55)
+ >>> tuple(fib_iterative_yield(-1))
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ """
+ if n < 0:
+ raise ValueError("n is negative")
+ a, b = 0, 1
+ yield a
+ for _ in range(n):
+ yield b
+ a, b = b, a + b
+
+
+def fib_iterative(n: int) -> list[int]:
+ """
+ Calculates the first n (0-indexed) Fibonacci numbers using iteration
+ >>> fib_iterative(0)
+ [0]
+ >>> fib_iterative(1)
+ [0, 1]
+ >>> fib_iterative(5)
+ [0, 1, 1, 2, 3, 5]
+ >>> fib_iterative(10)
+ [0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55]
+ >>> fib_iterative(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ """
+ if n < 0:
+ raise ValueError("n is negative")
+ if n == 0:
+ return [0]
+ fib = [0, 1]
+ for _ in range(n - 1):
+ fib.append(fib[-1] + fib[-2])
+ return fib
+
+
+def fib_recursive(n: int) -> list[int]:
+ """
+ Calculates the first n (0-indexed) Fibonacci numbers using recursion
+ >>> fib_iterative(0)
+ [0]
+ >>> fib_iterative(1)
+ [0, 1]
+ >>> fib_iterative(5)
+ [0, 1, 1, 2, 3, 5]
+ >>> fib_iterative(10)
+ [0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55]
+ >>> fib_iterative(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ """
+
+ def fib_recursive_term(i: int) -> int:
+ """
+ Calculates the i-th (0-indexed) Fibonacci number using recursion
+ >>> fib_recursive_term(0)
+ 0
+ >>> fib_recursive_term(1)
+ 1
+ >>> fib_recursive_term(5)
+ 5
+ >>> fib_recursive_term(10)
+ 55
+ >>> fib_recursive_term(-1)
+ Traceback (most recent call last):
+ ...
+ Exception: n is negative
+ """
+ if i < 0:
+ raise ValueError("n is negative")
+ if i < 2:
+ return i
+ return fib_recursive_term(i - 1) + fib_recursive_term(i - 2)
+
+ if n < 0:
+ raise ValueError("n is negative")
+ return [fib_recursive_term(i) for i in range(n + 1)]
+
+
+def fib_recursive_cached(n: int) -> list[int]:
+ """
+ Calculates the first n (0-indexed) Fibonacci numbers using recursion
+ >>> fib_iterative(0)
+ [0]
+ >>> fib_iterative(1)
+ [0, 1]
+ >>> fib_iterative(5)
+ [0, 1, 1, 2, 3, 5]
+ >>> fib_iterative(10)
+ [0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55]
+ >>> fib_iterative(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ """
+
+ @functools.cache
+ def fib_recursive_term(i: int) -> int:
+ """
+ Calculates the i-th (0-indexed) Fibonacci number using recursion
+ """
+ if i < 0:
+ raise ValueError("n is negative")
+ if i < 2:
+ return i
+ return fib_recursive_term(i - 1) + fib_recursive_term(i - 2)
+
+ if n < 0:
+ raise ValueError("n is negative")
+ return [fib_recursive_term(i) for i in range(n + 1)]
+
+
+def fib_memoization(n: int) -> list[int]:
+ """
+ Calculates the first n (0-indexed) Fibonacci numbers using memoization
+ >>> fib_memoization(0)
+ [0]
+ >>> fib_memoization(1)
+ [0, 1]
+ >>> fib_memoization(5)
+ [0, 1, 1, 2, 3, 5]
+ >>> fib_memoization(10)
+ [0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55]
+ >>> fib_iterative(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ """
+ if n < 0:
+ raise ValueError("n is negative")
+ # Cache must be outside recursuive function
+ # other it will reset every time it calls itself.
+ cache: dict[int, int] = {0: 0, 1: 1, 2: 1} # Prefilled cache
+
+ def rec_fn_memoized(num: int) -> int:
+ if num in cache:
+ return cache[num]
+
+ value = rec_fn_memoized(num - 1) + rec_fn_memoized(num - 2)
+ cache[num] = value
+ return value
+
+ return [rec_fn_memoized(i) for i in range(n + 1)]
+
+
+def fib_binet(n: int) -> list[int]:
+ """
+ Calculates the first n (0-indexed) Fibonacci numbers using a simplified form
+ of Binet's formula:
+ https://en.m.wikipedia.org/wiki/Fibonacci_number#Computation_by_rounding
+
+ NOTE 1: this function diverges from fib_iterative at around n = 71, likely
+ due to compounding floating-point arithmetic errors
+
+ NOTE 2: this function doesn't accept n >= 1475 because it overflows
+ thereafter due to the size limitations of Python floats
+ >>> fib_binet(0)
+ [0]
+ >>> fib_binet(1)
+ [0, 1]
+ >>> fib_binet(5)
+ [0, 1, 1, 2, 3, 5]
+ >>> fib_binet(10)
+ [0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55]
+ >>> fib_binet(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ >>> fib_binet(1475)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is too large
+ """
+ if n < 0:
+ raise ValueError("n is negative")
+ if n >= 1475:
+ raise ValueError("n is too large")
+ sqrt_5 = sqrt(5)
+ phi = (1 + sqrt_5) / 2
+ return [round(phi**i / sqrt_5) for i in range(n + 1)]
+
+
+def matrix_pow_np(m: ndarray, power: int) -> ndarray:
+ """
+ Raises a matrix to the power of 'power' using binary exponentiation.
+
+ Args:
+ m: Matrix as a numpy array.
+ power: The power to which the matrix is to be raised.
+
+ Returns:
+ The matrix raised to the power.
+
+ Raises:
+ ValueError: If power is negative.
+
+ >>> m = np.array([[1, 1], [1, 0]], dtype=int)
+ >>> matrix_pow_np(m, 0) # Identity matrix when raised to the power of 0
+ array([[1, 0],
+ [0, 1]])
+
+ >>> matrix_pow_np(m, 1) # Same matrix when raised to the power of 1
+ array([[1, 1],
+ [1, 0]])
+
+ >>> matrix_pow_np(m, 5)
+ array([[8, 5],
+ [5, 3]])
+
+ >>> matrix_pow_np(m, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: power is negative
+ """
+ result = np.array([[1, 0], [0, 1]], dtype=int) # Identity Matrix
+ base = m
+ if power < 0: # Negative power is not allowed
+ raise ValueError("power is negative")
+ while power:
+ if power % 2 == 1:
+ result = np.dot(result, base)
+ base = np.dot(base, base)
+ power //= 2
+ return result
+
+
+def fib_matrix_np(n: int) -> int:
+ """
+ Calculates the n-th Fibonacci number using matrix exponentiation.
+ https://www.nayuki.io/page/fast-fibonacci-algorithms#:~:text=
+ Summary:%20The%20two%20fast%20Fibonacci%20algorithms%20are%20matrix
+
+ Args:
+ n: Fibonacci sequence index
+
+ Returns:
+ The n-th Fibonacci number.
+
+ Raises:
+ ValueError: If n is negative.
+
+ >>> fib_matrix_np(0)
+ 0
+ >>> fib_matrix_np(1)
+ 1
+ >>> fib_matrix_np(5)
+ 5
+ >>> fib_matrix_np(10)
+ 55
+ >>> fib_matrix_np(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: n is negative
+ """
+ if n < 0:
+ raise ValueError("n is negative")
+ if n == 0:
+ return 0
+
+ m = np.array([[1, 1], [1, 0]], dtype=int)
+ result = matrix_pow_np(m, n - 1)
+ return int(result[0, 0])
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ # Time on an M1 MacBook Pro -- Fastest to slowest
+ num = 30
+ time_func(fib_iterative_yield, num) # 0.0012 ms
+ time_func(fib_iterative, num) # 0.0031 ms
+ time_func(fib_binet, num) # 0.0062 ms
+ time_func(fib_memoization, num) # 0.0100 ms
+ time_func(fib_recursive_cached, num) # 0.0153 ms
+ time_func(fib_recursive, num) # 257.0910 ms
+ time_func(fib_matrix_np, num) # 0.0000 ms
diff --git a/maths/fibonacci_sequence_recursion.py b/maths/fibonacci_sequence_recursion.py
deleted file mode 100644
index 794b9fc0bd3a..000000000000
--- a/maths/fibonacci_sequence_recursion.py
+++ /dev/null
@@ -1,22 +0,0 @@
-# Fibonacci Sequence Using Recursion
-
-
-def recur_fibo(n: int) -> int:
- """
- >>> [recur_fibo(i) for i in range(12)]
- [0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89]
- """
- return n if n <= 1 else recur_fibo(n - 1) + recur_fibo(n - 2)
-
-
-def main() -> None:
- limit = int(input("How many terms to include in fibonacci series: "))
- if limit > 0:
- print(f"The first {limit} terms of the fibonacci series are as follows:")
- print([recur_fibo(n) for n in range(limit)])
- else:
- print("Please enter a positive integer: ")
-
-
-if __name__ == "__main__":
- main()
diff --git a/maths/find_max.py b/maths/find_max.py
index 4d92e37eb2e1..4765d300634e 100644
--- a/maths/find_max.py
+++ b/maths/find_max.py
@@ -1,25 +1,84 @@
-# NguyenU
+from __future__ import annotations
-def find_max(nums):
+def find_max_iterative(nums: list[int | float]) -> int | float:
"""
>>> for nums in ([3, 2, 1], [-3, -2, -1], [3, -3, 0], [3.0, 3.1, 2.9]):
- ... find_max(nums) == max(nums)
+ ... find_max_iterative(nums) == max(nums)
True
True
True
True
+ >>> find_max_iterative([2, 4, 9, 7, 19, 94, 5])
+ 94
+ >>> find_max_iterative([])
+ Traceback (most recent call last):
+ ...
+ ValueError: find_max_iterative() arg is an empty sequence
"""
+ if len(nums) == 0:
+ raise ValueError("find_max_iterative() arg is an empty sequence")
max_num = nums[0]
for x in nums:
- if x > max_num:
+ if x > max_num: # noqa: PLR1730
max_num = x
return max_num
-def main():
- print(find_max([2, 4, 9, 7, 19, 94, 5])) # 94
+# Divide and Conquer algorithm
+def find_max_recursive(nums: list[int | float], left: int, right: int) -> int | float:
+ """
+ find max value in list
+ :param nums: contains elements
+ :param left: index of first element
+ :param right: index of last element
+ :return: max in nums
+
+ >>> for nums in ([3, 2, 1], [-3, -2, -1], [3, -3, 0], [3.0, 3.1, 2.9]):
+ ... find_max_recursive(nums, 0, len(nums) - 1) == max(nums)
+ True
+ True
+ True
+ True
+ >>> nums = [1, 3, 5, 7, 9, 2, 4, 6, 8, 10]
+ >>> find_max_recursive(nums, 0, len(nums) - 1) == max(nums)
+ True
+ >>> find_max_recursive([], 0, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: find_max_recursive() arg is an empty sequence
+ >>> find_max_recursive(nums, 0, len(nums)) == max(nums)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> find_max_recursive(nums, -len(nums), -1) == max(nums)
+ True
+ >>> find_max_recursive(nums, -len(nums) - 1, -1) == max(nums)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ """
+ if len(nums) == 0:
+ raise ValueError("find_max_recursive() arg is an empty sequence")
+ if (
+ left >= len(nums)
+ or left < -len(nums)
+ or right >= len(nums)
+ or right < -len(nums)
+ ):
+ raise IndexError("list index out of range")
+ if left == right:
+ return nums[left]
+ mid = (left + right) >> 1 # the middle
+ left_max = find_max_recursive(nums, left, mid) # find max in range[left, mid]
+ right_max = find_max_recursive(
+ nums, mid + 1, right
+ ) # find max in range[mid + 1, right]
+
+ return left_max if left_max >= right_max else right_max
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod(verbose=True)
diff --git a/maths/find_max_recursion.py b/maths/find_max_recursion.py
deleted file mode 100644
index 03fb81950dcb..000000000000
--- a/maths/find_max_recursion.py
+++ /dev/null
@@ -1,25 +0,0 @@
-# Divide and Conquer algorithm
-def find_max(nums, left, right):
- """
- find max value in list
- :param nums: contains elements
- :param left: index of first element
- :param right: index of last element
- :return: max in nums
-
- >>> nums = [1, 3, 5, 7, 9, 2, 4, 6, 8, 10]
- >>> find_max(nums, 0, len(nums) - 1) == max(nums)
- True
- """
- if left == right:
- return nums[left]
- mid = (left + right) >> 1 # the middle
- left_max = find_max(nums, left, mid) # find max in range[left, mid]
- right_max = find_max(nums, mid + 1, right) # find max in range[mid + 1, right]
-
- return left_max if left_max >= right_max else right_max
-
-
-if __name__ == "__main__":
- nums = [1, 3, 5, 7, 9, 2, 4, 6, 8, 10]
- assert find_max(nums, 0, len(nums) - 1) == 10
diff --git a/maths/find_min.py b/maths/find_min.py
index 2af2e44ba353..762562e36ef9 100644
--- a/maths/find_min.py
+++ b/maths/find_min.py
@@ -1,26 +1,87 @@
-def find_min(nums):
+from __future__ import annotations
+
+
+def find_min_iterative(nums: list[int | float]) -> int | float:
"""
Find Minimum Number in a List
:param nums: contains elements
:return: min number in list
>>> for nums in ([3, 2, 1], [-3, -2, -1], [3, -3, 0], [3.0, 3.1, 2.9]):
- ... find_min(nums) == min(nums)
+ ... find_min_iterative(nums) == min(nums)
True
True
True
True
+ >>> find_min_iterative([0, 1, 2, 3, 4, 5, -3, 24, -56])
+ -56
+ >>> find_min_iterative([])
+ Traceback (most recent call last):
+ ...
+ ValueError: find_min_iterative() arg is an empty sequence
"""
+ if len(nums) == 0:
+ raise ValueError("find_min_iterative() arg is an empty sequence")
min_num = nums[0]
for num in nums:
- if min_num > num:
- min_num = num
+ min_num = min(min_num, num)
return min_num
-def main():
- assert find_min([0, 1, 2, 3, 4, 5, -3, 24, -56]) == -56
+# Divide and Conquer algorithm
+def find_min_recursive(nums: list[int | float], left: int, right: int) -> int | float:
+ """
+ find min value in list
+ :param nums: contains elements
+ :param left: index of first element
+ :param right: index of last element
+ :return: min in nums
+
+ >>> for nums in ([3, 2, 1], [-3, -2, -1], [3, -3, 0], [3.0, 3.1, 2.9]):
+ ... find_min_recursive(nums, 0, len(nums) - 1) == min(nums)
+ True
+ True
+ True
+ True
+ >>> nums = [1, 3, 5, 7, 9, 2, 4, 6, 8, 10]
+ >>> find_min_recursive(nums, 0, len(nums) - 1) == min(nums)
+ True
+ >>> find_min_recursive([], 0, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: find_min_recursive() arg is an empty sequence
+ >>> find_min_recursive(nums, 0, len(nums)) == min(nums)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ >>> find_min_recursive(nums, -len(nums), -1) == min(nums)
+ True
+ >>> find_min_recursive(nums, -len(nums) - 1, -1) == min(nums)
+ Traceback (most recent call last):
+ ...
+ IndexError: list index out of range
+ """
+ if len(nums) == 0:
+ raise ValueError("find_min_recursive() arg is an empty sequence")
+ if (
+ left >= len(nums)
+ or left < -len(nums)
+ or right >= len(nums)
+ or right < -len(nums)
+ ):
+ raise IndexError("list index out of range")
+ if left == right:
+ return nums[left]
+ mid = (left + right) >> 1 # the middle
+ left_min = find_min_recursive(nums, left, mid) # find min in range[left, mid]
+ right_min = find_min_recursive(
+ nums, mid + 1, right
+ ) # find min in range[mid + 1, right]
+
+ return left_min if left_min <= right_min else right_min
if __name__ == "__main__":
- main()
+ import doctest
+
+ doctest.testmod(verbose=True)
diff --git a/maths/find_min_recursion.py b/maths/find_min_recursion.py
deleted file mode 100644
index 4488967cc57a..000000000000
--- a/maths/find_min_recursion.py
+++ /dev/null
@@ -1,25 +0,0 @@
-# Divide and Conquer algorithm
-def find_min(nums, left, right):
- """
- find min value in list
- :param nums: contains elements
- :param left: index of first element
- :param right: index of last element
- :return: min in nums
-
- >>> nums = [1, 3, 5, 7, 9, 2, 4, 6, 8, 10]
- >>> find_min(nums, 0, len(nums) - 1) == min(nums)
- True
- """
- if left == right:
- return nums[left]
- mid = (left + right) >> 1 # the middle
- left_min = find_min(nums, left, mid) # find min in range[left, mid]
- right_min = find_min(nums, mid + 1, right) # find min in range[mid + 1, right]
-
- return left_min if left_min <= right_min else right_min
-
-
-if __name__ == "__main__":
- nums = [1, 3, 5, 7, 9, 2, 4, 6, 8, 10]
- assert find_min(nums, 0, len(nums) - 1) == 1
diff --git a/maths/floor.py b/maths/floor.py
index 482250f5e59e..8bbcb21aa6e4 100644
--- a/maths/floor.py
+++ b/maths/floor.py
@@ -3,7 +3,7 @@
"""
-def floor(x) -> int:
+def floor(x: float) -> int:
"""
Return the floor of x as an Integral.
:param x: the number
diff --git a/maths/gamma.py b/maths/gamma.py
index 69cd819ef186..e328cd8b22b7 100644
--- a/maths/gamma.py
+++ b/maths/gamma.py
@@ -1,52 +1,46 @@
+"""
+Gamma function is a very useful tool in math and physics.
+It helps calculating complex integral in a convenient way.
+for more info: https://en.wikipedia.org/wiki/Gamma_function
+In mathematics, the gamma function is one commonly
+used extension of the factorial function to complex numbers.
+The gamma function is defined for all complex numbers except
+the non-positive integers
+Python's Standard Library math.gamma() function overflows around gamma(171.624).
+"""
+
import math
from numpy import inf
from scipy.integrate import quad
-def gamma(num: float) -> float:
+def gamma_iterative(num: float) -> float:
"""
- https://en.wikipedia.org/wiki/Gamma_function
- In mathematics, the gamma function is one commonly
- used extension of the factorial function to complex numbers.
- The gamma function is defined for all complex numbers except the non-positive
- integers
-
+ Calculates the value of Gamma function of num
+ where num is either an integer (1, 2, 3..) or a half-integer (0.5, 1.5, 2.5 ...).
- >>> gamma(-1)
+ >>> gamma_iterative(-1)
Traceback (most recent call last):
...
ValueError: math domain error
-
-
-
- >>> gamma(0)
+ >>> gamma_iterative(0)
Traceback (most recent call last):
...
ValueError: math domain error
-
-
- >>> gamma(9)
+ >>> gamma_iterative(9)
40320.0
-
>>> from math import gamma as math_gamma
- >>> all(.99999999 < gamma(i) / math_gamma(i) <= 1.000000001
+ >>> all(.99999999 < gamma_iterative(i) / math_gamma(i) <= 1.000000001
... for i in range(1, 50))
True
-
-
- >>> from math import gamma as math_gamma
- >>> gamma(-1)/math_gamma(-1) <= 1.000000001
+ >>> gamma_iterative(-1)/math_gamma(-1) <= 1.000000001
Traceback (most recent call last):
...
ValueError: math domain error
-
-
- >>> from math import gamma as math_gamma
- >>> gamma(3.3) - math_gamma(3.3) <= 0.00000001
+ >>> gamma_iterative(3.3) - math_gamma(3.3) <= 0.00000001
True
"""
-
if num <= 0:
raise ValueError("math domain error")
@@ -57,7 +51,66 @@ def integrand(x: float, z: float) -> float:
return math.pow(x, z - 1) * math.exp(-x)
+def gamma_recursive(num: float) -> float:
+ """
+ Calculates the value of Gamma function of num
+ where num is either an integer (1, 2, 3..) or a half-integer (0.5, 1.5, 2.5 ...).
+ Implemented using recursion
+ Examples:
+ >>> from math import isclose, gamma as math_gamma
+ >>> gamma_recursive(0.5)
+ 1.7724538509055159
+ >>> gamma_recursive(1)
+ 1.0
+ >>> gamma_recursive(2)
+ 1.0
+ >>> gamma_recursive(3.5)
+ 3.3233509704478426
+ >>> gamma_recursive(171.5)
+ 9.483367566824795e+307
+ >>> all(isclose(gamma_recursive(num), math_gamma(num))
+ ... for num in (0.5, 2, 3.5, 171.5))
+ True
+ >>> gamma_recursive(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: math domain error
+ >>> gamma_recursive(-1.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: math domain error
+ >>> gamma_recursive(-4)
+ Traceback (most recent call last):
+ ...
+ ValueError: math domain error
+ >>> gamma_recursive(172)
+ Traceback (most recent call last):
+ ...
+ OverflowError: math range error
+ >>> gamma_recursive(1.1)
+ Traceback (most recent call last):
+ ...
+ NotImplementedError: num must be an integer or a half-integer
+ """
+ if num <= 0:
+ raise ValueError("math domain error")
+ if num > 171.5:
+ raise OverflowError("math range error")
+ elif num - int(num) not in (0, 0.5):
+ raise NotImplementedError("num must be an integer or a half-integer")
+ elif num == 0.5:
+ return math.sqrt(math.pi)
+ else:
+ return 1.0 if num == 1 else (num - 1) * gamma_recursive(num - 1)
+
+
if __name__ == "__main__":
from doctest import testmod
testmod()
+ num = 1.0
+ while num:
+ num = float(input("Gamma of: "))
+ print(f"gamma_iterative({num}) = {gamma_iterative(num)}")
+ print(f"gamma_recursive({num}) = {gamma_recursive(num)}")
+ print("\nEnter 0 to exit...")
diff --git a/maths/gaussian.py b/maths/gaussian.py
index a5dba50a927d..b1e62ea77fe2 100644
--- a/maths/gaussian.py
+++ b/maths/gaussian.py
@@ -1,21 +1,22 @@
"""
Reference: https://en.wikipedia.org/wiki/Gaussian_function
"""
+
from numpy import exp, pi, sqrt
-def gaussian(x, mu: float = 0.0, sigma: float = 1.0) -> int:
+def gaussian(x, mu: float = 0.0, sigma: float = 1.0) -> float:
"""
- >>> gaussian(1)
+ >>> float(gaussian(1))
0.24197072451914337
- >>> gaussian(24)
+ >>> float(gaussian(24))
3.342714441794458e-126
- >>> gaussian(1, 4, 2)
+ >>> float(gaussian(1, 4, 2))
0.06475879783294587
- >>> gaussian(1, 5, 3)
+ >>> float(gaussian(1, 5, 3))
0.05467002489199788
Supports NumPy Arrays
@@ -28,7 +29,7 @@ def gaussian(x, mu: float = 0.0, sigma: float = 1.0) -> int:
5.05227108e-15, 1.02797736e-18, 7.69459863e-23, 2.11881925e-27,
2.14638374e-32, 7.99882776e-38, 1.09660656e-43])
- >>> gaussian(15)
+ >>> float(gaussian(15))
5.530709549844416e-50
>>> gaussian([1,2, 'string'])
@@ -46,13 +47,13 @@ def gaussian(x, mu: float = 0.0, sigma: float = 1.0) -> int:
...
OverflowError: (34, 'Result too large')
- >>> gaussian(10**-326)
+ >>> float(gaussian(10**-326))
0.3989422804014327
- >>> gaussian(2523, mu=234234, sigma=3425)
+ >>> float(gaussian(2523, mu=234234, sigma=3425))
0.0
"""
- return 1 / sqrt(2 * pi * sigma ** 2) * exp(-((x - mu) ** 2) / (2 * sigma ** 2))
+ return 1 / sqrt(2 * pi * sigma**2) * exp(-((x - mu) ** 2) / (2 * sigma**2))
if __name__ == "__main__":
diff --git a/maths/gcd_of_n_numbers.py b/maths/gcd_of_n_numbers.py
new file mode 100644
index 000000000000..63236c236ada
--- /dev/null
+++ b/maths/gcd_of_n_numbers.py
@@ -0,0 +1,109 @@
+"""
+Gcd of N Numbers
+Reference: https://en.wikipedia.org/wiki/Greatest_common_divisor
+"""
+
+from collections import Counter
+
+
+def get_factors(
+ number: int, factors: Counter | None = None, factor: int = 2
+) -> Counter:
+ """
+ this is a recursive function for get all factors of number
+ >>> get_factors(45)
+ Counter({3: 2, 5: 1})
+ >>> get_factors(2520)
+ Counter({2: 3, 3: 2, 5: 1, 7: 1})
+ >>> get_factors(23)
+ Counter({23: 1})
+ >>> get_factors(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: number must be integer and greater than zero
+ >>> get_factors(-1)
+ Traceback (most recent call last):
+ ...
+ TypeError: number must be integer and greater than zero
+ >>> get_factors(1.5)
+ Traceback (most recent call last):
+ ...
+ TypeError: number must be integer and greater than zero
+
+ factor can be all numbers from 2 to number that we check if number % factor == 0
+ if it is equal to zero, we check again with number // factor
+ else we increase factor by one
+ """
+
+ match number:
+ case int(number) if number == 1:
+ return Counter({1: 1})
+ case int(num) if number > 0:
+ number = num
+ case _:
+ raise TypeError("number must be integer and greater than zero")
+
+ factors = factors or Counter()
+
+ if number == factor: # break condition
+ # all numbers are factors of itself
+ factors[factor] += 1
+ return factors
+
+ if number % factor > 0:
+ # if it is greater than zero
+ # so it is not a factor of number and we check next number
+ return get_factors(number, factors, factor + 1)
+
+ factors[factor] += 1
+ # else we update factors (that is Counter(dict-like) type) and check again
+ return get_factors(number // factor, factors, factor)
+
+
+def get_greatest_common_divisor(*numbers: int) -> int:
+ """
+ get gcd of n numbers:
+ >>> get_greatest_common_divisor(18, 45)
+ 9
+ >>> get_greatest_common_divisor(23, 37)
+ 1
+ >>> get_greatest_common_divisor(2520, 8350)
+ 10
+ >>> get_greatest_common_divisor(-10, 20)
+ Traceback (most recent call last):
+ ...
+ Exception: numbers must be integer and greater than zero
+ >>> get_greatest_common_divisor(1.5, 2)
+ Traceback (most recent call last):
+ ...
+ Exception: numbers must be integer and greater than zero
+ >>> get_greatest_common_divisor(1, 2, 3, 4, 5, 6, 7, 8, 9, 10)
+ 1
+ >>> get_greatest_common_divisor("1", 2, 3, 4, 5, 6, 7, 8, 9, 10)
+ Traceback (most recent call last):
+ ...
+ Exception: numbers must be integer and greater than zero
+ """
+
+ # we just need factors, not numbers itself
+ try:
+ same_factors, *factors = map(get_factors, numbers)
+ except TypeError as e:
+ raise Exception("numbers must be integer and greater than zero") from e
+
+ for factor in factors:
+ same_factors &= factor
+ # get common factor between all
+ # `&` return common elements with smaller value (for Counter type)
+
+ # now, same_factors is something like {2: 2, 3: 4} that means 2 * 2 * 3 * 3 * 3 * 3
+ mult = 1
+ # power each factor and multiply
+ # for {2: 2, 3: 4}, it is [4, 81] and then 324
+ for m in [factor**power for factor, power in same_factors.items()]:
+ mult *= m
+ return mult
+
+
+if __name__ == "__main__":
+ print(get_greatest_common_divisor(18, 45)) # 9
diff --git a/maths/geometric_mean.py b/maths/geometric_mean.py
new file mode 100644
index 000000000000..240d519ad398
--- /dev/null
+++ b/maths/geometric_mean.py
@@ -0,0 +1,55 @@
+"""
+The Geometric Mean of n numbers is defined as the n-th root of the product
+of those numbers. It is used to measure the central tendency of the numbers.
+https://en.wikipedia.org/wiki/Geometric_mean
+"""
+
+
+def compute_geometric_mean(*args: int) -> float:
+ """
+ Return the geometric mean of the argument numbers.
+ >>> compute_geometric_mean(2,8)
+ 4.0
+ >>> compute_geometric_mean('a', 4)
+ Traceback (most recent call last):
+ ...
+ TypeError: Not a Number
+ >>> compute_geometric_mean(5, 125)
+ 25.0
+ >>> compute_geometric_mean(1, 0)
+ 0.0
+ >>> compute_geometric_mean(1, 5, 25, 5)
+ 5.0
+ >>> compute_geometric_mean(2, -2)
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: Cannot Compute Geometric Mean for these numbers.
+ >>> compute_geometric_mean(-5, 25, 1)
+ -5.0
+ """
+ product = 1
+ for number in args:
+ if not isinstance(number, int) and not isinstance(number, float):
+ raise TypeError("Not a Number")
+ product *= number
+ # Cannot calculate the even root for negative product.
+ # Frequently they are restricted to being positive.
+ if product < 0 and len(args) % 2 == 0:
+ raise ArithmeticError("Cannot Compute Geometric Mean for these numbers.")
+ mean = abs(product) ** (1 / len(args))
+ # Since python calculates complex roots for negative products with odd roots.
+ if product < 0:
+ mean = -mean
+ # Since it does floating point arithmetic, it gives 64**(1/3) as 3.99999996
+ possible_mean = float(round(mean))
+ # To check if the rounded number is actually the mean.
+ if possible_mean ** len(args) == product:
+ mean = possible_mean
+ return mean
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod(name="compute_geometric_mean")
+ print(compute_geometric_mean(-3, -27))
diff --git a/maths/germain_primes.py b/maths/germain_primes.py
new file mode 100644
index 000000000000..078d1967f41a
--- /dev/null
+++ b/maths/germain_primes.py
@@ -0,0 +1,72 @@
+"""
+A Sophie Germain prime is any prime p, where 2p + 1 is also prime.
+The second number, 2p + 1 is called a safe prime.
+
+Examples of Germain primes include: 2, 3, 5, 11, 23
+
+Their corresponding safe primes: 5, 7, 11, 23, 47
+https://en.wikipedia.org/wiki/Safe_and_Sophie_Germain_primes
+"""
+
+from maths.prime_check import is_prime
+
+
+def is_germain_prime(number: int) -> bool:
+ """Checks if input number and 2*number + 1 are prime.
+
+ >>> is_germain_prime(3)
+ True
+ >>> is_germain_prime(11)
+ True
+ >>> is_germain_prime(4)
+ False
+ >>> is_germain_prime(23)
+ True
+ >>> is_germain_prime(13)
+ False
+ >>> is_germain_prime(20)
+ False
+ >>> is_germain_prime('abc')
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be a positive integer. Input value: abc
+ """
+ if not isinstance(number, int) or number < 1:
+ msg = f"Input value must be a positive integer. Input value: {number}"
+ raise TypeError(msg)
+
+ return is_prime(number) and is_prime(2 * number + 1)
+
+
+def is_safe_prime(number: int) -> bool:
+ """Checks if input number and (number - 1)/2 are prime.
+ The smallest safe prime is 5, with the Germain prime is 2.
+
+ >>> is_safe_prime(5)
+ True
+ >>> is_safe_prime(11)
+ True
+ >>> is_safe_prime(1)
+ False
+ >>> is_safe_prime(2)
+ False
+ >>> is_safe_prime(3)
+ False
+ >>> is_safe_prime(47)
+ True
+ >>> is_safe_prime('abc')
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value must be a positive integer. Input value: abc
+ """
+ if not isinstance(number, int) or number < 1:
+ msg = f"Input value must be a positive integer. Input value: {number}"
+ raise TypeError(msg)
+
+ return (number - 1) % 2 == 0 and is_prime(number) and is_prime((number - 1) // 2)
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/maths/hardy_ramanujanalgo.py b/maths/hardy_ramanujanalgo.py
index 90e4913c70a7..31ec76fbe10b 100644
--- a/maths/hardy_ramanujanalgo.py
+++ b/maths/hardy_ramanujanalgo.py
@@ -4,9 +4,9 @@
import math
-def exactPrimeFactorCount(n):
+def exact_prime_factor_count(n: int) -> int:
"""
- >>> exactPrimeFactorCount(51242183)
+ >>> exact_prime_factor_count(51242183)
3
"""
count = 0
@@ -36,8 +36,8 @@ def exactPrimeFactorCount(n):
if __name__ == "__main__":
n = 51242183
- print(f"The number of distinct prime factors is/are {exactPrimeFactorCount(n)}")
- print("The value of log(log(n)) is {:.4f}".format(math.log(math.log(n))))
+ print(f"The number of distinct prime factors is/are {exact_prime_factor_count(n)}")
+ print(f"The value of log(log(n)) is {math.log(math.log(n)):.4f}")
"""
The number of distinct prime factors is/are 3
diff --git a/maths/integer_square_root.py b/maths/integer_square_root.py
new file mode 100644
index 000000000000..27e874a43c79
--- /dev/null
+++ b/maths/integer_square_root.py
@@ -0,0 +1,73 @@
+"""
+Integer Square Root Algorithm -- An efficient method to calculate the square root of a
+non-negative integer 'num' rounded down to the nearest integer. It uses a binary search
+approach to find the integer square root without using any built-in exponent functions
+or operators.
+* https://en.wikipedia.org/wiki/Integer_square_root
+* https://docs.python.org/3/library/math.html#math.isqrt
+Note:
+ - This algorithm is designed for non-negative integers only.
+ - The result is rounded down to the nearest integer.
+ - The algorithm has a time complexity of O(log(x)).
+ - Original algorithm idea based on binary search.
+"""
+
+
+def integer_square_root(num: int) -> int:
+ """
+ Returns the integer square root of a non-negative integer num.
+ Args:
+ num: A non-negative integer.
+ Returns:
+ The integer square root of num.
+ Raises:
+ ValueError: If num is not an integer or is negative.
+ >>> [integer_square_root(i) for i in range(18)]
+ [0, 1, 1, 1, 2, 2, 2, 2, 2, 3, 3, 3, 3, 3, 3, 3, 4, 4]
+ >>> integer_square_root(625)
+ 25
+ >>> integer_square_root(2_147_483_647)
+ 46340
+ >>> from math import isqrt
+ >>> all(integer_square_root(i) == isqrt(i) for i in range(20))
+ True
+ >>> integer_square_root(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: num must be non-negative integer
+ >>> integer_square_root(1.5)
+ Traceback (most recent call last):
+ ...
+ ValueError: num must be non-negative integer
+ >>> integer_square_root("0")
+ Traceback (most recent call last):
+ ...
+ ValueError: num must be non-negative integer
+ """
+ if not isinstance(num, int) or num < 0:
+ raise ValueError("num must be non-negative integer")
+
+ if num < 2:
+ return num
+
+ left_bound = 0
+ right_bound = num // 2
+
+ while left_bound <= right_bound:
+ mid = left_bound + (right_bound - left_bound) // 2
+ mid_squared = mid * mid
+ if mid_squared == num:
+ return mid
+
+ if mid_squared < num:
+ left_bound = mid + 1
+ else:
+ right_bound = mid - 1
+
+ return right_bound
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/interquartile_range.py b/maths/interquartile_range.py
new file mode 100644
index 000000000000..e91a651647d4
--- /dev/null
+++ b/maths/interquartile_range.py
@@ -0,0 +1,67 @@
+"""
+An implementation of interquartile range (IQR) which is a measure of statistical
+dispersion, which is the spread of the data.
+
+The function takes the list of numeric values as input and returns the IQR.
+
+Script inspired by this Wikipedia article:
+https://en.wikipedia.org/wiki/Interquartile_range
+"""
+
+from __future__ import annotations
+
+
+def find_median(nums: list[int | float]) -> float:
+ """
+ This is the implementation of the median.
+ :param nums: The list of numeric nums
+ :return: Median of the list
+ >>> find_median(nums=([1, 2, 2, 3, 4]))
+ 2
+ >>> find_median(nums=([1, 2, 2, 3, 4, 4]))
+ 2.5
+ >>> find_median(nums=([-1, 2, 0, 3, 4, -4]))
+ 1.5
+ >>> find_median(nums=([1.1, 2.2, 2, 3.3, 4.4, 4]))
+ 2.65
+ """
+ div, mod = divmod(len(nums), 2)
+ if mod:
+ return nums[div]
+ return (nums[div] + nums[(div) - 1]) / 2
+
+
+def interquartile_range(nums: list[int | float]) -> float:
+ """
+ Return the interquartile range for a list of numeric values.
+ :param nums: The list of numeric values.
+ :return: interquartile range
+
+ >>> interquartile_range(nums=[4, 1, 2, 3, 2])
+ 2.0
+ >>> interquartile_range(nums = [-2, -7, -10, 9, 8, 4, -67, 45])
+ 17.0
+ >>> interquartile_range(nums = [-2.1, -7.1, -10.1, 9.1, 8.1, 4.1, -67.1, 45.1])
+ 17.2
+ >>> interquartile_range(nums = [0, 0, 0, 0, 0])
+ 0.0
+ >>> interquartile_range(nums=[])
+ Traceback (most recent call last):
+ ...
+ ValueError: The list is empty. Provide a non-empty list.
+ """
+ if not nums:
+ raise ValueError("The list is empty. Provide a non-empty list.")
+ nums.sort()
+ length = len(nums)
+ div, mod = divmod(length, 2)
+ q1 = find_median(nums[:div])
+ half_length = sum((div, mod))
+ q3 = find_median(nums[half_length:length])
+ return q3 - q1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/is_int_palindrome.py b/maths/is_int_palindrome.py
new file mode 100644
index 000000000000..63dc9e2138e8
--- /dev/null
+++ b/maths/is_int_palindrome.py
@@ -0,0 +1,34 @@
+def is_int_palindrome(num: int) -> bool:
+ """
+ Returns whether `num` is a palindrome or not
+ (see for reference https://en.wikipedia.org/wiki/Palindromic_number).
+
+ >>> is_int_palindrome(-121)
+ False
+ >>> is_int_palindrome(0)
+ True
+ >>> is_int_palindrome(10)
+ False
+ >>> is_int_palindrome(11)
+ True
+ >>> is_int_palindrome(101)
+ True
+ >>> is_int_palindrome(120)
+ False
+ """
+ if num < 0:
+ return False
+
+ num_copy: int = num
+ rev_num: int = 0
+ while num > 0:
+ rev_num = rev_num * 10 + (num % 10)
+ num //= 10
+
+ return num_copy == rev_num
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/is_ip_v4_address_valid.py b/maths/is_ip_v4_address_valid.py
new file mode 100644
index 000000000000..305afabffed3
--- /dev/null
+++ b/maths/is_ip_v4_address_valid.py
@@ -0,0 +1,75 @@
+"""
+wiki: https://en.wikipedia.org/wiki/IPv4
+
+Is IP v4 address valid?
+A valid IP address must be four octets in the form of A.B.C.D,
+where A, B, C and D are numbers from 0-255
+for example: 192.168.23.1, 172.255.255.255 are valid IP address
+ 192.168.256.0, 256.192.3.121 are invalid IP address
+"""
+
+
+def is_ip_v4_address_valid(ip: str) -> bool:
+ """
+ print "Valid IP address" If IP is valid.
+ or
+ print "Invalid IP address" If IP is invalid.
+
+ >>> is_ip_v4_address_valid("192.168.0.23")
+ True
+
+ >>> is_ip_v4_address_valid("192.256.15.8")
+ False
+
+ >>> is_ip_v4_address_valid("172.100.0.8")
+ True
+
+ >>> is_ip_v4_address_valid("255.256.0.256")
+ False
+
+ >>> is_ip_v4_address_valid("1.2.33333333.4")
+ False
+
+ >>> is_ip_v4_address_valid("1.2.-3.4")
+ False
+
+ >>> is_ip_v4_address_valid("1.2.3")
+ False
+
+ >>> is_ip_v4_address_valid("1.2.3.4.5")
+ False
+
+ >>> is_ip_v4_address_valid("1.2.A.4")
+ False
+
+ >>> is_ip_v4_address_valid("0.0.0.0")
+ True
+
+ >>> is_ip_v4_address_valid("1.2.3.")
+ False
+
+ >>> is_ip_v4_address_valid("1.2.3.05")
+ False
+ """
+ octets = ip.split(".")
+ if len(octets) != 4:
+ return False
+
+ for octet in octets:
+ if not octet.isdigit():
+ return False
+
+ number = int(octet)
+ if len(str(number)) != len(octet):
+ return False
+
+ if not 0 <= number <= 255:
+ return False
+
+ return True
+
+
+if __name__ == "__main__":
+ ip = input().strip()
+ valid_or_invalid = "valid" if is_ip_v4_address_valid(ip) else "invalid"
+ print(f"{ip} is a {valid_or_invalid} IPv4 address.")
diff --git a/maths/is_square_free.py b/maths/is_square_free.py
index 8d83d95ffb67..a336c37e8dbc 100644
--- a/maths/is_square_free.py
+++ b/maths/is_square_free.py
@@ -1,8 +1,9 @@
"""
References: wikipedia:square free number
-python/black : True
-flake8 : True
+psf/black : True
+ruff : True
"""
+
from __future__ import annotations
@@ -15,7 +16,7 @@ def is_square_free(factors: list[int]) -> bool:
False
These are wrong but should return some value
- it simply checks for repition in the numbers.
+ it simply checks for repetition in the numbers.
>>> is_square_free([1, 3, 4, 'sd', 0.0])
True
diff --git a/maths/jaccard_similarity.py b/maths/jaccard_similarity.py
index 4f24d308f340..6b6243458fa8 100644
--- a/maths/jaccard_similarity.py
+++ b/maths/jaccard_similarity.py
@@ -14,7 +14,11 @@
"""
-def jaccard_similariy(setA, setB, alternativeUnion=False):
+def jaccard_similarity(
+ set_a: set[str] | list[str] | tuple[str],
+ set_b: set[str] | list[str] | tuple[str],
+ alternative_union=False,
+):
"""
Finds the jaccard similarity between two sets.
Essentially, its intersection over union.
@@ -24,8 +28,8 @@ def jaccard_similariy(setA, setB, alternativeUnion=False):
of a set with itself be 1/2 instead of 1. [MMDS 2nd Edition, Page 77]
Parameters:
- :setA (set,list,tuple): A non-empty set/list
- :setB (set,list,tuple): A non-empty set/list
+ :set_a (set,list,tuple): A non-empty set/list
+ :set_b (set,list,tuple): A non-empty set/list
:alternativeUnion (boolean): If True, use sum of number of
items as union
@@ -33,48 +37,59 @@ def jaccard_similariy(setA, setB, alternativeUnion=False):
(float) The jaccard similarity between the two sets.
Examples:
- >>> setA = {'a', 'b', 'c', 'd', 'e'}
- >>> setB = {'c', 'd', 'e', 'f', 'h', 'i'}
- >>> jaccard_similariy(setA,setB)
+ >>> set_a = {'a', 'b', 'c', 'd', 'e'}
+ >>> set_b = {'c', 'd', 'e', 'f', 'h', 'i'}
+ >>> jaccard_similarity(set_a, set_b)
0.375
-
- >>> jaccard_similariy(setA,setA)
+ >>> jaccard_similarity(set_a, set_a)
1.0
-
- >>> jaccard_similariy(setA,setA,True)
+ >>> jaccard_similarity(set_a, set_a, True)
0.5
-
- >>> setA = ['a', 'b', 'c', 'd', 'e']
- >>> setB = ('c', 'd', 'e', 'f', 'h', 'i')
- >>> jaccard_similariy(setA,setB)
+ >>> set_a = ['a', 'b', 'c', 'd', 'e']
+ >>> set_b = ('c', 'd', 'e', 'f', 'h', 'i')
+ >>> jaccard_similarity(set_a, set_b)
+ 0.375
+ >>> set_a = ('c', 'd', 'e', 'f', 'h', 'i')
+ >>> set_b = ['a', 'b', 'c', 'd', 'e']
+ >>> jaccard_similarity(set_a, set_b)
0.375
+ >>> set_a = ('c', 'd', 'e', 'f', 'h', 'i')
+ >>> set_b = ['a', 'b', 'c', 'd']
+ >>> jaccard_similarity(set_a, set_b, True)
+ 0.2
+ >>> set_a = {'a', 'b'}
+ >>> set_b = ['c', 'd']
+ >>> jaccard_similarity(set_a, set_b)
+ Traceback (most recent call last):
+ ...
+ ValueError: Set a and b must either both be sets or be either a list or a tuple.
"""
- if isinstance(setA, set) and isinstance(setB, set):
+ if isinstance(set_a, set) and isinstance(set_b, set):
+ intersection_length = len(set_a.intersection(set_b))
- intersection = len(setA.intersection(setB))
-
- if alternativeUnion:
- union = len(setA) + len(setB)
+ if alternative_union:
+ union_length = len(set_a) + len(set_b)
else:
- union = len(setA.union(setB))
-
- return intersection / union
+ union_length = len(set_a.union(set_b))
- if isinstance(setA, (list, tuple)) and isinstance(setB, (list, tuple)):
+ return intersection_length / union_length
- intersection = [element for element in setA if element in setB]
+ elif isinstance(set_a, (list, tuple)) and isinstance(set_b, (list, tuple)):
+ intersection = [element for element in set_a if element in set_b]
- if alternativeUnion:
- union = len(setA) + len(setB)
+ if alternative_union:
+ return len(intersection) / (len(set_a) + len(set_b))
else:
- union = setA + [element for element in setB if element not in setA]
-
- return len(intersection) / len(union)
+ # Cast set_a to list because tuples cannot be mutated
+ union = list(set_a) + [element for element in set_b if element not in set_a]
+ return len(intersection) / len(union)
+ raise ValueError(
+ "Set a and b must either both be sets or be either a list or a tuple."
+ )
if __name__ == "__main__":
-
- setA = {"a", "b", "c", "d", "e"}
- setB = {"c", "d", "e", "f", "h", "i"}
- print(jaccard_similariy(setA, setB))
+ set_a = {"a", "b", "c", "d", "e"}
+ set_b = {"c", "d", "e", "f", "h", "i"}
+ print(jaccard_similarity(set_a, set_b))
diff --git a/maths/joint_probability_distribution.py b/maths/joint_probability_distribution.py
new file mode 100644
index 000000000000..6fbcea40c358
--- /dev/null
+++ b/maths/joint_probability_distribution.py
@@ -0,0 +1,124 @@
+"""
+Calculate joint probability distribution
+https://en.wikipedia.org/wiki/Joint_probability_distribution
+"""
+
+
+def joint_probability_distribution(
+ x_values: list[int],
+ y_values: list[int],
+ x_probabilities: list[float],
+ y_probabilities: list[float],
+) -> dict:
+ """
+ >>> joint_distribution = joint_probability_distribution(
+ ... [1, 2], [-2, 5, 8], [0.7, 0.3], [0.3, 0.5, 0.2]
+ ... )
+ >>> from math import isclose
+ >>> isclose(joint_distribution.pop((1, 8)), 0.14)
+ True
+ >>> joint_distribution
+ {(1, -2): 0.21, (1, 5): 0.35, (2, -2): 0.09, (2, 5): 0.15, (2, 8): 0.06}
+ """
+ return {
+ (x, y): x_prob * y_prob
+ for x, x_prob in zip(x_values, x_probabilities)
+ for y, y_prob in zip(y_values, y_probabilities)
+ }
+
+
+# Function to calculate the expectation (mean)
+def expectation(values: list, probabilities: list) -> float:
+ """
+ >>> from math import isclose
+ >>> isclose(expectation([1, 2], [0.7, 0.3]), 1.3)
+ True
+ """
+ return sum(x * p for x, p in zip(values, probabilities))
+
+
+# Function to calculate the variance
+def variance(values: list[int], probabilities: list[float]) -> float:
+ """
+ >>> from math import isclose
+ >>> isclose(variance([1,2],[0.7,0.3]), 0.21)
+ True
+ """
+ mean = expectation(values, probabilities)
+ return sum((x - mean) ** 2 * p for x, p in zip(values, probabilities))
+
+
+# Function to calculate the covariance
+def covariance(
+ x_values: list[int],
+ y_values: list[int],
+ x_probabilities: list[float],
+ y_probabilities: list[float],
+) -> float:
+ """
+ >>> covariance([1, 2], [-2, 5, 8], [0.7, 0.3], [0.3, 0.5, 0.2])
+ -2.7755575615628914e-17
+ """
+ mean_x = expectation(x_values, x_probabilities)
+ mean_y = expectation(y_values, y_probabilities)
+ return sum(
+ (x - mean_x) * (y - mean_y) * px * py
+ for x, px in zip(x_values, x_probabilities)
+ for y, py in zip(y_values, y_probabilities)
+ )
+
+
+# Function to calculate the standard deviation
+def standard_deviation(variance: float) -> float:
+ """
+ >>> standard_deviation(0.21)
+ 0.458257569495584
+ """
+ return variance**0.5
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ # Input values for X and Y
+ x_vals = input("Enter values of X separated by spaces: ").split()
+ y_vals = input("Enter values of Y separated by spaces: ").split()
+
+ # Convert input values to integers
+ x_values = [int(x) for x in x_vals]
+ y_values = [int(y) for y in y_vals]
+
+ # Input probabilities for X and Y
+ x_probs = input("Enter probabilities for X separated by spaces: ").split()
+ y_probs = input("Enter probabilities for Y separated by spaces: ").split()
+ assert len(x_values) == len(x_probs)
+ assert len(y_values) == len(y_probs)
+
+ # Convert input probabilities to floats
+ x_probabilities = [float(p) for p in x_probs]
+ y_probabilities = [float(p) for p in y_probs]
+
+ # Calculate the joint probability distribution
+ jpd = joint_probability_distribution(
+ x_values, y_values, x_probabilities, y_probabilities
+ )
+
+ # Print the joint probability distribution
+ print(
+ "\n".join(
+ f"P(X={x}, Y={y}) = {probability}" for (x, y), probability in jpd.items()
+ )
+ )
+ mean_xy = expectation(
+ [x * y for x in x_values for y in y_values],
+ [px * py for px in x_probabilities for py in y_probabilities],
+ )
+ print(f"x mean: {expectation(x_values, x_probabilities) = }")
+ print(f"y mean: {expectation(y_values, y_probabilities) = }")
+ print(f"xy mean: {mean_xy}")
+ print(f"x: {variance(x_values, x_probabilities) = }")
+ print(f"y: {variance(y_values, y_probabilities) = }")
+ print(f"{covariance(x_values, y_values, x_probabilities, y_probabilities) = }")
+ print(f"x: {standard_deviation(variance(x_values, x_probabilities)) = }")
+ print(f"y: {standard_deviation(variance(y_values, y_probabilities)) = }")
diff --git a/maths/josephus_problem.py b/maths/josephus_problem.py
new file mode 100644
index 000000000000..271292ba1d9f
--- /dev/null
+++ b/maths/josephus_problem.py
@@ -0,0 +1,130 @@
+"""
+The Josephus problem is a famous theoretical problem related to a certain
+counting-out game. This module provides functions to solve the Josephus problem
+for num_people and a step_size.
+
+The Josephus problem is defined as follows:
+- num_people are standing in a circle.
+- Starting with a specified person, you count around the circle,
+ skipping a fixed number of people (step_size).
+- The person at which you stop counting is eliminated from the circle.
+- The counting continues until only one person remains.
+
+For more information about the Josephus problem, refer to:
+https://en.wikipedia.org/wiki/Josephus_problem
+"""
+
+
+def josephus_recursive(num_people: int, step_size: int) -> int:
+ """
+ Solve the Josephus problem for num_people and a step_size recursively.
+
+ Args:
+ num_people: A positive integer representing the number of people.
+ step_size: A positive integer representing the step size for elimination.
+
+ Returns:
+ The position of the last person remaining.
+
+ Raises:
+ ValueError: If num_people or step_size is not a positive integer.
+
+ Examples:
+ >>> josephus_recursive(7, 3)
+ 3
+ >>> josephus_recursive(10, 2)
+ 4
+ >>> josephus_recursive(0, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ >>> josephus_recursive(1.9, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ >>> josephus_recursive(-2, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ >>> josephus_recursive(7, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ >>> josephus_recursive(7, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ >>> josephus_recursive(1_000, 0.01)
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ >>> josephus_recursive("cat", "dog")
+ Traceback (most recent call last):
+ ...
+ ValueError: num_people or step_size is not a positive integer.
+ """
+ if (
+ not isinstance(num_people, int)
+ or not isinstance(step_size, int)
+ or num_people <= 0
+ or step_size <= 0
+ ):
+ raise ValueError("num_people or step_size is not a positive integer.")
+
+ if num_people == 1:
+ return 0
+
+ return (josephus_recursive(num_people - 1, step_size) + step_size) % num_people
+
+
+def find_winner(num_people: int, step_size: int) -> int:
+ """
+ Find the winner of the Josephus problem for num_people and a step_size.
+
+ Args:
+ num_people (int): Number of people.
+ step_size (int): Step size for elimination.
+
+ Returns:
+ int: The position of the last person remaining (1-based index).
+
+ Examples:
+ >>> find_winner(7, 3)
+ 4
+ >>> find_winner(10, 2)
+ 5
+ """
+ return josephus_recursive(num_people, step_size) + 1
+
+
+def josephus_iterative(num_people: int, step_size: int) -> int:
+ """
+ Solve the Josephus problem for num_people and a step_size iteratively.
+
+ Args:
+ num_people (int): The number of people in the circle.
+ step_size (int): The number of steps to take before eliminating someone.
+
+ Returns:
+ int: The position of the last person standing.
+
+ Examples:
+ >>> josephus_iterative(5, 2)
+ 3
+ >>> josephus_iterative(7, 3)
+ 4
+ """
+ circle = list(range(1, num_people + 1))
+ current = 0
+
+ while len(circle) > 1:
+ current = (current + step_size - 1) % len(circle)
+ circle.pop(current)
+
+ return circle[0]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/juggler_sequence.py b/maths/juggler_sequence.py
new file mode 100644
index 000000000000..7f65d1dff925
--- /dev/null
+++ b/maths/juggler_sequence.py
@@ -0,0 +1,63 @@
+"""
+== Juggler Sequence ==
+Juggler sequence start with any positive integer n. The next term is
+obtained as follows:
+ If n term is even, the next term is floor value of square root of n .
+ If n is odd, the next term is floor value of 3 time the square root of n.
+
+https://en.wikipedia.org/wiki/Juggler_sequence
+"""
+
+# Author : Akshay Dubey (https://github.com/itsAkshayDubey)
+import math
+
+
+def juggler_sequence(number: int) -> list[int]:
+ """
+ >>> juggler_sequence(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value of [number=0] must be a positive integer
+ >>> juggler_sequence(1)
+ [1]
+ >>> juggler_sequence(2)
+ [2, 1]
+ >>> juggler_sequence(3)
+ [3, 5, 11, 36, 6, 2, 1]
+ >>> juggler_sequence(5)
+ [5, 11, 36, 6, 2, 1]
+ >>> juggler_sequence(10)
+ [10, 3, 5, 11, 36, 6, 2, 1]
+ >>> juggler_sequence(25)
+ [25, 125, 1397, 52214, 228, 15, 58, 7, 18, 4, 2, 1]
+ >>> juggler_sequence(6.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=6.0] must be an integer
+ >>> juggler_sequence(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value of [number=-1] must be a positive integer
+ """
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+ if number < 1:
+ msg = f"Input value of [number={number}] must be a positive integer"
+ raise ValueError(msg)
+ sequence = [number]
+ while number != 1:
+ if number % 2 == 0:
+ number = math.floor(math.sqrt(number))
+ else:
+ number = math.floor(
+ math.sqrt(number) * math.sqrt(number) * math.sqrt(number)
+ )
+ sequence.append(number)
+ return sequence
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/kadanes.py b/maths/kadanes.py
deleted file mode 100644
index d239d4a2589b..000000000000
--- a/maths/kadanes.py
+++ /dev/null
@@ -1,65 +0,0 @@
-"""
-Kadane's algorithm to get maximum subarray sum
-https://medium.com/@rsinghal757/kadanes-algorithm-dynamic-programming-how-and-why-does-it-work-3fd8849ed73d
-https://en.wikipedia.org/wiki/Maximum_subarray_problem
-"""
-test_data: tuple = ([-2, -8, -9], [2, 8, 9], [-1, 0, 1], [0, 0], [])
-
-
-def negative_exist(arr: list) -> int:
- """
- >>> negative_exist([-2,-8,-9])
- -2
- >>> [negative_exist(arr) for arr in test_data]
- [-2, 0, 0, 0, 0]
- """
- arr = arr or [0]
- max = arr[0]
- for i in arr:
- if i >= 0:
- return 0
- elif max <= i:
- max = i
- return max
-
-
-def kadanes(arr: list) -> int:
- """
- If negative_exist() returns 0 than this function will execute
- else it will return the value return by negative_exist function
-
- For example: arr = [2, 3, -9, 8, -2]
- Initially we set value of max_sum to 0 and max_till_element to 0 than when
- max_sum is less than max_till particular element it will assign that value to
- max_sum and when value of max_till_sum is less than 0 it will assign 0 to i
- and after that whole process, return the max_sum
- So the output for above arr is 8
-
- >>> kadanes([2, 3, -9, 8, -2])
- 8
- >>> [kadanes(arr) for arr in test_data]
- [-2, 19, 1, 0, 0]
- """
- max_sum = negative_exist(arr)
- if max_sum < 0:
- return max_sum
-
- max_sum = 0
- max_till_element = 0
-
- for i in arr:
- max_till_element += i
- if max_sum <= max_till_element:
- max_sum = max_till_element
- if max_till_element < 0:
- max_till_element = 0
- return max_sum
-
-
-if __name__ == "__main__":
- try:
- print("Enter integer values sepatated by spaces")
- arr = [int(x) for x in input().split()]
- print(f"Maximum subarray sum of {arr} is {kadanes(arr)}")
- except ValueError:
- print("Please enter integer values.")
diff --git a/maths/karatsuba.py b/maths/karatsuba.py
index df29c77a5cf2..0e063fb44b83 100644
--- a/maths/karatsuba.py
+++ b/maths/karatsuba.py
@@ -1,7 +1,7 @@
-""" Multiply two numbers using Karatsuba algorithm """
+"""Multiply two numbers using Karatsuba algorithm"""
-def karatsuba(a, b):
+def karatsuba(a: int, b: int) -> int:
"""
>>> karatsuba(15463, 23489) == 15463 * 23489
True
@@ -10,18 +10,18 @@ def karatsuba(a, b):
"""
if len(str(a)) == 1 or len(str(b)) == 1:
return a * b
- else:
- m1 = max(len(str(a)), len(str(b)))
- m2 = m1 // 2
- a1, a2 = divmod(a, 10 ** m2)
- b1, b2 = divmod(b, 10 ** m2)
+ m1 = max(len(str(a)), len(str(b)))
+ m2 = m1 // 2
- x = karatsuba(a2, b2)
- y = karatsuba((a1 + a2), (b1 + b2))
- z = karatsuba(a1, b1)
+ a1, a2 = divmod(a, 10**m2)
+ b1, b2 = divmod(b, 10**m2)
- return (z * 10 ** (2 * m2)) + ((y - z - x) * 10 ** (m2)) + (x)
+ x = karatsuba(a2, b2)
+ y = karatsuba((a1 + a2), (b1 + b2))
+ z = karatsuba(a1, b1)
+
+ return (z * 10 ** (2 * m2)) + ((y - z - x) * 10 ** (m2)) + (x)
def main():
diff --git a/maths/kth_lexicographic_permutation.py b/maths/kth_lexicographic_permutation.py
index 23eab626fbf8..b85558aca6d4 100644
--- a/maths/kth_lexicographic_permutation.py
+++ b/maths/kth_lexicographic_permutation.py
@@ -1,17 +1,17 @@
-def kthPermutation(k, n):
+def kth_permutation(k, n):
"""
Finds k'th lexicographic permutation (in increasing order) of
0,1,2,...n-1 in O(n^2) time.
Examples:
First permutation is always 0,1,2,...n
- >>> kthPermutation(0,5)
+ >>> kth_permutation(0,5)
[0, 1, 2, 3, 4]
The order of permutation of 0,1,2,3 is [0,1,2,3], [0,1,3,2], [0,2,1,3],
[0,2,3,1], [0,3,1,2], [0,3,2,1], [1,0,2,3], [1,0,3,2], [1,2,0,3],
[1,2,3,0], [1,3,0,2]
- >>> kthPermutation(10,4)
+ >>> kth_permutation(10,4)
[1, 3, 0, 2]
"""
# Factorails from 1! to (n-1)!
diff --git a/maths/largest_of_very_large_numbers.py b/maths/largest_of_very_large_numbers.py
index d2dc0af18126..edee50371e02 100644
--- a/maths/largest_of_very_large_numbers.py
+++ b/maths/largest_of_very_large_numbers.py
@@ -4,14 +4,27 @@
def res(x, y):
+ """
+ Reduces large number to a more manageable number
+ >>> res(5, 7)
+ 4.892790030352132
+ >>> res(0, 5)
+ 0
+ >>> res(3, 0)
+ 1
+ >>> res(-1, 5)
+ Traceback (most recent call last):
+ ...
+ ValueError: math domain error
+ """
if 0 not in (x, y):
# We use the relation x^y = y*log10(x), where 10 is the base.
return y * math.log10(x)
- else:
- if x == 0: # 0 raised to any number is 0
- return 0
- elif y == 0:
- return 1 # any number raised to 0 is 1
+ elif x == 0: # 0 raised to any number is 0
+ return 0
+ elif y == 0:
+ return 1 # any number raised to 0 is 1
+ raise AssertionError("This should never happen")
if __name__ == "__main__": # Main function
diff --git a/maths/least_common_multiple.py b/maths/least_common_multiple.py
index 0d087643e869..a5c4bf8e3625 100644
--- a/maths/least_common_multiple.py
+++ b/maths/least_common_multiple.py
@@ -1,6 +1,8 @@
import unittest
from timeit import timeit
+from maths.greatest_common_divisor import greatest_common_divisor
+
def least_common_multiple_slow(first_num: int, second_num: int) -> int:
"""
@@ -20,26 +22,6 @@ def least_common_multiple_slow(first_num: int, second_num: int) -> int:
return common_mult
-def greatest_common_divisor(a: int, b: int) -> int:
- """
- Calculate Greatest Common Divisor (GCD).
- see greatest_common_divisor.py
- >>> greatest_common_divisor(24, 40)
- 8
- >>> greatest_common_divisor(1, 1)
- 1
- >>> greatest_common_divisor(1, 800)
- 1
- >>> greatest_common_divisor(11, 37)
- 1
- >>> greatest_common_divisor(3, 5)
- 1
- >>> greatest_common_divisor(16, 4)
- 4
- """
- return b if a == 0 else greatest_common_divisor(b % a, a)
-
-
def least_common_multiple_fast(first_num: int, second_num: int) -> int:
"""
Find the least common multiple of two numbers.
@@ -67,8 +49,7 @@ def benchmark():
class TestLeastCommonMultiple(unittest.TestCase):
-
- test_inputs = [
+ test_inputs = (
(10, 20),
(13, 15),
(4, 31),
@@ -78,16 +59,16 @@ class TestLeastCommonMultiple(unittest.TestCase):
(12, 25),
(10, 25),
(6, 9),
- ]
- expected_results = [20, 195, 124, 210, 1462, 60, 300, 50, 18]
+ )
+ expected_results = (20, 195, 124, 210, 1462, 60, 300, 50, 18)
def test_lcm_function(self):
for i, (first_num, second_num) in enumerate(self.test_inputs):
slow_result = least_common_multiple_slow(first_num, second_num)
fast_result = least_common_multiple_fast(first_num, second_num)
with self.subTest(i=i):
- self.assertEqual(slow_result, self.expected_results[i])
- self.assertEqual(fast_result, self.expected_results[i])
+ assert slow_result == self.expected_results[i]
+ assert fast_result == self.expected_results[i]
if __name__ == "__main__":
diff --git a/maths/line_length.py b/maths/line_length.py
index 1d386b44b50d..ed2efc31e96e 100644
--- a/maths/line_length.py
+++ b/maths/line_length.py
@@ -1,14 +1,15 @@
+from __future__ import annotations
+
import math
-from typing import Callable, Union
+from collections.abc import Callable
def line_length(
- fnc: Callable[[Union[int, float]], Union[int, float]],
- x_start: Union[int, float],
- x_end: Union[int, float],
+ fnc: Callable[[float], float],
+ x_start: float,
+ x_end: float,
steps: int = 100,
) -> float:
-
"""
Approximates the arc length of a line segment by treating the curve as a
sequence of linear lines and summing their lengths
@@ -38,8 +39,7 @@ def line_length(
fx1 = fnc(x_start)
length = 0.0
- for i in range(steps):
-
+ for _ in range(steps):
# Approximates curve as a sequence of linear lines and sums their length
x2 = (x_end - x_start) / steps + x1
fx2 = fnc(x2)
diff --git a/maths/liouville_lambda.py b/maths/liouville_lambda.py
new file mode 100644
index 000000000000..1ed228dd5434
--- /dev/null
+++ b/maths/liouville_lambda.py
@@ -0,0 +1,46 @@
+"""
+== Liouville Lambda Function ==
+The Liouville Lambda function, denoted by λ(n)
+and λ(n) is 1 if n is the product of an even number of prime numbers,
+and -1 if it is the product of an odd number of primes.
+
+https://en.wikipedia.org/wiki/Liouville_function
+"""
+
+# Author : Akshay Dubey (https://github.com/itsAkshayDubey)
+from maths.prime_factors import prime_factors
+
+
+def liouville_lambda(number: int) -> int:
+ """
+ This functions takes an integer number as input.
+ returns 1 if n has even number of prime factors and -1 otherwise.
+ >>> liouville_lambda(10)
+ 1
+ >>> liouville_lambda(11)
+ -1
+ >>> liouville_lambda(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive integer
+ >>> liouville_lambda(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive integer
+ >>> liouville_lambda(11.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=11.0] must be an integer
+ """
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+ if number < 1:
+ raise ValueError("Input must be a positive integer")
+ return -1 if len(prime_factors(number)) % 2 else 1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/lucas_lehmer_primality_test.py b/maths/lucas_lehmer_primality_test.py
index 15e25cbfe996..af5c81133044 100644
--- a/maths/lucas_lehmer_primality_test.py
+++ b/maths/lucas_lehmer_primality_test.py
@@ -1,13 +1,13 @@
"""
- In mathematics, the Lucas–Lehmer test (LLT) is a primality test for Mersenne
- numbers. https://en.wikipedia.org/wiki/Lucas%E2%80%93Lehmer_primality_test
+In mathematics, the Lucas-Lehmer test (LLT) is a primality test for Mersenne
+numbers. https://en.wikipedia.org/wiki/Lucas%E2%80%93Lehmer_primality_test
- A Mersenne number is a number that is one less than a power of two.
- That is M_p = 2^p - 1
- https://en.wikipedia.org/wiki/Mersenne_prime
+A Mersenne number is a number that is one less than a power of two.
+That is M_p = 2^p - 1
+https://en.wikipedia.org/wiki/Mersenne_prime
- The Lucas–Lehmer test is the primality test used by the
- Great Internet Mersenne Prime Search (GIMPS) to locate large primes.
+The Lucas-Lehmer test is the primality test used by the
+Great Internet Mersenne Prime Search (GIMPS) to locate large primes.
"""
@@ -30,9 +30,9 @@ def lucas_lehmer_test(p: int) -> bool:
return True
s = 4
- M = (1 << p) - 1
- for i in range(p - 2):
- s = ((s * s) - 2) % M
+ m = (1 << p) - 1
+ for _ in range(p - 2):
+ s = ((s * s) - 2) % m
return s == 0
diff --git a/maths/lucas_series.py b/maths/lucas_series.py
index 02eae8d8c658..cae6c2815aec 100644
--- a/maths/lucas_series.py
+++ b/maths/lucas_series.py
@@ -3,7 +3,7 @@
"""
-def recursive_lucas_number(n):
+def recursive_lucas_number(n_th_number: int) -> int:
"""
Returns the nth lucas number
>>> recursive_lucas_number(1)
@@ -19,17 +19,19 @@ def recursive_lucas_number(n):
...
TypeError: recursive_lucas_number accepts only integer arguments.
"""
- if n == 1:
- return n
- if n == 0:
- return 2
- if not isinstance(n, int):
+ if not isinstance(n_th_number, int):
raise TypeError("recursive_lucas_number accepts only integer arguments.")
+ if n_th_number == 0:
+ return 2
+ if n_th_number == 1:
+ return 1
- return recursive_lucas_number(n - 1) + recursive_lucas_number(n - 2)
+ return recursive_lucas_number(n_th_number - 1) + recursive_lucas_number(
+ n_th_number - 2
+ )
-def dynamic_lucas_number(n: int) -> int:
+def dynamic_lucas_number(n_th_number: int) -> int:
"""
Returns the nth lucas number
>>> dynamic_lucas_number(1)
@@ -45,14 +47,10 @@ def dynamic_lucas_number(n: int) -> int:
...
TypeError: dynamic_lucas_number accepts only integer arguments.
"""
- if not isinstance(n, int):
+ if not isinstance(n_th_number, int):
raise TypeError("dynamic_lucas_number accepts only integer arguments.")
- if n == 0:
- return 2
- if n == 1:
- return 1
a, b = 2, 1
- for i in range(n):
+ for _ in range(n_th_number):
a, b = b, a + b
return a
@@ -62,7 +60,6 @@ def dynamic_lucas_number(n: int) -> int:
testmod()
n = int(input("Enter the number of terms in lucas series:\n").strip())
- n = int(input("Enter the number of terms in lucas series:\n").strip())
print("Using recursive function to calculate lucas series:")
print(" ".join(str(recursive_lucas_number(i)) for i in range(n)))
print("\nUsing dynamic function to calculate lucas series:")
diff --git a/maths/maclaurin_series.py b/maths/maclaurin_series.py
new file mode 100644
index 000000000000..6ec5551a5e6e
--- /dev/null
+++ b/maths/maclaurin_series.py
@@ -0,0 +1,123 @@
+"""
+https://en.wikipedia.org/wiki/Taylor_series#Trigonometric_functions
+"""
+
+from math import factorial, pi
+
+
+def maclaurin_sin(theta: float, accuracy: int = 30) -> float:
+ """
+ Finds the maclaurin approximation of sin
+
+ :param theta: the angle to which sin is found
+ :param accuracy: the degree of accuracy wanted minimum
+ :return: the value of sine in radians
+
+
+ >>> from math import isclose, sin
+ >>> all(isclose(maclaurin_sin(x, 50), sin(x)) for x in range(-25, 25))
+ True
+ >>> maclaurin_sin(10)
+ -0.5440211108893691
+ >>> maclaurin_sin(-10)
+ 0.5440211108893704
+ >>> maclaurin_sin(10, 15)
+ -0.544021110889369
+ >>> maclaurin_sin(-10, 15)
+ 0.5440211108893704
+ >>> maclaurin_sin("10")
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_sin() requires either an int or float for theta
+ >>> maclaurin_sin(10, -30)
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_sin() requires a positive int for accuracy
+ >>> maclaurin_sin(10, 30.5)
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_sin() requires a positive int for accuracy
+ >>> maclaurin_sin(10, "30")
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_sin() requires a positive int for accuracy
+ """
+
+ if not isinstance(theta, (int, float)):
+ raise ValueError("maclaurin_sin() requires either an int or float for theta")
+
+ if not isinstance(accuracy, int) or accuracy <= 0:
+ raise ValueError("maclaurin_sin() requires a positive int for accuracy")
+
+ theta = float(theta)
+ div = theta // (2 * pi)
+ theta -= 2 * div * pi
+ return sum(
+ (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1) for r in range(accuracy)
+ )
+
+
+def maclaurin_cos(theta: float, accuracy: int = 30) -> float:
+ """
+ Finds the maclaurin approximation of cos
+
+ :param theta: the angle to which cos is found
+ :param accuracy: the degree of accuracy wanted
+ :return: the value of cosine in radians
+
+
+ >>> from math import isclose, cos
+ >>> all(isclose(maclaurin_cos(x, 50), cos(x)) for x in range(-25, 25))
+ True
+ >>> maclaurin_cos(5)
+ 0.2836621854632268
+ >>> maclaurin_cos(-5)
+ 0.2836621854632265
+ >>> maclaurin_cos(10, 15)
+ -0.8390715290764524
+ >>> maclaurin_cos(-10, 15)
+ -0.8390715290764521
+ >>> maclaurin_cos("10")
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_cos() requires either an int or float for theta
+ >>> maclaurin_cos(10, -30)
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_cos() requires a positive int for accuracy
+ >>> maclaurin_cos(10, 30.5)
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_cos() requires a positive int for accuracy
+ >>> maclaurin_cos(10, "30")
+ Traceback (most recent call last):
+ ...
+ ValueError: maclaurin_cos() requires a positive int for accuracy
+ """
+
+ if not isinstance(theta, (int, float)):
+ raise ValueError("maclaurin_cos() requires either an int or float for theta")
+
+ if not isinstance(accuracy, int) or accuracy <= 0:
+ raise ValueError("maclaurin_cos() requires a positive int for accuracy")
+
+ theta = float(theta)
+ div = theta // (2 * pi)
+ theta -= 2 * div * pi
+ return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r) for r in range(accuracy))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ print(maclaurin_sin(10))
+ print(maclaurin_sin(-10))
+ print(maclaurin_sin(10, 15))
+ print(maclaurin_sin(-10, 15))
+
+ print(maclaurin_cos(5))
+ print(maclaurin_cos(-5))
+ print(maclaurin_cos(10, 15))
+ print(maclaurin_cos(-10, 15))
diff --git a/maths/manhattan_distance.py b/maths/manhattan_distance.py
new file mode 100644
index 000000000000..413991468a49
--- /dev/null
+++ b/maths/manhattan_distance.py
@@ -0,0 +1,126 @@
+def manhattan_distance(point_a: list, point_b: list) -> float:
+ """
+ Expectts two list of numbers representing two points in the same
+ n-dimensional space
+
+ https://en.wikipedia.org/wiki/Taxicab_geometry
+
+ >>> manhattan_distance([1,1], [2,2])
+ 2.0
+ >>> manhattan_distance([1.5,1.5], [2,2])
+ 1.0
+ >>> manhattan_distance([1.5,1.5], [2.5,2])
+ 1.5
+ >>> manhattan_distance([-3, -3, -3], [0, 0, 0])
+ 9.0
+ >>> manhattan_distance([1,1], None)
+ Traceback (most recent call last):
+ ...
+ ValueError: Missing an input
+ >>> manhattan_distance([1,1], [2, 2, 2])
+ Traceback (most recent call last):
+ ...
+ ValueError: Both points must be in the same n-dimensional space
+ >>> manhattan_distance([1,"one"], [2, 2, 2])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found str
+ >>> manhattan_distance(1, [2, 2, 2])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found int
+ >>> manhattan_distance([1,1], "not_a_list")
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found str
+ """
+
+ _validate_point(point_a)
+ _validate_point(point_b)
+ if len(point_a) != len(point_b):
+ raise ValueError("Both points must be in the same n-dimensional space")
+
+ return float(sum(abs(a - b) for a, b in zip(point_a, point_b)))
+
+
+def _validate_point(point: list[float]) -> None:
+ """
+ >>> _validate_point(None)
+ Traceback (most recent call last):
+ ...
+ ValueError: Missing an input
+ >>> _validate_point([1,"one"])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found str
+ >>> _validate_point(1)
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found int
+ >>> _validate_point("not_a_list")
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found str
+ """
+ if point:
+ if isinstance(point, list):
+ for item in point:
+ if not isinstance(item, (int, float)):
+ msg = (
+ "Expected a list of numbers as input, found "
+ f"{type(item).__name__}"
+ )
+ raise TypeError(msg)
+ else:
+ msg = f"Expected a list of numbers as input, found {type(point).__name__}"
+ raise TypeError(msg)
+ else:
+ raise ValueError("Missing an input")
+
+
+def manhattan_distance_one_liner(point_a: list, point_b: list) -> float:
+ """
+ Version with one liner
+
+ >>> manhattan_distance_one_liner([1,1], [2,2])
+ 2.0
+ >>> manhattan_distance_one_liner([1.5,1.5], [2,2])
+ 1.0
+ >>> manhattan_distance_one_liner([1.5,1.5], [2.5,2])
+ 1.5
+ >>> manhattan_distance_one_liner([-3, -3, -3], [0, 0, 0])
+ 9.0
+ >>> manhattan_distance_one_liner([1,1], None)
+ Traceback (most recent call last):
+ ...
+ ValueError: Missing an input
+ >>> manhattan_distance_one_liner([1,1], [2, 2, 2])
+ Traceback (most recent call last):
+ ...
+ ValueError: Both points must be in the same n-dimensional space
+ >>> manhattan_distance_one_liner([1,"one"], [2, 2, 2])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found str
+ >>> manhattan_distance_one_liner(1, [2, 2, 2])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found int
+ >>> manhattan_distance_one_liner([1,1], "not_a_list")
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a list of numbers as input, found str
+ """
+
+ _validate_point(point_a)
+ _validate_point(point_b)
+ if len(point_a) != len(point_b):
+ raise ValueError("Both points must be in the same n-dimensional space")
+
+ return float(sum(abs(x - y) for x, y in zip(point_a, point_b)))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/matrix_exponentiation.py b/maths/matrix_exponentiation.py
index 033ceb3f28a0..7cdac9d34674 100644
--- a/maths/matrix_exponentiation.py
+++ b/maths/matrix_exponentiation.py
@@ -5,7 +5,7 @@
"""
Matrix Exponentiation is a technique to solve linear recurrences in logarithmic time.
You read more about it here:
-http://zobayer.blogspot.com/2010/11/matrix-exponentiation.html
+https://zobayer.blogspot.com/2010/11/matrix-exponentiation.html
https://www.hackerearth.com/practice/notes/matrix-exponentiation-1/
"""
@@ -39,6 +39,21 @@ def modular_exponentiation(a, b):
def fibonacci_with_matrix_exponentiation(n, f1, f2):
+ """
+ Returns the nth number of the Fibonacci sequence that
+ starts with f1 and f2
+ Uses the matrix exponentiation
+ >>> fibonacci_with_matrix_exponentiation(1, 5, 6)
+ 5
+ >>> fibonacci_with_matrix_exponentiation(2, 10, 11)
+ 11
+ >>> fibonacci_with_matrix_exponentiation(13, 0, 1)
+ 144
+ >>> fibonacci_with_matrix_exponentiation(10, 5, 9)
+ 411
+ >>> fibonacci_with_matrix_exponentiation(9, 2, 3)
+ 89
+ """
# Trivial Cases
if n == 1:
return f1
@@ -50,21 +65,34 @@ def fibonacci_with_matrix_exponentiation(n, f1, f2):
def simple_fibonacci(n, f1, f2):
+ """
+ Returns the nth number of the Fibonacci sequence that
+ starts with f1 and f2
+ Uses the definition
+ >>> simple_fibonacci(1, 5, 6)
+ 5
+ >>> simple_fibonacci(2, 10, 11)
+ 11
+ >>> simple_fibonacci(13, 0, 1)
+ 144
+ >>> simple_fibonacci(10, 5, 9)
+ 411
+ >>> simple_fibonacci(9, 2, 3)
+ 89
+ """
# Trivial Cases
if n == 1:
return f1
elif n == 2:
return f2
- fn_1 = f1
- fn_2 = f2
n -= 2
while n > 0:
- fn_1, fn_2 = fn_1 + fn_2, fn_1
+ f2, f1 = f1 + f2, f2
n -= 1
- return fn_1
+ return f2
def matrix_exponentiation_time():
diff --git a/other/max_sum_sliding_window.py b/maths/max_sum_sliding_window.py
similarity index 85%
rename from other/max_sum_sliding_window.py
rename to maths/max_sum_sliding_window.py
index 4be7d786f215..c7492978a6c9 100644
--- a/other/max_sum_sliding_window.py
+++ b/maths/max_sum_sliding_window.py
@@ -1,45 +1,48 @@
-"""
-Given an array of integer elements and an integer 'k', we are required to find the
-maximum sum of 'k' consecutive elements in the array.
-
-Instead of using a nested for loop, in a Brute force approach we will use a technique
-called 'Window sliding technique' where the nested loops can be converted to a single
-loop to reduce time complexity.
-"""
-from typing import List
-
-
-def max_sum_in_array(array: List[int], k: int) -> int:
- """
- Returns the maximum sum of k consecutive elements
- >>> arr = [1, 4, 2, 10, 2, 3, 1, 0, 20]
- >>> k = 4
- >>> max_sum_in_array(arr, k)
- 24
- >>> k = 10
- >>> max_sum_in_array(arr,k)
- Traceback (most recent call last):
- ...
- ValueError: Invalid Input
- >>> arr = [1, 4, 2, 10, 2, 13, 1, 0, 2]
- >>> k = 4
- >>> max_sum_in_array(arr, k)
- 27
- """
- if len(array) < k or k < 0:
- raise ValueError("Invalid Input")
- max_sum = current_sum = sum(array[:k])
- for i in range(len(array) - k):
- current_sum = current_sum - array[i] + array[i + k]
- max_sum = max(max_sum, current_sum)
- return max_sum
-
-
-if __name__ == "__main__":
- from doctest import testmod
- from random import randint
-
- testmod()
- array = [randint(-1000, 1000) for i in range(100)]
- k = randint(0, 110)
- print(f"The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}")
+"""
+Given an array of integer elements and an integer 'k', we are required to find the
+maximum sum of 'k' consecutive elements in the array.
+
+Instead of using a nested for loop, in a Brute force approach we will use a technique
+called 'Window sliding technique' where the nested loops can be converted to a single
+loop to reduce time complexity.
+"""
+
+from __future__ import annotations
+
+
+def max_sum_in_array(array: list[int], k: int) -> int:
+ """
+ Returns the maximum sum of k consecutive elements
+ >>> arr = [1, 4, 2, 10, 2, 3, 1, 0, 20]
+ >>> k = 4
+ >>> max_sum_in_array(arr, k)
+ 24
+ >>> k = 10
+ >>> max_sum_in_array(arr,k)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid Input
+ >>> arr = [1, 4, 2, 10, 2, 13, 1, 0, 2]
+ >>> k = 4
+ >>> max_sum_in_array(arr, k)
+ 27
+ """
+ if len(array) < k or k < 0:
+ raise ValueError("Invalid Input")
+ max_sum = current_sum = sum(array[:k])
+ for i in range(len(array) - k):
+ current_sum = current_sum - array[i] + array[i + k]
+ max_sum = max(max_sum, current_sum)
+ return max_sum
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+ from random import randint
+
+ testmod()
+ array = [randint(-1000, 1000) for i in range(100)]
+ k = randint(0, 110)
+ print(
+ f"The maximum sum of {k} consecutive elements is {max_sum_in_array(array, k)}"
+ )
diff --git a/maths/miller_rabin.py b/maths/miller_rabin.py
deleted file mode 100644
index fe992027190b..000000000000
--- a/maths/miller_rabin.py
+++ /dev/null
@@ -1,50 +0,0 @@
-import random
-
-from .binary_exp_mod import bin_exp_mod
-
-
-# This is a probabilistic check to test primality, useful for big numbers!
-# if it's a prime, it will return true
-# if it's not a prime, the chance of it returning true is at most 1/4**prec
-def is_prime(n, prec=1000):
- """
- >>> from .prime_check import prime_check
- >>> all(is_prime(i) == prime_check(i) for i in range(1000))
- True
- """
- if n < 2:
- return False
-
- if n % 2 == 0:
- return n == 2
-
- # this means n is odd
- d = n - 1
- exp = 0
- while d % 2 == 0:
- d /= 2
- exp += 1
-
- # n - 1=d*(2**exp)
- count = 0
- while count < prec:
- a = random.randint(2, n - 1)
- b = bin_exp_mod(a, d, n)
- if b != 1:
- flag = True
- for i in range(exp):
- if b == n - 1:
- flag = False
- break
- b = b * b
- b %= n
- if flag:
- return False
- count += 1
- return True
-
-
-if __name__ == "__main__":
- n = abs(int(input("Enter bound : ").strip()))
- print("Here's the list of primes:")
- print(", ".join(str(i) for i in range(n + 1) if is_prime(i)))
diff --git a/maths/minkowski_distance.py b/maths/minkowski_distance.py
new file mode 100644
index 000000000000..99f02e31e417
--- /dev/null
+++ b/maths/minkowski_distance.py
@@ -0,0 +1,45 @@
+def minkowski_distance(
+ point_a: list[float],
+ point_b: list[float],
+ order: int,
+) -> float:
+ """
+ This function calculates the Minkowski distance for a given order between
+ two n-dimensional points represented as lists. For the case of order = 1,
+ the Minkowski distance degenerates to the Manhattan distance. For
+ order = 2, the usual Euclidean distance is obtained.
+
+ https://en.wikipedia.org/wiki/Minkowski_distance
+
+ Note: due to floating point calculation errors the output of this
+ function may be inaccurate.
+
+ >>> minkowski_distance([1.0, 1.0], [2.0, 2.0], 1)
+ 2.0
+ >>> minkowski_distance([1.0, 2.0, 3.0, 4.0], [5.0, 6.0, 7.0, 8.0], 2)
+ 8.0
+ >>> import numpy as np
+ >>> bool(np.isclose(5.0, minkowski_distance([5.0], [0.0], 3)))
+ True
+ >>> minkowski_distance([1.0], [2.0], -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The order must be greater than or equal to 1.
+ >>> minkowski_distance([1.0], [1.0, 2.0], 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Both points must have the same dimension.
+ """
+ if order < 1:
+ raise ValueError("The order must be greater than or equal to 1.")
+
+ if len(point_a) != len(point_b):
+ raise ValueError("Both points must have the same dimension.")
+
+ return sum(abs(a - b) ** order for a, b in zip(point_a, point_b)) ** (1 / order)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/mobius_function.py b/maths/mobius_function.py
index 4fcf35f21813..8abdc4cafcb4 100644
--- a/maths/mobius_function.py
+++ b/maths/mobius_function.py
@@ -1,8 +1,8 @@
"""
References: https://en.wikipedia.org/wiki/M%C3%B6bius_function
References: wikipedia:square free number
-python/black : True
-flake8 : True
+psf/black : True
+ruff : True
"""
from maths.is_square_free import is_square_free
diff --git a/blockchain/modular_division.py b/maths/modular_division.py
similarity index 59%
rename from blockchain/modular_division.py
rename to maths/modular_division.py
index 8fcf6e37cbed..2f8f4479b27d 100644
--- a/blockchain/modular_division.py
+++ b/maths/modular_division.py
@@ -1,21 +1,23 @@
-# Modular Division :
-# An efficient algorithm for dividing b by a modulo n.
+from __future__ import annotations
-# GCD ( Greatest Common Divisor ) or HCF ( Highest Common Factor )
-# Given three integers a, b, and n, such that gcd(a,n)=1 and n>1, the algorithm should
-# return an integer x such that 0≤x≤n−1, and b/a=x(modn) (that is, b=ax(modn)).
+def modular_division(a: int, b: int, n: int) -> int:
+ """
+ Modular Division :
+ An efficient algorithm for dividing b by a modulo n.
-# Theorem:
-# a has a multiplicative inverse modulo n iff gcd(a,n) = 1
+ GCD ( Greatest Common Divisor ) or HCF ( Highest Common Factor )
+ Given three integers a, b, and n, such that gcd(a,n)=1 and n>1, the algorithm should
+ return an integer x such that 0≤x≤n-1, and b/a=x(modn) (that is, b=ax(modn)).
-# This find x = b*a^(-1) mod n
-# Uses ExtendedEuclid to find the inverse of a
+ Theorem:
+ a has a multiplicative inverse modulo n iff gcd(a,n) = 1
-def modular_division(a, b, n):
- """
+ This find x = b*a^(-1) mod n
+ Uses ExtendedEuclid to find the inverse of a
+
>>> modular_division(4,8,5)
2
@@ -26,15 +28,18 @@ def modular_division(a, b, n):
4
"""
- assert n > 1 and a > 0 and greatest_common_divisor(a, n) == 1
+ assert n > 1
+ assert a > 0
+ assert greatest_common_divisor(a, n) == 1
(d, t, s) = extended_gcd(n, a) # Implemented below
x = (b * s) % n
return x
-# This function find the inverses of a i.e., a^(-1)
-def invert_modulo(a, n):
+def invert_modulo(a: int, n: int) -> int:
"""
+ This function find the inverses of a i.e., a^(-1)
+
>>> invert_modulo(2, 5)
3
@@ -50,9 +55,11 @@ def invert_modulo(a, n):
# ------------------ Finding Modular division using invert_modulo -------------------
-# This function used the above inversion of a to find x = (b*a^(-1))mod n
-def modular_division2(a, b, n):
+
+def modular_division2(a: int, b: int, n: int) -> int:
"""
+ This function used the above inversion of a to find x = (b*a^(-1))mod n
+
>>> modular_division2(4,8,5)
2
@@ -68,22 +75,21 @@ def modular_division2(a, b, n):
return x
-# Extended Euclid's Algorithm : If d divides a and b and d = a*x + b*y for integers x
-# and y, then d = gcd(a,b)
-
-
-def extended_gcd(a, b):
+def extended_gcd(a: int, b: int) -> tuple[int, int, int]:
"""
- >>> extended_gcd(10, 6)
- (2, -1, 2)
+ Extended Euclid's Algorithm : If d divides a and b and d = a*x + b*y for integers x
+ and y, then d = gcd(a,b)
+ >>> extended_gcd(10, 6)
+ (2, -1, 2)
- >>> extended_gcd(7, 5)
- (1, -2, 3)
+ >>> extended_gcd(7, 5)
+ (1, -2, 3)
** extended_gcd function is used when d = gcd(a,b) is required in output
"""
- assert a >= 0 and b >= 0
+ assert a >= 0
+ assert b >= 0
if b == 0:
d, x, y = a, 1, 0
@@ -92,15 +98,16 @@ def extended_gcd(a, b):
x = q
y = p - q * (a // b)
- assert a % d == 0 and b % d == 0
+ assert a % d == 0
+ assert b % d == 0
assert d == a * x + b * y
return (d, x, y)
-# Extended Euclid
-def extended_euclid(a, b):
+def extended_euclid(a: int, b: int) -> tuple[int, int]:
"""
+ Extended Euclid
>>> extended_euclid(10, 6)
(-1, 2)
@@ -115,12 +122,11 @@ def extended_euclid(a, b):
return (y, x - k * y)
-# Euclid's Lemma : d divides a and b, if and only if d divides a-b and b
-# Euclid's Algorithm
-
-
-def greatest_common_divisor(a, b):
+def greatest_common_divisor(a: int, b: int) -> int:
"""
+ Euclid's Lemma : d divides a and b, if and only if d divides a-b and b
+ Euclid's Algorithm
+
>>> greatest_common_divisor(7,5)
1
diff --git a/maths/modular_exponential.py b/maths/modular_exponential.py
index 42987dbf3a24..a27e29ebc02a 100644
--- a/maths/modular_exponential.py
+++ b/maths/modular_exponential.py
@@ -1,8 +1,8 @@
"""
- Modular Exponential.
- Modular exponentiation is a type of exponentiation performed over a modulus.
- For more explanation, please check
- https://en.wikipedia.org/wiki/Modular_exponentiation
+Modular Exponential.
+Modular exponentiation is a type of exponentiation performed over a modulus.
+For more explanation, please check
+https://en.wikipedia.org/wiki/Modular_exponentiation
"""
"""Calculate Modular Exponential."""
diff --git a/maths/monte_carlo.py b/maths/monte_carlo.py
index 28027cbe4178..d174a0b188a2 100644
--- a/maths/monte_carlo.py
+++ b/maths/monte_carlo.py
@@ -1,10 +1,11 @@
"""
@author: MatteoRaso
"""
+
+from collections.abc import Callable
from math import pi, sqrt
from random import uniform
from statistics import mean
-from typing import Callable
def pi_estimator(iterations: int):
@@ -18,9 +19,10 @@ def pi_estimator(iterations: int):
5. Multiply this value by 4 to get your estimate of pi.
6. Print the estimated and numpy value of pi
"""
+
# A local function to see if a dot lands in the circle.
def is_in_circle(x: float, y: float) -> bool:
- distance_from_centre = sqrt((x ** 2) + (y ** 2))
+ distance_from_centre = sqrt((x**2) + (y**2))
# Our circle has a radius of 1, so a distance
# greater than 1 would land outside the circle.
return distance_from_centre <= 1
diff --git a/maths/monte_carlo_dice.py b/maths/monte_carlo_dice.py
index e8e3abe83a99..362f70b49828 100644
--- a/maths/monte_carlo_dice.py
+++ b/maths/monte_carlo_dice.py
@@ -7,15 +7,12 @@ class Dice:
NUM_SIDES = 6
def __init__(self):
- """ Initialize a six sided dice """
+ """Initialize a six sided dice"""
self.sides = list(range(1, Dice.NUM_SIDES + 1))
def roll(self):
return random.choice(self.sides)
- def _str_(self):
- return "Fair Dice"
-
def throw_dice(num_throws: int, num_dice: int = 2) -> list[float]:
"""
@@ -35,7 +32,7 @@ def throw_dice(num_throws: int, num_dice: int = 2) -> list[float]:
"""
dices = [Dice() for i in range(num_dice)]
count_of_sum = [0] * (len(dices) * Dice.NUM_SIDES + 1)
- for i in range(num_throws):
+ for _ in range(num_throws):
count_of_sum[sum(dice.roll() for dice in dices)] += 1
probability = [round((count * 100) / num_throws, 2) for count in count_of_sum]
return probability[num_dice:] # remove probability of sums that never appear
diff --git a/maths/newton_raphson.py b/maths/newton_raphson.py
deleted file mode 100644
index f2b7cb9766d2..000000000000
--- a/maths/newton_raphson.py
+++ /dev/null
@@ -1,54 +0,0 @@
-"""
- Author: P Shreyas Shetty
- Implementation of Newton-Raphson method for solving equations of kind
- f(x) = 0. It is an iterative method where solution is found by the expression
- x[n+1] = x[n] + f(x[n])/f'(x[n])
- If no solution exists, then either the solution will not be found when iteration
- limit is reached or the gradient f'(x[n]) approaches zero. In both cases, exception
- is raised. If iteration limit is reached, try increasing maxiter.
- """
-import math as m
-
-
-def calc_derivative(f, a, h=0.001):
- """
- Calculates derivative at point a for function f using finite difference
- method
- """
- return (f(a + h) - f(a - h)) / (2 * h)
-
-
-def newton_raphson(f, x0=0, maxiter=100, step=0.0001, maxerror=1e-6, logsteps=False):
-
- a = x0 # set the initial guess
- steps = [a]
- error = abs(f(a))
- f1 = lambda x: calc_derivative(f, x, h=step) # noqa: E731 Derivative of f(x)
- for _ in range(maxiter):
- if f1(a) == 0:
- raise ValueError("No converging solution found")
- a = a - f(a) / f1(a) # Calculate the next estimate
- if logsteps:
- steps.append(a)
- if error < maxerror:
- break
- else:
- raise ValueError("Iteration limit reached, no converging solution found")
- if logsteps:
- # If logstep is true, then log intermediate steps
- return a, error, steps
- return a, error
-
-
-if __name__ == "__main__":
- from matplotlib import pyplot as plt
-
- f = lambda x: m.tanh(x) ** 2 - m.exp(3 * x) # noqa: E731
- solution, error, steps = newton_raphson(
- f, x0=10, maxiter=1000, step=1e-6, logsteps=True
- )
- plt.plot([abs(f(x)) for x in steps])
- plt.xlabel("step")
- plt.ylabel("error")
- plt.show()
- print(f"solution = {{{solution:f}}}, error = {{{error:f}}}")
diff --git a/maths/number_of_digits.py b/maths/number_of_digits.py
index 3c0eb7b3863f..bb9c0d248fd1 100644
--- a/maths/number_of_digits.py
+++ b/maths/number_of_digits.py
@@ -16,7 +16,15 @@ def num_digits(n: int) -> int:
1
>>> num_digits(-123456)
6
+ >>> num_digits('123') # Raises a TypeError for non-integer input
+ Traceback (most recent call last):
+ ...
+ TypeError: Input must be an integer
"""
+
+ if not isinstance(n, int):
+ raise TypeError("Input must be an integer")
+
digits = 0
n = abs(n)
while True:
@@ -42,7 +50,15 @@ def num_digits_fast(n: int) -> int:
1
>>> num_digits_fast(-123456)
6
+ >>> num_digits('123') # Raises a TypeError for non-integer input
+ Traceback (most recent call last):
+ ...
+ TypeError: Input must be an integer
"""
+
+ if not isinstance(n, int):
+ raise TypeError("Input must be an integer")
+
return 1 if n == 0 else math.floor(math.log(abs(n), 10) + 1)
@@ -61,99 +77,37 @@ def num_digits_faster(n: int) -> int:
1
>>> num_digits_faster(-123456)
6
+ >>> num_digits('123') # Raises a TypeError for non-integer input
+ Traceback (most recent call last):
+ ...
+ TypeError: Input must be an integer
"""
+
+ if not isinstance(n, int):
+ raise TypeError("Input must be an integer")
+
return len(str(abs(n)))
def benchmark() -> None:
"""
- Benchmark code for comparing 3 functions,
- with 3 different length int values.
+ Benchmark multiple functions, with three different length int values.
"""
- print("\nFor small_num = ", small_num, ":")
- print(
- "> num_digits()",
- "\t\tans =",
- num_digits(small_num),
- "\ttime =",
- timeit("z.num_digits(z.small_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> num_digits_fast()",
- "\tans =",
- num_digits_fast(small_num),
- "\ttime =",
- timeit("z.num_digits_fast(z.small_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> num_digits_faster()",
- "\tans =",
- num_digits_faster(small_num),
- "\ttime =",
- timeit("z.num_digits_faster(z.small_num)", setup="import __main__ as z"),
- "seconds",
- )
-
- print("\nFor medium_num = ", medium_num, ":")
- print(
- "> num_digits()",
- "\t\tans =",
- num_digits(medium_num),
- "\ttime =",
- timeit("z.num_digits(z.medium_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> num_digits_fast()",
- "\tans =",
- num_digits_fast(medium_num),
- "\ttime =",
- timeit("z.num_digits_fast(z.medium_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> num_digits_faster()",
- "\tans =",
- num_digits_faster(medium_num),
- "\ttime =",
- timeit("z.num_digits_faster(z.medium_num)", setup="import __main__ as z"),
- "seconds",
- )
-
- print("\nFor large_num = ", large_num, ":")
- print(
- "> num_digits()",
- "\t\tans =",
- num_digits(large_num),
- "\ttime =",
- timeit("z.num_digits(z.large_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> num_digits_fast()",
- "\tans =",
- num_digits_fast(large_num),
- "\ttime =",
- timeit("z.num_digits_fast(z.large_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> num_digits_faster()",
- "\tans =",
- num_digits_faster(large_num),
- "\ttime =",
- timeit("z.num_digits_faster(z.large_num)", setup="import __main__ as z"),
- "seconds",
- )
+ from collections.abc import Callable
+
+ def benchmark_a_function(func: Callable, value: int) -> None:
+ call = f"{func.__name__}({value})"
+ timing = timeit(f"__main__.{call}", setup="import __main__")
+ print(f"{call}: {func(value)} -- {timing} seconds")
+
+ for value in (262144, 1125899906842624, 1267650600228229401496703205376):
+ for func in (num_digits, num_digits_fast, num_digits_faster):
+ benchmark_a_function(func, value)
+ print()
if __name__ == "__main__":
- small_num = 262144
- medium_num = 1125899906842624
- large_num = 1267650600228229401496703205376
- benchmark()
import doctest
doctest.testmod()
+ benchmark()
diff --git a/maths/numerical_analysis/__init__.py b/maths/numerical_analysis/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/maths/numerical_analysis/adams_bashforth.py b/maths/numerical_analysis/adams_bashforth.py
new file mode 100644
index 000000000000..26244a58552f
--- /dev/null
+++ b/maths/numerical_analysis/adams_bashforth.py
@@ -0,0 +1,231 @@
+"""
+Use the Adams-Bashforth methods to solve Ordinary Differential Equations.
+
+https://en.wikipedia.org/wiki/Linear_multistep_method
+Author : Ravi Kumar
+"""
+
+from collections.abc import Callable
+from dataclasses import dataclass
+
+import numpy as np
+
+
+@dataclass
+class AdamsBashforth:
+ """
+ args:
+ func: An ordinary differential equation (ODE) as function of x and y.
+ x_initials: List containing initial required values of x.
+ y_initials: List containing initial required values of y.
+ step_size: The increment value of x.
+ x_final: The final value of x.
+
+ Returns: Solution of y at each nodal point
+
+ >>> def f(x, y):
+ ... return x + y
+ >>> AdamsBashforth(f, [0, 0.2, 0.4], [0, 0.2, 1], 0.2, 1) # doctest: +ELLIPSIS
+ AdamsBashforth(func=..., x_initials=[0, 0.2, 0.4], y_initials=[0, 0.2, 1], step...)
+ >>> AdamsBashforth(f, [0, 0.2, 1], [0, 0, 0.04], 0.2, 1).step_2()
+ Traceback (most recent call last):
+ ...
+ ValueError: The final value of x must be greater than the initial values of x.
+
+ >>> AdamsBashforth(f, [0, 0.2, 0.3], [0, 0, 0.04], 0.2, 1).step_3()
+ Traceback (most recent call last):
+ ...
+ ValueError: x-values must be equally spaced according to step size.
+
+ >>> AdamsBashforth(f,[0,0.2,0.4,0.6,0.8],[0,0,0.04,0.128,0.307],-0.2,1).step_5()
+ Traceback (most recent call last):
+ ...
+ ValueError: Step size must be positive.
+ """
+
+ func: Callable[[float, float], float]
+ x_initials: list[float]
+ y_initials: list[float]
+ step_size: float
+ x_final: float
+
+ def __post_init__(self) -> None:
+ if self.x_initials[-1] >= self.x_final:
+ raise ValueError(
+ "The final value of x must be greater than the initial values of x."
+ )
+
+ if self.step_size <= 0:
+ raise ValueError("Step size must be positive.")
+
+ if not all(
+ round(x1 - x0, 10) == self.step_size
+ for x0, x1 in zip(self.x_initials, self.x_initials[1:])
+ ):
+ raise ValueError("x-values must be equally spaced according to step size.")
+
+ def step_2(self) -> np.ndarray:
+ """
+ >>> def f(x, y):
+ ... return x
+ >>> AdamsBashforth(f, [0, 0.2], [0, 0], 0.2, 1).step_2()
+ array([0. , 0. , 0.06, 0.16, 0.3 , 0.48])
+
+ >>> AdamsBashforth(f, [0, 0.2, 0.4], [0, 0, 0.04], 0.2, 1).step_2()
+ Traceback (most recent call last):
+ ...
+ ValueError: Insufficient initial points information.
+ """
+
+ if len(self.x_initials) != 2 or len(self.y_initials) != 2:
+ raise ValueError("Insufficient initial points information.")
+
+ x_0, x_1 = self.x_initials[:2]
+ y_0, y_1 = self.y_initials[:2]
+
+ n = int((self.x_final - x_1) / self.step_size)
+ y = np.zeros(n + 2)
+ y[0] = y_0
+ y[1] = y_1
+
+ for i in range(n):
+ y[i + 2] = y[i + 1] + (self.step_size / 2) * (
+ 3 * self.func(x_1, y[i + 1]) - self.func(x_0, y[i])
+ )
+ x_0 = x_1
+ x_1 += self.step_size
+
+ return y
+
+ def step_3(self) -> np.ndarray:
+ """
+ >>> def f(x, y):
+ ... return x + y
+ >>> y = AdamsBashforth(f, [0, 0.2, 0.4], [0, 0, 0.04], 0.2, 1).step_3()
+ >>> float(y[3])
+ 0.15533333333333332
+
+ >>> AdamsBashforth(f, [0, 0.2], [0, 0], 0.2, 1).step_3()
+ Traceback (most recent call last):
+ ...
+ ValueError: Insufficient initial points information.
+ """
+ if len(self.x_initials) != 3 or len(self.y_initials) != 3:
+ raise ValueError("Insufficient initial points information.")
+
+ x_0, x_1, x_2 = self.x_initials[:3]
+ y_0, y_1, y_2 = self.y_initials[:3]
+
+ n = int((self.x_final - x_2) / self.step_size)
+ y = np.zeros(n + 4)
+ y[0] = y_0
+ y[1] = y_1
+ y[2] = y_2
+
+ for i in range(n + 1):
+ y[i + 3] = y[i + 2] + (self.step_size / 12) * (
+ 23 * self.func(x_2, y[i + 2])
+ - 16 * self.func(x_1, y[i + 1])
+ + 5 * self.func(x_0, y[i])
+ )
+ x_0 = x_1
+ x_1 = x_2
+ x_2 += self.step_size
+
+ return y
+
+ def step_4(self) -> np.ndarray:
+ """
+ >>> def f(x,y):
+ ... return x + y
+ >>> y = AdamsBashforth(
+ ... f, [0, 0.2, 0.4, 0.6], [0, 0, 0.04, 0.128], 0.2, 1).step_4()
+ >>> float(y[4])
+ 0.30699999999999994
+ >>> float(y[5])
+ 0.5771083333333333
+
+ >>> AdamsBashforth(f, [0, 0.2, 0.4], [0, 0, 0.04], 0.2, 1).step_4()
+ Traceback (most recent call last):
+ ...
+ ValueError: Insufficient initial points information.
+ """
+
+ if len(self.x_initials) != 4 or len(self.y_initials) != 4:
+ raise ValueError("Insufficient initial points information.")
+
+ x_0, x_1, x_2, x_3 = self.x_initials[:4]
+ y_0, y_1, y_2, y_3 = self.y_initials[:4]
+
+ n = int((self.x_final - x_3) / self.step_size)
+ y = np.zeros(n + 4)
+ y[0] = y_0
+ y[1] = y_1
+ y[2] = y_2
+ y[3] = y_3
+
+ for i in range(n):
+ y[i + 4] = y[i + 3] + (self.step_size / 24) * (
+ 55 * self.func(x_3, y[i + 3])
+ - 59 * self.func(x_2, y[i + 2])
+ + 37 * self.func(x_1, y[i + 1])
+ - 9 * self.func(x_0, y[i])
+ )
+ x_0 = x_1
+ x_1 = x_2
+ x_2 = x_3
+ x_3 += self.step_size
+
+ return y
+
+ def step_5(self) -> np.ndarray:
+ """
+ >>> def f(x,y):
+ ... return x + y
+ >>> y = AdamsBashforth(
+ ... f, [0, 0.2, 0.4, 0.6, 0.8], [0, 0.02140, 0.02140, 0.22211, 0.42536],
+ ... 0.2, 1).step_5()
+ >>> float(y[-1])
+ 0.05436839444444452
+
+ >>> AdamsBashforth(f, [0, 0.2, 0.4], [0, 0, 0.04], 0.2, 1).step_5()
+ Traceback (most recent call last):
+ ...
+ ValueError: Insufficient initial points information.
+ """
+
+ if len(self.x_initials) != 5 or len(self.y_initials) != 5:
+ raise ValueError("Insufficient initial points information.")
+
+ x_0, x_1, x_2, x_3, x_4 = self.x_initials[:5]
+ y_0, y_1, y_2, y_3, y_4 = self.y_initials[:5]
+
+ n = int((self.x_final - x_4) / self.step_size)
+ y = np.zeros(n + 6)
+ y[0] = y_0
+ y[1] = y_1
+ y[2] = y_2
+ y[3] = y_3
+ y[4] = y_4
+
+ for i in range(n + 1):
+ y[i + 5] = y[i + 4] + (self.step_size / 720) * (
+ 1901 * self.func(x_4, y[i + 4])
+ - 2774 * self.func(x_3, y[i + 3])
+ - 2616 * self.func(x_2, y[i + 2])
+ - 1274 * self.func(x_1, y[i + 1])
+ + 251 * self.func(x_0, y[i])
+ )
+ x_0 = x_1
+ x_1 = x_2
+ x_2 = x_3
+ x_3 = x_4
+ x_4 += self.step_size
+
+ return y
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/arithmetic_analysis/bisection.py b/maths/numerical_analysis/bisection.py
similarity index 90%
rename from arithmetic_analysis/bisection.py
rename to maths/numerical_analysis/bisection.py
index 0ef691678702..e359cc170072 100644
--- a/arithmetic_analysis/bisection.py
+++ b/maths/numerical_analysis/bisection.py
@@ -1,4 +1,4 @@
-from typing import Callable
+from collections.abc import Callable
def bisection(function: Callable[[float], float], a: float, b: float) -> float:
@@ -8,7 +8,7 @@ def bisection(function: Callable[[float], float], a: float, b: float) -> float:
1.0000000149011612
>>> bisection(lambda x: x ** 3 - 1, 2, 1000)
Traceback (most recent call last):
- ...
+ ...
ValueError: could not find root in given interval.
>>> bisection(lambda x: x ** 2 - 4 * x + 3, 0, 2)
1.0
@@ -16,7 +16,7 @@ def bisection(function: Callable[[float], float], a: float, b: float) -> float:
3.0
>>> bisection(lambda x: x ** 2 - 4 * x + 3, 4, 1000)
Traceback (most recent call last):
- ...
+ ...
ValueError: could not find root in given interval.
"""
start: float = a
@@ -32,7 +32,7 @@ def bisection(function: Callable[[float], float], a: float, b: float) -> float:
raise ValueError("could not find root in given interval.")
else:
mid: float = start + (end - start) / 2.0
- while abs(start - mid) > 10 ** -7: # until precisely equals to 10^-7
+ while abs(start - mid) > 10**-7: # until precisely equals to 10^-7
if function(mid) == 0:
return mid
elif function(mid) * function(start) < 0:
@@ -44,7 +44,7 @@ def bisection(function: Callable[[float], float], a: float, b: float) -> float:
def f(x: float) -> float:
- return x ** 3 - 2 * x - 5
+ return x**3 - 2 * x - 5
if __name__ == "__main__":
diff --git a/maths/bisection.py b/maths/numerical_analysis/bisection_2.py
similarity index 97%
rename from maths/bisection.py
rename to maths/numerical_analysis/bisection_2.py
index 93cc2247b64e..68ba6577ce29 100644
--- a/maths/bisection.py
+++ b/maths/numerical_analysis/bisection_2.py
@@ -1,5 +1,5 @@
"""
-Given a function on floating number f(x) and two floating numbers ‘a’ and ‘b’ such that
+Given a function on floating number f(x) and two floating numbers `a` and `b` such that
f(a) * f(b) < 0 and f(x) is continuous in [a, b].
Here f(x) represents algebraic or transcendental equation.
Find root of function in interval [a, b] (Or find a value of x such that f(x) is 0)
@@ -32,7 +32,7 @@ def bisection(a: float, b: float) -> float:
3.158203125
>>> bisection(2, 3)
Traceback (most recent call last):
- ...
+ ...
ValueError: Wrong space!
"""
# Bolzano theory in order to find if there is a root between a and b
diff --git a/other/integeration_by_simpson_approx.py b/maths/numerical_analysis/integration_by_simpson_approx.py
similarity index 82%
rename from other/integeration_by_simpson_approx.py
rename to maths/numerical_analysis/integration_by_simpson_approx.py
index da0e1cffde02..043f3a9a72af 100644
--- a/other/integeration_by_simpson_approx.py
+++ b/maths/numerical_analysis/integration_by_simpson_approx.py
@@ -4,7 +4,7 @@
Purpose : You have one function f(x) which takes float integer and returns
float you have to integrate the function in limits a to b.
-The approximation proposed by Thomas Simpsons in 1743 is one way to calculate
+The approximation proposed by Thomas Simpson in 1743 is one way to calculate
integration.
( read article : https://cp-algorithms.com/num_methods/simpson-integration.html )
@@ -35,12 +35,11 @@ def f(x: float) -> float:
def simpson_integration(function, a: float, b: float, precision: int = 4) -> float:
-
"""
Args:
function : the function which's integration is desired
a : the lower limit of integration
- b : upper limit of integraion
+ b : upper limit of integration
precision : precision of the result,error required default is 4
Returns:
@@ -89,24 +88,20 @@ def simpson_integration(function, a: float, b: float, precision: int = 4) -> flo
AssertionError: precision should be positive integer your input : -1
"""
- assert callable(
- function
- ), f"the function(object) passed should be callable your input : {function}"
- assert isinstance(a, float) or isinstance(
- a, int
- ), f"a should be float or integer your input : {a}"
- assert isinstance(function(a), float) or isinstance(function(a), int), (
+ assert callable(function), (
+ f"the function(object) passed should be callable your input : {function}"
+ )
+ assert isinstance(a, (float, int)), f"a should be float or integer your input : {a}"
+ assert isinstance(function(a), (float, int)), (
"the function should return integer or float return type of your function, "
f"{type(a)}"
)
- assert isinstance(b, float) or isinstance(
- b, int
- ), f"b should be float or integer your input : {b}"
- assert (
- isinstance(precision, int) and precision > 0
- ), f"precision should be positive integer your input : {precision}"
-
- # just applying the formula of simpson for approximate integraion written in
+ assert isinstance(b, (float, int)), f"b should be float or integer your input : {b}"
+ assert isinstance(precision, int) and precision > 0, (
+ f"precision should be positive integer your input : {precision}"
+ )
+
+ # just applying the formula of simpson for approximate integration written in
# mentioned article in first comment of this file and above this function
h = (b - a) / N_STEPS
diff --git a/arithmetic_analysis/intersection.py b/maths/numerical_analysis/intersection.py
similarity index 89%
rename from arithmetic_analysis/intersection.py
rename to maths/numerical_analysis/intersection.py
index 204dd5d8a935..325abeaca996 100644
--- a/arithmetic_analysis/intersection.py
+++ b/maths/numerical_analysis/intersection.py
@@ -1,5 +1,5 @@
import math
-from typing import Callable
+from collections.abc import Callable
def intersection(function: Callable[[float], float], x0: float, x1: float) -> float:
@@ -10,7 +10,7 @@ def intersection(function: Callable[[float], float], x0: float, x1: float) -> fl
0.9999999999954654
>>> intersection(lambda x: x ** 3 - 1, 5, 5)
Traceback (most recent call last):
- ...
+ ...
ZeroDivisionError: float division by zero, could not find root
>>> intersection(lambda x: x ** 3 - 1, 100, 200)
1.0000000000003888
@@ -24,7 +24,7 @@ def intersection(function: Callable[[float], float], x0: float, x1: float) -> fl
0.0
>>> intersection(math.cos, -math.pi, math.pi)
Traceback (most recent call last):
- ...
+ ...
ZeroDivisionError: float division by zero, could not find root
"""
x_n: float = x0
@@ -35,13 +35,18 @@ def intersection(function: Callable[[float], float], x0: float, x1: float) -> fl
x_n2: float = x_n1 - (
function(x_n1) / ((function(x_n1) - function(x_n)) / (x_n1 - x_n))
)
- if abs(x_n2 - x_n1) < 10 ** -5:
+ if abs(x_n2 - x_n1) < 10**-5:
return x_n2
x_n = x_n1
x_n1 = x_n2
def f(x: float) -> float:
+ """
+ function is f(x) = x^3 - 2x - 5
+ >>> f(2)
+ -1.0
+ """
return math.pow(x, 3) - (2 * x) - 5
diff --git a/maths/numerical_analysis/nevilles_method.py b/maths/numerical_analysis/nevilles_method.py
new file mode 100644
index 000000000000..25c93ac6c531
--- /dev/null
+++ b/maths/numerical_analysis/nevilles_method.py
@@ -0,0 +1,55 @@
+"""
+Python program to show how to interpolate and evaluate a polynomial
+using Neville's method.
+Neville's method evaluates a polynomial that passes through a
+given set of x and y points for a particular x value (x0) using the
+Newton polynomial form.
+Reference:
+ https://rpubs.com/aaronsc32/nevilles-method-polynomial-interpolation
+"""
+
+
+def neville_interpolate(x_points: list, y_points: list, x0: int) -> list:
+ """
+ Interpolate and evaluate a polynomial using Neville's method.
+ Arguments:
+ x_points, y_points: Iterables of x and corresponding y points through
+ which the polynomial passes.
+ x0: The value of x to evaluate the polynomial for.
+ Return Value: A list of the approximated value and the Neville iterations
+ table respectively.
+ >>> import pprint
+ >>> neville_interpolate((1,2,3,4,6), (6,7,8,9,11), 5)[0]
+ 10.0
+ >>> pprint.pprint(neville_interpolate((1,2,3,4,6), (6,7,8,9,11), 99)[1])
+ [[0, 6, 0, 0, 0],
+ [0, 7, 0, 0, 0],
+ [0, 8, 104.0, 0, 0],
+ [0, 9, 104.0, 104.0, 0],
+ [0, 11, 104.0, 104.0, 104.0]]
+ >>> neville_interpolate((1,2,3,4,6), (6,7,8,9,11), 99)[0]
+ 104.0
+ >>> neville_interpolate((1,2,3,4,6), (6,7,8,9,11), '')
+ Traceback (most recent call last):
+ ...
+ TypeError: unsupported operand type(s) for -: 'str' and 'int'
+ """
+ n = len(x_points)
+ q = [[0] * n for i in range(n)]
+ for i in range(n):
+ q[i][1] = y_points[i]
+
+ for i in range(2, n):
+ for j in range(i, n):
+ q[j][i] = (
+ (x0 - x_points[j - i + 1]) * q[j][i - 1]
+ - (x0 - x_points[j]) * q[j - 1][i - 1]
+ ) / (x_points[j] - x_points[j - i + 1])
+
+ return [q[n - 1][n - 1], q]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/arithmetic_analysis/newton_forward_interpolation.py b/maths/numerical_analysis/newton_forward_interpolation.py
similarity index 88%
rename from arithmetic_analysis/newton_forward_interpolation.py
rename to maths/numerical_analysis/newton_forward_interpolation.py
index d32e3efbd1f2..466f6c18cf59 100644
--- a/arithmetic_analysis/newton_forward_interpolation.py
+++ b/maths/numerical_analysis/newton_forward_interpolation.py
@@ -1,10 +1,11 @@
# https://www.geeksforgeeks.org/newton-forward-backward-interpolation/
+from __future__ import annotations
import math
# for calculating u value
-def ucal(u, p):
+def ucal(u: float, p: int) -> float:
"""
>>> ucal(1, 2)
0
@@ -19,10 +20,10 @@ def ucal(u, p):
return temp
-def main():
+def main() -> None:
n = int(input("enter the numbers of values: "))
- y = []
- for i in range(n):
+ y: list[list[float]] = []
+ for _ in range(n):
y.append([])
for i in range(n):
for j in range(n):
diff --git a/maths/numerical_analysis/newton_raphson.py b/maths/numerical_analysis/newton_raphson.py
new file mode 100644
index 000000000000..10fb244bf426
--- /dev/null
+++ b/maths/numerical_analysis/newton_raphson.py
@@ -0,0 +1,114 @@
+"""
+The Newton-Raphson method (aka the Newton method) is a root-finding algorithm that
+approximates a root of a given real-valued function f(x). It is an iterative method
+given by the formula
+
+x_{n + 1} = x_n + f(x_n) / f'(x_n)
+
+with the precision of the approximation increasing as the number of iterations increase.
+
+Reference: https://en.wikipedia.org/wiki/Newton%27s_method
+"""
+
+from collections.abc import Callable
+
+RealFunc = Callable[[float], float]
+
+
+def calc_derivative(f: RealFunc, x: float, delta_x: float = 1e-3) -> float:
+ """
+ Approximate the derivative of a function f(x) at a point x using the finite
+ difference method
+
+ >>> import math
+ >>> tolerance = 1e-5
+ >>> derivative = calc_derivative(lambda x: x**2, 2)
+ >>> math.isclose(derivative, 4, abs_tol=tolerance)
+ True
+ >>> derivative = calc_derivative(math.sin, 0)
+ >>> math.isclose(derivative, 1, abs_tol=tolerance)
+ True
+ """
+ return (f(x + delta_x / 2) - f(x - delta_x / 2)) / delta_x
+
+
+def newton_raphson(
+ f: RealFunc,
+ x0: float = 0,
+ max_iter: int = 100,
+ step: float = 1e-6,
+ max_error: float = 1e-6,
+ log_steps: bool = False,
+) -> tuple[float, float, list[float]]:
+ """
+ Find a root of the given function f using the Newton-Raphson method.
+
+ :param f: A real-valued single-variable function
+ :param x0: Initial guess
+ :param max_iter: Maximum number of iterations
+ :param step: Step size of x, used to approximate f'(x)
+ :param max_error: Maximum approximation error
+ :param log_steps: bool denoting whether to log intermediate steps
+
+ :return: A tuple containing the approximation, the error, and the intermediate
+ steps. If log_steps is False, then an empty list is returned for the third
+ element of the tuple.
+
+ :raises ZeroDivisionError: The derivative approaches 0.
+ :raises ArithmeticError: No solution exists, or the solution isn't found before the
+ iteration limit is reached.
+
+ >>> import math
+ >>> tolerance = 1e-15
+ >>> root, *_ = newton_raphson(lambda x: x**2 - 5*x + 2, 0.4, max_error=tolerance)
+ >>> math.isclose(root, (5 - math.sqrt(17)) / 2, abs_tol=tolerance)
+ True
+ >>> root, *_ = newton_raphson(lambda x: math.log(x) - 1, 2, max_error=tolerance)
+ >>> math.isclose(root, math.e, abs_tol=tolerance)
+ True
+ >>> root, *_ = newton_raphson(math.sin, 1, max_error=tolerance)
+ >>> math.isclose(root, 0, abs_tol=tolerance)
+ True
+ >>> newton_raphson(math.cos, 0)
+ Traceback (most recent call last):
+ ...
+ ZeroDivisionError: No converging solution found, zero derivative
+ >>> newton_raphson(lambda x: x**2 + 1, 2)
+ Traceback (most recent call last):
+ ...
+ ArithmeticError: No converging solution found, iteration limit reached
+ """
+
+ def f_derivative(x: float) -> float:
+ return calc_derivative(f, x, step)
+
+ a = x0 # Set initial guess
+ steps = []
+ for _ in range(max_iter):
+ if log_steps: # Log intermediate steps
+ steps.append(a)
+
+ error = abs(f(a))
+ if error < max_error:
+ return a, error, steps
+
+ if f_derivative(a) == 0:
+ raise ZeroDivisionError("No converging solution found, zero derivative")
+ a -= f(a) / f_derivative(a) # Calculate next estimate
+ raise ArithmeticError("No converging solution found, iteration limit reached")
+
+
+if __name__ == "__main__":
+ import doctest
+ from math import exp, tanh
+
+ doctest.testmod()
+
+ def func(x: float) -> float:
+ return tanh(x) ** 2 - exp(3 * x)
+
+ solution, err, steps = newton_raphson(
+ func, x0=10, max_iter=100, step=1e-6, log_steps=True
+ )
+ print(f"{solution=}, {err=}")
+ print("\n".join(str(x) for x in steps))
diff --git a/maths/numerical_integration.py b/maths/numerical_analysis/numerical_integration.py
similarity index 87%
rename from maths/numerical_integration.py
rename to maths/numerical_analysis/numerical_integration.py
index 87184a76b740..f64436ec48c1 100644
--- a/maths/numerical_integration.py
+++ b/maths/numerical_analysis/numerical_integration.py
@@ -2,16 +2,17 @@
Approximates the area under the curve using the trapezoidal rule
"""
-from typing import Callable, Union
+from __future__ import annotations
+
+from collections.abc import Callable
def trapezoidal_area(
- fnc: Callable[[Union[int, float]], Union[int, float]],
- x_start: Union[int, float],
- x_end: Union[int, float],
+ fnc: Callable[[float], float],
+ x_start: float,
+ x_end: float,
steps: int = 100,
) -> float:
-
"""
Treats curve as a collection of linear lines and sums the area of the
trapezium shape they form
@@ -38,8 +39,7 @@ def trapezoidal_area(
fx1 = fnc(x_start)
area = 0.0
- for i in range(steps):
-
+ for _ in range(steps):
# Approximates small segments of curve as linear and solve
# for trapezoidal area
x2 = (x_end - x_start) / steps + x1
@@ -55,7 +55,7 @@ def trapezoidal_area(
if __name__ == "__main__":
def f(x):
- return x ** 3
+ return x**3
print("f(x) = x^3")
print("The area between the curve, x = -10, x = 10 and the x axis is:")
diff --git a/maths/numerical_analysis/proper_fractions.py b/maths/numerical_analysis/proper_fractions.py
new file mode 100644
index 000000000000..774ce9a24876
--- /dev/null
+++ b/maths/numerical_analysis/proper_fractions.py
@@ -0,0 +1,40 @@
+from math import gcd
+
+
+def proper_fractions(denominator: int) -> list[str]:
+ """
+ this algorithm returns a list of proper fractions, in the
+ range between 0 and 1, which can be formed with the given denominator
+ https://en.wikipedia.org/wiki/Fraction#Proper_and_improper_fractions
+
+ >>> proper_fractions(10)
+ ['1/10', '3/10', '7/10', '9/10']
+ >>> proper_fractions(5)
+ ['1/5', '2/5', '3/5', '4/5']
+ >>> proper_fractions(-15)
+ Traceback (most recent call last):
+ ...
+ ValueError: The Denominator Cannot be less than 0
+ >>> proper_fractions(0)
+ []
+ >>> proper_fractions(1.2)
+ Traceback (most recent call last):
+ ...
+ ValueError: The Denominator must be an integer
+ """
+
+ if denominator < 0:
+ raise ValueError("The Denominator Cannot be less than 0")
+ elif isinstance(denominator, float):
+ raise ValueError("The Denominator must be an integer")
+ return [
+ f"{numerator}/{denominator}"
+ for numerator in range(1, denominator)
+ if gcd(numerator, denominator) == 1
+ ]
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/maths/runge_kutta.py b/maths/numerical_analysis/runge_kutta.py
similarity index 89%
rename from maths/runge_kutta.py
rename to maths/numerical_analysis/runge_kutta.py
index 383797daa5ac..3a25b0fb0173 100644
--- a/maths/runge_kutta.py
+++ b/maths/numerical_analysis/runge_kutta.py
@@ -19,15 +19,15 @@ def runge_kutta(f, y0, x0, h, x_end):
... return y
>>> y0 = 1
>>> y = runge_kutta(f, y0, 0.0, 0.01, 5)
- >>> y[-1]
+ >>> float(y[-1])
148.41315904125113
"""
- N = int(np.ceil((x_end - x0) / h))
- y = np.zeros((N + 1,))
+ n = int(np.ceil((x_end - x0) / h))
+ y = np.zeros((n + 1,))
y[0] = y0
x = x0
- for k in range(N):
+ for k in range(n):
k1 = f(x, y[k])
k2 = f(x + 0.5 * h, y[k] + 0.5 * h * k1)
k3 = f(x + 0.5 * h, y[k] + 0.5 * h * k2)
diff --git a/maths/numerical_analysis/runge_kutta_fehlberg_45.py b/maths/numerical_analysis/runge_kutta_fehlberg_45.py
new file mode 100644
index 000000000000..0fbd60a35c1a
--- /dev/null
+++ b/maths/numerical_analysis/runge_kutta_fehlberg_45.py
@@ -0,0 +1,114 @@
+"""
+Use the Runge-Kutta-Fehlberg method to solve Ordinary Differential Equations.
+"""
+
+from collections.abc import Callable
+
+import numpy as np
+
+
+def runge_kutta_fehlberg_45(
+ func: Callable,
+ x_initial: float,
+ y_initial: float,
+ step_size: float,
+ x_final: float,
+) -> np.ndarray:
+ """
+ Solve an Ordinary Differential Equations using Runge-Kutta-Fehlberg Method (rkf45)
+ of order 5.
+
+ https://en.wikipedia.org/wiki/Runge%E2%80%93Kutta%E2%80%93Fehlberg_method
+
+ args:
+ func: An ordinary differential equation (ODE) as function of x and y.
+ x_initial: The initial value of x.
+ y_initial: The initial value of y.
+ step_size: The increment value of x.
+ x_final: The final value of x.
+
+ Returns:
+ Solution of y at each nodal point
+
+ # exact value of y[1] is tan(0.2) = 0.2027100937470787
+ >>> def f(x, y):
+ ... return 1 + y**2
+ >>> y = runge_kutta_fehlberg_45(f, 0, 0, 0.2, 1)
+ >>> float(y[1])
+ 0.2027100937470787
+ >>> def f(x,y):
+ ... return x
+ >>> y = runge_kutta_fehlberg_45(f, -1, 0, 0.2, 0)
+ >>> float(y[1])
+ -0.18000000000000002
+ >>> y = runge_kutta_fehlberg_45(5, 0, 0, 0.1, 1)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'int' object is not callable
+ >>> def f(x, y):
+ ... return x + y
+ >>> y = runge_kutta_fehlberg_45(f, 0, 0, 0.2, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The final value of x must be greater than initial value of x.
+ >>> def f(x, y):
+ ... return x
+ >>> y = runge_kutta_fehlberg_45(f, -1, 0, -0.2, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Step size must be positive.
+ """
+ if x_initial >= x_final:
+ raise ValueError(
+ "The final value of x must be greater than initial value of x."
+ )
+
+ if step_size <= 0:
+ raise ValueError("Step size must be positive.")
+
+ n = int((x_final - x_initial) / step_size)
+ y = np.zeros(
+ (n + 1),
+ )
+ x = np.zeros(n + 1)
+ y[0] = y_initial
+ x[0] = x_initial
+ for i in range(n):
+ k1 = step_size * func(x[i], y[i])
+ k2 = step_size * func(x[i] + step_size / 4, y[i] + k1 / 4)
+ k3 = step_size * func(
+ x[i] + (3 / 8) * step_size, y[i] + (3 / 32) * k1 + (9 / 32) * k2
+ )
+ k4 = step_size * func(
+ x[i] + (12 / 13) * step_size,
+ y[i] + (1932 / 2197) * k1 - (7200 / 2197) * k2 + (7296 / 2197) * k3,
+ )
+ k5 = step_size * func(
+ x[i] + step_size,
+ y[i] + (439 / 216) * k1 - 8 * k2 + (3680 / 513) * k3 - (845 / 4104) * k4,
+ )
+ k6 = step_size * func(
+ x[i] + step_size / 2,
+ y[i]
+ - (8 / 27) * k1
+ + 2 * k2
+ - (3544 / 2565) * k3
+ + (1859 / 4104) * k4
+ - (11 / 40) * k5,
+ )
+ y[i + 1] = (
+ y[i]
+ + (16 / 135) * k1
+ + (6656 / 12825) * k3
+ + (28561 / 56430) * k4
+ - (9 / 50) * k5
+ + (2 / 55) * k6
+ )
+ x[i + 1] = step_size + x[i]
+ return y
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/numerical_analysis/runge_kutta_gills.py b/maths/numerical_analysis/runge_kutta_gills.py
new file mode 100644
index 000000000000..5d9672679813
--- /dev/null
+++ b/maths/numerical_analysis/runge_kutta_gills.py
@@ -0,0 +1,90 @@
+"""
+Use the Runge-Kutta-Gill's method of order 4 to solve Ordinary Differential Equations.
+
+https://www.geeksforgeeks.org/gills-4th-order-method-to-solve-differential-equations/
+Author : Ravi Kumar
+"""
+
+from collections.abc import Callable
+from math import sqrt
+
+import numpy as np
+
+
+def runge_kutta_gills(
+ func: Callable[[float, float], float],
+ x_initial: float,
+ y_initial: float,
+ step_size: float,
+ x_final: float,
+) -> np.ndarray:
+ """
+ Solve an Ordinary Differential Equations using Runge-Kutta-Gills Method of order 4.
+
+ args:
+ func: An ordinary differential equation (ODE) as function of x and y.
+ x_initial: The initial value of x.
+ y_initial: The initial value of y.
+ step_size: The increment value of x.
+ x_final: The final value of x.
+
+ Returns:
+ Solution of y at each nodal point
+
+ >>> def f(x, y):
+ ... return (x-y)/2
+ >>> y = runge_kutta_gills(f, 0, 3, 0.2, 5)
+ >>> float(y[-1])
+ 3.4104259225717537
+
+ >>> def f(x,y):
+ ... return x
+ >>> y = runge_kutta_gills(f, -1, 0, 0.2, 0)
+ >>> y
+ array([ 0. , -0.18, -0.32, -0.42, -0.48, -0.5 ])
+
+ >>> def f(x, y):
+ ... return x + y
+ >>> y = runge_kutta_gills(f, 0, 0, 0.2, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The final value of x must be greater than initial value of x.
+
+ >>> def f(x, y):
+ ... return x
+ >>> y = runge_kutta_gills(f, -1, 0, -0.2, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Step size must be positive.
+ """
+ if x_initial >= x_final:
+ raise ValueError(
+ "The final value of x must be greater than initial value of x."
+ )
+
+ if step_size <= 0:
+ raise ValueError("Step size must be positive.")
+
+ n = int((x_final - x_initial) / step_size)
+ y = np.zeros(n + 1)
+ y[0] = y_initial
+ for i in range(n):
+ k1 = step_size * func(x_initial, y[i])
+ k2 = step_size * func(x_initial + step_size / 2, y[i] + k1 / 2)
+ k3 = step_size * func(
+ x_initial + step_size / 2,
+ y[i] + (-0.5 + 1 / sqrt(2)) * k1 + (1 - 1 / sqrt(2)) * k2,
+ )
+ k4 = step_size * func(
+ x_initial + step_size, y[i] - (1 / sqrt(2)) * k2 + (1 + 1 / sqrt(2)) * k3
+ )
+
+ y[i + 1] = y[i] + (k1 + (2 - sqrt(2)) * k2 + (2 + sqrt(2)) * k3 + k4) / 6
+ x_initial += step_size
+ return y
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/numerical_analysis/secant_method.py b/maths/numerical_analysis/secant_method.py
new file mode 100644
index 000000000000..9fff8222cdde
--- /dev/null
+++ b/maths/numerical_analysis/secant_method.py
@@ -0,0 +1,30 @@
+"""
+Implementing Secant method in Python
+Author: dimgrichr
+"""
+
+from math import exp
+
+
+def f(x: float) -> float:
+ """
+ >>> f(5)
+ 39.98652410600183
+ """
+ return 8 * x - 2 * exp(-x)
+
+
+def secant_method(lower_bound: float, upper_bound: float, repeats: int) -> float:
+ """
+ >>> secant_method(1, 3, 2)
+ 0.2139409276214589
+ """
+ x0 = lower_bound
+ x1 = upper_bound
+ for _ in range(repeats):
+ x0, x1 = x1, x1 - (f(x1) * (x1 - x0)) / (f(x1) - f(x0))
+ return x1
+
+
+if __name__ == "__main__":
+ print(f"Example: {secant_method(1, 3, 2)}")
diff --git a/maths/numerical_analysis/simpson_rule.py b/maths/numerical_analysis/simpson_rule.py
new file mode 100644
index 000000000000..e75fb557a2f5
--- /dev/null
+++ b/maths/numerical_analysis/simpson_rule.py
@@ -0,0 +1,86 @@
+"""
+Numerical integration or quadrature for a smooth function f with known values at x_i
+
+This method is the classical approach of summing 'Equally Spaced Abscissas'
+
+method 2:
+"Simpson Rule"
+
+"""
+
+
+def method_2(boundary: list[int], steps: int) -> float:
+ # "Simpson Rule"
+ # int(f) = delta_x/2 * (b-a)/3*(f1 + 4f2 + 2f_3 + ... + fn)
+ """
+ Calculate the definite integral of a function using Simpson's Rule.
+ :param boundary: A list containing the lower and upper bounds of integration.
+ :param steps: The number of steps or resolution for the integration.
+ :return: The approximate integral value.
+
+ >>> round(method_2([0, 2, 4], 10), 10)
+ 2.6666666667
+ >>> round(method_2([2, 0], 10), 10)
+ -0.2666666667
+ >>> round(method_2([-2, -1], 10), 10)
+ 2.172
+ >>> round(method_2([0, 1], 10), 10)
+ 0.3333333333
+ >>> round(method_2([0, 2], 10), 10)
+ 2.6666666667
+ >>> round(method_2([0, 2], 100), 10)
+ 2.5621226667
+ >>> round(method_2([0, 1], 1000), 10)
+ 0.3320026653
+ >>> round(method_2([0, 2], 0), 10)
+ Traceback (most recent call last):
+ ...
+ ZeroDivisionError: Number of steps must be greater than zero
+ >>> round(method_2([0, 2], -10), 10)
+ Traceback (most recent call last):
+ ...
+ ZeroDivisionError: Number of steps must be greater than zero
+ """
+ if steps <= 0:
+ raise ZeroDivisionError("Number of steps must be greater than zero")
+
+ h = (boundary[1] - boundary[0]) / steps
+ a = boundary[0]
+ b = boundary[1]
+ x_i = make_points(a, b, h)
+ y = 0.0
+ y += (h / 3.0) * f(a)
+ cnt = 2
+ for i in x_i:
+ y += (h / 3) * (4 - 2 * (cnt % 2)) * f(i)
+ cnt += 1
+ y += (h / 3.0) * f(b)
+ return y
+
+
+def make_points(a, b, h):
+ x = a + h
+ while x < (b - h):
+ yield x
+ x = x + h
+
+
+def f(x): # enter your function here
+ y = (x - 0) * (x - 0)
+ return y
+
+
+def main():
+ a = 0.0 # Lower bound of integration
+ b = 1.0 # Upper bound of integration
+ steps = 10.0 # number of steps or resolution
+ boundary = [a, b] # boundary of integration
+ y = method_2(boundary, steps)
+ print(f"y = {y}")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ main()
diff --git a/maths/square_root.py b/maths/numerical_analysis/square_root.py
similarity index 79%
rename from maths/square_root.py
rename to maths/numerical_analysis/square_root.py
index b324c723037c..4462ccb75261 100644
--- a/maths/square_root.py
+++ b/maths/numerical_analysis/square_root.py
@@ -19,14 +19,13 @@ def get_initial_point(a: float) -> float:
def square_root_iterative(
- a: float, max_iter: int = 9999, tolerance: float = 0.00000000000001
+ a: float, max_iter: int = 9999, tolerance: float = 1e-14
) -> float:
"""
- Square root is aproximated using Newtons method.
+ Square root approximated using Newton's method.
https://en.wikipedia.org/wiki/Newton%27s_method
- >>> all(abs(square_root_iterative(i)-math.sqrt(i)) <= .00000000000001
- ... for i in range(500))
+ >>> all(abs(square_root_iterative(i) - math.sqrt(i)) <= 1e-14 for i in range(500))
True
>>> square_root_iterative(-1)
@@ -49,7 +48,7 @@ def square_root_iterative(
value = get_initial_point(a)
- for i in range(max_iter):
+ for _ in range(max_iter):
prev_value = value
value = value - fx(value, a) / fx_derivative(value)
if abs(prev_value - value) < tolerance:
diff --git a/maths/odd_sieve.py b/maths/odd_sieve.py
new file mode 100644
index 000000000000..06605ca54296
--- /dev/null
+++ b/maths/odd_sieve.py
@@ -0,0 +1,42 @@
+from itertools import compress, repeat
+from math import ceil, sqrt
+
+
+def odd_sieve(num: int) -> list[int]:
+ """
+ Returns the prime numbers < `num`. The prime numbers are calculated using an
+ odd sieve implementation of the Sieve of Eratosthenes algorithm
+ (see for reference https://en.wikipedia.org/wiki/Sieve_of_Eratosthenes).
+
+ >>> odd_sieve(2)
+ []
+ >>> odd_sieve(3)
+ [2]
+ >>> odd_sieve(10)
+ [2, 3, 5, 7]
+ >>> odd_sieve(20)
+ [2, 3, 5, 7, 11, 13, 17, 19]
+ """
+
+ if num <= 2:
+ return []
+ if num == 3:
+ return [2]
+
+ # Odd sieve for numbers in range [3, num - 1]
+ sieve = bytearray(b"\x01") * ((num >> 1) - 1)
+
+ for i in range(3, int(sqrt(num)) + 1, 2):
+ if sieve[(i >> 1) - 1]:
+ i_squared = i**2
+ sieve[(i_squared >> 1) - 1 :: i] = repeat(
+ 0, ceil((num - i_squared) / (i << 1))
+ )
+
+ return [2, *list(compress(range(3, num, 2), sieve))]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/perfect_cube.py b/maths/perfect_cube.py
index 9ad287e41e75..a732b7cce6c8 100644
--- a/maths/perfect_cube.py
+++ b/maths/perfect_cube.py
@@ -11,6 +11,45 @@ def perfect_cube(n: int) -> bool:
return (val * val * val) == n
+def perfect_cube_binary_search(n: int) -> bool:
+ """
+ Check if a number is a perfect cube or not using binary search.
+ Time complexity : O(Log(n))
+ Space complexity: O(1)
+
+ >>> perfect_cube_binary_search(27)
+ True
+ >>> perfect_cube_binary_search(64)
+ True
+ >>> perfect_cube_binary_search(4)
+ False
+ >>> perfect_cube_binary_search("a")
+ Traceback (most recent call last):
+ ...
+ TypeError: perfect_cube_binary_search() only accepts integers
+ >>> perfect_cube_binary_search(0.1)
+ Traceback (most recent call last):
+ ...
+ TypeError: perfect_cube_binary_search() only accepts integers
+ """
+ if not isinstance(n, int):
+ raise TypeError("perfect_cube_binary_search() only accepts integers")
+ if n < 0:
+ n = -n
+ left = 0
+ right = n
+ while left <= right:
+ mid = left + (right - left) // 2
+ if mid * mid * mid == n:
+ return True
+ elif mid * mid * mid < n:
+ left = mid + 1
+ else:
+ right = mid - 1
+ return False
+
+
if __name__ == "__main__":
- print(perfect_cube(27))
- print(perfect_cube(4))
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/perfect_number.py b/maths/perfect_number.py
index 148e988fb4c5..52c816cc7895 100644
--- a/maths/perfect_number.py
+++ b/maths/perfect_number.py
@@ -14,21 +14,73 @@
def perfect(number: int) -> bool:
"""
+ Check if a number is a perfect number.
+
+ A perfect number is a positive integer that is equal to the sum of its proper
+ divisors (excluding itself).
+
+ Args:
+ number: The number to be checked.
+
+ Returns:
+ True if the number is a perfect number otherwise, False.
+ Start from 1 because dividing by 0 will raise ZeroDivisionError.
+ A number at most can be divisible by the half of the number except the number
+ itself. For example, 6 is at most can be divisible by 3 except by 6 itself.
+ Examples:
>>> perfect(27)
False
>>> perfect(28)
True
>>> perfect(29)
False
-
- Start from 1 because dividing by 0 will raise ZeroDivisionError.
- A number at most can be divisible by the half of the number except the number
- itself. For example, 6 is at most can be divisible by 3 except by 6 itself.
+ >>> perfect(6)
+ True
+ >>> perfect(12)
+ False
+ >>> perfect(496)
+ True
+ >>> perfect(8128)
+ True
+ >>> perfect(0)
+ False
+ >>> perfect(-1)
+ False
+ >>> perfect(33550336) # Large perfect number
+ True
+ >>> perfect(33550337) # Just above a large perfect number
+ False
+ >>> perfect(1) # Edge case: 1 is not a perfect number
+ False
+ >>> perfect("123") # String representation of a number
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be an integer
+ >>> perfect(12.34)
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be an integer
+ >>> perfect("Hello")
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be an integer
"""
+ if not isinstance(number, int):
+ raise ValueError("number must be an integer")
+ if number <= 0:
+ return False
return sum(i for i in range(1, number // 2 + 1) if number % i == 0) == number
if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
print("Program to check whether a number is a Perfect number or not...")
- number = int(input("Enter number: ").strip())
+ try:
+ number = int(input("Enter a positive integer: ").strip())
+ except ValueError:
+ msg = "number must be an integer"
+ raise ValueError(msg)
+
print(f"{number} is {'' if perfect(number) else 'not '}a Perfect Number.")
diff --git a/maths/perfect_square.py b/maths/perfect_square.py
index 4393dcfbc774..107e68528068 100644
--- a/maths/perfect_square.py
+++ b/maths/perfect_square.py
@@ -58,9 +58,9 @@ def perfect_square_binary_search(n: int) -> bool:
right = n
while left <= right:
mid = (left + right) // 2
- if mid ** 2 == n:
+ if mid**2 == n:
return True
- elif mid ** 2 > n:
+ elif mid**2 > n:
right = mid - 1
else:
left = mid + 1
diff --git a/maths/persistence.py b/maths/persistence.py
new file mode 100644
index 000000000000..c61a69a7c27d
--- /dev/null
+++ b/maths/persistence.py
@@ -0,0 +1,82 @@
+def multiplicative_persistence(num: int) -> int:
+ """
+ Return the persistence of a given number.
+
+ https://en.wikipedia.org/wiki/Persistence_of_a_number
+
+ >>> multiplicative_persistence(217)
+ 2
+ >>> multiplicative_persistence(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: multiplicative_persistence() does not accept negative values
+ >>> multiplicative_persistence("long number")
+ Traceback (most recent call last):
+ ...
+ ValueError: multiplicative_persistence() only accepts integral values
+ """
+
+ if not isinstance(num, int):
+ raise ValueError("multiplicative_persistence() only accepts integral values")
+ if num < 0:
+ raise ValueError("multiplicative_persistence() does not accept negative values")
+
+ steps = 0
+ num_string = str(num)
+
+ while len(num_string) != 1:
+ numbers = [int(i) for i in num_string]
+
+ total = 1
+ for i in range(len(numbers)):
+ total *= numbers[i]
+
+ num_string = str(total)
+
+ steps += 1
+ return steps
+
+
+def additive_persistence(num: int) -> int:
+ """
+ Return the persistence of a given number.
+
+ https://en.wikipedia.org/wiki/Persistence_of_a_number
+
+ >>> additive_persistence(199)
+ 3
+ >>> additive_persistence(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: additive_persistence() does not accept negative values
+ >>> additive_persistence("long number")
+ Traceback (most recent call last):
+ ...
+ ValueError: additive_persistence() only accepts integral values
+ """
+
+ if not isinstance(num, int):
+ raise ValueError("additive_persistence() only accepts integral values")
+ if num < 0:
+ raise ValueError("additive_persistence() does not accept negative values")
+
+ steps = 0
+ num_string = str(num)
+
+ while len(num_string) != 1:
+ numbers = [int(i) for i in num_string]
+
+ total = 0
+ for i in range(len(numbers)):
+ total += numbers[i]
+
+ num_string = str(total)
+
+ steps += 1
+ return steps
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/pi_generator.py b/maths/pi_generator.py
new file mode 100644
index 000000000000..97f2c540c1ce
--- /dev/null
+++ b/maths/pi_generator.py
@@ -0,0 +1,87 @@
+def calculate_pi(limit: int) -> str:
+ """
+ https://en.wikipedia.org/wiki/Leibniz_formula_for_%CF%80
+ Leibniz Formula for Pi
+
+ The Leibniz formula is the special case arctan(1) = pi / 4.
+ Leibniz's formula converges extremely slowly: it exhibits sublinear convergence.
+
+ Convergence (https://en.wikipedia.org/wiki/Leibniz_formula_for_%CF%80#Convergence)
+
+ We cannot try to prove against an interrupted, uncompleted generation.
+ https://en.wikipedia.org/wiki/Leibniz_formula_for_%CF%80#Unusual_behaviour
+ The errors can in fact be predicted, but those calculations also approach infinity
+ for accuracy.
+
+ Our output will be a string so that we can definitely store all digits.
+
+ >>> import math
+ >>> float(calculate_pi(15)) == math.pi
+ True
+
+ Since we cannot predict errors or interrupt any infinite alternating series
+ generation since they approach infinity, or interrupt any alternating series, we'll
+ need math.isclose()
+
+ >>> math.isclose(float(calculate_pi(50)), math.pi)
+ True
+ >>> math.isclose(float(calculate_pi(100)), math.pi)
+ True
+
+ Since math.pi contains only 16 digits, here are some tests with known values:
+
+ >>> calculate_pi(50)
+ '3.14159265358979323846264338327950288419716939937510'
+ >>> calculate_pi(80)
+ '3.14159265358979323846264338327950288419716939937510582097494459230781640628620899'
+ """
+ # Variables used for the iteration process
+ q = 1
+ r = 0
+ t = 1
+ k = 1
+ n = 3
+ m = 3
+
+ decimal = limit
+ counter = 0
+
+ result = ""
+
+ # We can't compare against anything if we make a generator,
+ # so we'll stick with plain return logic
+ while counter != decimal + 1:
+ if 4 * q + r - t < n * t:
+ result += str(n)
+ if counter == 0:
+ result += "."
+
+ if decimal == counter:
+ break
+
+ counter += 1
+ nr = 10 * (r - n * t)
+ n = ((10 * (3 * q + r)) // t) - 10 * n
+ q *= 10
+ r = nr
+ else:
+ nr = (2 * q + r) * m
+ nn = (q * (7 * k) + 2 + (r * m)) // (t * m)
+ q *= k
+ t *= m
+ m += 2
+ k += 1
+ n = nn
+ r = nr
+ return result
+
+
+def main() -> None:
+ print(f"{calculate_pi(50) = }")
+ import doctest
+
+ doctest.testmod()
+
+
+if __name__ == "__main__":
+ main()
diff --git a/maths/pi_monte_carlo_estimation.py b/maths/pi_monte_carlo_estimation.py
index 20b46dddc6e5..29b679907239 100644
--- a/maths/pi_monte_carlo_estimation.py
+++ b/maths/pi_monte_carlo_estimation.py
@@ -11,7 +11,7 @@ def is_in_unit_circle(self) -> bool:
True, if the point lies in the unit circle
False, otherwise
"""
- return (self.x ** 2 + self.y ** 2) <= 1
+ return (self.x**2 + self.y**2) <= 1
@classmethod
def random_unit_square(cls):
@@ -47,7 +47,7 @@ def estimate_pi(number_of_simulations: int) -> float:
raise ValueError("At least one simulation is necessary to estimate PI.")
number_in_unit_circle = 0
- for simulation_index in range(number_of_simulations):
+ for _ in range(number_of_simulations):
random_point = Point.random_unit_square()
if random_point.is_in_unit_circle():
diff --git a/maths/points_are_collinear_3d.py b/maths/points_are_collinear_3d.py
new file mode 100644
index 000000000000..c7adddda9494
--- /dev/null
+++ b/maths/points_are_collinear_3d.py
@@ -0,0 +1,126 @@
+"""
+Check if three points are collinear in 3D.
+
+In short, the idea is that we are able to create a triangle using three points,
+and the area of that triangle can determine if the three points are collinear or not.
+
+
+First, we create two vectors with the same initial point from the three points,
+then we will calculate the cross-product of them.
+
+The length of the cross vector is numerically equal to the area of a parallelogram.
+
+Finally, the area of the triangle is equal to half of the area of the parallelogram.
+
+Since we are only differentiating between zero and anything else,
+we can get rid of the square root when calculating the length of the vector,
+and also the division by two at the end.
+
+From a second perspective, if the two vectors are parallel and overlapping,
+we can't get a nonzero perpendicular vector,
+since there will be an infinite number of orthogonal vectors.
+
+To simplify the solution we will not calculate the length,
+but we will decide directly from the vector whether it is equal to (0, 0, 0) or not.
+
+
+Read More:
+ https://math.stackexchange.com/a/1951650
+"""
+
+Vector3d = tuple[float, float, float]
+Point3d = tuple[float, float, float]
+
+
+def create_vector(end_point1: Point3d, end_point2: Point3d) -> Vector3d:
+ """
+ Pass two points to get the vector from them in the form (x, y, z).
+
+ >>> create_vector((0, 0, 0), (1, 1, 1))
+ (1, 1, 1)
+ >>> create_vector((45, 70, 24), (47, 32, 1))
+ (2, -38, -23)
+ >>> create_vector((-14, -1, -8), (-7, 6, 4))
+ (7, 7, 12)
+ """
+ x = end_point2[0] - end_point1[0]
+ y = end_point2[1] - end_point1[1]
+ z = end_point2[2] - end_point1[2]
+ return (x, y, z)
+
+
+def get_3d_vectors_cross(ab: Vector3d, ac: Vector3d) -> Vector3d:
+ """
+ Get the cross of the two vectors AB and AC.
+
+ I used determinant of 2x2 to get the determinant of the 3x3 matrix in the process.
+
+ Read More:
+ https://en.wikipedia.org/wiki/Cross_product
+ https://en.wikipedia.org/wiki/Determinant
+
+ >>> get_3d_vectors_cross((3, 4, 7), (4, 9, 2))
+ (-55, 22, 11)
+ >>> get_3d_vectors_cross((1, 1, 1), (1, 1, 1))
+ (0, 0, 0)
+ >>> get_3d_vectors_cross((-4, 3, 0), (3, -9, -12))
+ (-36, -48, 27)
+ >>> get_3d_vectors_cross((17.67, 4.7, 6.78), (-9.5, 4.78, -19.33))
+ (-123.2594, 277.15110000000004, 129.11260000000001)
+ """
+ x = ab[1] * ac[2] - ab[2] * ac[1] # *i
+ y = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j
+ z = ab[0] * ac[1] - ab[1] * ac[0] # *k
+ return (x, y, z)
+
+
+def is_zero_vector(vector: Vector3d, accuracy: int) -> bool:
+ """
+ Check if vector is equal to (0, 0, 0) or not.
+
+ Since the algorithm is very accurate, we will never get a zero vector,
+ so we need to round the vector axis,
+ because we want a result that is either True or False.
+ In other applications, we can return a float that represents the collinearity ratio.
+
+ >>> is_zero_vector((0, 0, 0), accuracy=10)
+ True
+ >>> is_zero_vector((15, 74, 32), accuracy=10)
+ False
+ >>> is_zero_vector((-15, -74, -32), accuracy=10)
+ False
+ """
+ return tuple(round(x, accuracy) for x in vector) == (0, 0, 0)
+
+
+def are_collinear(a: Point3d, b: Point3d, c: Point3d, accuracy: int = 10) -> bool:
+ """
+ Check if three points are collinear or not.
+
+ 1- Create two vectors AB and AC.
+ 2- Get the cross vector of the two vectors.
+ 3- Calculate the length of the cross vector.
+ 4- If the length is zero then the points are collinear, else they are not.
+
+ The use of the accuracy parameter is explained in is_zero_vector docstring.
+
+ >>> are_collinear((4.802293498137402, 3.536233125455244, 0),
+ ... (-2.186788107953106, -9.24561398001649, 7.141509524846482),
+ ... (1.530169574640268, -2.447927606600034, 3.343487096469054))
+ True
+ >>> are_collinear((-6, -2, 6),
+ ... (6.200213806439997, -4.930157614926678, -4.482371908289856),
+ ... (-4.085171149525941, -2.459889509029438, 4.354787180795383))
+ True
+ >>> are_collinear((2.399001826862445, -2.452009976680793, 4.464656666157666),
+ ... (-3.682816335934376, 5.753788986533145, 9.490993909044244),
+ ... (1.962903518985307, 3.741415730125627, 7))
+ False
+ >>> are_collinear((1.875375340689544, -7.268426006071538, 7.358196269835993),
+ ... (-3.546599383667157, -4.630005261513976, 3.208784032924246),
+ ... (-2.564606140206386, 3.937845170672183, 7))
+ False
+ """
+ ab = create_vector(a, b)
+ ac = create_vector(a, c)
+ return is_zero_vector(get_3d_vectors_cross(ab, ac), accuracy)
diff --git a/maths/pollard_rho.py b/maths/pollard_rho.py
new file mode 100644
index 000000000000..e8bc89cef6c5
--- /dev/null
+++ b/maths/pollard_rho.py
@@ -0,0 +1,148 @@
+from __future__ import annotations
+
+from math import gcd
+
+
+def pollard_rho(
+ num: int,
+ seed: int = 2,
+ step: int = 1,
+ attempts: int = 3,
+) -> int | None:
+ """
+ Use Pollard's Rho algorithm to return a nontrivial factor of ``num``.
+ The returned factor may be composite and require further factorization.
+ If the algorithm will return None if it fails to find a factor within
+ the specified number of attempts or within the specified number of steps.
+ If ``num`` is prime, this algorithm is guaranteed to return None.
+ https://en.wikipedia.org/wiki/Pollard%27s_rho_algorithm
+
+ >>> pollard_rho(18446744073709551617)
+ 274177
+ >>> pollard_rho(97546105601219326301)
+ 9876543191
+ >>> pollard_rho(100)
+ 2
+ >>> pollard_rho(17)
+ >>> pollard_rho(17**3)
+ 17
+ >>> pollard_rho(17**3, attempts=1)
+ >>> pollard_rho(3*5*7)
+ 21
+ >>> pollard_rho(1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The input value cannot be less than 2
+ """
+ # A value less than 2 can cause an infinite loop in the algorithm.
+ if num < 2:
+ raise ValueError("The input value cannot be less than 2")
+
+ # Because of the relationship between ``f(f(x))`` and ``f(x)``, this
+ # algorithm struggles to find factors that are divisible by two.
+ # As a workaround, we specifically check for two and even inputs.
+ # See: https://math.stackexchange.com/a/2856214/165820
+ if num > 2 and num % 2 == 0:
+ return 2
+
+ # Pollard's Rho algorithm requires a function that returns pseudorandom
+ # values between 0 <= X < ``num``. It doesn't need to be random in the
+ # sense that the output value is cryptographically secure or difficult
+ # to calculate, it only needs to be random in the sense that all output
+ # values should be equally likely to appear.
+ # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num``
+ # However, the success of Pollard's algorithm isn't guaranteed and is
+ # determined in part by the initial seed and the chosen random function.
+ # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num``
+ # where ``C`` is a value that we can modify between each attempt.
+ def rand_fn(value: int, step: int, modulus: int) -> int:
+ """
+ Returns a pseudorandom value modulo ``modulus`` based on the
+ input ``value`` and attempt-specific ``step`` size.
+
+ >>> rand_fn(0, 0, 0)
+ Traceback (most recent call last):
+ ...
+ ZeroDivisionError: integer division or modulo by zero
+ >>> rand_fn(1, 2, 3)
+ 0
+ >>> rand_fn(0, 10, 7)
+ 3
+ >>> rand_fn(1234, 1, 17)
+ 16
+ """
+ return (pow(value, 2) + step) % modulus
+
+ for _ in range(attempts):
+ # These track the position within the cycle detection logic.
+ tortoise = seed
+ hare = seed
+
+ while True:
+ # At each iteration, the tortoise moves one step and the hare moves two.
+ tortoise = rand_fn(tortoise, step, num)
+ hare = rand_fn(hare, step, num)
+ hare = rand_fn(hare, step, num)
+
+ # At some point both the tortoise and the hare will enter a cycle whose
+ # length ``p`` is a divisor of ``num``. Once in that cycle, at some point
+ # the tortoise and hare will end up on the same value modulo ``p``.
+ # We can detect when this happens because the position difference between
+ # the tortoise and the hare will share a common divisor with ``num``.
+ divisor = gcd(hare - tortoise, num)
+
+ if divisor == 1:
+ # No common divisor yet, just keep searching.
+ continue
+ # We found a common divisor!
+ elif divisor == num:
+ # Unfortunately, the divisor is ``num`` itself and is useless.
+ break
+ else:
+ # The divisor is a nontrivial factor of ``num``!
+ return divisor
+
+ # If we made it here, then this attempt failed.
+ # We need to pick a new starting seed for the tortoise and hare
+ # in addition to a new step value for the random function.
+ # To keep this example implementation deterministic, the
+ # new values will be generated based on currently available
+ # values instead of using something like ``random.randint``.
+
+ # We can use the hare's position as the new seed.
+ # This is actually what Richard Brent's the "optimized" variant does.
+ seed = hare
+
+ # The new step value for the random function can just be incremented.
+ # At first the results will be similar to what the old function would
+ # have produced, but the value will quickly diverge after a bit.
+ step += 1
+
+ # We haven't found a divisor within the requested number of attempts.
+ # We were unlucky or ``num`` itself is actually prime.
+ return None
+
+
+if __name__ == "__main__":
+ import argparse
+
+ parser = argparse.ArgumentParser()
+ parser.add_argument(
+ "num",
+ type=int,
+ help="The value to find a divisor of",
+ )
+ parser.add_argument(
+ "--attempts",
+ type=int,
+ default=3,
+ help="The number of attempts before giving up",
+ )
+ args = parser.parse_args()
+
+ divisor = pollard_rho(args.num, attempts=args.attempts)
+ if divisor is None:
+ print(f"{args.num} is probably prime")
+ else:
+ quotient = args.num // divisor
+ print(f"{args.num} = {divisor} * {quotient}")
diff --git a/maths/polynomial_evaluation.py b/maths/polynomial_evaluation.py
index e929a2d02972..90a51f521e01 100644
--- a/maths/polynomial_evaluation.py
+++ b/maths/polynomial_evaluation.py
@@ -1,18 +1,18 @@
-from typing import Sequence
+from collections.abc import Sequence
def evaluate_poly(poly: Sequence[float], x: float) -> float:
"""Evaluate a polynomial f(x) at specified point x and return the value.
Arguments:
- poly -- the coeffiecients of a polynomial as an iterable in order of
+ poly -- the coefficients of a polynomial as an iterable in order of
ascending degree
x -- the point at which to evaluate the polynomial
>>> evaluate_poly((0.0, 0.0, 5.0, 9.3, 7.0), 10.0)
79800.0
"""
- return sum(c * (x ** i) for i, c in enumerate(poly))
+ return sum(c * (x**i) for i, c in enumerate(poly))
def horner(poly: Sequence[float], x: float) -> float:
@@ -26,7 +26,7 @@ def horner(poly: Sequence[float], x: float) -> float:
https://en.wikipedia.org/wiki/Horner's_method
Arguments:
- poly -- the coeffiecients of a polynomial as an iterable in order of
+ poly -- the coefficients of a polynomial as an iterable in order of
ascending degree
x -- the point at which to evaluate the polynomial
@@ -45,7 +45,7 @@ def horner(poly: Sequence[float], x: float) -> float:
>>> poly = (0.0, 0.0, 5.0, 9.3, 7.0) # f(x) = 7.0x^4 + 9.3x^3 + 5.0x^2
>>> x = -13.0
>>> # f(-13) = 7.0(-13)^4 + 9.3(-13)^3 + 5.0(-13)^2 = 180339.9
- >>> print(evaluate_poly(poly, x))
+ >>> evaluate_poly(poly, x)
180339.9
"""
poly = (0.0, 0.0, 5.0, 9.3, 7.0)
diff --git a/maths/polynomials/__init__.py b/maths/polynomials/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/maths/polynomials/single_indeterminate_operations.py b/maths/polynomials/single_indeterminate_operations.py
new file mode 100644
index 000000000000..e31e6caa3988
--- /dev/null
+++ b/maths/polynomials/single_indeterminate_operations.py
@@ -0,0 +1,188 @@
+"""
+
+This module implements a single indeterminate polynomials class
+with some basic operations
+
+Reference: https://en.wikipedia.org/wiki/Polynomial
+
+"""
+
+from __future__ import annotations
+
+from collections.abc import MutableSequence
+
+
+class Polynomial:
+ def __init__(self, degree: int, coefficients: MutableSequence[float]) -> None:
+ """
+ The coefficients should be in order of degree, from smallest to largest.
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> p = Polynomial(2, [1, 2, 3, 4])
+ Traceback (most recent call last):
+ ...
+ ValueError: The number of coefficients should be equal to the degree + 1.
+
+ """
+ if len(coefficients) != degree + 1:
+ raise ValueError(
+ "The number of coefficients should be equal to the degree + 1."
+ )
+
+ self.coefficients: list[float] = list(coefficients)
+ self.degree = degree
+
+ def __add__(self, polynomial_2: Polynomial) -> Polynomial:
+ """
+ Polynomial addition
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> q = Polynomial(2, [1, 2, 3])
+ >>> p + q
+ 6x^2 + 4x + 2
+ """
+
+ if self.degree > polynomial_2.degree:
+ coefficients = self.coefficients[:]
+ for i in range(polynomial_2.degree + 1):
+ coefficients[i] += polynomial_2.coefficients[i]
+ return Polynomial(self.degree, coefficients)
+ else:
+ coefficients = polynomial_2.coefficients[:]
+ for i in range(self.degree + 1):
+ coefficients[i] += self.coefficients[i]
+ return Polynomial(polynomial_2.degree, coefficients)
+
+ def __sub__(self, polynomial_2: Polynomial) -> Polynomial:
+ """
+ Polynomial subtraction
+ >>> p = Polynomial(2, [1, 2, 4])
+ >>> q = Polynomial(2, [1, 2, 3])
+ >>> p - q
+ 1x^2
+ """
+ return self + polynomial_2 * Polynomial(0, [-1])
+
+ def __neg__(self) -> Polynomial:
+ """
+ Polynomial negation
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> -p
+ - 3x^2 - 2x - 1
+ """
+ return Polynomial(self.degree, [-c for c in self.coefficients])
+
+ def __mul__(self, polynomial_2: Polynomial) -> Polynomial:
+ """
+ Polynomial multiplication
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> q = Polynomial(2, [1, 2, 3])
+ >>> p * q
+ 9x^4 + 12x^3 + 10x^2 + 4x + 1
+ """
+ coefficients: list[float] = [0] * (self.degree + polynomial_2.degree + 1)
+ for i in range(self.degree + 1):
+ for j in range(polynomial_2.degree + 1):
+ coefficients[i + j] += (
+ self.coefficients[i] * polynomial_2.coefficients[j]
+ )
+
+ return Polynomial(self.degree + polynomial_2.degree, coefficients)
+
+ def evaluate(self, substitution: float) -> float:
+ """
+ Evaluates the polynomial at x.
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> p.evaluate(2)
+ 17
+ """
+ result: int | float = 0
+ for i in range(self.degree + 1):
+ result += self.coefficients[i] * (substitution**i)
+ return result
+
+ def __str__(self) -> str:
+ """
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> print(p)
+ 3x^2 + 2x + 1
+ """
+ polynomial = ""
+ for i in range(self.degree, -1, -1):
+ if self.coefficients[i] == 0:
+ continue
+ elif self.coefficients[i] > 0:
+ if polynomial:
+ polynomial += " + "
+ else:
+ polynomial += " - "
+
+ if i == 0:
+ polynomial += str(abs(self.coefficients[i]))
+ elif i == 1:
+ polynomial += str(abs(self.coefficients[i])) + "x"
+ else:
+ polynomial += str(abs(self.coefficients[i])) + "x^" + str(i)
+
+ return polynomial
+
+ def __repr__(self) -> str:
+ """
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> p
+ 3x^2 + 2x + 1
+ """
+ return self.__str__()
+
+ def derivative(self) -> Polynomial:
+ """
+ Returns the derivative of the polynomial.
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> p.derivative()
+ 6x + 2
+ """
+ coefficients: list[float] = [0] * self.degree
+ for i in range(self.degree):
+ coefficients[i] = self.coefficients[i + 1] * (i + 1)
+ return Polynomial(self.degree - 1, coefficients)
+
+ def integral(self, constant: float = 0) -> Polynomial:
+ """
+ Returns the integral of the polynomial.
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> p.integral()
+ 1.0x^3 + 1.0x^2 + 1.0x
+ """
+ coefficients: list[float] = [0] * (self.degree + 2)
+ coefficients[0] = constant
+ for i in range(self.degree + 1):
+ coefficients[i + 1] = self.coefficients[i] / (i + 1)
+ return Polynomial(self.degree + 1, coefficients)
+
+ def __eq__(self, polynomial_2: object) -> bool:
+ """
+ Checks if two polynomials are equal.
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> q = Polynomial(2, [1, 2, 3])
+ >>> p == q
+ True
+ """
+ if not isinstance(polynomial_2, Polynomial):
+ return False
+
+ if self.degree != polynomial_2.degree:
+ return False
+
+ for i in range(self.degree + 1):
+ if self.coefficients[i] != polynomial_2.coefficients[i]:
+ return False
+
+ return True
+
+ def __ne__(self, polynomial_2: object) -> bool:
+ """
+ Checks if two polynomials are not equal.
+ >>> p = Polynomial(2, [1, 2, 3])
+ >>> q = Polynomial(2, [1, 2, 3])
+ >>> p != q
+ False
+ """
+ return not self.__eq__(polynomial_2)
diff --git a/maths/power_using_recursion.py b/maths/power_using_recursion.py
index f82097f6d8ec..eb775b161ae8 100644
--- a/maths/power_using_recursion.py
+++ b/maths/power_using_recursion.py
@@ -15,18 +15,45 @@
def power(base: int, exponent: int) -> float:
"""
- power(3, 4)
+ Calculate the power of a base raised to an exponent.
+
+ >>> power(3, 4)
81
>>> power(2, 0)
1
>>> all(power(base, exponent) == pow(base, exponent)
... for base in range(-10, 10) for exponent in range(10))
True
+ >>> power('a', 1)
+ 'a'
+ >>> power('a', 2)
+ Traceback (most recent call last):
+ ...
+ TypeError: can't multiply sequence by non-int of type 'str'
+ >>> power('a', 'b')
+ Traceback (most recent call last):
+ ...
+ TypeError: unsupported operand type(s) for -: 'str' and 'int'
+ >>> power(2, -1)
+ Traceback (most recent call last):
+ ...
+ RecursionError: maximum recursion depth exceeded
+ >>> power(0, 0)
+ 1
+ >>> power(0, 1)
+ 0
+ >>> power(5,6)
+ 15625
+ >>> power(23, 12)
+ 21914624432020321
"""
return base * power(base, (exponent - 1)) if exponent else 1
if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
print("Raise base to the power of exponent using recursion...")
base = int(input("Enter the base: ").strip())
exponent = int(input("Enter the exponent: ").strip())
diff --git a/maths/prime_check.py b/maths/prime_check.py
index e2bcb7b8f151..a757c4108f24 100644
--- a/maths/prime_check.py
+++ b/maths/prime_check.py
@@ -3,55 +3,87 @@
import math
import unittest
+import pytest
-def prime_check(number: int) -> bool:
- """Checks to see if a number is a prime.
+
+def is_prime(number: int) -> bool:
+ """Checks to see if a number is a prime in O(sqrt(n)).
A number is prime if it has exactly two factors: 1 and itself.
+
+ >>> is_prime(0)
+ False
+ >>> is_prime(1)
+ False
+ >>> is_prime(2)
+ True
+ >>> is_prime(3)
+ True
+ >>> is_prime(27)
+ False
+ >>> is_prime(87)
+ False
+ >>> is_prime(563)
+ True
+ >>> is_prime(2999)
+ True
+ >>> is_prime(67483)
+ False
+ >>> is_prime(16.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: is_prime() only accepts positive integers
+ >>> is_prime(-4)
+ Traceback (most recent call last):
+ ...
+ ValueError: is_prime() only accepts positive integers
"""
+ # precondition
+ if not isinstance(number, int) or not number >= 0:
+ raise ValueError("is_prime() only accepts positive integers")
+
if 1 < number < 4:
# 2 and 3 are primes
return True
- elif number < 2 or not number % 2:
- # Negatives, 0, 1 and all even numbers are not primes
+ elif number < 2 or number % 2 == 0 or number % 3 == 0:
+ # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes
return False
- odd_numbers = range(3, int(math.sqrt(number) + 1), 2)
- return not any(not number % i for i in odd_numbers)
+ # All primes number are in format of 6k +/- 1
+ for i in range(5, int(math.sqrt(number) + 1), 6):
+ if number % i == 0 or number % (i + 2) == 0:
+ return False
+ return True
class Test(unittest.TestCase):
def test_primes(self):
- self.assertTrue(prime_check(2))
- self.assertTrue(prime_check(3))
- self.assertTrue(prime_check(5))
- self.assertTrue(prime_check(7))
- self.assertTrue(prime_check(11))
- self.assertTrue(prime_check(13))
- self.assertTrue(prime_check(17))
- self.assertTrue(prime_check(19))
- self.assertTrue(prime_check(23))
- self.assertTrue(prime_check(29))
+ assert is_prime(2)
+ assert is_prime(3)
+ assert is_prime(5)
+ assert is_prime(7)
+ assert is_prime(11)
+ assert is_prime(13)
+ assert is_prime(17)
+ assert is_prime(19)
+ assert is_prime(23)
+ assert is_prime(29)
def test_not_primes(self):
- self.assertFalse(
- prime_check(-19),
- "Negative numbers are excluded by definition of prime numbers.",
- )
- self.assertFalse(
- prime_check(0),
- "Zero doesn't have any positive factors, primes must have exactly two.",
+ with pytest.raises(ValueError):
+ is_prime(-19)
+ assert not is_prime(0), (
+ "Zero doesn't have any positive factors, primes must have exactly two."
)
- self.assertFalse(
- prime_check(1),
- "One only has 1 positive factor, primes must have exactly two.",
+ assert not is_prime(1), (
+ "One only has 1 positive factor, primes must have exactly two."
)
- self.assertFalse(prime_check(2 * 2))
- self.assertFalse(prime_check(2 * 3))
- self.assertFalse(prime_check(3 * 3))
- self.assertFalse(prime_check(3 * 5))
- self.assertFalse(prime_check(3 * 5 * 7))
+ assert not is_prime(2 * 2)
+ assert not is_prime(2 * 3)
+ assert not is_prime(3 * 3)
+ assert not is_prime(3 * 5)
+ assert not is_prime(3 * 5 * 7)
if __name__ == "__main__":
diff --git a/maths/prime_factors.py b/maths/prime_factors.py
index e520ae3a6d04..47abcf10e618 100644
--- a/maths/prime_factors.py
+++ b/maths/prime_factors.py
@@ -1,6 +1,7 @@
"""
python/black : True
"""
+
from __future__ import annotations
diff --git a/maths/prime_numbers.py b/maths/prime_numbers.py
index 38bebddeee41..5ad12baf3dc3 100644
--- a/maths/prime_numbers.py
+++ b/maths/prime_numbers.py
@@ -1,8 +1,8 @@
import math
-from typing import Generator
+from collections.abc import Generator
-def slow_primes(max: int) -> Generator[int, None, None]:
+def slow_primes(max_n: int) -> Generator[int]:
"""
Return a list of all primes numbers up to max.
>>> list(slow_primes(0))
@@ -17,10 +17,10 @@ def slow_primes(max: int) -> Generator[int, None, None]:
[2, 3, 5, 7, 11]
>>> list(slow_primes(33))
[2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31]
- >>> list(slow_primes(10000))[-1]
- 9973
+ >>> list(slow_primes(1000))[-1]
+ 997
"""
- numbers: Generator = (i for i in range(1, (max + 1)))
+ numbers: Generator = (i for i in range(1, (max_n + 1)))
for i in (n for n in numbers if n > 1):
for j in range(2, i):
if (i % j) == 0:
@@ -29,7 +29,7 @@ def slow_primes(max: int) -> Generator[int, None, None]:
yield i
-def primes(max: int) -> Generator[int, None, None]:
+def primes(max_n: int) -> Generator[int]:
"""
Return a list of all primes numbers up to max.
>>> list(primes(0))
@@ -44,10 +44,10 @@ def primes(max: int) -> Generator[int, None, None]:
[2, 3, 5, 7, 11]
>>> list(primes(33))
[2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31]
- >>> list(primes(10000))[-1]
- 9973
+ >>> list(primes(1000))[-1]
+ 997
"""
- numbers: Generator = (i for i in range(1, (max + 1)))
+ numbers: Generator = (i for i in range(1, (max_n + 1)))
for i in (n for n in numbers if n > 1):
# only need to check for factors up to sqrt(i)
bound = int(math.sqrt(i)) + 1
@@ -58,13 +58,52 @@ def primes(max: int) -> Generator[int, None, None]:
yield i
+def fast_primes(max_n: int) -> Generator[int]:
+ """
+ Return a list of all primes numbers up to max.
+ >>> list(fast_primes(0))
+ []
+ >>> list(fast_primes(-1))
+ []
+ >>> list(fast_primes(-10))
+ []
+ >>> list(fast_primes(25))
+ [2, 3, 5, 7, 11, 13, 17, 19, 23]
+ >>> list(fast_primes(11))
+ [2, 3, 5, 7, 11]
+ >>> list(fast_primes(33))
+ [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31]
+ >>> list(fast_primes(1000))[-1]
+ 997
+ """
+ numbers: Generator = (i for i in range(1, (max_n + 1), 2))
+ # It's useless to test even numbers as they will not be prime
+ if max_n > 2:
+ yield 2 # Because 2 will not be tested, it's necessary to yield it now
+ for i in (n for n in numbers if n > 1):
+ bound = int(math.sqrt(i)) + 1
+ for j in range(3, bound, 2):
+ # As we removed the even numbers, we don't need them now
+ if (i % j) == 0:
+ break
+ else:
+ yield i
+
+
+def benchmark():
+ """
+ Let's benchmark our functions side-by-side...
+ """
+ from timeit import timeit
+
+ setup = "from __main__ import slow_primes, primes, fast_primes"
+ print(timeit("slow_primes(1_000_000_000_000)", setup=setup, number=1_000_000))
+ print(timeit("primes(1_000_000_000_000)", setup=setup, number=1_000_000))
+ print(timeit("fast_primes(1_000_000_000_000)", setup=setup, number=1_000_000))
+
+
if __name__ == "__main__":
number = int(input("Calculate primes up to:\n>> ").strip())
for ret in primes(number):
print(ret)
-
- # Let's benchmark them side-by-side...
- from timeit import timeit
-
- print(timeit("slow_primes(1_000_000)", setup="from __main__ import slow_primes"))
- print(timeit("primes(1_000_000)", setup="from __main__ import primes"))
+ benchmark()
diff --git a/maths/prime_sieve_eratosthenes.py b/maths/prime_sieve_eratosthenes.py
index 8d60e48c2140..32eef9165bba 100644
--- a/maths/prime_sieve_eratosthenes.py
+++ b/maths/prime_sieve_eratosthenes.py
@@ -1,12 +1,10 @@
-# flake8: noqa
-
"""
Sieve of Eratosthenes
-Input : n =10
+Input: n = 10
Output: 2 3 5 7
-Input : n = 20
+Input: n = 20
Output: 2 3 5 7 11 13 17 19
you can read in detail about this at
@@ -14,34 +12,43 @@
"""
-def prime_sieve_eratosthenes(num):
+def prime_sieve_eratosthenes(num: int) -> list[int]:
"""
- print the prime numbers up to n
+ Print the prime numbers up to n
>>> prime_sieve_eratosthenes(10)
- 2,3,5,7,
+ [2, 3, 5, 7]
>>> prime_sieve_eratosthenes(20)
- 2,3,5,7,11,13,17,19,
+ [2, 3, 5, 7, 11, 13, 17, 19]
+ >>> prime_sieve_eratosthenes(2)
+ [2]
+ >>> prime_sieve_eratosthenes(1)
+ []
+ >>> prime_sieve_eratosthenes(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive integer
"""
- primes = [True for i in range(num + 1)]
- p = 2
+ if num <= 0:
+ raise ValueError("Input must be a positive integer")
+ primes = [True] * (num + 1)
+
+ p = 2
while p * p <= num:
if primes[p]:
for i in range(p * p, num + 1, p):
primes[i] = False
p += 1
- for prime in range(2, num + 1):
- if primes[prime]:
- print(prime, end=",")
+ return [prime for prime in range(2, num + 1) if primes[prime]]
if __name__ == "__main__":
import doctest
doctest.testmod()
- num = int(input())
- prime_sieve_eratosthenes(num)
+ user_num = int(input("Enter a positive integer: ").strip())
+ print(prime_sieve_eratosthenes(user_num))
diff --git a/maths/primelib.py b/maths/primelib.py
new file mode 100644
index 000000000000..9f031efc50a9
--- /dev/null
+++ b/maths/primelib.py
@@ -0,0 +1,841 @@
+"""
+Created on Thu Oct 5 16:44:23 2017
+
+@author: Christian Bender
+
+This Python library contains some useful functions to deal with
+prime numbers and whole numbers.
+
+Overview:
+
+is_prime(number)
+sieve_er(N)
+get_prime_numbers(N)
+prime_factorization(number)
+greatest_prime_factor(number)
+smallest_prime_factor(number)
+get_prime(n)
+get_primes_between(pNumber1, pNumber2)
+
+----
+
+is_even(number)
+is_odd(number)
+kg_v(number1, number2) // least common multiple
+get_divisors(number) // all divisors of 'number' inclusive 1, number
+is_perfect_number(number)
+
+NEW-FUNCTIONS
+
+simplify_fraction(numerator, denominator)
+factorial (n) // n!
+fib (n) // calculate the n-th fibonacci term.
+
+-----
+
+goldbach(number) // Goldbach's assumption
+
+"""
+
+from math import sqrt
+
+from maths.greatest_common_divisor import gcd_by_iterative
+
+
+def is_prime(number: int) -> bool:
+ """
+ input: positive integer 'number'
+ returns true if 'number' is prime otherwise false.
+
+ >>> is_prime(3)
+ True
+ >>> is_prime(10)
+ False
+ >>> is_prime(97)
+ True
+ >>> is_prime(9991)
+ False
+ >>> is_prime(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and positive
+ >>> is_prime("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and positive
+ """
+
+ # precondition
+ assert isinstance(number, int) and (number >= 0), (
+ "'number' must been an int and positive"
+ )
+
+ status = True
+
+ # 0 and 1 are none primes.
+ if number <= 1:
+ status = False
+
+ for divisor in range(2, round(sqrt(number)) + 1):
+ # if 'number' divisible by 'divisor' then sets 'status'
+ # of false and break up the loop.
+ if number % divisor == 0:
+ status = False
+ break
+
+ # precondition
+ assert isinstance(status, bool), "'status' must been from type bool"
+
+ return status
+
+
+# ------------------------------------------
+
+
+def sieve_er(n):
+ """
+ input: positive integer 'N' > 2
+ returns a list of prime numbers from 2 up to N.
+
+ This function implements the algorithm called
+ sieve of erathostenes.
+
+ >>> sieve_er(8)
+ [2, 3, 5, 7]
+ >>> sieve_er(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'N' must been an int and > 2
+ >>> sieve_er("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'N' must been an int and > 2
+ """
+
+ # precondition
+ assert isinstance(n, int) and (n > 2), "'N' must been an int and > 2"
+
+ # beginList: contains all natural numbers from 2 up to N
+ begin_list = list(range(2, n + 1))
+
+ ans = [] # this list will be returns.
+
+ # actual sieve of erathostenes
+ for i in range(len(begin_list)):
+ for j in range(i + 1, len(begin_list)):
+ if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0):
+ begin_list[j] = 0
+
+ # filters actual prime numbers.
+ ans = [x for x in begin_list if x != 0]
+
+ # precondition
+ assert isinstance(ans, list), "'ans' must been from type list"
+
+ return ans
+
+
+# --------------------------------
+
+
+def get_prime_numbers(n):
+ """
+ input: positive integer 'N' > 2
+ returns a list of prime numbers from 2 up to N (inclusive)
+ This function is more efficient as function 'sieveEr(...)'
+
+ >>> get_prime_numbers(8)
+ [2, 3, 5, 7]
+ >>> get_prime_numbers(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'N' must been an int and > 2
+ >>> get_prime_numbers("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'N' must been an int and > 2
+ """
+
+ # precondition
+ assert isinstance(n, int) and (n > 2), "'N' must been an int and > 2"
+
+ ans = []
+
+ # iterates over all numbers between 2 up to N+1
+ # if a number is prime then appends to list 'ans'
+ for number in range(2, n + 1):
+ if is_prime(number):
+ ans.append(number)
+
+ # precondition
+ assert isinstance(ans, list), "'ans' must been from type list"
+
+ return ans
+
+
+# -----------------------------------------
+
+
+def prime_factorization(number):
+ """
+ input: positive integer 'number'
+ returns a list of the prime number factors of 'number'
+
+ >>> prime_factorization(0)
+ [0]
+ >>> prime_factorization(8)
+ [2, 2, 2]
+ >>> prime_factorization(287)
+ [7, 41]
+ >>> prime_factorization(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 0
+ >>> prime_factorization("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 0
+ """
+
+ # precondition
+ assert isinstance(number, int) and number >= 0, "'number' must been an int and >= 0"
+
+ ans = [] # this list will be returns of the function.
+
+ # potential prime number factors.
+
+ factor = 2
+
+ quotient = number
+
+ if number in {0, 1}:
+ ans.append(number)
+
+ # if 'number' not prime then builds the prime factorization of 'number'
+ elif not is_prime(number):
+ while quotient != 1:
+ if is_prime(factor) and (quotient % factor == 0):
+ ans.append(factor)
+ quotient /= factor
+ else:
+ factor += 1
+
+ else:
+ ans.append(number)
+
+ # precondition
+ assert isinstance(ans, list), "'ans' must been from type list"
+
+ return ans
+
+
+# -----------------------------------------
+
+
+def greatest_prime_factor(number):
+ """
+ input: positive integer 'number' >= 0
+ returns the greatest prime number factor of 'number'
+
+ >>> greatest_prime_factor(0)
+ 0
+ >>> greatest_prime_factor(8)
+ 2
+ >>> greatest_prime_factor(287)
+ 41
+ >>> greatest_prime_factor(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 0
+ >>> greatest_prime_factor("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 0
+ """
+
+ # precondition
+ assert isinstance(number, int) and (number >= 0), (
+ "'number' must been an int and >= 0"
+ )
+
+ ans = 0
+
+ # prime factorization of 'number'
+ prime_factors = prime_factorization(number)
+
+ ans = max(prime_factors)
+
+ # precondition
+ assert isinstance(ans, int), "'ans' must been from type int"
+
+ return ans
+
+
+# ----------------------------------------------
+
+
+def smallest_prime_factor(number):
+ """
+ input: integer 'number' >= 0
+ returns the smallest prime number factor of 'number'
+
+ >>> smallest_prime_factor(0)
+ 0
+ >>> smallest_prime_factor(8)
+ 2
+ >>> smallest_prime_factor(287)
+ 7
+ >>> smallest_prime_factor(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 0
+ >>> smallest_prime_factor("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 0
+ """
+
+ # precondition
+ assert isinstance(number, int) and (number >= 0), (
+ "'number' must been an int and >= 0"
+ )
+
+ ans = 0
+
+ # prime factorization of 'number'
+ prime_factors = prime_factorization(number)
+
+ ans = min(prime_factors)
+
+ # precondition
+ assert isinstance(ans, int), "'ans' must been from type int"
+
+ return ans
+
+
+# ----------------------
+
+
+def is_even(number):
+ """
+ input: integer 'number'
+ returns true if 'number' is even, otherwise false.
+
+ >>> is_even(0)
+ True
+ >>> is_even(8)
+ True
+ >>> is_even(287)
+ False
+ >>> is_even(-1)
+ False
+ >>> is_even("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int
+ """
+
+ # precondition
+ assert isinstance(number, int), "'number' must been an int"
+ assert isinstance(number % 2 == 0, bool), "compare must been from type bool"
+
+ return number % 2 == 0
+
+
+# ------------------------
+
+
+def is_odd(number):
+ """
+ input: integer 'number'
+ returns true if 'number' is odd, otherwise false.
+
+ >>> is_odd(0)
+ False
+ >>> is_odd(8)
+ False
+ >>> is_odd(287)
+ True
+ >>> is_odd(-1)
+ True
+ >>> is_odd("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int
+ """
+
+ # precondition
+ assert isinstance(number, int), "'number' must been an int"
+ assert isinstance(number % 2 != 0, bool), "compare must been from type bool"
+
+ return number % 2 != 0
+
+
+# ------------------------
+
+
+def goldbach(number):
+ """
+ Goldbach's assumption
+ input: a even positive integer 'number' > 2
+ returns a list of two prime numbers whose sum is equal to 'number'
+
+ >>> goldbach(8)
+ [3, 5]
+ >>> goldbach(824)
+ [3, 821]
+ >>> goldbach(0)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int, even and > 2
+ >>> goldbach(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int, even and > 2
+ >>> goldbach("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int, even and > 2
+ """
+
+ # precondition
+ assert isinstance(number, int) and (number > 2) and is_even(number), (
+ "'number' must been an int, even and > 2"
+ )
+
+ ans = [] # this list will returned
+
+ # creates a list of prime numbers between 2 up to 'number'
+ prime_numbers = get_prime_numbers(number)
+ len_pn = len(prime_numbers)
+
+ # run variable for while-loops.
+ i = 0
+ j = None
+
+ # exit variable. for break up the loops
+ loop = True
+
+ while i < len_pn and loop:
+ j = i + 1
+
+ while j < len_pn and loop:
+ if prime_numbers[i] + prime_numbers[j] == number:
+ loop = False
+ ans.append(prime_numbers[i])
+ ans.append(prime_numbers[j])
+
+ j += 1
+
+ i += 1
+
+ # precondition
+ assert (
+ isinstance(ans, list)
+ and (len(ans) == 2)
+ and (ans[0] + ans[1] == number)
+ and is_prime(ans[0])
+ and is_prime(ans[1])
+ ), "'ans' must contains two primes. And sum of elements must been eq 'number'"
+
+ return ans
+
+
+# ----------------------------------------------
+
+
+def kg_v(number1, number2):
+ """
+ Least common multiple
+ input: two positive integer 'number1' and 'number2'
+ returns the least common multiple of 'number1' and 'number2'
+
+ >>> kg_v(8,10)
+ 40
+ >>> kg_v(824,67)
+ 55208
+ >>> kg_v(1, 10)
+ 10
+ >>> kg_v(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: kg_v() missing 1 required positional argument: 'number2'
+ >>> kg_v(10,-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number1' and 'number2' must been positive integer.
+ >>> kg_v("test","test2")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number1' and 'number2' must been positive integer.
+ """
+
+ # precondition
+ assert (
+ isinstance(number1, int)
+ and isinstance(number2, int)
+ and (number1 >= 1)
+ and (number2 >= 1)
+ ), "'number1' and 'number2' must been positive integer."
+
+ ans = 1 # actual answer that will be return.
+
+ # for kgV (x,1)
+ if number1 > 1 and number2 > 1:
+ # builds the prime factorization of 'number1' and 'number2'
+ prime_fac_1 = prime_factorization(number1)
+ prime_fac_2 = prime_factorization(number2)
+
+ elif number1 == 1 or number2 == 1:
+ prime_fac_1 = []
+ prime_fac_2 = []
+ ans = max(number1, number2)
+
+ count1 = 0
+ count2 = 0
+
+ done = [] # captured numbers int both 'primeFac1' and 'primeFac2'
+
+ # iterates through primeFac1
+ for n in prime_fac_1:
+ if n not in done:
+ if n in prime_fac_2:
+ count1 = prime_fac_1.count(n)
+ count2 = prime_fac_2.count(n)
+
+ for _ in range(max(count1, count2)):
+ ans *= n
+
+ else:
+ count1 = prime_fac_1.count(n)
+
+ for _ in range(count1):
+ ans *= n
+
+ done.append(n)
+
+ # iterates through primeFac2
+ for n in prime_fac_2:
+ if n not in done:
+ count2 = prime_fac_2.count(n)
+
+ for _ in range(count2):
+ ans *= n
+
+ done.append(n)
+
+ # precondition
+ assert isinstance(ans, int) and (ans >= 0), (
+ "'ans' must been from type int and positive"
+ )
+
+ return ans
+
+
+# ----------------------------------
+
+
+def get_prime(n):
+ """
+ Gets the n-th prime number.
+ input: positive integer 'n' >= 0
+ returns the n-th prime number, beginning at index 0
+
+ >>> get_prime(0)
+ 2
+ >>> get_prime(8)
+ 23
+ >>> get_prime(824)
+ 6337
+ >>> get_prime(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been a positive int
+ >>> get_prime("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been a positive int
+ """
+
+ # precondition
+ assert isinstance(n, int) and (n >= 0), "'number' must been a positive int"
+
+ index = 0
+ ans = 2 # this variable holds the answer
+
+ while index < n:
+ index += 1
+
+ ans += 1 # counts to the next number
+
+ # if ans not prime then
+ # runs to the next prime number.
+ while not is_prime(ans):
+ ans += 1
+
+ # precondition
+ assert isinstance(ans, int) and is_prime(ans), (
+ "'ans' must been a prime number and from type int"
+ )
+
+ return ans
+
+
+# ---------------------------------------------------
+
+
+def get_primes_between(p_number_1, p_number_2):
+ """
+ input: prime numbers 'pNumber1' and 'pNumber2'
+ pNumber1 < pNumber2
+ returns a list of all prime numbers between 'pNumber1' (exclusive)
+ and 'pNumber2' (exclusive)
+
+ >>> get_primes_between(3, 67)
+ [5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61]
+ >>> get_primes_between(0)
+ Traceback (most recent call last):
+ ...
+ TypeError: get_primes_between() missing 1 required positional argument: 'p_number_2'
+ >>> get_primes_between(0, 1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: The arguments must been prime numbers and 'pNumber1' < 'pNumber2'
+ >>> get_primes_between(-1, 3)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and positive
+ >>> get_primes_between("test","test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and positive
+ """
+
+ # precondition
+ assert (
+ is_prime(p_number_1) and is_prime(p_number_2) and (p_number_1 < p_number_2)
+ ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'"
+
+ number = p_number_1 + 1 # jump to the next number
+
+ ans = [] # this list will be returns.
+
+ # if number is not prime then
+ # fetch the next prime number.
+ while not is_prime(number):
+ number += 1
+
+ while number < p_number_2:
+ ans.append(number)
+
+ number += 1
+
+ # fetch the next prime number.
+ while not is_prime(number):
+ number += 1
+
+ # precondition
+ assert (
+ isinstance(ans, list)
+ and ans[0] != p_number_1
+ and ans[len(ans) - 1] != p_number_2
+ ), "'ans' must been a list without the arguments"
+
+ # 'ans' contains not 'pNumber1' and 'pNumber2' !
+ return ans
+
+
+# ----------------------------------------------------
+
+
+def get_divisors(n):
+ """
+ input: positive integer 'n' >= 1
+ returns all divisors of n (inclusive 1 and 'n')
+
+ >>> get_divisors(8)
+ [1, 2, 4, 8]
+ >>> get_divisors(824)
+ [1, 2, 4, 8, 103, 206, 412, 824]
+ >>> get_divisors(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'n' must been int and >= 1
+ >>> get_divisors("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'n' must been int and >= 1
+ """
+
+ # precondition
+ assert isinstance(n, int) and (n >= 1), "'n' must been int and >= 1"
+
+ ans = [] # will be returned.
+
+ for divisor in range(1, n + 1):
+ if n % divisor == 0:
+ ans.append(divisor)
+
+ # precondition
+ assert ans[0] == 1 and ans[len(ans) - 1] == n, "Error in function getDivisiors(...)"
+
+ return ans
+
+
+# ----------------------------------------------------
+
+
+def is_perfect_number(number):
+ """
+ input: positive integer 'number' > 1
+ returns true if 'number' is a perfect number otherwise false.
+
+ >>> is_perfect_number(28)
+ True
+ >>> is_perfect_number(824)
+ False
+ >>> is_perfect_number(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 1
+ >>> is_perfect_number("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'number' must been an int and >= 1
+ """
+
+ # precondition
+ assert isinstance(number, int) and (number > 1), (
+ "'number' must been an int and >= 1"
+ )
+
+ divisors = get_divisors(number)
+
+ # precondition
+ assert (
+ isinstance(divisors, list)
+ and (divisors[0] == 1)
+ and (divisors[len(divisors) - 1] == number)
+ ), "Error in help-function getDivisiors(...)"
+
+ # summed all divisors up to 'number' (exclusive), hence [:-1]
+ return sum(divisors[:-1]) == number
+
+
+# ------------------------------------------------------------
+
+
+def simplify_fraction(numerator, denominator):
+ """
+ input: two integer 'numerator' and 'denominator'
+ assumes: 'denominator' != 0
+ returns: a tuple with simplify numerator and denominator.
+
+ >>> simplify_fraction(10, 20)
+ (1, 2)
+ >>> simplify_fraction(10, -1)
+ (10, -1)
+ >>> simplify_fraction("test","test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: The arguments must been from type int and 'denominator' != 0
+ """
+
+ # precondition
+ assert (
+ isinstance(numerator, int)
+ and isinstance(denominator, int)
+ and (denominator != 0)
+ ), "The arguments must been from type int and 'denominator' != 0"
+
+ # build the greatest common divisor of numerator and denominator.
+ gcd_of_fraction = gcd_by_iterative(abs(numerator), abs(denominator))
+
+ # precondition
+ assert (
+ isinstance(gcd_of_fraction, int)
+ and (numerator % gcd_of_fraction == 0)
+ and (denominator % gcd_of_fraction == 0)
+ ), "Error in function gcd_by_iterative(...,...)"
+
+ return (numerator // gcd_of_fraction, denominator // gcd_of_fraction)
+
+
+# -----------------------------------------------------------------
+
+
+def factorial(n):
+ """
+ input: positive integer 'n'
+ returns the factorial of 'n' (n!)
+
+ >>> factorial(0)
+ 1
+ >>> factorial(20)
+ 2432902008176640000
+ >>> factorial(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'n' must been a int and >= 0
+ >>> factorial("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'n' must been a int and >= 0
+ """
+
+ # precondition
+ assert isinstance(n, int) and (n >= 0), "'n' must been a int and >= 0"
+
+ ans = 1 # this will be return.
+
+ for factor in range(1, n + 1):
+ ans *= factor
+
+ return ans
+
+
+# -------------------------------------------------------------------
+
+
+def fib(n: int) -> int:
+ """
+ input: positive integer 'n'
+ returns the n-th fibonacci term , indexing by 0
+
+ >>> fib(0)
+ 1
+ >>> fib(5)
+ 8
+ >>> fib(20)
+ 10946
+ >>> fib(99)
+ 354224848179261915075
+ >>> fib(-1)
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'n' must been an int and >= 0
+ >>> fib("test")
+ Traceback (most recent call last):
+ ...
+ AssertionError: 'n' must been an int and >= 0
+ """
+
+ # precondition
+ assert isinstance(n, int) and (n >= 0), "'n' must been an int and >= 0"
+
+ tmp = 0
+ fib1 = 1
+ ans = 1 # this will be return
+
+ for _ in range(n - 1):
+ tmp = ans
+ ans += fib1
+ fib1 = tmp
+
+ return ans
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/print_multiplication_table.py b/maths/print_multiplication_table.py
new file mode 100644
index 000000000000..dbe4a4be0ee8
--- /dev/null
+++ b/maths/print_multiplication_table.py
@@ -0,0 +1,26 @@
+def multiplication_table(number: int, number_of_terms: int) -> str:
+ """
+ Prints the multiplication table of a given number till the given number of terms
+
+ >>> print(multiplication_table(3, 5))
+ 3 * 1 = 3
+ 3 * 2 = 6
+ 3 * 3 = 9
+ 3 * 4 = 12
+ 3 * 5 = 15
+
+ >>> print(multiplication_table(-4, 6))
+ -4 * 1 = -4
+ -4 * 2 = -8
+ -4 * 3 = -12
+ -4 * 4 = -16
+ -4 * 5 = -20
+ -4 * 6 = -24
+ """
+ return "\n".join(
+ f"{number} * {i} = {number * i}" for i in range(1, number_of_terms + 1)
+ )
+
+
+if __name__ == "__main__":
+ print(multiplication_table(number=5, number_of_terms=10))
diff --git a/maths/pythagoras.py b/maths/pythagoras.py
index 69a17731a0fd..7770e981d44d 100644
--- a/maths/pythagoras.py
+++ b/maths/pythagoras.py
@@ -14,17 +14,13 @@ def __repr__(self) -> str:
def distance(a: Point, b: Point) -> float:
- return math.sqrt(abs((b.x - a.x) ** 2 + (b.y - a.y) ** 2 + (b.z - a.z) ** 2))
-
-
-def test_distance() -> None:
"""
>>> point1 = Point(2, -1, 7)
>>> point2 = Point(1, -3, 5)
>>> print(f"Distance from {point1} to {point2} is {distance(point1, point2)}")
Distance from Point(2, -1, 7) to Point(1, -3, 5) is 3.0
"""
- pass
+ return math.sqrt(abs((b.x - a.x) ** 2 + (b.y - a.y) ** 2 + (b.z - a.z) ** 2))
if __name__ == "__main__":
diff --git a/maths/qr_decomposition.py b/maths/qr_decomposition.py
index 5e15fede4f2a..670b49206aa7 100644
--- a/maths/qr_decomposition.py
+++ b/maths/qr_decomposition.py
@@ -1,7 +1,7 @@
import numpy as np
-def qr_householder(A):
+def qr_householder(a: np.ndarray):
"""Return a QR-decomposition of the matrix A using Householder reflection.
The QR-decomposition decomposes the matrix A of shape (m, n) into an
@@ -37,14 +37,14 @@ def qr_householder(A):
>>> np.allclose(np.triu(R), R)
True
"""
- m, n = A.shape
+ m, n = a.shape
t = min(m, n)
- Q = np.eye(m)
- R = A.copy()
+ q = np.eye(m)
+ r = a.copy()
for k in range(t - 1):
# select a column of modified matrix A':
- x = R[k:, [k]]
+ x = r[k:, [k]]
# construct first basis vector
e1 = np.zeros_like(x)
e1[0] = 1.0
@@ -55,14 +55,14 @@ def qr_householder(A):
v /= np.linalg.norm(v)
# construct the Householder matrix
- Q_k = np.eye(m - k) - 2.0 * v @ v.T
+ q_k = np.eye(m - k) - 2.0 * v @ v.T
# pad with ones and zeros as necessary
- Q_k = np.block([[np.eye(k), np.zeros((k, m - k))], [np.zeros((m - k, k)), Q_k]])
+ q_k = np.block([[np.eye(k), np.zeros((k, m - k))], [np.zeros((m - k, k)), q_k]])
- Q = Q @ Q_k.T
- R = Q_k @ R
+ q = q @ q_k.T
+ r = q_k @ r
- return Q, R
+ return q, r
if __name__ == "__main__":
diff --git a/maths/quadratic_equations_complex_numbers.py b/maths/quadratic_equations_complex_numbers.py
index 01a411bc560d..1035171e4ec3 100644
--- a/maths/quadratic_equations_complex_numbers.py
+++ b/maths/quadratic_equations_complex_numbers.py
@@ -30,8 +30,8 @@ def quadratic_roots(a: int, b: int, c: int) -> tuple[complex, complex]:
def main():
- solutions = quadratic_roots(a=5, b=6, c=1)
- print("The solutions are: {} and {}".format(*solutions))
+ solution1, solution2 = quadratic_roots(a=5, b=6, c=1)
+ print(f"The solutions are: {solution1} and {solution2}")
if __name__ == "__main__":
diff --git a/maths/radians.py b/maths/radians.py
index 465467a3ba08..b8ac61cb135c 100644
--- a/maths/radians.py
+++ b/maths/radians.py
@@ -3,7 +3,7 @@
def radians(degree: float) -> float:
"""
- Coverts the given angle from degrees to radians
+ Converts the given angle from degrees to radians
https://en.wikipedia.org/wiki/Radian
>>> radians(180)
@@ -16,7 +16,7 @@ def radians(degree: float) -> float:
1.9167205845401725
>>> from math import radians as math_radians
- >>> all(abs(radians(i)-math_radians(i)) <= 0.00000001 for i in range(-2, 361))
+ >>> all(abs(radians(i) - math_radians(i)) <= 1e-8 for i in range(-2, 361))
True
"""
diff --git a/maths/radix2_fft.py b/maths/radix2_fft.py
index de87071e5440..d41dc82d5588 100644
--- a/maths/radix2_fft.py
+++ b/maths/radix2_fft.py
@@ -39,7 +39,7 @@ class FFT:
>>> x = FFT(A, B)
Print product
- >>> print(x.product) # 2x + 3x^2 + 8x^3 + 4x^4 + 6x^5
+ >>> x.product # 2x + 3x^2 + 8x^3 + 4x^4 + 6x^5
[(-0+0j), (2+0j), (3+0j), (8+0j), (6+0j), (8+0j)]
__str__ test
@@ -49,10 +49,10 @@ class FFT:
A*B = 0*x^(-0+0j) + 1*x^(2+0j) + 2*x^(3+0j) + 3*x^(8+0j) + 4*x^(6+0j) + 5*x^(8+0j)
"""
- def __init__(self, polyA=[0], polyB=[0]):
+ def __init__(self, poly_a=None, poly_b=None):
# Input as list
- self.polyA = list(polyA)[:]
- self.polyB = list(polyB)[:]
+ self.polyA = list(poly_a or [0])[:]
+ self.polyB = list(poly_b or [0])[:]
# Remove leading zero coefficients
while self.polyA[-1] == 0:
@@ -64,44 +64,40 @@ def __init__(self, polyA=[0], polyB=[0]):
self.len_B = len(self.polyB)
# Add 0 to make lengths equal a power of 2
- self.C_max_length = int(
+ self.c_max_length = int(
2 ** np.ceil(np.log2(len(self.polyA) + len(self.polyB) - 1))
)
- while len(self.polyA) < self.C_max_length:
+ while len(self.polyA) < self.c_max_length:
self.polyA.append(0)
- while len(self.polyB) < self.C_max_length:
+ while len(self.polyB) < self.c_max_length:
self.polyB.append(0)
# A complex root used for the fourier transform
- self.root = complex(mpmath.root(x=1, n=self.C_max_length, k=1))
+ self.root = complex(mpmath.root(x=1, n=self.c_max_length, k=1))
# The product
self.product = self.__multiply()
# Discrete fourier transform of A and B
- def __DFT(self, which):
- if which == "A":
- dft = [[x] for x in self.polyA]
- else:
- dft = [[x] for x in self.polyB]
+ def __dft(self, which):
+ dft = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB]
# Corner case
if len(dft) <= 1:
return dft[0]
- #
- next_ncol = self.C_max_length // 2
+ next_ncol = self.c_max_length // 2
while next_ncol > 0:
new_dft = [[] for i in range(next_ncol)]
- root = self.root ** next_ncol
+ root = self.root**next_ncol
# First half of next step
current_root = 1
- for j in range(self.C_max_length // (next_ncol * 2)):
+ for j in range(self.c_max_length // (next_ncol * 2)):
for i in range(next_ncol):
new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j])
current_root *= root
# Second half of next step
current_root = 1
- for j in range(self.C_max_length // (next_ncol * 2)):
+ for j in range(self.c_max_length // (next_ncol * 2)):
for i in range(next_ncol):
new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j])
current_root *= root
@@ -112,65 +108,65 @@ def __DFT(self, which):
# multiply the DFTs of A and B and find A*B
def __multiply(self):
- dftA = self.__DFT("A")
- dftB = self.__DFT("B")
- inverseC = [[dftA[i] * dftB[i] for i in range(self.C_max_length)]]
- del dftA
- del dftB
+ dft_a = self.__dft("A")
+ dft_b = self.__dft("B")
+ inverce_c = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length)]]
+ del dft_a
+ del dft_b
# Corner Case
- if len(inverseC[0]) <= 1:
- return inverseC[0]
+ if len(inverce_c[0]) <= 1:
+ return inverce_c[0]
# Inverse DFT
next_ncol = 2
- while next_ncol <= self.C_max_length:
- new_inverseC = [[] for i in range(next_ncol)]
+ while next_ncol <= self.c_max_length:
+ new_inverse_c = [[] for i in range(next_ncol)]
root = self.root ** (next_ncol // 2)
current_root = 1
# First half of next step
- for j in range(self.C_max_length // next_ncol):
+ for j in range(self.c_max_length // next_ncol):
for i in range(next_ncol // 2):
# Even positions
- new_inverseC[i].append(
+ new_inverse_c[i].append(
(
- inverseC[i][j]
- + inverseC[i][j + self.C_max_length // next_ncol]
+ inverce_c[i][j]
+ + inverce_c[i][j + self.c_max_length // next_ncol]
)
/ 2
)
# Odd positions
- new_inverseC[i + next_ncol // 2].append(
+ new_inverse_c[i + next_ncol // 2].append(
(
- inverseC[i][j]
- - inverseC[i][j + self.C_max_length // next_ncol]
+ inverce_c[i][j]
+ - inverce_c[i][j + self.c_max_length // next_ncol]
)
/ (2 * current_root)
)
current_root *= root
# Update
- inverseC = new_inverseC
+ inverce_c = new_inverse_c
next_ncol *= 2
# Unpack
- inverseC = [round(x[0].real, 8) + round(x[0].imag, 8) * 1j for x in inverseC]
+ inverce_c = [round(x[0].real, 8) + round(x[0].imag, 8) * 1j for x in inverce_c]
# Remove leading 0's
- while inverseC[-1] == 0:
- inverseC.pop()
- return inverseC
+ while inverce_c[-1] == 0:
+ inverce_c.pop()
+ return inverce_c
# Overwrite __str__ for print(); Shows A, B and A*B
def __str__(self):
- A = "A = " + " + ".join(
+ a = "A = " + " + ".join(
f"{coef}*x^{i}" for coef, i in enumerate(self.polyA[: self.len_A])
)
- B = "B = " + " + ".join(
+ b = "B = " + " + ".join(
f"{coef}*x^{i}" for coef, i in enumerate(self.polyB[: self.len_B])
)
- C = "A*B = " + " + ".join(
+ c = "A*B = " + " + ".join(
f"{coef}*x^{i}" for coef, i in enumerate(self.product)
)
- return "\n".join((A, B, C))
+ return f"{a}\n{b}\n{c}"
# Unit tests
diff --git a/maths/remove_digit.py b/maths/remove_digit.py
new file mode 100644
index 000000000000..db14ac902a6f
--- /dev/null
+++ b/maths/remove_digit.py
@@ -0,0 +1,37 @@
+def remove_digit(num: int) -> int:
+ """
+
+ returns the biggest possible result
+ that can be achieved by removing
+ one digit from the given number
+
+ >>> remove_digit(152)
+ 52
+ >>> remove_digit(6385)
+ 685
+ >>> remove_digit(-11)
+ 1
+ >>> remove_digit(2222222)
+ 222222
+ >>> remove_digit("2222222")
+ Traceback (most recent call last):
+ TypeError: only integers accepted as input
+ >>> remove_digit("string input")
+ Traceback (most recent call last):
+ TypeError: only integers accepted as input
+ """
+
+ if not isinstance(num, int):
+ raise TypeError("only integers accepted as input")
+ else:
+ num_str = str(abs(num))
+ num_transpositions = [list(num_str) for char in range(len(num_str))]
+ for index in range(len(num_str)):
+ num_transpositions[index].pop(index)
+ return max(
+ int("".join(list(transposition))) for transposition in num_transpositions
+ )
+
+
+if __name__ == "__main__":
+ __import__("doctest").testmod()
diff --git a/maths/segmented_sieve.py b/maths/segmented_sieve.py
index c1cc497ad33e..125390edc588 100644
--- a/maths/segmented_sieve.py
+++ b/maths/segmented_sieve.py
@@ -3,8 +3,37 @@
import math
-def sieve(n):
- """Segmented Sieve."""
+def sieve(n: int) -> list[int]:
+ """
+ Segmented Sieve.
+
+ Examples:
+ >>> sieve(8)
+ [2, 3, 5, 7]
+
+ >>> sieve(27)
+ [2, 3, 5, 7, 11, 13, 17, 19, 23]
+
+ >>> sieve(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Number 0 must instead be a positive integer
+
+ >>> sieve(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Number -1 must instead be a positive integer
+
+ >>> sieve(22.2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Number 22.2 must instead be a positive integer
+ """
+
+ if n <= 0 or isinstance(n, float):
+ msg = f"Number {n} must instead be a positive integer"
+ raise ValueError(msg)
+
in_prime = []
start = 2
end = int(math.sqrt(n)) # Size of every segment
@@ -15,20 +44,16 @@ def sieve(n):
if temp[start] is True:
in_prime.append(start)
for i in range(start * start, end + 1, start):
- if temp[i] is True:
- temp[i] = False
+ temp[i] = False
start += 1
prime += in_prime
low = end + 1
- high = low + end - 1
- if high > n:
- high = n
+ high = min(2 * end, n)
while low <= n:
temp = [True] * (high - low + 1)
for each in in_prime:
-
t = math.floor(low / each) * each
if t < low:
t += each
@@ -41,11 +66,14 @@ def sieve(n):
prime.append(j + low)
low = high + 1
- high = low + end - 1
- if high > n:
- high = n
+ high = min(high + end, n)
return prime
-print(sieve(10 ** 6))
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ print(f"{sieve(10**6) = }")
diff --git a/maths/series/arithmetic.py b/maths/series/arithmetic.py
new file mode 100644
index 000000000000..dc28c5c7bc5f
--- /dev/null
+++ b/maths/series/arithmetic.py
@@ -0,0 +1,77 @@
+"""
+Arithmetic mean
+Reference: https://en.wikipedia.org/wiki/Arithmetic_mean
+
+Arithmetic series
+Reference: https://en.wikipedia.org/wiki/Arithmetic_series
+(The URL above will redirect you to arithmetic progression)
+"""
+
+
+def is_arithmetic_series(series: list) -> bool:
+ """
+ checking whether the input series is arithmetic series or not
+ >>> is_arithmetic_series([2, 4, 6])
+ True
+ >>> is_arithmetic_series([3, 6, 12, 24])
+ False
+ >>> is_arithmetic_series([1, 2, 3])
+ True
+ >>> is_arithmetic_series(4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series is not valid, valid series - [2, 4, 6]
+ >>> is_arithmetic_series([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input list must be a non empty list
+ """
+ if not isinstance(series, list):
+ raise ValueError("Input series is not valid, valid series - [2, 4, 6]")
+ if len(series) == 0:
+ raise ValueError("Input list must be a non empty list")
+ if len(series) == 1:
+ return True
+ common_diff = series[1] - series[0]
+ for index in range(len(series) - 1):
+ if series[index + 1] - series[index] != common_diff:
+ return False
+ return True
+
+
+def arithmetic_mean(series: list) -> float:
+ """
+ return the arithmetic mean of series
+
+ >>> arithmetic_mean([2, 4, 6])
+ 4.0
+ >>> arithmetic_mean([3, 6, 9, 12])
+ 7.5
+ >>> arithmetic_mean(4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series is not valid, valid series - [2, 4, 6]
+ >>> arithmetic_mean([4, 8, 1])
+ 4.333333333333333
+ >>> arithmetic_mean([1, 2, 3])
+ 2.0
+ >>> arithmetic_mean([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input list must be a non empty list
+
+ """
+ if not isinstance(series, list):
+ raise ValueError("Input series is not valid, valid series - [2, 4, 6]")
+ if len(series) == 0:
+ raise ValueError("Input list must be a non empty list")
+ answer = 0
+ for val in series:
+ answer += val
+ return answer / len(series)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/series/geometric.py b/maths/series/geometric.py
new file mode 100644
index 000000000000..7b6239b1585d
--- /dev/null
+++ b/maths/series/geometric.py
@@ -0,0 +1,83 @@
+"""
+Geometric Mean
+Reference : https://en.wikipedia.org/wiki/Geometric_mean
+
+Geometric series
+Reference: https://en.wikipedia.org/wiki/Geometric_series
+"""
+
+
+def is_geometric_series(series: list) -> bool:
+ """
+ checking whether the input series is geometric series or not
+ >>> is_geometric_series([2, 4, 8])
+ True
+ >>> is_geometric_series([3, 6, 12, 24])
+ True
+ >>> is_geometric_series([1, 2, 3])
+ False
+ >>> is_geometric_series([0, 0, 3])
+ False
+ >>> is_geometric_series([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input list must be a non empty list
+ >>> is_geometric_series(4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series is not valid, valid series - [2, 4, 8]
+ """
+ if not isinstance(series, list):
+ raise ValueError("Input series is not valid, valid series - [2, 4, 8]")
+ if len(series) == 0:
+ raise ValueError("Input list must be a non empty list")
+ if len(series) == 1:
+ return True
+ try:
+ common_ratio = series[1] / series[0]
+ for index in range(len(series) - 1):
+ if series[index + 1] / series[index] != common_ratio:
+ return False
+ except ZeroDivisionError:
+ return False
+ return True
+
+
+def geometric_mean(series: list) -> float:
+ """
+ return the geometric mean of series
+
+ >>> geometric_mean([2, 4, 8])
+ 3.9999999999999996
+ >>> geometric_mean([3, 6, 12, 24])
+ 8.48528137423857
+ >>> geometric_mean([4, 8, 16])
+ 7.999999999999999
+ >>> geometric_mean(4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series is not valid, valid series - [2, 4, 8]
+ >>> geometric_mean([1, 2, 3])
+ 1.8171205928321397
+ >>> geometric_mean([0, 2, 3])
+ 0.0
+ >>> geometric_mean([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input list must be a non empty list
+
+ """
+ if not isinstance(series, list):
+ raise ValueError("Input series is not valid, valid series - [2, 4, 8]")
+ if len(series) == 0:
+ raise ValueError("Input list must be a non empty list")
+ answer = 1
+ for value in series:
+ answer *= value
+ return pow(answer, 1 / len(series))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/series/geometric_series.py b/maths/series/geometric_series.py
index d12382e6d8c4..55c42fd90e99 100644
--- a/maths/series/geometric_series.py
+++ b/maths/series/geometric_series.py
@@ -1,7 +1,6 @@
"""
This is a pure Python implementation of the Geometric Series algorithm
https://en.wikipedia.org/wiki/Geometric_series
-
Run the doctests with the following command:
python3 -m doctest -v geometric_series.py
or
@@ -10,9 +9,17 @@
python3 geometric_series.py
"""
+from __future__ import annotations
+
+
+def geometric_series(
+ nth_term: float,
+ start_term_a: float,
+ common_ratio_r: float,
+) -> list[float]:
+ """
+ Pure Python implementation of Geometric Series algorithm
-def geometric_series(nth_term: int, start_term_a: int, common_ratio_r: int) -> list:
- """Pure Python implementation of Geometric Series algorithm
:param nth_term: The last term (nth term of Geometric Series)
:param start_term_a : The first term of Geometric Series
:param common_ratio_r : The common ratio between all the terms
@@ -20,15 +27,15 @@ def geometric_series(nth_term: int, start_term_a: int, common_ratio_r: int) -> l
ration with first term with increase in power till last term (nth term)
Examples:
>>> geometric_series(4, 2, 2)
- [2, '4.0', '8.0', '16.0']
+ [2, 4.0, 8.0, 16.0]
>>> geometric_series(4.0, 2.0, 2.0)
- [2.0, '4.0', '8.0', '16.0']
+ [2.0, 4.0, 8.0, 16.0]
>>> geometric_series(4.1, 2.1, 2.1)
- [2.1, '4.41', '9.261000000000001', '19.448100000000004']
+ [2.1, 4.41, 9.261000000000001, 19.448100000000004]
>>> geometric_series(4, 2, -2)
- [2, '-4.0', '8.0', '-16.0']
+ [2, -4.0, 8.0, -16.0]
>>> geometric_series(4, -2, 2)
- [-2, '-4.0', '-8.0', '-16.0']
+ [-2, -4.0, -8.0, -16.0]
>>> geometric_series(-4, 2, 2)
[]
>>> geometric_series(0, 100, 500)
@@ -38,26 +45,30 @@ def geometric_series(nth_term: int, start_term_a: int, common_ratio_r: int) -> l
>>> geometric_series(0, 0, 0)
[]
"""
- if "" in (nth_term, start_term_a, common_ratio_r):
- return ""
- series = []
+ if not all((nth_term, start_term_a, common_ratio_r)):
+ return []
+ series: list[float] = []
power = 1
multiple = common_ratio_r
for _ in range(int(nth_term)):
- if series == []:
+ if not series:
series.append(start_term_a)
else:
power += 1
- series.append(str(float(start_term_a) * float(multiple)))
+ series.append(float(start_term_a * multiple))
multiple = pow(float(common_ratio_r), power)
return series
if __name__ == "__main__":
- nth_term = input("Enter the last number (n term) of the Geometric Series")
- start_term_a = input("Enter the starting term (a) of the Geometric Series")
- common_ratio_r = input(
- "Enter the common ratio between two terms (r) of the Geometric Series"
+ import doctest
+
+ doctest.testmod()
+
+ nth_term = float(input("Enter the last number (n term) of the Geometric Series"))
+ start_term_a = float(input("Enter the starting term (a) of the Geometric Series"))
+ common_ratio_r = float(
+ input("Enter the common ratio between two terms (r) of the Geometric Series")
)
print("Formula of Geometric Series => a + ar + ar^2 ... +ar^n")
print(geometric_series(nth_term, start_term_a, common_ratio_r))
diff --git a/maths/series/harmonic.py b/maths/series/harmonic.py
new file mode 100644
index 000000000000..35792d38af9b
--- /dev/null
+++ b/maths/series/harmonic.py
@@ -0,0 +1,92 @@
+"""
+Harmonic mean
+Reference: https://en.wikipedia.org/wiki/Harmonic_mean
+
+Harmonic series
+Reference: https://en.wikipedia.org/wiki/Harmonic_series(mathematics)
+"""
+
+
+def is_harmonic_series(series: list) -> bool:
+ """
+ checking whether the input series is arithmetic series or not
+ >>> is_harmonic_series([ 1, 2/3, 1/2, 2/5, 1/3])
+ True
+ >>> is_harmonic_series([ 1, 2/3, 2/5, 1/3])
+ False
+ >>> is_harmonic_series([1, 2, 3])
+ False
+ >>> is_harmonic_series([1/2, 1/3, 1/4])
+ True
+ >>> is_harmonic_series([2/5, 2/10, 2/15, 2/20, 2/25])
+ True
+ >>> is_harmonic_series(4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series is not valid, valid series - [1, 2/3, 2]
+ >>> is_harmonic_series([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input list must be a non empty list
+ >>> is_harmonic_series([0])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series cannot have 0 as an element
+ >>> is_harmonic_series([1,2,0,6])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series cannot have 0 as an element
+ """
+ if not isinstance(series, list):
+ raise ValueError("Input series is not valid, valid series - [1, 2/3, 2]")
+ if len(series) == 0:
+ raise ValueError("Input list must be a non empty list")
+ if len(series) == 1 and series[0] != 0:
+ return True
+ rec_series = []
+ series_len = len(series)
+ for i in range(series_len):
+ if series[i] == 0:
+ raise ValueError("Input series cannot have 0 as an element")
+ rec_series.append(1 / series[i])
+ common_diff = rec_series[1] - rec_series[0]
+ for index in range(2, series_len):
+ if rec_series[index] - rec_series[index - 1] != common_diff:
+ return False
+ return True
+
+
+def harmonic_mean(series: list) -> float:
+ """
+ return the harmonic mean of series
+
+ >>> harmonic_mean([1, 4, 4])
+ 2.0
+ >>> harmonic_mean([3, 6, 9, 12])
+ 5.759999999999999
+ >>> harmonic_mean(4)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input series is not valid, valid series - [2, 4, 6]
+ >>> harmonic_mean([1, 2, 3])
+ 1.6363636363636365
+ >>> harmonic_mean([])
+ Traceback (most recent call last):
+ ...
+ ValueError: Input list must be a non empty list
+
+ """
+ if not isinstance(series, list):
+ raise ValueError("Input series is not valid, valid series - [2, 4, 6]")
+ if len(series) == 0:
+ raise ValueError("Input list must be a non empty list")
+ answer = 0
+ for val in series:
+ answer += 1 / val
+ return len(series) / answer
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/series/harmonic_series.py b/maths/series/harmonic_series.py
index 91b5944583e4..d42d13d912f1 100644
--- a/maths/series/harmonic_series.py
+++ b/maths/series/harmonic_series.py
@@ -33,8 +33,8 @@ def harmonic_series(n_term: str) -> list:
['1']
"""
if n_term == "":
- return n_term
- series = []
+ return []
+ series: list = []
for temp in range(int(n_term)):
series.append(f"1/{temp + 1}" if series else "1")
return series
diff --git a/maths/series/hexagonal_numbers.py b/maths/series/hexagonal_numbers.py
new file mode 100644
index 000000000000..582b1989b7c6
--- /dev/null
+++ b/maths/series/hexagonal_numbers.py
@@ -0,0 +1,42 @@
+"""
+A hexagonal number sequence is a sequence of figurate numbers
+where the nth hexagonal number hₙ is the number of distinct dots
+in a pattern of dots consisting of the outlines of regular
+hexagons with sides up to n dots, when the hexagons are overlaid
+so that they share one vertex.
+
+ Calculates the hexagonal numbers sequence with a formula
+ hₙ = n(2n-1)
+ where:
+ hₙ --> is nth element of the sequence
+ n --> is the number of element in the sequence
+ reference-->"Hexagonal number" Wikipedia
+
+"""
+
+
+def hexagonal_numbers(length: int) -> list[int]:
+ """
+ :param len: max number of elements
+ :type len: int
+ :return: Hexagonal numbers as a list
+
+ Tests:
+ >>> hexagonal_numbers(10)
+ [0, 1, 6, 15, 28, 45, 66, 91, 120, 153]
+ >>> hexagonal_numbers(5)
+ [0, 1, 6, 15, 28]
+ >>> hexagonal_numbers(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Length must be a positive integer.
+ """
+
+ if length <= 0 or not isinstance(length, int):
+ raise ValueError("Length must be a positive integer.")
+ return [n * (2 * n - 1) for n in range(length)]
+
+
+if __name__ == "__main__":
+ print(hexagonal_numbers(length=5))
+ print(hexagonal_numbers(length=10))
diff --git a/maths/series/p_series.py b/maths/series/p_series.py
index 04019aed5a85..93812f443857 100644
--- a/maths/series/p_series.py
+++ b/maths/series/p_series.py
@@ -1,48 +1,51 @@
"""
This is a pure Python implementation of the P-Series algorithm
https://en.wikipedia.org/wiki/Harmonic_series_(mathematics)#P-series
-
For doctests run following command:
python -m doctest -v p_series.py
or
python3 -m doctest -v p_series.py
-
For manual testing run:
python3 p_series.py
"""
+from __future__ import annotations
-def p_series(nth_term: int, power: int) -> list:
- """Pure Python implementation of P-Series algorithm
+def p_series(nth_term: float | str, power: float | str) -> list[str]:
+ """
+ Pure Python implementation of P-Series algorithm
:return: The P-Series starting from 1 to last (nth) term
-
Examples:
>>> p_series(5, 2)
- [1, '1/4', '1/9', '1/16', '1/25']
+ ['1', '1 / 4', '1 / 9', '1 / 16', '1 / 25']
>>> p_series(-5, 2)
[]
>>> p_series(5, -2)
- [1, '1/0.25', '1/0.1111111111111111', '1/0.0625', '1/0.04']
+ ['1', '1 / 0.25', '1 / 0.1111111111111111', '1 / 0.0625', '1 / 0.04']
>>> p_series("", 1000)
- ''
+ ['']
>>> p_series(0, 0)
[]
>>> p_series(1, 1)
- [1]
+ ['1']
"""
if nth_term == "":
- return nth_term
+ return [""]
nth_term = int(nth_term)
power = int(power)
- series = []
+ series: list[str] = []
for temp in range(int(nth_term)):
- series.append(f"1/{pow(temp + 1, int(power))}" if series else 1)
+ series.append(f"1 / {pow(temp + 1, int(power))}" if series else "1")
return series
if __name__ == "__main__":
- nth_term = input("Enter the last number (nth term) of the P-Series")
- power = input("Enter the power for P-Series")
+ import doctest
+
+ doctest.testmod()
+
+ nth_term = int(input("Enter the last number (nth term) of the P-Series"))
+ power = int(input("Enter the power for P-Series"))
print("Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p")
print(p_series(nth_term, power))
diff --git a/maths/sieve_of_eratosthenes.py b/maths/sieve_of_eratosthenes.py
index faf6fc0f9a98..3923dc3e1612 100644
--- a/maths/sieve_of_eratosthenes.py
+++ b/maths/sieve_of_eratosthenes.py
@@ -8,54 +8,59 @@
Reference: https://en.wikipedia.org/wiki/Sieve_of_Eratosthenes
doctest provider: Bruno Simas Hadlich (https://github.com/brunohadlich)
-Also thanks Dmitry (https://github.com/LizardWizzard) for finding the problem
+Also thanks to Dmitry (https://github.com/LizardWizzard) for finding the problem
"""
+from __future__ import annotations
import math
-def sieve(n):
+def prime_sieve(num: int) -> list[int]:
"""
Returns a list with all prime numbers up to n.
- >>> sieve(50)
+ >>> prime_sieve(50)
[2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47]
- >>> sieve(25)
+ >>> prime_sieve(25)
[2, 3, 5, 7, 11, 13, 17, 19, 23]
- >>> sieve(10)
+ >>> prime_sieve(10)
[2, 3, 5, 7]
- >>> sieve(9)
+ >>> prime_sieve(9)
[2, 3, 5, 7]
- >>> sieve(2)
+ >>> prime_sieve(2)
[2]
- >>> sieve(1)
+ >>> prime_sieve(1)
[]
"""
- l = [True] * (n + 1) # noqa: E741
+ if num <= 0:
+ msg = f"{num}: Invalid input, please enter a positive integer."
+ raise ValueError(msg)
+
+ sieve = [True] * (num + 1)
prime = []
start = 2
- end = int(math.sqrt(n))
+ end = int(math.sqrt(num))
while start <= end:
# If start is a prime
- if l[start] is True:
+ if sieve[start] is True:
prime.append(start)
# Set multiples of start be False
- for i in range(start * start, n + 1, start):
- if l[i] is True:
- l[i] = False
+ for i in range(start * start, num + 1, start):
+ if sieve[i] is True:
+ sieve[i] = False
start += 1
- for j in range(end + 1, n + 1):
- if l[j] is True:
+ for j in range(end + 1, num + 1):
+ if sieve[j] is True:
prime.append(j)
return prime
if __name__ == "__main__":
- print(sieve(int(input("Enter n: ").strip())))
+ print(prime_sieve(int(input("Enter a positive integer: ").strip())))
diff --git a/maths/sigmoid.py b/maths/sigmoid.py
index 147588e8871f..cb45bde2702c 100644
--- a/maths/sigmoid.py
+++ b/maths/sigmoid.py
@@ -11,7 +11,7 @@
import numpy as np
-def sigmoid(vector: np.array) -> np.array:
+def sigmoid(vector: np.ndarray) -> np.ndarray:
"""
Implements the sigmoid function
diff --git a/maths/signum.py b/maths/signum.py
new file mode 100644
index 000000000000..c89753e76637
--- /dev/null
+++ b/maths/signum.py
@@ -0,0 +1,58 @@
+"""
+Signum function -- https://en.wikipedia.org/wiki/Sign_function
+"""
+
+
+def signum(num: float) -> int:
+ """
+ Applies signum function on the number
+
+ Custom test cases:
+ >>> signum(-10)
+ -1
+ >>> signum(10)
+ 1
+ >>> signum(0)
+ 0
+ >>> signum(-20.5)
+ -1
+ >>> signum(20.5)
+ 1
+ >>> signum(-1e-6)
+ -1
+ >>> signum(1e-6)
+ 1
+ >>> signum("Hello")
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'str' and 'int'
+ >>> signum([])
+ Traceback (most recent call last):
+ ...
+ TypeError: '<' not supported between instances of 'list' and 'int'
+ """
+ if num < 0:
+ return -1
+ return 1 if num else 0
+
+
+def test_signum() -> None:
+ """
+ Tests the signum function
+ >>> test_signum()
+ """
+ assert signum(5) == 1
+ assert signum(-5) == -1
+ assert signum(0) == 0
+ assert signum(10.5) == 1
+ assert signum(-10.5) == -1
+ assert signum(1e-6) == 1
+ assert signum(-1e-6) == -1
+ assert signum(123456789) == 1
+ assert signum(-123456789) == -1
+
+
+if __name__ == "__main__":
+ print(signum(12))
+ print(signum(-12))
+ print(signum(0))
diff --git a/maths/simpson_rule.py b/maths/simpson_rule.py
deleted file mode 100644
index d66dc39a7171..000000000000
--- a/maths/simpson_rule.py
+++ /dev/null
@@ -1,51 +0,0 @@
-"""
-Numerical integration or quadrature for a smooth function f with known values at x_i
-
-This method is the classical approach of suming 'Equally Spaced Abscissas'
-
-method 2:
-"Simpson Rule"
-
-"""
-
-
-def method_2(boundary, steps):
- # "Simpson Rule"
- # int(f) = delta_x/2 * (b-a)/3*(f1 + 4f2 + 2f_3 + ... + fn)
- h = (boundary[1] - boundary[0]) / steps
- a = boundary[0]
- b = boundary[1]
- x_i = make_points(a, b, h)
- y = 0.0
- y += (h / 3.0) * f(a)
- cnt = 2
- for i in x_i:
- y += (h / 3) * (4 - 2 * (cnt % 2)) * f(i)
- cnt += 1
- y += (h / 3.0) * f(b)
- return y
-
-
-def make_points(a, b, h):
- x = a + h
- while x < (b - h):
- yield x
- x = x + h
-
-
-def f(x): # enter your function here
- y = (x - 0) * (x - 0)
- return y
-
-
-def main():
- a = 0.0 # Lower bound of integration
- b = 1.0 # Upper bound of integration
- steps = 10.0 # define number of steps or resolution
- boundary = [a, b] # define boundary of integration
- y = method_2(boundary, steps)
- print(f"y = {y}")
-
-
-if __name__ == "__main__":
- main()
diff --git a/maths/simultaneous_linear_equation_solver.py b/maths/simultaneous_linear_equation_solver.py
new file mode 100644
index 000000000000..9685a33e82fe
--- /dev/null
+++ b/maths/simultaneous_linear_equation_solver.py
@@ -0,0 +1,142 @@
+"""
+https://en.wikipedia.org/wiki/Augmented_matrix
+
+This algorithm solves simultaneous linear equations of the form
+λa + λb + λc + λd + ... = y as [λ, λ, λ, λ, ..., y]
+Where λ & y are individual coefficients, the no. of equations = no. of coefficients - 1
+
+Note in order to work there must exist 1 equation where all instances of λ and y != 0
+"""
+
+
+def simplify(current_set: list[list]) -> list[list]:
+ """
+ >>> simplify([[1, 2, 3], [4, 5, 6]])
+ [[1.0, 2.0, 3.0], [0.0, 0.75, 1.5]]
+ >>> simplify([[5, 2, 5], [5, 1, 10]])
+ [[1.0, 0.4, 1.0], [0.0, 0.2, -1.0]]
+ """
+ # Divide each row by magnitude of first term --> creates 'unit' matrix
+ duplicate_set = current_set.copy()
+ for row_index, row in enumerate(duplicate_set):
+ magnitude = row[0]
+ for column_index, column in enumerate(row):
+ if magnitude == 0:
+ current_set[row_index][column_index] = column
+ continue
+ current_set[row_index][column_index] = column / magnitude
+ # Subtract to cancel term
+ first_row = current_set[0]
+ final_set = [first_row]
+ current_set = current_set[1::]
+ for row in current_set:
+ temp_row = []
+ # If first term is 0, it is already in form we want, so we preserve it
+ if row[0] == 0:
+ final_set.append(row)
+ continue
+ for column_index in range(len(row)):
+ temp_row.append(first_row[column_index] - row[column_index])
+ final_set.append(temp_row)
+ # Create next recursion iteration set
+ if len(final_set[0]) != 3:
+ current_first_row = final_set[0]
+ current_first_column = []
+ next_iteration = []
+ for row in final_set[1::]:
+ current_first_column.append(row[0])
+ next_iteration.append(row[1::])
+ resultant = simplify(next_iteration)
+ for i in range(len(resultant)):
+ resultant[i].insert(0, current_first_column[i])
+ resultant.insert(0, current_first_row)
+ final_set = resultant
+ return final_set
+
+
+def solve_simultaneous(equations: list[list]) -> list:
+ """
+ >>> solve_simultaneous([[1, 2, 3],[4, 5, 6]])
+ [-1.0, 2.0]
+ >>> solve_simultaneous([[0, -3, 1, 7],[3, 2, -1, 11],[5, 1, -2, 12]])
+ [6.4, 1.2, 10.6]
+ >>> solve_simultaneous([])
+ Traceback (most recent call last):
+ ...
+ IndexError: solve_simultaneous() requires n lists of length n+1
+ >>> solve_simultaneous([[1, 2, 3],[1, 2]])
+ Traceback (most recent call last):
+ ...
+ IndexError: solve_simultaneous() requires n lists of length n+1
+ >>> solve_simultaneous([[1, 2, 3],["a", 7, 8]])
+ Traceback (most recent call last):
+ ...
+ ValueError: solve_simultaneous() requires lists of integers
+ >>> solve_simultaneous([[0, 2, 3],[4, 0, 6]])
+ Traceback (most recent call last):
+ ...
+ ValueError: solve_simultaneous() requires at least 1 full equation
+ """
+ if len(equations) == 0:
+ raise IndexError("solve_simultaneous() requires n lists of length n+1")
+ _length = len(equations) + 1
+ if any(len(item) != _length for item in equations):
+ raise IndexError("solve_simultaneous() requires n lists of length n+1")
+ for row in equations:
+ if any(not isinstance(column, (int, float)) for column in row):
+ raise ValueError("solve_simultaneous() requires lists of integers")
+ if len(equations) == 1:
+ return [equations[0][-1] / equations[0][0]]
+ data_set = equations.copy()
+ if any(0 in row for row in data_set):
+ temp_data = data_set.copy()
+ full_row = []
+ for row_index, row in enumerate(temp_data):
+ if 0 not in row:
+ full_row = data_set.pop(row_index)
+ break
+ if not full_row:
+ raise ValueError("solve_simultaneous() requires at least 1 full equation")
+ data_set.insert(0, full_row)
+ useable_form = data_set.copy()
+ simplified = simplify(useable_form)
+ simplified = simplified[::-1]
+ solutions: list = []
+ for row in simplified:
+ current_solution = row[-1]
+ if not solutions:
+ if row[-2] == 0:
+ solutions.append(0)
+ continue
+ solutions.append(current_solution / row[-2])
+ continue
+ temp_row = row.copy()[: len(row) - 1 :]
+ while temp_row[0] == 0:
+ temp_row.pop(0)
+ if len(temp_row) == 0:
+ solutions.append(0)
+ continue
+ temp_row = temp_row[1::]
+ temp_row = temp_row[::-1]
+ for column_index, column in enumerate(temp_row):
+ current_solution -= column * solutions[column_index]
+ solutions.append(current_solution)
+ final = []
+ for item in solutions:
+ final.append(float(round(item, 5)))
+ return final[::-1]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ eq = [
+ [2, 1, 1, 1, 1, 4],
+ [1, 2, 1, 1, 1, 5],
+ [1, 1, 2, 1, 1, 6],
+ [1, 1, 1, 2, 1, 7],
+ [1, 1, 1, 1, 2, 8],
+ ]
+ print(solve_simultaneous(eq))
+ print(solve_simultaneous([[4, 2]]))
diff --git a/maths/sin.py b/maths/sin.py
new file mode 100644
index 000000000000..b06e6c9f1e5d
--- /dev/null
+++ b/maths/sin.py
@@ -0,0 +1,64 @@
+"""
+Calculate sin function.
+
+It's not a perfect function so I am rounding the result to 10 decimal places by default.
+
+Formula: sin(x) = x - x^3/3! + x^5/5! - x^7/7! + ...
+Where: x = angle in randians.
+
+Source:
+ https://www.homeschoolmath.net/teaching/sine_calculator.php
+
+"""
+
+from math import factorial, radians
+
+
+def sin(
+ angle_in_degrees: float, accuracy: int = 18, rounded_values_count: int = 10
+) -> float:
+ """
+ Implement sin function.
+
+ >>> sin(0.0)
+ 0.0
+ >>> sin(90.0)
+ 1.0
+ >>> sin(180.0)
+ 0.0
+ >>> sin(270.0)
+ -1.0
+ >>> sin(0.68)
+ 0.0118679603
+ >>> sin(1.97)
+ 0.0343762121
+ >>> sin(64.0)
+ 0.8987940463
+ >>> sin(9999.0)
+ -0.9876883406
+ >>> sin(-689.0)
+ 0.5150380749
+ >>> sin(89.7)
+ 0.9999862922
+ """
+ # Simplify the angle to be between 360 and -360 degrees.
+ angle_in_degrees = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0)
+
+ # Converting from degrees to radians
+ angle_in_radians = radians(angle_in_degrees)
+
+ result = angle_in_radians
+ a = 3
+ b = -1
+
+ for _ in range(accuracy):
+ result += (b * (angle_in_radians**a)) / factorial(a)
+
+ b = -b # One positive term and the next will be negative and so on...
+ a += 2 # Increased by 2 for every term.
+
+ return round(result, rounded_values_count)
+
+
+if __name__ == "__main__":
+ __import__("doctest").testmod()
diff --git a/maths/sock_merchant.py b/maths/sock_merchant.py
new file mode 100644
index 000000000000..304efec9ba5e
--- /dev/null
+++ b/maths/sock_merchant.py
@@ -0,0 +1,20 @@
+from collections import Counter
+
+
+def sock_merchant(colors: list[int]) -> int:
+ """
+ >>> sock_merchant([10, 20, 20, 10, 10, 30, 50, 10, 20])
+ 3
+ >>> sock_merchant([1, 1, 3, 3])
+ 2
+ """
+ return sum(socks_by_color // 2 for socks_by_color in Counter(colors).values())
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ colors = [int(x) for x in input("Enter socks by color :").rstrip().split()]
+ print(f"sock_merchant({colors}) = {sock_merchant(colors)}")
diff --git a/maths/softmax.py b/maths/softmax.py
index e021a7f8a6fe..95c95e66f59e 100644
--- a/maths/softmax.py
+++ b/maths/softmax.py
@@ -28,7 +28,7 @@ def softmax(vector):
The softmax vector adds up to one. We need to ceil to mitigate for
precision
- >>> np.ceil(np.sum(softmax([1,2,3,4])))
+ >>> float(np.ceil(np.sum(softmax([1,2,3,4]))))
1.0
>>> vec = np.array([5,5])
@@ -41,13 +41,13 @@ def softmax(vector):
# Calculate e^x for each x in your vector where e is Euler's
# number (approximately 2.718)
- exponentVector = np.exp(vector)
+ exponent_vector = np.exp(vector)
# Add up the all the exponentials
- sumOfExponents = np.sum(exponentVector)
+ sum_of_exponents = np.sum(exponent_vector)
# Divide every exponent by the sum of all exponents
- softmax_vector = exponentVector / sumOfExponents
+ softmax_vector = exponent_vector / sum_of_exponents
return softmax_vector
diff --git a/maths/solovay_strassen_primality_test.py b/maths/solovay_strassen_primality_test.py
new file mode 100644
index 000000000000..b2d905b07bed
--- /dev/null
+++ b/maths/solovay_strassen_primality_test.py
@@ -0,0 +1,106 @@
+"""
+This script implements the Solovay-Strassen Primality test.
+
+This probabilistic primality test is based on Euler's criterion. It is similar
+to the Fermat test but uses quadratic residues. It can quickly identify
+composite numbers but may occasionally classify composite numbers as prime.
+
+More details and concepts about this can be found on:
+https://en.wikipedia.org/wiki/Solovay%E2%80%93Strassen_primality_test
+"""
+
+import random
+
+
+def jacobi_symbol(random_a: int, number: int) -> int:
+ """
+ Calculate the Jacobi symbol. The Jacobi symbol is a generalization
+ of the Legendre symbol, which can be used to simplify computations involving
+ quadratic residues. The Jacobi symbol is used in primality tests, like the
+ Solovay-Strassen test, because it helps determine if an integer is a
+ quadratic residue modulo a given modulus, providing valuable information
+ about the number's potential primality or compositeness.
+
+ Parameters:
+ random_a: A randomly chosen integer from 2 to n-2 (inclusive)
+ number: The number that is tested for primality
+
+ Returns:
+ jacobi_symbol: The Jacobi symbol is a mathematical function
+ used to determine whether an integer is a quadratic residue modulo
+ another integer (usually prime) or not.
+
+ >>> jacobi_symbol(2, 13)
+ -1
+ >>> jacobi_symbol(5, 19)
+ 1
+ >>> jacobi_symbol(7, 14)
+ 0
+ """
+
+ if random_a in (0, 1):
+ return random_a
+
+ random_a %= number
+ t = 1
+
+ while random_a != 0:
+ while random_a % 2 == 0:
+ random_a //= 2
+ r = number % 8
+ if r in (3, 5):
+ t = -t
+
+ random_a, number = number, random_a
+
+ if random_a % 4 == number % 4 == 3:
+ t = -t
+
+ random_a %= number
+
+ return t if number == 1 else 0
+
+
+def solovay_strassen(number: int, iterations: int) -> bool:
+ """
+ Check whether the input number is prime or not using
+ the Solovay-Strassen Primality test
+
+ Parameters:
+ number: The number that is tested for primality
+ iterations: The number of times that the test is run
+ which effects the accuracy
+
+ Returns:
+ result: True if number is probably prime and false
+ if not
+
+ >>> random.seed(10)
+ >>> solovay_strassen(13, 5)
+ True
+ >>> solovay_strassen(9, 10)
+ False
+ >>> solovay_strassen(17, 15)
+ True
+ """
+
+ if number <= 1:
+ return False
+ if number <= 3:
+ return True
+
+ for _ in range(iterations):
+ a = random.randint(2, number - 2)
+ x = jacobi_symbol(a, number)
+ y = pow(a, (number - 1) // 2, number)
+
+ if x == 0 or y != x % number:
+ return False
+
+ return True
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/spearman_rank_correlation_coefficient.py b/maths/spearman_rank_correlation_coefficient.py
new file mode 100644
index 000000000000..32ff6b9e3d71
--- /dev/null
+++ b/maths/spearman_rank_correlation_coefficient.py
@@ -0,0 +1,82 @@
+from collections.abc import Sequence
+
+
+def assign_ranks(data: Sequence[float]) -> list[int]:
+ """
+ Assigns ranks to elements in the array.
+
+ :param data: List of floats.
+ :return: List of ints representing the ranks.
+
+ Example:
+ >>> assign_ranks([3.2, 1.5, 4.0, 2.7, 5.1])
+ [3, 1, 4, 2, 5]
+
+ >>> assign_ranks([10.5, 8.1, 12.4, 9.3, 11.0])
+ [3, 1, 5, 2, 4]
+ """
+ ranked_data = sorted((value, index) for index, value in enumerate(data))
+ ranks = [0] * len(data)
+
+ for position, (_, index) in enumerate(ranked_data):
+ ranks[index] = position + 1
+
+ return ranks
+
+
+def calculate_spearman_rank_correlation(
+ variable_1: Sequence[float], variable_2: Sequence[float]
+) -> float:
+ """
+ Calculates Spearman's rank correlation coefficient.
+
+ :param variable_1: List of floats representing the first variable.
+ :param variable_2: List of floats representing the second variable.
+ :return: Spearman's rank correlation coefficient.
+
+ Example Usage:
+
+ >>> x = [1, 2, 3, 4, 5]
+ >>> y = [5, 4, 3, 2, 1]
+ >>> calculate_spearman_rank_correlation(x, y)
+ -1.0
+
+ >>> x = [1, 2, 3, 4, 5]
+ >>> y = [2, 4, 6, 8, 10]
+ >>> calculate_spearman_rank_correlation(x, y)
+ 1.0
+
+ >>> x = [1, 2, 3, 4, 5]
+ >>> y = [5, 1, 2, 9, 5]
+ >>> calculate_spearman_rank_correlation(x, y)
+ 0.6
+ """
+ n = len(variable_1)
+ rank_var1 = assign_ranks(variable_1)
+ rank_var2 = assign_ranks(variable_2)
+
+ # Calculate differences of ranks
+ d = [rx - ry for rx, ry in zip(rank_var1, rank_var2)]
+
+ # Calculate the sum of squared differences
+ d_squared = sum(di**2 for di in d)
+
+ # Calculate the Spearman's rank correlation coefficient
+ rho = 1 - (6 * d_squared) / (n * (n**2 - 1))
+
+ return rho
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ # Example usage:
+ print(
+ f"{calculate_spearman_rank_correlation([1, 2, 3, 4, 5], [2, 4, 6, 8, 10]) = }"
+ )
+
+ print(f"{calculate_spearman_rank_correlation([1, 2, 3, 4, 5], [5, 4, 3, 2, 1]) = }")
+
+ print(f"{calculate_spearman_rank_correlation([1, 2, 3, 4, 5], [5, 1, 2, 9, 5]) = }")
diff --git a/maths/special_numbers/__init__.py b/maths/special_numbers/__init__.py
new file mode 100644
index 000000000000..e69de29bb2d1
diff --git a/maths/special_numbers/armstrong_numbers.py b/maths/special_numbers/armstrong_numbers.py
new file mode 100644
index 000000000000..a3cb69b814de
--- /dev/null
+++ b/maths/special_numbers/armstrong_numbers.py
@@ -0,0 +1,99 @@
+"""
+An Armstrong number is equal to the sum of its own digits each raised to the
+power of the number of digits.
+
+For example, 370 is an Armstrong number because 3*3*3 + 7*7*7 + 0*0*0 = 370.
+
+Armstrong numbers are also called Narcissistic numbers and Pluperfect numbers.
+
+On-Line Encyclopedia of Integer Sequences entry: https://oeis.org/A005188
+"""
+
+PASSING = (1, 153, 370, 371, 1634, 24678051, 115132219018763992565095597973971522401)
+FAILING: tuple = (-153, -1, 0, 1.2, 200, "A", [], {}, None)
+
+
+def armstrong_number(n: int) -> bool:
+ """
+ Return True if n is an Armstrong number or False if it is not.
+
+ >>> all(armstrong_number(n) for n in PASSING)
+ True
+ >>> any(armstrong_number(n) for n in FAILING)
+ False
+ """
+ if not isinstance(n, int) or n < 1:
+ return False
+
+ # Initialization of sum and number of digits.
+ total = 0
+ number_of_digits = 0
+ temp = n
+ # Calculation of digits of the number
+ number_of_digits = len(str(n))
+ # Dividing number into separate digits and find Armstrong number
+ temp = n
+ while temp > 0:
+ rem = temp % 10
+ total += rem**number_of_digits
+ temp //= 10
+ return n == total
+
+
+def pluperfect_number(n: int) -> bool:
+ """Return True if n is a pluperfect number or False if it is not
+
+ >>> all(pluperfect_number(n) for n in PASSING)
+ True
+ >>> any(pluperfect_number(n) for n in FAILING)
+ False
+ """
+ if not isinstance(n, int) or n < 1:
+ return False
+
+ # Init a "histogram" of the digits
+ digit_histogram = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
+ digit_total = 0
+ total = 0
+ temp = n
+ while temp > 0:
+ temp, rem = divmod(temp, 10)
+ digit_histogram[rem] += 1
+ digit_total += 1
+
+ for cnt, i in zip(digit_histogram, range(len(digit_histogram))):
+ total += cnt * i**digit_total
+
+ return n == total
+
+
+def narcissistic_number(n: int) -> bool:
+ """Return True if n is a narcissistic number or False if it is not.
+
+ >>> all(narcissistic_number(n) for n in PASSING)
+ True
+ >>> any(narcissistic_number(n) for n in FAILING)
+ False
+ """
+ if not isinstance(n, int) or n < 1:
+ return False
+ expo = len(str(n)) # the power that all digits will be raised to
+ # check if sum of each digit multiplied expo times is equal to number
+ return n == sum(int(i) ** expo for i in str(n))
+
+
+def main():
+ """
+ Request that user input an integer and tell them if it is Armstrong number.
+ """
+ num = int(input("Enter an integer to see if it is an Armstrong number: ").strip())
+ print(f"{num} is {'' if armstrong_number(num) else 'not '}an Armstrong number.")
+ print(f"{num} is {'' if narcissistic_number(num) else 'not '}an Armstrong number.")
+ print(f"{num} is {'' if pluperfect_number(num) else 'not '}an Armstrong number.")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ main()
diff --git a/maths/special_numbers/automorphic_number.py b/maths/special_numbers/automorphic_number.py
new file mode 100644
index 000000000000..8ed9375632a4
--- /dev/null
+++ b/maths/special_numbers/automorphic_number.py
@@ -0,0 +1,59 @@
+"""
+== Automorphic Numbers ==
+A number n is said to be a Automorphic number if
+the square of n "ends" in the same digits as n itself.
+
+Examples of Automorphic Numbers: 0, 1, 5, 6, 25, 76, 376, 625, 9376, 90625, ...
+https://en.wikipedia.org/wiki/Automorphic_number
+"""
+
+# Author : Akshay Dubey (https://github.com/itsAkshayDubey)
+# Time Complexity : O(log10n)
+
+
+def is_automorphic_number(number: int) -> bool:
+ """
+ # doctest: +NORMALIZE_WHITESPACE
+ This functions takes an integer number as input.
+ returns True if the number is automorphic.
+ >>> is_automorphic_number(-1)
+ False
+ >>> is_automorphic_number(0)
+ True
+ >>> is_automorphic_number(5)
+ True
+ >>> is_automorphic_number(6)
+ True
+ >>> is_automorphic_number(7)
+ False
+ >>> is_automorphic_number(25)
+ True
+ >>> is_automorphic_number(259918212890625)
+ True
+ >>> is_automorphic_number(259918212890636)
+ False
+ >>> is_automorphic_number(740081787109376)
+ True
+ >>> is_automorphic_number(5.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=5.0] must be an integer
+ """
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+ if number < 0:
+ return False
+ number_square = number * number
+ while number > 0:
+ if number % 10 != number_square % 10:
+ return False
+ number //= 10
+ number_square //= 10
+ return True
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/bell_numbers.py b/maths/special_numbers/bell_numbers.py
new file mode 100644
index 000000000000..d573e7a3962d
--- /dev/null
+++ b/maths/special_numbers/bell_numbers.py
@@ -0,0 +1,81 @@
+"""
+Bell numbers represent the number of ways to partition a set into non-empty
+subsets. This module provides functions to calculate Bell numbers for sets of
+integers. In other words, the first (n + 1) Bell numbers.
+
+For more information about Bell numbers, refer to:
+https://en.wikipedia.org/wiki/Bell_number
+"""
+
+
+def bell_numbers(max_set_length: int) -> list[int]:
+ """
+ Calculate Bell numbers for the sets of lengths from 0 to max_set_length.
+ In other words, calculate first (max_set_length + 1) Bell numbers.
+
+ Args:
+ max_set_length (int): The maximum length of the sets for which
+ Bell numbers are calculated.
+
+ Returns:
+ list: A list of Bell numbers for sets of lengths from 0 to max_set_length.
+
+ Examples:
+ >>> bell_numbers(-2)
+ Traceback (most recent call last):
+ ...
+ ValueError: max_set_length must be non-negative
+ >>> bell_numbers(0)
+ [1]
+ >>> bell_numbers(1)
+ [1, 1]
+ >>> bell_numbers(5)
+ [1, 1, 2, 5, 15, 52]
+ """
+ if max_set_length < 0:
+ raise ValueError("max_set_length must be non-negative")
+
+ bell = [0] * (max_set_length + 1)
+ bell[0] = 1
+
+ for i in range(1, max_set_length + 1):
+ for j in range(i):
+ bell[i] += _binomial_coefficient(i - 1, j) * bell[j]
+
+ return bell
+
+
+def _binomial_coefficient(total_elements: int, elements_to_choose: int) -> int:
+ """
+ Calculate the binomial coefficient C(total_elements, elements_to_choose)
+
+ Args:
+ total_elements (int): The total number of elements.
+ elements_to_choose (int): The number of elements to choose.
+
+ Returns:
+ int: The binomial coefficient C(total_elements, elements_to_choose).
+
+ Examples:
+ >>> _binomial_coefficient(5, 2)
+ 10
+ >>> _binomial_coefficient(6, 3)
+ 20
+ """
+ if elements_to_choose in {0, total_elements}:
+ return 1
+
+ elements_to_choose = min(elements_to_choose, total_elements - elements_to_choose)
+
+ coefficient = 1
+ for i in range(elements_to_choose):
+ coefficient *= total_elements - i
+ coefficient //= i + 1
+
+ return coefficient
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/carmichael_number.py b/maths/special_numbers/carmichael_number.py
new file mode 100644
index 000000000000..c73908545702
--- /dev/null
+++ b/maths/special_numbers/carmichael_number.py
@@ -0,0 +1,86 @@
+"""
+== Carmichael Numbers ==
+A number n is said to be a Carmichael number if it
+satisfies the following modular arithmetic condition:
+
+ power(b, n-1) MOD n = 1,
+ for all b ranging from 1 to n such that b and
+ n are relatively prime, i.e, gcd(b, n) = 1
+
+Examples of Carmichael Numbers: 561, 1105, ...
+https://en.wikipedia.org/wiki/Carmichael_number
+"""
+
+from maths.greatest_common_divisor import greatest_common_divisor
+
+
+def power(x: int, y: int, mod: int) -> int:
+ """
+ Examples:
+ >>> power(2, 15, 3)
+ 2
+ >>> power(5, 1, 30)
+ 5
+ """
+
+ if y == 0:
+ return 1
+ temp = power(x, y // 2, mod) % mod
+ temp = (temp * temp) % mod
+ if y % 2 == 1:
+ temp = (temp * x) % mod
+ return temp
+
+
+def is_carmichael_number(n: int) -> bool:
+ """
+ Examples:
+ >>> is_carmichael_number(4)
+ False
+ >>> is_carmichael_number(561)
+ True
+ >>> is_carmichael_number(562)
+ False
+ >>> is_carmichael_number(900)
+ False
+ >>> is_carmichael_number(1105)
+ True
+ >>> is_carmichael_number(8911)
+ True
+ >>> is_carmichael_number(5.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Number 5.1 must instead be a positive integer
+
+ >>> is_carmichael_number(-7)
+ Traceback (most recent call last):
+ ...
+ ValueError: Number -7 must instead be a positive integer
+
+ >>> is_carmichael_number(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Number 0 must instead be a positive integer
+ """
+
+ if n <= 0 or not isinstance(n, int):
+ msg = f"Number {n} must instead be a positive integer"
+ raise ValueError(msg)
+
+ return all(
+ power(b, n - 1, n) == 1
+ for b in range(2, n)
+ if greatest_common_divisor(b, n) == 1
+ )
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ number = int(input("Enter number: ").strip())
+ if is_carmichael_number(number):
+ print(f"{number} is a Carmichael Number.")
+ else:
+ print(f"{number} is not a Carmichael Number.")
diff --git a/maths/special_numbers/catalan_number.py b/maths/special_numbers/catalan_number.py
new file mode 100644
index 000000000000..20c2cfb17c06
--- /dev/null
+++ b/maths/special_numbers/catalan_number.py
@@ -0,0 +1,53 @@
+"""
+
+Calculate the nth Catalan number
+
+Source:
+ https://en.wikipedia.org/wiki/Catalan_number
+
+"""
+
+
+def catalan(number: int) -> int:
+ """
+ :param number: nth catalan number to calculate
+ :return: the nth catalan number
+ Note: A catalan number is only defined for positive integers
+
+ >>> catalan(5)
+ 14
+ >>> catalan(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value of [number=0] must be > 0
+ >>> catalan(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value of [number=-1] must be > 0
+ >>> catalan(5.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=5.0] must be an integer
+ """
+
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+
+ if number < 1:
+ msg = f"Input value of [number={number}] must be > 0"
+ raise ValueError(msg)
+
+ current_number = 1
+
+ for i in range(1, number):
+ current_number *= 4 * i - 2
+ current_number //= i + 1
+
+ return current_number
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/hamming_numbers.py b/maths/special_numbers/hamming_numbers.py
new file mode 100644
index 000000000000..a473cc93883b
--- /dev/null
+++ b/maths/special_numbers/hamming_numbers.py
@@ -0,0 +1,55 @@
+"""
+A Hamming number is a positive integer of the form 2^i*3^j*5^k, for some
+non-negative integers i, j, and k. They are often referred to as regular numbers.
+More info at: https://en.wikipedia.org/wiki/Regular_number.
+"""
+
+
+def hamming(n_element: int) -> list:
+ """
+ This function creates an ordered list of n length as requested, and afterwards
+ returns the last value of the list. It must be given a positive integer.
+
+ :param n_element: The number of elements on the list
+ :return: The nth element of the list
+
+ >>> hamming(-5)
+ Traceback (most recent call last):
+ ...
+ ValueError: n_element should be a positive number
+ >>> hamming(5)
+ [1, 2, 3, 4, 5]
+ >>> hamming(10)
+ [1, 2, 3, 4, 5, 6, 8, 9, 10, 12]
+ >>> hamming(15)
+ [1, 2, 3, 4, 5, 6, 8, 9, 10, 12, 15, 16, 18, 20, 24]
+ """
+ n_element = int(n_element)
+ if n_element < 1:
+ my_error = ValueError("n_element should be a positive number")
+ raise my_error
+
+ hamming_list = [1]
+ i, j, k = (0, 0, 0)
+ index = 1
+ while index < n_element:
+ while hamming_list[i] * 2 <= hamming_list[-1]:
+ i += 1
+ while hamming_list[j] * 3 <= hamming_list[-1]:
+ j += 1
+ while hamming_list[k] * 5 <= hamming_list[-1]:
+ k += 1
+ hamming_list.append(
+ min(hamming_list[i] * 2, hamming_list[j] * 3, hamming_list[k] * 5)
+ )
+ index += 1
+ return hamming_list
+
+
+if __name__ == "__main__":
+ n = input("Enter the last number (nth term) of the Hamming Number Series: ")
+ print("Formula of Hamming Number Series => 2^i * 3^j * 5^k")
+ hamming_numbers = hamming(int(n))
+ print("-----------------------------------------------------")
+ print(f"The list with nth numbers is: {hamming_numbers}")
+ print("-----------------------------------------------------")
diff --git a/maths/special_numbers/happy_number.py b/maths/special_numbers/happy_number.py
new file mode 100644
index 000000000000..eac3167e304b
--- /dev/null
+++ b/maths/special_numbers/happy_number.py
@@ -0,0 +1,48 @@
+def is_happy_number(number: int) -> bool:
+ """
+ A happy number is a number which eventually reaches 1 when replaced by the sum of
+ the square of each digit.
+
+ :param number: The number to check for happiness.
+ :return: True if the number is a happy number, False otherwise.
+
+ >>> is_happy_number(19)
+ True
+ >>> is_happy_number(2)
+ False
+ >>> is_happy_number(23)
+ True
+ >>> is_happy_number(1)
+ True
+ >>> is_happy_number(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: number=0 must be a positive integer
+ >>> is_happy_number(-19)
+ Traceback (most recent call last):
+ ...
+ ValueError: number=-19 must be a positive integer
+ >>> is_happy_number(19.1)
+ Traceback (most recent call last):
+ ...
+ ValueError: number=19.1 must be a positive integer
+ >>> is_happy_number("happy")
+ Traceback (most recent call last):
+ ...
+ ValueError: number='happy' must be a positive integer
+ """
+ if not isinstance(number, int) or number <= 0:
+ msg = f"{number=} must be a positive integer"
+ raise ValueError(msg)
+
+ seen = set()
+ while number != 1 and number not in seen:
+ seen.add(number)
+ number = sum(int(digit) ** 2 for digit in str(number))
+ return number == 1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/harshad_numbers.py b/maths/special_numbers/harshad_numbers.py
new file mode 100644
index 000000000000..417120bd840e
--- /dev/null
+++ b/maths/special_numbers/harshad_numbers.py
@@ -0,0 +1,166 @@
+"""
+A harshad number (or more specifically an n-harshad number) is a number that's
+divisible by the sum of its digits in some given base n.
+Reference: https://en.wikipedia.org/wiki/Harshad_number
+"""
+
+
+def int_to_base(number: int, base: int) -> str:
+ """
+ Convert a given positive decimal integer to base 'base'.
+ Where 'base' ranges from 2 to 36.
+
+ Examples:
+ >>> int_to_base(0, 21)
+ '0'
+ >>> int_to_base(23, 2)
+ '10111'
+ >>> int_to_base(58, 5)
+ '213'
+ >>> int_to_base(167, 16)
+ 'A7'
+ >>> # bases below 2 and beyond 36 will error
+ >>> int_to_base(98, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ >>> int_to_base(98, 37)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ >>> int_to_base(-99, 16)
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be a positive integer
+ """
+
+ if base < 2 or base > 36:
+ raise ValueError("'base' must be between 2 and 36 inclusive")
+
+ digits = "0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ"
+ result = ""
+
+ if number < 0:
+ raise ValueError("number must be a positive integer")
+
+ while number > 0:
+ number, remainder = divmod(number, base)
+ result = digits[remainder] + result
+
+ if result == "":
+ result = "0"
+
+ return result
+
+
+def sum_of_digits(num: int, base: int) -> str:
+ """
+ Calculate the sum of digit values in a positive integer
+ converted to the given 'base'.
+ Where 'base' ranges from 2 to 36.
+
+ Examples:
+ >>> sum_of_digits(103, 12)
+ '13'
+ >>> sum_of_digits(1275, 4)
+ '30'
+ >>> sum_of_digits(6645, 2)
+ '1001'
+ >>> # bases below 2 and beyond 36 will error
+ >>> sum_of_digits(543, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ >>> sum_of_digits(543, 37)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ """
+
+ if base < 2 or base > 36:
+ raise ValueError("'base' must be between 2 and 36 inclusive")
+
+ num_str = int_to_base(num, base)
+ res = sum(int(char, base) for char in num_str)
+ res_str = int_to_base(res, base)
+ return res_str
+
+
+def harshad_numbers_in_base(limit: int, base: int) -> list[str]:
+ """
+ Finds all Harshad numbers smaller than num in base 'base'.
+ Where 'base' ranges from 2 to 36.
+
+ Examples:
+ >>> harshad_numbers_in_base(15, 2)
+ ['1', '10', '100', '110', '1000', '1010', '1100']
+ >>> harshad_numbers_in_base(12, 34)
+ ['1', '2', '3', '4', '5', '6', '7', '8', '9', 'A', 'B']
+ >>> harshad_numbers_in_base(12, 4)
+ ['1', '2', '3', '10', '12', '20', '21']
+ >>> # bases below 2 and beyond 36 will error
+ >>> harshad_numbers_in_base(234, 37)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ >>> harshad_numbers_in_base(234, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ >>> harshad_numbers_in_base(-12, 6)
+ []
+ """
+
+ if base < 2 or base > 36:
+ raise ValueError("'base' must be between 2 and 36 inclusive")
+
+ if limit < 0:
+ return []
+
+ numbers = [
+ int_to_base(i, base)
+ for i in range(1, limit)
+ if i % int(sum_of_digits(i, base), base) == 0
+ ]
+
+ return numbers
+
+
+def is_harshad_number_in_base(num: int, base: int) -> bool:
+ """
+ Determines whether n in base 'base' is a harshad number.
+ Where 'base' ranges from 2 to 36.
+
+ Examples:
+ >>> is_harshad_number_in_base(18, 10)
+ True
+ >>> is_harshad_number_in_base(21, 10)
+ True
+ >>> is_harshad_number_in_base(-21, 5)
+ False
+ >>> # bases below 2 and beyond 36 will error
+ >>> is_harshad_number_in_base(45, 37)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ >>> is_harshad_number_in_base(45, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: 'base' must be between 2 and 36 inclusive
+ """
+
+ if base < 2 or base > 36:
+ raise ValueError("'base' must be between 2 and 36 inclusive")
+
+ if num < 0:
+ return False
+
+ n = int_to_base(num, base)
+ d = sum_of_digits(num, base)
+ return int(n, base) % int(d, base) == 0
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/hexagonal_number.py b/maths/special_numbers/hexagonal_number.py
new file mode 100644
index 000000000000..3677ab95ee00
--- /dev/null
+++ b/maths/special_numbers/hexagonal_number.py
@@ -0,0 +1,49 @@
+"""
+== Hexagonal Number ==
+The nth hexagonal number hn is the number of distinct dots
+in a pattern of dots consisting of the outlines of regular
+hexagons with sides up to n dots, when the hexagons are
+overlaid so that they share one vertex.
+
+https://en.wikipedia.org/wiki/Hexagonal_number
+"""
+
+# Author : Akshay Dubey (https://github.com/itsAkshayDubey)
+
+
+def hexagonal(number: int) -> int:
+ """
+ :param number: nth hexagonal number to calculate
+ :return: the nth hexagonal number
+ Note: A hexagonal number is only defined for positive integers
+ >>> hexagonal(4)
+ 28
+ >>> hexagonal(11)
+ 231
+ >>> hexagonal(22)
+ 946
+ >>> hexagonal(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive integer
+ >>> hexagonal(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input must be a positive integer
+ >>> hexagonal(11.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=11.0] must be an integer
+ """
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+ if number < 1:
+ raise ValueError("Input must be a positive integer")
+ return number * (2 * number - 1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/krishnamurthy_number.py b/maths/special_numbers/krishnamurthy_number.py
similarity index 92%
rename from maths/krishnamurthy_number.py
rename to maths/special_numbers/krishnamurthy_number.py
index c88f68a07f27..c1d8a8fc5f56 100644
--- a/maths/krishnamurthy_number.py
+++ b/maths/special_numbers/krishnamurthy_number.py
@@ -33,12 +33,12 @@ def krishnamurthy(number: int) -> bool:
True
"""
- factSum = 0
+ fact_sum = 0
duplicate = number
while duplicate > 0:
duplicate, digit = divmod(duplicate, 10)
- factSum += factorial(digit)
- return factSum == number
+ fact_sum += factorial(digit)
+ return fact_sum == number
if __name__ == "__main__":
diff --git a/maths/special_numbers/perfect_number.py b/maths/special_numbers/perfect_number.py
new file mode 100644
index 000000000000..a022dc677638
--- /dev/null
+++ b/maths/special_numbers/perfect_number.py
@@ -0,0 +1,79 @@
+"""
+== Perfect Number ==
+In number theory, a perfect number is a positive integer that is equal to the sum of
+its positive divisors, excluding the number itself.
+For example: 6 ==> divisors[1, 2, 3, 6]
+ Excluding 6, the sum(divisors) is 1 + 2 + 3 = 6
+ So, 6 is a Perfect Number
+
+Other examples of Perfect Numbers: 28, 486, ...
+
+https://en.wikipedia.org/wiki/Perfect_number
+"""
+
+
+def perfect(number: int) -> bool:
+ """
+ Check if a number is a perfect number.
+
+ A perfect number is a positive integer that is equal to the sum of its proper
+ divisors (excluding itself).
+
+ Args:
+ number: The number to be checked.
+
+ Returns:
+ True if the number is a perfect number, False otherwise.
+
+ Start from 1 because dividing by 0 will raise ZeroDivisionError.
+ A number at most can be divisible by the half of the number except the number
+ itself. For example, 6 is at most can be divisible by 3 except by 6 itself.
+
+ Examples:
+ >>> perfect(27)
+ False
+ >>> perfect(28)
+ True
+ >>> perfect(29)
+ False
+ >>> perfect(6)
+ True
+ >>> perfect(12)
+ False
+ >>> perfect(496)
+ True
+ >>> perfect(8128)
+ True
+ >>> perfect(0)
+ False
+ >>> perfect(-1)
+ False
+ >>> perfect(12.34)
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be an integer
+ >>> perfect("Hello")
+ Traceback (most recent call last):
+ ...
+ ValueError: number must be an integer
+ """
+ if not isinstance(number, int):
+ raise ValueError("number must be an integer")
+ if number <= 0:
+ return False
+ return sum(i for i in range(1, number // 2 + 1) if number % i == 0) == number
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ print("Program to check whether a number is a Perfect number or not...")
+ try:
+ number = int(input("Enter a positive integer: ").strip())
+ except ValueError:
+ msg = "number must be an integer"
+ print(msg)
+ raise ValueError(msg)
+
+ print(f"{number} is {'' if perfect(number) else 'not '}a Perfect Number.")
diff --git a/maths/special_numbers/polygonal_numbers.py b/maths/special_numbers/polygonal_numbers.py
new file mode 100644
index 000000000000..7a7dc91acb26
--- /dev/null
+++ b/maths/special_numbers/polygonal_numbers.py
@@ -0,0 +1,32 @@
+def polygonal_num(num: int, sides: int) -> int:
+ """
+ Returns the `num`th `sides`-gonal number. It is assumed that `num` >= 0 and
+ `sides` >= 3 (see for reference https://en.wikipedia.org/wiki/Polygonal_number).
+
+ >>> polygonal_num(0, 3)
+ 0
+ >>> polygonal_num(3, 3)
+ 6
+ >>> polygonal_num(5, 4)
+ 25
+ >>> polygonal_num(2, 5)
+ 5
+ >>> polygonal_num(-1, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid input: num must be >= 0 and sides must be >= 3.
+ >>> polygonal_num(0, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid input: num must be >= 0 and sides must be >= 3.
+ """
+ if num < 0 or sides < 3:
+ raise ValueError("Invalid input: num must be >= 0 and sides must be >= 3.")
+
+ return ((sides - 2) * num**2 - (sides - 4) * num) // 2
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/pronic_number.py b/maths/special_numbers/pronic_number.py
new file mode 100644
index 000000000000..cf4d3d2eb24b
--- /dev/null
+++ b/maths/special_numbers/pronic_number.py
@@ -0,0 +1,55 @@
+"""
+== Pronic Number ==
+A number n is said to be a Proic number if
+there exists an integer m such that n = m * (m + 1)
+
+Examples of Proic Numbers: 0, 2, 6, 12, 20, 30, 42, 56, 72, 90, 110 ...
+https://en.wikipedia.org/wiki/Pronic_number
+"""
+
+# Author : Akshay Dubey (https://github.com/itsAkshayDubey)
+
+
+def is_pronic(number: int) -> bool:
+ """
+ # doctest: +NORMALIZE_WHITESPACE
+ This functions takes an integer number as input.
+ returns True if the number is pronic.
+ >>> is_pronic(-1)
+ False
+ >>> is_pronic(0)
+ True
+ >>> is_pronic(2)
+ True
+ >>> is_pronic(5)
+ False
+ >>> is_pronic(6)
+ True
+ >>> is_pronic(8)
+ False
+ >>> is_pronic(30)
+ True
+ >>> is_pronic(32)
+ False
+ >>> is_pronic(2147441940)
+ True
+ >>> is_pronic(9223372033963249500)
+ True
+ >>> is_pronic(6.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=6.0] must be an integer
+ """
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+ if number < 0 or number % 2 == 1:
+ return False
+ number_sqrt = int(number**0.5)
+ return number == number_sqrt * (number_sqrt + 1)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/special_numbers/proth_number.py b/maths/special_numbers/proth_number.py
new file mode 100644
index 000000000000..47747ed260f7
--- /dev/null
+++ b/maths/special_numbers/proth_number.py
@@ -0,0 +1,75 @@
+"""
+Calculate the nth Proth number
+Source:
+ https://handwiki.org/wiki/Proth_number
+"""
+
+import math
+
+
+def proth(number: int) -> int:
+ """
+ :param number: nth number to calculate in the sequence
+ :return: the nth number in Proth number
+ Note: indexing starts at 1 i.e. proth(1) gives the first Proth number of 3
+ >>> proth(6)
+ 25
+ >>> proth(0)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value of [number=0] must be > 0
+ >>> proth(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Input value of [number=-1] must be > 0
+ >>> proth(6.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=6.0] must be an integer
+ """
+
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+
+ if number < 1:
+ msg = f"Input value of [number={number}] must be > 0"
+ raise ValueError(msg)
+ elif number == 1:
+ return 3
+ elif number == 2:
+ return 5
+ else:
+ """
+ +1 for binary starting at 0 i.e. 2^0, 2^1, etc.
+ +1 to start the sequence at the 3rd Proth number
+ Hence, we have a +2 in the below statement
+ """
+ block_index = int(math.log(number // 3, 2)) + 2
+
+ proth_list = [3, 5]
+ proth_index = 2
+ increment = 3
+ for block in range(1, block_index):
+ for _ in range(increment):
+ proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1])
+ proth_index += 1
+ increment *= 2
+
+ return proth_list[number - 1]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+
+ for number in range(11):
+ value = 0
+ try:
+ value = proth(number)
+ except ValueError:
+ print(f"ValueError: there is no {number}th Proth number")
+ continue
+
+ print(f"The {number}th Proth number: {value}")
diff --git a/maths/special_numbers/triangular_numbers.py b/maths/special_numbers/triangular_numbers.py
new file mode 100644
index 000000000000..5be89e6108b2
--- /dev/null
+++ b/maths/special_numbers/triangular_numbers.py
@@ -0,0 +1,43 @@
+"""
+A triangular number or triangle number counts objects arranged in an
+equilateral triangle. This module provides a function to generate n'th
+triangular number.
+
+For more information about triangular numbers, refer to:
+https://en.wikipedia.org/wiki/Triangular_number
+"""
+
+
+def triangular_number(position: int) -> int:
+ """
+ Generate the triangular number at the specified position.
+
+ Args:
+ position (int): The position of the triangular number to generate.
+
+ Returns:
+ int: The triangular number at the specified position.
+
+ Raises:
+ ValueError: If `position` is negative.
+
+ Examples:
+ >>> triangular_number(1)
+ 1
+ >>> triangular_number(3)
+ 6
+ >>> triangular_number(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: param `position` must be non-negative
+ """
+ if position < 0:
+ raise ValueError("param `position` must be non-negative")
+
+ return position * (position + 1) // 2
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/ugly_numbers.py b/maths/special_numbers/ugly_numbers.py
similarity index 94%
rename from maths/ugly_numbers.py
rename to maths/special_numbers/ugly_numbers.py
index 4451a68cdaad..c6ceb784622a 100644
--- a/maths/ugly_numbers.py
+++ b/maths/special_numbers/ugly_numbers.py
@@ -1,54 +1,54 @@
-"""
-Ugly numbers are numbers whose only prime factors are 2, 3 or 5. The sequence
-1, 2, 3, 4, 5, 6, 8, 9, 10, 12, 15, … shows the first 11 ugly numbers. By convention,
-1 is included.
-Given an integer n, we have to find the nth ugly number.
-
-For more details, refer this article
-https://www.geeksforgeeks.org/ugly-numbers/
-"""
-
-
-def ugly_numbers(n: int) -> int:
- """
- Returns the nth ugly number.
- >>> ugly_numbers(100)
- 1536
- >>> ugly_numbers(0)
- 1
- >>> ugly_numbers(20)
- 36
- >>> ugly_numbers(-5)
- 1
- >>> ugly_numbers(-5.5)
- Traceback (most recent call last):
- ...
- TypeError: 'float' object cannot be interpreted as an integer
- """
- ugly_nums = [1]
-
- i2, i3, i5 = 0, 0, 0
- next_2 = ugly_nums[i2] * 2
- next_3 = ugly_nums[i3] * 3
- next_5 = ugly_nums[i5] * 5
-
- for i in range(1, n):
- next_num = min(next_2, next_3, next_5)
- ugly_nums.append(next_num)
- if next_num == next_2:
- i2 += 1
- next_2 = ugly_nums[i2] * 2
- if next_num == next_3:
- i3 += 1
- next_3 = ugly_nums[i3] * 3
- if next_num == next_5:
- i5 += 1
- next_5 = ugly_nums[i5] * 5
- return ugly_nums[-1]
-
-
-if __name__ == "__main__":
- from doctest import testmod
-
- testmod(verbose=True)
- print(f"{ugly_numbers(200) = }")
+"""
+Ugly numbers are numbers whose only prime factors are 2, 3 or 5. The sequence
+1, 2, 3, 4, 5, 6, 8, 9, 10, 12, 15, … shows the first 11 ugly numbers. By convention,
+1 is included.
+Given an integer n, we have to find the nth ugly number.
+
+For more details, refer this article
+https://www.geeksforgeeks.org/ugly-numbers/
+"""
+
+
+def ugly_numbers(n: int) -> int:
+ """
+ Returns the nth ugly number.
+ >>> ugly_numbers(100)
+ 1536
+ >>> ugly_numbers(0)
+ 1
+ >>> ugly_numbers(20)
+ 36
+ >>> ugly_numbers(-5)
+ 1
+ >>> ugly_numbers(-5.5)
+ Traceback (most recent call last):
+ ...
+ TypeError: 'float' object cannot be interpreted as an integer
+ """
+ ugly_nums = [1]
+
+ i2, i3, i5 = 0, 0, 0
+ next_2 = ugly_nums[i2] * 2
+ next_3 = ugly_nums[i3] * 3
+ next_5 = ugly_nums[i5] * 5
+
+ for _ in range(1, n):
+ next_num = min(next_2, next_3, next_5)
+ ugly_nums.append(next_num)
+ if next_num == next_2:
+ i2 += 1
+ next_2 = ugly_nums[i2] * 2
+ if next_num == next_3:
+ i3 += 1
+ next_3 = ugly_nums[i3] * 3
+ if next_num == next_5:
+ i5 += 1
+ next_5 = ugly_nums[i5] * 5
+ return ugly_nums[-1]
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod(verbose=True)
+ print(f"{ugly_numbers(200) = }")
diff --git a/maths/special_numbers/weird_number.py b/maths/special_numbers/weird_number.py
new file mode 100644
index 000000000000..5c9240d0ea4e
--- /dev/null
+++ b/maths/special_numbers/weird_number.py
@@ -0,0 +1,101 @@
+"""
+https://en.wikipedia.org/wiki/Weird_number
+
+Fun fact: The set of weird numbers has positive asymptotic density.
+"""
+
+from math import sqrt
+
+
+def factors(number: int) -> list[int]:
+ """
+ >>> factors(12)
+ [1, 2, 3, 4, 6]
+ >>> factors(1)
+ [1]
+ >>> factors(100)
+ [1, 2, 4, 5, 10, 20, 25, 50]
+
+ # >>> factors(-12)
+ # [1, 2, 3, 4, 6]
+ """
+
+ values = [1]
+ for i in range(2, int(sqrt(number)) + 1, 1):
+ if number % i == 0:
+ values.append(i)
+ if int(number // i) != i:
+ values.append(int(number // i))
+ return sorted(values)
+
+
+def abundant(n: int) -> bool:
+ """
+ >>> abundant(0)
+ True
+ >>> abundant(1)
+ False
+ >>> abundant(12)
+ True
+ >>> abundant(13)
+ False
+ >>> abundant(20)
+ True
+
+ # >>> abundant(-12)
+ # True
+ """
+ return sum(factors(n)) > n
+
+
+def semi_perfect(number: int) -> bool:
+ """
+ >>> semi_perfect(0)
+ True
+ >>> semi_perfect(1)
+ True
+ >>> semi_perfect(12)
+ True
+ >>> semi_perfect(13)
+ False
+
+ # >>> semi_perfect(-12)
+ # True
+ """
+ values = factors(number)
+ r = len(values)
+ subset = [[0 for i in range(number + 1)] for j in range(r + 1)]
+ for i in range(r + 1):
+ subset[i][0] = True
+
+ for i in range(1, number + 1):
+ subset[0][i] = False
+
+ for i in range(1, r + 1):
+ for j in range(1, number + 1):
+ if j < values[i - 1]:
+ subset[i][j] = subset[i - 1][j]
+ else:
+ subset[i][j] = subset[i - 1][j] or subset[i - 1][j - values[i - 1]]
+
+ return subset[r][number] != 0
+
+
+def weird(number: int) -> bool:
+ """
+ >>> weird(0)
+ False
+ >>> weird(70)
+ True
+ >>> weird(77)
+ False
+ """
+ return abundant(number) and not semi_perfect(number)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod(verbose=True)
+ for number in (69, 70, 71):
+ print(f"{number} is {'' if weird(number) else 'not '}weird.")
diff --git a/maths/sum_of_arithmetic_series.py b/maths/sum_of_arithmetic_series.py
index 74eef0f18a12..3e381b8c20a8 100644
--- a/maths/sum_of_arithmetic_series.py
+++ b/maths/sum_of_arithmetic_series.py
@@ -1,5 +1,5 @@
# DarkCoder
-def sum_of_series(first_term, common_diff, num_of_terms):
+def sum_of_series(first_term: int, common_diff: int, num_of_terms: int) -> float:
"""
Find the sum of n terms in an arithmetic progression.
@@ -8,9 +8,9 @@ def sum_of_series(first_term, common_diff, num_of_terms):
>>> sum_of_series(1, 10, 100)
49600.0
"""
- sum = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff)
+ total = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff)
# formula for sum of series
- return sum
+ return total
def main():
diff --git a/maths/sum_of_digits.py b/maths/sum_of_digits.py
index 64da00d4634c..d5488bb9e9e0 100644
--- a/maths/sum_of_digits.py
+++ b/maths/sum_of_digits.py
@@ -1,10 +1,6 @@
-from timeit import timeit
-
-
def sum_of_digits(n: int) -> int:
"""
Find the sum of digits of a number.
-
>>> sum_of_digits(12345)
15
>>> sum_of_digits(123)
@@ -14,18 +10,17 @@ def sum_of_digits(n: int) -> int:
>>> sum_of_digits(0)
0
"""
- n = -n if n < 0 else n
+ n = abs(n)
res = 0
while n > 0:
res += n % 10
- n = n // 10
+ n //= 10
return res
def sum_of_digits_recursion(n: int) -> int:
"""
Find the sum of digits of a number using recursion
-
>>> sum_of_digits_recursion(12345)
15
>>> sum_of_digits_recursion(123)
@@ -35,14 +30,13 @@ def sum_of_digits_recursion(n: int) -> int:
>>> sum_of_digits_recursion(0)
0
"""
- n = -n if n < 0 else n
+ n = abs(n)
return n if n < 10 else n % 10 + sum_of_digits(n // 10)
def sum_of_digits_compact(n: int) -> int:
"""
Find the sum of digits of a number
-
>>> sum_of_digits_compact(12345)
15
>>> sum_of_digits_compact(123)
@@ -57,93 +51,24 @@ def sum_of_digits_compact(n: int) -> int:
def benchmark() -> None:
"""
- Benchmark code for comparing 3 functions,
- with 3 different length int values.
+ Benchmark multiple functions, with three different length int values.
"""
- print("\nFor small_num = ", small_num, ":")
- print(
- "> sum_of_digits()",
- "\t\tans =",
- sum_of_digits(small_num),
- "\ttime =",
- timeit("z.sum_of_digits(z.small_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> sum_of_digits_recursion()",
- "\tans =",
- sum_of_digits_recursion(small_num),
- "\ttime =",
- timeit("z.sum_of_digits_recursion(z.small_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> sum_of_digits_compact()",
- "\tans =",
- sum_of_digits_compact(small_num),
- "\ttime =",
- timeit("z.sum_of_digits_compact(z.small_num)", setup="import __main__ as z"),
- "seconds",
- )
+ from collections.abc import Callable
+ from timeit import timeit
- print("\nFor medium_num = ", medium_num, ":")
- print(
- "> sum_of_digits()",
- "\t\tans =",
- sum_of_digits(medium_num),
- "\ttime =",
- timeit("z.sum_of_digits(z.medium_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> sum_of_digits_recursion()",
- "\tans =",
- sum_of_digits_recursion(medium_num),
- "\ttime =",
- timeit("z.sum_of_digits_recursion(z.medium_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> sum_of_digits_compact()",
- "\tans =",
- sum_of_digits_compact(medium_num),
- "\ttime =",
- timeit("z.sum_of_digits_compact(z.medium_num)", setup="import __main__ as z"),
- "seconds",
- )
+ def benchmark_a_function(func: Callable, value: int) -> None:
+ call = f"{func.__name__}({value})"
+ timing = timeit(f"__main__.{call}", setup="import __main__")
+ print(f"{call:56} = {func(value)} -- {timing:.4f} seconds")
- print("\nFor large_num = ", large_num, ":")
- print(
- "> sum_of_digits()",
- "\t\tans =",
- sum_of_digits(large_num),
- "\ttime =",
- timeit("z.sum_of_digits(z.large_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> sum_of_digits_recursion()",
- "\tans =",
- sum_of_digits_recursion(large_num),
- "\ttime =",
- timeit("z.sum_of_digits_recursion(z.large_num)", setup="import __main__ as z"),
- "seconds",
- )
- print(
- "> sum_of_digits_compact()",
- "\tans =",
- sum_of_digits_compact(large_num),
- "\ttime =",
- timeit("z.sum_of_digits_compact(z.large_num)", setup="import __main__ as z"),
- "seconds",
- )
+ for value in (262144, 1125899906842624, 1267650600228229401496703205376):
+ for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact):
+ benchmark_a_function(func, value)
+ print()
if __name__ == "__main__":
- small_num = 262144
- medium_num = 1125899906842624
- large_num = 1267650600228229401496703205376
- benchmark()
import doctest
doctest.testmod()
+ benchmark()
diff --git a/maths/sum_of_geometric_progression.py b/maths/sum_of_geometric_progression.py
index f29dd8005cff..9079f35af6d9 100644
--- a/maths/sum_of_geometric_progression.py
+++ b/maths/sum_of_geometric_progression.py
@@ -25,4 +25,4 @@ def sum_of_geometric_progression(
return num_of_terms * first_term
# Formula for finding sum of n terms of a GeometricProgression
- return (first_term / (1 - common_ratio)) * (1 - common_ratio ** num_of_terms)
+ return (first_term / (1 - common_ratio)) * (1 - common_ratio**num_of_terms)
diff --git a/maths/sum_of_harmonic_series.py b/maths/sum_of_harmonic_series.py
new file mode 100644
index 000000000000..9e0d6b19b95a
--- /dev/null
+++ b/maths/sum_of_harmonic_series.py
@@ -0,0 +1,29 @@
+def sum_of_harmonic_progression(
+ first_term: float, common_difference: float, number_of_terms: int
+) -> float:
+ """
+ https://en.wikipedia.org/wiki/Harmonic_progression_(mathematics)
+
+ Find the sum of n terms in an harmonic progression. The calculation starts with the
+ first_term and loops adding the common difference of Arithmetic Progression by which
+ the given Harmonic Progression is linked.
+
+ >>> sum_of_harmonic_progression(1 / 2, 2, 2)
+ 0.75
+ >>> sum_of_harmonic_progression(1 / 5, 5, 5)
+ 0.45666666666666667
+ """
+ arithmetic_progression = [1 / first_term]
+ first_term = 1 / first_term
+ for _ in range(number_of_terms - 1):
+ first_term += common_difference
+ arithmetic_progression.append(first_term)
+ harmonic_series = [1 / step for step in arithmetic_progression]
+ return sum(harmonic_series)
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(sum_of_harmonic_progression(1 / 2, 2, 2))
diff --git a/maths/sumset.py b/maths/sumset.py
new file mode 100644
index 000000000000..fa18f9e24b4c
--- /dev/null
+++ b/maths/sumset.py
@@ -0,0 +1,37 @@
+"""
+
+Calculates the SumSet of two sets of numbers (A and B)
+
+Source:
+ https://en.wikipedia.org/wiki/Sumset
+
+"""
+
+
+def sumset(set_a: set, set_b: set) -> set:
+ """
+ :param first set: a set of numbers
+ :param second set: a set of numbers
+ :return: the nth number in Sylvester's sequence
+
+ >>> sumset({1, 2, 3}, {4, 5, 6})
+ {5, 6, 7, 8, 9}
+
+ >>> sumset({1, 2, 3}, {4, 5, 6, 7})
+ {5, 6, 7, 8, 9, 10}
+
+ >>> sumset({1, 2, 3, 4}, 3)
+ Traceback (most recent call last):
+ ...
+ AssertionError: The input value of [set_b=3] is not a set
+ """
+ assert isinstance(set_a, set), f"The input value of [set_a={set_a}] is not a set"
+ assert isinstance(set_b, set), f"The input value of [set_b={set_b}] is not a set"
+
+ return {a + b for a in set_a for b in set_b}
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/maths/sylvester_sequence.py b/maths/sylvester_sequence.py
new file mode 100644
index 000000000000..607424c6a90b
--- /dev/null
+++ b/maths/sylvester_sequence.py
@@ -0,0 +1,44 @@
+"""
+
+Calculates the nth number in Sylvester's sequence
+
+Source:
+ https://en.wikipedia.org/wiki/Sylvester%27s_sequence
+
+"""
+
+
+def sylvester(number: int) -> int:
+ """
+ :param number: nth number to calculate in the sequence
+ :return: the nth number in Sylvester's sequence
+
+ >>> sylvester(8)
+ 113423713055421844361000443
+
+ >>> sylvester(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: The input value of [n=-1] has to be > 0
+
+ >>> sylvester(8.0)
+ Traceback (most recent call last):
+ ...
+ AssertionError: The input value of [n=8.0] is not an integer
+ """
+ assert isinstance(number, int), f"The input value of [n={number}] is not an integer"
+
+ if number == 1:
+ return 2
+ elif number < 1:
+ msg = f"The input value of [n={number}] has to be > 0"
+ raise ValueError(msg)
+ else:
+ num = sylvester(number - 1)
+ lower = num - 1
+ upper = num
+ return lower * upper + 1
+
+
+if __name__ == "__main__":
+ print(f"The 8th number in Sylvester's sequence: {sylvester(8)}")
diff --git a/maths/tanh.py b/maths/tanh.py
new file mode 100644
index 000000000000..011d6f17e22b
--- /dev/null
+++ b/maths/tanh.py
@@ -0,0 +1,43 @@
+"""
+This script demonstrates the implementation of the tangent hyperbolic
+or tanh function.
+
+The function takes a vector of K real numbers as input and
+then (e^x - e^(-x))/(e^x + e^(-x)). After through tanh, the
+element of the vector mostly -1 between 1.
+
+Script inspired from its corresponding Wikipedia article
+https://en.wikipedia.org/wiki/Activation_function
+"""
+
+import numpy as np
+
+
+def tangent_hyperbolic(vector: np.ndarray) -> np.ndarray:
+ """
+ Implements the tanh function
+
+ Parameters:
+ vector: np.ndarray
+
+ Returns:
+ tanh (np.array): The input numpy array after applying tanh.
+
+ mathematically (e^x - e^(-x))/(e^x + e^(-x)) can be written as (2/(1+e^(-2x))-1
+
+ Examples:
+ >>> tangent_hyperbolic(np.array([1,5,6,-0.67]))
+ array([ 0.76159416, 0.9999092 , 0.99998771, -0.58497988])
+
+ >>> tangent_hyperbolic(np.array([8,10,2,-0.98,13]))
+ array([ 0.99999977, 1. , 0.96402758, -0.7530659 , 1. ])
+
+ """
+
+ return (2 / (1 + np.exp(-2 * vector))) - 1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/test_prime_check.py b/maths/test_prime_check.py
index b6389684af9e..3ea3b2f1f88b 100644
--- a/maths/test_prime_check.py
+++ b/maths/test_prime_check.py
@@ -1,6 +1,6 @@
"""
Minimalist file that allows pytest to find and run the Test unittest. For details, see:
-http://doc.pytest.org/en/latest/goodpractices.html#conventions-for-python-test-discovery
+https://doc.pytest.org/en/latest/goodpractices.html#conventions-for-python-test-discovery
"""
from .prime_check import Test
diff --git a/maths/three_sum.py b/maths/three_sum.py
new file mode 100644
index 000000000000..09956f8415a0
--- /dev/null
+++ b/maths/three_sum.py
@@ -0,0 +1,47 @@
+"""
+https://en.wikipedia.org/wiki/3SUM
+"""
+
+
+def three_sum(nums: list[int]) -> list[list[int]]:
+ """
+ Find all unique triplets in a sorted array of integers that sum up to zero.
+
+ Args:
+ nums: A sorted list of integers.
+
+ Returns:
+ A list of lists containing unique triplets that sum up to zero.
+
+ >>> three_sum([-1, 0, 1, 2, -1, -4])
+ [[-1, -1, 2], [-1, 0, 1]]
+ >>> three_sum([1, 2, 3, 4])
+ []
+ """
+ nums.sort()
+ ans = []
+ for i in range(len(nums) - 2):
+ if i == 0 or (nums[i] != nums[i - 1]):
+ low, high, c = i + 1, len(nums) - 1, 0 - nums[i]
+ while low < high:
+ if nums[low] + nums[high] == c:
+ ans.append([nums[i], nums[low], nums[high]])
+
+ while low < high and nums[low] == nums[low + 1]:
+ low += 1
+ while low < high and nums[high] == nums[high - 1]:
+ high -= 1
+
+ low += 1
+ high -= 1
+ elif nums[low] + nums[high] < c:
+ low += 1
+ else:
+ high -= 1
+ return ans
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/maths/trapezoidal_rule.py b/maths/trapezoidal_rule.py
index 9a4ddc8af66b..21b10b239b5f 100644
--- a/maths/trapezoidal_rule.py
+++ b/maths/trapezoidal_rule.py
@@ -1,17 +1,26 @@
"""
Numerical integration or quadrature for a smooth function f with known values at x_i
+"""
-This method is the classical approach of suming 'Equally Spaced Abscissas'
-
-method 1:
-"extended trapezoidal rule"
-"""
+def trapezoidal_rule(boundary, steps):
+ """
+ Implements the extended trapezoidal rule for numerical integration.
+ The function f(x) is provided below.
+ :param boundary: List containing the lower and upper bounds of integration [a, b]
+ :param steps: The number of steps (intervals) used in the approximation
+ :return: The numerical approximation of the integral
-def method_1(boundary, steps):
- # "extended trapezoidal rule"
- # int(f) = dx/2 * (f1 + 2f2 + ... + fn)
+ >>> abs(trapezoidal_rule([0, 1], 10) - 0.33333) < 0.01
+ True
+ >>> abs(trapezoidal_rule([0, 1], 100) - 0.33333) < 0.01
+ True
+ >>> abs(trapezoidal_rule([0, 2], 1000) - 2.66667) < 0.01
+ True
+ >>> abs(trapezoidal_rule([1, 2], 1000) - 2.33333) < 0.01
+ True
+ """
h = (boundary[1] - boundary[0]) / steps
a = boundary[0]
b = boundary[1]
@@ -19,32 +28,78 @@ def method_1(boundary, steps):
y = 0.0
y += (h / 2.0) * f(a)
for i in x_i:
- # print(i)
y += h * f(i)
y += (h / 2.0) * f(b)
return y
def make_points(a, b, h):
+ """
+ Generates points between a and b with step size h for trapezoidal integration.
+
+ :param a: The lower bound of integration
+ :param b: The upper bound of integration
+ :param h: The step size
+ :yield: The next x-value in the range (a, b)
+
+ >>> list(make_points(0, 1, 0.1)) # doctest: +NORMALIZE_WHITESPACE
+ [0.1, 0.2, 0.30000000000000004, 0.4, 0.5, 0.6, 0.7, 0.7999999999999999, \
+ 0.8999999999999999]
+ >>> list(make_points(0, 10, 2.5))
+ [2.5, 5.0, 7.5]
+ >>> list(make_points(0, 10, 2))
+ [2, 4, 6, 8]
+ >>> list(make_points(1, 21, 5))
+ [6, 11, 16]
+ >>> list(make_points(1, 5, 2))
+ [3]
+ >>> list(make_points(1, 4, 3))
+ []
+ """
x = a + h
- while x < (b - h):
+ while x <= (b - h):
yield x
- x = x + h
+ x += h
-def f(x): # enter your function here
- y = (x - 0) * (x - 0)
- return y
+def f(x):
+ """
+ This is the function to integrate, f(x) = (x - 0)^2 = x^2.
+
+ :param x: The input value
+ :return: The value of f(x)
+
+ >>> f(0)
+ 0
+ >>> f(1)
+ 1
+ >>> f(0.5)
+ 0.25
+ """
+ return x**2
def main():
- a = 0.0 # Lower bound of integration
- b = 1.0 # Upper bound of integration
- steps = 10.0 # define number of steps or resolution
- boundary = [a, b] # define boundary of integration
- y = method_1(boundary, steps)
+ """
+ Main function to test the trapezoidal rule.
+ :a: Lower bound of integration
+ :b: Upper bound of integration
+ :steps: define number of steps or resolution
+ :boundary: define boundary of integration
+
+ >>> main()
+ y = 0.3349999999999999
+ """
+ a = 0.0
+ b = 1.0
+ steps = 10.0
+ boundary = [a, b]
+ y = trapezoidal_rule(boundary, steps)
print(f"y = {y}")
if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
main()
diff --git a/other/triplet_sum.py b/maths/triplet_sum.py
similarity index 93%
rename from other/triplet_sum.py
rename to maths/triplet_sum.py
index 0e78bb52bb72..e74f67daad47 100644
--- a/other/triplet_sum.py
+++ b/maths/triplet_sum.py
@@ -1,89 +1,90 @@
-"""
-Given an array of integers and another integer target,
-we are required to find a triplet from the array such that it's sum is equal to
-the target.
-"""
-from __future__ import annotations
-
-from itertools import permutations
-from random import randint
-from timeit import repeat
-
-
-def make_dataset() -> tuple[list[int], int]:
- arr = [randint(-1000, 1000) for i in range(10)]
- r = randint(-5000, 5000)
- return (arr, r)
-
-
-dataset = make_dataset()
-
-
-def triplet_sum1(arr: list[int], target: int) -> tuple[int, int, int]:
- """
- Returns a triplet in the array with sum equal to target,
- else (0, 0, 0).
- >>> triplet_sum1([13, 29, 7, 23, 5], 35)
- (5, 7, 23)
- >>> triplet_sum1([37, 9, 19, 50, 44], 65)
- (9, 19, 37)
- >>> arr = [6, 47, 27, 1, 15]
- >>> target = 11
- >>> triplet_sum1(arr, target)
- (0, 0, 0)
- """
- for triplet in permutations(arr, 3):
- if sum(triplet) == target:
- return tuple(sorted(triplet))
- return (0, 0, 0)
-
-
-def triplet_sum2(arr: list[int], target: int) -> tuple[int, int, int]:
- """
- Returns a triplet in the array with sum equal to target,
- else (0, 0, 0).
- >>> triplet_sum2([13, 29, 7, 23, 5], 35)
- (5, 7, 23)
- >>> triplet_sum2([37, 9, 19, 50, 44], 65)
- (9, 19, 37)
- >>> arr = [6, 47, 27, 1, 15]
- >>> target = 11
- >>> triplet_sum2(arr, target)
- (0, 0, 0)
- """
- arr.sort()
- n = len(arr)
- for i in range(n - 1):
- left, right = i + 1, n - 1
- while left < right:
- if arr[i] + arr[left] + arr[right] == target:
- return (arr[i], arr[left], arr[right])
- elif arr[i] + arr[left] + arr[right] < target:
- left += 1
- elif arr[i] + arr[left] + arr[right] > target:
- right -= 1
- return (0, 0, 0)
-
-
-def solution_times() -> tuple[float, float]:
- setup_code = """
-from __main__ import dataset, triplet_sum1, triplet_sum2
-"""
- test_code1 = """
-triplet_sum1(*dataset)
-"""
- test_code2 = """
-triplet_sum2(*dataset)
-"""
- times1 = repeat(setup=setup_code, stmt=test_code1, repeat=5, number=10000)
- times2 = repeat(setup=setup_code, stmt=test_code2, repeat=5, number=10000)
- return (min(times1), min(times2))
-
-
-if __name__ == "__main__":
- from doctest import testmod
-
- testmod()
- times = solution_times()
- print(f"The time for naive implementation is {times[0]}.")
- print(f"The time for optimized implementation is {times[1]}.")
+"""
+Given an array of integers and another integer target,
+we are required to find a triplet from the array such that it's sum is equal to
+the target.
+"""
+
+from __future__ import annotations
+
+from itertools import permutations
+from random import randint
+from timeit import repeat
+
+
+def make_dataset() -> tuple[list[int], int]:
+ arr = [randint(-1000, 1000) for i in range(10)]
+ r = randint(-5000, 5000)
+ return (arr, r)
+
+
+dataset = make_dataset()
+
+
+def triplet_sum1(arr: list[int], target: int) -> tuple[int, ...]:
+ """
+ Returns a triplet in the array with sum equal to target,
+ else (0, 0, 0).
+ >>> triplet_sum1([13, 29, 7, 23, 5], 35)
+ (5, 7, 23)
+ >>> triplet_sum1([37, 9, 19, 50, 44], 65)
+ (9, 19, 37)
+ >>> arr = [6, 47, 27, 1, 15]
+ >>> target = 11
+ >>> triplet_sum1(arr, target)
+ (0, 0, 0)
+ """
+ for triplet in permutations(arr, 3):
+ if sum(triplet) == target:
+ return tuple(sorted(triplet))
+ return (0, 0, 0)
+
+
+def triplet_sum2(arr: list[int], target: int) -> tuple[int, int, int]:
+ """
+ Returns a triplet in the array with sum equal to target,
+ else (0, 0, 0).
+ >>> triplet_sum2([13, 29, 7, 23, 5], 35)
+ (5, 7, 23)
+ >>> triplet_sum2([37, 9, 19, 50, 44], 65)
+ (9, 19, 37)
+ >>> arr = [6, 47, 27, 1, 15]
+ >>> target = 11
+ >>> triplet_sum2(arr, target)
+ (0, 0, 0)
+ """
+ arr.sort()
+ n = len(arr)
+ for i in range(n - 1):
+ left, right = i + 1, n - 1
+ while left < right:
+ if arr[i] + arr[left] + arr[right] == target:
+ return (arr[i], arr[left], arr[right])
+ elif arr[i] + arr[left] + arr[right] < target:
+ left += 1
+ elif arr[i] + arr[left] + arr[right] > target:
+ right -= 1
+ return (0, 0, 0)
+
+
+def solution_times() -> tuple[float, float]:
+ setup_code = """
+from __main__ import dataset, triplet_sum1, triplet_sum2
+"""
+ test_code1 = """
+triplet_sum1(*dataset)
+"""
+ test_code2 = """
+triplet_sum2(*dataset)
+"""
+ times1 = repeat(setup=setup_code, stmt=test_code1, repeat=5, number=10000)
+ times2 = repeat(setup=setup_code, stmt=test_code2, repeat=5, number=10000)
+ return (min(times1), min(times2))
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
+ times = solution_times()
+ print(f"The time for naive implementation is {times[0]}.")
+ print(f"The time for optimized implementation is {times[1]}.")
diff --git a/maths/twin_prime.py b/maths/twin_prime.py
new file mode 100644
index 000000000000..912b10b366c0
--- /dev/null
+++ b/maths/twin_prime.py
@@ -0,0 +1,46 @@
+"""
+== Twin Prime ==
+A number n+2 is said to be a Twin prime of number n if
+both n and n+2 are prime.
+
+Examples of Twin pairs: (3, 5), (5, 7), (11, 13), (17, 19), (29, 31), (41, 43), ...
+https://en.wikipedia.org/wiki/Twin_prime
+"""
+
+# Author : Akshay Dubey (https://github.com/itsAkshayDubey)
+from maths.prime_check import is_prime
+
+
+def twin_prime(number: int) -> int:
+ """
+ # doctest: +NORMALIZE_WHITESPACE
+ This functions takes an integer number as input.
+ returns n+2 if n and n+2 are prime numbers and -1 otherwise.
+ >>> twin_prime(3)
+ 5
+ >>> twin_prime(4)
+ -1
+ >>> twin_prime(5)
+ 7
+ >>> twin_prime(17)
+ 19
+ >>> twin_prime(0)
+ -1
+ >>> twin_prime(6.0)
+ Traceback (most recent call last):
+ ...
+ TypeError: Input value of [number=6.0] must be an integer
+ """
+ if not isinstance(number, int):
+ msg = f"Input value of [number={number}] must be an integer"
+ raise TypeError(msg)
+ if is_prime(number) and is_prime(number + 2):
+ return number + 2
+ else:
+ return -1
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/other/two_pointer.py b/maths/two_pointer.py
similarity index 100%
rename from other/two_pointer.py
rename to maths/two_pointer.py
index ff234cddc9e4..8a6d8eb7aff0 100644
--- a/other/two_pointer.py
+++ b/maths/two_pointer.py
@@ -17,6 +17,7 @@
[1]: https://github.com/TheAlgorithms/Python/blob/master/other/two_sum.py
"""
+
from __future__ import annotations
@@ -43,7 +44,6 @@ def two_pointer(nums: list[int], target: int) -> list[int]:
j = len(nums) - 1
while i < j:
-
if nums[i] + nums[j] == target:
return [i, j]
elif nums[i] + nums[j] < target:
diff --git a/other/two_sum.py b/maths/two_sum.py
similarity index 96%
rename from other/two_sum.py
rename to maths/two_sum.py
index 5209acbc7e44..58c933a5078a 100644
--- a/other/two_sum.py
+++ b/maths/two_sum.py
@@ -11,6 +11,7 @@
Because nums[0] + nums[1] = 2 + 7 = 9,
return [0, 1].
"""
+
from __future__ import annotations
@@ -31,7 +32,7 @@ def two_sum(nums: list[int], target: int) -> list[int]:
>>> two_sum([3 * i for i in range(10)], 19)
[]
"""
- chk_map = {}
+ chk_map: dict[int, int] = {}
for index, val in enumerate(nums):
compl = target - val
if compl in chk_map:
diff --git a/maths/volume.py b/maths/volume.py
index 41d2331db3cb..08bdf72b013b 100644
--- a/maths/volume.py
+++ b/maths/volume.py
@@ -1,13 +1,16 @@
"""
-Find Volumes of Various Shapes.
+Find the volume of various shapes.
-Wikipedia reference: https://en.wikipedia.org/wiki/Volume
+* https://en.wikipedia.org/wiki/Volume
+* https://en.wikipedia.org/wiki/Spherical_cap
"""
-from math import pi, pow
-from typing import Union
+from __future__ import annotations
-def vol_cube(side_length: Union[int, float]) -> float:
+from math import pi, pow # noqa: A004
+
+
+def vol_cube(side_length: float) -> float:
"""
Calculate the Volume of a Cube.
@@ -15,117 +18,549 @@ def vol_cube(side_length: Union[int, float]) -> float:
1.0
>>> vol_cube(3)
27.0
+ >>> vol_cube(0)
+ 0.0
+ >>> vol_cube(1.6)
+ 4.096000000000001
+ >>> vol_cube(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_cube() only accepts non-negative values
"""
+ if side_length < 0:
+ raise ValueError("vol_cube() only accepts non-negative values")
return pow(side_length, 3)
+def vol_spherical_cap(height: float, radius: float) -> float:
+ """
+ Calculate the volume of the spherical cap.
+
+ >>> vol_spherical_cap(1, 2)
+ 5.235987755982988
+ >>> vol_spherical_cap(1.6, 2.6)
+ 16.621119532592402
+ >>> vol_spherical_cap(0, 0)
+ 0.0
+ >>> vol_spherical_cap(-1, 2)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_spherical_cap() only accepts non-negative values
+ >>> vol_spherical_cap(1, -2)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_spherical_cap() only accepts non-negative values
+ """
+ if height < 0 or radius < 0:
+ raise ValueError("vol_spherical_cap() only accepts non-negative values")
+ # Volume is 1/3 pi * height squared * (3 * radius - height)
+ return 1 / 3 * pi * pow(height, 2) * (3 * radius - height)
+
+
+def vol_spheres_intersect(
+ radius_1: float, radius_2: float, centers_distance: float
+) -> float:
+ r"""
+ Calculate the volume of the intersection of two spheres.
+
+ The intersection is composed by two spherical caps and therefore its volume is the
+ sum of the volumes of the spherical caps.
+ First, it calculates the heights :math:`(h_1, h_2)` of the spherical caps,
+ then the two volumes and it returns the sum.
+ The height formulas are
+
+ .. math::
+ h_1 = \frac{(radius_1 - radius_2 + centers\_distance)
+ \cdot (radius_1 + radius_2 - centers\_distance)}
+ {2 \cdot centers\_distance}
+
+ h_2 = \frac{(radius_2 - radius_1 + centers\_distance)
+ \cdot (radius_2 + radius_1 - centers\_distance)}
+ {2 \cdot centers\_distance}
+
+ if `centers_distance` is 0 then it returns the volume of the smallers sphere
+
+ :return: ``vol_spherical_cap`` (:math:`h_1`, :math:`radius_2`)
+ + ``vol_spherical_cap`` (:math:`h_2`, :math:`radius_1`)
+
+ >>> vol_spheres_intersect(2, 2, 1)
+ 21.205750411731103
+ >>> vol_spheres_intersect(2.6, 2.6, 1.6)
+ 40.71504079052372
+ >>> vol_spheres_intersect(0, 0, 0)
+ 0.0
+ >>> vol_spheres_intersect(-2, 2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_spheres_intersect() only accepts non-negative values
+ >>> vol_spheres_intersect(2, -2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_spheres_intersect() only accepts non-negative values
+ >>> vol_spheres_intersect(2, 2, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_spheres_intersect() only accepts non-negative values
+ """
+ if radius_1 < 0 or radius_2 < 0 or centers_distance < 0:
+ raise ValueError("vol_spheres_intersect() only accepts non-negative values")
+ if centers_distance == 0:
+ return vol_sphere(min(radius_1, radius_2))
+
+ h1 = (
+ (radius_1 - radius_2 + centers_distance)
+ * (radius_1 + radius_2 - centers_distance)
+ / (2 * centers_distance)
+ )
+ h2 = (
+ (radius_2 - radius_1 + centers_distance)
+ * (radius_2 + radius_1 - centers_distance)
+ / (2 * centers_distance)
+ )
+
+ return vol_spherical_cap(h1, radius_2) + vol_spherical_cap(h2, radius_1)
+
+
+def vol_spheres_union(
+ radius_1: float, radius_2: float, centers_distance: float
+) -> float:
+ r"""
+ Calculate the volume of the union of two spheres that possibly intersect.
+
+ It is the sum of sphere :math:`A` and sphere :math:`B` minus their intersection.
+ First, it calculates the volumes :math:`(v_1, v_2)` of the spheres,
+ then the volume of the intersection :math:`i` and
+ it returns the sum :math:`v_1 + v_2 - i`.
+ If `centers_distance` is 0 then it returns the volume of the larger sphere
+
+ :return: ``vol_sphere`` (:math:`radius_1`) + ``vol_sphere`` (:math:`radius_2`)
+ - ``vol_spheres_intersect``
+ (:math:`radius_1`, :math:`radius_2`, :math:`centers\_distance`)
+
+ >>> vol_spheres_union(2, 2, 1)
+ 45.814892864851146
+ >>> vol_spheres_union(1.56, 2.2, 1.4)
+ 48.77802773671288
+ >>> vol_spheres_union(0, 2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_spheres_union() only accepts non-negative values, non-zero radius
+ >>> vol_spheres_union('1.56', '2.2', '1.4')
+ Traceback (most recent call last):
+ ...
+ TypeError: '<=' not supported between instances of 'str' and 'int'
+ >>> vol_spheres_union(1, None, 1)
+ Traceback (most recent call last):
+ ...
+ TypeError: '<=' not supported between instances of 'NoneType' and 'int'
+ """
+
+ if radius_1 <= 0 or radius_2 <= 0 or centers_distance < 0:
+ raise ValueError(
+ "vol_spheres_union() only accepts non-negative values, non-zero radius"
+ )
+
+ if centers_distance == 0:
+ return vol_sphere(max(radius_1, radius_2))
+
+ return (
+ vol_sphere(radius_1)
+ + vol_sphere(radius_2)
+ - vol_spheres_intersect(radius_1, radius_2, centers_distance)
+ )
+
+
def vol_cuboid(width: float, height: float, length: float) -> float:
"""
Calculate the Volume of a Cuboid.
- :return multiple of width, length and height
+
+ :return: multiple of `width`, `length` and `height`
>>> vol_cuboid(1, 1, 1)
1.0
>>> vol_cuboid(1, 2, 3)
6.0
+ >>> vol_cuboid(1.6, 2.6, 3.6)
+ 14.976
+ >>> vol_cuboid(0, 0, 0)
+ 0.0
+ >>> vol_cuboid(-1, 2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_cuboid() only accepts non-negative values
+ >>> vol_cuboid(1, -2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_cuboid() only accepts non-negative values
+ >>> vol_cuboid(1, 2, -3)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_cuboid() only accepts non-negative values
"""
+ if width < 0 or height < 0 or length < 0:
+ raise ValueError("vol_cuboid() only accepts non-negative values")
return float(width * height * length)
def vol_cone(area_of_base: float, height: float) -> float:
- """
- Calculate the Volume of a Cone.
+ r"""
+ | Calculate the Volume of a Cone.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Cone
- Wikipedia reference: https://en.wikipedia.org/wiki/Cone
- :return (1/3) * area_of_base * height
+ :return: :math:`\frac{1}{3} \cdot area\_of\_base \cdot height`
>>> vol_cone(10, 3)
10.0
>>> vol_cone(1, 1)
0.3333333333333333
+ >>> vol_cone(1.6, 1.6)
+ 0.8533333333333335
+ >>> vol_cone(0, 0)
+ 0.0
+ >>> vol_cone(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_cone() only accepts non-negative values
+ >>> vol_cone(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_cone() only accepts non-negative values
"""
+ if height < 0 or area_of_base < 0:
+ raise ValueError("vol_cone() only accepts non-negative values")
return area_of_base * height / 3.0
def vol_right_circ_cone(radius: float, height: float) -> float:
- """
- Calculate the Volume of a Right Circular Cone.
+ r"""
+ | Calculate the Volume of a Right Circular Cone.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Cone
- Wikipedia reference: https://en.wikipedia.org/wiki/Cone
- :return (1/3) * pi * radius^2 * height
+ :return: :math:`\frac{1}{3} \cdot \pi \cdot radius^2 \cdot height`
>>> vol_right_circ_cone(2, 3)
12.566370614359172
+ >>> vol_right_circ_cone(0, 0)
+ 0.0
+ >>> vol_right_circ_cone(1.6, 1.6)
+ 4.289321169701265
+ >>> vol_right_circ_cone(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_right_circ_cone() only accepts non-negative values
+ >>> vol_right_circ_cone(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_right_circ_cone() only accepts non-negative values
"""
+ if height < 0 or radius < 0:
+ raise ValueError("vol_right_circ_cone() only accepts non-negative values")
return pi * pow(radius, 2) * height / 3.0
def vol_prism(area_of_base: float, height: float) -> float:
- """
- Calculate the Volume of a Prism.
- Wikipedia reference: https://en.wikipedia.org/wiki/Prism_(geometry)
- :return V = Bh
+ r"""
+ | Calculate the Volume of a Prism.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Prism_(geometry)
+
+ :return: :math:`V = B \cdot h`
>>> vol_prism(10, 2)
20.0
>>> vol_prism(11, 1)
11.0
+ >>> vol_prism(1.6, 1.6)
+ 2.5600000000000005
+ >>> vol_prism(0, 0)
+ 0.0
+ >>> vol_prism(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_prism() only accepts non-negative values
+ >>> vol_prism(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_prism() only accepts non-negative values
"""
+ if height < 0 or area_of_base < 0:
+ raise ValueError("vol_prism() only accepts non-negative values")
return float(area_of_base * height)
def vol_pyramid(area_of_base: float, height: float) -> float:
- """
- Calculate the Volume of a Pyramid.
- Wikipedia reference: https://en.wikipedia.org/wiki/Pyramid_(geometry)
- :return (1/3) * Bh
+ r"""
+ | Calculate the Volume of a Pyramid.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Pyramid_(geometry)
+
+ :return: :math:`\frac{1}{3} \cdot B \cdot h`
>>> vol_pyramid(10, 3)
10.0
>>> vol_pyramid(1.5, 3)
1.5
+ >>> vol_pyramid(1.6, 1.6)
+ 0.8533333333333335
+ >>> vol_pyramid(0, 0)
+ 0.0
+ >>> vol_pyramid(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_pyramid() only accepts non-negative values
+ >>> vol_pyramid(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_pyramid() only accepts non-negative values
"""
+ if height < 0 or area_of_base < 0:
+ raise ValueError("vol_pyramid() only accepts non-negative values")
return area_of_base * height / 3.0
def vol_sphere(radius: float) -> float:
- """
- Calculate the Volume of a Sphere.
- Wikipedia reference: https://en.wikipedia.org/wiki/Sphere
- :return (4/3) * pi * r^3
+ r"""
+ | Calculate the Volume of a Sphere.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Sphere
+
+ :return: :math:`\frac{4}{3} \cdot \pi \cdot r^3`
>>> vol_sphere(5)
523.5987755982989
>>> vol_sphere(1)
4.1887902047863905
+ >>> vol_sphere(1.6)
+ 17.15728467880506
+ >>> vol_sphere(0)
+ 0.0
+ >>> vol_sphere(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_sphere() only accepts non-negative values
"""
+ if radius < 0:
+ raise ValueError("vol_sphere() only accepts non-negative values")
+ # Volume is 4/3 * pi * radius cubed
return 4 / 3 * pi * pow(radius, 3)
+def vol_hemisphere(radius: float) -> float:
+ r"""
+ | Calculate the volume of a hemisphere
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Hemisphere
+ | Other references: https://www.cuemath.com/geometry/hemisphere
+
+ :return: :math:`\frac{2}{3} \cdot \pi \cdot radius^3`
+
+ >>> vol_hemisphere(1)
+ 2.0943951023931953
+ >>> vol_hemisphere(7)
+ 718.377520120866
+ >>> vol_hemisphere(1.6)
+ 8.57864233940253
+ >>> vol_hemisphere(0)
+ 0.0
+ >>> vol_hemisphere(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_hemisphere() only accepts non-negative values
+ """
+ if radius < 0:
+ raise ValueError("vol_hemisphere() only accepts non-negative values")
+ # Volume is radius cubed * pi * 2/3
+ return pow(radius, 3) * pi * 2 / 3
+
+
def vol_circular_cylinder(radius: float, height: float) -> float:
- """Calculate the Volume of a Circular Cylinder.
- Wikipedia reference: https://en.wikipedia.org/wiki/Cylinder
- :return pi * radius^2 * height
+ r"""
+ | Calculate the Volume of a Circular Cylinder.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Cylinder
+
+ :return: :math:`\pi \cdot radius^2 \cdot height`
>>> vol_circular_cylinder(1, 1)
3.141592653589793
>>> vol_circular_cylinder(4, 3)
150.79644737231007
+ >>> vol_circular_cylinder(1.6, 1.6)
+ 12.867963509103795
+ >>> vol_circular_cylinder(0, 0)
+ 0.0
+ >>> vol_circular_cylinder(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_circular_cylinder() only accepts non-negative values
+ >>> vol_circular_cylinder(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_circular_cylinder() only accepts non-negative values
+ """
+ if height < 0 or radius < 0:
+ raise ValueError("vol_circular_cylinder() only accepts non-negative values")
+ # Volume is radius squared * height * pi
+ return pow(radius, 2) * height * pi
+
+
+def vol_hollow_circular_cylinder(
+ inner_radius: float, outer_radius: float, height: float
+) -> float:
+ """
+ Calculate the Volume of a Hollow Circular Cylinder.
+
+ >>> vol_hollow_circular_cylinder(1, 2, 3)
+ 28.274333882308138
+ >>> vol_hollow_circular_cylinder(1.6, 2.6, 3.6)
+ 47.50088092227767
+ >>> vol_hollow_circular_cylinder(-1, 2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_hollow_circular_cylinder() only accepts non-negative values
+ >>> vol_hollow_circular_cylinder(1, -2, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_hollow_circular_cylinder() only accepts non-negative values
+ >>> vol_hollow_circular_cylinder(1, 2, -3)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_hollow_circular_cylinder() only accepts non-negative values
+ >>> vol_hollow_circular_cylinder(2, 1, 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: outer_radius must be greater than inner_radius
+ >>> vol_hollow_circular_cylinder(0, 0, 0)
+ Traceback (most recent call last):
+ ...
+ ValueError: outer_radius must be greater than inner_radius
+ """
+ # Volume - (outer_radius squared - inner_radius squared) * pi * height
+ if inner_radius < 0 or outer_radius < 0 or height < 0:
+ raise ValueError(
+ "vol_hollow_circular_cylinder() only accepts non-negative values"
+ )
+ if outer_radius <= inner_radius:
+ raise ValueError("outer_radius must be greater than inner_radius")
+ return pi * (pow(outer_radius, 2) - pow(inner_radius, 2)) * height
+
+
+def vol_conical_frustum(height: float, radius_1: float, radius_2: float) -> float:
+ """
+ | Calculate the Volume of a Conical Frustum.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Frustum
+
+ >>> vol_conical_frustum(45, 7, 28)
+ 48490.482608158454
+ >>> vol_conical_frustum(1, 1, 2)
+ 7.330382858376184
+ >>> vol_conical_frustum(1.6, 2.6, 3.6)
+ 48.7240076620753
+ >>> vol_conical_frustum(0, 0, 0)
+ 0.0
+ >>> vol_conical_frustum(-2, 2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_conical_frustum() only accepts non-negative values
+ >>> vol_conical_frustum(2, -2, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_conical_frustum() only accepts non-negative values
+ >>> vol_conical_frustum(2, 2, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_conical_frustum() only accepts non-negative values
+ """
+ # Volume is 1/3 * pi * height *
+ # (radius_1 squared + radius_2 squared + radius_1 * radius_2)
+ if radius_1 < 0 or radius_2 < 0 or height < 0:
+ raise ValueError("vol_conical_frustum() only accepts non-negative values")
+ return (
+ 1
+ / 3
+ * pi
+ * height
+ * (pow(radius_1, 2) + pow(radius_2, 2) + radius_1 * radius_2)
+ )
+
+
+def vol_torus(torus_radius: float, tube_radius: float) -> float:
+ r"""
+ | Calculate the Volume of a Torus.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Torus
+
+ :return: :math:`2 \pi^2 \cdot torus\_radius \cdot tube\_radius^2`
+
+ >>> vol_torus(1, 1)
+ 19.739208802178716
+ >>> vol_torus(4, 3)
+ 710.6115168784338
+ >>> vol_torus(3, 4)
+ 947.4820225045784
+ >>> vol_torus(1.6, 1.6)
+ 80.85179925372404
+ >>> vol_torus(0, 0)
+ 0.0
+ >>> vol_torus(-1, 1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_torus() only accepts non-negative values
+ >>> vol_torus(1, -1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_torus() only accepts non-negative values
+ """
+ if torus_radius < 0 or tube_radius < 0:
+ raise ValueError("vol_torus() only accepts non-negative values")
+ return 2 * pow(pi, 2) * torus_radius * pow(tube_radius, 2)
+
+
+def vol_icosahedron(tri_side: float) -> float:
+ """
+ | Calculate the Volume of an Icosahedron.
+ | Wikipedia reference: https://en.wikipedia.org/wiki/Regular_icosahedron
+
+ >>> from math import isclose
+ >>> isclose(vol_icosahedron(2.5), 34.088984228514256)
+ True
+ >>> isclose(vol_icosahedron(10), 2181.694990624912374)
+ True
+ >>> isclose(vol_icosahedron(5), 272.711873828114047)
+ True
+ >>> isclose(vol_icosahedron(3.49), 92.740688412033628)
+ True
+ >>> vol_icosahedron(0)
+ 0.0
+ >>> vol_icosahedron(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_icosahedron() only accepts non-negative values
+ >>> vol_icosahedron(-0.2)
+ Traceback (most recent call last):
+ ...
+ ValueError: vol_icosahedron() only accepts non-negative values
"""
- return pi * pow(radius, 2) * height
+ if tri_side < 0:
+ raise ValueError("vol_icosahedron() only accepts non-negative values")
+ return tri_side**3 * (3 + 5**0.5) * 5 / 12
def main():
"""Print the Results of Various Volume Calculations."""
print("Volumes:")
- print("Cube: " + str(vol_cube(2))) # = 8
- print("Cuboid: " + str(vol_cuboid(2, 2, 2))) # = 8
- print("Cone: " + str(vol_cone(2, 2))) # ~= 1.33
- print("Right Circular Cone: " + str(vol_right_circ_cone(2, 2))) # ~= 8.38
- print("Prism: " + str(vol_prism(2, 2))) # = 4
- print("Pyramid: " + str(vol_pyramid(2, 2))) # ~= 1.33
- print("Sphere: " + str(vol_sphere(2))) # ~= 33.5
- print("Circular Cylinder: " + str(vol_circular_cylinder(2, 2))) # ~= 25.1
+ print(f"Cube: {vol_cube(2) = }") # = 8
+ print(f"Cuboid: {vol_cuboid(2, 2, 2) = }") # = 8
+ print(f"Cone: {vol_cone(2, 2) = }") # ~= 1.33
+ print(f"Right Circular Cone: {vol_right_circ_cone(2, 2) = }") # ~= 8.38
+ print(f"Prism: {vol_prism(2, 2) = }") # = 4
+ print(f"Pyramid: {vol_pyramid(2, 2) = }") # ~= 1.33
+ print(f"Sphere: {vol_sphere(2) = }") # ~= 33.5
+ print(f"Hemisphere: {vol_hemisphere(2) = }") # ~= 16.75
+ print(f"Circular Cylinder: {vol_circular_cylinder(2, 2) = }") # ~= 25.1
+ print(f"Torus: {vol_torus(2, 2) = }") # ~= 157.9
+ print(f"Conical Frustum: {vol_conical_frustum(2, 2, 4) = }") # ~= 58.6
+ print(f"Spherical cap: {vol_spherical_cap(1, 2) = }") # ~= 5.24
+ print(f"Spheres intersetion: {vol_spheres_intersect(2, 2, 1) = }") # ~= 21.21
+ print(f"Spheres union: {vol_spheres_union(2, 2, 1) = }") # ~= 45.81
+ print(
+ f"Hollow Circular Cylinder: {vol_hollow_circular_cylinder(1, 2, 3) = }"
+ ) # ~= 28.3
+ print(f"Icosahedron: {vol_icosahedron(2.5) = }") # ~=34.09
if __name__ == "__main__":
diff --git a/maths/zellers_congruence.py b/maths/zellers_congruence.py
index 2d4a22a0a5ba..b958ed3b8659 100644
--- a/maths/zellers_congruence.py
+++ b/maths/zellers_congruence.py
@@ -3,25 +3,26 @@
def zeller(date_input: str) -> str:
-
"""
- Zellers Congruence Algorithm
- Find the day of the week for nearly any Gregorian or Julian calendar date
+ | Zellers Congruence Algorithm
+ | Find the day of the week for nearly any Gregorian or Julian calendar date
>>> zeller('01-31-2010')
'Your date 01-31-2010, is a Sunday!'
- Validate out of range month
+ Validate out of range month:
+
>>> zeller('13-31-2010')
Traceback (most recent call last):
- ...
+ ...
ValueError: Month must be between 1 - 12
>>> zeller('.2-31-2010')
Traceback (most recent call last):
- ...
+ ...
ValueError: invalid literal for int() with base 10: '.2'
Validate out of range date:
+
>>> zeller('01-33-2010')
Traceback (most recent call last):
...
@@ -32,30 +33,35 @@ def zeller(date_input: str) -> str:
ValueError: invalid literal for int() with base 10: '.4'
Validate second separator:
+
>>> zeller('01-31*2010')
Traceback (most recent call last):
...
ValueError: Date separator must be '-' or '/'
Validate first separator:
+
>>> zeller('01^31-2010')
Traceback (most recent call last):
...
ValueError: Date separator must be '-' or '/'
Validate out of range year:
+
>>> zeller('01-31-8999')
Traceback (most recent call last):
...
ValueError: Year out of range. There has to be some sort of limit...right?
Test null input:
+
>>> zeller()
Traceback (most recent call last):
...
TypeError: zeller() missing 1 required positional argument: 'date_input'
- Test length of date_input:
+ Test length of `date_input`:
+
>>> zeller('')
Traceback (most recent call last):
...
diff --git a/matrix/binary_search_matrix.py b/matrix/binary_search_matrix.py
new file mode 100644
index 000000000000..6f203b7a3484
--- /dev/null
+++ b/matrix/binary_search_matrix.py
@@ -0,0 +1,57 @@
+def binary_search(array: list, lower_bound: int, upper_bound: int, value: int) -> int:
+ """
+ This function carries out Binary search on a 1d array and
+ return -1 if it do not exist
+ array: A 1d sorted array
+ value : the value meant to be searched
+ >>> matrix = [1, 4, 7, 11, 15]
+ >>> binary_search(matrix, 0, len(matrix) - 1, 1)
+ 0
+ >>> binary_search(matrix, 0, len(matrix) - 1, 23)
+ -1
+ """
+
+ r = int((lower_bound + upper_bound) // 2)
+ if array[r] == value:
+ return r
+ if lower_bound >= upper_bound:
+ return -1
+ if array[r] < value:
+ return binary_search(array, r + 1, upper_bound, value)
+ else:
+ return binary_search(array, lower_bound, r - 1, value)
+
+
+def mat_bin_search(value: int, matrix: list) -> list:
+ """
+ This function loops over a 2d matrix and calls binarySearch on
+ the selected 1d array and returns [-1, -1] is it do not exist
+ value : value meant to be searched
+ matrix = a sorted 2d matrix
+ >>> matrix = [[1, 4, 7, 11, 15],
+ ... [2, 5, 8, 12, 19],
+ ... [3, 6, 9, 16, 22],
+ ... [10, 13, 14, 17, 24],
+ ... [18, 21, 23, 26, 30]]
+ >>> target = 1
+ >>> mat_bin_search(target, matrix)
+ [0, 0]
+ >>> target = 34
+ >>> mat_bin_search(target, matrix)
+ [-1, -1]
+ """
+ index = 0
+ if matrix[index][0] == value:
+ return [index, 0]
+ while index < len(matrix) and matrix[index][0] < value:
+ r = binary_search(matrix[index], 0, len(matrix[index]) - 1, value)
+ if r != -1:
+ return [index, r]
+ index += 1
+ return [-1, -1]
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/matrix/count_islands_in_matrix.py b/matrix/count_islands_in_matrix.py
index ad9c67fb8c1b..64c595e8499d 100644
--- a/matrix/count_islands_in_matrix.py
+++ b/matrix/count_islands_in_matrix.py
@@ -3,13 +3,13 @@
# connections.
-class matrix: # Public class to implement a graph
- def __init__(self, row: int, col: int, graph: list):
+class Matrix: # Public class to implement a graph
+ def __init__(self, row: int, col: int, graph: list[list[bool]]) -> None:
self.ROW = row
self.COL = col
self.graph = graph
- def is_safe(self, i, j, visited) -> bool:
+ def is_safe(self, i: int, j: int, visited: list[list[bool]]) -> bool:
return (
0 <= i < self.ROW
and 0 <= j < self.COL
@@ -17,13 +17,14 @@ def is_safe(self, i, j, visited) -> bool:
and self.graph[i][j]
)
- def diffs(self, i, j, visited): # Checking all 8 elements surrounding nth element
- rowNbr = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order
- colNbr = [-1, 0, 1, -1, 1, -1, 0, 1]
+ def diffs(self, i: int, j: int, visited: list[list[bool]]) -> None:
+ # Checking all 8 elements surrounding nth element
+ row_nbr = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order
+ col_nbr = [-1, 0, 1, -1, 1, -1, 0, 1]
visited[i][j] = True # Make those cells visited
for k in range(8):
- if self.is_safe(i + rowNbr[k], j + colNbr[k], visited):
- self.diffs(i + rowNbr[k], j + colNbr[k], visited)
+ if self.is_safe(i + row_nbr[k], j + col_nbr[k], visited):
+ self.diffs(i + row_nbr[k], j + col_nbr[k], visited)
def count_islands(self) -> int: # And finally, count all islands.
visited = [[False for j in range(self.COL)] for i in range(self.ROW)]
diff --git a/matrix/count_negative_numbers_in_sorted_matrix.py b/matrix/count_negative_numbers_in_sorted_matrix.py
new file mode 100644
index 000000000000..2799ff3b45fe
--- /dev/null
+++ b/matrix/count_negative_numbers_in_sorted_matrix.py
@@ -0,0 +1,151 @@
+"""
+Given an matrix of numbers in which all rows and all columns are sorted in decreasing
+order, return the number of negative numbers in grid.
+
+Reference: https://leetcode.com/problems/count-negative-numbers-in-a-sorted-matrix
+"""
+
+
+def generate_large_matrix() -> list[list[int]]:
+ """
+ >>> generate_large_matrix() # doctest: +ELLIPSIS
+ [[1000, ..., -999], [999, ..., -1001], ..., [2, ..., -1998]]
+ """
+ return [list(range(1000 - i, -1000 - i, -1)) for i in range(1000)]
+
+
+grid = generate_large_matrix()
+test_grids = (
+ [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]],
+ [[3, 2], [1, 0]],
+ [[7, 7, 6]],
+ [[7, 7, 6], [-1, -2, -3]],
+ grid,
+)
+
+
+def validate_grid(grid: list[list[int]]) -> None:
+ """
+ Validate that the rows and columns of the grid is sorted in decreasing order.
+ >>> for grid in test_grids:
+ ... validate_grid(grid)
+ """
+ assert all(row == sorted(row, reverse=True) for row in grid)
+ assert all(list(col) == sorted(col, reverse=True) for col in zip(*grid))
+
+
+def find_negative_index(array: list[int]) -> int:
+ """
+ Find the smallest negative index
+
+ >>> find_negative_index([0,0,0,0])
+ 4
+ >>> find_negative_index([4,3,2,-1])
+ 3
+ >>> find_negative_index([1,0,-1,-10])
+ 2
+ >>> find_negative_index([0,0,0,-1])
+ 3
+ >>> find_negative_index([11,8,7,-3,-5,-9])
+ 3
+ >>> find_negative_index([-1,-1,-2,-3])
+ 0
+ >>> find_negative_index([5,1,0])
+ 3
+ >>> find_negative_index([-5,-5,-5])
+ 0
+ >>> find_negative_index([0])
+ 1
+ >>> find_negative_index([])
+ 0
+ """
+ left = 0
+ right = len(array) - 1
+
+ # Edge cases such as no values or all numbers are negative.
+ if not array or array[0] < 0:
+ return 0
+
+ while right + 1 > left:
+ mid = (left + right) // 2
+ num = array[mid]
+
+ # Num must be negative and the index must be greater than or equal to 0.
+ if num < 0 and array[mid - 1] >= 0:
+ return mid
+
+ if num >= 0:
+ left = mid + 1
+ else:
+ right = mid - 1
+ # No negative numbers so return the last index of the array + 1 which is the length.
+ return len(array)
+
+
+def count_negatives_binary_search(grid: list[list[int]]) -> int:
+ """
+ An O(m logn) solution that uses binary search in order to find the boundary between
+ positive and negative numbers
+
+ >>> [count_negatives_binary_search(grid) for grid in test_grids]
+ [8, 0, 0, 3, 1498500]
+ """
+ total = 0
+ bound = len(grid[0])
+
+ for i in range(len(grid)):
+ bound = find_negative_index(grid[i][:bound])
+ total += bound
+ return (len(grid) * len(grid[0])) - total
+
+
+def count_negatives_brute_force(grid: list[list[int]]) -> int:
+ """
+ This solution is O(n^2) because it iterates through every column and row.
+
+ >>> [count_negatives_brute_force(grid) for grid in test_grids]
+ [8, 0, 0, 3, 1498500]
+ """
+ return len([number for row in grid for number in row if number < 0])
+
+
+def count_negatives_brute_force_with_break(grid: list[list[int]]) -> int:
+ """
+ Similar to the brute force solution above but uses break in order to reduce the
+ number of iterations.
+
+ >>> [count_negatives_brute_force_with_break(grid) for grid in test_grids]
+ [8, 0, 0, 3, 1498500]
+ """
+ total = 0
+ for row in grid:
+ for i, number in enumerate(row):
+ if number < 0:
+ total += len(row) - i
+ break
+ return total
+
+
+def benchmark() -> None:
+ """Benchmark our functions next to each other"""
+ from timeit import timeit
+
+ print("Running benchmarks")
+ setup = (
+ "from __main__ import count_negatives_binary_search, "
+ "count_negatives_brute_force, count_negatives_brute_force_with_break, grid"
+ )
+ for func in (
+ "count_negatives_binary_search", # took 0.7727 seconds
+ "count_negatives_brute_force_with_break", # took 4.6505 seconds
+ "count_negatives_brute_force", # took 12.8160 seconds
+ ):
+ time = timeit(f"{func}(grid=grid)", setup=setup, number=500)
+ print(f"{func}() took {time:0.4f} seconds")
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ benchmark()
diff --git a/matrix/count_paths.py b/matrix/count_paths.py
new file mode 100644
index 000000000000..4861ad5fd0aa
--- /dev/null
+++ b/matrix/count_paths.py
@@ -0,0 +1,75 @@
+"""
+Given a grid, where you start from the top left position [0, 0],
+you want to find how many paths you can take to get to the bottom right position.
+
+start here -> 0 0 0 0
+ 1 1 0 0
+ 0 0 0 1
+ 0 1 0 0 <- finish here
+how many 'distinct' paths can you take to get to the finish?
+Using a recursive depth-first search algorithm below, you are able to
+find the number of distinct unique paths (count).
+
+'*' will demonstrate a path
+In the example above, there are two distinct paths:
+1. 2.
+ * * * 0 * * * *
+ 1 1 * 0 1 1 * *
+ 0 0 * 1 0 0 * 1
+ 0 1 * * 0 1 * *
+"""
+
+
+def depth_first_search(grid: list[list[int]], row: int, col: int, visit: set) -> int:
+ """
+ Recursive Backtracking Depth First Search Algorithm
+
+ Starting from top left of a matrix, count the number of
+ paths that can reach the bottom right of a matrix.
+ 1 represents a block (inaccessible)
+ 0 represents a valid space (accessible)
+
+ 0 0 0 0
+ 1 1 0 0
+ 0 0 0 1
+ 0 1 0 0
+ >>> grid = [[0, 0, 0, 0], [1, 1, 0, 0], [0, 0, 0, 1], [0, 1, 0, 0]]
+ >>> depth_first_search(grid, 0, 0, set())
+ 2
+
+ 0 0 0 0 0
+ 0 1 1 1 0
+ 0 1 1 1 0
+ 0 0 0 0 0
+ >>> grid = [[0, 0, 0, 0, 0], [0, 1, 1, 1, 0], [0, 1, 1, 1, 0], [0, 0, 0, 0, 0]]
+ >>> depth_first_search(grid, 0, 0, set())
+ 2
+ """
+ row_length, col_length = len(grid), len(grid[0])
+ if (
+ min(row, col) < 0
+ or row == row_length
+ or col == col_length
+ or (row, col) in visit
+ or grid[row][col] == 1
+ ):
+ return 0
+ if row == row_length - 1 and col == col_length - 1:
+ return 1
+
+ visit.add((row, col))
+
+ count = 0
+ count += depth_first_search(grid, row + 1, col, visit)
+ count += depth_first_search(grid, row - 1, col, visit)
+ count += depth_first_search(grid, row, col + 1, visit)
+ count += depth_first_search(grid, row, col - 1, visit)
+
+ visit.remove((row, col))
+ return count
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/matrix/cramers_rule_2x2.py b/matrix/cramers_rule_2x2.py
new file mode 100644
index 000000000000..081035bec002
--- /dev/null
+++ b/matrix/cramers_rule_2x2.py
@@ -0,0 +1,83 @@
+# https://www.chilimath.com/lessons/advanced-algebra/cramers-rule-with-two-variables
+# https://en.wikipedia.org/wiki/Cramer%27s_rule
+
+
+def cramers_rule_2x2(equation1: list[int], equation2: list[int]) -> tuple[float, float]:
+ """
+ Solves the system of linear equation in 2 variables.
+ :param: equation1: list of 3 numbers
+ :param: equation2: list of 3 numbers
+ :return: String of result
+ input format : [a1, b1, d1], [a2, b2, d2]
+ determinant = [[a1, b1], [a2, b2]]
+ determinant_x = [[d1, b1], [d2, b2]]
+ determinant_y = [[a1, d1], [a2, d2]]
+
+ >>> cramers_rule_2x2([2, 3, 0], [5, 1, 0])
+ (0.0, 0.0)
+ >>> cramers_rule_2x2([0, 4, 50], [2, 0, 26])
+ (13.0, 12.5)
+ >>> cramers_rule_2x2([11, 2, 30], [1, 0, 4])
+ (4.0, -7.0)
+ >>> cramers_rule_2x2([4, 7, 1], [1, 2, 0])
+ (2.0, -1.0)
+
+ >>> cramers_rule_2x2([1, 2, 3], [2, 4, 6])
+ Traceback (most recent call last):
+ ...
+ ValueError: Infinite solutions. (Consistent system)
+ >>> cramers_rule_2x2([1, 2, 3], [2, 4, 7])
+ Traceback (most recent call last):
+ ...
+ ValueError: No solution. (Inconsistent system)
+ >>> cramers_rule_2x2([1, 2, 3], [11, 22])
+ Traceback (most recent call last):
+ ...
+ ValueError: Please enter a valid equation.
+ >>> cramers_rule_2x2([0, 1, 6], [0, 0, 3])
+ Traceback (most recent call last):
+ ...
+ ValueError: No solution. (Inconsistent system)
+ >>> cramers_rule_2x2([0, 0, 6], [0, 0, 3])
+ Traceback (most recent call last):
+ ...
+ ValueError: Both a & b of two equations can't be zero.
+ >>> cramers_rule_2x2([1, 2, 3], [1, 2, 3])
+ Traceback (most recent call last):
+ ...
+ ValueError: Infinite solutions. (Consistent system)
+ >>> cramers_rule_2x2([0, 4, 50], [0, 3, 99])
+ Traceback (most recent call last):
+ ...
+ ValueError: No solution. (Inconsistent system)
+ """
+
+ # Check if the input is valid
+ if not len(equation1) == len(equation2) == 3:
+ raise ValueError("Please enter a valid equation.")
+ if equation1[0] == equation1[1] == equation2[0] == equation2[1] == 0:
+ raise ValueError("Both a & b of two equations can't be zero.")
+
+ # Extract the coefficients
+ a1, b1, c1 = equation1
+ a2, b2, c2 = equation2
+
+ # Calculate the determinants of the matrices
+ determinant = a1 * b2 - a2 * b1
+ determinant_x = c1 * b2 - c2 * b1
+ determinant_y = a1 * c2 - a2 * c1
+
+ # Check if the system of linear equations has a solution (using Cramer's rule)
+ if determinant == 0:
+ if determinant_x == determinant_y == 0:
+ raise ValueError("Infinite solutions. (Consistent system)")
+ else:
+ raise ValueError("No solution. (Inconsistent system)")
+ elif determinant_x == determinant_y == 0:
+ # Trivial solution (Inconsistent system)
+ return (0.0, 0.0)
+ else:
+ x = determinant_x / determinant
+ y = determinant_y / determinant
+ # Non-Trivial Solution (Consistent system)
+ return (x, y)
diff --git a/matrix/inverse_of_matrix.py b/matrix/inverse_of_matrix.py
index 9deca6c3c08e..e53d90df8253 100644
--- a/matrix/inverse_of_matrix.py
+++ b/matrix/inverse_of_matrix.py
@@ -2,22 +2,25 @@
from decimal import Decimal
+from numpy import array
+
def inverse_of_matrix(matrix: list[list[float]]) -> list[list[float]]:
"""
A matrix multiplied with its inverse gives the identity matrix.
- This function finds the inverse of a 2x2 matrix.
+ This function finds the inverse of a 2x2 and 3x3 matrix.
If the determinant of a matrix is 0, its inverse does not exist.
Sources for fixing inaccurate float arithmetic:
https://stackoverflow.com/questions/6563058/how-do-i-use-accurate-float-arithmetic-in-python
https://docs.python.org/3/library/decimal.html
+ Doctests for 2x2
>>> inverse_of_matrix([[2, 5], [2, 0]])
[[0.0, 0.5], [0.2, -0.2]]
>>> inverse_of_matrix([[2.5, 5], [1, 2]])
Traceback (most recent call last):
- ...
+ ...
ValueError: This matrix has no inverse.
>>> inverse_of_matrix([[12, -16], [-9, 0]])
[[0.0, -0.1111111111111111], [-0.0625, -0.08333333333333333]]
@@ -25,16 +28,128 @@ def inverse_of_matrix(matrix: list[list[float]]) -> list[list[float]]:
[[0.16666666666666666, -0.0625], [-0.3333333333333333, 0.25]]
>>> inverse_of_matrix([[10, 5], [3, 2.5]])
[[0.25, -0.5], [-0.3, 1.0]]
+
+ Doctests for 3x3
+ >>> inverse_of_matrix([[2, 5, 7], [2, 0, 1], [1, 2, 3]])
+ [[2.0, 5.0, -4.0], [1.0, 1.0, -1.0], [-5.0, -12.0, 10.0]]
+ >>> inverse_of_matrix([[1, 2, 2], [1, 2, 2], [3, 2, -1]])
+ Traceback (most recent call last):
+ ...
+ ValueError: This matrix has no inverse.
+
+ >>> inverse_of_matrix([[],[]])
+ Traceback (most recent call last):
+ ...
+ ValueError: Please provide a matrix of size 2x2 or 3x3.
+
+ >>> inverse_of_matrix([[1, 2], [3, 4], [5, 6]])
+ Traceback (most recent call last):
+ ...
+ ValueError: Please provide a matrix of size 2x2 or 3x3.
+
+ >>> inverse_of_matrix([[1, 2, 1], [0,3, 4]])
+ Traceback (most recent call last):
+ ...
+ ValueError: Please provide a matrix of size 2x2 or 3x3.
+
+ >>> inverse_of_matrix([[1, 2, 3], [7, 8, 9], [7, 8, 9]])
+ Traceback (most recent call last):
+ ...
+ ValueError: This matrix has no inverse.
+
+ >>> inverse_of_matrix([[1, 0, 0], [0, 1, 0], [0, 0, 1]])
+ [[1.0, 0.0, 0.0], [0.0, 1.0, 0.0], [0.0, 0.0, 1.0]]
"""
- D = Decimal # An abbreviation to be conciseness
- # Calculate the determinant of the matrix
- determinant = D(matrix[0][0]) * D(matrix[1][1]) - D(matrix[1][0]) * D(matrix[0][1])
- if determinant == 0:
- raise ValueError("This matrix has no inverse.")
- # Creates a copy of the matrix with swapped positions of the elements
- swapped_matrix = [[0.0, 0.0], [0.0, 0.0]]
- swapped_matrix[0][0], swapped_matrix[1][1] = matrix[1][1], matrix[0][0]
- swapped_matrix[1][0], swapped_matrix[0][1] = -matrix[1][0], -matrix[0][1]
- # Calculate the inverse of the matrix
- return [[float(D(n) / determinant) or 0.0 for n in row] for row in swapped_matrix]
+ d = Decimal
+
+ # Check if the provided matrix has 2 rows and 2 columns
+ # since this implementation only works for 2x2 matrices
+ if len(matrix) == 2 and len(matrix[0]) == 2 and len(matrix[1]) == 2:
+ # Calculate the determinant of the matrix
+ determinant = float(
+ d(matrix[0][0]) * d(matrix[1][1]) - d(matrix[1][0]) * d(matrix[0][1])
+ )
+ if determinant == 0:
+ raise ValueError("This matrix has no inverse.")
+
+ # Creates a copy of the matrix with swapped positions of the elements
+ swapped_matrix = [[0.0, 0.0], [0.0, 0.0]]
+ swapped_matrix[0][0], swapped_matrix[1][1] = matrix[1][1], matrix[0][0]
+ swapped_matrix[1][0], swapped_matrix[0][1] = -matrix[1][0], -matrix[0][1]
+
+ # Calculate the inverse of the matrix
+ return [
+ [(float(d(n)) / determinant) or 0.0 for n in row] for row in swapped_matrix
+ ]
+ elif (
+ len(matrix) == 3
+ and len(matrix[0]) == 3
+ and len(matrix[1]) == 3
+ and len(matrix[2]) == 3
+ ):
+ # Calculate the determinant of the matrix using Sarrus rule
+ determinant = float(
+ (
+ (d(matrix[0][0]) * d(matrix[1][1]) * d(matrix[2][2]))
+ + (d(matrix[0][1]) * d(matrix[1][2]) * d(matrix[2][0]))
+ + (d(matrix[0][2]) * d(matrix[1][0]) * d(matrix[2][1]))
+ )
+ - (
+ (d(matrix[0][2]) * d(matrix[1][1]) * d(matrix[2][0]))
+ + (d(matrix[0][1]) * d(matrix[1][0]) * d(matrix[2][2]))
+ + (d(matrix[0][0]) * d(matrix[1][2]) * d(matrix[2][1]))
+ )
+ )
+ if determinant == 0:
+ raise ValueError("This matrix has no inverse.")
+
+ # Creating cofactor matrix
+ cofactor_matrix = [
+ [d(0.0), d(0.0), d(0.0)],
+ [d(0.0), d(0.0), d(0.0)],
+ [d(0.0), d(0.0), d(0.0)],
+ ]
+ cofactor_matrix[0][0] = (d(matrix[1][1]) * d(matrix[2][2])) - (
+ d(matrix[1][2]) * d(matrix[2][1])
+ )
+ cofactor_matrix[0][1] = -(
+ (d(matrix[1][0]) * d(matrix[2][2])) - (d(matrix[1][2]) * d(matrix[2][0]))
+ )
+ cofactor_matrix[0][2] = (d(matrix[1][0]) * d(matrix[2][1])) - (
+ d(matrix[1][1]) * d(matrix[2][0])
+ )
+ cofactor_matrix[1][0] = -(
+ (d(matrix[0][1]) * d(matrix[2][2])) - (d(matrix[0][2]) * d(matrix[2][1]))
+ )
+ cofactor_matrix[1][1] = (d(matrix[0][0]) * d(matrix[2][2])) - (
+ d(matrix[0][2]) * d(matrix[2][0])
+ )
+ cofactor_matrix[1][2] = -(
+ (d(matrix[0][0]) * d(matrix[2][1])) - (d(matrix[0][1]) * d(matrix[2][0]))
+ )
+ cofactor_matrix[2][0] = (d(matrix[0][1]) * d(matrix[1][2])) - (
+ d(matrix[0][2]) * d(matrix[1][1])
+ )
+ cofactor_matrix[2][1] = -(
+ (d(matrix[0][0]) * d(matrix[1][2])) - (d(matrix[0][2]) * d(matrix[1][0]))
+ )
+ cofactor_matrix[2][2] = (d(matrix[0][0]) * d(matrix[1][1])) - (
+ d(matrix[0][1]) * d(matrix[1][0])
+ )
+
+ # Transpose the cofactor matrix (Adjoint matrix)
+ adjoint_matrix = array(cofactor_matrix)
+ for i in range(3):
+ for j in range(3):
+ adjoint_matrix[i][j] = cofactor_matrix[j][i]
+
+ # Inverse of the matrix using the formula (1/determinant) * adjoint matrix
+ inverse_matrix = array(cofactor_matrix)
+ for i in range(3):
+ for j in range(3):
+ inverse_matrix[i][j] /= d(determinant)
+
+ # Calculate the inverse of the matrix
+ return [[float(d(n)) or 0.0 for n in row] for row in inverse_matrix]
+ raise ValueError("Please provide a matrix of size 2x2 or 3x3.")
diff --git a/matrix/largest_square_area_in_matrix.py b/matrix/largest_square_area_in_matrix.py
new file mode 100644
index 000000000000..16263fb798f1
--- /dev/null
+++ b/matrix/largest_square_area_in_matrix.py
@@ -0,0 +1,188 @@
+"""
+Question:
+Given a binary matrix mat of size n * m, find out the maximum size square
+sub-matrix with all 1s.
+
+---
+Example 1:
+
+Input:
+n = 2, m = 2
+mat = [[1, 1],
+ [1, 1]]
+
+Output:
+2
+
+Explanation: The maximum size of the square
+sub-matrix is 2. The matrix itself is the
+maximum sized sub-matrix in this case.
+---
+Example 2
+
+Input:
+n = 2, m = 2
+mat = [[0, 0],
+ [0, 0]]
+Output: 0
+
+Explanation: There is no 1 in the matrix.
+
+
+Approach:
+We initialize another matrix (dp) with the same dimensions
+as the original one initialized with all 0's.
+
+dp_array(i,j) represents the side length of the maximum square whose
+bottom right corner is the cell with index (i,j) in the original matrix.
+
+Starting from index (0,0), for every 1 found in the original matrix,
+we update the value of the current element as
+
+dp_array(i,j)=dp_array(dp(i-1,j),dp_array(i-1,j-1),dp_array(i,j-1)) + 1.
+"""
+
+
+def largest_square_area_in_matrix_top_down_approch(
+ rows: int, cols: int, mat: list[list[int]]
+) -> int:
+ """
+ Function updates the largest_square_area[0], if recursive call found
+ square with maximum area.
+
+ We aren't using dp_array here, so the time complexity would be exponential.
+
+ >>> largest_square_area_in_matrix_top_down_approch(2, 2, [[1,1], [1,1]])
+ 2
+ >>> largest_square_area_in_matrix_top_down_approch(2, 2, [[0,0], [0,0]])
+ 0
+ """
+
+ def update_area_of_max_square(row: int, col: int) -> int:
+ # BASE CASE
+ if row >= rows or col >= cols:
+ return 0
+
+ right = update_area_of_max_square(row, col + 1)
+ diagonal = update_area_of_max_square(row + 1, col + 1)
+ down = update_area_of_max_square(row + 1, col)
+
+ if mat[row][col]:
+ sub_problem_sol = 1 + min([right, diagonal, down])
+ largest_square_area[0] = max(largest_square_area[0], sub_problem_sol)
+ return sub_problem_sol
+ else:
+ return 0
+
+ largest_square_area = [0]
+ update_area_of_max_square(0, 0)
+ return largest_square_area[0]
+
+
+def largest_square_area_in_matrix_top_down_approch_with_dp(
+ rows: int, cols: int, mat: list[list[int]]
+) -> int:
+ """
+ Function updates the largest_square_area[0], if recursive call found
+ square with maximum area.
+
+ We are using dp_array here, so the time complexity would be O(N^2).
+
+ >>> largest_square_area_in_matrix_top_down_approch_with_dp(2, 2, [[1,1], [1,1]])
+ 2
+ >>> largest_square_area_in_matrix_top_down_approch_with_dp(2, 2, [[0,0], [0,0]])
+ 0
+ """
+
+ def update_area_of_max_square_using_dp_array(
+ row: int, col: int, dp_array: list[list[int]]
+ ) -> int:
+ if row >= rows or col >= cols:
+ return 0
+ if dp_array[row][col] != -1:
+ return dp_array[row][col]
+
+ right = update_area_of_max_square_using_dp_array(row, col + 1, dp_array)
+ diagonal = update_area_of_max_square_using_dp_array(row + 1, col + 1, dp_array)
+ down = update_area_of_max_square_using_dp_array(row + 1, col, dp_array)
+
+ if mat[row][col]:
+ sub_problem_sol = 1 + min([right, diagonal, down])
+ largest_square_area[0] = max(largest_square_area[0], sub_problem_sol)
+ dp_array[row][col] = sub_problem_sol
+ return sub_problem_sol
+ else:
+ return 0
+
+ largest_square_area = [0]
+ dp_array = [[-1] * cols for _ in range(rows)]
+ update_area_of_max_square_using_dp_array(0, 0, dp_array)
+
+ return largest_square_area[0]
+
+
+def largest_square_area_in_matrix_bottom_up(
+ rows: int, cols: int, mat: list[list[int]]
+) -> int:
+ """
+ Function updates the largest_square_area, using bottom up approach.
+
+ >>> largest_square_area_in_matrix_bottom_up(2, 2, [[1,1], [1,1]])
+ 2
+ >>> largest_square_area_in_matrix_bottom_up(2, 2, [[0,0], [0,0]])
+ 0
+
+ """
+ dp_array = [[0] * (cols + 1) for _ in range(rows + 1)]
+ largest_square_area = 0
+ for row in range(rows - 1, -1, -1):
+ for col in range(cols - 1, -1, -1):
+ right = dp_array[row][col + 1]
+ diagonal = dp_array[row + 1][col + 1]
+ bottom = dp_array[row + 1][col]
+
+ if mat[row][col] == 1:
+ dp_array[row][col] = 1 + min(right, diagonal, bottom)
+ largest_square_area = max(dp_array[row][col], largest_square_area)
+ else:
+ dp_array[row][col] = 0
+
+ return largest_square_area
+
+
+def largest_square_area_in_matrix_bottom_up_space_optimization(
+ rows: int, cols: int, mat: list[list[int]]
+) -> int:
+ """
+ Function updates the largest_square_area, using bottom up
+ approach. with space optimization.
+
+ >>> largest_square_area_in_matrix_bottom_up_space_optimization(2, 2, [[1,1], [1,1]])
+ 2
+ >>> largest_square_area_in_matrix_bottom_up_space_optimization(2, 2, [[0,0], [0,0]])
+ 0
+ """
+ current_row = [0] * (cols + 1)
+ next_row = [0] * (cols + 1)
+ largest_square_area = 0
+ for row in range(rows - 1, -1, -1):
+ for col in range(cols - 1, -1, -1):
+ right = current_row[col + 1]
+ diagonal = next_row[col + 1]
+ bottom = next_row[col]
+
+ if mat[row][col] == 1:
+ current_row[col] = 1 + min(right, diagonal, bottom)
+ largest_square_area = max(current_row[col], largest_square_area)
+ else:
+ current_row[col] = 0
+ next_row = current_row
+
+ return largest_square_area
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
+ print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
diff --git a/matrix/matrix_based_game.py b/matrix/matrix_based_game.py
new file mode 100644
index 000000000000..6181086c6704
--- /dev/null
+++ b/matrix/matrix_based_game.py
@@ -0,0 +1,284 @@
+"""
+Matrix-Based Game Script
+=========================
+This script implements a matrix-based game where players interact with a grid of
+elements. The primary goals are to:
+- Identify connected elements of the same type from a selected position.
+- Remove those elements, adjust the matrix by simulating gravity, and reorganize empty
+ columns.
+- Calculate and display the score based on the number of elements removed in each move.
+
+Functions:
+-----------
+1. `find_repeat`: Finds all connected elements of the same type.
+2. `increment_score`: Calculates the score for a given move.
+3. `move_x`: Simulates gravity in a column.
+4. `move_y`: Reorganizes the matrix by shifting columns leftward when a column becomes
+ empty.
+5. `play`: Executes a single move, updating the matrix and returning the score.
+
+Input Format:
+--------------
+1. Matrix size (`lines`): Integer specifying the size of the matrix (N x N).
+2. Matrix content (`matrix`): Rows of the matrix, each consisting of characters.
+3. Number of moves (`movs`): Integer indicating the number of moves.
+4. List of moves (`movements`): A comma-separated string of coordinates for each move.
+
+(0,0) position starts from first left column to last right, and below row to up row
+
+
+Example Input:
+---------------
+4
+RRBG
+RBBG
+YYGG
+XYGG
+2
+0 1,1 1
+
+Example (0,0) = X
+
+Output:
+--------
+The script outputs the total score after processing all moves.
+
+Usage:
+-------
+Run the script and provide the required inputs as prompted.
+
+"""
+
+
+def validate_matrix_size(size: int) -> None:
+ """
+ >>> validate_matrix_size(-1)
+ Traceback (most recent call last):
+ ...
+ ValueError: Matrix size must be a positive integer.
+ """
+ if not isinstance(size, int) or size <= 0:
+ raise ValueError("Matrix size must be a positive integer.")
+
+
+def validate_matrix_content(matrix: list[str], size: int) -> None:
+ """
+ Validates that the number of elements in the matrix matches the given size.
+
+ >>> validate_matrix_content(['aaaa', 'aaaa', 'aaaa', 'aaaa'], 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: The matrix dont match with size.
+ >>> validate_matrix_content(['aa%', 'aaa', 'aaa'], 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Matrix rows can only contain letters and numbers.
+ >>> validate_matrix_content(['aaa', 'aaa', 'aaaa'], 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Each row in the matrix must have exactly 3 characters.
+ """
+ print(matrix)
+ if len(matrix) != size:
+ raise ValueError("The matrix dont match with size.")
+ for row in matrix:
+ if len(row) != size:
+ msg = f"Each row in the matrix must have exactly {size} characters."
+ raise ValueError(msg)
+ if not all(char.isalnum() for char in row):
+ raise ValueError("Matrix rows can only contain letters and numbers.")
+
+
+def validate_moves(moves: list[tuple[int, int]], size: int) -> None:
+ """
+ >>> validate_moves([(1, 2), (-1, 0)], 3)
+ Traceback (most recent call last):
+ ...
+ ValueError: Move is out of bounds for a matrix.
+ """
+ for move in moves:
+ x, y = move
+ if not (0 <= x < size and 0 <= y < size):
+ raise ValueError("Move is out of bounds for a matrix.")
+
+
+def parse_moves(input_str: str) -> list[tuple[int, int]]:
+ """
+ >>> parse_moves("0 1, 1 1")
+ [(0, 1), (1, 1)]
+ >>> parse_moves("0 1, 1 1, 2")
+ Traceback (most recent call last):
+ ...
+ ValueError: Each move must have exactly two numbers.
+ >>> parse_moves("0 1, 1 1, 2 4 5 6")
+ Traceback (most recent call last):
+ ...
+ ValueError: Each move must have exactly two numbers.
+ """
+ moves = []
+ for pair in input_str.split(","):
+ parts = pair.strip().split()
+ if len(parts) != 2:
+ raise ValueError("Each move must have exactly two numbers.")
+ x, y = map(int, parts)
+ moves.append((x, y))
+ return moves
+
+
+def find_repeat(
+ matrix_g: list[list[str]], row: int, column: int, size: int
+) -> set[tuple[int, int]]:
+ """
+ Finds all connected elements of the same type from a given position.
+
+ >>> find_repeat([['A', 'B', 'A'], ['A', 'B', 'A'], ['A', 'A', 'A']], 0, 0, 3)
+ {(1, 2), (2, 1), (0, 0), (2, 0), (0, 2), (2, 2), (1, 0)}
+ >>> find_repeat([['-', '-', '-'], ['-', '-', '-'], ['-', '-', '-']], 1, 1, 3)
+ set()
+ """
+
+ column = size - 1 - column
+ visited = set()
+ repeated = set()
+
+ if (color := matrix_g[column][row]) != "-":
+
+ def dfs(row_n: int, column_n: int) -> None:
+ if row_n < 0 or row_n >= size or column_n < 0 or column_n >= size:
+ return
+ if (row_n, column_n) in visited:
+ return
+ visited.add((row_n, column_n))
+ if matrix_g[row_n][column_n] == color:
+ repeated.add((row_n, column_n))
+ dfs(row_n - 1, column_n)
+ dfs(row_n + 1, column_n)
+ dfs(row_n, column_n - 1)
+ dfs(row_n, column_n + 1)
+
+ dfs(column, row)
+
+ return repeated
+
+
+def increment_score(count: int) -> int:
+ """
+ Calculates the score for a move based on the number of elements removed.
+
+ >>> increment_score(3)
+ 6
+ >>> increment_score(0)
+ 0
+ """
+ return int(count * (count + 1) / 2)
+
+
+def move_x(matrix_g: list[list[str]], column: int, size: int) -> list[list[str]]:
+ """
+ Simulates gravity in a specific column.
+
+ >>> move_x([['-', 'A'], ['-', '-'], ['-', 'C']], 1, 2)
+ [['-', '-'], ['-', 'A'], ['-', 'C']]
+ """
+
+ new_list = []
+
+ for row in range(size):
+ if matrix_g[row][column] != "-":
+ new_list.append(matrix_g[row][column])
+ else:
+ new_list.insert(0, matrix_g[row][column])
+ for row in range(size):
+ matrix_g[row][column] = new_list[row]
+ return matrix_g
+
+
+def move_y(matrix_g: list[list[str]], size: int) -> list[list[str]]:
+ """
+ Shifts all columns leftward when an entire column becomes empty.
+
+ >>> move_y([['-', 'A'], ['-', '-'], ['-', 'C']], 2)
+ [['A', '-'], ['-', '-'], ['-', 'C']]
+ """
+
+ empty_columns = []
+
+ for column in range(size - 1, -1, -1):
+ if all(matrix_g[row][column] == "-" for row in range(size)):
+ empty_columns.append(column)
+
+ for column in empty_columns:
+ for col in range(column + 1, size):
+ for row in range(size):
+ matrix_g[row][col - 1] = matrix_g[row][col]
+ for row in range(size):
+ matrix_g[row][-1] = "-"
+
+ return matrix_g
+
+
+def play(
+ matrix_g: list[list[str]], pos_x: int, pos_y: int, size: int
+) -> tuple[list[list[str]], int]:
+ """
+ Processes a single move, updating the matrix and calculating the score.
+
+ >>> play([['R', 'G'], ['R', 'G']], 0, 0, 2)
+ ([['G', '-'], ['G', '-']], 3)
+ """
+
+ same_colors = find_repeat(matrix_g, pos_x, pos_y, size)
+
+ if len(same_colors) != 0:
+ for pos in same_colors:
+ matrix_g[pos[0]][pos[1]] = "-"
+ for column in range(size):
+ matrix_g = move_x(matrix_g, column, size)
+
+ matrix_g = move_y(matrix_g, size)
+
+ return (matrix_g, increment_score(len(same_colors)))
+
+
+def process_game(size: int, matrix: list[str], moves: list[tuple[int, int]]) -> int:
+ """Processes the game logic for the given matrix and moves.
+
+ Args:
+ size (int): Size of the game board.
+ matrix (List[str]): Initial game matrix.
+ moves (List[Tuple[int, int]]): List of moves as (x, y) coordinates.
+
+ Returns:
+ int: The total score obtained.
+ >>> process_game(3, ['aaa', 'bbb', 'ccc'], [(0, 0)])
+ 6
+ """
+
+ game_matrix = [list(row) for row in matrix]
+ total_score = 0
+
+ for move in moves:
+ pos_x, pos_y = move
+ game_matrix, score = play(game_matrix, pos_x, pos_y, size)
+ total_score += score
+
+ return total_score
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod(verbose=True)
+ try:
+ size = int(input("Enter the size of the matrix: "))
+ validate_matrix_size(size)
+ print(f"Enter the {size} rows of the matrix:")
+ matrix = [input(f"Row {i + 1}: ") for i in range(size)]
+ validate_matrix_content(matrix, size)
+ moves_input = input("Enter the moves (e.g., '0 0, 1 1'): ")
+ moves = parse_moves(moves_input)
+ validate_moves(moves, size)
+ score = process_game(size, matrix, moves)
+ print(f"Total score: {score}")
+ except ValueError as e:
+ print(f"{e}")
diff --git a/matrix/matrix_class.py b/matrix/matrix_class.py
index 57a2fc45ffd1..a5940a38e836 100644
--- a/matrix/matrix_class.py
+++ b/matrix/matrix_class.py
@@ -1,358 +1,366 @@
-# An OOP approach to representing and manipulating matrices
-
-
-class Matrix:
- """
- Matrix object generated from a 2D array where each element is an array representing
- a row.
- Rows can contain type int or float.
- Common operations and information available.
- >>> rows = [
- ... [1, 2, 3],
- ... [4, 5, 6],
- ... [7, 8, 9]
- ... ]
- >>> matrix = Matrix(rows)
- >>> print(matrix)
- [[1. 2. 3.]
- [4. 5. 6.]
- [7. 8. 9.]]
-
- Matrix rows and columns are available as 2D arrays
- >>> print(matrix.rows)
- [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
- >>> print(matrix.columns())
- [[1, 4, 7], [2, 5, 8], [3, 6, 9]]
-
- Order is returned as a tuple
- >>> matrix.order
- (3, 3)
-
- Squareness and invertability are represented as bool
- >>> matrix.is_square
- True
- >>> matrix.is_invertable()
- False
-
- Identity, Minors, Cofactors and Adjugate are returned as Matrices. Inverse can be
- a Matrix or Nonetype
- >>> print(matrix.identity())
- [[1. 0. 0.]
- [0. 1. 0.]
- [0. 0. 1.]]
- >>> print(matrix.minors())
- [[-3. -6. -3.]
- [-6. -12. -6.]
- [-3. -6. -3.]]
- >>> print(matrix.cofactors())
- [[-3. 6. -3.]
- [6. -12. 6.]
- [-3. 6. -3.]]
- >>> # won't be apparent due to the nature of the cofactor matrix
- >>> print(matrix.adjugate())
- [[-3. 6. -3.]
- [6. -12. 6.]
- [-3. 6. -3.]]
- >>> print(matrix.inverse())
- None
-
- Determinant is an int, float, or Nonetype
- >>> matrix.determinant()
- 0
-
- Negation, scalar multiplication, addition, subtraction, multiplication and
- exponentiation are available and all return a Matrix
- >>> print(-matrix)
- [[-1. -2. -3.]
- [-4. -5. -6.]
- [-7. -8. -9.]]
- >>> matrix2 = matrix * 3
- >>> print(matrix2)
- [[3. 6. 9.]
- [12. 15. 18.]
- [21. 24. 27.]]
- >>> print(matrix + matrix2)
- [[4. 8. 12.]
- [16. 20. 24.]
- [28. 32. 36.]]
- >>> print(matrix - matrix2)
- [[-2. -4. -6.]
- [-8. -10. -12.]
- [-14. -16. -18.]]
- >>> print(matrix ** 3)
- [[468. 576. 684.]
- [1062. 1305. 1548.]
- [1656. 2034. 2412.]]
-
- Matrices can also be modified
- >>> matrix.add_row([10, 11, 12])
- >>> print(matrix)
- [[1. 2. 3.]
- [4. 5. 6.]
- [7. 8. 9.]
- [10. 11. 12.]]
- >>> matrix2.add_column([8, 16, 32])
- >>> print(matrix2)
- [[3. 6. 9. 8.]
- [12. 15. 18. 16.]
- [21. 24. 27. 32.]]
- >>> print(matrix * matrix2)
- [[90. 108. 126. 136.]
- [198. 243. 288. 304.]
- [306. 378. 450. 472.]
- [414. 513. 612. 640.]]
-
- """
-
- def __init__(self, rows):
- error = TypeError(
- "Matrices must be formed from a list of zero or more lists containing at "
- "least one and the same number of values, each of which must be of type "
- "int or float."
- )
- if len(rows) != 0:
- cols = len(rows[0])
- if cols == 0:
- raise error
- for row in rows:
- if len(row) != cols:
- raise error
- for value in row:
- if not isinstance(value, (int, float)):
- raise error
- self.rows = rows
- else:
- self.rows = []
-
- # MATRIX INFORMATION
- def columns(self):
- return [[row[i] for row in self.rows] for i in range(len(self.rows[0]))]
-
- @property
- def num_rows(self):
- return len(self.rows)
-
- @property
- def num_columns(self):
- return len(self.rows[0])
-
- @property
- def order(self):
- return (self.num_rows, self.num_columns)
-
- @property
- def is_square(self):
- return self.order[0] == self.order[1]
-
- def identity(self):
- values = [
- [0 if column_num != row_num else 1 for column_num in range(self.num_rows)]
- for row_num in range(self.num_rows)
- ]
- return Matrix(values)
-
- def determinant(self):
- if not self.is_square:
- return None
- if self.order == (0, 0):
- return 1
- if self.order == (1, 1):
- return self.rows[0][0]
- if self.order == (2, 2):
- return (self.rows[0][0] * self.rows[1][1]) - (
- self.rows[0][1] * self.rows[1][0]
- )
- else:
- return sum(
- self.rows[0][column] * self.cofactors().rows[0][column]
- for column in range(self.num_columns)
- )
-
- def is_invertable(self):
- return bool(self.determinant())
-
- def get_minor(self, row, column):
- values = [
- [
- self.rows[other_row][other_column]
- for other_column in range(self.num_columns)
- if other_column != column
- ]
- for other_row in range(self.num_rows)
- if other_row != row
- ]
- return Matrix(values).determinant()
-
- def get_cofactor(self, row, column):
- if (row + column) % 2 == 0:
- return self.get_minor(row, column)
- return -1 * self.get_minor(row, column)
-
- def minors(self):
- return Matrix(
- [
- [self.get_minor(row, column) for column in range(self.num_columns)]
- for row in range(self.num_rows)
- ]
- )
-
- def cofactors(self):
- return Matrix(
- [
- [
- self.minors().rows[row][column]
- if (row + column) % 2 == 0
- else self.minors().rows[row][column] * -1
- for column in range(self.minors().num_columns)
- ]
- for row in range(self.minors().num_rows)
- ]
- )
-
- def adjugate(self):
- values = [
- [self.cofactors().rows[column][row] for column in range(self.num_columns)]
- for row in range(self.num_rows)
- ]
- return Matrix(values)
-
- def inverse(self):
- determinant = self.determinant()
- return None if not determinant else self.adjugate() * (1 / determinant)
-
- def __repr__(self):
- return str(self.rows)
-
- def __str__(self):
- if self.num_rows == 0:
- return "[]"
- if self.num_rows == 1:
- return "[[" + ". ".join(self.rows[0]) + "]]"
- return (
- "["
- + "\n ".join(
- [
- "[" + ". ".join([str(value) for value in row]) + ".]"
- for row in self.rows
- ]
- )
- + "]"
- )
-
- # MATRIX MANIPULATION
- def add_row(self, row, position=None):
- type_error = TypeError("Row must be a list containing all ints and/or floats")
- if not isinstance(row, list):
- raise type_error
- for value in row:
- if not isinstance(value, (int, float)):
- raise type_error
- if len(row) != self.num_columns:
- raise ValueError(
- "Row must be equal in length to the other rows in the matrix"
- )
- if position is None:
- self.rows.append(row)
- else:
- self.rows = self.rows[0:position] + [row] + self.rows[position:]
-
- def add_column(self, column, position=None):
- type_error = TypeError(
- "Column must be a list containing all ints and/or floats"
- )
- if not isinstance(column, list):
- raise type_error
- for value in column:
- if not isinstance(value, (int, float)):
- raise type_error
- if len(column) != self.num_rows:
- raise ValueError(
- "Column must be equal in length to the other columns in the matrix"
- )
- if position is None:
- self.rows = [self.rows[i] + [column[i]] for i in range(self.num_rows)]
- else:
- self.rows = [
- self.rows[i][0:position] + [column[i]] + self.rows[i][position:]
- for i in range(self.num_rows)
- ]
-
- # MATRIX OPERATIONS
- def __eq__(self, other):
- if not isinstance(other, Matrix):
- raise TypeError("A Matrix can only be compared with another Matrix")
- return self.rows == other.rows
-
- def __ne__(self, other):
- return not self == other
-
- def __neg__(self):
- return self * -1
-
- def __add__(self, other):
- if self.order != other.order:
- raise ValueError("Addition requires matrices of the same order")
- return Matrix(
- [
- [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns)]
- for i in range(self.num_rows)
- ]
- )
-
- def __sub__(self, other):
- if self.order != other.order:
- raise ValueError("Subtraction requires matrices of the same order")
- return Matrix(
- [
- [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns)]
- for i in range(self.num_rows)
- ]
- )
-
- def __mul__(self, other):
- if isinstance(other, (int, float)):
- return Matrix([[element * other for element in row] for row in self.rows])
- elif isinstance(other, Matrix):
- if self.num_columns != other.num_rows:
- raise ValueError(
- "The number of columns in the first matrix must "
- "be equal to the number of rows in the second"
- )
- return Matrix(
- [
- [Matrix.dot_product(row, column) for column in other.columns()]
- for row in self.rows
- ]
- )
- else:
- raise TypeError(
- "A Matrix can only be multiplied by an int, float, or another matrix"
- )
-
- def __pow__(self, other):
- if not isinstance(other, int):
- raise TypeError("A Matrix can only be raised to the power of an int")
- if not self.is_square:
- raise ValueError("Only square matrices can be raised to a power")
- if other == 0:
- return self.identity()
- if other < 0:
- if self.is_invertable:
- return self.inverse() ** (-other)
- raise ValueError(
- "Only invertable matrices can be raised to a negative power"
- )
- result = self
- for i in range(other - 1):
- result *= self
- return result
-
- @classmethod
- def dot_product(cls, row, column):
- return sum(row[i] * column[i] for i in range(len(row)))
-
-
-if __name__ == "__main__":
- import doctest
-
- doctest.testmod()
+# An OOP approach to representing and manipulating matrices
+
+from __future__ import annotations
+
+
+class Matrix:
+ """
+ Matrix object generated from a 2D array where each element is an array representing
+ a row.
+ Rows can contain type int or float.
+ Common operations and information available.
+ >>> rows = [
+ ... [1, 2, 3],
+ ... [4, 5, 6],
+ ... [7, 8, 9]
+ ... ]
+ >>> matrix = Matrix(rows)
+ >>> print(matrix)
+ [[1. 2. 3.]
+ [4. 5. 6.]
+ [7. 8. 9.]]
+
+ Matrix rows and columns are available as 2D arrays
+ >>> matrix.rows
+ [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
+ >>> matrix.columns()
+ [[1, 4, 7], [2, 5, 8], [3, 6, 9]]
+
+ Order is returned as a tuple
+ >>> matrix.order
+ (3, 3)
+
+ Squareness and invertability are represented as bool
+ >>> matrix.is_square
+ True
+ >>> matrix.is_invertable()
+ False
+
+ Identity, Minors, Cofactors and Adjugate are returned as Matrices. Inverse can be
+ a Matrix or Nonetype
+ >>> print(matrix.identity())
+ [[1. 0. 0.]
+ [0. 1. 0.]
+ [0. 0. 1.]]
+ >>> print(matrix.minors())
+ [[-3. -6. -3.]
+ [-6. -12. -6.]
+ [-3. -6. -3.]]
+ >>> print(matrix.cofactors())
+ [[-3. 6. -3.]
+ [6. -12. 6.]
+ [-3. 6. -3.]]
+ >>> # won't be apparent due to the nature of the cofactor matrix
+ >>> print(matrix.adjugate())
+ [[-3. 6. -3.]
+ [6. -12. 6.]
+ [-3. 6. -3.]]
+ >>> matrix.inverse()
+ Traceback (most recent call last):
+ ...
+ TypeError: Only matrices with a non-zero determinant have an inverse
+
+ Determinant is an int, float, or Nonetype
+ >>> matrix.determinant()
+ 0
+
+ Negation, scalar multiplication, addition, subtraction, multiplication and
+ exponentiation are available and all return a Matrix
+ >>> print(-matrix)
+ [[-1. -2. -3.]
+ [-4. -5. -6.]
+ [-7. -8. -9.]]
+ >>> matrix2 = matrix * 3
+ >>> print(matrix2)
+ [[3. 6. 9.]
+ [12. 15. 18.]
+ [21. 24. 27.]]
+ >>> print(matrix + matrix2)
+ [[4. 8. 12.]
+ [16. 20. 24.]
+ [28. 32. 36.]]
+ >>> print(matrix - matrix2)
+ [[-2. -4. -6.]
+ [-8. -10. -12.]
+ [-14. -16. -18.]]
+ >>> print(matrix ** 3)
+ [[468. 576. 684.]
+ [1062. 1305. 1548.]
+ [1656. 2034. 2412.]]
+
+ Matrices can also be modified
+ >>> matrix.add_row([10, 11, 12])
+ >>> print(matrix)
+ [[1. 2. 3.]
+ [4. 5. 6.]
+ [7. 8. 9.]
+ [10. 11. 12.]]
+ >>> matrix2.add_column([8, 16, 32])
+ >>> print(matrix2)
+ [[3. 6. 9. 8.]
+ [12. 15. 18. 16.]
+ [21. 24. 27. 32.]]
+ >>> print(matrix * matrix2)
+ [[90. 108. 126. 136.]
+ [198. 243. 288. 304.]
+ [306. 378. 450. 472.]
+ [414. 513. 612. 640.]]
+ """
+
+ def __init__(self, rows: list[list[int]]):
+ error = TypeError(
+ "Matrices must be formed from a list of zero or more lists containing at "
+ "least one and the same number of values, each of which must be of type "
+ "int or float."
+ )
+ if len(rows) != 0:
+ cols = len(rows[0])
+ if cols == 0:
+ raise error
+ for row in rows:
+ if len(row) != cols:
+ raise error
+ for value in row:
+ if not isinstance(value, (int, float)):
+ raise error
+ self.rows = rows
+ else:
+ self.rows = []
+
+ # MATRIX INFORMATION
+ def columns(self) -> list[list[int]]:
+ return [[row[i] for row in self.rows] for i in range(len(self.rows[0]))]
+
+ @property
+ def num_rows(self) -> int:
+ return len(self.rows)
+
+ @property
+ def num_columns(self) -> int:
+ return len(self.rows[0])
+
+ @property
+ def order(self) -> tuple[int, int]:
+ return self.num_rows, self.num_columns
+
+ @property
+ def is_square(self) -> bool:
+ return self.order[0] == self.order[1]
+
+ def identity(self) -> Matrix:
+ values = [
+ [0 if column_num != row_num else 1 for column_num in range(self.num_rows)]
+ for row_num in range(self.num_rows)
+ ]
+ return Matrix(values)
+
+ def determinant(self) -> int:
+ if not self.is_square:
+ return 0
+ if self.order == (0, 0):
+ return 1
+ if self.order == (1, 1):
+ return int(self.rows[0][0])
+ if self.order == (2, 2):
+ return int(
+ (self.rows[0][0] * self.rows[1][1])
+ - (self.rows[0][1] * self.rows[1][0])
+ )
+ else:
+ return sum(
+ self.rows[0][column] * self.cofactors().rows[0][column]
+ for column in range(self.num_columns)
+ )
+
+ def is_invertable(self) -> bool:
+ return bool(self.determinant())
+
+ def get_minor(self, row: int, column: int) -> int:
+ values = [
+ [
+ self.rows[other_row][other_column]
+ for other_column in range(self.num_columns)
+ if other_column != column
+ ]
+ for other_row in range(self.num_rows)
+ if other_row != row
+ ]
+ return Matrix(values).determinant()
+
+ def get_cofactor(self, row: int, column: int) -> int:
+ if (row + column) % 2 == 0:
+ return self.get_minor(row, column)
+ return -1 * self.get_minor(row, column)
+
+ def minors(self) -> Matrix:
+ return Matrix(
+ [
+ [self.get_minor(row, column) for column in range(self.num_columns)]
+ for row in range(self.num_rows)
+ ]
+ )
+
+ def cofactors(self) -> Matrix:
+ return Matrix(
+ [
+ [
+ self.minors().rows[row][column]
+ if (row + column) % 2 == 0
+ else self.minors().rows[row][column] * -1
+ for column in range(self.minors().num_columns)
+ ]
+ for row in range(self.minors().num_rows)
+ ]
+ )
+
+ def adjugate(self) -> Matrix:
+ values = [
+ [self.cofactors().rows[column][row] for column in range(self.num_columns)]
+ for row in range(self.num_rows)
+ ]
+ return Matrix(values)
+
+ def inverse(self) -> Matrix:
+ determinant = self.determinant()
+ if not determinant:
+ raise TypeError("Only matrices with a non-zero determinant have an inverse")
+ return self.adjugate() * (1 / determinant)
+
+ def __repr__(self) -> str:
+ return str(self.rows)
+
+ def __str__(self) -> str:
+ if self.num_rows == 0:
+ return "[]"
+ if self.num_rows == 1:
+ return "[[" + ". ".join(str(self.rows[0])) + "]]"
+ return (
+ "["
+ + "\n ".join(
+ [
+ "[" + ". ".join([str(value) for value in row]) + ".]"
+ for row in self.rows
+ ]
+ )
+ + "]"
+ )
+
+ # MATRIX MANIPULATION
+ def add_row(self, row: list[int], position: int | None = None) -> None:
+ type_error = TypeError("Row must be a list containing all ints and/or floats")
+ if not isinstance(row, list):
+ raise type_error
+ for value in row:
+ if not isinstance(value, (int, float)):
+ raise type_error
+ if len(row) != self.num_columns:
+ raise ValueError(
+ "Row must be equal in length to the other rows in the matrix"
+ )
+ if position is None:
+ self.rows.append(row)
+ else:
+ self.rows = self.rows[0:position] + [row] + self.rows[position:]
+
+ def add_column(self, column: list[int], position: int | None = None) -> None:
+ type_error = TypeError(
+ "Column must be a list containing all ints and/or floats"
+ )
+ if not isinstance(column, list):
+ raise type_error
+ for value in column:
+ if not isinstance(value, (int, float)):
+ raise type_error
+ if len(column) != self.num_rows:
+ raise ValueError(
+ "Column must be equal in length to the other columns in the matrix"
+ )
+ if position is None:
+ self.rows = [self.rows[i] + [column[i]] for i in range(self.num_rows)]
+ else:
+ self.rows = [
+ self.rows[i][0:position] + [column[i]] + self.rows[i][position:]
+ for i in range(self.num_rows)
+ ]
+
+ # MATRIX OPERATIONS
+ def __eq__(self, other: object) -> bool:
+ if not isinstance(other, Matrix):
+ return NotImplemented
+ return self.rows == other.rows
+
+ def __ne__(self, other: object) -> bool:
+ return not self == other
+
+ def __neg__(self) -> Matrix:
+ return self * -1
+
+ def __add__(self, other: Matrix) -> Matrix:
+ if self.order != other.order:
+ raise ValueError("Addition requires matrices of the same order")
+ return Matrix(
+ [
+ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns)]
+ for i in range(self.num_rows)
+ ]
+ )
+
+ def __sub__(self, other: Matrix) -> Matrix:
+ if self.order != other.order:
+ raise ValueError("Subtraction requires matrices of the same order")
+ return Matrix(
+ [
+ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns)]
+ for i in range(self.num_rows)
+ ]
+ )
+
+ def __mul__(self, other: Matrix | float) -> Matrix:
+ if isinstance(other, (int, float)):
+ return Matrix(
+ [[int(element * other) for element in row] for row in self.rows]
+ )
+ elif isinstance(other, Matrix):
+ if self.num_columns != other.num_rows:
+ raise ValueError(
+ "The number of columns in the first matrix must "
+ "be equal to the number of rows in the second"
+ )
+ return Matrix(
+ [
+ [Matrix.dot_product(row, column) for column in other.columns()]
+ for row in self.rows
+ ]
+ )
+ else:
+ raise TypeError(
+ "A Matrix can only be multiplied by an int, float, or another matrix"
+ )
+
+ def __pow__(self, other: int) -> Matrix:
+ if not isinstance(other, int):
+ raise TypeError("A Matrix can only be raised to the power of an int")
+ if not self.is_square:
+ raise ValueError("Only square matrices can be raised to a power")
+ if other == 0:
+ return self.identity()
+ if other < 0:
+ if self.is_invertable():
+ return self.inverse() ** (-other)
+ raise ValueError(
+ "Only invertable matrices can be raised to a negative power"
+ )
+ result = self
+ for _ in range(other - 1):
+ result *= self
+ return result
+
+ @classmethod
+ def dot_product(cls, row: list[int], column: list[int]) -> int:
+ return sum(row[i] * column[i] for i in range(len(row)))
+
+
+if __name__ == "__main__":
+ import doctest
+
+ doctest.testmod()
diff --git a/matrix/matrix_equalization.py b/matrix/matrix_equalization.py
new file mode 100644
index 000000000000..e7e76505cf63
--- /dev/null
+++ b/matrix/matrix_equalization.py
@@ -0,0 +1,55 @@
+from sys import maxsize
+
+
+def array_equalization(vector: list[int], step_size: int) -> int:
+ """
+ This algorithm equalizes all elements of the input vector
+ to a common value, by making the minimal number of
+ "updates" under the constraint of a step size (step_size).
+
+ >>> array_equalization([1, 1, 6, 2, 4, 6, 5, 1, 7, 2, 2, 1, 7, 2, 2], 4)
+ 4
+ >>> array_equalization([22, 81, 88, 71, 22, 81, 632, 81, 81, 22, 92], 2)
+ 5
+ >>> array_equalization([0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 5)
+ 0
+ >>> array_equalization([22, 22, 22, 33, 33, 33], 2)
+ 2
+ >>> array_equalization([1, 2, 3], 0)
+ Traceback (most recent call last):
+ ValueError: Step size must be positive and non-zero.
+ >>> array_equalization([1, 2, 3], -1)
+ Traceback (most recent call last):
+ ValueError: Step size must be positive and non-zero.
+ >>> array_equalization([1, 2, 3], 0.5)
+ Traceback (most recent call last):
+ ValueError: Step size must be an integer.
+ >>> array_equalization([1, 2, 3], maxsize)
+ 1
+ """
+ if step_size <= 0:
+ raise ValueError("Step size must be positive and non-zero.")
+ if not isinstance(step_size, int):
+ raise ValueError("Step size must be an integer.")
+
+ unique_elements = set(vector)
+ min_updates = maxsize
+
+ for element in unique_elements:
+ elem_index = 0
+ updates = 0
+ while elem_index < len(vector):
+ if vector[elem_index] != element:
+ updates += 1
+ elem_index += step_size
+ else:
+ elem_index += 1
+ min_updates = min(min_updates, updates)
+
+ return min_updates
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ testmod()
diff --git a/matrix/matrix_multiplication_recursion.py b/matrix/matrix_multiplication_recursion.py
new file mode 100644
index 000000000000..57c4d80de017
--- /dev/null
+++ b/matrix/matrix_multiplication_recursion.py
@@ -0,0 +1,181 @@
+# @Author : ojas-wani
+# @File : matrix_multiplication_recursion.py
+# @Date : 10/06/2023
+
+
+"""
+Perform matrix multiplication using a recursive algorithm.
+https://en.wikipedia.org/wiki/Matrix_multiplication
+"""
+
+# type Matrix = list[list[int]] # psf/black currenttly fails on this line
+Matrix = list[list[int]]
+
+matrix_1_to_4 = [
+ [1, 2],
+ [3, 4],
+]
+
+matrix_5_to_8 = [
+ [5, 6],
+ [7, 8],
+]
+
+matrix_5_to_9_high = [
+ [5, 6],
+ [7, 8],
+ [9],
+]
+
+matrix_5_to_9_wide = [
+ [5, 6],
+ [7, 8, 9],
+]
+
+matrix_count_up = [
+ [1, 2, 3, 4],
+ [5, 6, 7, 8],
+ [9, 10, 11, 12],
+ [13, 14, 15, 16],
+]
+
+matrix_unordered = [
+ [5, 8, 1, 2],
+ [6, 7, 3, 0],
+ [4, 5, 9, 1],
+ [2, 6, 10, 14],
+]
+matrices = (
+ matrix_1_to_4,
+ matrix_5_to_8,
+ matrix_5_to_9_high,
+ matrix_5_to_9_wide,
+ matrix_count_up,
+ matrix_unordered,
+)
+
+
+def is_square(matrix: Matrix) -> bool:
+ """
+ >>> is_square([])
+ True
+ >>> is_square(matrix_1_to_4)
+ True
+ >>> is_square(matrix_5_to_9_high)
+ False
+ """
+ len_matrix = len(matrix)
+ return all(len(row) == len_matrix for row in matrix)
+
+
+def matrix_multiply(matrix_a: Matrix, matrix_b: Matrix) -> Matrix:
+ """
+ >>> matrix_multiply(matrix_1_to_4, matrix_5_to_8)
+ [[19, 22], [43, 50]]
+ """
+ return [
+ [sum(a * b for a, b in zip(row, col)) for col in zip(*matrix_b)]
+ for row in matrix_a
+ ]
+
+
+def matrix_multiply_recursive(matrix_a: Matrix, matrix_b: Matrix) -> Matrix:
+ """
+ :param matrix_a: A square Matrix.
+ :param matrix_b: Another square Matrix with the same dimensions as matrix_a.
+ :return: Result of matrix_a * matrix_b.
+ :raises ValueError: If the matrices cannot be multiplied.
+
+ >>> matrix_multiply_recursive([], [])
+ []
+ >>> matrix_multiply_recursive(matrix_1_to_4, matrix_5_to_8)
+ [[19, 22], [43, 50]]
+ >>> matrix_multiply_recursive(matrix_count_up, matrix_unordered)
+ [[37, 61, 74, 61], [105, 165, 166, 129], [173, 269, 258, 197], [241, 373, 350, 265]]
+ >>> matrix_multiply_recursive(matrix_1_to_4, matrix_5_to_9_wide)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid matrix dimensions
+ >>> matrix_multiply_recursive(matrix_1_to_4, matrix_5_to_9_high)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid matrix dimensions
+ >>> matrix_multiply_recursive(matrix_1_to_4, matrix_count_up)
+ Traceback (most recent call last):
+ ...
+ ValueError: Invalid matrix dimensions
+ """
+ if not matrix_a or not matrix_b:
+ return []
+ if not all(
+ (len(matrix_a) == len(matrix_b), is_square(matrix_a), is_square(matrix_b))
+ ):
+ raise ValueError("Invalid matrix dimensions")
+
+ # Initialize the result matrix with zeros
+ result = [[0] * len(matrix_b[0]) for _ in range(len(matrix_a))]
+
+ # Recursive multiplication of matrices
+ def multiply(
+ i_loop: int,
+ j_loop: int,
+ k_loop: int,
+ matrix_a: Matrix,
+ matrix_b: Matrix,
+ result: Matrix,
+ ) -> None:
+ """
+ :param matrix_a: A square Matrix.
+ :param matrix_b: Another square Matrix with the same dimensions as matrix_a.
+ :param result: Result matrix
+ :param i: Index used for iteration during multiplication.
+ :param j: Index used for iteration during multiplication.
+ :param k: Index used for iteration during multiplication.
+ >>> 0 > 1 # Doctests in inner functions are never run
+ True
+ """
+ if i_loop >= len(matrix_a):
+ return
+ if j_loop >= len(matrix_b[0]):
+ return multiply(i_loop + 1, 0, 0, matrix_a, matrix_b, result)
+ if k_loop >= len(matrix_b):
+ return multiply(i_loop, j_loop + 1, 0, matrix_a, matrix_b, result)
+ result[i_loop][j_loop] += matrix_a[i_loop][k_loop] * matrix_b[k_loop][j_loop]
+ return multiply(i_loop, j_loop, k_loop + 1, matrix_a, matrix_b, result)
+
+ # Perform the recursive matrix multiplication
+ multiply(0, 0, 0, matrix_a, matrix_b, result)
+ return result
+
+
+if __name__ == "__main__":
+ from doctest import testmod
+
+ failure_count, test_count = testmod()
+ if not failure_count:
+ matrix_a = matrices[0]
+ for matrix_b in matrices[1:]:
+ print("Multiplying:")
+ for row in matrix_a:
+ print(row)
+ print("By:")
+ for row in matrix_b:
+ print(row)
+ print("Result:")
+ try:
+ result = matrix_multiply_recursive(matrix_a, matrix_b)
+ for row in result:
+ print(row)
+ assert result == matrix_multiply(matrix_a, matrix_b)
+ except ValueError as e:
+ print(f"{e!r}")
+ print()
+ matrix_a = matrix_b
+
+ print("Benchmark:")
+ from functools import partial
+ from timeit import timeit
+
+ mytimeit = partial(timeit, globals=globals(), number=100_000)
+ for func in ("matrix_multiply", "matrix_multiply_recursive"):
+ print(f"{func:>25}(): {mytimeit(f'{func}(matrix_count_up, matrix_unordered)')}")
diff --git a/matrix/matrix_operation.py b/matrix/matrix_operation.py
index dca01f9c3183..d63e758f1838 100644
--- a/matrix/matrix_operation.py
+++ b/matrix/matrix_operation.py
@@ -4,8 +4,10 @@
from __future__ import annotations
+from typing import Any
-def add(*matrix_s: list[list]) -> list[list]:
+
+def add(*matrix_s: list[list[int]]) -> list[list[int]]:
"""
>>> add([[1,2],[3,4]],[[2,3],[4,5]])
[[3, 5], [7, 9]]
@@ -13,19 +15,28 @@ def add(*matrix_s: list[list]) -> list[list]:
[[3.2, 5.4], [7, 9]]
>>> add([[1, 2], [4, 5]], [[3, 7], [3, 4]], [[3, 5], [5, 7]])
[[7, 14], [12, 16]]
+ >>> add([3], [4, 5])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a matrix, got int/list instead
"""
if all(_check_not_integer(m) for m in matrix_s):
for i in matrix_s[1:]:
_verify_matrix_sizes(matrix_s[0], i)
return [[sum(t) for t in zip(*m)] for m in zip(*matrix_s)]
+ raise TypeError("Expected a matrix, got int/list instead")
-def subtract(matrix_a: list[list], matrix_b: list[list]) -> list[list]:
+def subtract(matrix_a: list[list[int]], matrix_b: list[list[int]]) -> list[list[int]]:
"""
>>> subtract([[1,2],[3,4]],[[2,3],[4,5]])
[[-1, -1], [-1, -1]]
>>> subtract([[1,2.5],[3,4]],[[2,3],[4,5.5]])
[[-1, -0.5], [-1, -1.5]]
+ >>> subtract([3], [4, 5])
+ Traceback (most recent call last):
+ ...
+ TypeError: Expected a matrix, got int/list instead
"""
if (
_check_not_integer(matrix_a)
@@ -33,9 +44,10 @@ def subtract(matrix_a: list[list], matrix_b: list[list]) -> list[list]:
and _verify_matrix_sizes(matrix_a, matrix_b)
):
return [[i - j for i, j in zip(*m)] for m in zip(matrix_a, matrix_b)]
+ raise TypeError("Expected a matrix, got int/list instead")
-def scalar_multiply(matrix: list[list], n: int) -> list[list]:
+def scalar_multiply(matrix: list[list[int]], n: float) -> list[list[float]]:
"""
>>> scalar_multiply([[1,2],[3,4]],5)
[[5, 10], [15, 20]]
@@ -45,7 +57,7 @@ def scalar_multiply(matrix: list[list], n: int) -> list[list]:
return [[x * n for x in row] for row in matrix]
-def multiply(matrix_a: list[list], matrix_b: list[list]) -> list[list]:
+def multiply(matrix_a: list[list[int]], matrix_b: list[list[int]]) -> list[list[int]]:
"""
>>> multiply([[1,2],[3,4]],[[5,5],[7,5]])
[[19, 15], [43, 35]]
@@ -58,16 +70,17 @@ def multiply(matrix_a: list[list], matrix_b: list[list]) -> list[list]:
rows, cols = _verify_matrix_sizes(matrix_a, matrix_b)
if cols[0] != rows[1]:
- raise ValueError(
- f"Cannot multiply matrix of dimensions ({rows[0]},{cols[0]}) "
- f"and ({rows[1]},{cols[1]})"
+ msg = (
+ "Cannot multiply matrix of dimensions "
+ f"({rows[0]},{cols[0]}) and ({rows[1]},{cols[1]})"
)
+ raise ValueError(msg)
return [
[sum(m * n for m, n in zip(i, j)) for j in zip(*matrix_b)] for i in matrix_a
]
-def identity(n: int) -> list[list]:
+def identity(n: int) -> list[list[int]]:
"""
:param n: dimension for nxn matrix
:type n: int
@@ -79,21 +92,28 @@ def identity(n: int) -> list[list]:
return [[int(row == column) for column in range(n)] for row in range(n)]
-def transpose(matrix: list[list], return_map: bool = True) -> list[list]:
+def transpose(
+ matrix: list[list[int]], return_map: bool = True
+) -> list[list[int]] | map[list[int]]:
"""
>>> transpose([[1,2],[3,4]]) # doctest: +ELLIPSIS