Skip to content

[pull] master from TheAlgorithms:master #131

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 2 commits into from
Jul 11, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 5 additions & 6 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,9 +16,9 @@ repos:
- id: auto-walrus

- repo: https://github.com/astral-sh/ruff-pre-commit
rev: v0.11.11
rev: v0.12.3
hooks:
- id: ruff
- id: ruff-check
- id: ruff-format

- repo: https://github.com/codespell-project/codespell
Expand All @@ -29,7 +29,7 @@ repos:
- tomli

- repo: https://github.com/tox-dev/pyproject-fmt
rev: "v2.6.0"
rev: v2.6.0
hooks:
- id: pyproject-fmt

Expand All @@ -53,12 +53,11 @@ repos:
args:
- --explicit-package-bases
- --ignore-missing-imports
- --install-types # See mirrors-mypy README.md
- --install-types
- --non-interactive
additional_dependencies: [types-requests]

- repo: https://github.com/pre-commit/mirrors-prettier
rev: "v4.0.0-alpha.8"
rev: v4.0.0-alpha.8
hooks:
- id: prettier
types_or: [toml, yaml]
33 changes: 14 additions & 19 deletions web_programming/fetch_well_rx_price.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,21 +5,19 @@

"""

from urllib.error import HTTPError

import httpx
from bs4 import BeautifulSoup
from requests import exceptions, get

BASE_URL = "https://www.wellrx.com/prescriptions/{0}/{1}/?freshSearch=true"
BASE_URL = "https://www.wellrx.com/prescriptions/{}/{}/?freshSearch=true"


def fetch_pharmacy_and_price_list(drug_name: str, zip_code: str) -> list | None:
"""[summary]

This function will take input of drug name and zipcode,
then request to the BASE_URL site.
Get the page data and scrape it to the generate the
list of lowest prices for the prescription drug.
Get the page data and scrape it to generate the
list of the lowest prices for the prescription drug.

Args:
drug_name (str): [Drug name]
Expand All @@ -28,12 +26,12 @@ def fetch_pharmacy_and_price_list(drug_name: str, zip_code: str) -> list | None:
Returns:
list: [List of pharmacy name and price]

>>> fetch_pharmacy_and_price_list(None, None)

>>> fetch_pharmacy_and_price_list(None, 30303)

>>> fetch_pharmacy_and_price_list("eliquis", None)

>>> print(fetch_pharmacy_and_price_list(None, None))
None
>>> print(fetch_pharmacy_and_price_list(None, 30303))
None
>>> print(fetch_pharmacy_and_price_list("eliquis", None))
None
"""

try:
Expand All @@ -42,25 +40,22 @@ def fetch_pharmacy_and_price_list(drug_name: str, zip_code: str) -> list | None:
return None

request_url = BASE_URL.format(drug_name, zip_code)
response = get(request_url, timeout=10)

# Is the response ok?
response.raise_for_status()
response = httpx.get(request_url, timeout=10).raise_for_status()

# Scrape the data using bs4
soup = BeautifulSoup(response.text, "html.parser")

# This list will store the name and price.
pharmacy_price_list = []

# Fetch all the grids that contains the items.
# Fetch all the grids that contain the items.
grid_list = soup.find_all("div", {"class": "grid-x pharmCard"})
if grid_list and len(grid_list) > 0:
for grid in grid_list:
# Get the pharmacy price.
pharmacy_name = grid.find("p", {"class": "list-title"}).text

# Get price of the drug.
# Get the price of the drug.
price = grid.find("span", {"p", "price price-large"}).text

pharmacy_price_list.append(
Expand All @@ -72,7 +67,7 @@ def fetch_pharmacy_and_price_list(drug_name: str, zip_code: str) -> list | None:

return pharmacy_price_list

except (HTTPError, exceptions.RequestException, ValueError):
except (httpx.HTTPError, ValueError):
return None


Expand Down