jamtur01's picture
Upload folder using huggingface_hub
9c6594c verified
# Licensed under the Apache License, Version 2.0 (the "License");
# http://www.apache.org/licenses/LICENSE-2.0
#
import glob
import logging
import os
import re
import requests
def _download_file(file_url: str, folder: str) -> str:
"""Download a file from URL to a particular folder."""
fname = os.path.basename(file_url)
file_path = os.path.join(folder, fname)
if os.path.isfile(file_path):
logging.warning(f'given file "{file_path}" already exists and will be overwritten with {file_url}')
# see: https://stackoverflow.com/a/34957875
rq = requests.get(file_url, timeout=10)
with open(file_path, "wb") as outfile:
outfile.write(rq.content)
return fname
def _search_all_occurrences(list_files: list[str], pattern: str) -> list[str]:
"""Search for all occurrences of specific pattern in a collection of files.
Args:
list_files: list of files to be scanned
pattern: pattern for search, reg. expression
"""
collected = []
for file_path in list_files:
with open(file_path, encoding="UTF-8") as fopem:
body = fopem.read()
found = re.findall(pattern, body)
collected += found
return collected
def _replace_remote_with_local(file_path: str, docs_folder: str, pairs_url_path: list[tuple[str, str]]) -> None:
"""Replace all URL with local files in a given file.
Args:
file_path: file for replacement
docs_folder: the location of docs related to the project root
pairs_url_path: pairs of URL and local file path to be swapped
"""
# drop the default/global path to the docs
relt_path = os.path.dirname(file_path).replace(docs_folder, "")
# filter the path starting with / as not empty folder names
depth = len([p for p in relt_path.split(os.path.sep) if p])
with open(file_path, encoding="UTF-8") as fopen:
body = fopen.read()
for url, fpath in pairs_url_path:
if depth:
path_up = [".."] * depth
fpath = os.path.join(*path_up, fpath)
body = body.replace(url, fpath)
with open(file_path, "w", encoding="UTF-8") as fw:
fw.write(body)
def fetch_external_assets(
docs_folder: str = "docs/source",
assets_folder: str = "fetched-s3-assets",
file_pattern: str = "*.rst",
retrieve_pattern: str = r"https?://[-a-zA-Z0-9_]+\.s3\.[-a-zA-Z0-9()_\\+.\\/=]+",
) -> None:
"""Search all URL in docs, download these files locally and replace online with local version.
Args:
docs_folder: the location of docs related to the project root
assets_folder: a folder inside ``docs_folder`` to be created and saving online assets
file_pattern: what kind of files shall be scanned
retrieve_pattern: pattern for reg. expression to search URL/S3 resources
"""
list_files = glob.glob(os.path.join(docs_folder, "**", file_pattern), recursive=True)
if not list_files:
logging.warning(f'no files were listed in folder "{docs_folder}" and pattern "{file_pattern}"')
return
urls = _search_all_occurrences(list_files, pattern=retrieve_pattern)
if not urls:
logging.info(f"no resources/assets were match in {docs_folder} for {retrieve_pattern}")
return
target_folder = os.path.join(docs_folder, assets_folder)
os.makedirs(target_folder, exist_ok=True)
pairs_url_file = []
for i, url in enumerate(set(urls)):
logging.info(f" >> downloading ({i}/{len(urls)}): {url}")
fname = _download_file(url, target_folder)
pairs_url_file.append((url, os.path.join(assets_folder, fname)))
for fpath in list_files:
_replace_remote_with_local(fpath, docs_folder, pairs_url_file)