Spaces:
Runtime error
Runtime error
| # coding=utf-8 | |
| # Copyright 2022-present, the HuggingFace Inc. team. | |
| # | |
| # Licensed under the Apache License, Version 2.0 (the "License"); | |
| # you may not use this file except in compliance with the License. | |
| # You may obtain a copy of the License at | |
| # | |
| # http://www.apache.org/licenses/LICENSE-2.0 | |
| # | |
| # Unless required by applicable law or agreed to in writing, software | |
| # distributed under the License is distributed on an "AS IS" BASIS, | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
| # See the License for the specific language governing permissions and | |
| # limitations under the License. | |
| """Contains command to scan the HF cache directory. | |
| Usage: | |
| huggingface-cli scan-cache | |
| huggingface-cli scan-cache -v | |
| huggingface-cli scan-cache -vvv | |
| huggingface-cli scan-cache --dir ~/.cache/huggingface/hub | |
| """ | |
| import time | |
| from argparse import Namespace, _SubParsersAction | |
| from typing import Optional | |
| from ..utils import CacheNotFound, HFCacheInfo, scan_cache_dir | |
| from . import BaseHuggingfaceCLICommand | |
| from ._cli_utils import ANSI, tabulate | |
| class ScanCacheCommand(BaseHuggingfaceCLICommand): | |
| def register_subcommand(parser: _SubParsersAction): | |
| scan_cache_parser = parser.add_parser("scan-cache", help="Scan cache directory.") | |
| scan_cache_parser.add_argument( | |
| "--dir", | |
| type=str, | |
| default=None, | |
| help="cache directory to scan (optional). Default to the default HuggingFace cache.", | |
| ) | |
| scan_cache_parser.add_argument( | |
| "-v", | |
| "--verbose", | |
| action="count", | |
| default=0, | |
| help="show a more verbose output", | |
| ) | |
| scan_cache_parser.set_defaults(func=ScanCacheCommand) | |
| def __init__(self, args: Namespace) -> None: | |
| self.verbosity: int = args.verbose | |
| self.cache_dir: Optional[str] = args.dir | |
| def run(self): | |
| try: | |
| t0 = time.time() | |
| hf_cache_info = scan_cache_dir(self.cache_dir) | |
| t1 = time.time() | |
| except CacheNotFound as exc: | |
| cache_dir = exc.cache_dir | |
| print(f"Cache directory not found: {cache_dir}") | |
| return | |
| self._print_hf_cache_info_as_table(hf_cache_info) | |
| print( | |
| f"\nDone in {round(t1-t0,1)}s. Scanned {len(hf_cache_info.repos)} repo(s)" | |
| f" for a total of {ANSI.red(hf_cache_info.size_on_disk_str)}." | |
| ) | |
| if len(hf_cache_info.warnings) > 0: | |
| message = f"Got {len(hf_cache_info.warnings)} warning(s) while scanning." | |
| if self.verbosity >= 3: | |
| print(ANSI.gray(message)) | |
| for warning in hf_cache_info.warnings: | |
| print(ANSI.gray(warning)) | |
| else: | |
| print(ANSI.gray(message + " Use -vvv to print details.")) | |
| def _print_hf_cache_info_as_table(self, hf_cache_info: HFCacheInfo) -> None: | |
| print(get_table(hf_cache_info, verbosity=self.verbosity)) | |
| def get_table(hf_cache_info: HFCacheInfo, *, verbosity: int = 0) -> str: | |
| """Generate a table from the [`HFCacheInfo`] object. | |
| Pass `verbosity=0` to get a table with a single row per repo, with columns | |
| "repo_id", "repo_type", "size_on_disk", "nb_files", "last_accessed", "last_modified", "refs", "local_path". | |
| Pass `verbosity=1` to get a table with a row per repo and revision (thus multiple rows can appear for a single repo), with columns | |
| "repo_id", "repo_type", "revision", "size_on_disk", "nb_files", "last_modified", "refs", "local_path". | |
| Example: | |
| ```py | |
| >>> from huggingface_hub.utils import scan_cache_dir | |
| >>> from huggingface_hub.commands.scan_cache import get_table | |
| >>> hf_cache_info = scan_cache_dir() | |
| HFCacheInfo(...) | |
| >>> print(get_table(hf_cache_info, verbosity=0)) | |
| REPO ID REPO TYPE SIZE ON DISK NB FILES LAST_ACCESSED LAST_MODIFIED REFS LOCAL PATH | |
| --------------------------------------------------- --------- ------------ -------- ------------- ------------- ---- -------------------------------------------------------------------------------------------------- | |
| roberta-base model 2.7M 5 1 day ago 1 week ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--roberta-base | |
| suno/bark model 8.8K 1 1 week ago 1 week ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--suno--bark | |
| t5-base model 893.8M 4 4 days ago 7 months ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--t5-base | |
| t5-large model 3.0G 4 5 weeks ago 5 months ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--t5-large | |
| >>> print(get_table(hf_cache_info, verbosity=1)) | |
| REPO ID REPO TYPE REVISION SIZE ON DISK NB FILES LAST_MODIFIED REFS LOCAL PATH | |
| --------------------------------------------------- --------- ---------------------------------------- ------------ -------- ------------- ---- ----------------------------------------------------------------------------------------------------------------------------------------------------- | |
| roberta-base model e2da8e2f811d1448a5b465c236feacd80ffbac7b 2.7M 5 1 week ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--roberta-base\\snapshots\\e2da8e2f811d1448a5b465c236feacd80ffbac7b | |
| suno/bark model 70a8a7d34168586dc5d028fa9666aceade177992 8.8K 1 1 week ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--suno--bark\\snapshots\\70a8a7d34168586dc5d028fa9666aceade177992 | |
| t5-base model a9723ea7f1b39c1eae772870f3b547bf6ef7e6c1 893.8M 4 7 months ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--t5-base\\snapshots\\a9723ea7f1b39c1eae772870f3b547bf6ef7e6c1 | |
| t5-large model 150ebc2c4b72291e770f58e6057481c8d2ed331a 3.0G 4 5 months ago main C:\\Users\\admin\\.cache\\huggingface\\hub\\models--t5-large\\snapshots\\150ebc2c4b72291e770f58e6057481c8d2ed331a ``` | |
| ``` | |
| Args: | |
| hf_cache_info ([`HFCacheInfo`]): | |
| The HFCacheInfo object to print. | |
| verbosity (`int`, *optional*): | |
| The verbosity level. Defaults to 0. | |
| Returns: | |
| `str`: The table as a string. | |
| """ | |
| if verbosity == 0: | |
| return tabulate( | |
| rows=[ | |
| [ | |
| repo.repo_id, | |
| repo.repo_type, | |
| "{:>12}".format(repo.size_on_disk_str), | |
| repo.nb_files, | |
| repo.last_accessed_str, | |
| repo.last_modified_str, | |
| ", ".join(sorted(repo.refs)), | |
| str(repo.repo_path), | |
| ] | |
| for repo in sorted(hf_cache_info.repos, key=lambda repo: repo.repo_path) | |
| ], | |
| headers=[ | |
| "REPO ID", | |
| "REPO TYPE", | |
| "SIZE ON DISK", | |
| "NB FILES", | |
| "LAST_ACCESSED", | |
| "LAST_MODIFIED", | |
| "REFS", | |
| "LOCAL PATH", | |
| ], | |
| ) | |
| else: | |
| return tabulate( | |
| rows=[ | |
| [ | |
| repo.repo_id, | |
| repo.repo_type, | |
| revision.commit_hash, | |
| "{:>12}".format(revision.size_on_disk_str), | |
| revision.nb_files, | |
| revision.last_modified_str, | |
| ", ".join(sorted(revision.refs)), | |
| str(revision.snapshot_path), | |
| ] | |
| for repo in sorted(hf_cache_info.repos, key=lambda repo: repo.repo_path) | |
| for revision in sorted(repo.revisions, key=lambda revision: revision.commit_hash) | |
| ], | |
| headers=[ | |
| "REPO ID", | |
| "REPO TYPE", | |
| "REVISION", | |
| "SIZE ON DISK", | |
| "NB FILES", | |
| "LAST_MODIFIED", | |
| "REFS", | |
| "LOCAL PATH", | |
| ], | |
| ) | |