import os
from huggingface_hub import snapshot_download
from transformers.utils.hub import move_cache

hf_token: str = os.getenv("HF_TOKEN")
if hf_token is None:
    raise ValueError("HF_TOKEN is not set")

hf_token = hf_token.strip()
if hf_token == "":
    raise ValueError("HF_TOKEN is empty")


# This is about 2.47 GB
snapshot_download(
    repo_id="meta-llama/Llama-3.2-1B-Instruct",
    revision="9213176726f574b556790deb65791e0c5aa438b6",
    token=hf_token,
)


# This is about 3.67 GB
snapshot_download(
    repo_id="sail/Sailor-4B-Chat",
    revision="89a866a7041e6ec023dd462adeca8e28dd53c83e",
    token=hf_token,
)

# https://github.com/huggingface/transformers/issues/20428
move_cache()