Update README.md
Browse files
README.md
CHANGED
@@ -19,6 +19,8 @@ Feel free to send in pull requests or use this code however you'd like.\
|
|
19 |
|
20 |
- [EXL2 Private Quant V3](https://colab.research.google.com/drive/1Vc7d6JU3Z35OVHmtuMuhT830THJnzNfS?usp=sharing) **(COLAB)**
|
21 |
|
|
|
|
|
22 |
## work in progress/not tested ([unfinished](https://huggingface.co/Anthonyg5005/hf-scripts/tree/unfinished) branch)
|
23 |
|
24 |
- EXL2 Private Quant V4
|
@@ -29,9 +31,6 @@ Feel free to send in pull requests or use this code however you'd like.\
|
|
29 |
- Windows/Linux support (don't have mac)
|
30 |
- Colab version will use this with markdown parameters
|
31 |
|
32 |
-
- [Upload folder](https://huggingface.co/Anthonyg5005/hf-scripts/blob/unfinished/upload%20to%20hub.py)
|
33 |
-
- Uploads user specified folder to specified repo, can create private repos too
|
34 |
-
|
35 |
## other recommended files
|
36 |
|
37 |
- [Download models (download HF Hub models) [Oobabooga]](https://github.com/oobabooga/text-generation-webui/blob/main/download-model.py)
|
@@ -43,6 +42,9 @@ Feel free to send in pull requests or use this code however you'd like.\
|
|
43 |
|
44 |
- EXL2 Private Quant
|
45 |
- Allows you to quantize to exl2 using colab. This version creates a exl2 quant to upload to private repo. Should work on any Linux jupyterlab server with CUDA, ROCM should be supported by exl2 but not tested.
|
|
|
|
|
|
|
46 |
|
47 |
- Download models
|
48 |
- Make sure you have [requests](https://pypi.org/project/requests/) and [tqdm](https://pypi.org/project/tqdm/) installed. You can install them with '`pip install requests tqdm`'. To use the script, open a terminal and run '`python download-model.py USER/MODEL:BRANCH`'. There's also a '`--help`' flag to show the available arguments. To download from private repositories, make sure to login using '`huggingface-cli login`' or (not recommended) `HF_TOKEN` environment variable.
|
|
|
19 |
|
20 |
- [EXL2 Private Quant V3](https://colab.research.google.com/drive/1Vc7d6JU3Z35OVHmtuMuhT830THJnzNfS?usp=sharing) **(COLAB)**
|
21 |
|
22 |
+
- [Upload folder to repo](https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/upload%20folder%20to%20repo.py)
|
23 |
+
|
24 |
## work in progress/not tested ([unfinished](https://huggingface.co/Anthonyg5005/hf-scripts/tree/unfinished) branch)
|
25 |
|
26 |
- EXL2 Private Quant V4
|
|
|
31 |
- Windows/Linux support (don't have mac)
|
32 |
- Colab version will use this with markdown parameters
|
33 |
|
|
|
|
|
|
|
34 |
## other recommended files
|
35 |
|
36 |
- [Download models (download HF Hub models) [Oobabooga]](https://github.com/oobabooga/text-generation-webui/blob/main/download-model.py)
|
|
|
42 |
|
43 |
- EXL2 Private Quant
|
44 |
- Allows you to quantize to exl2 using colab. This version creates a exl2 quant to upload to private repo. Should work on any Linux jupyterlab server with CUDA, ROCM should be supported by exl2 but not tested.
|
45 |
+
|
46 |
+
- Upload folder to repo
|
47 |
+
- Uploads user specified folder to specified repo, can create private repos too. Not the same as git commit and push, instead uploads any additional files.
|
48 |
|
49 |
- Download models
|
50 |
- Make sure you have [requests](https://pypi.org/project/requests/) and [tqdm](https://pypi.org/project/tqdm/) installed. You can install them with '`pip install requests tqdm`'. To use the script, open a terminal and run '`python download-model.py USER/MODEL:BRANCH`'. There's also a '`--help`' flag to show the available arguments. To download from private repositories, make sure to login using '`huggingface-cli login`' or (not recommended) `HF_TOKEN` environment variable.
|