Upload with huggingface_hub
Browse files- .gitattributes +4 -0
- README.md +34 -0
- ggml-model-q4_0.bin.4 +3 -0
- ggml-model-q4_0.bin.5 +3 -0
- ggml-model-q4_0.bin.6 +3 -0
- ggml-model-q4_0.bin.7 +3 -0
.gitattributes
CHANGED
@@ -35,3 +35,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
35 |
ggml-model-q4_0.bin.1 filter=lfs diff=lfs merge=lfs -text
|
36 |
ggml-model-q4_0.bin.2 filter=lfs diff=lfs merge=lfs -text
|
37 |
ggml-model-q4_0.bin.3 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
35 |
ggml-model-q4_0.bin.1 filter=lfs diff=lfs merge=lfs -text
|
36 |
ggml-model-q4_0.bin.2 filter=lfs diff=lfs merge=lfs -text
|
37 |
ggml-model-q4_0.bin.3 filter=lfs diff=lfs merge=lfs -text
|
38 |
+
ggml-model-q4_0.bin.4 filter=lfs diff=lfs merge=lfs -text
|
39 |
+
ggml-model-q4_0.bin.5 filter=lfs diff=lfs merge=lfs -text
|
40 |
+
ggml-model-q4_0.bin.6 filter=lfs diff=lfs merge=lfs -text
|
41 |
+
ggml-model-q4_0.bin.7 filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
### alpaca.cpp 65B ggml model weight
|
3 |
+
### make 65B ggml story
|
4 |
+
#### 1. clone 65B model data
|
5 |
+
```shell
|
6 |
+
git clone https://huggingface.co/datasets/nyanko7/LLaMA-65B/
|
7 |
+
```
|
8 |
+
#### 2. clone alpaca.cpp
|
9 |
+
```shell
|
10 |
+
git clone https://github.com/antimatter15/alpaca.cpp
|
11 |
+
```
|
12 |
+
#### 3. weight quantize.sh
|
13 |
+
```shell
|
14 |
+
|
15 |
+
mv LLaMA-65B/tokenizer.model ./
|
16 |
+
python convert-pth-to-ggml.py ../LLaMA-65B/ 1
|
17 |
+
|
18 |
+
cd alpaca.cpp
|
19 |
+
mkdir -p models/65B
|
20 |
+
mv ../LLaMA-65B/ggml-model-f16.bin models/65B/
|
21 |
+
mv ../LLaMA-65B/ggml-model-f16.bin.* models/65B/
|
22 |
+
bash quantize.sh 65B
|
23 |
+
```
|
24 |
+
|
25 |
+
#### 4. upload weight file
|
26 |
+
##### Upload is slower. The upload is taking almost 2 days, I decided to curve the upload
|
27 |
+
##### I using https://tmp.link/ as temp store
|
28 |
+
##### I using colab and huggingface api upload
|
29 |
+
|
30 |
+
### run
|
31 |
+
```shell
|
32 |
+
git clone https://github.com/antimatter15/
|
33 |
+
./chat -m alpaca.cpp_65b_ggml/ggml-model-q4_0.bin
|
34 |
+
```
|
ggml-model-q4_0.bin.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ee474051e418c5732b7949190b084d9d679db447f83c1de0d2a82daaa1a0cfa
|
3 |
+
size 5105954665
|
ggml-model-q4_0.bin.5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a45aa05e7212bd6782790722d68056c5419667ea6b564ccc94bbcb8111d79b8b
|
3 |
+
size 5105954665
|
ggml-model-q4_0.bin.6
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a58fda714b759c28ad5e4c1d8bf8fda7b158fd5e4c4a49f851f36342fa97a105
|
3 |
+
size 5105954665
|
ggml-model-q4_0.bin.7
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3540cfcbcda33c223c6b0d606034adbd78f17e0e5de1582b78795e78754f7a8
|
3 |
+
size 5105954665
|