Liu-Xiang commited on
Commit
8e8677d
1 Parent(s): a4b8a93

Training in progress, step 176

Browse files
adapter_config.json CHANGED
@@ -19,12 +19,12 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "gate_proj",
23
- "q_proj",
24
- "k_proj",
25
  "up_proj",
26
- "down_proj",
27
  "o_proj",
 
 
 
 
28
  "v_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
22
  "up_proj",
 
23
  "o_proj",
24
+ "q_proj",
25
+ "down_proj",
26
+ "k_proj",
27
+ "gate_proj",
28
  "v_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae5e3cd03d1c696300386f43d6727e2b9621e02d023d72984aaea5bda1e03b8d
3
  size 50503848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3817f9fc0a99dc28ac16366afab67ab13b7938683b63f0b325b8bf77adc7ab2
3
  size 50503848
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:040237bff03311f4af156509ef0f6a9e4644b5cf5652e82039ab80f596cccb37
3
  size 1999828576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:917ce394801a3f401146bfa2ecd22c70598b19c4e236330f4ae722fc5189323c
3
  size 1999828576
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6946240fc8cd0c0a212fc6a0ea098dcfffc9a179ba5702314c89476e21e74154
3
  size 200291080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2213f0055f1f3591aff24c1875ee43377aa0f860738e4543d19661c70c65bbb4
3
  size 200291080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40f35c16cc80ea40f987a35d01f6091227a4ca15db77c062fc8d69104f9887ee
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee97143aaf5b9c27672cbff1fe90ee761a2c8b9964d50574862150c72ab0e190
3
  size 4792