Text Generation
GGUF
medical
Elkhayyat17 commited on
Commit
1034235
1 Parent(s): 590d423

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -10
README.md CHANGED
@@ -24,8 +24,7 @@ quantized_by: Elkhayyat
24
 
25
 
26
 
27
- # CodeLlama 7B - GGUF
28
- - Model creator: [Meta](https://huggingface.co/meta-llama)
29
  - Original model: [meta](meta-llama/Llama-2-7b-chat-hf)
30
 
31
  <!-- description start -->
@@ -107,7 +106,7 @@ The following clients/libraries will automatically download models for you, prov
107
 
108
  ### In `text-generation-webui`
109
 
110
- Under Download Model, you can enter the model repo: Elkhayyat17/llama2-Med-gguf and below it, a specific filename to download, such as: codellama-7b.q4_K_M.gguf.
111
 
112
  Then click Download.
113
 
@@ -158,7 +157,7 @@ Windows CLI users: Use `set HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1` before running
158
  Make sure you are using `llama.cpp` from commit [d0cee0d36d5be95a0d9088b674dbb27354107221](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
159
 
160
  ```shell
161
- ./main -ngl 32 -m codellama-7b.q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "{prompt}"
162
  ```
163
 
164
  Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
@@ -231,7 +230,7 @@ from transformers import AutoTokenizer
231
  import transformers
232
  import torch
233
 
234
- model = "codellama/CodeLlama-7b-hf"
235
 
236
  tokenizer = AutoTokenizer.from_pretrained(model)
237
  pipeline = transformers.pipeline(
@@ -256,25 +255,20 @@ for seq in sequences:
256
  ```
257
 
258
  ## Model Details
259
- *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs).
260
 
261
  **Model Developers** Meta
262
 
263
  **Variations**
264
  ## Training Data
265
 
266
- All experiments reported here and the released models have been trained and fine-tuned using the same data as Llama 2 with different weights (see Section 2 and Table 1 in the [research paper](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/) for details).
267
 
268
  ## Evaluation Results
269
 
270
- See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper.
271
 
272
 
273
  ## Ethical Considerations and Limitations
274
 
275
- Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model.
276
 
277
- Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-user-guide](https://ai.meta.com/llama/responsible-user-guide).
278
 
279
  <!-- original-model-card end -->
280
  ---
 
24
 
25
 
26
 
27
+ # Llama2-Med - GGUF
 
28
  - Original model: [meta](meta-llama/Llama-2-7b-chat-hf)
29
 
30
  <!-- description start -->
 
106
 
107
  ### In `text-generation-webui`
108
 
109
+ Under Download Model, you can enter the model repo: Elkhayyat17/llama2-Med-gguf and below it, a specific filename to download, such as: ggml-model-Q4_K_M.gguf
110
 
111
  Then click Download.
112
 
 
157
  Make sure you are using `llama.cpp` from commit [d0cee0d36d5be95a0d9088b674dbb27354107221](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
158
 
159
  ```shell
160
+ ./main -ngl 32 -m c ggml-model-Q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "{prompt}"
161
  ```
162
 
163
  Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
 
230
  import transformers
231
  import torch
232
 
233
+ model = "meta-llama/Llama-2-7b-chat-hf"
234
 
235
  tokenizer = AutoTokenizer.from_pretrained(model)
236
  pipeline = transformers.pipeline(
 
255
  ```
256
 
257
  ## Model Details
 
258
 
259
  **Model Developers** Meta
260
 
261
  **Variations**
262
  ## Training Data
263
 
 
264
 
265
  ## Evaluation Results
266
 
 
267
 
268
 
269
  ## Ethical Considerations and Limitations
270
 
 
271
 
 
272
 
273
  <!-- original-model-card end -->
274
  ---