Llama-2-7b-guanaco
π Article | π» Colab | π Script
This is a llama-2-7b-chat-hf
model fine-tuned using QLoRA (4-bit precision) on the mlabonne/guanaco-llama2
dataset.
π§ Training
It was trained on a Google Colab notebook with a T4 GPU and high RAM.
π» Usage
# pip install transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mlabonne/llama-2-7b-miniguanaco"
prompt = "What is a large language model?"
tokenizer = AutoTokenizer.from_pretrained(model)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
sequences = pipeline(
f'<s>[INST] {prompt} [/INST]',
do_sample=True,
top_k=10,
num_return_sequences=1,
eos_token_id=tokenizer.eos_token_id,
max_length=200,
)
for seq in sequences:
print(f"Result: {seq['generated_text']}")
Output:
A large language model is a type of artificial intelligence (AI) model that is trained to generate human-like language. The models can be trained on text from a specific genre, such as news articles, or on a large corpus of text, such as the internet. They can then be used to generate text, such as articles, stories or even entire books. These models are often used in applications such as chatbots, language translation and content generation. They have been used to write books such as: "The Last Days of New Paris" by China MiΓ©ville.
The large models are also used for many other applications such as:
- Translation
- Summarization
- Sentiment Analysis
- Text classification
- Generative writing (creates articles, stories, and more.)
- Conversational language generation.
- Downloads last month
- 26