model generates inconsistent output

#40
by ahoo1260 - opened

I want to use Llama-3.2-1B-Instruct model, and although I have set "temperature": 0.0, "top_p":0.0 and "top_k":0, it still generates inconsistent output. This is how my pipeline looks like:

pipe = pipeline(
    "text-generation",
    model=model_id,
    torch_dtype=torch.bfloat16,
    device_map="mps",
        model_kwargs={"temperature": 0.0,
                  "do_sample":True,
                              "top_p":0.0,
                              "top_k":0,},
)

Any idea how to solve this issue?

Sign up or log in to comment