Would you be willing to do this on a smaller model?

#2
by DazzlingXeno - opened

As above, for people like myself who cannot seem to get Llamacpp working on Windows? Say on Mistral small 22b, Qwen 32b or similar around the 20-35b range?

Cheers

Fancy seeing you here ;D

Fancy seeing you here ;D

Hehehe

I do plan to try other models, yeah.

Qwen 32b

I actually gave Qwen-2.5-72b a try, but the result wasn't so good. I suspect the Qwen models haven't been trained on as many books as Mistral and Command-R. So the resulting model couldn't easily write in the style of .

Mistral small 22b

I'm actually working on a 22b at the moment, experimental/custom architecture. If it's coherent enough, I'll release it.

Other than that I was planning to try gemma2-27b as it can be cohered into writing pretty well, though it only has an 8192 context.

cannot seem to get Llamacpp working on Windows

This should be fixable by the way. I haven't got a windows machine myself, but I'm pretty sure llama.cpp can run on pretty much anything. There were some big changes with llama.cpp earlier this month where they deprecated make, and it was a bit unstable for a week after that. But I'm betting the latest main branch will build fine with cmake

I do plan to try other models, yeah.

Qwen 32b

I actually gave Qwen-2.5-72b a try, but the result wasn't so good. I suspect the Qwen models haven't been trained on as many books as Mistral and Command-R. So the resulting model couldn't easily write in the style of .

Mistral small 22b

I'm actually working on a 22b at the moment, experimental/custom architecture. If it's coherent enough, I'll release it.

Other than that I was planning to try gemma2-27b as it can be cohered into writing pretty well, though it only has an 8192 context.

cannot seem to get Llamacpp working on Windows

This should be fixable by the way. I haven't got a windows machine myself, but I'm pretty sure llama.cpp can run on pretty much anything. There were some big changes with llama.cpp earlier this month where they deprecated make, and it was a bit unstable for a week after that. But I'm betting the latest main branch will build fine with cmake

That sounds great I look forward to seeing whatever you release.

I tried a few months ago and had some luck with LlamaCPP In that I got it to load a model but I just couldn't get it to work with CUDA/CuBLAS no matter what I did. I'll be honest though I am really strapped for time having 2 young kids and working full time. But I'll give it another gom if you know of any decent guides out there (unlike as you don't use Windows) let me know.

Anyway Merry Christmas!

Sign up or log in to comment