Update README.md
Browse filesProctora is a MoE model made of
- OpenPipe/mistral-ft-optimized-1227 as a base model
- SanjiWatsuki/Kunoichi-7B as a first expert dedicated to RP tasks.
- samir-fama/SamirGPT-v1 as a second expert for factual answers.
I do not have yet any metrics for this model but subjective ones. It was made at first out of curiosity and experimentation.
My goal is to produce a model excellent at being a game master for RPG sessions. However being dissatisfied with the existing evaluation tool-sets, I decided to create my own (still a WIP on 01/16/24). And among my collection of models of small/medium models Proctora gave me the best results to evaluate the answers produced by other LLMs. Therefore, I surprisingly settled with it and gave it an appropriate name according to the task.
TLDR: Proctora is a tool for a tool!
I doubt this model will be useful for the community. I publish it for the sake of transparency in my creative process.