David Cody Taupo Lingan PRO
rizzware
AI & ML interests
Incoming @ ? [hopefully an open source GPU acceleration team]
Recent Activity
liked
a Space
about 1 month ago
huggingface/open-source-ai-year-in-review-2024
new activity
5 months ago
databricks/dbrx-instruct:Fine-tune dbrx via Hugging Face Trainer vs. LLM-Foundry
posted
an
update
5 months ago
Question about LightEval π€:
I've been searching for an LLM evaluation suite that can, out-of-the-box, compare the outputs of a model(s) without any enhancements vs. the same model with better prompt engineering, vs. the same model with RAG vs. the same model with fine-tuning.
I unfortunately have not found a tool that fits my exact description, but of course I ran into LightEval.
A huge pain-point of building large-scale projects that use LLMs is that prior to building an MVP, it is difficult to evaluate whether better prompt engineering, or RAG, or fine-tuning, or some combination of all is needed for satisfactory LLM output in terms of the project's given use case.
Time and resources is then wasted R&D'ing exactly what LLM enhancements are needed.
I believe an out-of-the-box solution to compare models w/ or w/out the aforementioned LLM enhancements could help teams of any size better decide what LLM enhancements are needed prior to building.
I wanted to know if the LightEval team or Hugging Face in general is thinking about such a tool.
Organizations
rizzware's activity
Fine-tune dbrx via Hugging Face Trainer vs. LLM-Foundry
#58 opened 9 months ago
by
rizzware
Fine-tune dbrx via Hugging Face Trainer vs. LLM-Foundry
#58 opened 9 months ago
by
rizzware