A while ago Tianjin University released a study, which defined a LLM model evaluation and implementation taxonomy.
The image below shows the taxonomy of major categories and sub-categories used by the study for LLM evaluation.
Considering the taxonomy, much focus has been on Evaluation Organisation, Knowledge & Capability and Specialised LLMs from a technology perspective.
From an enterprise perspective, the concerns and questions raised centre around Alignment and Safety.
This survey aims to create an extensive perspective on how LLMs should be evaluated.
The study categorises the evaluation of LLMs into three major groups:
- Knowledge and Capability Evaluation,
- Alignment Evaluation and
- Safety Evaluation.
What I also found interesting from the study, was the focus on more complex metrics like reasoning and tool learning, as seen below:
LLMs have introduced significant complexity and have shown certain tendencies by revealing behaviours indicative of risks and demonstrating abilities to perform higher-order tasks in current evaluations.
Consequently, creating a taxonomy like this to reference can help ensure that due diligence is followed when assessing LLMs.
Find the full study here.