.. | ||
data_dir | ||
Genera_evaluation.md | ||
InternLM2_7B_chat_eval.py | ||
metric.py | ||
Professional_evaluation.md | ||
Qwen1_5-0_5B-Chat_eval.py | ||
qwen_generation_utils.py | ||
README_EN.md | ||
README.md |
EmoLLM Evaluation
General Metrics Evaluation
- For specific metrics and methods, see General_evaluation.md
Metric | Value |
---|---|
ROUGE-1 | 27.23% |
ROUGE-2 | 8.55% |
ROUGE-L | 17.05% |
BLEU-1 | 26.65% |
BLEU-2 | 13.11% |
BLEU-3 | 7.19% |
BLEU-4 | 4.05% |
Professional Metrics Evaluation
- For specific metrics and methods, see Professional_evaluation_EN.md
Metric | Value |
---|---|
Comprehensiveness | 1.32 |
Professionalism | 2.20 |
Authenticity | 2.10 |
Safety | 1.00 |