| .. | ||
| data_dir | ||
| Genera_evaluation.md | ||
| InternLM2_7B_chat_eval.py | ||
| metric.py | ||
| Professional_evaluation.md | ||
| Qwen1_5-0_5B-Chat_eval.py | ||
| qwen_generation_utils.py | ||
| README_EN.md | ||
| README.md | ||
EmoLLM Evaluation
General Metrics Evaluation
- For specific metrics and methods, see General_evaluation.md
| Metric | Value |
|---|---|
| ROUGE-1 | 27.23% |
| ROUGE-2 | 8.55% |
| ROUGE-L | 17.05% |
| BLEU-1 | 26.65% |
| BLEU-2 | 13.11% |
| BLEU-3 | 7.19% |
| BLEU-4 | 4.05% |
Professional Metrics Evaluation
- For specific metrics and methods, see Professional_evaluation_EN.md
| Metric | Value |
|---|---|
| Comprehensiveness | 1.32 |
| Professionalism | 2.20 |
| Authenticity | 2.10 |
| Safety | 1.00 |