An overview of NLP evaluation metrics and distance/similarity measures
Free Friday Talk, Ebot7 GmbH, Munich, Germany
Evaluation in NLP is an open problem especially when it comes to the evaluation of generated text. It’s a really hard problem and therefore there is no single go to evaluation metric available. What makes this problem even more hard is that metrics need to evaluate competing goals i.e Correctness (quality)/Specicity (diversity).