论文标题
CLEVAL:文本检测和识别任务的角色级评估
CLEval: Character-Level Evaluation for Text Detection and Recognition Tasks
论文作者
论文摘要
尽管文本检测和识别方法最近取得了成功,但现有的评估指标未能在这些方法之间提供公平可靠的比较。此外,没有考虑OCR任务特征的端到端评估指标。以前的端到端度量包含在检测和识别任务中应用的二进制评分过程中的级联错误。忽略部分正确的结果会增加定量和定性分析之间的差距,并防止细粒度评估。基于字符是文本的关键要素的事实,我们在此提出了一个字符级评估度量标准(CLEVAL)。在Cleval中,\ textIt {instance匹配}进程处理拆分并合并检测案例,\ textit {评分过程}进行字符级评估。通过汇总字符级别的得分,CLEVAL度量可以从末端表现的角度对由检测和识别组成的端到端结果进行精细元素评估。我们认为,我们的指标可以在开发和分析最先进的文本检测和识别方法中发挥关键作用。评估代码可在https://github.com/clovaai/cleval上公开获得。
Despite the recent success of text detection and recognition methods, existing evaluation metrics fail to provide a fair and reliable comparison among those methods. In addition, there exists no end-to-end evaluation metric that takes characteristics of OCR tasks into account. Previous end-to-end metric contains cascaded errors from the binary scoring process applied in both detection and recognition tasks. Ignoring partially correct results raises a gap between quantitative and qualitative analysis, and prevents fine-grained assessment. Based on the fact that character is a key element of text, we hereby propose a Character-Level Evaluation metric (CLEval). In CLEval, the \textit{instance matching} process handles split and merge detection cases, and the \textit{scoring process} conducts character-level evaluation. By aggregating character-level scores, the CLEval metric provides a fine-grained evaluation of end-to-end results composed of the detection and recognition as well as individual evaluations for each module from the end-performance perspective. We believe that our metrics can play a key role in developing and analyzing state-of-the-art text detection and recognition methods. The evaluation code is publicly available at https://github.com/clovaai/CLEval.