论文标题
严格评估模型推断精度使用语言基数
Rigorous Assessment of Model Inference Accuracy using Language Cardinality
论文作者
论文摘要
诸如有限状态自动机之类的模型通过捕获可观察到的事件在执行过程中可观察到的事件的序列被广泛用于抽象软件系统的行为。然而,模型在实践中很少存在,而当它们这样做时,模型很容易过时。此外,手动构建和维护模型是昂贵且容易出错的。结果,已经提出了从执行轨迹自动构建模型的各种模型推理方法来解决这些问题。 但是,对推断模型进行系统和可靠的准确性评估仍然是一个空旷的问题。即使给出了参考模型,大多数现有的模型精度评估方法也可能返回误导和偏见的结果。这主要是由于它们依赖于有限数量的随机生成轨迹上的统计估计器,从而引入了有关估计的可避免的不确定性,并且对随机跟踪生成过程的参数敏感。 本文通过基于分析组合技术来开发系统的方法来解决此问题,该方法通过用确定性的准确性度量替换统计估计来最大程度地减少模型准确性评估中的偏差和不确定性。我们通过评估最先进的推理工具推断出的模型的准确性来证明我们的方法的一致性和适用性,以针对已建立的规范挖掘基准的参考模型推断。
Models such as finite state automata are widely used to abstract the behavior of software systems by capturing the sequences of events observable during their execution. Nevertheless, models rarely exist in practice and, when they do, get easily outdated; moreover, manually building and maintaining models is costly and error-prone. As a result, a variety of model inference methods that automatically construct models from execution traces have been proposed to address these issues. However, performing a systematic and reliable accuracy assessment of inferred models remains an open problem. Even when a reference model is given, most existing model accuracy assessment methods may return misleading and biased results. This is mainly due to their reliance on statistical estimators over a finite number of randomly generated traces, introducing avoidable uncertainty about the estimation and being sensitive to the parameters of the random trace generative process. This paper addresses this problem by developing a systematic approach based on analytic combinatorics that minimizes bias and uncertainty in model accuracy assessment by replacing statistical estimation with deterministic accuracy measures. We experimentally demonstrate the consistency and applicability of our approach by assessing the accuracy of models inferred by state-of-the-art inference tools against reference models from established specification mining benchmarks.