论文标题
MAVVStream:使用提取的视频内容扩展情况监视的情况数据库功能
MavVStream: Extending Database Capabilities for Situation Monitoring Using Extracted Video Contents
论文作者
论文摘要
基于查询的视频情况检测(与手动或定制算法相反)对于诸如流量监控,监视1和其他类型的环境/基础架构监视等不同应用至关重要。视频内容在不同的对象类型和背景信息方面很复杂。因此,除了使用最新的视觉技术(包括基于深度学习)提取复杂内容外,它们的表示以及查询构成了各种挑战。一旦我们有代表性来容纳提取的内容,对此的临时查询将需要新的操作员,以及他们的语义和算法来进行有效的计算。扩展数据库框架(表示和实时查询)用于处理一次视频内容上的查询一次至关重要,这是朝着该方向迈出的第一步。在本文中,我们将传统关系扩展到R ++(向量属性)和ARRABLE,以适应视频内容,并将CQL(连续查询语言)与一些新操作员扩展到扩展表示形式。本文讨论了向后兼容,易用性,新操作员(包括空间和时间)以及有效执行的算法。查询类是根据其复杂性来确定的,以评估视频内容。已经使用了大量的大型和大型视频数据集(一些文献中的一些数据集)来展示如何在可用数据集上使用我们的作品。通过手动基础真理,有效的评估以及算法的鲁棒性的正确性。我们的主要贡献是为一个问题造成一个框架,该框架是基于新思想的大数据分析的一部分变得非常重要的。
Query-based video situation detection (as opposed to manual or customized algorithms) is critical for diverse applications such as traffic monitoring, surveillance1 , and other types of environmental/infrastructure monitoring. Video contents are complex in terms of disparate object types and background information. Therefore, in addition to extracting complex contents using the latest vision technologies (including deep learning-based), their representation as well as querying pose different kinds of challenges. Once we have a representation to accommodate extracted contents, ad-hoc querying on that will need new operators, along with their semantics and algorithms for their efficient computation. Extending database framework (representation and real-time querying) for processing queries on video contents extracted only once is critical and this effort is an initial step in that direction. In this paper, we extend the traditional relation to R++ (vector attributes) and arrables to accommodate video contents and extend CQL (Continuous Query Language) with a few new operators to query situations on the extended representation. Backward compatibility, ease-of-use, new operators (including spatial and temporal), and algorithms for efficient execution are discussed in this paper. Classes of queries are identified based on their complexity to evaluate with respect to video content. A large number of small and large video datasets have been used (some from the literature) to show how our work can be used on available datasets. Correctness of queries with manual ground truth, efficient evaluation as well as robustness of algorithms is demonstrated. Our main contribution is couching a framework for a problem that is becoming very important as part of big data analytics based on a novel idea.