论文标题
如果可以的话,请抓住我:检测秘密地理位置的新任务(CGL)
Catch Me if You Can: A Novel Task for Detection of Covert Geo-Locations (CGL)
论文作者
论文摘要
大多数视觉场景理解计算机视觉领域的任务都涉及对场景中存在的对象的识别。图像区域之类的场景区域,转弯和其他模糊区域也包含至关重要的信息,用于特定的监视任务。本文提出的任务涉及设计智能视觉援助以识别图像中此类位置的任务,该辅助是有可能从对手身上造成迫在眉睫的威胁,或者是作为需要进一步调查的目标区域而出现的。隐藏在遮挡物体后面的秘密位置(CGL)是隐藏的3D位置,从视点(相机)无法检测到。因此,这涉及描述围绕遮挡物体外边界的投影的特定图像区域,作为围绕潜在藏身处访问的位置。 CGL检测发现了在军事反企业行动中的应用,对探索性机器人的路径计划进行了监视。给定RGB图像,目标是识别2D场景中的所有CGL。识别此类区域将需要了解遮盖物品的3D边界(支柱,家具),以及与现场相邻地区的空间位置。我们将其作为一项新任务,称为秘密地理位置(CGL)检测。将图像的任何区域的分类为CGL(作为隐藏藏身处的遮挡对象的边界子细分)需要检查遮挡物体的边界之间的3D关系及其邻域及其周围环境。我们的方法成功地从单个RGB图像中提取相关的深度特征,并定量地对现有对象检测和分割模型进行了显着改进,该模型对CGL检测进行了调整和训练。我们还介绍了一个新型的手工注释的CGL检测数据集,其中包含1.5K现实世界图像进行实验。
Most visual scene understanding tasks in the field of computer vision involve identification of the objects present in the scene. Image regions like hideouts, turns, & other obscured regions of the scene also contain crucial information, for specific surveillance tasks. Task proposed in this paper involves the design of an intelligent visual aid for identification of such locations in an image, which has either the potential to create an imminent threat from an adversary or appear as the target zones needing further investigation. Covert places (CGL) for hiding behind an occluding object are concealed 3D locations, not detectable from the viewpoint (camera). Hence this involves delineating specific image regions around the projections of outer boundary of the occluding objects, as places to be accessed around the potential hideouts. CGL detection finds applications in military counter-insurgency operations, surveillance with path planning for an exploratory robot. Given an RGB image, the goal is to identify all CGLs in the 2D scene. Identification of such regions would require knowledge about the 3D boundaries of obscuring items (pillars, furniture), their spatial location with respect to the neighboring regions of the scene. We propose this as a novel task, termed Covert Geo-Location (CGL) Detection. Classification of any region of an image as a CGL (as boundary sub-segments of an occluding object that conceals the hideout) requires examining the 3D relation between boundaries of occluding objects and their neighborhoods & surroundings. Our method successfully extracts relevant depth features from a single RGB image and quantitatively yields significant improvement over existing object detection and segmentation models adapted and trained for CGL detection. We also introduce a novel hand-annotated CGL detection dataset containing 1.5K real-world images for experimentation.