Bridging the Semantic Gap in Image and Video Analysis by Halina Kwaśnicka & Lakhmi C. Jain

Bridging the Semantic Gap in Image and Video Analysis by Halina Kwaśnicka & Lakhmi C. Jain

Author:Halina Kwaśnicka & Lakhmi C. Jain
Language: eng
Format: epub
Publisher: Springer International Publishing, Cham


5.6 Testing Scenarios

As an simple illustration, generic mug model will be used (similar to presented in previous sections), with two distinct modalities of incoming data: RGB image and RGB-D image. In first scenario, edge-based analysis [37] is applied, with two basic concepts—Cylinder and Arc, corresponding to two mug parts—body and handle. Second scenario uses depth data [38], and two basic surface concepts—Cylinder and Thorus.

5.6.1 Data Acquisition

As an input data, Complex scene 3 from WUT Visual Perception Dataset [39] was used, which contains a recorded trajectory (77 points) “around the table”. The set was acquired using Microsoft Kinect sensor, and it contains, for every recorded position, a pair of images, aligned with each other: RGB image (Fig. 5.9a) and depth map (Fig. 5.9b). The selected scene contains three cylindrical objects—two mugs and one coffee jar, as well as some other kitchen utensils.

The data was acquired with hand-held sensor, thus there is no ground-truth position data and the trajectory was recovered using visual odometry solution [5].

Fig. 5.9Test scene: a RGB image, b depth map



Download



Copyright Disclaimer:
This site does not store any files on its server. We only index and link to content provided by other sites. Please contact the content providers to delete copyright contents if any and email us, we'll remove relevant links or contents immediately.