Personal Multimedia Preservation by Vasileios Mezaris Claudia Niederée & Robert H. Logie

Personal Multimedia Preservation by Vasileios Mezaris Claudia Niederée & Robert H. Logie

Author:Vasileios Mezaris, Claudia Niederée & Robert H. Logie
Language: eng
Format: epub
Publisher: Springer International Publishing, Cham


An overview of this method is illustrated in Fig. 5.14. This method takes as input a photo collection that has been captured by a user attending an event and a textual description of the event that contains the name, location and time of the event. Web photos that are related to the event are retrieved and downloaded, their visual content is analysed and a selection of them is provided to the user as additional contextual content.

First, the user’s photo collection is fed into an image analysis component. Using the Caffe framework [181] and the 22-layer GoogLeNet Deep Convolutional Neural Network (DCNN) pre-trained model [377], the loss1/classifier and loss3/classifier layers’ output is extracted (resulting in two 1000-dimensional vectors). The L*a*b* colourspace colour histogram of each photo is also calculated, using 8 bins for the L* plane, 32 bins for the a* plane and 32 bins for the b* plane (resulting in a 72-dimensional vector). The 2072-dimensional concatenated feature vector of the DCNN layers output and the L*a*b* histogram are used to represent each photo in the user collection. The textual description that must be provided by the user consists of the name of the event (e.g. Olympic Games), the location (e.g. London) where the event took place and the date that the event took place (e.g. 2012). This information is then used to create a set of queries which are sent to web search engines, and a pool of data is collected, from which the contextual information will be selected. The following three web queries are performed:1.A query to the Google text search engine; the P first web pages and the Y first YouTube links of the search results are kept. Moreover, the photos contained in the returned web pages are also collected. The query to the Google text search engine is formed by combining the textual description provided by the user, using AND (&) (i.e. event & location & time).



Download



Copyright Disclaimer:
This site does not store any files on its server. We only index and link to content provided by other sites. Please contact the content providers to delete copyright contents if any and email us, we'll remove relevant links or contents immediately.