The rapid growth in the amount of photos/videos online necessitates for social media companies to automatically extract knowledge structures (concepts) from photos and videos to provide diverse multimedia-related services such as event detection and summarization. However, real-world photos and videos aggregated in social media sharing platforms (e.g., Flickr and Instagram) are complex and noisy, and extracting semantics and sentics from the multimedia content alone is a very difficult task because suitable concepts may be exhibited in different representations. Since semantics and sentics knowledge structures are very useful in multimedia search, retrieval, and recommendation, it is desirable to analyze UGCs from multiple modalities for a better understanding. To this end, we first present the EventBuilder system that deals with semantics understanding and automatically generates a multimedia summary for a given event in real-time by leveraging different social media such as Wikipedia and Flickr. Subsequently, we present the EventSensor system that aims to address sentics understanding and produces a multimedia summary for a given mood. It extracts concepts and mood tags from visual content and textual metadata of UGCs, and exploits them in supporting several significant multimedia-related services such as a musical multimedia summary. Moreover, EventSensor supports sentics-based event summarization by leveraging EventBuilder as its semantics engine component. Experimental results confirm that both EventBuilder and EventSensor outperform their baselines and efficiently summarize knowledge structures on the YFCC100M dataset.
CITATION STYLE
Shah, R., & Zimmermann, R. (2017). Event Understanding (pp. 59–99). https://doi.org/10.1007/978-3-319-61807-4_3
Mendeley helps you to discover research relevant for your work.