Abstract
The boom of mobile devices and cloud services has led to an explosion of personal photo and video data. However, due to the missing user-generated metadata such as titles or descriptions, it usually takes a user a lot of swipes to find some video on the cell phone. To solve the problem, we present an innovative idea called Visual Memory QA which allow a user not only to search but also to ask questions about her daily life captured in the personal videos. The proposed system automatically analyzes the content of personal videos without usergenerated metadata, and offers a conversational interface to accept and answer questions. To the best of our knowledge, it is the first to answer personal questions discovered in personal photos or videos. The example questions are "what was the lat time we went hiking in the forest near San Francisco?"; "did we have pizza last week?"; "with whom did I have dinner in AAAI 2015?".
Cite
CITATION STYLE
Jiang, L., Cao, L. L., Kalantidis, Y., Farfade, S., & Hauptmann, A. G. (2017). Visual memory QA: Your personal photo and video search agent. In 31st AAAI Conference on Artificial Intelligence, AAAI 2017 (pp. 5093–5094). AAAI press. https://doi.org/10.1609/aaai.v31i1.10537
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.