The 2nd ACM International Workshop on Interactive Multimedia on Mobile and Portable Devices (IMMPD2012)
In Conjunction with ACM Multimedia 2012, Nara, Japan
Prof. Winston Hsu, National Taiwan University, Taipei, Taiwan
Title: Emerging Challenges and Opportunities in Exploiting Mobile Photos and Videos
Abstract: The prevalence of mobile cameras and social media sharing services have drastically increased the volume of photos and videos, which pose a new challenge for developing efficient retrieval, mining, and visualization approaches. Meanwhile, the mobile devices have almost been the must in daily life for capturing important moments, browsing the photos and videos shared by the families and friends, or working as Q&A devices for understanding the objects of interests (e.g., landmarks, products, video programs, or restaurant reviews, etc.) by taking a snapshot and searching the related information. In this talk, we will first look into the emerging opportunities and promising solutions to manipulate such large-scale data through mobile devices. We will point the properties that we can leverage from mobile devices and go through the challenges (e.g., small mobile screen size, low bandwidth, limited storage and computing power), which further pose new constraints in photo/video content analysis and retrieval in the emerging mobile platform. Meanwhile, instead of treating the user-generated photos/videos as independent workloads, we will demonstrate the exciting applications in mobile recommendation and human activity mining as, in a holistic manner, analyzing such sheer amount of photos and videos, which are generally to record people activities and those the most memorable for the users. We will also review emerging industrial and academic research problems and the pilot experiments over million-scale photo/video collections.
Bio: Dr. Winston Hsu is an active researcher dedicated to novel algorithms and systems for ultra-large-scale image/video retrieval, social media mining/recommendation, multimedia analytics, and mobile and cloud-based multimedia applications. He is keen to realizing advanced researches towards business deliverables. He has been recognized with technical awards and best paper awards in multimedia research community. He delivered keynote speeches and lectured several state-of-the-art tutorials in top conferences. He is in the technical committee for top conferences (ACM Multimedia, SIGIR, and WWW) and the organizing committee for ACM Multimedia and ICME. He is on the Editorial Board for IEEE Multimedia Magazine and Journal of Multimedia (JMM). He is an Associate Professor in the Department of Computer Science and Information Engineering, NTU and the founder of MiRA (Multimedia indexing, Retrieval, and Analysis) Research Group in Communication and Multimedia Lab (CMLab). He received his Ph.D. (2007) degree from Columbia University, New York. Before that, he worked for years as a founding engineer in the multimedia software company, CyberLink Corp., where experiencing Engineer, Project Leader, and R&D Manager. He is recently awarded 2011 Ta-You Wu Memorial Award, a national and prestigious recognition for young researchers.
With the development of silicon technologies, mobile and portable electronics devices, such as personal computers, mobile phones, digital cameras, and PDA, have become ubiquitous for people's daily life. These devices provide multimedia sources for entertainment, communication, and so on. How to design user interfaces of these products that enable natural, intuitive and fun interaction is one of the main challenges the multimedia community is facing.
Considering that mobile and portable devices are usually supplied with multiple sensors (e.g., camera and microphone), how to employ multimodal information for interaction has recently received much attention in both academia and industry. But interactive multimedia is still an under-explored field. Many challenges exist when moving to multimodal interaction: for example, how to annotate and search huge data acquired by using multiple sensors, especially in the unconstrained end-user environments? how to effectively extract and select representative multimedia features for human behavior recognition? and how to select the fusion strategy of multimodal data for a given application? To address these challenges, we must adapt the existing approaches or find new solutions suitable for multimedia interaction on mobile and portable devices.
This workshop will bring together researchers from both academia and industry in domains including computer vision, audio and speech processing, machine learning, pattern recognition, communications, human-computer interaction, and media technology to share and discuss recent advances in interactive multimedia. Topics include, but are not limited to: