Loading...
Please wait, while we are loading the content...
Similar Documents
Proceedings of the 2016 ACM workshop on Vision and Language Integration Meets Multimedia Fusion (iV&L-MM '16)
| Content Provider | ACM Digital Library |
|---|---|
| Editor | Moens, Marie-Francine Tuytelaars, Tinne Pastra, Katerina Saenko, Kate |
| Copyright Year | 2016 |
| Abstract | It is our great pleasure to welcome you to the ACM Multimedia 2016 Workshop Vision and Language Integration Meets Multimedia Fusion (iV&L-MM 2016) in Amsterdam, The Netherlands on October 16, 2016. Multimodal information fusion both at the signal and the semantics levels is a core part in most multimedia applications, including multimedia indexing, retrieval, summarization and others. Early or late fusion of modality-specific processing results has been addressed in multimedia prototypes since their very early days, through various methodologies including rule-based approaches, information-theoretic models and machine learning. Vision and Language are two of the predominant modalities that are being fused and which have attracted special attention in international challenges with a long history of results, such as TRECVid, ImageClef and others. During the last decade, vision-language semantic integration has attracted attention from traditionally non-interdisciplinary research communities, such as Computer Vision and Natural Language Processing. This is due to the fact that one modality can greatly assist the processing of another providing cues for disambiguation, complementary information and noise/error filtering. The latest boom of deep learning methods has opened up new directions in joint modelling of visual and co-occurring verbal information in multimedia discourse. The proceedings contain seven selected long papers, which have been orally presented at the workshop, and three abstracts of the invited keynote speeches. The papers and abstracts discuss data collection, representation learning, deep learning approaches, matrix and tensor factorization methods and graph based clustering with regard to the fusion of multimedia data. A variety of applications is presented including image captioning, summarization of news, video hyperlinking, sub-shot segmentation of user generated video, cross-modal classification, cross-modal questionanswering, and the detection of misleading metadata of user generated video. The call for papers attracted submissions from Europe, Asia, Australia and the United States. We received 15 long papers of which the program committee reviewed and accepted 7, resulting in an acceptance rate of about 47%. The accepted long papers are orally presented at the workshop. We also encourage attendees to attend the keynote talk presentations. These valuable and insightful talks can and will guide us to a better understanding of the future: Explain and Answer: Relating Natural Language and Visual Recognition, Marcus Rohrbach (University of California Berkeley, USA) Jointly Representing Images and Text: Dependency Graphs, Word Senses, and Multimodal Embeddings, Frank Keller (University of Edinburgh, UK) Beyond Language and Vision, Towards Truly Multimedia Integration, Tat-Seng Chua (National University of Singapore, Singapore). |
| ISBN | 9781450345194 |
| Language | English |
| Publisher | Association for Computing Machinery (ACM) |
| Publisher Date | 2016-10-16 |
| Access Restriction | Subscribed |
| Content Type | Text |
| Resource Type | Conference Proceedings |