Surgical Quality Assessment and Gradient-based Signatures

The preprints of our paper on endoscopic image to video linking with fast parameterized gradient-based signatures and the demo paper of a tool for supporting surgical quality assessment (SQA), to appear in The Third International Conference on Multimedia Big Data (BigMM 2017), is available for download:

4th International Workshop on Mobile Multimedia Computing (MMC 2017)

We are very happy that the fourth iteration of our successful International Workshop on Mobile Multimedia Computing (MMC) has been accepted to be held at the International Conference  on Multimedia & Expo (ICME) 2017 in Hong Kong in July 2017. We are looking forward to receive your papers until February 27, 2017.

More information can be found on the MMC 2017 workshop website.

Tutorial at ACM CHIIR 2017

We are very happy that our tutorial on Interactive Search in Video & Lifelogging Repositories (presented together with Frank Hopfgartner from the University of Glasgow, UK) has been accepted for the ACM SIGIR Conference on Human Information Interaction & Retrieval (CHIIR), to be held in Oslo, Norway from March 7-11, 2017.

We are looking forward to an interesting tutorial. More information on the conference can be found here:

New Paper Preprints (MMM 2017)

The preprints of our papers accepted for the International Conference on MultiMedia Modeling (MMM) 2017, to be held in Reykjavik, Iceland, January 4-6 2017, are now available.

Deep Learning of Shot Classification in Gynecologic Surgery VideosAn Evaluation of Video Browsing on Tablets with the ThumbBrowserStoryboard-based Video Browsing Using Color and Concept IndicesCollaborative Feature Maps for Interactive Video Search

Interactive Search in Video & Lifelog Repos

The slides of our tutorial at the IEEE International Conference on Multimedia & Expo (ICME 2016), held on July 11, 2016 in Seattle, USA, can be found here:

Finding the Chameleon in Your Videos

CollabrisOur demo paper Finding the Chameleon in Your Video Collection presented at the ACM Multimedia Systems 2016 Conference (MMSYS 2016) is now available for download. It describes an interactive video retrieval tool that is optimized for search by different types (sketch, example segment, and semantic concepts) and makes uses CNNs (convolutional neural networks) as well as optical flow estimation and temporal features signatures to analyze the content in the video. The interface is designed according to research results from several years on content-based search in video.

Results of the NGVB Project

The Next Generation Video Browsing (NGVB) project, an FWF funded translational research project (TRP 273-N15), which ran for three years, has ended in January this year. In the NGVB project we investigated how to design better tools for content-based search in videos, in particular when used on tablet devices, and how these search tools could benefit from the powerful features provided by tablets (e.g., multi-touch interaction and high computing power). We focused on improving video content navigation and visualization of the video structure through 3D models and color-sorted arrangement of images. To this end, we performed several initial user studies in order to find out how users search with common video player software and used the results for the design of novel search features. In order to evaluate these novel search features we developed software tools with new interaction models and search features and performed comparative user studies with them. The results of these studies show that the newly proposed features and interfaces provide significantly better search performance and are easier to use for the vast majority of tested users.

The research work of the NGVB project brought forward several interesting results and I would like to thank all involved people (in particular Dr. Marco A. Hudelist and Dr. Claudiu Cobarzan). The figure below shows an overview of research prototypes (i.e., interfaces) for video and image browsing that evolved from the project. More details can be found on our project website (including related research papers).


Domain-Specific Compression of Endoscopic Video

Domain Specific  Video Compression
We are very happy to announce that our paper on
Domain-Specific Video Compression for Long-term Archiving of Endoscopic Surgery Videos“,
authored by Bernd Münzer, Klaus Schoeffmann, and Laszlo Böszörmenyi,
has been accepted to the 29th IEEE International Symposium on Computer-Based Medical Systems (CBMS 2016), to be held in June 2016 in Dublin/Belfast.

A preprint of the paper is available here.

VBS Winners 2016

Congratulations again to Kai Uwe Barthel, Nico Hezel, and Radek Mackowiak (and Florian Barthel), from HTW Berlin, Germany, for winning the Video Browser Showdown 2016 competition with their interactive video search system described in Navigating a graph of scenes for exploring large video collections!

vibro2016 HTW

CFP: Content-Based Image and Multimedia Analysis and Indexing for Healthcare @ CBMI 2016

Multimedia information indexing and retrieval nowadays is more and more penetrating an important domain for society: healthcare. Feature-based classification approaches which are being developed for medical image classification for computer-aided diagnosis borrow the approaches from classical CBIR in feature engineering and cascaded classification. Deep learning classifiers which are being extensively studied and applied for concept recognition in multimedia data, image and video understanding are being applied for prediction of patients categories on the basis of physiological parameters such as gaze fixations. Information fusion approaches which are necessary for understanding and content – based indexing of highly dimensional multimedia data are applied for fusion of different modalities in medical image recognition. Video analysis and summarization approaches are being developed for automatic visual reporting in surgery. Similarly, video content analysis and retrieval in archived video data collected from surgeries becomes more and more important and provides the basis for later usage of these valuable data for scenarios such as case comparisons/similarity search, teaching of new operation techniques, as well as quality control/error inspection. Finally, the multimedia nowadays is more and more multimodal – not only image, video, textual and sound modalities supply the information, but also and specifically in medical and healthcare applications, a large variety of different sensors either measuring the context or physiological parameters are deployed. Future multimedia becomes multimodal and this happens in the healthcare domain in priority.

We are now looking for papers (6 pages, IEEE style) for this exciting special session (deadline: February 1, 2016).
More information can be found here.