% Categories: EndoViP2 % Encoding: utf-8 @InProceedings{SchoeffmannTutorialACMMM2017, author = {Schoeffmann,Klaus and Münzer,Bernd and Riegler,Michael and Halvorsen,Paal}, booktitle = {MM ’17 Proceedings of the 2017 ACM on Multimedia Conference}, title = {Medical Multimedia Information Systems (MMIS)}, year = {2017}, address = {New York, NY, USA}, editor = {Liu, Qiong and Lienhart, Rainer and Wang, Haohong}, month = {oct}, pages = {1957-1958}, publisher = {ACM}, abstract = {In hospitals all around the world, medical multimedia information systems have gained high importance over the last few years. One of the reasons is that an increasing number of interventions are performed in a minimally invasive way. These endoscopic inspections and surgeries are performed with a tiny camera -- the endoscope -- which produces a video signal that is used to control the intervention. Apart from the viewing purpose, the video signal is also used for automatic content analysis during the intervention as well as for post-surgical usage, such as communicating operation techniques, planning future interventions, and medical forensics. Another reason is video documentation, which is even enforced by law in some countries. The problem, however, is the sheer amount of unstructured medical videos that are added to the multimedia archive on a daily basis. Without proper management and a multimedia information system, the medical videos cannot be used efficiently for post-surgical scenarios. It is therefore already foreseeable that medical multimedia information systems will gain even more attraction in the next few years. In this tutorial we will introduce the audience to this challenging new field, describe the domain-specific characteristics and challenges of medical multimedia data, introduce related use cases, and talk about existing works -- contributed by the medical imaging and robotics community, but also already partly from the multimedia community -- as well as the many open issues and challenges that bear high research potential.}, doi = {10.1145/3123266.3130142}, isbn10 = {978-1-4503-4906-2}, keywords = {endoscopic video, medical image processing, medical multimedia}, language = {EN}, location = {Mountain View, CA}, talkdate = {2017.10.27}, talktype = {registered}, url = {https://dl.acm.org/citation.cfm?id=3130142} } @InProceedings{Schoeffmann2017MMM, author = {Schoeffmann, Klaus and Primus, Manfred Jürgen and Muenzer, Bernd and Petscharnig, Stefan and Karisch, Christoph and Xu, Qing and Huerst, Wolfgang}, booktitle = {MultiMedia Modeling: 23rd International Conference, MMM 2017, Reykjavik, Iceland, January 4-6, 2017, Proceedings, Part II}, title = {Collaborative Feature Maps for Interactive Video Search}, year = {2017}, address = {Cham}, editor = {Amsaleg, Laurent and Guðmundsson, Gylfi Þór and Gurrin, Cathal and Jónsson, Björn Þór and Satoh, Shin’ichi}, month = {jan}, pages = {457-462}, publisher = {Springer International Publishing}, abstract = {This extended demo paper summarizes our interface used for the Video Browser Showdown (VBS) 2017 competition, where visual and textual known-item search (KIS) tasks, as well as ad-hoc video search (AVS) tasks in a 600-h video archive need to be solved interactively. To this end, we propose a very flexible distributed video search system that combines many ideas of related work in a novel and collaborative way, such that several users can work together and explore the video archive in a complementary manner. The main interface is a perspective Feature Map, which shows keyframes of shots arranged according to a selected content similarity feature (e.g., color, motion, semantic concepts, etc.). This Feature Map is accompanied by additional views, which allow users to search and filter according to a particular content feature. For collaboration of several users we provide a cooperative heatmap that shows a synchronized view of inspection actions of all users. Moreover, we use collaborative re-ranking of shots (in specific views) based on retrieved results of other users.}, doi = {10.1007/978-3-319-51814-5_41}, language = {EN}, location = {Reykjavik, Iceland}, talkdate = {2017.01.04}, talktype = {registered}, url = {https://link.springer.com/chapter/10.1007/978-3-319-51814-5_41#copyrightInformation} } @InProceedings{Muenzer2017c, author = {Münzer, Bernd and Primus, Manfred Jürgen and Kletz, Sabrina and Petscharnig, Stefan and Schoeffmann, Klaus}, booktitle = {IEEE International Symposium on Multimedia (ISM2017)}, title = {Static vs. Dynamic Content Descriptors for Video Retrieval in Laparoscopy}, year = {2017}, address = {Taichung, Taiwan}, editor = {Chang, Kang-Ming and Chang, Wen-Thong}, month = {dec}, pages = {8}, publisher = {IEEE}, abstract = {The domain of minimally invasive surgery has recently attracted attention from the Multimedia community due to the fact that systematic video documentation is on the rise in this medical field. The vastly growing volumes of video archives demand for effective and efficient techniques to retrieve specific information from large video collections with visually very homogeneous content. One specific challenge in this context is to retrieve scenes showing similar surgical actions, i.e., similarity search. Although this task has a high and constantly growing relevance for surgeons and other health professionals, it has rarely been investigated in the literature so far for this particular domain. In this paper, we propose and evaluate a number of both static and dynamic content descriptors for this purpose. The former only take into account individual images, while the latter consider the motion within a scene. Our experimental results show that although static descriptors achieve the highest overall performance, dynamic descriptors are much more discriminative for certain classes of surgical actions. We conclude that the two approaches have complementary strengths and further research should investigate methods to combine them.}, language = {EN}, location = {Taichung, Taiwan}, talkdate = {2017.12.12}, talktype = {registered} } @InProceedings{Muenzer2017b, author = {Münzer, Bernd and Schoeffmann, Klaus and Böszörmenyi, Laszlo}, booktitle = {IEEE International Symposium on Multimedia (ISM2017)}, title = {EndoXplore: A Web-based Video Explorer for Endoscopic Videos}, year = {2017}, address = {Taichung, Taiwan}, editor = {Chang, Kang-Ming and Chang, Wen-Thong}, month = {dec}, pages = {2}, publisher = {IEEE}, abstract = {The rapidly increasing volume of videos recorded in the course of endoscopic screenings and surgeries poses demanding challenges to video retrieval and browsing systems. Surgeons typically have to use standard video players to retrospectively review their procedures, which is an extremely cumbersome and time-consuming process. We present an HTML5-based video explorer that is specially tailored to this purpose and enables a time-efficient post-operative review of procedures. It incorporates various interactive browsing mechanisms as well as domain-specific content-based features based on previous research results. Preliminary interviews with surgeons indicate that this tool can considerably improve retrieval and browsing efficiency for users in the medical domain and allows surgeons to more easily and quickly revisit specific moments in recordings of their endoscopic surgeries.}, language = {EN}, location = {Taichung, Taiwan}, talkdate = {2017.12.11}, talktype = {poster} } @InProceedings{Muenzer2017a, author = {Münzer, Bernd and Primus, Manfred Jürgen and Hudelist, Marco and Beecks, Christian and Hürst, Wolfgang and Schoeffmann, Klaus}, booktitle = {2017 IEEE International Conference on Multimedia \& Expo Workshops (ICMEW)}, title = {When content-based video retrieval and human computation unite: Towards effective collaborative video search}, year = {2017}, address = {Hongkong, China}, editor = {Chan, Yui-Lam and Rahardja, Susanto}, month = {jul}, pages = {214-219}, publisher = {IEEE}, abstract = {Although content-based retrieval methods achieved very good results for large-scale video collections in recent years, they still suffer from various deficiencies. On the other hand, plain human perception is a very powerful ability that still outperforms automatic methods in appropriate settings, but is very limited when it comes to large-scale data collections. In this paper, we propose to take the best from both worlds by combining an advanced content-based retrieval system featuring various query modalities with a straightforward mobile tool that is optimized for fast human perception in a sequential manner. In this collaborative system with multiple users, both subsystems benefit from each other: The results of issued queries are used to re-rank the video list on the tablet tool, which in turn notifies the retrieval tool about parts of the dataset that have already been inspected in detail and can be omitted in subsequent queries. The preliminary experiments show promising results in terms of search performance.}, doi = {10.1109/ICMEW.2017.8026262}, language = {EN}, location = {Hongkong}, talkdate = {2017.07.10}, talktype = {registered} } @Article{Muenzer2017, author = {Münzer, Bernd and Schoeffmann, Klaus and Böszörmenyi, Laszlo}, journal = {Multimedia Tools and Applications}, title = {Content-based processing and analysis of endoscopic images and videos: A survey}, year = {2017}, month = {jan}, pages = {1-40}, abstract = {In recent years, digital endoscopy has established as key technology for medical screenings and minimally invasive surgery. Since then, various research communities with manifold backgrounds have picked up on the idea of processing and automatically analyzing the inherently available video signal that is produced by the endoscopic camera. Proposed works mainly include image processing techniques, pattern recognition, machine learning methods and Computer Vision algorithms. While most contributions deal with real-time assistance at procedure time, the post-procedural processing of recorded videos is still in its infancy. Many post-processing problems are based on typical Multimedia methods like indexing, retrieval, summarization and video interaction, but have only been sparsely addressed so far for this domain. The goals of this survey are (1) to introduce this research field to a broader audience in the Multimedia community to stimulate further research, (2) to describe domain-specific characteristics of endoscopic videos that need to be addressed in a pre-processing step, and (3) to systematically bring together the very diverse research results for the first time to provide a broader overview of related research that is currently not perceived as belonging together.}, address = {Berlin, Heidelberg, New York}, doi = {10.1007/s11042-016-4219-z}, language = {EN}, publisher = {Springer}, url = {https://link.springer.com/article/10.1007/s11042-016-4219-z} } @InProceedings{Muenzer2016, author = {Münzer, Bernd and Schoeffmann, Klaus and Böszörmenyi, Laszlo}, booktitle = {29th International Symposium on Computer-Based Medical Systems (CBMS'16)}, title = {Domain-Specific Video Compression for Long-term Archiving of Endoscopic Surgery Videos}, year = {2016}, address = {Dublin, Ireland}, editor = {Kane, B and Marshall, A and Soda, P}, month = {jun}, pages = {312-317}, publisher = {IEEE}, doi = {10.1109/CBMS.2016.28}, language = {EN}, location = {Dublin}, talkdate = {2016.06.23}, talktype = {registered} }