[205] | Zhiming Zhao, Ian Taylor, Radu Prodan, Editorial for FGCS Special issue on "Time-critical Applications on Software-defined Infrastructures", In Future Generation Computer Systems, Elsevier BV, vol. 112, pp. 1170-1171, 2020.
[bib][url] [doi] |
[204] | Laurens Versluis, Roland Matha, Sacheendra Talluri, Tim Hegeman, Radu Prodan, Ewa Deelman, Alexandru Iosup, The Workflow Trace Archive: Open-Access Data From Public and Private Computing Infrastructures, In IEEE Transactions on Parallel and Distributed Systems, Institute of Electrical and Electronics Engineers (IEEE), vol. 31, no. 9, pp. 2170-2184, 2020.
[bib][url] [doi] [abstract]
Abstract: Realistic, relevant, and reproducible experiments often need input traces collected from real-world environments. In this work, we focus on traces of workflows—common in datacenters, clouds, and HPC infrastructures. We show that the state-of-the-art in using workflow-traces raises important issues: (1) the use of realistic traces is infrequent and (2) the use of realistic, open-access traces even more so. Alleviating these issues, we introduce the Workflow Trace Archive (WTA), an open-access archive of workflow traces from diverse computing infrastructures and tooling to parse, validate, and analyze traces. The WTA includes >48 million workflows captured from >10 computing infrastructures, representing a broad diversity of trace domains and characteristics. To emphasize the importance of trace diversity, we characterize the WTA contents and analyze in simulation the impact of trace diversity on experiment results. Our results indicate significant differences in characteristics, properties, and workflow structures between workload sources, domains, and fields.
|
[203] | Ennio Torre, Juan J. Durillo, Vincenzo de Maio, Prateek Agrawal, Shajulin Benedict, Nishant Saurabh, Radu Prodan, A dynamic evolutionary multi-objective virtual machine placement heuristic for cloud data centers, In Information and Software Technology, Elsevier BV, vol. 128, pp. 106390, 2020.
[bib][url] [doi] [abstract]
Abstract: Minimizing the resource wastage reduces the energy cost of operating a data center, but may also lead to a considerably high resource overcommitment affecting the Quality of Service (QoS) of the running applications. The effective tradeoff between resource wastage and overcommitment is a challenging task in virtualized Clouds and depends on the allocation of virtual machines (VMs) to physical resources. We propose in this paper a multi-objective method for dynamic VM placement, which exploits live migration mechanisms to simultaneously optimize the resource wastage, overcommitment ratio and migration energy. Our optimization algorithm uses a novel evolutionary meta-heuristic based on an island population model to approximate the Pareto optimal set of VM placements with good accuracy and diversity. Simulation results using traces collected from a real Google cluster demonstrate that our method outperforms related approaches by reducing the migration energy by up to 57% with a QoS increase below 6%.
|
[202] | Nishant Saurabh, Shajulin Benedict, Jorge G. Barbosa, Radu Prodan, Expelliarmus: Semantic-centric virtual machine image management in IaaS Clouds, In Journal of Parallel and Distributed Computing, Elsevier BV, vol. 146, pp. 107-121, 2020.
[bib][url] [doi] [abstract]
Abstract: Infrastructure-as-a-service (IaaS) Clouds concurrently accommodate diverse sets of user requests, requiring an efficient strategy for storing and retrieving virtual machine images (VMIs) at a large scale. The VMI storage management requires dealing with multiple VMIs, typically in the magnitude of gigabytes, which entails VMI sprawl issues hindering the elastic resource management and provisioning. Unfortunately, existing techniques to facilitate VMI management overlook VMI semantics (i.e at the level of base image and software packages), with either restricted possibility to identify and extract reusable functionalities or with higher VMI publishing and retrieval overheads. In this paper, we propose Expelliarmus, a novel VMI management system that helps to minimize VMI storage, publishing and retrieval overheads. To achieve this goal, Expelliarmus incorporates three complementary features. First, it models VMIs as semantic graphs to facilitate their similarity computation. Second, it provides a semantically-aware VMI decomposition and base image selection to extract and store non-redundant base image and software packages. Third, it assembles VMIs based on the required software packages upon user request. We evaluate Expelliarmus through a representative set of synthetic Cloud VMIs on a real test-bed. Experimental results show that our semantic-centric approach is able to optimize the repository size by 2.3 - 22 times compared to state-of-the-art systems (e.g. IBM’s Mirage and Hemera) with significant VMI publishing and slight retrieval performance improvement.
|
[201] | Andrew Perkis, Christian Timmerer, Sabina Baraković, Jasmina Baraković Husić, Søren Bech, Sebastian Bosse, Jean Botev, Kjell Brunnström, Luis Cruz, Katrien De Moor, Andrea de Polo Saibanti, Wouter Durnez, Sebastian Egger-Lampl, Ulrich Engelke, Tiago H. Falk, Jesús Gutiérrez, Asim Hameed, Andrew Hines, Tanja Kojic, Dragan Kukolj, Eirini Liotou, Dragorad Milovanovic, Sebastian Möller, Niall Murray, Babak Naderi, Manuela Pereira, Stuart Perry, Antonio Pinheiro, Andres Pinilla, Alexander Raake, Sarvesh Rajesh Agrawal, Ulrich Reiter, Rafael Rodrigues, Raimund Schatz, Peter Schelkens, Steven Schmidt, Saeed Shafiee Sabet, Ashutosh Singla, Lea Skorin-Kapov, Mirko Suznjevic, Stefan Uhrig, Sara Vlahović, Jan-Niklas Voigt-Antons, Saman Zadtootaghaj, QUALINET White Paper on Definitions of Immersive Media Experience (IMEx), In , 2020.
[bib] [abstract]
Abstract: With the coming of age of virtual/augmented reality and interactive media, numerous definitions, frameworks, and models of immersion have emerged across different fields ranging from computer graphics to literary works. Immersion is oftentimes used interchangeably with presence as both concepts are closely related. However, there are noticeable interdisciplinary differences regarding definitions, scope, and constituents that are required to be addressed so that a coherent understanding of the concepts can be achieved. Such consensus is vital for paving the directionality of the future of immersive media experiences (IMEx) and all related matters. The aim of this white paper is to provide a survey of definitions of immersion and presence which leads to a definition of immersive media experience (IMEx). The Quality of Experience (QoE) for immersive media is described by establishing a relationship between the concepts of QoE and IMEx followed by application areas of immersive media experience. Influencing factors on immersive media experience are elaborated as well as the assessment of immersive media experience. Finally, standardization activities related to IMEx are highlighted and the white paper is concluded with an outlook related to future developments.
|
[200] | Roland Matha, Sasko Ristov, Thomas Fahringer, Radu Prodan, Simplified Workflow Simulation on Clouds based on Computation and Communication Noisiness, In IEEE Transactions on Parallel and Distributed Systems, Institute of Electrical and Electronics Engineers (IEEE), vol. 31, no. 7, pp. 1559-1574, 2020.
[bib][url] [doi] [abstract]
Abstract: Many researchers rely on simulations to analyze and validate their researched methods on Cloud infrastructures. However, determining relevant simulation parameters and correctly instantiating them to match the real Cloud performance is a difficult and costly operation, as minor configuration changes can easily generate an unreliable inaccurate simulation result. Using legacy values experimentally determined by other researchers can reduce the configuration costs, but is still inaccurate as the underlying public Clouds and the number of active tenants are highly different and dynamic in time. To overcome these deficiencies, we propose a novel model that simulates the dynamic Cloud performance by introducing noise in the computation and communication tasks, determined by a small set of runtime execution data. Although the estimating method is apparently costly, a comprehensive sensitivity analysis shows that the configuration parameters determined for a certain simulation setup can be used for other simulations too, thereby reducing the tuning cost by up to 82.46%, while declining the simulation accuracy by only 1.98% in average. Extensive evaluation also shows that our novel model outperforms other state-of-the-art dynamic Cloud simulation models, leading up to 22% lower makespan inaccuracy.
|
[199] | Vincenzo De Maio, Dragi Kimovski, Multi-objective scheduling of extreme data scientific workflows in Fog, In Future Generation Computer Systems, Elsevier BV, vol. 106, pp. 171-184, 2020.
[bib][url] [doi] [abstract]
Abstract: The concept of “extreme data” is a recent re-incarnation of the “big data” problem, which is distinguished by the massive amounts of information that must be analyzed with strict time requirements. In the past decade, the Cloud data centers have been envisioned as the essential computing architectures for enabling extreme data workflows. However, the Cloud data centers are often geographically distributed. Such geographical distribution increases offloading latency, making it unsuitable for processing of workflows with strict latency requirements, as the data transfer times could be very high. Fog computing emerged as a promising solution to this issue, as it allows partial workflow processing in lower-network layers. Performing data processing on the Fog significantly reduces data transfer latency, allowing to meet the workflows’ strict latency requirements. However, the Fog layer is highly heterogeneous and loosely connected, which affects reliability and response time of task offloading. In this work, we investigate the potential of Fog for scheduling of extreme data workflows with strict response time requirements. Moreover, we propose a novel Pareto-based approach for task offloading in Fog, called Multi-objective Workflow Offloading (MOWO). MOWO considers three optimization objectives, namely response time, reliability, and financial cost. We evaluate MOWO workflow scheduler on a set of real-world biomedical, meteorological and astronomy workflows representing examples of extreme data application with strict latency requirements.
|
[198] | Jeroen van der Hooft, Maria Torres Vega, Tim Wauters, Christian Timmerer, Ali C. Begen, Filip De Turck, Raimund Schatz, From Capturing to Rendering: Volumetric Media Delivery with Six Degrees of Freedom, In IEEE Communications Magazine, Institute of Electrical and Electronics Engineers (IEEE), vol. 58, no. 10, pp. 49-55, 2020.
[bib][url] [doi] [abstract]
Abstract: Technological improvements are rapidly advancing holographic-type content distribution. Significant research efforts have been made to meet the low-latency and high-bandwidth requirements set forward by interactive applications such as remote surgery and virtual reality. Recent research made six degrees of freedom (6DoF) for immersive media possible, where users may both move their heads and change their position within a scene. In this article, we present the status and challenges of 6DoF applications based on volumetric media, focusing on the key aspects required to deliver such services. Furthermore, we present results from a subjective study to highlight relevant directions for future research.
|
[197] | Samira Hayat, Evsen Yanmaz, Christian Bettstetter, Timothy X. Brown, Multi-objective drone path planning for search and rescue with quality-of-service requirements, In Autonomous Robots, Springer Science and Business Media LLC, vol. 44, no. 7, pp. 1183-1198, 2020.
[bib][url] [doi] [abstract]
Abstract: We incorporate communication into the multi-UAV path planning problem for search and rescue missions to enable dynamic task allocation via information dissemination. Communication is not treated as a constraint but a mission goal. While achieving this goal, our aim is to avoid compromising the area coverage goal and the overall mission time. We define the mission tasks as: search, inform, and monitor at the best possible link quality. Building on our centralized simultaneous inform and connect (SIC) path planning strategy, we propose two adaptive strategies: (1) SIC with QoS (SICQ): optimizes search, inform, and monitor tasks simultaneously and (2) SIC following QoS (SIC+): first optimizes search and inform tasks together and then finds the optimum positions for monitoring. Both strategies utilize information as soon as it becomes available to determine UAV tasks. The strategies can be tuned to prioritize certain tasks in relation to others. We illustrate that more tasks can be performed in the given mission time by efficient incorporation of communication in the path design. We also observe that the quality of the resultant paths improves in terms of connectivity.
|
[196] | Negin Ghamsarian, Klaus Schoeffmann, Morteza Khademi, Blind MV-based video steganalysis based on joint inter-frame and intra-frame statistics, In Multimedia Tools and Applications, Springer Science and Business Media LLC, vol. 80, no. 6, pp. 1-23, 2020.
[bib][url] [doi] [abstract]
Abstract: Despite all its irrefutable benefits, the development of steganography methods has sparked ever-increasing concerns over steganography abuse in recent decades. To prevent the inimical usage of steganography, steganalysis approaches have been introduced. Since motion vector manipulation leads to random and indirect changes in the statistics of videos, MV-based video steganography has been the center of attention in recent years. In this paper, we propose a 54-dimentional feature set exploiting spatio-temporal features of motion vectors to blindly detect MV-based stego videos. The idea behind the proposed features originates from two facts. First, there are strong dependencies among neighboring MVs due to utilizing rate-distortion optimization techniques and belonging to the same rigid object or static background. Accordingly, MV manipulation can leave important clues on the differences between each MV and the MVs belonging to the neighboring blocks. Second, a majority of MVs in original videos are locally optimal after decoding concerning the Lagrangian multiplier, notwithstanding the information loss during compression. Motion vector alteration during information embedding can affect these statistics that can be utilized for steganalysis. Experimental results have shown that our features’ performance far exceeds that of state-of-the-art steganalysis methods. This outstanding performance lies in the utilization of complementary spatio-temporal statistics affected by MV manipulation as well as feature dimensionality reduction applied to prevent overfitting. Moreover, unlike other existing MV-based steganalysis methods, our proposed features can be adjusted to various settings of the state-of-the-art video codec standards such as sub-pixel motion estimation and variable-block-size motion estimation.
|
[195] | Ekrem Cetinkaya, M. Furkan KIRAÇ, Image denoising using deep convolutional autoencoder with feature pyramids, In TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, vol. 28, no. 4, pp. 2096-2109, 2020.
[bib][url] [doi] [abstract]
Abstract: Image denoising is 1 of the fundamental problems in the image processing field since it is the preliminary stepfor many computer vision applications. Various approaches have been used for image denoising throughout the yearsfrom spatial filtering to model-based approaches. Having outperformed all traditional methods, neural-network-baseddiscriminative methods have gained popularity in recent years. However, most of these methods still struggle to achieveflexibility against various noise levels and types. In this paper, a deep convolutional autoencoder combined with a variantof feature pyramid network is proposed for image denoising. Simulated data generated by Blender software along withcorrupted natural images are used during training to improve robustness against various noise levels. Experimental resultsshow that the proposed method can achieve competitive performance in blind Gaussian denoising with significantly lesstraining time required compared to state of the art methods. Extensive experiments showed the proposed method givespromising performance in a wide range of noise levels with a single network.
|
[194] | Hanna Borgli, Vajira Thambawita, Pia H. Smedsrud, Steven Hicks, Debesh Jha, Sigrun L. Eskeland, Kristin Ranheim Randel, Konstantin Pogorelov, Mathias Lux, Duc Tien Dang Nguyen, Dag Johansen, Carsten Griwodz, H\aakon K. Stensland, Enrique Garcia-Ceja, Peter T. Schmidt, Hugo L. Hammer, Michael A. Riegler, Paal Halvorsen, Thomas de Lange, HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy, In Scientific Data, Springer Science and Business Media LLC, vol. 7, no. 1, 2020.
[bib][url] [doi] [abstract]
Abstract: Artificial intelligence is currently a hot topic in medicine. However, medical data is often sparse and hard to obtain due to legal restrictions and lack of medical personnel for the cumbersome and tedious process to manually label training data. These constraints make it difficult to develop systems for automatic analysis, like detecting disease or other lesions. In this respect, this article presents HyperKvasir, the largest image and video dataset of the gastrointestinal tract available today. The data is collected during real gastro- and colonoscopy examinations at Bærum Hospital in Norway and partly labeled by experienced gastrointestinal endoscopists. The dataset contains 110,079 images and 374 videos, and represents anatomical landmarks as well as pathological and normal findings. The total number of images and video frames together is around 1 million. Initial experiments demonstrate the potential benefits of artificial intelligence-based computer-assisted diagnosis systems. The HyperKvasir dataset can play a valuable role in developing better algorithms and computer-assisted examination systems not only for gastro- and colonoscopy, but also for other fields in medicine.
|
[193] | Neha Bhadwal, Prateek Agrawal, Vishu Madaan, A Machine Translation System from Hindi to Sanskrit Language using Rule based Approach, In Scalable Computing: Practice and Experience, Scalable Computing: Practice and Experience, vol. 21, no. 3, pp. 543-554, 2020.
[bib][url] [doi] [abstract]
Abstract: Machine Translation is an area of Natural Language Processing which can replace the laborious task of manual translation. Sanskrit language is among the ancient Indo-Aryan languages. There are numerous works of art and literature in Sanskrit. It has also been a medium for creating treatise of philosophical work as well as works on logic, astronomy and mathematics. On the other hand, Hindi is the most prominent language of India. Moreover,it is among the most widely spoken languages across the world. This paper is an effort to bridge the language barrier between Hindi and Sanskrit language such that any text in Hindi can be translated to Sanskrit. The technique used for achieving the aforesaid objective is rule-based machine translation. The salient linguistic features of the two languages are used to perform the translation. The results are produced in the form of two confusion matrices wherein a total of 50 random sentences and 100 tokens (Hindi words or phrases) were taken for system evaluation. The semantic evaluation of 100 tokens produce an accuracy of 94% while the pragmatic analysis of 50 sentences produce an accuracy of around 86%. Hence, the proposed system can be used to understand the whole translation process and can further be employed as a tool for learning as well as teaching. Further, this application can be embedded in local communication based assisting Internet of Things (IoT) devices like Alexa or Google Assistant.
|
[192] | Abdelhak Bentaleb, Christian Timmerer, Ali C. Begen, Roger Zimmermann, Performance Analysis of ACTE: a Bandwidth Prediction Method for Low-Latency Chunked Streaming, In ACM Transactions on Multimedia Computing, Communications, and Applications, Association for Computing Machinery (ACM), vol. 16, no. 2s, pp. 1-24, 2020.
[bib][url] [doi] [abstract]
Abstract: HTTP adaptive streaming with chunked transfer encoding can offer low-latency streaming without sacrificing the coding efficiency.This allows media segments to be delivered while still being packaged. However, conventional schemes often make widely inaccurate bandwidth measurements due to the presence of idle periods between the chunks and hence this is causing sub-optimal adaptation decisions. To address this issue, we earlier proposed ACTE (ABR for Chunked Transfer Encoding), a bandwidth prediction scheme for low-latency chunked streaming. While ACTE was a significant step forward, in this study we focus on two still remaining open areas, namely (i) quantifying the impact of encoding parameters, including chunk and segment durations, bitrate levels, minimum interval between IDR-frames and frame rate onACTE, and (ii) exploring the impact of video content complexity on ACTE. We thoroughly investigate these questions and report on our findings. We also discuss some additional issues that arise in the context of pursuing very low latency HTTP video streaming.
|
[191] | Michal Barcis, Agata Barcis, Hermann Hellwagner, Information Distribution in Multi-Robot Systems: Utility-Based Evaluation Model, In Sensors, MDPI AG, vol. 20, no. 3, 2020.
[bib][url] [doi] [abstract]
Abstract: This work addresses the problem of information distribution in multi-robot systems, with an emphasis on multi-UAV (unmanned aerial vehicle) applications. We present an analytical model that helps evaluate and compare different information distribution schemes in a robotic mission. It serves as a unified framework to represent the usefulness (utility) of each message exchanged by the robots. It can be used either on its own in order to assess the information distribution efficacy or as a building block of solutions aimed at optimizing information distribution. Moreover, we present multiple examples of instantiating the model for specific missions. They illustrate various approaches to defining the utility of different information types. Finally, we introduce a proof of concept showing the applicability of the model in a robotic system by implementing it in Robot Operating System 2 (ROS 2) and performing a simple simulated mission using a network emulator. We believe the introduced model can serve as a basis for further research on generic solutions for assessing or optimizing information distribution.
|
[190] | Prateek Agrawal, Deepak Chaudhary, Vishu Madaan, Anatoliy Zabrovskiy, Radu Prodan, Dragi Kimovski, Christian Timmerer, Automated bank cheque verification using image processing and deep learning methods, In Multimedia Tools and Applications, Springer Science and Business Media LLC, vol. 80, no. 4, pp. 5319-5350, 2020.
[bib][url] [doi] [abstract]
Abstract: Automated bank cheque verification using image processing is an attempt to complement the present cheque truncation system, as well as to provide an alternate methodology for the processing of bank cheques with minimal human intervention. When it comes to the clearance of the bank cheques and monetary transactions, this should not only be reliable and robust but also save time which is one of the major factor for the countries having large population. In order to perform the task of cheque verification, we developed a tool which acquires the cheque leaflet key components, essential for the task of cheque clearance using image processing and deep learning methods. These components include the bank branch code, cheque number, legal as well as courtesy amount, account number, and signature patterns. our innovation aims at benefiting the banking system by re-innovating the other competent cheque-based monetary transaction system which requires automated system intervention. For this research, we used institute of development and research in banking technology (IDRBT) cheque dataset and deep learning based convolutional neural networks (CNN) which gave us an accuracy of 99.14% for handwritten numeric character recognition. It resulted in improved accuracy and precise assessment of the handwritten components of bank cheque. For machine printed script, we used MATLAB in-built OCR method and the accuracy achieved is satisfactory (97.7%) also for verification of Signature we have used Scale Invariant Feature Transform (SIFT) for extraction of features and Support Vector Machine (SVM) as classifier, the accuracy achieved for signature verification is 98.10%.
|
[189] | Prateek Agrawal, Anatoliy Zabrovskiy, Adithyan Ilangovan, Christian Timmerer, Radu Prodan, FastTTPS: fast approach for video transcoding time prediction and scheduling for HTTP adaptive streaming videos, In Cluster Computing, Springer Science and Business Media LLC, pp. 1-17, 2020.
[bib][url] [doi] [abstract]
Abstract: HTTP adaptive streaming of video content becomes an integrated part of the Internet and dominates other streaming protocols and solutions. The duration of creating video content for adaptive streaming ranges from seconds or up to several hours or days, due to the plethora of video transcoding parameters and video source types. Although, the computing resources of different transcoding platforms and services constantly increase, accurate and fast transcoding time prediction and scheduling is still crucial. We propose in this paper a novel method called fast video transcoding time prediction and scheduling (FastTTPS) of x264 encoded videos based on three phases: (i) transcoding data engineering, (ii) transcoding time prediction, and (iii) transcoding scheduling. The first phase is responsible for video sequence selection, segmentation and feature data collection required for predicting the transcoding time. The second phase develops an artificial neural network (ANN) model for segment transcoding time prediction based on transcoding parameters and derived video complexity features. The third phase compares a number of parallel schedulers to map the predicted transcoding segments on the underlying high-performance computing resources. Experimental results show that our predictive ANN model minimizes the transcoding mean absolute error (MAE) and mean square error (MSE) by up to 1.7 and 26.8, respectively. In terms of scheduling, our method reduces the transcoding time by up to 38% using a Max–Min algorithm compared to the actual transcoding time without prediction information.
|
[188] | Jeroen van der Hooft, Maria Torres Vega, Tim Wauters, Hemanth Kumar Ravuri, C. Timmerer, Hermann Hellwagner, Filip De Turck, Towards 6DoF virtual reality video streaming: status and challenges, In IEEE COMSOC MMTC COMMUNICATIONS - FRONTIERS, vol. 14, no. 5, pp. 30-37, 2019.
[bib][url] [abstract]
Abstract: In the last few years, delivery of immersive video with six degrees of freedom (6DoF) has become an important topic for content providers. Recent technological advancements have resulted in affordable head-mounted displays, allowing a broad range of users to enjoy Virtual Reality (VR) content. Service providers such as Facebook1and YouTube2were among the first to provide 360°video, using the principle of HTTP Adaptive Streaming (HAS) to deliver the content to the enduser. In HAS, the content is encoded using several quality representations, temporally segmented into chunks of one to ten seconds and stored on one or multiple servers within a content delivery network. Based on the perceived network conditions, the device characteristics, and the user's preferences, the client can then decide on the quality of each of these segments[1]. Having the ability to adapt the video quality, this approach actively avoids buffer starvation, and therefore results in smoother playback of the requested content and a higher Quality of Experience (QoE) for the end user[2]. The introduction of 360° video provides the user with three degrees of freedom to move within an immersive world, allowing changes in the yaw, roll, and pitch.In the last few years, multiple solutions have been proposed to efficiently deliver VR content through HAS, focusing, for instance, on foveas-and tile-based encoding, improved viewport prediction (i.e., prediction of the user’s head movement in the near future in order to buffer useful high-quality content), and application layer optimizations [3]. In these works, however, the location of the user remains fixed to the position of the camera within the scene. Recently, significant research efforts have been made to realize 6DoF for streamed video content, i.e., the user may experience three additional degrees of freedom by being able to change the viewing position in a video scene. These efforts are promising, but significant research contributions will be required in order to realize its full potential. In this paper, an overview of existing 6DoF solutions is presented, and key challenges and opportunities are highlighted.
|
[187] | Christian Timmerer, Anatoliy Zabrovskiy, Automating QoS and QoE Evaluation of HTTP Adaptive Streaming Systems, In ZTE COMMUNICATIONS, vol. 17, no. 1, pp. 18-24, 2019.
[bib][url] [doi] [abstract]
Abstract: Streaming audio and video content currently accounts for the majority of the In⁃ternet traffic and is typically deployed over the top of the existing infrastructure. We arefacing the challenge of a plethora of media players and adaptation algorithms showing dif⁃ferent behavior but lacking a common framework for both objective and subjective evalua⁃tion of such systems. This paper aims to close this gap byproposing such a framework,de⁃scribing its architecture,providing an example evaluation, anddiscussing open issues.
|
[186] | Wen Ji, Zhu Li, H. Vincent Poor, Christian Timmerer, Wenwu Zhu, Guest Editorial Multimedia Economics for Future Networks: Theory, Methods, and Applications, In IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, vol. 37, no. 7, pp. 1473-1477, 2019.
[bib][url] [doi] [abstract]
Abstract: With the growing integration of telecommunication networks, Internet of Things (IoT), and 5G networks, there is a tremendous demand for multimedia services over heterogeneous networks. According to recent survey reports, mobile video traffic accounted for 60 percent of total mobile data traffic in 2016, and it will reach up to 78 percent by the end of 2021. Users’ daily lives are inundated with multimedia services, such as online video streaming (e.g., YouTube and Netflix), social networks (e.g., Facebook, Instagram, and Twitter), IoT and machine generated video (e.g, surveillance cameras), and multimedia service providers (e.g., Over-the-Top (OTT) services). Multimedia data is thus becoming the dominant traffic in the near future for both wired and wireless networks.
|
[185] | Christian Timmerer, MPEG column: 124th MPEG meeting in Macau, China, In SIGMultimedia Records, ACM, vol. 10, no. 4, New York, NY, USA, pp. 8:8-8:8, 2019.
[bib][url] [doi] |
[184] | Klaus Schöffmann, Björn Þór Jónsson, Cathal Gurrin, Dataset Column: Report from the MMM 2019 Special Session on Multimedia Datasets for Repeatable Experimentation (MDRE 2019), In ACM SIGMM Records, vol. 11, no. 3, 2019.
[bib][url] [abstract]
Abstract: Information retrieval and multimedia content access have a long history of comparative evaluation, and many of the advances in the area over the past decade can be attributed to the availability of open datasets that support comparative and repeatable experimentation. Sharing data and code to allow other researchers to replicate research results is needed in the multimedia modeling field, as it helps to improve the performance of systems and the reproducibility of published papers.This report summarizes the special session on Multimedia Datasets for Repeatable Experimentation (MDRE 2019), which was organized at the 25th International Conference on MultiMedia Modeling (MMM 2019), which was held in January 2019 in Thessaloniki, Greece.The intent of these special sessions is to be a venue for releasing datasets to the multimedia community and discussing dataset related issues. The presentation mode in 2019 was to have short presentations (8 minutes) with some questions, and an additional panel discussion after all the presentations, which was moderated by Björn Þór Jónsson. In the following we summarize the special session, including its talks, questions, and discussions.
|
[183] | Cathal Gurrin, Hideo Joho, Liting Zhou, Duc-Tien Dang-Nguyen, Luca Piras, Jakub Lokoc, Klaus Schöffmann, Andreas Leibetseder, Aaron Duane, Michael Riegler, Minh-Triet Tran, Wolfgang Hürst, Comparing Approaches to Interactive Lifelog Search at the Lifelog Search Challenge (LSC2018), In ITE Transactions on Media Technology and Applications, vol. 7, no. 2, pp. 46-59, 2019.
[bib][url] [doi] [abstract]
Abstract: The Lifelog Search Challenge (LSC) is an international content retrieval competition that evaluates search for personal lifelog data. At the LSC, content-based search is performed over a multi-modal dataset, continuously recorded by a lifelogger over 27 days, consisting of multimedia content, biometric data, human activity data, and information activities data. In this work, we report on the first LSC that took place in Yokohama, Japan in 2018 as a special workshop at ACM International Conference on Multimedia Retrieval 2018 (ICMR 2018). We describe the general idea of this challenge, summarise the participating search systems as well as the evaluation procedure, and analyse the search performance of the teams in various aspects. We try to identify reasons why some systems performed better than others and provide an outlook as well as open issues for upcoming iterations of the challenge.
|
[182] | Luca Rossetto, Fabian Berns, Klaus Schöffmann, George M. Awad, Christian Beecks, The V3C1 Dataset: Advancing the State of the Art in Video Retrieval, In ACM SIGMM Records, vol. 11, no. 2, 2019.
[bib][url] [abstract]
Abstract: Standardized datasets are of vital importance in multimedia research, as they form the basis for reproducible experiments and evaluations. In the area of video retrieval, widely used datasets such as the IACC [5], which has formed the basis for the TRECVID Ad-Hoc Video Search Task and other retrieval-related challenges, have started to show their age. For example, IACC is no longer representative of video content as it is found in the wild [7]. This is illustrated by the figures below, showing the distribution of video age and duration across various datasets in comparison with a sample drawn from Vimeo and Youtube.
|
[181] | Laura Ricci, Alexander Iosup, Radu Aurel Prodan, EDITORIAL Special Issue on Large Scale Cooperative Virtual Environments, In Journal of Grid Computing, vol. 17, pp. 1-2, 2019.
[bib][url] [doi] |