The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
Video summarization is an important multimedia task for applications such as video indexing and retrieval, video surveillance, human-computer interaction and video "storyboarding". In this paper, we present a new approach for automatic summarization of video collections that leverages a structured minimum-risk classifier and efficient submodular inference. To test the accuracy of the predicted...
The representation of data in terms of its statistical properties is valuable in many applications. This work uses statistics obtained from 4D scene geometry to characterize, in terms of redundancy, the content produced by lenslet-based light field cameras and by high-density arrays of cameras for the JPEG Pleno Call for Proposals on Light Field Coding. This paper proposes a novel so-called geometric...
Erasure coding has been extensively employed for data availability protection in production storage systems by maintaining a low degree of data redundancy. However, how to mitigate the parity update overhead of partial stripe writes in erasure-coded storage systems is still a critical concern. In this paper, we reconsider this problem from two new perspectives: data correlation and stripe organization,...
Erasure coding has been widely adopted to protect data storage against failures in production data centers. Given the hierarchical nature of data centers, characterizing the effects of erasure coding and redundancy placement on the reliability of erasure-coded data centers is critical yet largely unexplored. This paper presents a comprehensive simulation analysis of reliability on erasure-coded data...
Failure tolerant data encoding and storage is of paramount importance for data centers, supercomputers, data transfers, and many aspects of information technology. Reed-Solomon failure erasure codes and their variants are the basis for many applications in this field. Efficient implementation of these codes is challenging because they require computations in Galois fields, which are not supported...
The importance of the task of countering the means of unauthorized access is to preserve the integrity of restricted access information circulating in computer networks determines the relevance of investigating perspective methods of cryptographic transformations, which are characterized by high speed and reliability of encryption. The methods of information security in the telecommunication system...
The problem of increasing noise immunity for sending messages usually solved by introducing the additional redundancy or by re-transmission of distorted fragment of the message. At the same time, the resulting delay is not always acceptable. An original concept for irredundant code efficiency rising is discussed. It is shown that some genetic-like algorithm applying to the class of permutative equivalent...
Modern distributed storage systems often store redundant data in multiple replications or erasure coding according to their access frequencies. Multiple replications scheme is well-performance for hot data while erasure coding scheme is storage-efficient for warm and cold data. When hot data turn cold, an encoding procedure starts to do the conversion. However, due to sequential striping, current...
This paper proposes novel soft error detection and mitigation technique in reduced instruction set computer (RISC) based pipeline processors. We leveraged the data encoding techniques (re-computing with rotated operands (RERO)) in conjunction with back pressure controlling mechanism in pipeline architecture. In order to alleviate the performance degradation due to potential stalling, we exploited...
With the increase in data volumes, it is prudent to classify data depending on its criticality. One might prefer a cheap storage for a year old system logs but a highly fault tolerant storage for personal photos. The existing solutions include storing these two sets of data in two different systems or choosing a system with a fault tolerance level required by the most critical data. This means a higher...
In this paper, an investigation of multiview video coding schemes is presented, based on different inter-view prediction structures and various group of pictures (GOP) sizes. The studied inter-view prediction schemes in this paper are: A new approach namely “PIP”, recently proposed (PBI), and the benchmark structure (MVC). Results of the conducted tests allow to order the studied inter-view structures...
The present work describes the PPM — Prediction by Partial (string) Matching algorithm for lossless compression of text. It studies the procedure of extending the source alphabet for the PPM encoding in order to allow using of symbols not yet present in the source alphabet at the very beginning of the process of text encoding. The work describes the procedure of extending the alphabet and presents...
In large-scale distributed computing clusters, such as Amazon EC2, there are several types of “system noise” that can result in major degradation of performance: system failures, bottlenecks due to limited communication bandwidth, latency due to straggler nodes, etc. On the other hand, these systems enjoy abundance of redundancy — a vast number of computing nodes and large storage capacity. There...
In this work, we establish the pointwise optimality of side information version of SWLZ algorithm for stationary ergodic sources. We also obtain a pointwise upper bound on the redundancy rate of this side information version of SWLZ algorithm for a subclass of φ-mixing sources, which includes Markov sources as a special case. This upper bound obtained differs only by a constant factor from the best...
Today, many real-world machine learning and data analytics problems are of a scale that requires distributed optimization; unlike in centralized computing, these systems are vulnerable to network and node failures. Recently, coding-theoretic ideas have been applied to mitigate node failures in such distributed computing networks. Relaxing the exact recovery requirement of such techniques, we propose...
Using an age of information (AoI) metric, we examine the transmission of coded updates through a binary erasure channel to a monitor/receiver. We start by deriving the average status update age of an infinite incremental redundancy (IIR) system in which the transmission of a k-symbol update continues until k symbols are received. This system is then compared to a fixed redundancy (FR) system in which...
We consider the problem of constructing codes that can correct δ deletions occurring in an arbitrary binary string of length n bits. Varshamov-Tenengolts (VT) codes can correct all possible single deletions (δ = 1) with an asymptotically optimal redundancy. Finding similar codes for δ ≥ 2 deletions is an open problem. We propose a new family of codes, that we call Guess & Check (GC) codes, that...
Deduplication finds and removes long-range data duplicates. It is commonly used in cloud and enterprise server settings and has been successfully applied to primary, backup, and archival storage. Despite its practical importance as a source-coding technique, its analysis from the point of view of information theory is missing. This paper provides such an information-theoretic analysis of data deduplication...
We consider a cache network, where a single server is connected to multiple users via a shared bottleneck link. The server has a set of files, which can be cached by each user in a prefetching phase. In a following delivery phase, each user requests a file and the server delivers user demands as efficiently as possible by taking into account their cache contents. We focus on an important and commonly...
Mutually Uncorrelated (MU) codes are a class of codes in which no proper prefix of one codeword is a suffix of another codeword. These codes were originally studied for synchronization purposes and recently, Yazdi et al. showed their applicability to enable random access in DNA storage. In this work we follow the research of Yazdi et al. and study MU codes along with their extensions to correct errors...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.