The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
Monitoring is a common requirement in everyday modern life, from anti-virus scanners on personal computers to securing large multinational systems' health and status. Such monitoring is a fundamental process which must be undertaken where resources are offered as a service as in the cloud, to make sure that the service has been delivered to meet all contractual obligations. However, with a dynamic...
Infrastructure-as-a-Service environments are becoming increasingly popular. When there is a failure, many applications require service restoration within a few seconds. Reaction to failures in Cloud is still slow for many applications. Monitoring is limited to instance metrics that are not conducive to precise diagnosis due to complexity of virtualization in physical hosts. Interferences among different...
The Domain Name Service (DNS) is a vital service in the Internet. Much more than a simple translation mechanism, it also allows higher profile functionalities such as load balancing and enhanced content distribution. In the scope of cloud computing, DNS is foreseen as an elastic and robust service, supporting failover mechanisms, decentralised configuration and multi-tenant isolation.
In order to provide the cloud computing research community with a full-system level datacenter server emulator with programmable hardware and software, and stimulate more innovative research work, this poster and demo shows a scientific research platform, Titian2, designed and implemented at ICT of CAS. Titian2 has the ability of on-line profiling and measuring, and the scalability of connecting with...
A new energy-proportional computing model extends Barroso and Hölzle's original definition for fixed-resource systems to aid in the design of more efficient modern systems with reconfigurable resources that can be varied at runtime.
Application Performance Monitoring (APM) is crucial to any enterprise to ensure that the performance of its application or service is meeting the service level agreements. There is an inherent need for the APM of traditional networks to evolve, with the advances in the internet technologies. In this paper, we present a novel approach and framework for realizing a futuristic APM for Software Defined...
Provisioning of high I/O capabilities for high-end HPC architectures is generally considered a challenge. A good understanding of the characteristics of the utilisation of modern I/O systems can help address the increasing performance gap between I/O and computation. In this paper we present results from an analysis of server-side performance counters that had been collected for multiple years on...
As the number of data centers increases, it is urgent to reduce their energy consumption. Although servers are becoming more energy-efficient, their idle consumption remains high, which is an issue as data centers are often over-provisioned. This work proposes a novel approach for building data centers with heterogeneous machines carefully chosen for their performance and energy efficiency ratios...
DC/OS is a widely used distributed operating system that abstracts the resources of light-weighted virtualized datacenters, which is based on Mesos distributed systems kernel and user space services such as Marathon. It automates resource management and process scheduling, thus significantly impacts the performance of datacenters. In this paper, we propose Themis, a flexible, automatic and distributed...
In recent years, the fifth revision of the HyperText Markup Language standard (HTML5) has promoted the porting of traditional desktop applications to the Web, thus allowing the rise of Rich Internet Applications (RIAs). Additionally, the progressive support to Web Real-Time Communication (WebRTC) standard in modern browsers for real-time communications is accelerating this process, enabling the migration...
MapReduce is a popular computing model for parallel data processing on large-scale datasets, which can vary from gigabytes to terabytes and petabytes. Though Hadoop MapReduce normally uses Hadoop Distributed File System (HDFS) local file system, it can be configured to use a remote file system. Then, an interesting question is raised: for a given application, which is the best running platform among...
TACC Stats is a continuous monitoring tool for HPC systems that collects data at the core and process level for every job executing on a monitored system. That data can be aggregated at the system, group, user, application, job, node, or core level. TACC Stats has been in production use for about 5 years and is now used by numerous HPC systems around the world. This paper reports on a major new version...
Nowadays, the explosion of cloud-based applications is leading to a much higher demand on both computing and network infrastructure resources than only a few years ago. Enhancing the user experience, by reducing the latency and increasing network stability, becomes an important challenge for cloud operators. In this paper, we propose a unified protocol architecture, based on the Locator/Identifier...
With supercomputer system scaling up, the performance gap between compute and storage system increases dramatically. The traditional speedup only measures the performance of compute system. In this paper, we firstly propose the speedup metric taking into account the I/O constraint. The new metric unifies the computing and I/O performance, and evaluates practical speedup of parallel application under...
With the promise of providing flexible and elastic computing resources on demand, the cloud computing has been attracting enterprises and individuals to migrate workloads in the legacy environment to the public/private/hybrid clouds. Also, cloud customers want to migrate between cloud providers with different requirements such as cost, performance, and manageability. However the workload migration...
This paper makes a short overview of current state of the art monitoring tools for cloud and big data frameworks. In order to effectively create, test and deploy new algorithms or frameworks one needs suitable monitoring solutions. Hence we aim on creating a critical overview for some of the monitoring solutions existing on the market. Also we present relevant metrics used for monitoring cloud and...
Public and private cloud computing environments employ virtualization methods to consolidate application workloads onto shared servers. Modern servers typically have one or more sockets each with one or more computing cores, a multi-level caching hierarchy, a memory subsystem, and an interconnect to the memory of other sockets. While resource management methods may manage application performance by...
The energy consumption of Cloud computing continues to be an area of significant concern as data center growth continues to increase. This paper reports on an energy efficient interoperable Cloud architecture realized as a Cloud toolbox that focuses on reducing the energy consumption of Cloud applications holistically across all deployments models. The architecture supports energy efficiency at service...
A key feature of Cloud computing is its agility and flexibility to support the scalability needs of business solutions. Currently, the agility is only limited to the scalability of the compute, memory and storage. To improve an application's agility, we need to monitor & measure solution level metrics and associate the performance of the metrics to the business agility needs of the solution...
Software-Defined Infrastructure (SDI) presents an approach for integrated management of virtualized heterogeneous resources. Monitoring and measurement is an essential component for effective control and management. This paper presents an architecture of a system, named MonArch, based on SDI that provides integrated monitoring and measurement functionalities. Unlike existing cloud and network monitoring...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.