- Cloud Computing and Resource Management
- IoT and Edge/Fog Computing
- Software System Performance and Reliability
- Distributed systems and fault tolerance
- Advanced Data Storage Technologies
- Green IT and Sustainability
- Cloud Data Security Solutions
- Distributed and Parallel Computing Systems
- Advanced Queuing Theory Analysis
- Software-Defined Networks and 5G
- Mobile Health and mHealth Applications
- Process Optimization and Integration
- Smart Grid Security and Resilience
- Blockchain Technology Applications and Security
Universidade Federal de Pernambuco
2017-2022
Guaranteeing high levels of availability is a huge challenge for cloud providers. The authors look at the causes failures and recommend ways to prevent them minimize their effects when they occur.
Cloud computing has gained popularity in recent years due to its pay-as-you-go business model, high availability of services, and scalability. Service unavailability does not affect just user experience but is also translated into direct costs for cloud providers companies. Part this SLA breaches, once interruption time greater than those signed the contract generate financial penalties. Thus, have tried identify failure points estimate their services. This paper proposes models assess...
E-health systems can be used to monitor people in real-time, offering a range of multimedia-based health services, at the same time reducing cost since cheaper devices compose it. However, any downtime, mainly case critical result patient problems and worst case, loss life. In this paper, we use an interdisciplinary approach combining stochastic models with optimisation algorithms analyse how failures impact e-health monitoring system availability. We propose surrogate estimate availability...
A data center is divided into three basic subsystems: information technology (IT), power, and cooling. Cooling plays an important role related to availability, a failure in this subsystem may cause interruption of services. Generally, redundant cooling implemented based on replacing the failed component by standby one. However, it also can be rotation computer room air conditioners (CRACs). This paper proposes scalable models that represent behavior evaluate impact failures availability....
The cloud data center is a complex system composed of power, cooling, and IT subsystems. power subsystem crucial to feed the equipment. Power disruptions may result in service unavailability. This paper analyzes impact failures on services regarding different architecture configurations based TIA-942 standard such as non-redundant, redundant, concurrently maintainable, fault tolerant. We model both subsystems, IT, through Stochastic Petri Net (SPN). availability results show that tolerant...
E-health systems can be used to monitor people in real-time, offering a range of multimedia-based health services, at the same time reducing cost since cheaper devices compose it. However, any downtime, mainly case critical result patient problems and worst case, loss life. In this paper, we use an interdisciplinary approach combining stochastic models with optimisation algorithms analyse how failures impact e-health monitoring system availability. We propose surrogate estimate availability...
Large data centers are complex systems that depend on several generations of hardware and software components, ranging from legacy mainframes rack-based appliances to modular blade servers modern rack scale design solutions. To cope with this heterogeneity, the center manager must coordinate a multitude tools, protocols, standards. Currently, managers, standardization bodies, hardware/software manufacturers joining efforts develop promote Redfish as main management standard for centers, even...
The emergence of new computing paradigms such as fog and edge provides the Internet Things with needed connectivity high availability. In context e-health systems, wearable sensors are being used to continuously collect information about our health, forward it for processing by Medical (IoMT). E-health systems designed assist subjects in real-time providing them a range multimedia-based health services personalised treatment promise reducing economic burden on systems. Nonetheless, any...
Traditional data center infrastructure suffers from a lack of standard and ubiquitous management solutions. Despite the achieved advances, existing tools interoperability are sometimes hardware dependent. Vendors already actively participating in specification design new software interfaces within different forums. Nevertheless, complexity variety components that includes servers, cooling, networking, power hardware, coupled with introduction defined paradigm, led to parallel development...
Summary Next‐generation cloud data centers are based on software‐defined center infrastructures that promote flexibility, automation, optimization, and scalability. The Redfish standard the Intel Rack Scale Design technology enable infrastructure disaggregate bare‐metal compute, storage, networking resources into virtual pools to dynamically compose create performance‐optimized (vPODs) tailored workload‐specific demands. This article proposes four chassis design configurations Distributed...
Many enterprises rely on cloud infrastructure to host their critical applications (such as trading, banking transaction, airline reservation system, and credit card authorization). The unavailability of these may lead severe consequences that go beyond the financial losses, reaching provider reputation too. However, maintain high availability in a data center is difficult task due its complexity. power subsystem crucial for entire operation because it supplies all other subsystems, including...
Making data centers highly available remains a challenge that must be considered since the design phase. The problem is selecting right strategies and components for achieving this goal given limited investment. Furthermore, center designers currently lack reliable specialized tools to accomplish task. In paper, we disclose formal method chooses optimize availability of while considering budget as constraint. For that, make use stochastic models represent cloud infrastructure based on...
Summary Emergency call services are expected to be highly available in order minimize the loss of urgent calls and, as a consequence, life due lack timely medical response. This service availability depends heavily on cloud data center which it is hosted. However, information alone cannot provide sufficient understanding how failures impact and users' perception. In this paper, we evaluate an emergency system, considering service‐level metrics such number affected per failure time takes...
Summary To assess the availability of different data center configurations, understand main root causes failures and represent its low‐level details, such as subsystem's behavior their interconnections, we have proposed, in previous works, a set stochastic models to architectures (considering three subsystems: power, cooling, IT) based on TIA‐942 standard. In this paper, propose Data Center Availability (DCAV), web‐based software system allow operators evaluate infrastructure through...
Cooling plays a very important role in data centre availability by mitigating the overheating of Information Technology (IT) equipment. While many existing works evaluated performance cooling sub-systems centres, only few studies have considered relationship between and IT sub-systems. This work provides efficient models (using Stochastic Petri Nets (SPNs)) to represent sub-system analyse impact its failures terms service downtime financial cost. We provide an model, diminishing state...
Users pay for running their applications on cloud infrastructure, and in return they expect high availability, minimal data loss case of failure. From a provider perspective, any hardware or software failure must be detected recovered as quickly possible to maintain users' trust avoid financial losses. user's failures transparent should not impact application performance. In order recover failed application, providers perform checkpoints, periodically save data, which can then following...
Users pay for running their applications on cloud infrastructure, and in return they expect high availability, minimal data loss case of failure. From a provider perspective, any hardware or software failure must be detected recovered as quickly possible to maintain users' trust avoid financial losses. user's failures transparent should not impact application performance. In order recover failed application, providers perform checkpoints, periodically save data, which can then following...
Cooling plays a very important role in data centre availability by mitigating the overheating of Information Technology (IT) equipment. While many existing works evaluated performance cooling sub-systems centres, only few studies have considered relationship between and IT sub-systems. This work provides efficient models (using Stochastic Petri Nets (SPNs)) to represent sub-system analyse impact its failures terms service downtime financial cost. We provide an model, diminishing state...