Academia.eduAcademia.edu

Data Replication

description1,578 papers
group581 followers
lightbulbAbout this topic
Data replication is the process of storing copies of data in multiple locations or systems to ensure consistency, availability, and reliability. It is commonly used in database management and distributed computing to enhance data accessibility and fault tolerance.
lightbulbAbout this topic
Data replication is the process of storing copies of data in multiple locations or systems to ensure consistency, availability, and reliability. It is commonly used in database management and distributed computing to enhance data accessibility and fault tolerance.

Key research themes

1. How do data replication protocols balance availability, consistency, and efficiency in distributed systems?

This research area investigates the design, analysis, and performance evaluation of data replication protocols that ensure consistent and available access to replicated data under various failure conditions and system constraints. It matters because the trade-offs between availability, fault-tolerance, communication overhead, and consistency dominate the effectiveness of replicated data management in distributed and cloud environments. Understanding these protocols aids in deploying resilient, high-performance distributed systems.

Key finding: The paper analyzes original available copy protocols and two variants (naive and optimistic), demonstrating through Markov models that these variants nearly match the original in availability and reliability while not... Read more
Key finding: This study differentiates failure unavailability (due to site failures) from conflict unavailability (due to concurrent access conflicts) and reviews techniques to improve replica availability by refining replica control (RC)... Read more
Key finding: This research develops a genetic programming approach to automatically generating and evolving data replication strategies optimizing the trade-off between availability and operation cost. It demonstrates that novel,... Read more
Key finding: This survey synthesizes the mechanisms of various replication techniques in grid environments, highlighting their impact on availability, fault tolerance, and performance in geographically distributed systems. It emphasizes... Read more

2. What middleware-level approaches integrate transactional concurrency control and group communication to enable scalable, consistent data replication?

This theme explores middleware designs that lie between applications and databases to achieve consistent and scalable data replication without requiring intrusive modifications to underlying database systems. The research examines leveraging transactional protocols with group communication primitives to reduce redundant computation, maintain one-copy serializability, and optimize communication overhead, important for systems like web farms and distributed object platforms.

Key finding: Proposes a middleware-level replication engine combining transactional concurrency control with group communication to maintain one-copy serializability. Introduces protocols that execute transactions at a primary site to... Read more
Key finding: Describes the implementation and performance evaluation of a replication framework supporting pessimistic and optimistic active replication using atomic broadcast primitives. The prototype confirms that in large-scale... Read more
Key finding: Introduces EA2-IMDG that leverages in-memory data grids (IMDGs) to reduce latency and improve scalability of replication and task scheduling in grid systems. By distributing data in RAM across nodes, it minimizes disk I/O... Read more
Key finding: Proposes ICTSDC, which tightly couples task scheduling and data replication by leveraging a self-adaptive Dwarf Mongoose Optimization (SADMO) algorithm. The model optimizes objectives such as bottleneck reduction, migration... Read more

3. How are data replication strategies in cloud environments optimized for multi-objective goals including provider cost, energy consumption, performance, and SLA satisfaction?

This research theme focuses on dynamic and static replication strategies in cloud systems that consider economic factors, energy efficiency, and SLA requirements alongside performance metrics. Approaches include elastic replica management, economic modeling, heuristic optimization, and data mining-based methods to balance replication overhead with provider profit and tenant QoS demands, addressing the challenges created by cloud heterogeneity and large-scale distributed data.

Key finding: Proposes a dynamic replication strategy balancing provider profit and tenant SLA satisfaction, using a cost model that enables replication only when necessary. By incorporating both response time and economic benefit in... Read more
Key finding: Surveys cloud data replication strategies across multiple dimensions including static/dynamic operation, workload balancing approaches, replica factor adjustment, and objective functions. Emphasizes that effective cloud... Read more
Key finding: Introduces E2ARS, a static, multi-objective replication strategy that jointly reduces cloud provider energy consumption and expenditure under SLA constraints. Employs optimization algorithms that leverage cloud heterogeneity... Read more
Key finding: Develops a hybrid replication strategy based on quorum voting structures that balances availability and access operation costs across varying scenarios. The approach supports flexible configuration of read/write quorums to... Read more
Key finding: Proposes a novel algorithm combining particle swarm optimization (PSO) with fuzzy logic system for replica placement and replacement in cloud environments. The method optimizes conflicting objectives such as service time,... Read more
Key finding: Introduces GUEES, a hybrid algorithm combining Sealion Optimization Model and Grey Wolf Optimizer to identify frequent data access patterns for informed data replication. By prioritizing data queues and evaluating storage... Read more

All papers in Data Replication

The (latest) crisis in confidence in social psychology has generated much heated discussion about the importance of replication, including how it should be carried out as well as interpreted by scholars in the field. For example, what... more
In a much-publicized paper, Zhong and Liljenquist (2006) reported evidence that feelings of moral cleanliness are grounded in feelings of physical cleanliness: a threat to people’s moral purity leads them to seek, literally, to cleanse... more
by Emma Boyland and 
1 more
Food advert exposure has been shown to influence calorie intake and food choice in 9-11 year olds. However, little is known about the effect of food advertisements on feeding behaviour in younger children. Therefore, we conducted a study... more
Today Wikipedia is one of the most famous online encyclopedias providing very useful information in various areas of the human knowledge. The most important thing is that anyone can write and edit the articles in the encyclopedia or we... more
THE SYNCHRONIZATION OF DATABASE AT THE FACULTY OF ENGINEERING AND THE RECTORATE OFFICE, HASANUDDIN UNIVERSITY, MAKASSAR
The sounds produced when we touch textured surfaces frequently provide information regarding the structure of those surfaces. It has recently been demonstrated that the perception of the texture of the hands can be modified simply by... more
Abstract: Despite the numerous studies on the determinants of subjective wellbeing (SWB), there are still under-researched areas as follows: a full-model approach allowing un-confounded and robust estimations, extension of... more
Replication is an essential cornerstone of cloud storage where 24x7 availability is needed. Failures are normal rather than exceptional in the cloud computing environments. Aiming to provide high reliability and cost effective storage,... more
Cloud computing is growing rapidly over the years and it faces challenges especially in resource management. Resource management in cloud computing is necessary due to its distributed nature with different user demands. Quality of Service... more
Trichotillomania (TTM) is a chronic impulse control disorder characterized by repetitive hair-pulling resulting in alopecia. Although this condition is frequently observed in children and adolescents, research on pediatric TTM has been... more
Commentary on: Rolf A. Zwaan, Alexander Etz, Richard E. Lucas, and M. Brent Donnellan (2017). Making Replication Mainstream. Behavioral and Brain Sciences, published online: 25 October 2017, pp. 1-50; forthcoming at... more
I propose a simple and computationally undemanding method for extending Pemstein, Meserve, and Melton (2010)'s Unified Democracy Scores (UDS) from 2015 all the way back to the beginning of the 19th century (and in some cases to the late... more
Geospatial data sharing is an increasingly important subject as large amount of data is produced by variety of sources, stored in incompatible formats, and accessible through different GIS applications. Past efforts to enable sharing have... more
The sounds produced when we touch textured surfaces frequently provide information regarding the structure of those surfaces. It has recently been demonstrated that the perception of the texture of the hands can be modified simply by... more
As we delve deeper into the ‘Digital Age’, we witness an explosive growth in the volume, velocity, and variety of the data available on the Internet. For example, in 2012 about 2.5 quintillion bytes of data was created on a daily basis... more
One of the basic services in grids is the transfer of data between remote machines. Files may be transferred at the explicit request of the user or as part of delegated resource management services, such as data replication or job... more
In data-intensive distributed systems, replication is the most widely used approach to offer high data availability, low bandwidth consumption, increased fault-tolerance and improved scalability of the overall system. Replication-based... more
As a pillar of Data Access and Research Transparency (DA-RT), analytic transparency calls for radical honesty about how political scientists infer conclusions from their data. However, honesty about one's research practices often means... more
As we delve deeper into the ‗Digital Age', we witness an explosive growth in the volume, velocity, and variety of the data available on the Internet. For example, in 2012 about 2.5 quintillion bytes of data was created on a daily basis... more
INTRODUCTION: While surgical resection has been shown to improve short-term local disease control, it remains debated whether surgical resection is associated with improved overall survival in patients with malignant primary osseous... more
Mobile ad hoc networks (MANET) are becoming an integral part of the ubiquitous computing and communication environment, providing new infrastructure for multimedia applications such as video phone, multimedia-on-demand, and others. In... more
A sample of 93 veterans (92.4% males), with a median age of 41, (Mean=43.5) attending clinics for problem drinking, drug abuse and other mental disorders was screened for problems associated with the diagnosis of pathological gambling.... more
Grid computing is a type of distributed computing system that provides access to various computational resources which are shared by different organizations, in order to create an integrated powerful virtual computer. Nowadays, grid is... more
The neurohormone oxytocin (OT) has been one the most studied peptides in behavioural sciences over the past two decades. Primarily known for its crucial role in labour and lactation, a rapidly growing literature suggests that intranasal... more
As the use of the Internet continues to grow explosively, edge computing has emerged as an important technique for delivering Web content over the Internet. Edge computing moves data and computation closer to end-users for fast local... more
With traditional PLM (Product Lifecycle Management), people think towards the future: first comes product development, then manufacturing, then support and finally disposal -- data flows only in forward direction. With the CL2M... more
by T. McLaughlin and 
1 more
The purpose of this study was to use Direct Instruction (DI) flashcards to teach high use sight words. The participant was a second grade student with a learning disorder and Attention Deficit Hyperactive Disorder (ADHD). The study was... more
The Digital Imaging and Communications in Medicine (DICOM) standard defines Radiology medical device interoperability and image data exchange between modalities, image databases - Picture Archiving and Communication Systems (PACS) - and... more
Grid Computing is a type of parallel and distributed systems that is designed to provide reliable access to data and computational resources in wide area networks. These resources are distributed in different geographical locations,... more
by Jorge Peña and 
1 more
This study examined how avatar body size (normal, obese) and opponent character body size (normal, obese) influenced physical activity while male participants played an exergame. Males operating normal weight avatars showed more physical... more
In this paper, a review for consistency of data replication protocols has been investigated. A brief deliberation about consistency models in data replication is shown. Also we debate on propagation techniques such as eager and lazy... more
In this paper we investigate the performance issues of data replication in a loosely coupled distributed database system, where a set of database servers are connected via a network. A database replication scheme, Replication with... more
In data grids, many distributed scientific and engineering applications often require access to a large amount of data (terabytes or petabytes). Data access time depends on bandwidth, especially in a cluster grid. Network bandwidth within... more
Agradeço à minha mãe, minha irmã e ao meu padrasto, que são o meu pilar e minha motivação para continuar sempre alcançando meus objetivos e realizando meus sonhos.
With traditional PLM (Product Lifecycle Management), people think towards the future: first comes product development, then manufacturing, then support and finally disposal -data flows only in forward direction. With the CL 2 M... more
Data Grids seek to harness geographically distributed resources for large-scale data-intensive problems. Such problems, involving loosely coupled jobs and large data-sets, are found in fields like high-energy physics, astronomy and... more
Data Grids seek to harness geographically distributed resources for large-scale data-intensive problems. Such problems, involving loosely coupled jobs and large data-sets, are found in fields like high-energy physics, astronomy and... more
Cloud computing is a concept of providing user and application oriented services in a virtual environment. Users can use the various cloud services as per their requirements dynamically. Different users have different requirements in... more
Background: Swedish male smokers are more likely than female smokers to switch to smokeless tobacco (snus) and males' smoking cessation rate is higher than that of females. These results have fuelled international debate over promoting... more
In peer-to-peer content distribution the lack of a central authority makes authentication difficult. Without authentication, adversary nodes can spoof identity and falsify messages in the overlay. This enables malicious nodes to launch... more
Comparative statistical analyses often require data harmonization, yet the social sciences do not have clear operationalization frameworks that guide and homogenize variable coding decisions across disciplines. When faced with a need to... more
INTRODUCTION: While surgical resection has been shown to improve short-term local disease control, it remains debated whether surgical resection is associated with improved overall survival in patients with malignant primary osseous... more
Replication is one of the popular tools to determine the availability degree of resources (e.g., data files) in data grids. Since data grids involve limited file storages and high computing costs, replication process is very essential in... more
With "peer production" becoming commonplace and new standards like RDF or OWL paving the way for the much anticipated semantic web, a new breed of very large scale semantic systems is about to appear. Traditional semantic reconciliation... more
tion. Abstract: In order to meet their temporal constraints, current applications such as Web-based services and electronic commerce use the technique of data replication. To take the replication benefit, we need to develop con- currency... more
Due to the third-generation mobile networks and high bandwidth wireless networks, grid computing has moved from traditional parallel and distributed model to mobility-based model. This paradigm shift has given rise to in-creased use of... more
Data replication is gaining increased importance due to the increasing demand for availability, performance and fault tolerance in databases. The main challenge for deploying replicated databases on a large scale is to resolve conflicting... more
Download research papers for free!