Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Ijcet: International Journal of Computer Engineering & Technology (Ijcet)

Download as pdf or txt
Download as pdf or txt
You are on page 1of 12

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & ISSN

0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

TECHNOLOGY (IJCET)

ISSN 0976 6367(Print) ISSN 0976 6375(Online) Volume 4, Issue 6, November - December (2013), pp. 314-325 IAEME: www.iaeme.com/ijcet.asp Journal Impact Factor (2013): 6.1302 (Calculated by GISI) www.jifactor.com

IJCET
IAEME

QoS ENSURED OPTIMAL REPLICA PLACEMENT IN GRAPH BASED DATA GRIDS


N. Devakirubai1 and
1

S. Palani2

Dept. of Computer Science and Engg., 2Dept. of Electronics and Communication Engg. Jayam College of Engineering and Technology, Dharmapuri.

ABSTRACT This paper studies the QoS ensured optimal replica placement problem. Most of the studies on replica placement concerns only about minimizing the data access time, without considering the quality of service issue. In heterogeneous environments, the quality assurance is important. We propose an algorithm named Replica Placement in Graph Topology Grid (RPGTG), which optimally place the replicas and also ensures QoS. In this paper we study optimization problem, which addresses the issues, minimizing the data access time, load balancing among replica servers and avoiding unnecessary replications. Along with these issues we restrict the QoS in terms of number of hops and deadline to complete the request. Keywords: Data Grid, Data Replication, Graph topology grid, Replica Placement, Quality of Service. 1. INTRODUCTION

Grid can be defined as utilizing the unused resources of computers in a network, to solve a single problem that is too intensive for any stand-alone machine. Grid computing is commonly used for scientific or technical problem which requires huge number of computer processing cycles and need to access very large amounts of data. The real environment architecture of grid is general graph [1]. When a single computer is not able to perform a large computational task, it is divided up among individual machines, which may span multiple administrative domains and geographically distant. These machines run calculations in parallel and return back the results to the original machine. This is cost effective than Supercomputers and called as Computational Grid.

314

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

Data grid is an architecture or group of services which enable users to access, modify and transfer very huge amount of data which are geographically distributed. This architecture provides the services like storage systems, data access and meta data services [2]. The main consideration of the data grid is to maximize the data availability to users. The major issues while accessing data in a data grid is network latency and bandwidth problems. Due to the dynamic changes in the size of the grid, maximum data availability becomes the challenging issue. This challenge can be overcome by Data Replication, which increases data availability, reduced bandwidth consumption, increased fault tolerance, improved scalability and improved response time. Copying data files at more than one place in a grid is called as Data Replication. Data Replication is required for fast and efficient data access and also ensures storage availability and network bandwidth availability [3]. Issues to be considered in data replication are balancing the number of replicas in grid sites, placing replicas in the appropriate site, maintaining the consistency of the replicated data and avoiding unnecessary replications. The purpose of data replication is to place the data so that the user can access the data fast and efficiently. We are proposing an algorithm called Replica Placement on Graph Topology Grid (RPGTG) to meet the replica placement issues. As a first step, the communication with in the architecture used by our algorithm is discussed. The replication technique is completely based on the architecture of the grid. In all the grid architectures, like Multitier architecture (tree architecture), peer to peer architecture, graph architecture and hybrid architecture, the nodes are considered as sites. Each site is composed of two basic elements: Storage Element (SE) and Computing Element (CE). CE follows the access load of site and SE follows the availability of storage device represented by their storage load [4]. Placing replica in the parent node of a node which generates the maximum request is the common practice. So the access load can be defined as the cumulative total number of requests responded by a node, where the requests are received from all of its children. Though in the real environment the sites are arranged in the graph topology, for better managing of replicas, the graph structure will be viewed as tree structure, with the help of Breadth First Search (BFS) and Depth First Search (DFS) algorithm. BFS determines the level of each node and DFS determines the adjacent node list of each node. Then the replica selection and placement algorithms are implemented on the resulted tree structure. The replacement algorithm monitors the storage load of each replica server. The above algorithm solved the replica placement issue and an optimal placement is achieved. While an average performance metric is important, the above said algorithm can not provide the QoS guarantees. In this paper, we improve the optimal replica placement algorithm to support Quality of Service. 2. LITERATURE SURVEY

In [5] Ding et al. considers grid topology as clusters, with each cluster having varied storage and computational capabilities. They proposed a data placement algorithm, which automatically stores new data objects and a self-tuning data replication algorithm, which fastly adapts to system changes and performs an efficient and effective data replication. This replication algorithm outperforms the general threshold based algorithms in terms of efficiency and load balancing. In [6] Rasool et al. proposed a two way replication strategy. The multi-tier sibling tree architecture is used which is a mixture of the architectures, presented by Ranganathan and Lin. It's a hierarchical model in which all the siblings are connected to each other as well. In this two way replication (TWR) scheme the most popular data is identified and placed to its proper host in a bottom up manner in this they are closer to the clients. In top down manner the less popular files are identified and are placed to one tier below the root node, in this way they are close to the roots. In
315

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

this approach, replica selection is done by using the closest policy which tries to provide the data from the nearest site. The drawback of the research is that it only considers the homogeneous data grid nodes and cannot be applied to heterogeneous nodes while the nodes in a data grid are normally heterogeneous. In [7] Lin et al. have addressed the problem of placement of a new replica in a proper place by considering a priority list. The proposed replica placement algorithm finds out the minimum number of replicas when the maximum workload capacity of each replica is given. The hierarchical model is different from other related works, because in this model they assume a logical connection between all siblings of a parent and a request can be served from a node present in sibling ring. If requested data is not present in sibling ring then parent ring is searched. This architecture is called a Sibling Tree model, which is an extension of a normal tree structure. The hierarchical model assumes a logical connection between the siblings and actually all connections from one sibling to another physically involves the parent i.e. at most two hops. This means the actual time taken to serve a request is infected more than it is presented, as this logical connection is assumed physical and already the time complexity is too high. The problem of network congestion or bandwidth consumption is not mentioned in proposed model. In [8] Golda. J et al. they have worked for ensuring data availability, improving fault tolerance, reducing file access time, minimizing file transfer cost and controlling network bandwidth usage. They proposed two data replication algorithms in their paper. First, when geographically distributed storage servers are concerned for data replication, Centralized Data Replication algorithm is used. Second when the proximity among the storage servers are really high, then the Distributed Data Replication algorithm is used. In these two cases, the final replication decision is based on the characteristics such as need count of the files, minimum number of hops, size of the data files, bandwidth of file transmission, request rates of files across the sites, file location and capacity of storage servers. The drawback in this research is grid topology is not graph based. In [9] Garmehi. M et al stated that data replication is one of the best-known strategies to achieve high levels of availability and fault tolerance, as well as minimal access times for large, distributed user communication using a worldwide data grid (DG). One of the challenges in data replication is to select the candidate sites where replicas should be placed and which are known as optimal placement of replicas (OPR). In that paper an algorithm is proposed, which is formulated by using dynamic programming method to find optimal placement k replicas of an object over DG systems, such that the overall cost (i.e. storage cost plus read cost) is minimized. In [4] Jianzhong Li et al. they proposed a load balancing replication strategy, Fair-Share Replication (FSR) that takes into account both the access load and storage load of the replica servers before placing a replica. They use P2P concepts in multi-tier data grid with unique path for data search, and leverage the neighborhood of replica servers in replication process. A sibling node of a candidate replica server is selected for replica placement if its access and storage loads are less than the candidate. Experiment results show that FSR performs better than fast spread replication using random access pattern. In [10] Tang et al. present a dynamic replication algorithm for multitier Data Grids. They propose two dynamic replica algorithms: Single Bottom Up and Aggregate Bottom Up. ABU achieved great performance improvements for all access patterns even if the available storage size of the replication server is very small. They compared the two algorithms to Fast Spread dynamic replication strategy and found ABU proves to be superior. As for SBU, although the average response time of Fast Spread is better in most cases, Fast Spreads replication frequency is too high to be applicable in the real world. Performance results show both algorithms reduce the average response time of data access compared to a static replication strategy in a multitier Data Grid.

316

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

In [11] Yaser. N et al. proposed a dynamic data replication policy. They developed a replication strategy for 3-level hierarchical structure, called HRS (Hierarchical Replication Strategy). In this replication algorithm, replicated file stored in the site with the largest number of access for the file and main factor for replica selection is bandwidth. The replication strategy proposed in this paper allows placing data in a manner that will suggest a faster access to files require by grid jobs, therefore there is an increase in the job executions performance. If the available storage for replication is not sufficient, the proposed algorithm only replicates those files that do not exist in the local LAN. In [12] Abdullah et al. presented a P2P model for higher availability, reliability, and scalability. Then have developed their own data grid simulator to test the proposed replication strategy, taking response time, number of hops and average bandwidth consumption as basic parameters for evaluation. In this research they are studying four replication strategies, out of which two are existing strategies: "requester node placement strategy" and "path node placement strategy", and two are newly proposed in this research: "path and requester node placement strategy", and "Nhop distance node placement". In the "requester node placement strategy", when a required file is found then it is only replicated to the requester node. In the "Path node placement strategy" the file is replicated to all the nodes on the path from the requester node to provider node. The newly proposed strategy "Path and requester node placement strategy" is a combination of the first two strategies. In "N-hop distance node placement" a file is replicated to all neighbors' of provider nodes within an n hop distance. The results of their simulation show that new strategies have shown better performance than existing ones in terms of performance, success rates and response time. However, the proposed strategies use more bandwidth than the existing strategies. The drawback of the research is that the storage loads of replica servers are not considered in their strategies, because the file is replicated to all the nodes on the path from the requester node to provider node. In [13] Sashi et al. presented a modified form of BHR (Bandwidth Hierarchy Replication) to overcome its limitations. In the modified BHR model a network region is defined as a network topological space where sites are located closely. Whenever the required replica is present in the same region, the job completion will be fast. Again, the Modified BHR model is based on tree structure which is not very suitable in real data grid environment. Some of the first authors to introduce actual QoS constraints in the problem were Tang and Xu [14]. In their approach, the QoS corresponds to the latency requirements of each client. Different access policies are considered. First, a replica-aware policy in a general graph with heterogeneous nodes is proven to be NP-complete. When the clients do not know where the replicas are (replicablind policy), the graph is simplified to a tree (fixed routing scheme) with the Closest policy, and in this case again it is possible to find an optimal dynamic programming algorithm. In [15], Wang et al. deal with the QoS aware replica placement problem on grid systems. In their general graph model, QoS is a parameter of communication cost. Their research includes heterogeneous nodes and communication links. They proposed a heuristic algorithm and compared to the results of Tang and Xu [14]. 3. ARCHITECTURE OF REPLICA PLACEMENT IN GRAPH TOPOLOGY GRID ALGORITHM

Architecture for the proposed algorithm Replica Placement in Graph Topology Grid (RPGTG) is explained in this section. Since the true representation of grid is general graph, where there is no root node, we consider the grid topology as graph. For example, the CMS Data Challenge 2002 grid configuration [16]. Due to better managing of replica servers and their related nodes, we convert the graph structure to hierarchical structure. Here the data grid is modeled to have three tiers,
317

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

where the tier 0 machines have enormous storage capacity. The tier 1 machines are called as Regional Servers have computing and storage resources. The tier 2 machines are called as Local Servers and tier 3 machines are workstations. A network topological space where sites are closely located is called as network region. Job completion will be fast, if the required replica is found within the network region. To facilitate dynamic file replication in the multi-tier data grid, the following services are available in the system [17]. The Replica Catalog (RC) and Replica Manager (RM) are located at Regional servers and the Local Replica Catalog (LRC) and the Local Replica Manager (LRM) are the local services which are distributed on every machine in the system. The RC and RM at the Regional Servers manage the LRC and LRM of sites which are connected to them. Replica Catalog helps to recognize the physical locations of data files. Physical File Name (PFN) is a URL pointing to a physical copy of the file. The Logical File Name (LFN) is a system-wide unique identifier of PFN. Replica Catalog has a database that stores the mappings between LFN and PFN. This database acts as a registry and keeps track of where files are placed in the grid. Another database in Replica Catalog stores the breadth-first and depth-first order of each node, by which each node can be aware of the nodes which are located at its sub tree. The level indicates the distance of a particular node from the root node. Replica Location Service (RLS) in Replica Catalog invokes the above two databases to store mapping between LFN and PFN. The function of Replica Manager is to perform replication and creating replicas. Replica Manager contains the following components: Reservation Manager, which authenticate the reservation requests for resources based on the resource availability. Allocation Manager is a component that communicates with the resource manager to allocate the resources based on the reservations. Replica selector, selects the replica with the minimum communication cost, where communication cost is obtained by dividing the size of replica i with available bandwidth between grid site a and b. Catalog Service is a component invoked by optimizer, which interacts with RLS to get the information from the databases in Replica Catalog. The component Replicator replicates the chosen replica on the optimal place found by our proposed algorithm. Replica Selection component has the following components: Available Bandwidth Estimation (ABWE) is a monitoring tool used to estimate the Round Trip Time (RTT), against which the parameter d is compared to choose the nodes which can respond before their deadline d and ABWE also estimates the available bandwidth between the host pair of nodes. The source node uses ABWE for sending continuously examined packets to the nodes which has the required replica. These nodes in turn reply about the delay and available bandwidth. The source node collects the replies and chooses the target node with the minimum communication cost. The second component Threshold Controller [18] determines the threshold value by comparing the average aggregated access request count and available server capacity.

318

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME
Site 0 (Requester Site) Request (F,M,d, h) Replica Catalog (RC) Replica Location Service (RLS) Candidate Site IDs Replica Manager (RM) Request (F,M) Catalog Service Replica Selector (RS) Replica Placement Component (RPC) ABWE Tool DB (File ID,Site ID) Return available bandwidth of all CS Sends packets to receive the available RPC calculates communication cost using available bandwidth RPC chooses site with min cost ( distance < h) RPC computes RTT Yes If d>RTT No No (one of Sibling site ID Forwards Request Yes No Storage Management Component Checks for available free space in the site Replica Replacement component Reservation Manager Reserves resources

Search to Find PFN of site that contains desired replica

Search to Find the level and DFS order of site that contains desired replica DB (Site ID,Level,order)

Allocation Manager Place Replica File at Target Site Allocation resources Replicator

Req.site has to access file remotely

Perform Replication

Threshold controller Checks whether the site has min. access load

Yes

LCA (target site, Requester site)

Site 1

Site 2 (CS)

Site 3 (CS)

Target Site Site (n-1) Site n

Fig. 1 Message Passing between the components of RPGTG Algorithm The average aggregated access count is calculated by dividing the total number of aggregated access counts for a file at the replica server at each level with the number of the replica servers at their level. File is called as popular file when its access count exceeds the threshold value. The number of replica creation is reduced when the threshold value is greater and when the request rate decreases. The third component in Replica Selector is the Replica Placement Management. The proposed algorithm uses this component to find the minimum distance between source and destination, which is limited by the parameter number of hops h. After considering the access rate of each replica server, this component chooses the replica server with minimum access load. Target nodes which have their access rate lesser than the threshold value are selected for replication. 4. 4.1 REPLICA PLACEMENT IN GRAPH TOPOLOGY GRID ALGORITHM

Determining the Level and Adjacent Nodes Of Each Node Every node in the graph structure is a site in real data grid structure, and every edge demonstrates the relations between sites. Graph structure is traversed by BFS algorithm by which the level of each node can be determined. Tree T is obtained after traversing our graph structure by the BFS algorithm. The tree T is traversed by DFS algorithm which determines the number of adjacent nodes of each node. List L is maintained to store the DFS order of nodes. The identification number given to any node is smaller than the identification number given to its descendents. The above mentioned identification number is assumed to be the grid site IDs. After traversing the tree by the
319

No Site ID returned by LCA( )

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

BFS and DFS algorithms, one of the database in the RC get data about SiteID, level, depth_order of each node. So now the Local Server and Regional Server of our architecture are aware of the gird sites that are placed on its sub tree. 4.2 Finding the Lowest Common Ancestor (LCA) To find the LCA(u,v) in List L, find the first occurrence of u and v, this part of numbers form the interval I. The minimum number in I is the value for LCA(u,v). 4.3 Popular Files The RM in Regional Servers aggregate the access record of each file from lower to upper level to determine the threshold value. The most Popular files are the one which are frequently accessed by clients and whose access count exceeds the threshold value. The threshold value is set based on the average aggregated access count at the replica servers in each level. The average aggregated access count is calculated by dividing the total number of aggregated access count for a file at the replica servers at a level by the number of replica servers at that level. 4.4

Functioning of Proposed Algorithm When a grid site needs a popular file which is not stored locally, the request will be sent to the Local Server with information like requested file (F), requesting machine (M), deadline (d) and number of hops (h). Whenever Local Server receives a file request, it stores LFN of that file, PFN of the requestor site at its RC in the database that contains (File ID, Site ID). If the requested file is not available in Local Server, the request is forwarded to Regional Server. The request is forwarded to the Catalog Service in RM for determining which grid sites have the requested replica. Catalog Service contacts the RLS at the RC, which invokes the two databases: database which stores File ID, Site ID and database which stores Site ID, Level, Order information. The RLS searches in database and returns back the site information that contains the desired replica and these sites are called as candidate sites. Among these candidate sites, the site with minimum distance from the requester site is chosen as target site. This is done by the Replica Selector Component, which selects the target site with minimum communication cost. The Replica Placement component in Replica selector, computes the minimum distance between the requester site and the candidate sites which are all satisfies the condition (distance < h). Replica Placement Component computes Round Trip Time (RTT) through ABWE (Available Bandwidth Estimation) Tool. ABWE is a low network intrusive monitoring application based on packet pair techniques and designed to work in continuous mode. The requester site use ABWE for sending continuously probe packets to all other candidate sites. The candidate sites reply the available bandwidth to the Replica Placement Component, which in turn uses the available bandwidth information to calculate the communication cost of each candidate site. The communication cost is calculated by dividing the sizei with the bandwidthab (available bandwidth between grid site a and grid site b). The Replica Selector now chooses the site with the minimum communication cost. As each request has the deadline, our proposed algorithm now decides whether to perform replication or to access the file remotely by comparing the requests deadline. If (requests deadline > RTT) then the requester can access the file remotely or else replication will be performed. Consider if the above condition results false, then replication has to be performed; now the question arises about where to replicate the file. To replicate on the requester site the Storage Management Component is contacted to check for the availability of space, if there is space for storing replication the Replica Placement Component contacts the Threshold controller to decide whether this site has minimum access load, the request is sent to replicator to perform the replication
320

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

process. Replicator forwards this request to Reservation Manager through Storage Management Component. If Reservation Manager succeeds in making reservations, RM calls allocation manager. Once Allocation Manager finishes allocating the resource reservation, RM starts the file transferring from the identified target machine to the requestor site. If Reservation Manager does not succeed in making reservations on requester site, the proposed algorithm finds LCA of the identified target site and requester site. To replicate on the site identified by LCA algorithm, the Storage Management Component is contacted to check for the availability of space, if there is space for storing replication, the Replica Placement Component contacts the Threshold controller to decide whether this site has minimum access load, if yes, replication is performed or else replication is performed on one of its sibling node whose access load is lesser than the threshold value. If no grid sites within the Local Server and Regional Server contain the desired replica, the request will be sent upward towards root. Replacement is performed when a remote replica has been selected for replication to the requester sites storage element and the storage element might not have sufficient space. In this case one or more replica must be deleted by LRU algorithm. 4.5 Algorithm To Determine The Level And Adjacent Nodes Of Each Node 4.5.1 Run the BFS algorithm to traverse the graph structure. 4.5.2 Let T denotes a tree which is traversed by the Step 4.5.1. 4.5.3 Maintain the adjacent of each node on the tree T. 4.5.4 Run DFS algorithm to traverse the tree T. 4.5.5 Maintain the List L of nodes in the same order that they are visited. 4.5.6 Keep the sub tree of every node. 4.6 4.6.1 4.6.2 4.6.3 Algorithm To replicate and replica placement Submit request (job) to grid. Every request sent to Replica manager of Regional servers. Replica manager query Replica Catalog to determine which grid site contains the desired replica (Candidate sites). 4.6.4 If the file not found in lower level its Manager send Request to upper level. 4.6.5 Determine the communication cost between requester site and candidate sites (distance<h). 4.6.6 Compute the Round Trip Time (RTT). 4.6.7 If (d>RTT) then access the file from the remote place or else replicate. 4.6.8 Check the storage element of the site selected for replication. If no storage space available invoke the Replica Replacement algorithm, otherwise 4.6.9 The Threshold Controller checks whether the site has minimum access load, if yes, it communicates with Reservation Manager. 4.6.10 If Reservation Manager succeeds in making reservations, Allocation Manager is called to allocate resources. 4.6.11 Once Allocation Manager allocated resources, Replication Placement is performed. 4.6.12 If Reservation Manager is not succeeded, then the Lowest Common Ancestor algorithm (LCA) is invoked. 4.6.13 LCA returns a site, with this site ID, repeat steps 4.6.8 and 4.6.9. 4.6.14 If the Threshold Controller results maximum access load, choose one of the sibling node and continue step 4.6.10 and 4.6.11.

321

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

4.7 Replica Replacement 4.7.1 If (target site doesn't have enough free space) Then Do Replacement 4.7.2 Replacement: Sort Files in SE using LRU and delete the files in ascending order until the space found for the new replica. 5. 5.1 PERFORMANCE EVALUATION

Simulation tool OptorSim is used as the simulator tool to evaluate the performance of our proposed algorithm. OptorSim [19] is a simulation package written in Java. It was developed to study the effectiveness of replica optimization algorithms within a Data Grid environment [20] and to represent the structure of a real European Data Grid. The simulation was constructed assuming that the Grid contains several sites; each consists of zero or more Computing Elements (CEs) and zero or more Storage Elements (SEs). CEs run jobs by processing data files, which are stored in the SEs. A Resource Broker (RB) controls scheduling of jobs to Grid Sites, and schedules jobs to CEs according to scheduling algorithm. Each site handles its file content with Replica Manager (RM), within which a Replica Optimizer (RO) contains the replication algorithm which drives automatic creation and deletion of replicas [21]. There are two types of algorithms in OptorSim: the scheduling algorithm used by the RB to schedule jobs to CEs and the replication algorithm used by RM at each site to manage replication. Each scheduling and replication algorithm is implemented as a separate Resource Broker and Replica Optimizer class respectively. We have made changes only in Replica Optimizer Class and the default Resource Broker class is used. There are three options for Replication Algorithms in OptorSim. First, one can choose No Replication which never replicates a file and all replicas are taken from the master site where the data were produced at the beginning of the simulation and the distribution of files does not change during simulation. Second, one can use LRU or LFU algorithm that always tries to replicate and, if necessary, deletes Least Recently Used files or Least Frequently Used files. Third, one can use an economic model in which algorithm only deletes files if they are less valuable than a new file. There are currently two types of the economic model: the binomial economic model, where file values are predicted by ranking the files in a binomial distribution according to their popularity in the recent past, and the Zipf economic model, where a Zipf-like distribution is used instead [16].We have compared our proposed algorithm with all of these algorithms. 5.2 Simulation Results The simulated grid used in our experiments has 20 sites, 18 of them have Storage Element (SE) and Computing Element (CE) and 2 of them have only SE. The capacity of sites 14 and 19 that only have SEs are 100 GB (all master files are stored in these two sites at the beginning of simulation) and the other ones are 70 GB and 50 GB. The SEs of Regional Server are 70 GB. Also there are 8 routers that do not have SEs and CEs. We ran six jobs totally 100 times and evaluated the impacts of file access pattern. We tested RPGTG and the other algorithms in 2 types of access pattern: 1. Random Access, 2. Random Zipf Access. The performance evaluation metrics that we used in our simulation are: Mean Job Execution Time and Effective Network Usage.

322

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

Mean Job Time of all Jobs on Grid The mean job time of all jobs on grid is defined as the combined total time in milliseconds of all the jobs run divided by the number of jobs completed. Mean Job Time of all jobs on Grid is calculated as dividing the sum of the difference in Job arrival time and Job departure time with number of jobs completed. Note that for all the components, total job time is defined as the sum of the entire individual job times, including their queuing times [22]. We have compared Mean Job Time of our proposed algorithm with other existing ones in both cases when the QoS values are larger and smaller. When the QoS values are large, the Mean Job Time of our proposed algorithm is higher when compared by having the lesser QoS values. The comparison results are shown in Fig 2 and Fig. 3. The simulation results show that RPGTG has the lowest value of Mean Job Execution Time in both Random and zipf access patterns and also in both cases of QoS values. One of the important factors that decrease the grid sites job execution time is having their required files locally stored on their storage element. It should be noted that, according to zipf access pattern, a few files are requested many times. So, as mentioned before, in our proposed architecture the physical location of sites and the files that requested by them are registered in specific databases. Additionally, the proposed algorithm selects the most popular file. By this features the proposed algorithm has the lowest value of Mean Job Execution Time in comparison with LFU, LRU, and No replication. But when Random access pattern is used, LFU and LRU have shorter mean job. As Mean Job Execution Time is the most important evaluation metric, RPGTG can be considered as the superior strategy.

5.2.1

Fig. 2 Average Mean Job Execution Time with smaller QoS values

Fig. 3 Average Mean Job Execution Time with larger QoS values

5.1.1 Effective Network Usage (ENU) This is effectively the ratio of files transferred to files requested, so a low value indicates that the optimization strategy used is better at putting files in the right places. It ranges from 0 to 1. It can be measured by using equation. ENU is calculated by dividing the sum of remote file accesses and file replications with the sum of remote file accesses and local file accesses. By this assumption the bandwidth consumption is minimized and used effectively. The No Replication strategy performs the worst because it always accesses files remotely. We have compared Effective Network Usage of our proposed algorithm with other existing ones in both cases when the QoS values are larger and smaller. When the QoS values are large, the Effective Network Usage of our proposed algorithm is higher when compared by having the lesser
323

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

QoS values. The comparison results are shown in Fig 4 and Fig. 5. The simulation results show that RPGTG has the lowest value of Effective Network Usage in both Random and zipf access patterns and also in both cases of QoS values.

Fig. 4 Average Effective Network Usage with smaller QoS values 6. CONCLUSION

Fig. 5 Average Effective Network Usage with larger QoS values

Data Replication is a mainstream technique for fast data access in Data Grid. Grid computing infrastructure has various types of resources and also the performance of these resources is discrete. So before replicating, considering the topology of the grid and to satisfy the quality of service issues are essential. In this paper, we consider QoS requirements deadlines and number of hops before placing the replica at an optimal place. Also we achieved minimal data access cost and effective network usage. This work can be extended to maintain the consistency of the replica where as this work assumes the replica as read only. 7. [1] [2] [3] REFERENCES Zeinab Fadaie, Amir Masoud Rahmani, A new Replica Placement Algorithm in Data Grid, International Journal of Computer Science, Vol 9, Issue 2, (2012), Pp 491-507 J. Zhang, B.S. Lee, X. Tang, C.K.Yeo, A Model To Predict The Optimal Performance Of The Hierarchical Data Grid, Future Generation Computer Systems 26 (2010), Pp 1-11 Y. Yuan, Y. Wu, G. Yang, F. Yu, Dynamic Data Replication Based On Local Optimization Principle In Data Grid, Grid And Cooperative Computing, (2007) GCC 2007, Sixth International Conference On. Q. Rasool, J. Li, G. S. Oreku, Sh. Zhang, D. Yang, A Load Balancing Replica Placement Strategy In Data Grid, Digital Information Management, ICDIM 2008. Third International Conference, Nov. (2008), Pp. 751-756 Y. Ding, Y. Lu, Automatic Data Placement And Replication In Grids, In: High Performance Computing, Hipc, International Conference On, (2009), Pp. 3039 Q. Rasool, J. Li, S. Zhang, Replica Placement In Multi-Tier Data Grid, In: 2009 Eighth IEEE International Conference On Dependable, Autonomic And Secure Computing, (2009), Pp. 103108.
324

[4]

[5] [6]

International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), IAEME

[7] [8]

[9] [10] [11] [12]

[13] [14] [15]

[16]

[17] [18] [19] [20]

[21]

[22] [23]

[24]

Y.F. Lin, J.J. Wu, P. Liu, A List-Based Strategy For Optimal Replica Placement In Data Grid Systems, In: 37th International Conference On Parallel Processing, (2008), Pp. 198205. P. Golda Jeyasheeli, K. Ramar, And M. Archanaa Centralized and Distributed Replica Placement Algorithms For Data Grid, European Journal Of Scientific Research, Vol. 79, No. 1, (2012), Pp. 68-81. M. Garmehi And Y. Mansouri, Optimal Placement Replication On Data Grid Environments, 10th International Conference On Information Technology, ICIT 2007. M. Tang B.S Lee, C.K. Yeo and X. Tang, Dynamic Replication Algorithms for the Multi-Tier Data Grid, Future Generation Computer Systems, Vol. 21, (2005), Pp. 775-790. Yaser Nemati, Faramarz Samsami, Mehdi Nikhkhah (2012), A Novel Data Replication Policy In Data Grid,, Australian Journal Of Basic And Applied Sciences, 6(7): 339-344. Abdullah, M. Othman, H. Ibrahim, M.N. Sulaiman, A.T. Othman, (2008) Decentralized Replication Strategies For P2P Based Scientific Data Grid, Information Technology, Itsim, International Symposium On, Pp. 18. Sashi. K, A.S. Thanamani, (2011) Dynamic Replication In A Data Grid Using A Modified BHR Region Based Algorithm Future Generation Computer Systems 27 (2), Pp. 202210. X. Tang and J. Xu. QoS-Aware Replica Placement for Content Distribution. IEEE Trans. Parallel Distributed Systems, 16(10):921-932, 2005. H. Wang, P. Liu, and J.-J. Wu, A QoS-aware Heuristic Algorithm for Replica Placement, in Proceedings of the 7th International Conference on Grid Computing, GRID2006. IEEE Computer Society, 2006, pp. 96103 D.G. Cameron, A.P. Millar, C. Nicholson, Optorsim: A Simulation Tool For Scheduling And Replica Optimization In Data Grids, Proc. Computing In High Energy And Nuclear Physics (CHEP), (2004). Dogan, A Study On Performance Of Dynamic File Replication Algorithms For Realtime File Access In Data Grids Future Generation Computer Systems, September (2009), Pp. 829-839. M. Shorfuzzaman, P. Graham, R. Eskicioglu, Adaptive Replica Placement In Hierarchical Data Grids, Journal Of Physics: Conference Series 256 (2010). Optorsim A Replica Optimiser Simulation, Http:// Grid-Datamanagement.Web.Cern.Ch/Grid-Data Management/ Optimization/Optor W.H. Bell, D.G. Cameron, L. Capozza, A.P.Millar, K. Stockinger, F. Zini, Simulation Of Dynamic Grid Replication Strategies In Optorsim, Int. Journal Of High Performance Computing Applications, 17(4), (2003). R. M. Vozmediano, Application Layer Multicast For Efficient Grid File Transfer, International Journal Of Computer Science And Applications, Technomathematics Research Foundation, (2009), Pp.70-84. D.G. Cameron, R. Schiaffino, J. Ferguson, A.P. Millar, C. Nicholson, K. Stockinger, F. Zini, Optorsim V2.1 Installation And User Guide, October (2006). M. Pushpalatha, T. Ramarao, Revathi Venkataraman and Sorna Lakshmi, Mobility Aware Data Replication using Minimum Dominating Set in Mobile Ad Hoc Networks, International Journal of Computer Engineering & Technology (IJCET), Volume 3, Issue 2, 2011, pp. 645 - 658, ISSN Print: 0976 6367, ISSN Online: 0976 6375. Priya Deshpande, Brijesh Khundhawala and Prasanna Joeg, Dynamic Data Replication and Job Scheduling Based on Popularity and Category, International Journal of Computer Engineering & Technology (IJCET), Volume 4, Issue 5, 2013, pp. 109 - 114, ISSN Print: 0976 6367, ISSN Online: 0976 6375.

325

You might also like