Replication is a very useful technique in distributed systems, grid community and clustering systems. Now a days a lot of algorithm development has been focused for safe replication services. Replication can improve performance,... more
Replication is a very useful technique in distributed systems, grid community and clustering systems. Now a days a lot of algorithm development has been focused for safe replication services. Replication can improve performance, reliability, portability of entire database. In this paper, a persistent layer has been developed and proposed which supports heterogeneous system. The persistent layer work on asynchronous model, hence it is known as asynchronous replication. The implementation of this algorithm has built on Java based technology thus it is very easy to deploy any OS without hassle and configurable files makes the whole system easy to maintain and cost effective. The experimental servers (both main server and replication server) used both windows and Linux OS’s. Finally some experiments have been carried out by different data taken from references. The results show that, the proposed model outperforms all available replication models.
Software driven solutions are limited to the amount of memory size and storage capacity, but the sizes of databases are increasing every day. Hence, now a day, handling data and accessing it in an acceptable time is one of the biggest... more
Software driven solutions are limited to the amount of memory size and storage capacity, but the sizes of databases are increasing every day. Hence, now a day, handling data and accessing it in an acceptable time is one of the biggest challenges especially in a large database system. In a database, the records can be categorized according to the access frequencies; some records are very frequently accessed (hot data), some records are hardly accessed (cold data) and other records accessed occasionally (warm data). In a conventional database we keep all hot, warm and cold records in a single database. In case of record access (query, update etc.) a query might takes longer time even if a good data accessing algorithm (clustering/mining) incorporate with the database. Thus categorizing of the data set, i. e. clustering in terms of access frequency may improve data accessibility. In this paper, we are proposing a data clustering mechanism based on data access frequency. Finally, the expected result shows how and why data accessibility time should outperform other available data clustering techniques.
The frequent use of web based application plays a vital role in our everyday life. MVC (Model View Controller) architecture is used as programmed architectural pattern in order to implement user interfaces. Application software developers... more
The frequent use of web based application plays a vital role in our everyday life. MVC (Model View Controller) architecture is used as programmed architectural pattern in order to implement user interfaces. Application software developers utilize MVC (Model View Controller) Architecture for developing web based application. The sizes of databases are increasing day by day in relation with time. Therefore, if we take into account the concept of huge centralized database systems, it has become one of the most challenging criterions for accessing data in acceptable time. Basically, in centralized databases, the records can be classified into two categories considering the access frequency of data. Those records that are being accessed frequently are known as Level 1 data, on the contrary, those accessed in lesser frequency is considered as Level 2 Data. In this paper, we will try to enhance and optimize the performance of MVC architecture based on two parameters namely response time and throughput. The response time and throughput is improved based on the proposed database search algorithm using B+ tree. If the database search engine is idle, the database search engine will look forward to discover whether the intended data is in level 1, otherwise it will search for level 2 data. The level 2 data will be included as level 1 data inside the database or vice versa, for insertion and update operation. However, whether the data is level 1 or level 2 data will be depended upon user choice. Thus, the overall response time as well as throughput will be optimally increased.
The sizes of databases are increasing every day. Hence, now days, accessing data in an acceptable time is one of the biggest challenges in centralized database. In centralized databases, the records can be categorized according to the... more
The sizes of databases are increasing every day. Hence, now days, accessing data in an acceptable time is one of the biggest challenges in centralized database. In centralized databases, the records can be categorized according to the access frequencies; least accessed records (cold data) and most accessed records (hot data). In a study it shows that more than 90% cases query are requested for hot data, and in case of insertion operation, 99% are done on hot data. Thus categorizing of the data set may improve data accessibility. In this paper, we are proposing a data clustering mechanism based on data access frequency. We have considered only the hot data and the cold data. Here we divided the whole database into two separate files. The first file contains only hot data and the second file contains only the cold data. The time period of hot and cold data will vary for different application domains. The database engine will have direct access on the first database file and in case of unavailability of data; the database engine will look for the second database file. Finally, the experiment result shows how and why data accessibility time should outperform than other available data clustering techniques.
The regular utilization of web-based applications is crucial in our everyday life. The Model View Controller (MVC) architecture serves as a structured programming design that developers utilize to create user interfaces. This pattern is... more
The regular utilization of web-based applications is crucial in our everyday life. The Model View Controller (MVC) architecture serves as a structured programming design that developers utilize to create user interfaces. This pattern is commonly applied by application software developers to construct web-based applications. The use of a MVC framework of PHP Scripting language is often essential for application software development. There is a significant argument regarding the most suitable PHP MVC such as Codeigniter & Laravel and Phalcon frameworks since not all frameworks cater to everyone's needs. It's a fact that not all MVC frameworks are created equal and different frameworks can be combined for specific scenarios. Selecting the appropriate MVC framework can pose a challenge at times. In this context, our paper focuses on conducting a comparative analysis of different PHP frameworks. The widely used PHP MVC frameworks are picked to compare the performance on basic Operation of Relational databases and different type of Application software to calculate execution time. In this experiment a large (Big Data) dataset was used.
Background: COVID-19 (Coronavirus) disease is one of the most critical human diseases in the contemporary world and influences human life exceptionally severely. The Coronavirus widespread caused by the novel SARS-CoV-2 has significantly... more
Background: COVID-19 (Coronavirus) disease is one of the most critical human diseases in the contemporary world and influences human life exceptionally severely. The Coronavirus widespread caused by the novel SARS-CoV-2 has significantly affected human health, social and the economy especially in countries struggling with financial assets for medical testing and treatment. The statistically data analysis (hematological characteristics) of CBC Datasets in COVID-19 positive and COVID-19 negative patients. Objective: Data analysis of CBC test for explore the hematological characteristics. Methodology: One Hundred sixteen CBC test results were collected from the department of pathology in National institute of cardiovascular diseases & hospital (NICVDH),
Binary particle swarm optimization Transfer function The 0–1 knapsack problem The truss optimization problem a b s t r a c t Many real-world problems belong to the family of discrete optimization problems. Most of these problems are... more
Binary particle swarm optimization Transfer function The 0–1 knapsack problem The truss optimization problem a b s t r a c t Many real-world problems belong to the family of discrete optimization problems. Most of these problems are NP-hard and difficult to solve efficiently using classical linear and convex optimization methods. In addition, the computational difficulties of these optimization tasks increase rapidly with the increasing number of decision variables. A further difficulty can be also caused by the search space being intrinsi-cally multimodal and non-convex. In such a case, it is more desirable to have an effective optimization method that can cope better with these problem characteristics. Binary particle swarm optimization (BPSO) is a simple and effective discrete optimization method. The original BPSO and its variants have been used to solve a number of classic discrete optimization problems. However, it is reported that the original BPSO and its variants are unable to provide satisfactory results due to the use of inappropriate transfer functions. More specifically, these transfer functions are unable to provide BPSO a good balance between exploration and exploitation in the search space, limiting their performances. To overcome this problem, this paper proposes to employ a time-varying transfer function in the BPSO, namely TV T-BPSO. To understand the search behaviour of the TV T-BPSO, we provide a systematic analysis of its exploration and exploitation capability. Our experimental results demonstrate that TV T-BPSO outperforms existing BPSO variants on both low-dimensional and high-dimensional classical 0–1 knapsack problems, as well as a 200-member truss problem, suggesting that TV T-BPSO is able to better scale to high dimensional combinatorial problems than the existing BPSO variants and other metaheuristic algorithms.
Blind spots (or bad sampling points) in indoor areas are the positions where no signal exists (or the signal is too weak) and the existence of a receiver within the blind spot decelerates the performance of the communication system.... more
Blind spots (or bad sampling points) in indoor areas are the positions where no signal exists (or the signal is too weak) and the existence of a receiver within the blind spot decelerates the performance of the communication system. Therefore, it is one of the fundamental requirements to eliminate the blind spots from the indoor area and obtain the maximum coverage while designing the wireless networks. In this regard, this paper combines ray-tracing (RT), genetic algorithm (GA), depth first search (DFS), and branch-and-bound method as a new technique that guarantees the removal of blind spots and subsequently determines the optimal wireless coverage using minimum number of transmitters. The proposed system outperforms the existing techniques in terms of algorithmic complexity and demonstrates that the computation time can be reduced as high as 99% and 75%, respectively, as compared to existing algorithms. Moreover, in terms of experimental analysis, the coverage prediction successf...
The advent of technology with the increasing use of wireless network has led to the development of Wireless Body Area Network (WBAN) to continuously monitor the change of physiological data in a cost efficient manner. As numerous... more
The advent of technology with the increasing use of wireless network has led to the development of Wireless Body Area Network (WBAN) to continuously monitor the change of physiological data in a cost efficient manner. As numerous researches on wave propagation characterization have been done in intrabody communication, this study has given emphasis on the wave propagation characterization between the control units (CUs) and wireless access point (AP) in a hospital scenario. Ray tracing is a tool to predict the rays to characterize the wave propagation. It takes huge simulation time, especially when multiple transmitters are involved to transmit physiological data in a realistic hospital environment. Therefore, this study has developed an accelerated ray tracing method based on the nearest neighbor cell and prior knowledge of intersection techniques. Beside this, Red-Black tree is used to store and provide a faster retrieval mechanism of objects in the hospital environment. To prove ...
The amount of data has been increasing exponentially in every sector such as banking securities, healthcare, education, manufacturing, consumer-trade, transportation, and energy. Most of these data are noise, different in shapes, and... more
The amount of data has been increasing exponentially in every sector such as banking securities, healthcare, education, manufacturing, consumer-trade, transportation, and energy. Most of these data are noise, different in shapes, and outliers. In such cases, it is challenging to find the desired data clusters using conventional clustering algorithms. DBSCAN is a popular clustering algorithm which is widely used for noisy, arbitrary shape, and outlier data. However, its performance highly depends on the proper selection of cluster radius (Eps) and the minimum number of points (MinPts) that are required for forming clusters for the given dataset. In the case of real-world clustering problems, it is a difficult task to select the exact value of Eps and (MinPts) to perform the clustering on unknown datasets. To address these, this paper proposes a dynamic DBSCAN algorithm that calculates the suitable value for (Eps) and (MinPts) dynamically by which the clustering quality of the given p...
This paper analyzed the new compact design and negative refractive index (NRI) metamaterial for wideband applications. The proposed metamaterial exhibits NRI and wideband characteristics of the x-axis wave propagation. It displayed the... more
This paper analyzed the new compact design and negative refractive index (NRI) metamaterial for wideband applications. The proposed metamaterial exhibits NRI and wideband characteristics of the x-axis wave propagation. It displayed the NRI property at the frequency of 1.54 GHz and wideband from 1.26 GHz to 7.08 GHz frequency (L, S, and C band). Moreover, the response of the 1×2 horizontal and 2×1 vertical array structure showed the wideband frequency in the 7.17 GHz to 13.62 GHz and 1.46 GHz to 9.53 GHz, respectively. Electromagnetic simulation software called CST has been used to design the metamaterial unit cell. The metamaterial has been displayed the multi-band characteristics such as L, S, C, X and K u bands with negative index material properties.
Due to the attractive features of millimeter band, its uses are greatly expanding in the indoor wireless communication systems. As the distance between the transmitter and receiver is much shorter in indoor environments than that of the... more
Due to the attractive features of millimeter band, its uses are greatly expanding in the indoor wireless communication systems. As the distance between the transmitter and receiver is much shorter in indoor environments than that of the outdoor environment, the radio wave paths of the millimeter band frequencies are highly influenced by the building materials as well as by the human movements. Ray tracing is widely used method to characterize the radio wave propagation for the planning and establishment of the indoor wireless network. Accurate object modeling for the realistic environment and computational time are two classical problems of the ray tracing model. Because, large number of rays that travels in a complex and convoluted indoor environment must be traced. Therefore, an accurate and efficient ray tracing method is proposed here, which is based on the surface separation, object address distribution and surface skipping techniques. The proposed approach considers the effect...