The conventional procedures of clustering algorithms are incapable of overcoming the difficulty of managing and analyzing the rapid growth of generated data from different sources. Using the concept of parallel clustering is one of the robust solutions to this problem. Apache Hadoop architecture is one of the assortment ecosystems that provide the capability to store and process the data in a distributed and parallel fashion. In this paper, a parallel model is designed to process the k-means clustering algorithm in the Apache Hadoop ecosystem by connecting three nodes, one is for server (name) nodes and the other two are for clients (data) nodes. The aim is to speed up the time of managing the massive scale of healthcare insurance dataset with the size of 11 GB and also using machine learning algorithms, which are provided by the Mahout Framework. The experimental results depict that the proposed model can efficiently process large datasets. The parallel k-means algorithm outperforms the sequential k-means algorithm based on the execution time of the algorithm, where the required time to execute a data size of 11 GB is around 1.847 hours using the parallel k-means algorithm, while it equals 68.567 hours using the sequential k-means algorithm. As a result, we deduce that when the nodes number in the parallel system increases, the computation time of the proposed algorithm decreases.
The vast majority of EC applications are the web-based deployed in 3-tire Server-Client environment, the data within such application often resides within several heterogeneous data sources. Building a single application that can access each data sources can be a matter of challenging; this paper concerns with developing a software program that runs transparently against heterogeneous environment for an EC-application.
In this paper, we employ the maximum likelihood estimator in addition to the shrinkage estimation procedure to estimate the system reliability (
So far, APT (Advanced Persistent Threats) is a constant concern for information security. Despite that, many approaches have been used in order to detect APT attacks, such as change controlling, sandboxing and network traffic analysis. However, success of 100% couldn’t be achieved. Current studies have illustrated that APTs adopt many complex techniques to evade all detection types. This paper describes and analyzes APT problems by analyzing the most common techniques, tools and pathways used by attackers. In addition, it highlights the weaknesses and strengths of the existing security solutions that have been used since the threat was identified in 2006 until 2019. Furthermore, this research proposes a new framework that can be u
... Show MoreSeveral remote sensor network (WSN) tasks require sensor information join. This in-processing Join is configured in parallel sensor hub to save battery power and limit the communication cost. Hence, a parallel join system is proposed for sensor networks. The proposed parallel join algorithm organizes in section-situated databases. A novel join method has been proposed for remote WSNs to limit the aggregate communication cost and enhance execution. This approach depends on two procedures; section-situated databases and parallel join algorithm utilized to store sensor information and speed up processing respectively. A segment arranged databases store information table in segmented shrewd. The Parallel-Joining WSN algorithm is effectively
... Show MoreCrime is considered as an unlawful activity of all kinds and it is punished by law. Crimes have an impact on a society's quality of life and economic development. With a large rise in crime globally, there is a necessity to analyze crime data to bring down the rate of crime. This encourages the police and people to occupy the required measures and more effectively restricting the crimes. The purpose of this research is to develop predictive models that can aid in crime pattern analysis and thus support the Boston department's crime prevention efforts. The geographical location factor has been adopted in our model, and this is due to its being an influential factor in several situations, whether it is traveling to a specific area or livin
... Show MoreWith the development of high-speed network technologies, there has been a recent rise in the transfer of significant amounts of sensitive data across the Internet and other open channels. The data will be encrypted using the same key for both Triple Data Encryption Standard (TDES) and Advanced Encryption Standard (AES), with block cipher modes called cipher Block Chaining (CBC) and Electronic CodeBook (ECB). Block ciphers are often used for secure data storage in fixed hard drives, portable devices, and safe network data transport. Therefore, to assess the security of the encryption method, it is necessary to become familiar with and evaluate the algorithms of cryptographic systems. Block cipher users need to be sure that the ciphers the
... Show MoreThis paper proposes a completion that can allow fracturing four zones in a single trip in the well called “Y” (for confidential reasons) of the field named “X” (for confidential reasons). The steps to design a well completion for multiple fracturing are first to select the best completion method then the required equipment and the materials that it is made of. After that, the completion schematic must be drawn by using Power Draw in this case, and the summary installation procedures explained. The data used to design the completion are the well trajectory, the reservoir data (including temperature, pressure and fluid properties), the production and injection strategy. The results suggest that multi-stage hydraulic fracturing can
... Show MoreHuman detection represents a main problem of interest when using video based monitoring. In this paper, artificial neural networks, namely multilayer perceptron (MLP) and radial basis function (RBF) are used to detect humans among different objects in a sequence of frames (images) using classification approach. The classification used is based on the shape of the object instead of depending on the contents of the frame. Initially, background subtraction is depended to extract objects of interest from the frame, then statistical and geometric information are obtained from vertical and horizontal projections of the objects that are detected to stand for the shape of the object. Next to this step, two ty
... Show MoreTransportation and distribution are the most important elements in the work system for any company, which are of great importance in the success of the chain work. Al-Rabee factory is one of the largest ice cream factories in Iraq and it is considered one of the most productive and diversified factories with products where its products cover most areas of the capital Baghdad, however, it lacks a distribution system based on scientific and mathematical methods to work in the transportation and distribution processes, moreover, these processes need a set of important data that cannot in any way be separated from the reality of fuzziness industrial environment in Iraq, which led to use the fuzzy sets theory to reduce the levels of uncertainty.
... Show More