The region-based association analysis has been proposed to capture the collective behavior of sets of variants by testing the association of each set instead of individual variants with the disease. Such an analysis typically involves a list of unphased multiple-locus genotypes with potentially sparse frequencies in cases and controls. To tackle the problem of the sparse distribution, a two-stage approach was proposed in literature: In the first stage, haplotypes are computationally inferred from genotypes, followed by a haplotype coclassification. In the second stage, the association analysis is performed on the inferred haplotype groups. If a haplotype is unevenly distributed between the case and control samples, this haplotype is labeled as a risk haplotype. Unfortunately, the in-silico reconstruction of haplotypes might produce a proportion of false haplotypes which hamper the detection of rare but true haplotypes. Here, to address the issue, we propose an alternative approach: In Stage 1, we cluster genotypes instead of inferred haplotypes and estimate the risk genotypes based on a finite mixture model. In Stage 2, we infer risk haplotypes from risk genotypes inferred from the previous stage. To estimate the finite mixture model, we propose an EM algorithm with a novel data partition-based initialization. The performance of the proposed procedure is assessed by simulation studies and a real data analysis. Compared to the existing multiple Z-test procedure, we find that the power of genome-wide association studies can be increased by using the proposed procedure.
Within the framework of big data, energy issues are highly significant. Despite the significance of energy, theoretical studies focusing primarily on the issue of energy within big data analytics in relation to computational intelligent algorithms are scarce. The purpose of this study is to explore the theoretical aspects of energy issues in big data analytics in relation to computational intelligent algorithms since this is critical in exploring the emperica aspects of big data. In this chapter, we present a theoretical study of energy issues related to applications of computational intelligent algorithms in big data analytics. This work highlights that big data analytics using computational intelligent algorithms generates a very high amo
... Show MoreThe research aims to demonstrate the dual use of analysis to predict financial failure according to the Altman model and stress tests to achieve integration in banking risk management. On the bank’s ability to withstand crises, especially in light of its low rating according to the Altman model, and the possibility of its failure in the future, thus proving or denying the research hypothesis, the research reached a set of conclusions, the most important of which (the bank, according to the Altman model, is threatened with failure in the near future, as it is located within the red zone according to the model’s description, and will incur losses if it is exposed to crises in the future according to the analysis of stress tests
... Show MoreThe research was performed in order to investigate the prevalence of Eimeria spp in buffalo. Coccidiosis, is a common livestock disease include water buffaloes and nothing is known about the most pathogenic species of Eimeria. Since the highest prevalence of oocyst shedding and incidence of disease occurs in buffalo calves less than one year of age. The omnipresent occurrence and negative effects of the infection on health and buffalo growth output are taken into account. Therefore, both farmers and veterinarians should pay greater attention to infections with Eimeria spp. And there is little analysis of data reported in Iraq and the world regarding Eimeria infection in river buffalo spp.
Wireless sensor networks (WSNs) represent one of the key technologies in internet of things (IoTs) networks. Since WSNs have finite energy sources, there is ongoing research work to develop new strategies for minimizing power consumption or enhancing traditional techniques. In this paper, a novel Gaussian mixture models (GMMs) algorithm is proposed for mobile wireless sensor networks (MWSNs) for energy saving. Performance evaluation of the clustering process with the GMM algorithm shows a remarkable energy saving in the network of up to 92%. In addition, a comparison with another clustering strategy that uses the K-means algorithm has been made, and the developed method has outperformed K-means with superior performance, saving ener
... Show MoreThis research presents a study of using an additive for the objective of increasing the setting time of a material used in several aspects in the constructional field, this material is “Local-Gypsum” which is locally called “Joss”, and the additive used in this study is “Trees Glue Powder” denoted by “TGP”. Nine mixtures of Local-gypsum (joss) had been experimented in the current study to find their setting time, these mixes were divided into three groups according to their water-joss ratios (W/J) (0.3, 0.4 and 0.5), and each group was sub-divided into three sub-groups according to their TGP contents (0.0%, 0.3% and 0.6%). It was found that, when TGP is added with the
Ti6Al4V alloy is widely used in aerospace and medical applications. It is classified as a difficult to machine material due to its low thermal conductivity and high chemical reactivity. In this study, hybrid intelligent models have been developed to predict surface roughness when end milling Ti6Al4V alloy with a Physical Vapor Deposition PVD coated tool under dry cutting conditions. Back propagation neural network (BPNN) has been hybridized with two heuristic optimization techniques, namely: gravitational search algorithm (GSA) and genetic algorithm (GA). Taguchi method was used with an L27 orthogonal array to generate 27 experiment runs. Design expert software was used to do analysis of variances (ANOVA). The experimental data were
... Show MoreIn many scientific fields, Bayesian models are commonly used in recent research. This research presents a new Bayesian model for estimating parameters and forecasting using the Gibbs sampler algorithm. Posterior distributions are generated using the inverse gamma distribution and the multivariate normal distribution as prior distributions. The new method was used to investigate and summaries Bayesian statistics' posterior distribution. The theory and derivation of the posterior distribution are explained in detail in this paper. The proposed approach is applied to three simulation datasets of 100, 300, and 500 sample sizes. Also, the procedure was extended to the real dataset called the rock intensity dataset. The actual dataset is collecte
... Show MoreLost circulation or losses in drilling fluid is one of the most important problems in the oil and gas industry, and it appeared at the beginning of this industry, which caused many problems during the drilling process, which may lead to closing the well and stopping the drilling process. The drilling muds are relatively expensive, especially the muds that contain oil-based mud or that contain special additives, so it is not economically beneficial to waste and lose these muds. The treatment of drilling fluid losses is also somewhat expensive as a result of the wasted time that it caused, as well as the high cost of materials used in the treatment such as heavy materials, cement, and others. The best way to deal with drilling fluid losses
... Show More