The region-based association analysis has been proposed to capture the collective behavior of sets of variants by testing the association of each set instead of individual variants with the disease. Such an analysis typically involves a list of unphased multiple-locus genotypes with potentially sparse frequencies in cases and controls. To tackle the problem of the sparse distribution, a two-stage approach was proposed in literature: In the first stage, haplotypes are computationally inferred from genotypes, followed by a haplotype coclassification. In the second stage, the association analysis is performed on the inferred haplotype groups. If a haplotype is unevenly distributed between the case and control samples, this haplotype is labeled as a risk haplotype. Unfortunately, the in-silico reconstruction of haplotypes might produce a proportion of false haplotypes which hamper the detection of rare but true haplotypes. Here, to address the issue, we propose an alternative approach: In Stage 1, we cluster genotypes instead of inferred haplotypes and estimate the risk genotypes based on a finite mixture model. In Stage 2, we infer risk haplotypes from risk genotypes inferred from the previous stage. To estimate the finite mixture model, we propose an EM algorithm with a novel data partition-based initialization. The performance of the proposed procedure is assessed by simulation studies and a real data analysis. Compared to the existing multiple Z-test procedure, we find that the power of genome-wide association studies can be increased by using the proposed procedure.
The purpose of this work is to concurrently estimate the UVvisible spectra of binary combinations of piroxicam and mefenamic acid using the chemometric approach. To create the model, spectral data from 73 samples (with wavelengths between 200 and 400 nm) were employed. A two-layer artificial neural network model was created, with two neurons in the output layer and fourteen neurons in the hidden layer. The model was trained to simulate the concentrations and spectra of piroxicam and mefenamic acid. For piroxicam and mefenamic acid, respectively, the Levenberg-Marquardt algorithm with feed-forward back-propagation learning produced root mean square errors of prediction of 0.1679 μg/mL and 0.1154 μg/mL, with coefficients of determination of
... Show MoreIn this article, it is interesting to estimate and derive the three parameters which contain two scales parameters and one shape parameter of a new mixture distribution for the singly type one censored data which is the branch of right censored sample. Then to define some special mathematical and statistical properties for this new mixture distribution which is considered one of the continuous distributions characterized by its flexibility. Next, using maximum likelihood estimator method for singly type one censored data based on the Newton-Raphson matrix procedure to find and estimate values of these three parameter by utilizing the real data taken from the National Center for Research and Treatment of Hematology/University of Mus
... Show MoreIn this study, we made a comparison between LASSO & SCAD methods, which are two special methods for dealing with models in partial quantile regression. (Nadaraya & Watson Kernel) was used to estimate the non-parametric part ;in addition, the rule of thumb method was used to estimate the smoothing bandwidth (h). Penalty methods proved to be efficient in estimating the regression coefficients, but the SCAD method according to the mean squared error criterion (MSE) was the best after estimating the missing data using the mean imputation method
The Estimation Of The Reliability Function Depends On The Accuracy Of The Data Used To Estimate The Parameters Of The Probability distribution, and Because Some Data Suffer from a Skew in their Data to Estimate the Parameters and Calculate the Reliability Function in light of the Presence of Some Skew in the Data, there must be a Distribution that has flexibility in dealing with that Data. As in the data of Diyala Company for Electrical Industries, as it was observed that there was a positive twisting in the data collected from the Power and Machinery Department, which required distribution that deals with those data and searches for methods that accommodate this problem and lead to accurate estimates of the reliability function,
... Show MoreBubbled slabs can be exposed to damage or deterioration during its life. Therefore, the solution for strengthening must be provided. For the simulation of this case, the analysis of finite elements was carried out using ABAQUS 2017 software on six simply supported specimens, during which five are voided with 88 bubbles, and the other is solid. The slab specimens with symmetric boundary conditions were of dimensions 3200/570/150 mm. The solid slab and one bubbled slab are deemed references. Each of the other slabs was exposed to; (1) service charge, then unloaded (2) external prestressing and (3) loading to collapse under two line load. The external strengthening was applied using prestressed wire with four approaches, wh
... Show MoreThis research presents a method of using MATLAB in analyzing a nonhomogeneous soil (Gibson-type) by
estimating the displacements and stresses under the strip footing during applied incremental loading
sequences. This paper presents a two-dimensional finite element method. In this method, the soil is divided into a number of triangle elements. A model soil (Gibson-type) with linearly increasing modulus of elasticity with depth is presented. The influences of modulus of elasticity, incremental loading, width of footing, and depth of footing are considered in this paper. The results are compared with authors' conclusions of previous studies.
The performance quality and searching speed of Block Matching (BM) algorithm are affected by shapes and sizes of the search patterns used in the algorithm. In this paper, Kite Cross Hexagonal Search (KCHS) is proposed. This algorithm uses different search patterns (kite, cross, and hexagonal) to search for the best Motion Vector (MV). In first step, KCHS uses cross search pattern. In second step, it uses one of kite search patterns (up, down, left, or right depending on the first step). In subsequent steps, it uses large/small Hexagonal Search (HS) patterns. This new algorithm is compared with several known fast block matching algorithms. Comparisons are based on search points and Peak Signal to Noise Ratio (PSNR). According to resul
... Show MoreString matching is seen as one of the essential problems in computer science. A variety of computer applications provide the string matching service for their end users. The remarkable boost in the number of data that is created and kept by modern computational devices influences researchers to obtain even more powerful methods for coping with this problem. In this research, the Quick Search string matching algorithm are adopted to be implemented under the multi-core environment using OpenMP directive which can be employed to reduce the overall execution time of the program. English text, Proteins and DNA data types are utilized to examine the effect of parallelization and implementation of Quick Search string matching algorithm on multi-co
... Show MorePoverty phenomenon is very substantial topic that determines the future of societies and governments and the way that they deals with education, health and economy. Sometimes poverty takes multidimensional trends through education and health. The research aims at studying multidimensional poverty in Iraq by using panelized regression methods, to analyze Big Data sets from demographical surveys collected by the Central Statistical Organization in Iraq. We choose classical penalized regression method represented by The Ridge Regression, Moreover; we choose another penalized method which is the Smooth Integration of Counting and Absolute Deviation (SICA) to analyze Big Data sets related to the different poverty forms in Iraq. Euclidian Distanc
... Show More