The region-based association analysis has been proposed to capture the collective behavior of sets of variants by testing the association of each set instead of individual variants with the disease. Such an analysis typically involves a list of unphased multiple-locus genotypes with potentially sparse frequencies in cases and controls. To tackle the problem of the sparse distribution, a two-stage approach was proposed in literature: In the first stage, haplotypes are computationally inferred from genotypes, followed by a haplotype coclassification. In the second stage, the association analysis is performed on the inferred haplotype groups. If a haplotype is unevenly distributed between the case and control samples, this haplotype is labeled as a risk haplotype. Unfortunately, the in-silico reconstruction of haplotypes might produce a proportion of false haplotypes which hamper the detection of rare but true haplotypes. Here, to address the issue, we propose an alternative approach: In Stage 1, we cluster genotypes instead of inferred haplotypes and estimate the risk genotypes based on a finite mixture model. In Stage 2, we infer risk haplotypes from risk genotypes inferred from the previous stage. To estimate the finite mixture model, we propose an EM algorithm with a novel data partition-based initialization. The performance of the proposed procedure is assessed by simulation studies and a real data analysis. Compared to the existing multiple Z-test procedure, we find that the power of genome-wide association studies can be increased by using the proposed procedure.
In most manufacturing processes, and in spite of statistical control, several process capability indices refer to non conformance of the true mean (µc ) from the target mean ( µT ), and the variation is also high. In this paper, data have been analyzed and studied for a blow molded plastic product (Zahi Bottle) (ZB). WinQSB software was used to facilitate the statistical process control, and process capability analysis and some of capability indices. The relationship between different process capability indices and the true mean of the process were represented, and then with the standard deviation (σ ), of achievement of process capability value that can reduce the standard deviation value and improve production out of theoretical con
... Show MoreString matching is seen as one of the essential problems in computer science. A variety of computer applications provide the string matching service for their end users. The remarkable boost in the number of data that is created and kept by modern computational devices influences researchers to obtain even more powerful methods for coping with this problem. In this research, the Quick Search string matching algorithm are adopted to be implemented under the multi-core environment using OpenMP directive which can be employed to reduce the overall execution time of the program. English text, Proteins and DNA data types are utilized to examine the effect of parallelization and implementation of Quick Search string matching algorithm on multi-co
... Show MoreThe performance quality and searching speed of Block Matching (BM) algorithm are affected by shapes and sizes of the search patterns used in the algorithm. In this paper, Kite Cross Hexagonal Search (KCHS) is proposed. This algorithm uses different search patterns (kite, cross, and hexagonal) to search for the best Motion Vector (MV). In first step, KCHS uses cross search pattern. In second step, it uses one of kite search patterns (up, down, left, or right depending on the first step). In subsequent steps, it uses large/small Hexagonal Search (HS) patterns. This new algorithm is compared with several known fast block matching algorithms. Comparisons are based on search points and Peak Signal to Noise Ratio (PSNR). According to resul
... Show MoreMixture experiments are response variables based on the proportions of component for this mixture. In our research we will compare the scheffʼe model with the kronecker model for the mixture experiments, especially when the experimental area is restricted.
Because of the experience of the mixture of high correlation problem and the problem of multicollinearity between the explanatory variables, which has an effect on the calculation of the Fisher information matrix of the regression model.
to estimate the parameters of the mixture model, we used the (generalized inverse ) And the Stepwise Regression procedure
... Show MoreWe studied in this research how to find a method of estimating the quantity (Kinetically) of three kinds of Insecticide and their mixture, which are used in agriculture. The extracted insecticide from the polluted samples with these insect from air, soil, and the leaves of trees, have be used into the reaction with H2O2 and benzedine. The kinetic study of this reaction was formed in basic medium,( pH= 8.6), using UV. Spectra at (?= 420nm). The study showed that the reaction is the first order, and the speed of the reaction was used to estimate the concentration of insecticide in solution and mixture. The experiments of this study indicated that this method has the speed and efficiency for quantitatively estimating these
... Show MoreThe study investigated the behaviour of asphalt concrete mixes for aggregate gradations, according to the Iraqi specification using the Bailey method designed by an Excel spreadsheet. In mixing aggregates with varying gradations (coarse and fine aggregate), The Bailey method is a systematic methodology that offers aggregate interlocking as the backbone of the framework and a controlled gradation to complete the blends. Six types of gradation are used according to the bailey method considered in this study. Two-course prepared Asphalt Concrete Wearing and Asphalt Concrete binder, the Nominal Maximum Aggregate Sizes (NMAS) of the mixtures are 19 and 12.5 mm, respectively. The total number of specimens was 240 for both layers (15 samp
... Show MoreBecause of the experience of the mixture problem of high correlation and the existence of linear MultiCollinearity between the explanatory variables, because of the constraint of the unit and the interactions between them in the model, which increases the existence of links between the explanatory variables and this is illustrated by the variance inflation vector (VIF), L-Pseudo component to reduce the bond between the components of the mixture.
To estimate the parameters of the mixture model, we used in our research the use of methods that increase bias and reduce variance, such as the Ridge Regression Method and the Least Absolute Shrinkage and Selection Operator (LASSO) method a
... Show MoreIn this study, we made a comparison between LASSO & SCAD methods, which are two special methods for dealing with models in partial quantile regression. (Nadaraya & Watson Kernel) was used to estimate the non-parametric part ;in addition, the rule of thumb method was used to estimate the smoothing bandwidth (h). Penalty methods proved to be efficient in estimating the regression coefficients, but the SCAD method according to the mean squared error criterion (MSE) was the best after estimating the missing data using the mean imputation method
 
        