The region-based association analysis has been proposed to capture the collective behavior of sets of variants by testing the association of each set instead of individual variants with the disease. Such an analysis typically involves a list of unphased multiple-locus genotypes with potentially sparse frequencies in cases and controls. To tackle the problem of the sparse distribution, a two-stage approach was proposed in literature: In the first stage, haplotypes are computationally inferred from genotypes, followed by a haplotype coclassification. In the second stage, the association analysis is performed on the inferred haplotype groups. If a haplotype is unevenly distributed between the case and control samples, this haplotype is labeled as a risk haplotype. Unfortunately, the in-silico reconstruction of haplotypes might produce a proportion of false haplotypes which hamper the detection of rare but true haplotypes. Here, to address the issue, we propose an alternative approach: In Stage 1, we cluster genotypes instead of inferred haplotypes and estimate the risk genotypes based on a finite mixture model. In Stage 2, we infer risk haplotypes from risk genotypes inferred from the previous stage. To estimate the finite mixture model, we propose an EM algorithm with a novel data partition-based initialization. The performance of the proposed procedure is assessed by simulation studies and a real data analysis. Compared to the existing multiple Z-test procedure, we find that the power of genome-wide association studies can be increased by using the proposed procedure.
A hand gesture recognition system provides a robust and innovative solution to nonverbal communication through human–computer interaction. Deep learning models have excellent potential for usage in recognition applications. To overcome related issues, most previous studies have proposed new model architectures or have fine-tuned pre-trained models. Furthermore, these studies relied on one standard dataset for both training and testing. Thus, the accuracy of these studies is reasonable. Unlike these works, the current study investigates two deep learning models with intermediate layers to recognize static hand gesture images. Both models were tested on different datasets, adjusted to suit the dataset, and then trained under different m
... Show MoreThe issue of image captioning, which comprises automatic text generation to understand an image’s visual information, has become feasible with the developments in object recognition and image classification. Deep learning has received much interest from the scientific community and can be very useful in real-world applications. The proposed image captioning approach involves the use of Convolution Neural Network (CNN) pre-trained models combined with Long Short Term Memory (LSTM) to generate image captions. The process includes two stages. The first stage entails training the CNN-LSTM models using baseline hyper-parameters and the second stage encompasses training CNN-LSTM models by optimizing and adjusting the hyper-parameters of
... Show MoreIn this paper we introduce a brief review about Box-Jenkins models. The acronym ARIMA stands for “autoregressive integrated moving averageâ€. It is a good method to forecast for stationary and non stationary time series. According to the data which obtained from Baghdad Water Authority, we are modelling two series, the first one about pure water consumption and the second about the number of participants. Then we determine an optimal model by depending on choosing minimum MSE as criterion.
The Purpose of this research is a comparison between two types of multivariate GARCH models BEKK and DVECH to forecast using financial time series which are the series of daily Iraqi dinar exchange rate with dollar, the global daily of Oil price with dollar and the global daily of gold price with dollar for the period from 01/01/2014 till 01/01/2016.The estimation, testing and forecasting process has been computed through the program RATS. Three time series have been transferred to the three asset returns to get the Stationarity, some tests were conducted including Ljung- Box, Multivariate Q and Multivariate ARCH to Returns Series and Residuals Series for both models with comparison between the estimation and for
... Show MoreSoil improvement has developed as a realistic solution for enhancing soil properties so that structures can be constructed to meet project engineering requirements due to the limited availability of construction land in urban centers. The jet grouting method for soil improvement is a novel geotechnical alternative for problematic soils for which conventional foundation designs cannot provide acceptable and lasting solutions. The paper's methodology was based on constructing pile models using a low-pressure injection laboratory setup built and made locally to simulate the operation of field equipment. The setup design was based on previous research that systematically conducted unconfined compression testing (U.C.Ts.). Th
... Show MoreConcrete structures is affected by a deleterious reaction, which is known as Alkali Aggregate Reaction (AAR). AAR can be defined as a chemical reaction between the alkali content in the pore water solution of the cement paste and reactive forms of silica hold in the aggregate. This internal reaction produces expansion and cracking in concrete, which can lead to loss of strength and stiffness. Carbon fiber-reinforced polymer (CFRP) is one of the methods used to suppress further AAR expansion and rehabilitate and support damaged concrete structures. In this research, thirty-six cylindrical specimens were fabricated from non-reactive and reactive concrete, which contained fused silica as
This research sought to present a concept of cross-sectional data models, A crucial double data to take the impact of the change in time and obtained from the measured phenomenon of repeated observations in different time periods, Where the models of the panel data were defined by different types of fixed , random and mixed, and Comparing them by studying and analyzing the mathematical relationship between the influence of time with a set of basic variables Which are the main axes on which the research is based and is represented by the monthly revenue of the working individual and the profits it generates, which represents the variable response And its relationship to a set of explanatory variables represented by the
... Show MoreLongitudinal data is becoming increasingly common, especially in the medical and economic fields, and various methods have been analyzed and developed to analyze this type of data.
In this research, the focus was on compiling and analyzing this data, as cluster analysis plays an important role in identifying and grouping co-expressed subfiles over time and employing them on the nonparametric smoothing cubic B-spline model, which is characterized by providing continuous first and second derivatives, resulting in a smoother curve with fewer abrupt changes in slope. It is also more flexible and can pick up on more complex patterns and fluctuations in the data.
The longitudinal balanced data profile was compiled into subgroup
... Show MoreTwitter data analysis is an emerging field of research that utilizes data collected from Twitter to address many issues such as disaster response, sentiment analysis, and demographic studies. The success of data analysis relies on collecting accurate and representative data of the studied group or phenomena to get the best results. Various twitter analysis applications rely on collecting the locations of the users sending the tweets, but this information is not always available. There are several attempts at estimating location based aspects of a tweet. However, there is a lack of attempts on investigating the data collection methods that are focused on location. In this paper, we investigate the two methods for obtaining location-based dat
... Show MoreReal life scheduling problems require the decision maker to consider a number of criteria before arriving at any decision. In this paper, we consider the multi-criteria scheduling problem of n jobs on single machine to minimize a function of five criteria denoted by total completion times (∑), total tardiness (∑), total earliness (∑), maximum tardiness () and maximum earliness (). The single machine total tardiness problem and total earliness problem are already NP-hard, so the considered problem is strongly NP-hard.
We apply two local search algorithms (LSAs) descent method (DM) and simulated annealing method (SM) for the 1// (∑∑∑
... Show More