In this paper, double Sumudu and double Elzaki transforms methods are used to compute the numerical solutions for some types of fractional order partial differential equations with constant coefficients and explaining the efficiently of the method by illustrating some numerical examples that are computed by using Mathcad 15.and graphic in Matlab R2015a.
The necessary optimality conditions with Lagrange multipliers are studied and derived for a new class that includes the system of Caputo–Katugampola fractional derivatives to the optimal control problems with considering the end time free. The formula for the integral by parts has been proven for the left Caputo–Katugampola fractional derivative that contributes to the finding and deriving the necessary optimality conditions. Also, three special cases are obtained, including the study of the necessary optimality conditions when both the final time and the final state are fixed. According to convexity assumptions prove that necessary optimality conditions are sufficient optimality conditions.
... Show MoreThe Hartley transform generalizes to the fractional Hartley transform (FRHT) which gives various uses in different fields of image encryption. Unfortunately, the available literature of fractional Hartley transform is unable to provide its inversion theorem. So accordingly original function cannot retrieve directly, which restrict its applications. The intension of this paper is to propose inversion theorem of fractional Hartley transform to overcome this drawback. Moreover, some properties of fractional Hartley transform are discussed in this paper.
Presents here in the results of comparison between the theoretical equation stated by Huang and Menq and laboratory model tests used to study the bearing capacity of square footing on geogrid-reinforced loose sand by performing model tests. The effects of several parameters were studied in order to study the general behavior of improving the soil by using the geogrid. These parameters include depth of first layer of reinforcement, vertical spacing of reinforcement layers, number of reinforcement layers and types of reinforcement layers The results show that the theoretical equation can be used to estimate the bearing capacity of loose sand.
In this research the Empirical Bayes method is used to Estimate the affiliation parameter in the clinical trials and then we compare this with the Moment Estimates for this parameter using Monte Carlo stimulation , we assumed that the distribution of the observation is binomial distribution while the distribution with the unknown random parameters is beta distribution ,finally we conclude that the Empirical bayes method for the random affiliation parameter is efficient using Mean Squares Error (MSE) and for different Sample size .
This paper proposes a new method to tune a fractional order PID controller. This method utilizes both the analytic and numeric approach to determine the controller parameters. The control design specifications that must be achieved by the control system are gain crossover frequency, phase margin, and peak magnitude at the resonant frequency, where the latter is a new design specification suggested by this paper. These specifications results in three equations in five unknown variables. Assuming that certain relations exist between two variables and discretizing one of them, a performance index can be evaluated and the optimal controller parameters that minimize this performance index are selected. As a case study, a thir
... Show MoreA mixture model is used to model data that come from more than one component. In recent years, it became an effective tool in drawing inferences about the complex data that we might come across in real life. Moreover, it can represent a tremendous confirmatory tool in classification observations based on similarities amongst them. In this paper, several mixture regression-based methods were conducted under the assumption that the data come from a finite number of components. A comparison of these methods has been made according to their results in estimating component parameters. Also, observation membership has been inferred and assessed for these methods. The results showed that the flexible mixture model outperformed the others
... Show MoreA mixture model is used to model data that come from more than one component. In recent years, it became an effective tool in drawing inferences about the complex data that we might come across in real life. Moreover, it can represent a tremendous confirmatory tool in classification observations based on similarities amongst them. In this paper, several mixture regression-based methods were conducted under the assumption that the data come from a finite number of components. A comparison of these methods has been made according to their results in estimating component parameters. Also, observation membership has been inferred and assessed for these methods. The results showed that the flexible mixture model outperformed the
... Show MoreMultilevel models are among the most important models widely used in the application and analysis of data that are characterized by the fact that observations take a hierarchical form, In our research we examined the multilevel logistic regression model (intercept random and slope random model) , here the importance of the research highlights that the usual regression models calculate the total variance of the model and its inability to read variance and variations between levels ,however in the case of multi-level regression models, the calculation of the total variance is inaccurate and therefore these models calculate the variations for each level of the model, Where the research aims to estimate the parameters of this m
... Show MoreThis paper includes a comparison between denoising techniques by using statistical approach, principal component analysis with local pixel grouping (PCA-LPG), this procedure is iterated second time to further improve the denoising performance, and other enhancement filters were used. Like adaptive Wiener low pass-filter to a grayscale image that has been degraded by constant power additive noise, based on statistics estimated from a local neighborhood of each pixel. Performs Median filter of the input noisy image, each output pixel contains the Median value in the M-by-N neighborhood around the corresponding pixel in the input image, Gaussian low pass-filter and Order-statistic filter also be used. Experimental results shows LPG-PCA method
... Show MoreNeighShrink is an efficient image denoising algorithm based on the discrete wavelet
transform (DWT). Its disadvantage is to use a suboptimal universal threshold and identical
neighbouring window size in all wavelet subbands. Dengwen and Wengang proposed an
improved method, which can determine an optimal threshold and neighbouring window size
for every subband by the Stein’s unbiased risk estimate (SURE). Its denoising performance is
considerably superior to NeighShrink and also outperforms SURE-LET, which is an up-todate
denoising algorithm based on the SURE. In this paper different wavelet transform
families are used with this improved method, the results show that Haar wavelet has the
lowest performance among