Feature selection (FS) constitutes a series of processes used to decide which relevant features/attributes to include and which irrelevant features to exclude for predictive modeling. It is a crucial task that aids machine learning classifiers in reducing error rates, computation time, overfitting, and improving classification accuracy. It has demonstrated its efficacy in myriads of domains, ranging from its use for text classification (TC), text mining, and image recognition. While there are many traditional FS methods, recent research efforts have been devoted to applying metaheuristic algorithms as FS techniques for the TC task. However, there are few literature reviews concerning TC. Therefore, a comprehensive overview was systematicall
... Show MoreThis paper considers a new Double Integral transform called Double Sumudu-Elzaki transform DSET. The combining of the DSET with a semi-analytical method, namely the variational iteration method DSETVIM, to arrive numerical solution of nonlinear PDEs of Fractional Order derivatives. The proposed dual method property decreases the number of calculations required, so combining these two methods leads to calculating the solution's speed. The suggested technique is tested on four problems. The results demonstrated that solving these types of equations using the DSETVIM was more advantageous and efficient
Abstract
The Non - Homogeneous Poisson process is considered as one of the statistical subjects which had an importance in other sciences and a large application in different areas as waiting raws and rectifiable systems method , computer and communication systems and the theory of reliability and many other, also it used in modeling the phenomenon that occurred by unfixed way over time (all events that changed by time).
This research deals with some of the basic concepts that are related to the Non - Homogeneous Poisson process , This research carried out two models of the Non - Homogeneous Poisson process which are the power law model , and Musa –okumto , to estimate th
... Show MoreA condense study was done to compare between the ordinary estimators. In particular the maximum likelihood estimator and the robust estimator, to estimate the parameters of the mixed model of order one, namely ARMA(1,1) model.
Simulation study was done for a varieties the model. using: small, moderate and large sample sizes, were some new results were obtained. MAPE was used as a statistical criterion for comparison.
Abstract
The use of modern scientific methods and techniques, is considered important topics to solve many of the problems which face some sector, including industrial, service and health. The researcher always intends to use modern methods characterized by accuracy, clarity and speed to reach the optimal solution and be easy at the same time in terms of understanding and application.
the research presented this comparison between the two methods of solution for linear fractional programming models which are linear transformation for Charnas & Cooper , and denominator function restriction method through applied on the oil heaters and gas cookers plant , where the show after reac
... Show MoreCommunication of the human brain with the surroundings became reality by using Brain- Computer Interface (BCI) based mechanism. Electroencephalography (EEG) being the non-invasive method has become popular for interaction with the brain. Traditionally, the devices were used for clinical applications to detect various brain diseases but with the advancement in technologies, companies like Emotiv, NeuoSky are coming up with low cost, easily portable EEG based consumer graded devices that can be used in various application domains like gaming, education etc as these devices are comfortable to wear also. This paper reviews the fields where the EEG has shown its impact and the way it has p
The sensitive and important data are increased in the last decades rapidly, since the tremendous updating of networking infrastructure and communications. to secure this data becomes necessary with increasing volume of it, to satisfy securing for data, using different cipher techniques and methods to ensure goals of security that are integrity, confidentiality, and availability. This paper presented a proposed hybrid text cryptography method to encrypt a sensitive data by using different encryption algorithms such as: Caesar, Vigenère, Affine, and multiplicative. Using this hybrid text cryptography method aims to make the encryption process more secure and effective. The hybrid text cryptography method depends on circular queue. Using circ
... Show MoreThe present work involved designing and synthesizing of a series of new. compounds which their molecules are composed from two biologically active components namely sulfamethoxazole or β-lactam containing drugs and cyclic imides. The target new compounds were synthesized by two steps in the first one a series of six bis (N-drug phthalamic acid_4-yl) ketone (1-6) were prepared from the reaction of sulfamethoxazole or β-lactam containing drugs with benzophenone 3, 3′, 4, 4′ -tetracarboxylic dianhydride.
In the second step, compounds (1-6) were introduced in dehydration reaction via fusion process producing the target compounds bis (N-drug phthalimidyl-4-yl) ketone (7-12). The antibacterial and antifungal high
... Show More