Feature selection (FS) constitutes a series of processes used to decide which relevant features/attributes to include and which irrelevant features to exclude for predictive modeling. It is a crucial task that aids machine learning classifiers in reducing error rates, computation time, overfitting, and improving classification accuracy. It has demonstrated its efficacy in myriads of domains, ranging from its use for text classification (TC), text mining, and image recognition. While there are many traditional FS methods, recent research efforts have been devoted to applying metaheuristic algorithms as FS techniques for the TC task. However, there are few literature reviews concerning TC. Therefore, a comprehensive overview was systematically studied by exploring available studies of different metaheuristic algorithms used for FS to improve TC. This paper will contribute to the body of existing knowledge by answering four research questions (RQs): 1) What are the different approaches of FS that apply metaheuristic algorithms to improve TC? 2) Does applying metaheuristic algorithms for TC lead to better accuracy than the typical FS methods? 3) How effective are the modified, hybridized metaheuristic algorithms for text FS problems?, and 4) What are the gaps in the current studies and their future directions? These RQs led to a study of recent works on metaheuristic-based FS methods, their contributions, and limitations. Hence, a final list of thirty-seven (37) related articles was extracted and investigated to align with our RQs to generate new knowledge in the domain of study. Most of the conducted papers focused on addressing the TC in tandem with metaheuristic algorithms based on the wrapper and hybrid FS approaches. Future research should focus on using a hybrid-based FS approach as it intuitively handles complex optimization problems and potentiality provide new research opportunities in this rapidly developing field.
In this research study, I tried to trace the epic effect to learn how it was understood and how it was used. Following the descriptive and analytical approach in the research, the first chapter dealt with a presentation of the methodological framework of the problem, the goal, the limits of the research, the importance and the need for it and the definition of terms, as well as the theoretical framework which consisted of two topics, including the impact of the epic theater on the world theater and the second the effect of the epic theater on the Arab theater, This came by tracing the epic impact on the world stage of the Greeks, the Middle Ages, the Renaissance, and the Arab theater of the twentieth century.
As for the second
As a result of the development and global openness and the possibility of companies providing their services outside their spatial boundaries that were determined by them, and the transformation of the world due to the development of the means of communication into a large global market that accommodates all products from different regions and of the same type and production field, competition resulted between companies, and the race to obtain the largest market share It ensures the largest amount of profits, and it is natural for the advertising promotion by companies for their product to shift from an advertisement for one product to a competitive advertisement that calls on the recipient to leave the competing product and switch to it
... Show MoreClassification of imbalanced data is an important issue. Many algorithms have been developed for classification, such as Back Propagation (BP) neural networks, decision tree, Bayesian networks etc., and have been used repeatedly in many fields. These algorithms speak of the problem of imbalanced data, where there are situations that belong to more classes than others. Imbalanced data result in poor performance and bias to a class without other classes. In this paper, we proposed three techniques based on the Over-Sampling (O.S.) technique for processing imbalanced dataset and redistributing it and converting it into balanced dataset. These techniques are (Improved Synthetic Minority Over-Sampling Technique (Improved SMOTE), Border
... Show MoreBecause of the experience of the mixture problem of high correlation and the existence of linear MultiCollinearity between the explanatory variables, because of the constraint of the unit and the interactions between them in the model, which increases the existence of links between the explanatory variables and this is illustrated by the variance inflation vector (VIF), L-Pseudo component to reduce the bond between the components of the mixture.
To estimate the parameters of the mixture model, we used in our research the use of methods that increase bias and reduce variance, such as the Ridge Regression Method and the Least Absolute Shrinkage and Selection Operator (LASSO) method a
... Show MoreIn this paper, the methods of weighted residuals: Collocation Method (CM), Least Squares Method (LSM) and Galerkin Method (GM) are used to solve the thin film flow (TFF) equation. The weighted residual methods were implemented to get an approximate solution to the TFF equation. The accuracy of the obtained results is checked by calculating the maximum error remainder functions (MER). Moreover, the outcomes were examined in comparison with the 4th-order Runge-Kutta method (RK4) and good agreements have been achieved. All the evaluations have been successfully implemented by using the computer system Mathematica®10.
The purpose of this study is to measure the levels of quality control for some crude oil products in Iraqi refineries, and how they are close to the international standards, through the application of statistical methods in quality control of oil products in Iraqi refineries. Where the answers of the study sample were applied to a group of Iraqi refinery employees (Al-Dora refinery, Al-Nasiriyah refinery, and Al-Basra refinery) on the principles of quality management control, and according to the different personal characteristics (gender, age, academic qualification, number of years of experience, job level). In order to achieve the objectives of the study, a questionnaire that included (12) items, in order to collect preliminary inform
... Show MoreFractal geometry is receiving increase attention as a quantitative and qualitative model for natural phenomena description, which can establish an active classification technique when applied on satellite images. In this paper, a satellite image is used which was taken by Quick Bird that contains different visible classes. After pre-processing, this image passes through two stages: segmentation and classification. The segmentation carried out by hybrid two methods used to produce effective results; the two methods are Quadtree method that operated inside Horizontal-Vertical method. The hybrid method is segmented the image into two rectangular blocks, either horizontally or vertically depending on spectral uniformity crit
... Show MoreThe audience is one of the important practical elements in the theatrical show and its importance is not confined to its static activity as a receiver element only, rather it went beyond that issue as an effective and influential element in the proceedings of the show and the process of meaning construction, that it gains an active role in the construction and production of the connotation that influences and is influenced by the actor, where the communication channels are open between the two sides, consequently a kind of watching and joint interaction happens between them. Thus, it has become necessary for the actor to create a suitable environment for the onlookers in order for it to be an essential part of the show system. The
... Show MoreThis work implements an Electroencephalogram (EEG) signal classifier. The implemented method uses Orthogonal Polynomials (OP) to convert the EEG signal samples to moments. A Sparse Filter (SF) reduces the number of converted moments to increase the classification accuracy. A Support Vector Machine (SVM) is used to classify the reduced moments between two classes. The proposed method’s performance is tested and compared with two methods by using two datasets. The datasets are divided into 80% for training and 20% for testing, with 5 -fold used for cross-validation. The results show that this method overcomes the accuracy of other methods. The proposed method’s best accuracy is 95.6% and 99.5%, respectively. Finally, from the results, it
... Show More