Recent years have seen an explosion in graph data from a variety of scientific, social and technological fields. From these fields, emotion recognition is an interesting research area because it finds many applications in real life such as in effective social robotics to increase the interactivity of the robot with human, driver safety during driving, pain monitoring during surgery etc. A novel facial emotion recognition based on graph mining has been proposed in this paper to make a paradigm shift in the way of representing the face region, where the face region is represented as a graph of nodes and edges and the gSpan frequent sub-graphs mining algorithm is used to find the frequent sub-structures in the graph database of each emotion. To reduce the number of generated sub-graphs, overlap ratio metric is utilized for this purpose. After encoding the final selected sub-graphs, binary classification is then applied to classify the emotion of the queried input facial image using six levels of classification. Binary cat swarm intelligence is applied within each level of classification to select proper sub-graphs that give the highest accuracy in that level. Different experiments have been conducted using Surrey Audio-Visual Expressed Emotion (SAVEE) database and the final system accuracy was 90.00%. The results show significant accuracy improvements (about 2%) by the proposed system in comparison to current published works in SAVEE database.
The strong cryptography employed by PGP (Pretty Good Privacy) is one of the best available today. The PGP protocol is a hybrid cryptosystem that combines some of the best features of both conventional and public-key cryptography. This paper aim to improve PGP protocol by combined between the Random Genetic algorithm, NTRU (N-th degree Truncated polynomial Ring Unit) algorithm with PGP protocol stages in order to increase PGP protocol speed, security, and make it more difficult in front of the counterfeiter. This can be achieved by use the Genetic algorithm that only generates the keys according to the Random Genetic equations. The final keys that obtained from Genetic algorithm were observed to be purely random (according to the randomne
... Show MoreScheduling Timetables for courses in the big departments in the universities is a very hard problem and is often be solved by many previous works although results are partially optimal. This work implements the principle of an evolutionary algorithm by using genetic theories to solve the timetabling problem to get a random and full optimal timetable with the ability to generate a multi-solution timetable for each stage in the collage. The major idea is to generate course timetables automatically while discovering the area of constraints to get an optimal and flexible schedule with no redundancy through the change of a viable course timetable. The main contribution in this work is indicated by increasing the flexibility of generating opti
... Show More In this research, an adaptive Canny algorithm using fast Otsu multithresholding method is presented, in which fast Otsu multithresholding method is used to calculate the optimum maximum and minimum hysteresis values and used as automatic thresholding for the fourth stage of the Canny algorithm. The new adaptive Canny algorithm and the standard Canny algorithm (manual hysteresis value) was tested on standard image (Lena) and satellite image. The results approved the validity and accuracy of the new algorithm to find the images edges for personal and satellite images as pre-step for image segmentation.
In this paper we proposes the philosophy of the Darwinian selection as synthesis method called Genetic algorithm ( GA ), and include new merit function with simple form then its uses in other works for designing one of the kinds of multilayer optical filters called high reflection mirror. Here we intend to investigate solutions for many practical problems. This work appears designed high reflection mirror that have good performance with reduction the number of layers, which can enable one to controlling the errors effect of the thickness layers on the final product, where in this work we can yield such a solution in a very shorter time by controlling the length of the chromosome and optimal genetic operators . Res
... Show MoreIn this paper, the propose is to use the xtreme value distribution as the rate of occurrence of the non-homogenous Poisson process, in order to improve the rate of occurrence of the non-homogenous process, which has been called the Extreme value Process. To estimate the parameters of this process, it is proposed to use the Maximum Likelihood method, Method of Moment and a smart method represented by the Artificial Bee Colony:(ABC) algorithm to reach an estimator for this process which represents the best data representation. The results of the three methods are compared through a simulation of the model, and it is concluded that the estimator of (ABC) is better than the estimator of the maximum likelihood method and method of mo
... Show MoreBackground/Objectives: The purpose of current research aims to a modified image representation framework for Content-Based Image Retrieval (CBIR) through gray scale input image, Zernike Moments (ZMs) properties, Local Binary Pattern (LBP), Y Color Space, Slantlet Transform (SLT), and Discrete Wavelet Transform (DWT). Methods/Statistical analysis: This study surveyed and analysed three standard datasets WANG V1.0, WANG V2.0, and Caltech 101. The features an image of objects in this sets that belong to 101 classes-with approximately 40-800 images for every category. The suggested infrastructure within the study seeks to present a description and operationalization of the CBIR system through automated attribute extraction system premised on CN
... Show MoreSegmentation is one of the most computer vision processes importance, it aims to understand the image contents by partitioning it into segments that are more meaningful and easier to analyze. However, this process comes with a set of challenges including image skew, noise, and object clipping. In this paper, a solution is proposed to address the challenges encountered when using Optical Character Recognition to recognize mathematical expressions. The proposed method involves three stages: pre-processing, segmentation, and post-processing. During pre-processing, the mathematical expression image is transformed into a binary image, noise reduction techniques are applied, image component discontinuities are resolved, and skew corre
... Show MoreThe limitations of wireless sensor nodes are power, computational capabilities, and memory. This paper suggests a method to reduce the power consumption by a sensor node. This work is based on the analogy of the routing problem to distribute an electrical field in a physical media with a given density of charges. From this analogy a set of partial differential equations (Poisson's equation) is obtained. A finite difference method is utilized to solve this set numerically. Then a parallel implementation is presented. The parallel implementation is based on domain decomposition, where the original calculation domain is decomposed into several blocks, each of which given to a processing element. All nodes then execute computations in parall
... Show MoreAssessment the actual accuracy of laboratory devices prior to first use is very important to know the capabilities of such devices and employ them in multiple domains. As the manual of the device provides information and values in laboratory conditions for the accuracy of these devices, thus the actual evaluation process is necessary.
In this paper, the accuracy of laser scanner (stonex X-300) cameras were evaluated, so that those cameras attached to the device and lead supporting role in it. This is particularly because the device manual did not contain sufficient information about those cameras.
To know the accuracy when using these cameras in close range photogrammetry, laser scanning (stonex X-300) de
... Show MoreIn this paper, we prove that our proposed localization algorithm named Improved
Accuracy Distribution localization for wireless sensor networks (IADLoc) [1] is the
best when it is compared with the other localization algorithms by introducing many
cases of studies. The IADLoc is used to minimize the error rate of localization
without any additional cost and minimum energy consumption and also
decentralized implementation. The IADLoc is a range free and also range based
localization algorithm that uses both type of antenna (directional and omnidirectional)
it allows sensors to determine their location based on the region of
intersection (ROI) when the beacon nodes send the information to the sink node and
the la