In this paper, the botnet detection problem is defined as a feature selection problem and the genetic algorithm (GA) is used to search for the best significant combination of features from the entire search space of set of features. Furthermore, the Decision Tree (DT) classifier is used as an objective function to direct the ability of the proposed GA to locate the combination of features that can correctly classify the activities into normal traffics and botnet attacks. Two datasets namely the UNSW-NB15 and the Canadian Institute for Cybersecurity Intrusion Detection System 2017 (CICIDS2017), are used as evaluation datasets. The results reveal that the proposed DT-aware GA can effectively find the relevant features from the whole features set. Thus, it obtains efficient botnet detection results in terms of F-score, precision, detection rate, and number of relevant features, when compared with DT alone.
<p>The directing of a wheeled robot in an unknown moving environment with physical barriers is a difficult proposition. In particular, having an optimal or near-optimal path that avoids obstacles is a major challenge. In this paper, a modified neuro-controller mechanism is proposed for controlling the movement of an indoor mobile robot. The proposed mechanism is based on the design of a modified Elman neural network (MENN) with an effective element aware gate (MEEG) as the neuro-controller. This controller is updated to overcome the rigid and dynamic barriers in the indoor area. The proposed controller is implemented with a mobile robot known as Khepera IV in a practical manner. The practical results demonstrate that the propo
... Show MoreOften phenomena suffer from disturbances in their data as well as the difficulty of formulation, especially with a lack of clarity in the response, or the large number of essential differences plaguing the experimental units that have been taking this data from them. Thus emerged the need to include an estimation method implicit rating of these experimental units using the method of discrimination or create blocks for each item of these experimental units in the hope of controlling their responses and make it more homogeneous. Because of the development in the field of computers and taking the principle of the integration of sciences it has been found that modern algorithms used in the field of Computer Science genetic algorithm or ant colo
... Show MoreThe modern systems that have been based upon the hash function are more suitable compared to the conventional systems; however, the complicated algorithms for the generation of the invertible functions have a high level of time consumption. With the use of the GAs, the key strength is enhanced, which results in ultimately making the entire algorithm sufficient. Initially, the process of the key generation is performed by using the results of n-queen problem that is solved by the genetic algorithm, with the use of a random number generator and through the application of the GA operations. Ultimately, the encryption of the data is performed with the use of the Modified Reverse Encryption Algorithm (MREA). It was noticed that the
... Show MoreToday the Genetic Algorithm (GA) tops all the standard algorithms in solving complex nonlinear equations based on the laws of nature. However, permute convergence is considered one of the most significant drawbacks of GA, which is known as increasing the number of iterations needed to achieve a global optimum. To address this shortcoming, this paper proposes a new GA based on chaotic systems. In GA processes, we use the logistic map and the Linear Feedback Shift Register (LFSR) to generate chaotic values to use instead of each step requiring random values. The Chaos Genetic Algorithm (CGA) avoids local convergence more frequently than the traditional GA due to its diversity. The concept is using chaotic sequences with LFSR to gene
... Show MoreFor several applications, it is very important to have an edge detection technique matching human visual contour perception and less sensitive to noise. The edge detection algorithm describes in this paper based on the results obtained by Maximum a posteriori (MAP) and Maximum Entropy (ME) deblurring algorithms. The technique makes a trade-off between sharpening and smoothing the noisy image. One of the advantages of the described algorithm is less sensitive to noise than that given by Marr and Geuen techniques that considered to be the best edge detection algorithms in terms of matching human visual contour perception.
Finding communities of connected individuals in complex networks is challenging, yet crucial for understanding different real-world societies and their interactions. Recently attention has turned to discover the dynamics of such communities. However, detecting accurate community structures that evolve over time adds additional challenges. Almost all the state-of-the-art algorithms are designed based on seemingly the same principle while treating the problem as a coupled optimization model to simultaneously identify community structures and their evolution over time. Unlike all these studies, the current work aims to individually consider this three measures, i.e. intra-community score, inter-community score, and evolution of community over
... Show MoreWith the development of communication technologies for mobile devices and electronic communications, and went to the world of e-government, e-commerce and e-banking. It became necessary to control these activities from exposure to intrusion or misuse and to provide protection to them, so it's important to design powerful and efficient systems-do-this-purpose. It this paper it has been used several varieties of algorithm selection passive immune algorithm selection passive with real values, algorithm selection with passive detectors with a radius fixed, algorithm selection with passive detectors, variable- sized intrusion detection network type misuse where the algorithm generates a set of detectors to distinguish the self-samples. Practica
... Show MoreRivest Cipher 4 (RC4) is an efficient stream cipher that is commonly used in internet protocols. However, there are several flaws in the key scheduling algorithm (KSA) of RC4. The contribution of this paper is to overcome some of these weaknesses by proposing a new version of KSA coined as modified KSA . In the initial state of the array is suggested to contain random values instead of the identity permutation. Moreover, the permutation of the array is modified to depend on the key value itself. The proposed performance is assessed in terms of cipher secrecy, randomness test and time under a set of experiments with variable key size and different plaintext size. The results show that the RC4 with improves the randomness and secrecy with
... Show More