In the last few years, the Internet of Things (IoT) is gaining remarkable attention in both academic and industrial worlds. The main goal of the IoT is laying on describing everyday objects with different capabilities in an interconnected fashion to the Internet to share resources and to carry out the assigned tasks. Most of the IoT objects are heterogeneous in terms of the amount of energy, processing ability, memory storage, etc. However, one of the most important challenges facing the IoT networks is the energy-efficient task allocation. An efficient task allocation protocol in the IoT network should ensure the fair and efficient distribution of resources for all objects to collaborate dynamically with limited energy. The canonical definition for network lifetime in the IoT is to increase the period of cooperation between objects to carry out all the assigned tasks. The main contribution in this paper is to address the problem of task allocation in the IoT as an optimization problem with a lifetime-aware model. A genetic algorithm is proposed as a task allocation protocol. For the proposed algorithm, a problem-tailored individual representation and a modified uniform crossover are designed. Further, the individual initialization and perturbation operators (crossover and mutation) are designed so as to remedy the infeasibility of any solution located or reached by the proposed genetic algorithm. The results showed reasonable performance for the proposed genetic-based task allocation protocol. Further, the results prove the necessity for designing problem-specific operators instead of adopting the canonical counterparts.
The convergence speed is the most important feature of Back-Propagation (BP) algorithm. A lot of improvements were proposed to this algorithm since its presentation, in order to speed up the convergence phase. In this paper, a new modified BP algorithm called Speeding up Back-Propagation Learning (SUBPL) algorithm is proposed and compared to the standard BP. Different data sets were implemented and experimented to verify the improvement in SUBPL.
The goal of this work is demonstrating, through the gradient observation of a of type linear ( -systems), the possibility for reducing the effect of any disturbances (pollution, radiation, infection, etc.) asymptotically, by a suitable choice of related actuators of these systems. Thus, a class of ( -system) was developed based on finite time ( -system). Furthermore, definitions and some properties of this concept -system and asymptotically gradient controllable system ( -controllable) were stated and studied. More precisely, asymptotically gradient efficient actuators ensuring the weak asymptotically gradient compensation system ( -system) of known or unknown disturbances are examined. Consequently, under convenient hypo
... Show MoreThe increasing availability of computing power in the past two decades has been use to develop new techniques for optimizing solution of estimation problem. Today's computational capacity and the widespread availability of computers have enabled development of new generation of intelligent computing techniques, such as our interest algorithm, this paper presents one of new class of stochastic search algorithm (known as Canonical Genetic' Algorithm ‘CGA’) for optimizing the maximum likelihood function strategy is composed of three main steps: recombination, mutation, and selection. The experimental design is based on simulating the CGA with different values of are compared with those of moment method. Based on MSE value obtained from bot
... Show MoreAll modern critical approaches attempt to cover the meanings and overtones of the text, claiming that they are better than others in the analysis and attainment of the intended meanings of the text. The structural approach claims to be able to do so more than any other modern critical approach, as it claimed that it is possible to separate what is read from the reader, on the presumed belief that it is possible to read the text with a zero-memory. However, the studies in criticism of criticism state that each of these approaches is successful in dealing with the text in one or more aspects while failing in one or more aspects. Consequently, the criticism whether the approach possesses the text, or that the text rejects this possession, r
... Show MoreThe logistic regression model is an important statistical model showing the relationship between the binary variable and the explanatory variables. The large number of explanations that are usually used to illustrate the response led to the emergence of the problem of linear multiplicity between the explanatory variables that make estimating the parameters of the model not accurate.
... Show MoreResearch Hypothesis from the fact that kicks off the effect that agricultural production in Iraq plays an important role in overcoming the food problem and achieving food security, but he became far far away from the provision of sufficient quantities of food products and then securing the Iraqi consumer food basket by the challenges faced by the agricultural sector.
To prove the hypothesis research in its structure in three axes came, the first axis eating historical significance to the subject of food over time periods as well as to clarify the concept of food security, and the second axis touched on the most important challenges facing the agricultural sector in Iraq and prevent the achievement of food requirements for members of
In this paper, the methods of weighted residuals: Collocation Method (CM), Least Squares Method (LSM) and Galerkin Method (GM) are used to solve the thin film flow (TFF) equation. The weighted residual methods were implemented to get an approximate solution to the TFF equation. The accuracy of the obtained results is checked by calculating the maximum error remainder functions (MER). Moreover, the outcomes were examined in comparison with the 4th-order Runge-Kutta method (RK4) and good agreements have been achieved. All the evaluations have been successfully implemented by using the computer system Mathematica®10.
Most of the medical datasets suffer from missing data, due to the expense of some tests or human faults while recording these tests. This issue affects the performance of the machine learning models because the values of some features will be missing. Therefore, there is a need for a specific type of methods for imputing these missing data. In this research, the salp swarm algorithm (SSA) is used for generating and imputing the missing values in the pain in my ass (also known Pima) Indian diabetes disease (PIDD) dataset, the proposed algorithm is called (ISSA). The obtained results showed that the classification performance of three different classifiers which are support vector machine (SVM), K-nearest neighbour (KNN), and Naïve B
... Show More