The rise of edge-cloud continuum computing is a result of the growing significance of edge computing, which has become a complementary or substitute option for traditional cloud services. The convergence of networking and computers presents a notable challenge due to their distinct historical development. Task scheduling is a major challenge in the context of edge-cloud continuum computing. The selection of the execution location of tasks, is crucial in meeting the quality-of-service (QoS) requirements of applications. An efficient scheduling strategy for distributing workloads among virtual machines in the edge-cloud continuum data center is mandatory to ensure the fulfilment of QoS requirements for both customer and service provider. Existing research used metaheuristic algorithm to solve tak scheduling problem, however, must of the existing metaheuristics used suffers from falling into local mina due to their inefficiency to avoid unfeasible region in the solution search space. Therefore, there is a dire need for an efficient metaheuristic algorithm for task scheduling. This study proposed an FPA-ISFLA task scheduling model using hybrid flower pollination and improved shuffled frog leaping algorithms. The simulation results indicate that the FPA-ISFLA algorithm is superior to the PSO algorithm in terms of makespan time, resource utilization, and execution cost reduction, especially with an increasing number of tasks.
A novel median filter based on crow optimization algorithms (OMF) is suggested to reduce the random salt and pepper noise and improve the quality of the RGB-colored and gray images. The fundamental idea of the approach is that first, the crow optimization algorithm detects noise pixels, and that replacing them with an optimum median value depending on a criterion of maximization fitness function. Finally, the standard measure peak signal-to-noise ratio (PSNR), Structural Similarity, absolute square error and mean square error have been used to test the performance of suggested filters (original and improved median filter) used to removed noise from images. It achieves the simulation based on MATLAB R2019b and the resul
... Show MoreRegarding to the computer system security, the intrusion detection systems are fundamental components for discriminating attacks at the early stage. They monitor and analyze network traffics, looking for abnormal behaviors or attack signatures to detect intrusions in early time. However, many challenges arise while developing flexible and efficient network intrusion detection system (NIDS) for unforeseen attacks with high detection rate. In this paper, deep neural network (DNN) approach was proposed for anomaly detection NIDS. Dropout is the regularized technique used with DNN model to reduce the overfitting. The experimental results applied on NSL_KDD dataset. SoftMax output layer has been used with cross entropy loss funct
... Show MoreCloud computing describes computer services provided through the internet and includes a wide range of virtualization resources. Because cloud computing is made up of a sizable number of heterogeneous autonomous systems with an adaptable computational architecture, it has been widely adopted by many businesses. The scheduling and management of resource utilization, however, have become more difficult as a result of cloud computing. Task scheduling is crucial, and this procedure must schedule tasks on the virtual machine while using the least amount of time possible. Utilizing an effective scheduling strategy enhances and expedites cloud computing services. Optimization techniques are used to resolve cloud scheduling problems.
... Show MoreArtificial Intelligence Algorithms have been used in recent years in many scientific fields. We suggest employing flower pollination algorithm in the environmental field to find the best estimate of the semi-parametric regression function with measurement errors in the explanatory variables and the dependent variable, where measurement errors appear frequently in fields such as chemistry, biological sciences, medicine, and epidemiological studies, rather than an exact measurement. We estimate the regression function of the semi-parametric model by estimating the parametric model and estimating the non-parametric model, the parametric model is estimated by using an instrumental variables method (Wald method, Bartlett’s method, and Durbin
... Show MoreMost recognition system of human facial emotions are assessed solely on accuracy, even if other performance criteria are also thought to be important in the evaluation process such as sensitivity, precision, F-measure, and G-mean. Moreover, the most common problem that must be resolved in face emotion recognition systems is the feature extraction methods, which is comparable to traditional manual feature extraction methods. This traditional method is not able to extract features efficiently. In other words, there are redundant amount of features which are considered not significant, which affect the classification performance. In this work, a new system to recognize human facial emotions from images is proposed. The HOG (Histograms of Or
... Show MoreNurse scheduling problem is one of combinatorial optimization problems and it is one of NP-Hard problems which is difficult to be solved as optimal solution. In this paper, we had created an proposed algorithm which it is hybrid simulated annealing algorithm to solve nurse scheduling problem, developed the simulated annealing algorithm and Genetic algorithm. We can note that the proposed algorithm (Hybrid simulated Annealing Algorithm(GS-h)) is the best method among other methods which it is used in this paper because it satisfied minimum average of the total cost and maximum number of Solved , Best and Optimal problems. So we can note that the ratios of the optimal solution are 77% for the proposed algorithm(GS-h), 28.75% for Si
... Show MoreEstimating the semantic similarity between short texts plays an increasingly prominent role in many fields related to text mining and natural language processing applications, especially with the large increase in the volume of textual data that is produced daily. Traditional approaches for calculating the degree of similarity between two texts, based on the words they share, do not perform well with short texts because two similar texts may be written in different terms by employing synonyms. As a result, short texts should be semantically compared. In this paper, a semantic similarity measurement method between texts is presented which combines knowledge-based and corpus-based semantic information to build a semantic network that repre
... Show MoreTask scheduling in an important element in a distributed system. It is vital how the jobs are correctly assigned for each computer’s processor to improve performance. The presented approaches attempt to reduce the expense of optimizing the use of the CPU. These techniques mostly lack planning and in need to be comprehensive. To address this fault, a hybrid optimization scheduling technique is proposed for the hybridization of both First-Come First-Served (FCFS), and Shortest Job First (SJF). In addition, we propose to apply Simulated Annealing (SA) algorithm as an optimization technique to find optimal job’s execution sequence considering both job’s entrance time and job’s execution time to balance them to reduce the job
... Show MoreThe purified frog skin peptides were tested on leukemic patients lymphocytes, which revealed effects of cytotoxicity. Four frogs (Rana ridibunda) were stimulated by single intra-peritoneal injection of norepinephrine-HCl . Five different peptides;1(18) A, 2(19) L, 3(20) I,4(21) E and 5(22) Y were isolated and quantified. The peptide 3(20)I had 5.87% of hemolysis, while healthy human lymphocytes cytotoxic activity was for 2(19)L with inhibition( -10.4%).All peptides were subjected to polyacrylamide gel electrophoresis. The results revealed peptides 1(18)A, 2(19)L, 3(20)I which appeared as low as 10 KDa marker. Theoretically, the whole polypeptide had a molecular weight 7488.61 Dalt
... Show MoreSemantic segmentation is effective in numerous object classification tasks such as autonomous vehicles and scene understanding. With the advent in the deep learning domain, lots of efforts are seen in applying deep learning algorithms for semantic segmentation. Most of the algorithms gain the required accuracy while compromising on their storage and computational requirements. The work showcases the implementation of Convolutional Neural Network (CNN) using Discrete Cosine Transform (DCT), where DCT exhibit exceptional energy compaction properties. The proposed Adaptive Weight Wiener Filter (AWWF) rearranges the DCT coefficients by truncating the high frequency coefficients. AWWF-DCT model reinstate the convolutional l
... Show More