The emphasis of Master Production Scheduling (MPS) or tactic planning is on time and spatial disintegration of the cumulative planning targets and forecasts, along with the provision and forecast of the required resources. This procedure eventually becomes considerably difficult and slow as the number of resources, products and periods considered increases. A number of studies have been carried out to understand these impediments and formulate algorithms to optimise the production planning problem, or more specifically the master production scheduling (MPS) problem. These algorithms include an Evolutionary Algorithm called Genetic Algorithm, a Swarm Intelligence methodology called Gravitational Search Algorithm (GSA), Bat Algorithm (BAT), Teaching Learning Based Algorithm and Harmony Search Algorithm (GA, TLBO and HS). This segment of the research constitutes of two parts the first emphasises on the outcome of the five simulation algorithms and then identifies the best, while the second part is about comparing the best with the search algorithm work in 2009, the applicability of BAT Algorithm and Gravitational Search Algorithm (GSABAT) to planning problems, the technique used for calculating master production scheduling, and the very important results and recommendations for future studies
For businesses that provide delivery services, the efficiency of the delivery process in terms of punctuality is very important. In addition to increasing customer trust, efficient route management, and selection are required to reduce vehicle fuel costs and expedite delivery. Some small and medium businesses still use conventional methods to manage delivery routes. Decisions to manage delivery schedules and routes do not use any specific methods to expedite the delivery settlement process. This process is inefficient, takes a long time, increases costs and is prone to errors. Therefore, the Dijkstra algorithm has been used to improve the delivery management process. A delivery management system was developed to help managers and drivers
... Show MoreIn the last decade, 3D models gained interest in many applications, such as games, the medical field, and manufacture. It is necessary to protect these models from unauthorized copying, distribution, and editing. Digital watermarking is the best way to solve this problem. This paper introduces a robust watermarking method by embedding the watermark in the low-frequency domain, then selecting the coarsest level for embedding the watermark based on the strength factor. The invisibility of the watermark for the proposed algorithm is tested by using different measurements, such as HD and PSNR. The robustness was tested by using different types of attacks; the correlation coefficient was applied for the evaluati
... Show MoreEmotion recognition has important applications in human-computer interaction. Various sources such as facial expressions and speech have been considered for interpreting human emotions. The aim of this paper is to develop an emotion recognition system from facial expressions and speech using a hybrid of machine-learning algorithms in order to enhance the overall performance of human computer communication. For facial emotion recognition, a deep convolutional neural network is used for feature extraction and classification, whereas for speech emotion recognition, the zero-crossing rate, mean, standard deviation and mel frequency cepstral coefficient features are extracted. The extracted features are then fed to a random forest classifier. In
... Show MoreThe investigation of signature validation is crucial to the field of personal authenticity. The biometrics-based system has been developed to support some information security features.Aperson’s signature, an essential biometric trait of a human being, can be used to verify their identification. In this study, a mechanism for automatically verifying signatures has been suggested. The offline properties of handwritten signatures are highlighted in this study which aims to verify the authenticity of handwritten signatures whether they are real or forged using computer-based machine learning techniques. The main goal of developing such systems is to verify people through the validity of their signatures. In this research, images of a group o
... Show MoreThe lossy-FDNR based aclive fil ter has an important property among many design realizations. 'This includes a significant reduction in component count particularly in the number of OP-AMP which consumes power. However the· problem of this type is the large component spreads which affect the fdter performance.
In this paper Genetic Algorithm is applied to minimize the component spread (capacitance and resistance p,read). The minimization of these spreads allow the fil
... Show MoreDust is a frequent contributor to health risks and changes in the climate, one of the most dangerous issues facing people today. Desertification, drought, agricultural practices, and sand and dust storms from neighboring regions bring on this issue. Deep learning (DL) long short-term memory (LSTM) based regression was a proposed solution to increase the forecasting accuracy of dust and monitoring. The proposed system has two parts to detect and monitor the dust; at the first step, the LSTM and dense layers are used to build a system using to detect the dust, while at the second step, the proposed Wireless Sensor Networks (WSN) and Internet of Things (IoT) model is used as a forecasting and monitoring model. The experiment DL system
... Show More