Grass trimming operation is widely done in Malaysia for the purpose of maintaining highways. Large number of operators engaged in this work encounters high level of noise generated by back pack type grass trimmer used for this purpose. High level of noise exposure gives different kinds of ill effect on human operators. Exact nature of deteriorated work performance is not known. For predicting the work efficiency deterioration, fuzzy tool has been used in present research. It has been established that a fuzzy computing system will help in identification and analysis of fuzzy models fuzzy system offers a convenient way of representing the relationships between the inputs and outputs of a system in the form of IF-THEN rules. The paper presents a fuzzy model for predicting the effects of noise pollution on human work efficiency as a function of noise level, exposure time and age of the operators.
Sentiment analysis is one of the major fields in natural language processing whose main task is to extract sentiments, opinions, attitudes, and emotions from a subjective text. And for its importance in decision making and in people's trust with reviews on web sites, there are many academic researches to address sentiment analysis problems. Deep Learning (DL) is a powerful Machine Learning (ML) technique that has emerged with its ability of feature representation and differentiating data, leading to state-of-the-art prediction results. In recent years, DL has been widely used in sentiment analysis, however, there is scarce in its implementation in the Arabic language field. Most of the previous researches address other l
... Show More In this research, an adaptive Canny algorithm using fast Otsu multithresholding method is presented, in which fast Otsu multithresholding method is used to calculate the optimum maximum and minimum hysteresis values and used as automatic thresholding for the fourth stage of the Canny algorithm. The new adaptive Canny algorithm and the standard Canny algorithm (manual hysteresis value) was tested on standard image (Lena) and satellite image. The results approved the validity and accuracy of the new algorithm to find the images edges for personal and satellite images as pre-step for image segmentation.
Groupwise non-rigid image alignment is a difficult non-linear optimization problem involving many parameters and often large datasets. Previous methods have explored various metrics and optimization strategies. Good results have been previously achieved with simple metrics, requiring complex optimization, often with many unintuitive parameters that require careful tuning for each dataset. In this chapter, the problem is restructured to use a simpler, iterative optimization algorithm, with very few free parameters. The warps are refined using an iterative Levenberg-Marquardt minimization to the mean, based on updating the locations of a small number of points and incorporating a stiffness constraint. This optimization approach is eff
... Show MoreGender classification is a critical task in computer vision. This task holds substantial importance in various domains, including surveillance, marketing, and human-computer interaction. In this work, the face gender classification model proposed consists of three main phases: the first phase involves applying the Viola-Jones algorithm to detect facial images, which includes four steps: 1) Haar-like features, 2) Integral Image, 3) Adaboost Learning, and 4) Cascade Classifier. In the second phase, four pre-processing operations are employed, namely cropping, resizing, converting the image from(RGB) Color Space to (LAB) color space, and enhancing the images using (HE, CLAHE). The final phase involves utilizing Transfer lea
... Show MoreReflections are ubiquitous effects in photos taken through transparent glass mediums, and represent a big problem in photography that impacts severely the performance of computer vision algorithms. Reflection removal is widely needed in daily lives with the prevalence of camera-equipped smart phones, and it is important, but it is a hard problem. This paper addresses the problem of reflection separation from two images taken from different viewpoints in front of a transparent glass medium, and proposes algorithm that exploits the natural image prior (gradient sparsity prior), and robust regression method to remove reflections. The proposed algorithm is tested on real world images, and the quantitative and visual quality comparisons were
... Show MoreThe main objective of this work is to propose a new routing protocol for wireless sensor network employed to serve IoT systems. The routing protocol has to adapt with different requirements in order to enhance the performance of IoT applications. The link quality, node depth and energy are used as metrics to make routing decisions. Comparison with other protocols is essential to show the improvements achieved by this work, thus protocols designed to serve the same purpose such as AODV, REL and LABILE are chosen to compare the proposed routing protocol with. To add integrative and holistic, some of important features are added and tested such as actuating and mobility. These features are greatly required by some of IoT applications and im
... Show MoreThe application of ultrafiltration (UF) and nanofiltration (NF) processes in the handling of raw produced water have been investigated in the present study. Experiments of both ultrafiltration and nanofiltration processes are performed in a laboratory unit, which is operated in a cross-flow pattern. Various types of hollow fiber membranes were utilized in this study such as poly vinyl chloride (PVC) UF membrane, two different polyether sulfone (PES) NF membranes, and poly phenyl sulfone PPSU NF membrane. It was found that the turbidity of the treated water is higher than 95 % by using UF and NF membranes. The chemical oxygen demand COD (160 mg/l) and Oil content (26.8 mg/l) were found after treatment according to the allowable limits set
... Show MoreBusiness organizations have faced many challenges in recent times, most important of which is information technology, because it is widely spread and easy to use. Its use has led to an increase in the amount of data that business organizations deal with an unprecedented manner. The amount of data available through the internet is a problem that many parties seek to find solutions for. Why is it available there in this huge amount randomly? Many expectations have revealed that in 2017, there will be devices connected to the internet estimated at three times the population of the Earth, and in 2015 more than one and a half billion gigabytes of data was transferred every minute globally. Thus, the so-called data mining emerged as a
... Show More