Objective: Evaluation the national standards for exposure to chemical materials and dusts in The State
Company for Drugs Industry in Samarra.
Methodology: A descriptive evaluation design is employed through the present study from 25th May 2011
to 30th November 2011 in order to evaluate the national standards for exposure chemical materials and dusts
in The State Company for Drugs Industry in Samarra. A purposive (non-probability) sample is selected for the
study which includes (110) workers from the State Company for Drugs Industry in Samarra. Data were
gathered through the workers` interviewed according to the nature of work that they perform. The evaluation
questionnaire comprised of three parts which include the workers` demographic characteristic and other two
part which concern the national standards for exposure to chemical materials and dusts in workplace.
Reliability and validity of this tool is determined through application of a pilot study and panel of experts. Data
were analyzed through the application of descriptive statistical (frequencies and percentages) and inferential
statistical (mean of score).
Results: The findings of the study present that the national standards for exposure chemical materials and
standards for exposure to dusts that are applicable in the workplace, can be adopted as national
standards. So, there is no significant impact of occupational hazards that may affect workers and work
environment as a result of applicable of this standards.
Recommendations: The study recommends that increase awareness, training and health education
programs should be provided for all workers regularly and periodically in order to help them comply
with standards for exposure chemical materials and standards for exposure to dusts in order to avoid
hazards that affecting their health and work environment.
Among the metaheuristic algorithms, population-based algorithms are an explorative search algorithm superior to the local search algorithm in terms of exploring the search space to find globally optimal solutions. However, the primary downside of such algorithms is their low exploitative capability, which prevents the expansion of the search space neighborhood for more optimal solutions. The firefly algorithm (FA) is a population-based algorithm that has been widely used in clustering problems. However, FA is limited in terms of its premature convergence when no neighborhood search strategies are employed to improve the quality of clustering solutions in the neighborhood region and exploring the global regions in the search space. On the
... Show MoreScheduling Timetables for courses in the big departments in the universities is a very hard problem and is often be solved by many previous works although results are partially optimal. This work implements the principle of an evolutionary algorithm by using genetic theories to solve the timetabling problem to get a random and full optimal timetable with the ability to generate a multi-solution timetable for each stage in the collage. The major idea is to generate course timetables automatically while discovering the area of constraints to get an optimal and flexible schedule with no redundancy through the change of a viable course timetable. The main contribution in this work is indicated by increasing the flexibility of generating opti
... Show MoreIn this research estimated the parameters of Gumbel distribution Type 1 for Maximum values through the use of two estimation methods:- Moments (MoM) and Modification Moments(MM) Method. the Simulation used for comparison between each of the estimation methods to reach the best method to estimate the parameters where the simulation was to generate random data follow Gumbel distributiondepending on three models of the real values of the parameters for different sample sizes with samples of replicate (R=500).The results of the assessment were put in tables prepared for the purpose of comparison, which made depending on the mean squares error (MSE).
The need for information web-searching is needed by many users nowadays. They use the search engines to input their query or question and wait for the answer or best search results. As results to user query the search engines many times may be return irrelevant pages or not related to information need. This paper presents a proposed model to provide the user with efficient and effective result through search engine, based on modified chicken swarm algorithm and cosine similarity to eliminate and delete irrelevant pages(outliers) from the ranked list results, and to improve the results of the user's query . The proposed model is applied to Arabic dataset and use the ZAD corpus dataset for 27
... Show MoreDecision making is vital and important activity in field operations research ,engineering ,administration science and economic science with any industrial or service company or organization because the core of management process as well as improve him performance . The research includes decision making process when the objective function is fraction function and solve models fraction programming by using some fraction programming methods and using goal programming method aid programming ( win QSB )and the results explain the effect use the goal programming method in decision making process when the objective function is
fraction .
In this paper Volterra Runge-Kutta methods which include: method of order two and four will be applied to general nonlinear Volterra integral equations of the second kind. Moreover we study the convergent of the algorithms of Volterra Runge-Kutta methods. Finally, programs for each method are written in MATLAB language and a comparison between the two types has been made depending on the least square errors.
In this work, results from an optical technique (laser speckle technique) for measuring surface roughness was done by using statistical properties of speckle pattern from the point of view of computer image texture analysis. Four calibration relationships were used to cover wide range of measurement with the same laser speckle technique. The first one is based on intensity contrast of the speckle, the second is based on analysis of speckle binary image, the third is on size of speckle pattern spot, and the latest one is based on characterization of the energy feature of the gray level co-occurrence matrices for the speckle pattern. By these calibration relationships surface roughness of an object surface can be evaluated within the
... Show More<span lang="EN-US">The use of bio-signals analysis in human-robot interaction is rapidly increasing. There is an urgent demand for it in various applications, including health care, rehabilitation, research, technology, and manufacturing. Despite several state-of-the-art bio-signals analyses in human-robot interaction (HRI) research, it is unclear which one is the best. In this paper, the following topics will be discussed: robotic systems should be given priority in the rehabilitation and aid of amputees and disabled people; second, domains of feature extraction approaches now in use, which are divided into three main sections (time, frequency, and time-frequency). The various domains will be discussed, then a discussion of e
... Show MoreHiding technique for dynamic encryption text using encoding table and symmetric encryption method (AES algorithm) is presented in this paper. The encoding table is generated dynamically from MSB of the cover image points that used as the first phase of encryption. The Harris corner point algorithm is applied on cover image to generate the corner points which are used to generate dynamic AES key to second phase of text encryption. The embedded process in the LSB for the image pixels except the Harris corner points for more robust. Experimental results have demonstrated that the proposed scheme have embedding quality, error-free text recovery, and high value in PSNR.
Malware represents one of the dangerous threats to computer security. Dynamic analysis has difficulties in detecting unknown malware. This paper developed an integrated multi – layer detection approach to provide more accuracy in detecting malware. User interface integrated with Virus Total was designed as a first layer which represented a warning system for malware infection, Malware data base within malware samples as a second layer, Cuckoo as a third layer, Bull guard as a fourth layer and IDA pro as a fifth layer. The results showed that the use of fifth layers was better than the use of a single detector without merging. For example, the efficiency of the proposed approach is 100% compared with 18% and 63% of Virus Total and Bel
... Show More