Color image compression is a good way to encode digital images by decreasing the number of bits wanted to supply the image. The main objective is to reduce storage space, reduce transportation costs and maintain good quality. In current research work, a simple effective methodology is proposed for the purpose of compressing color art digital images and obtaining a low bit rate by compressing the matrix resulting from the scalar quantization process (reducing the number of bits from 24 to 8 bits) using displacement coding and then compressing the remainder using the Mabel ZF algorithm Welch LZW. The proposed methodology maintains the quality of the reconstructed image. Macroscopic and quantitative experimental results on technical color images show that the proposed methodology gives reconstructed images with a high PSNR value compared to standard image compression techniques.
Continuous turbidimetric analysis (CTA) for a distinctive analytical application by employing a homemade analyser (NAG Dual & Solo 0-180°) which contained two consecutive detection zones (measuring cells 1 & 2) is described. The analyser works based on light-emitting diodes as a light source and a set of solar cells as a light detector for turbidity measurements without needing further fibres or lenses. Formation of a turbid precipitated product with yellow colour due to the reaction between the warfarin and the precipitation reagent (Potassium dichromate) is what the developed method is based on. The CTA method was applied to determine the warfarin in pure form and pharmaceu
Abstract
This study investigates the mechanical compression properties of tin-lead and lead-free alloy spherical balls, using more than 500 samples to identify statistical variability in the properties in each alloy. Isothermal aging was done to study and compare the aging effect on the microstructure and properties.
The results showed significant elastic and plastic anisotropy of tin phase in lead-free tin based solder and that was compared with simulation using a Crystal Plasticity Finite Element (CPEF) method that has the anisotropy of Sn installed. The results and experiments were in good agreement, indicating the range of values expected with anisotropic properties.
Keywords<
... Show MoreIn this paper, several combination algorithms between Partial Update LMS (PU LMS) methods and previously proposed algorithm (New Variable Length LMS (NVLLMS)) have been developed. Then, the new sets of proposed algorithms were applied to an Acoustic Echo Cancellation system (AEC) in order to decrease the filter coefficients, decrease the convergence time, and enhance its performance in terms of Mean Square Error (MSE) and Echo Return Loss Enhancement (ERLE). These proposed algorithms will use the Echo Return Loss Enhancement (ERLE) to control the operation of filter's coefficient length variation. In addition, the time-varying step size is used.The total number of coefficients required was reduced by about 18% , 10% , 6%
... Show MoreAbsence or hypoplasia of the internal carotid artery (ICA) is a rare congenital anomaly that is mostly unilateral and highly associated with other intracranial vascular anomalies, of which saccular aneurysm is the most common. Blood flow to the circulation of the affected side is maintained by collateral pathways, some of which include the anterior communicating artery (Acom) as part of their anatomy. Therefore, temporary clipping during microsurgery on Acom aneurysms in patients with unilateral ICA anomalies could jeopardize these collaterals and place the patient at risk of ischemic damage. In this paper, we review the literature on cases with a unilaterally absent ICA associa
International companies are striving to reduce their costs and increase their profits, and these trends have produced many methods and techniques to achieve these goals. these methods is heuristic and the other Optimization.. The research includes an attempt to adapt some of these techniques in the Iraqi companies, and these techniques are to determine the optimal lot size using the algorithms Wagner-Whitin under the theory of constraints. The research adopted the case study methodology to objectively identify the problem of research, namely determining lot size optimal for each of the products of electronic measurement laboratory in Diyala and in light of the bottlenecks in w
... Show MoreThe paper presents a highly accurate power flow solution, reducing the possibility of ending at local minima, by using Real-Coded Genetic Algorithm (RCGA) with system reduction and restoration. The proposed method (RCGA) is modified to reduce the total computing time by reducing the system in size to that of the generator buses, which, for any realistic system, will be smaller in number, and the load buses are eliminated. Then solving the power flow problem for the generator buses only by real-coded GA to calculate the voltage phase angles, whereas the voltage magnitudes are specified resulted in reduced computation time for the solution. Then the system is restored by calculating the voltages of the load buses in terms
... Show MoreThe predilection for 5G telemedicine networks has piqued the interest of industry researchers and academics. The most significant barrier to global telemedicine adoption is to achieve a secure and efficient transport of patients, which has two critical responsibilities. The first is to get the patient to the nearest hospital as quickly as possible, and the second is to keep the connection secure while traveling to the hospital. As a result, a new network scheme has been suggested to expand the medical delivery system, which is an agile network scheme to securely redirect ambulance motorbikes to the nearest hospital in emergency cases. This research provides a secured and efficient telemedicine transport strategy compatible with the
... Show MorePlagiarism is becoming more of a problem in academics. It’s made worse by the ease with which a wide range of resources can be found on the internet, as well as the ease with which they can be copied and pasted. It is academic theft since the perpetrator has ”taken” and presented the work of others as his or her own. Manual detection of plagiarism by a human being is difficult, imprecise, and time-consuming because it is difficult for anyone to compare their work to current data. Plagiarism is a big problem in higher education, and it can happen on any topic. Plagiarism detection has been studied in many scientific articles, and methods for recognition have been created utilizing the Plagiarism analysis, Authorship identification, and
... Show More