Compressing an image and reconstructing it without degrading its original quality is one of the challenges that still exist now a day. A coding system that considers both quality and compression rate is implemented in this work. The implemented system applies a high synthetic entropy coding schema to store the compressed image at the smallest size as possible without affecting its original quality. This coding schema is applied with two transform-based techniques, one with Discrete Cosine Transform and the other with Discrete Wavelet Transform. The implemented system was tested with different standard color images and the obtained results with different evaluation metrics have been shown. A comparison was made with some previous related works to test the effectiveness of the implemented coding schema.
Background: Metabolic syndrome is a complex series of metabolic defects, characterized by high levels of serum glucose, hypertension, abdominal obesity, and dyslipidemia. The high mobility group AT-hook1, an architectural transcript factor, affects the homeostasis of glucose. No previous studies have been performed to examine whether HMGA1 can be secreted into the extracellular milieu.
Objectives: this case-control study aimed to examine whether HMGA1 secretes into the extracellular milieu and compares its serum level in two groups of metabolic syndrome (with and without diabetes) and a control group composed of apparently healthy individuals of Iraqi population with different nationalities.
... Show MoreError control schemes became a necessity in network-on-chip (NoC) to improve reliability as the on-chip interconnect errors increase with the continuous shrinking of geometry. Accordingly, many researchers are trying to present multi-bit error correction coding schemes that perform a high error correction capability with the simplest design possible to minimize area and power consumption. A recent work, Multi-bit Error Correcting Coding with Reduced Link Bandwidth (MECCRLB), showed a huge reduction in area and power consumption compared to a well-known scheme, namely, Hamming product code (HPC) with Type-II HARQ. Moreover, the authors showed that the proposed scheme can correct 11 random errors which is considered a high
... Show MoreHueckel edge detector study using binary step edge image is presented. The standard algorithm that Hueckel presented, in his paper without any alteration is adopted. This paper studies a fully analysis for the algorithm efficiency, time consuming and the expected results with slide window size and edge direction. An analysis for its behavior with the changing of the slide window size (disk size) is presented. The best result is acquired when the window size equals to four pixel.
Image Fusion is being used to gather important data from such an input image array and to place it in a single output picture to make it much more meaningful & usable than either of the input images. Image fusion boosts the quality and application of data. The accuracy of the image that has fused depending on the application. It is widely used in smart robotics, audio camera fusion, photonics, system control and output, construction and inspection of electronic circuits, complex computer, software diagnostics, also smart line assembling robots. In this paper provides a literature review of different image fusion techniques in the spatial domain and frequency domain, such as averaging, min-max, block substitution, Intensity-Hue-Saturation(IH
... Show MoreFeatures are the description of the image contents which could be corner, blob or edge. Scale-Invariant Feature Transform (SIFT) extraction and description patent algorithm used widely in computer vision, it is fragmented to four main stages. This paper introduces image feature extraction using SIFT and chooses the most descriptive features among them by blurring image using Gaussian function and implementing Otsu segmentation algorithm on image, then applying Scale-Invariant Feature Transform feature extraction algorithm on segmented portions. On the other hand the SIFT feature extraction algorithm preceded by gray image normalization and binary thresholding as another preprocessing step. SIFT is a strong algorithm and gives more accura
... Show MoreHuman skin detection, which usually performed before image processing, is the method of discovering skin-colored pixels and regions that may be of human faces or limbs in videos or photos. Many computer vision approaches have been developed for skin detection. A skin detector usually transforms a given pixel into a suitable color space and then uses a skin classifier to mark the pixel as a skin or a non-skin pixel. A skin classifier explains the decision boundary of the class of a skin color in the color space based on skin-colored pixels. The purpose of this research is to build a skin detection system that will distinguish between skin and non-skin pixels in colored still pictures. This performed by introducing a metric that measu
... Show MoreThe research shows that the visual image plays an important role when Farzdaq in the issue of aesthetic perception, it enables him to feel a sense of artistic and mental perception to raise astonishment and admiration through his ability to link the optics through the suggestive image to carry us to a new vision imagined full of visual images.