Digital image manipulation has become increasingly prevalent due to the widespread availability of sophisticated image editing tools. In copy-move forgery, a portion of an image is copied and pasted into another area within the same image. The proposed methodology begins with extracting the image's Local Binary Pattern (LBP) algorithm features. Two main statistical functions, Stander Deviation (STD) and Angler Second Moment (ASM), are computed for each LBP feature, capturing additional statistical information about the local textures. Next, a multi-level LBP feature selection is applied to select the most relevant features. This process involves performing LBP computation at multiple scales or levels, capturing textures at different resolutions. By considering features from multiple levels, the detection algorithm can better capture both global and local characteristics of the manipulated regions, enhancing the accuracy of forgery detection. To achieve a high accuracy rate, this paper presents a variety of scenarios based on a machine-learning approach. In Copy-Move detection, artifacts and their properties are used as image features and support Vector Machine (SVM) to determine whether an image is tampered with. The dataset is manipulated to train and test each classifier; the target is to learn the discriminative patterns that detect instances of copy-move forgery. Media Integration and Call Center Forgery (MICC-F2000) were utilized in this paper. Experimental evaluations demonstrate the effectiveness of the proposed methodology in detecting copy-move. The implementation phases in the proposed work have produced encouraging outcomes. In the case of the best-implemented scenario involving multiple trials, the detection stage achieved a copy-move accuracy of 97.8 %.
Currently, with the huge increase in modern communication and network applications, the speed of transformation and storing data in compact forms are pressing issues. Daily an enormous amount of images are stored and shared among people every moment, especially in the social media realm, but unfortunately, even with these marvelous applications, the limited size of sent data is still the main restriction's, where essentially all these applications utilized the well-known Joint Photographic Experts Group (JPEG) standard techniques, in the same way, the need for construction of universally accepted standard compression systems urgently required to play a key role in the immense revolution. This review is concerned with Different
... Show MoreIn this paper the behavior of the quality of the gradient that implemented on an image as a function of noise error is presented. The cross correlation coefficient (ccc) between the derivative of the original image before and after introducing noise error shows dramatic decline compared with the corresponding images before taking derivatives. Mathematical equations have been constructed to control the relation between (ccc) and the noise parameter.
The image caption is the process of adding an explicit, coherent description to the contents of the image. This is done by using the latest deep learning techniques, which include computer vision and natural language processing, to understand the contents of the image and give it an appropriate caption. Multiple datasets suitable for many applications have been proposed. The biggest challenge for researchers with natural language processing is that the datasets are incompatible with all languages. The researchers worked on translating the most famous English data sets with Google Translate to understand the content of the images in their mother tongue. In this paper, the proposed review aims to enhance the understanding o
... Show MoreIn this paper, a method is proposed to increase the compression ratio for the color images by
dividing the image into non-overlapping blocks and applying different compression ratio for these
blocks depending on the importance information of the block. In the region that contain important
information the compression ratio is reduced to prevent loss of the information, while in the
smoothness region which has not important information, high compression ratio is used .The
proposed method shows better results when compared with classical methods(wavelet and DCT).
The biosorption of lead (II) and chromium (III) onto dead anaerobic biomass (DAB) in single and binary systems has been studied using fixed bed adsorber. A general rate multi- component model (GRM) has been utilized to predict the fixed bed breakthrough curves for single and dual- component system. This model considers both external and internal mass transfer resistances as well as axial dispersion with non-liner multi-component isotherm (Langmuir model). The effects of important parameters, such as flow rate, initial concentration and bed height on the behavior of breakthrough curves have been studied. The equilibrium isotherm model parameters such as maximum uptake capacities for lead (II) and chromium (III) were found to be 35.12 and
... Show MoreIn this paper the use of a circular array antenna with adaptive system in conjunction with modified Linearly Constrained Minimum Variance Beam forming (LCMVB) algorithm is proposed to meet the requirement of Angle of Arrival (AOA) estimation in 2-D as well as the Signal to Noise Ratio (SNR) of estimated sources (Three Dimensional 3-D estimation), rather than interference cancelation as it is used for. The proposed system was simulated, tested and compared with the modified Multiple Signal Classification (MUSIC) technique for 2-D estimation. The results show the system has exhibited astonishing results for simultaneously estimating 3-D parameters with accuracy approximately equivalent to the MUSIC technique (for estimating elevation and a
... Show More