The image caption is the process of adding an explicit, coherent description to the contents of the image. This is done by using the latest deep learning techniques, which include computer vision and natural language processing, to understand the contents of the image and give it an appropriate caption. Multiple datasets suitable for many applications have been proposed. The biggest challenge for researchers with natural language processing is that the datasets are incompatible with all languages. The researchers worked on translating the most famous English data sets with Google Translate to understand the content of the images in their mother tongue. In this paper, the proposed review aims to enhance the understanding of image captioning strategies and to survey previous research related to image captioning while examining the most popular databases in different languages, mostly English, translating into other languages using the latest models for describing images, summarizing evaluation measures, and comparing them.
This research discusses the subject of identity in the urban environment as it attempts to answer a number of questions that come with the concept of identity. The first of these questions: What is identity? Can a definition or conceptual framework be developed for identity? What about individual, collective, cultural, ethnic, political and regional identity? Is there a definition of identity in the urban environment in particular? If there is a definition of identity, what about social mobility responsible for social change? How can we see identity through this kinetics? Can we assume that identity in the urban environment has a variable structure or is of variable shape with a more stable structure? Can we determine the spatial-tempora
... Show MoreDigital image manipulation has become increasingly prevalent due to the widespread availability of sophisticated image editing tools. In copy-move forgery, a portion of an image is copied and pasted into another area within the same image. The proposed methodology begins with extracting the image's Local Binary Pattern (LBP) algorithm features. Two main statistical functions, Stander Deviation (STD) and Angler Second Moment (ASM), are computed for each LBP feature, capturing additional statistical information about the local textures. Next, a multi-level LBP feature selection is applied to select the most relevant features. This process involves performing LBP computation at multiple scales or levels, capturing textures at different
... Show MoreThis paper is concerned with the design and implementation of an image compression method based on biorthogonal tap-9/7 discrete wavelet transform (DWT) and quadtree coding method. As a first step the color correlation is handled using YUV color representation instead of RGB. Then, the chromatic sub-bands are downsampled, and the data of each color band is transformed using wavelet transform. The produced wavelet sub-bands are quantized using hierarchal scalar quantization method. The detail quantized coefficient is coded using quadtree coding followed by Lempel-Ziv-Welch (LZW) encoding. While the approximation coefficients are coded using delta coding followed by LZW encoding. The test results indicated that the compression results are com
... Show MoreThe digital camera which contain light unit inside it is useful with low illumination but not for high. For different intensity; the quality of the image will not stay good but it will have dark or low intensity so we can not change the contrast and the intensity in order to increase the losses information in the bright and the dark regions. . In this search we study the regular illumination on the images using the tungsten light by changing the intensities. The result appears that the tungsten light gives nearly far intensity for the three color bands(RGB) and the illuminated band(L).the result depend on the statistical properties which represented by the voltage ,power and intensities and the effect of this parameter on the digital
... Show MoreImproved Merging Multi Convolutional Neural Networks Framework of Image Indexing and Retrieval