Preferred Language
Articles
/
iRe5Po8BVTCNdQwCz2Wy
Lossy Image Compression Using Hybrid Deep Learning Autoencoder Based On kmean Clusteri
...Show More Authors

Image compression plays an important role in reducing the size and storage of data while increasing the speed of its transmission through the Internet significantly. Image compression is an important research topic for several decades and recently, with the great successes achieved by deep learning in many areas of image processing, especially image compression, and its use is increasing Gradually in the field of image compression. The deep learning neural network has also achieved great success in the field of processing and compressing various images of different sizes. In this paper, we present a structure for image compression based on the use of a Convolutional AutoEncoder (CAE) for deep learning, inspired by the diversity of human eyes' observation of the different colors and features of images. We propose a multi-layer hybrid system for deep learning using the unsupervised CAE architecture and using the color clustering of the K-mean algorithm to compress images and determine their size and color intensity. The system is implemented using Kodak and Challenge on Learned Image Compression (CLIC) dataset for deep learning. Experimental results show that our proposed method is superior to the traditional compression methods of the autoencoder, and the proposed work has better performance in terms of performance speed and quality measures Peak Signal To Noise Ratio (PSNR) and Structural Similarity Index (SSIM) where the results achieved better performance and high efficiency With high compression bit rates and low Mean Squared Error (MSE) rate the results recorded the highest compression ratios that ranged between (0.7117 to 0.8707) for the Kodak dataset and (0.7191 to 0.9930) for CLIC dataset. The system achieved high accuracy and quality in comparison to the error coefficient, which was recorded (0.0126 to reach 0.0003) below, and this system is onsidered the most quality and accurate compared to the methods of deep learning compared to the deep learning methods of the autoencoder

Publication Date
Sat Jul 01 2017
Journal Name
Diyala Journal For Pure Science
Correlated Hierarchical Autoregressive Models Image Compression
...Show More Authors

View Publication
Crossref
Publication Date
Fri Dec 01 2023
Journal Name
Al-khwarizmi Engineering Journal
An Overview of Audio-Visual Source Separation Using Deep Learning
...Show More Authors

    In this article, the research presents a general overview of deep learning-based AVSS (audio-visual source separation) systems. AVSS has achieved exceptional results in a number of areas, including decreasing noise levels, boosting speech recognition, and improving audio quality. The advantages and disadvantages of each deep learning model are discussed throughout the research as it reviews various current experiments on AVSS. The TCD TIMIT dataset (which contains top-notch audio and video recordings created especially for speech recognition tasks) and the Voxceleb dataset (a sizable collection of brief audio-visual clips with human speech) are just a couple of the useful datasets summarized in the paper that can be used to test A

... Show More
View Publication Preview PDF
Scopus (2)
Crossref (1)
Scopus Crossref
Publication Date
Mon Jan 01 2024
Journal Name
Ieee Access
Transfer Learning and Hybrid Deep Convolutional Neural Networks Models for Autism Spectrum Disorder Classification From EEG Signals
...Show More Authors

View Publication
Scopus (18)
Crossref (27)
Scopus Clarivate Crossref
Publication Date
Mon Mar 31 2025
Journal Name
International Journal Of Advanced Technology And Engineering Exploration
Breast cancer survival rate prediction using multimodal deep learning with multigenetic features
...Show More Authors

Breast cancer is a heterogeneous disease characterized by molecular complexity. This research utilized three genetic expression profiles—gene expression, deoxyribonucleic acid (DNA) methylation, and micro ribonucleic acid (miRNA) expression—to deepen the understanding of breast cancer biology and contribute to the development of a reliable survival rate prediction model. During the preprocessing phase, principal component analysis (PCA) was applied to reduce the dimensionality of each dataset before computing consensus features across the three omics datasets. By integrating these datasets with the consensus features, the model's ability to uncover deep connections within the data was significantly improved. The proposed multimodal deep

... Show More
View Publication Preview PDF
Scopus Crossref
Publication Date
Sat Jan 01 2022
Journal Name
Indonesian Journal Of Electrical Engineering And Computer Science
Increasing validation accuracy of a face mask detection by new deep learning model-based classification
...Show More Authors

During COVID-19, wearing a mask was globally mandated in various workplaces, departments, and offices. New deep learning convolutional neural network (CNN) based classifications were proposed to increase the validation accuracy of face mask detection. This work introduces a face mask model that is able to recognize whether a person is wearing mask or not. The proposed model has two stages to detect and recognize the face mask; at the first stage, the Haar cascade detector is used to detect the face, while at the second stage, the proposed CNN model is used as a classification model that is built from scratch. The experiment was applied on masked faces (MAFA) dataset with images of 160x160 pixels size and RGB color. The model achieve

... Show More
View Publication
Scopus (5)
Crossref (4)
Scopus Crossref
Publication Date
Tue Jun 30 2015
Journal Name
Al-khwarizmi Engineering Journal
Multi-Focus Image Fusion Based on Pixel Significance Using Counterlet Transform
...Show More Authors

Abstract

 The objective of image fusion is to merge multiple sources of images together in such a way that the final representation contains higher amount of useful information than any input one.. In this paper, a weighted average fusion method is proposed. It depends on using weights that are extracted from source images using counterlet transform. The extraction method is done by making the approximated transformed coefficients equal to zero, then taking the inverse counterlet transform to get the details of the images to be fused. The performance of the proposed algorithm has been verified on several grey scale and color test  images, and compared with some present methods.

... Show More
View Publication Preview PDF
Publication Date
Sat May 01 2021
Journal Name
Journal Of Physics: Conference Series
Image Steganography using Dynamic Threshold based on Discrete Cosine Transform
...Show More Authors

The art of preventing the detection of hidden information messages is the way that steganography work. Several algorithms have been proposed for steganographic techniques. A major portion of these algorithms is specified for image steganography because the image has a high level of redundancy. This paper proposed an image steganography technique using a dynamic threshold produced by the discrete cosine coefficient. After dividing the green and blue channel of the cover image into 1*3-pixel blocks, check if any bits of green channel block less or equal to threshold then start to store the secret bits in blue channel block, and to increase the security not all bits in the chosen block used to store the secret bits. Firstly, store in the cente

... Show More
View Publication Preview PDF
Scopus (6)
Crossref (6)
Scopus Crossref
Publication Date
Sat Jun 01 2024
Journal Name
Journal Of Information Hiding And Multimedia Signal Processing
Upscale Gray Image using Mixing Transform Generation based on Tensor Product
...Show More Authors

The increased size of grayscale images or upscale plays a central role in various fields such as medicine, satellite imagery, and photography. This paper presents a technique for improving upscaling gray images using a new mixing wavelet generation by tensor product. The proposed technique employs a multi-resolution analysis provided by a new mixing wavelet transform algorithm to decompose the input image into different frequency components. After processing, the low-resolution input image is effectively transformed into a higher-resolution representation by adding a zeroes matrix. Discrete wavelets transform (Daubechies wavelet Haar) as a 2D matrix is used but is mixed using tensor product with another wavelet matrix’s size. MATLAB R2021

... Show More
Preview PDF
Scopus (1)
Scopus
Publication Date
Sat Oct 30 2021
Journal Name
Iraqi Journal Of Science
Small Binary Codebook Design for Image Compression Depending on Rotating Blocks
...Show More Authors

     The searching process using a binary codebook of combined Block Truncation Coding (BTC) method and Vector Quantization (VQ), i.e. a full codebook search for each input image vector to find the best matched code word in the codebook, requires a long time.   Therefore, in this paper, after designing a small binary codebook, we adopted a new method by rotating each binary code word in this codebook into 900 to 2700 step 900 directions. Then, we systematized each code word depending on its angle  to involve four types of binary code books (i.e. Pour when , Flat when  , Vertical when, or Zigzag). The proposed scheme was used for decreasing the time of the coding procedure, with very small distortion per block, by designing s

... Show More
Scopus (5)
Crossref (1)
Scopus Crossref
Publication Date
Tue Apr 02 2019
Journal Name
Artificial Intelligence Research
A three-stage learning algorithm for deep multilayer perceptron with effective weight initialisation based on sparse auto-encoder
...Show More Authors

A three-stage learning algorithm for deep multilayer perceptron (DMLP) with effective weight initialisation based on sparse auto-encoder is proposed in this paper, which aims to overcome difficulties in training deep neural networks with limited training data in high-dimensional feature space. At the first stage, unsupervised learning is adopted using sparse auto-encoder to obtain the initial weights of the feature extraction layers of the DMLP. At the second stage, error back-propagation is used to train the DMLP by fixing the weights obtained at the first stage for its feature extraction layers. At the third stage, all the weights of the DMLP obtained at the second stage are refined by error back-propagation. Network structures an

... Show More
View Publication
Crossref (1)
Crossref