In this paper, a method is proposed to increase the compression ratio for the color images by
dividing the image into non-overlapping blocks and applying different compression ratio for these
blocks depending on the importance information of the block. In the region that contain important
information the compression ratio is reduced to prevent loss of the information, while in the
smoothness region which has not important information, high compression ratio is used .The
proposed method shows better results when compared with classical methods(wavelet and DCT).
In this paper, the combined source coding with Multi Carrier Code Division Multiple Access (MC-CDMA) system is proposed, where the transmission of the compressed image produced from source coding through Additive White Gaussian Noise (AWGN) channel for a single user and multi users. In which the (MC-CDMA) system removes Inter Symbol Interference (ISI) and Inter Carrier Interference (ICI). The hybrid compression system of Discrete Cosine Transform (DCT) and predictive coding (PC) technique are integrated as a source coding. The simulation results indicates that the transmission system of a single user was much better than the transmission system of multi users. When the number of users increased, the Bit Error Rate (BER) increased. For a
... Show MoreThe Internet is providing vital communications between millions of individuals. It is also more and more utilized as one of the commerce tools; thus, security is of high importance for securing communications and protecting vital information. Cryptography algorithms are essential in the field of security. Brute force attacks are the major Data Encryption Standard attacks. This is the main reason that warranted the need to use the improved structure of the Data Encryption Standard algorithm. This paper proposes a new, improved structure for Data Encryption Standard to make it secure and immune to attacks. The improved structure of Data Encryption Standard was accomplished using standard Data Encryption Standard with a new way of two key gene
... Show MoreThe complexity of multimedia contents is significantly increasing in the current world. This leads to an exigent demand for developing highly effective systems to satisfy human needs. Until today, handwritten signature considered an important means that is used in banks and businesses to evidence identity, so there are many works tried to develop a method for recognition purpose. This paper introduced an efficient technique for offline signature recognition depending on extracting the local feature by utilizing the haar wavelet subbands and energy. Three different sets of features are utilized by partitioning the signature image into non overlapping blocks where different block sizes are used. CEDAR signature database is used as a dataset f
... Show MoreThis article aim to estimate the Return Stock Rate of the private banking sector, with two banks, by adopting a Partial Linear Model based on the Arbitrage Pricing Model (APT) theory, using Wavelet and Kernel Smoothers. The results have proved that the wavelet method is the best. Also, the results of the market portfolio impact and inflation rate have proved an adversely effectiveness on the rate of return, and direct impact of the money supply.
Entropy define as uncertainty measure has been transfared by using the cumulative distribution function and reliability function for the Burr type – xii. In the case of data which suffer from volatility to build a model the probability distribution on every failure of a sample after achieving limitations function, probabilistic distribution. Has been derived formula probability distribution of the new transfer application entropy on the probability distribution of continuous Burr Type-XII and tested a new function and found that it achieved the conditions function probability, been derived mean and function probabilistic aggregate in order to be approved in the generation of data for the purpose of implementation of simulation
... Show MoreEncryption of data is translating data to another shape or symbol which enables people only with an access to the secret key or a password that can read it. The data which are encrypted are generally referred to as cipher text, while data which are unencrypted are known plain text. Entropy can be used as a measure which gives the number of bits that are needed for coding the data of an image. As the values of pixel within an image are dispensed through further gray-levels, the entropy increases. The aim of this research is to compare between CAST-128 with proposed adaptive key and RSA encryption methods for video frames to determine the more accurate method with highest entropy. The first method is achieved by applying the "CAST-128" and
... Show MoreThis paper presents the application of a framework of fast and efficient compressive sampling based on the concept of random sampling of sparse Audio signal. It provides four important features. (i) It is universal with a variety of sparse signals. (ii) The number of measurements required for exact reconstruction is nearly optimal and much less then the sampling frequency and below the Nyquist frequency. (iii) It has very low complexity and fast computation. (iv) It is developed on the provable mathematical model from which we are able to quantify trade-offs among streaming capability, computation/memory requirement and quality of reconstruction of the audio signal. Compressed sensing CS is an attractive compression scheme due to its uni
... Show MoreThe effectiveness of detecting and matching of image features using multiple views of a specified scene using dynamic scene analysis is considered to be a critical first step for many applications in computer vision image processing. The Scale invariant feature transform (SIFT) can be applied very successfully of typical images captured by a digital camera.
In this paper, firstly the SIFT and its variants are systematically analyzed. Then, the performances are evaluated in many situations: change in rotation, change in blurs, change in scale and change in illumination. The outcome results show that each algorithm has its advantages when compared with other algorithms