Image pattern classification is considered a significant step for image and video processing.Although various image pattern algorithms have been proposed so far that achieved adequate classification,achieving higher accuracy while reducing the computation time remains challenging to date. A robust imagepattern classification method is essential to obtain the desired accuracy. This method can be accuratelyclassify image blocks into plain, edge, and texture (PET) using an efficient feature extraction mechanism.Moreover, to date, most of the existing studies are focused on evaluating their methods based on specificorthogonal moments, which limits the understanding of their potential application to various DiscreteOrthogonal Moments (DOMs). Therefore, finding a fast PET classification method that accurately clas-sify image pattern is crucial. To this end, this paper proposes a new scheme for accurate and fast imagepattern classification using an efficient DOM. To reduce the computational complexity of feature extraction,an election mechanism is proposed to reduce the number of processed block patterns. In addition, supportvector machine is used to classify the extracted features for different block patterns. The proposed scheme isevaluated by comparing the accuracy of the proposed method with the accuracy achieved by state-of-the-artmethods. In addition, we compare the performance of the proposed method based on different DOMs toget the robust one. The results show that the proposed method achieves the highest classification accuracycompared with the existing methods in all the scenarios considered
This article aims to determine the time-dependent heat coefficient together with the temperature solution for a type of semi-linear time-fractional inverse source problem by applying a method based on the finite difference scheme and Tikhonov regularization. An unconditionally stable implicit finite difference scheme is used as a direct (forward) solver. While by the MATLAB routine lsqnonlin from the optimization toolbox, the inverse problem is reformulated as nonlinear least square minimization and solved efficiently. Since the problem is generally incorrect or ill-posed that means any error inclusion in the input data will produce a large error in the output data. Therefore, the Tikhonov regularization technique is applie
... Show MoreThe laser micro-cutting process is the most widely commonly applied machining process which can be applied to practically all metallic and non-metallic materials. While this had challenges in cutting quality criteria such as geometrical precision, surface quality and numerous others. This article investigates the laser micro-cutting of PEEK composite material using nano-fiber laser, due to their significant importunity and efficiency of laser in various manufacturing processes. Design of experiential tool based on Response Surface Methodology (RSM)-Central Composite Design (CCD) used to generate the statistical model. This method was employed to analysis the influence of parameters including laser speed,
... Show MoreBiometrics is widely used with security systems nowadays; each biometric modality can be useful and has distinctive properties that provide uniqueness and ambiguity for security systems especially in communication and network technologies. This paper is about using biometric features of fingerprint, which is called (minutiae) to cipher a text message and ensure safe arrival of data at receiver end. The classical cryptosystems (Caesar, Vigenère, etc.) became obsolete methods for encryption because of the high-performance machines which focusing on repetition of the key in their attacks to break the cipher. Several Researchers of cryptography give efforts to modify and develop Vigenère cipher by enhancing its weaknesses.
... Show MoreTo achieve safe security to transfer data from the sender to receiver, cryptography is one way that is used for such purposes. However, to increase the level of data security, DNA as a new term was introduced to cryptography. The DNA can be easily used to store and transfer the data, and it becomes an effective procedure for such aims and used to implement the computation. A new cryptography system is proposed, consisting of two phases: the encryption phase and the decryption phase. The encryption phase includes six steps, starting by converting plaintext to their equivalent ASCII values and converting them to binary values. After that, the binary values are converted to DNA characters and then converted to their equivalent complementary DN
... Show MoreThe confirming of security and confidentiality of multimedia data is a serious challenge through the growing dependence on digital communication. This paper offers a new image cryptography based on the Chebyshev chaos polynomials map, via employing the randomness characteristic of chaos concept to improve security. The suggested method includes block shuffling, dynamic offset chaos key production, inter-layer XOR, and block 90 degree rotations to disorder the correlations intrinsic in image. The method is aimed for efficiency and scalability, accomplishing complexity order for n-pixels over specific cipher rounds. The experiment outcomes depict great resistant to cryptanalysis attacks, containing statistical, differential and brut
... Show MoreChaotic systems have been proved to be useful and effective for cryptography. Through this work, a new Feistel cipher depend upon chaos systems and Feistel network structure with dynamic secret key size according to the message size have been proposed. Compared with the classical traditional ciphers like Feistel-based structure ciphers, Data Encryption Standards (DES), is the common example of Feistel-based ciphers, the process of confusion and diffusion, will contains the dynamical permutation choice boxes, dynamical substitution choice boxes, which will be generated once and hence, considered static,
While using chaotic maps, in the suggested system, called
The efficiency of the Honeywords approach has been proven to be a significant tool for boosting password security. The suggested system utilizes the Meerkat Clan Algorithm (MCA) in conjunction with WordNet to produce honeywords, thereby enhancing the level of password security. The technique of generating honeywords involves data sources from WordNet, which contributes to the improvement of authenticity and diversity in the honeywords. The method encompasses a series of consecutive stages, which include the tokenization of passwords, the formation of alphabet tokens using the Meerkat Clan Algorithm (MCA), the handling of digit tokens, the creation of unique character tokens, and the consolidation of honeywords. The optimization of t
... Show MoreThis paper presents the design of a longitudinal controller for an autonomous unmanned aerial vehicle (UAV). This paper proposed the dual loop (inner-outer loop) control based on the intelligent algorithm. The inner feedback loop controller is a Linear Quadratic Regulator (LQR) to provide robust (adaptive) stability. In contrast, the outer loop controller is based on Fuzzy-PID (Proportional, Integral, and Derivative) algorithm to provide reference signal tracking. The proposed dual controller is to control the position (altitude) and velocity (airspeed) of an aircraft. An adaptive Unscented Kalman Filter (AUKF) is employed to track the reference signal and is decreased the Gaussian noise. The mathematical model of aircraft
... Show More<p>In this paper, a simple color image compression system has been proposed using image signal decomposition. Where, the RGB image color band is converted to the less correlated YUV color model and the pixel value (magnitude) in each band is decomposed into 2-values; most and least significant. According to the importance of the most significant value (MSV) that influenced by any simply modification happened, an adaptive lossless image compression system is proposed using bit plane (BP) slicing, delta pulse code modulation (Delta PCM), adaptive quadtree (QT) partitioning followed by an adaptive shift encoder. On the other hand, a lossy compression system is introduced to handle the least significant value (LSV), it is based on
... Show More