Steganography is a technique of concealing secret data within other quotidian files of the same or different types. Hiding data has been essential to digital information security. This work aims to design a stego method that can effectively hide a message inside the images of the video file. In this work, a video steganography model has been proposed through training a model to hiding video (or images) within another video using convolutional neural networks (CNN). By using a CNN in this approach, two main goals can be achieved for any steganographic methods which are, increasing security (hardness to observed and broken by used steganalysis program), this was achieved in this work as the weights and architecture are randomized. Thus, the exact way by which the network will hide the information is unable to be known to anyone who does not have the weights. The second goal is to increase hiding capacity, which has been achieved by using CNN as a strategy to make decisions to determine the best areas that are redundant and, as a result, gain more size to be hidden. Furthermore, In the proposed model, CNN is concurrently trained to generate the revealing and hiding processes, and it is designed to work as a pair mainly. This model has a good strategy for the patterns of images, which assists to make decisions to determine which is the parts of the cover image should be redundant, as well as more pixels are hidden there. The CNN implementation can be done by using Keras, along with tensor flow backend. In addition, random RGB images from the "ImageNet dataset" have been used for training the proposed model (About 45000 images of size (256x256)). The proposed model has been trained by CNN using random images taken from the database of ImageNet and can work on images taken from a wide range of sources. By saving space on an image by removing redundant areas, the quantity of hidden data can be raised (improve capacity). Since the weights and model architecture are randomized, the actual method in which the network will hide the data can't be known to anyone who does not have the weights. Furthermore, additional block-shuffling is incorporated as an encryption method to improved security; also, the image enhancement methods are used to improving the output quality. From results, the proposed method has achieved high-security level, high embedding capacity. In addition, the result approves that the system achieves good results in visibility and attacks, in which the proposed method successfully tricks observer and the steganalysis program.
Community detection is an important and interesting topic for better understanding and analyzing complex network structures. Detecting hidden partitions in complex networks is proven to be an NP-hard problem that may not be accurately resolved using traditional methods. So it is solved using evolutionary computation methods and modeled in the literature as an optimization problem. In recent years, many researchers have directed their research efforts toward addressing the problem of community structure detection by developing different algorithms and making use of single-objective optimization methods. In this study, we have continued that research line by improving the Particle Swarm Optimization (PSO) algorithm using a
... Show MoreThe speaker identification is one of the fundamental problems in speech processing and voice modeling. The speaker identification applications include authentication in critical security systems and the accuracy of the selection. Large-scale voice recognition applications are a major challenge. Quick search in the speaker database requires fast, modern techniques and relies on artificial intelligence to achieve the desired results from the system. Many efforts are made to achieve this through the establishment of variable-based systems and the development of new methodologies for speaker identification. Speaker identification is the process of recognizing who is speaking using the characteristics extracted from the speech's waves like pi
... Show MoreIn real situations all observations and measurements are not exact numbers but more or less non-exact, also called fuzzy. So, in this paper, we use approximate non-Bayesian computational methods to estimate inverse Weibull parameters and reliability function with fuzzy data. The maximum likelihood and moment estimations are obtained as non-Bayesian estimation. The maximum likelihood estimators have been derived numerically based on two iterative techniques namely “Newton-Raphson†and the “Expectation-Maximization†techniques. In addition, we provide compared numerically through Monte-Carlo simulation study to obtained estimates of the parameters and reliability function i
... Show MoreThis research includes the use of an artificial intelligence algorithm, which is one of the algorithms of biological systems which is the algorithm of genetic regulatory networks (GRNs), which is a dynamic system for a group of variables representing space within time. To construct this biological system, we use (ODEs) and to analyze the stationarity of the model we use Euler's method. And through the factors that affect the process of gene expression in terms of inhibition and activation of the transcription process on DNA, we will use TF transcription factors. The current research aims to use the latest methods of the artificial intelligence algorithm. To apply Gene Regulation Networks (GRNs), we used a progr
... Show MoreMany authors investigated the problem of the early visibility of the new crescent moon after the conjunction and proposed many criteria addressing this issue in the literature. This article presented a proposed criterion for early crescent moon sighting based on a deep-learned pattern recognizer artificial neural network (ANN) performance. Moon sight datasets were collected from various sources and used to learn the ANN. The new criterion relied on the crescent width and the arc of vision from the edge of the crescent bright limb. The result of that criterion was a control value indicating the moon's visibility condition, which separated the datasets into four regions: invisible, telescope only, probably visible, and certai
... Show MoreThe calculation of the oil density is more complex due to a wide range of pressuresand temperatures, which are always determined by specific conditions, pressure andtemperature. Therefore, the calculations that depend on oil components are moreaccurate and easier in finding such kind of requirements. The analyses of twenty liveoil samples are utilized. The three parameters Peng Robinson equation of state istuned to get match between measured and calculated oil viscosity. The Lohrenz-Bray-Clark (LBC) viscosity calculation technique is adopted to calculate the viscosity of oilfrom the given composition, pressure and temperature for 20 samples. The tunedequation of state is used to generate oil viscosity values for a range of temperatu
... Show MoreThe finishing operation of the electrochemical finishing technology (ECF) for tube of steel was investigated In this study. Experimental procedures included qualitative
and quantitative analyses for surface roughness and material removal. Qualitative analyses utilized finishing optimization of a specific specimen in various design and operating conditions; value of gap from 0.2 to 10mm, flow rate of electrolytes from 5 to 15liter/min, finishing time from 1 to 4min and the applied voltage from 6 to 12v, to find out the value of surface roughness and material removal at each electrochemical state. From the measured material removal for each process state was used to verify the relationship with finishing time of work piece. Electrochemi