The study presents the modification of the Broyden-Flecher-Goldfarb-Shanno (BFGS) update (H-Version) based on the determinant property of inverse of Hessian matrix (second derivative of the objective function), via updating of the vector s ( the difference between the next solution and the current solution), such that the determinant of the next inverse of Hessian matrix is equal to the determinant of the current inverse of Hessian matrix at every iteration. Moreover, the sequence of inverse of Hessian matrix generated by the method would never approach a near-singular matrix, such that the program would never break before the minimum value of the objective function is obtained. Moreover, the new modification of BFGS update (H-version) preserves the symmetric property and the positive definite property without any condition.
This work includes design, implementation and testing of a microcontroller – based spectrum analyzer system. Both hardware and software structures are built to verify the main functions that are required by such system. Their design utilizes the permissible and available tools to achieve the main functions of the system in such a way to be modularly permitting any adaptation for a specific changing in the application environment. The analysis technique, mainly, depends on the Fourier analysis based methods of spectral analysis with the necessary required preconditioning processes. The software required for waveform analysis has been prepared. The spectrum of the waveform has been displayed, and the instrument accuracy has been checked.
... Show MoreAlthough the Wiener filtering is the optimal tradeoff of inverse filtering and noise smoothing, in the case when the blurring filter is singular, the Wiener filtering actually amplify the noise. This suggests that a denoising step is needed to remove the amplified noise .Wavelet-based denoising scheme provides a natural technique for this purpose .
In this paper a new image restoration scheme is proposed, the scheme contains two separate steps : Fourier-domain inverse filtering and wavelet-domain image denoising. The first stage is Wiener filtering of the input image , the filtered image is inputted to adaptive threshold wavelet
... Show MoreThe need for an efficient method to find the furthermost appropriate document corresponding to a particular search query has become crucial due to the exponential development in the number of papers that are now readily available to us on the web. The vector space model (VSM) a perfect model used in “information retrieval”, represents these words as a vector in space and gives them weights via a popular weighting method known as term frequency inverse document frequency (TF-IDF). In this research, work has been proposed to retrieve the most relevant document focused on representing documents and queries as vectors comprising average term term frequency inverse sentence frequency (TF-ISF) weights instead of representing them as v
... Show MoreThe azo dye brilliant reactive red K-2BP (λmax = 534 nm) is widely used for coloring textiles because of its low-cost and tolerance fastness properties. Wastewaters treatment that contains the dye by conventional ways is usually inadequate due to its resistance to biological and chemical degradation. During this study, the continuous reactor of an advanced oxidation method supported the use of H2O2/sunlight, H2O2/UV, H2O2/TiO2/sunlight, and H2O2/TiO2/UV for decolorization of brilliant reactive red dye from the effluent. The existence of an optimum pH, H2O2 concentration, TiO2 concentration, and d
... Show MoreThe application of the test case prioritization method is a key part of system testing intended to think it through and sort out the issues early in the development stage. Traditional prioritization techniques frequently fail to take into account the complexities of big-scale test suites, growing systems and time constraints, therefore cannot fully fix this problem. The proposed study here will deal with a meta-heuristic hybrid method that focuses on addressing the challenges of the modern time. The strategy utilizes genetic algorithms alongside a black hole as a means to create a smooth tradeoff between exploring numerous possibilities and exploiting the best one. The proposed hybrid algorithm of genetic black hole (HGBH) uses the
... Show MoreTo ensure that a software/hardware product is of sufficient quality and functionality, it is essential to conduct thorough testing and evaluations of the numerous individual software components that make up the application. Many different approaches exist for testing software, including combinatorial testing and covering arrays. Because of the difficulty of dealing with difficulties like a two-way combinatorial explosion, this brings up yet another problem: time. Using client-server architectures, this research introduces a parallel implementation of the TWGH algorithm. Many studies have been conducted to demonstrate the efficiency of this technique. The findings of this experiment were used to determine the increase in speed and co
... Show MoreAbstract This study seeks to deal academically with how the EU treats clandestine immigration, through adopting a purley security approach, based on the European understanding of security threats posed to the security of communities and States in EU at all levels. So they agreed upon criminalizing this threat within the bloc while using repressive tools and steps to limit illegal immigrants flow to European territories. Accordingly, the EU gave the phenomen a security character. So it takes it from low politics level , that of employment and economic field to that of high politics, as a new security problem lying within a new security language embraced by ruling European elites, in other words EU touched on this issue as a speech act emp
... Show MoreThis paper deals with proposing new lifting scheme (HYBRID Algorithm) that is capable of preventing images and documents which are fraud through decomposing there in to the real colors value arrays (red, blue and green) to create retrieval keys for its properties and store it in the database and then check the document originality by retrieve the query image or document through the decomposition described above and compare the predicted color values (retrieval keys) of the query document with those stored in the database. The proposed algorithm has been developed from the two known lifting schemes (Haar and D4) by merging them to find out HYBRID lifting scheme. The validity and accuracy of the proposed algorithm have been ev
... Show MoreClassification of imbalanced data is an important issue. Many algorithms have been developed for classification, such as Back Propagation (BP) neural networks, decision tree, Bayesian networks etc., and have been used repeatedly in many fields. These algorithms speak of the problem of imbalanced data, where there are situations that belong to more classes than others. Imbalanced data result in poor performance and bias to a class without other classes. In this paper, we proposed three techniques based on the Over-Sampling (O.S.) technique for processing imbalanced dataset and redistributing it and converting it into balanced dataset. These techniques are (Improved Synthetic Minority Over-Sampling Technique (Improved SMOTE), Border
... Show More