The focus of this article is to add a new class of rank one of modified Quasi-Newton techniques to solve the problem of unconstrained optimization by updating the inverse Hessian matrix with an update of rank 1, where a diagonal matrix is the first component of the next inverse Hessian approximation, The inverse Hessian matrix is generated by the method proposed which is symmetric and it satisfies the condition of modified quasi-Newton, so the global convergence is retained. In addition, it is positive definite that guarantees the existence of the minimizer at every iteration of the objective function. We use the program MATLAB to solve an algorithm function to introduce the feasibility of the proposed procedure. Various numerical examples are given`.
In this paper, the methods of weighted residuals: Collocation Method (CM), Least Squares Method (LSM) and Galerkin Method (GM) are used to solve the thin film flow (TFF) equation. The weighted residual methods were implemented to get an approximate solution to the TFF equation. The accuracy of the obtained results is checked by calculating the maximum error remainder functions (MER). Moreover, the outcomes were examined in comparison with the 4th-order Runge-Kutta method (RK4) and good agreements have been achieved. All the evaluations have been successfully implemented by using the computer system Mathematica®10.
The continuous increases in the size of current telecommunication infrastructures have led to the many challenges that existing algorithms face in underlying optimization. The unrealistic assumptions and low efficiency of the traditional algorithms make them unable to solve large real-life problems at reasonable times.
The use of approximate optimization techniques, such as adaptive metaheuristic algorithms, has become more prevalent in a diverse research area. In this paper, we proposed the use of a self-adaptive differential evolution (jDE) algorithm to solve the radio network planning (RNP) problem in the context of the upcoming generation 5G. The experimental results prove the jDE with best vecto
This paper presents a linear fractional programming problem (LFPP) with rough interval coefficients (RICs) in the objective function. It shows that the LFPP with RICs in the objective function can be converted into a linear programming problem (LPP) with RICs by using the variable transformations. To solve this problem, we will make two LPP with interval coefficients (ICs). Next, those four LPPs can be constructed under these assumptions; the LPPs can be solved by the classical simplex method and used with MS Excel Solver. There is also argumentation about solving this type of linear fractional optimization programming problem. The derived theory can be applied to several numerical examples with its details, but we show only two examples
... Show MoreIn this paper, new approach based on coupled Laplace transformation with decomposition method is proposed to solve type of partial differential equation. Then it’s used to find the accurate solution for heat equation with initial conditions. Four examples introduced to illustrate the accuracy, efficiency of suggested method. The practical results show the importance of suggested method for solve differential equations with high accuracy and easy implemented.
The conjugate coefficient optimal is the very establishment of a variety of conjugate gradient methods. This paper proposes a new class coefficient of conjugate gradient (CG) methods for impulse noise removal, which is based on the quadratic model. Our proposed method ensures descent independent of the accuracy of the line search and it is globally convergent under some conditions, Numerical experiments are also presented for the impulse noise removal in images.
This paper investigates some exact and local search methods to solve the traveling salesman problem. The Branch and Bound technique (BABT) is proposed, as an exact method, with two models. In addition, the classical Genetic Algorithm (GA) and Simulated Annealing (SA) are discussed and applied as local search methods. To improve the performance of GA we propose two kinds of improvements for GA; the first is called improved GA (IGA) and the second is Hybrid GA (HGA).
The IGA gives best results than GA and SA, while the HGA is the best local search method for all within a reasonable time for 5 ≤ n ≤ 2000, where n is the number of visited cities. An effective method of reducing the size of the TSP matrix was proposed with
... Show MoreHuman beings are greatly inspired by nature. Nature has the ability to solve very complex problems in its own distinctive way. The problems around us are becoming more and more complex in the real time and at the same instance our mother nature is guiding us to solve these natural problems. Nature gives some of the logical and effective ways to find solutions to these problems. Nature acts as an optimized source for solving the complex problems. Decomposition is a basic strategy in traditional multi-objective optimization. However, it has not yet been widely used in multi-objective evolutionary optimization.
Although computational strategies for taking care of Multi-objective Optimization Problems (MOPs) h
... Show MoreThis paper has the interest of finding the approximate solution (APPS) of a nonlinear variable coefficients hyperbolic boundary value problem (NOLVCHBVP). The given boundary value problem is written in its discrete weak form (WEFM) and proved have a unique solution, which is obtained via the mixed Galerkin finite element with implicit method that reduces the problem to solve the Galerkin nonlinear algebraic system (GNAS). In this part, the predictor and the corrector techniques (PT and CT, respectively) are proved at first convergence and then are used to transform the obtained GNAS to a linear GLAS . Then the GLAS is solved using the Cholesky method (ChMe). The stability and the convergence of the method are stud
... Show More