The application of the test case prioritization method is a key part of system testing intended to think it through and sort out the issues early in the development stage. Traditional prioritization techniques frequently fail to take into account the complexities of big-scale test suites, growing systems and time constraints, therefore cannot fully fix this problem. The proposed study here will deal with a meta-heuristic hybrid method that focuses on addressing the challenges of the modern time. The strategy utilizes genetic algorithms alongside a black hole as a means to create a smooth tradeoff between exploring numerous possibilities and exploiting the best one. The proposed hybrid algorithm of genetic black hole (HGBH) uses the capabilities of considering the imperatives such as code coverage, fault finding rate and execution time from search algorithms in our hybrid approach to refine test cases considerations repetitively. The strategy accomplished this by putting experiments on a large-scale project of industrial software developed. The hybrid meta-heuristic technique ends up being better than the routine techniques. It helps in higher code coverage, which, in turn, enables to detect crucial defects at an early stage and also to allocate the testing resources in a better way. In particular, the best APFD value was 0.9321, which was achieved in 6 generations with 4.879 seconds the value to which the computer was run. Besides these, , the approach resulted in the mean value of APFD as 0.9247 and 0.9302 seconds which took from 10.509 seconds to 30.372 seconds. The carried out experiment proves the feasibility of this approach in implementing complex systems and consistently detecting the changes, enabling it to adapt to rapidly changing systems. In the end, this research provides us with a new hybrid meta-heuristic way of test case prioritization and optimization, which, in turn, helps to tackle the obstacles caused by large-scale test cases and constantly changing systems.
In this work, the performance of the receiver in a quantum cryptography system based on BB84 protocol is scaled by calculating the Quantum Bit Error Rate (QBER) of the receiver. To apply this performance test, an optical setup was arranged and a circuit was designed and implemented to calculate the QBER. This electronic circuit is used to calculate the number of counts per second generated by the avalanche photodiodes set in the receiver. The calculated counts per second are used to calculate the QBER for the receiver that gives an indication for the performance of the receiver. Minimum QBER, 6%, was obtained with avalanche photodiode excess voltage equals to 2V and laser diode power of 3.16 nW at avalanche photodiode temperature of -10
... Show MoreNumeral recognition is considered an essential preliminary step for optical character recognition, document understanding, and others. Although several handwritten numeral recognition algorithms have been proposed so far, achieving adequate recognition accuracy and execution time remain challenging to date. In particular, recognition accuracy depends on the features extraction mechanism. As such, a fast and robust numeral recognition method is essential, which meets the desired accuracy by extracting the features efficiently while maintaining fast implementation time. Furthermore, to date most of the existing studies are focused on evaluating their methods based on clean environments, thus limiting understanding of their potential a
... Show MoreRock type identification is very important task in Reservoir characterization in order to constrict robust reservoir models. There are several approaches have been introduced to define the rock type in reservoirs and each approach should relate the geological and petrophysical properties, such that each rock type is proportional to a unique hydraulic flow unit. A hydraulic flow unit is a reservoir zone that is laterally and vertically has similar flow and bedding characteristics. According to effect of rock type in reservoir performance, many empirical and statistical approaches introduced. In this paper Cluster Analysis technique is used to identify the rock groups in tertiary reservoir for Khabaz oil field by analyses variation o
... Show MoreThe aim of this research is to construct a three-dimensional maritime transport model to transport nonhomogeneous goods (k) and different transport modes (v) from their sources (i) to their destinations (j), while limiting the optimum quantities v ijk x to be transported at the lowest possible cost v ijk c and time v ijk t using the heuristic algorithm, Transport problems have been widely studied in computer science and process research and are one of the main problems of transport problems that are usually used to reduce the cost or times of transport of goods with a number of sources and a number of destinations and by means of transport to meet the conditions of supply and demand. Transport models are a key tool in logistics an
... Show MoreStandards play a vital role in documenting the values of new test results in the form of tables. They are one of the basic requirements that the standardization process aims for as a complement to standardizing test procedures, and contribute to knowing the current reality of the student. The degree of readiness and level as a result of practicing different exercises for sports activities, in addition to the possibility of adopting it for comparison with his group or similar groups, classification, prediction and selection. Developing the skill of handling the football in the educational field is an important matter for achieving distinguished performance among students. This skill requires a level of accuracy, speed and control, and
... Show MoreDrilling well design optimization reduces total Authorization for Expenditures (AFE) by decreasing well constructing time and expense. Well design is not a constant pattern during the life cycle of the field. It should be optimized by continuous improvements for all aspects of redesigning the well depending on the actual field conditions and problems. The core objective of this study is to deliver a general review of the well design optimization processes and the available studies and applications to employ the well design optimization to solve problems encountered with well design so that cost effectiveness and perfect drilling well performance are achievable. Well design optimization processes include unconventional design(slimhole) co
... Show MoreOptimizing the Access Point (AP) deployment is of great importance in wireless applications owing the requirement to provide efficient and cost-effective communication. Highly targeted by many researchers and academic industries, Quality of Service (QOS) is an important primary parameter and objective in mind along with AP placement and overall publishing cost. This study proposes and investigates a multi-level optimization algorithm based on Binary Particle Swarm Optimization (BPSO). It aims to an optimal multi-floor AP placement with effective coverage that makes it more capable of supporting QOS and cost effectiveness. Five pairs (coverage, AP placement) of weights, signal threshol