The need for an efficient method to find the furthermost appropriate document corresponding to a particular search query has become crucial due to the exponential development in the number of papers that are now readily available to us on the web. The vector space model (VSM) a perfect model used in “information retrieval”, represents these words as a vector in space and gives them weights via a popular weighting method known as term frequency inverse document frequency (TF-IDF). In this research, work has been proposed to retrieve the most relevant document focused on representing documents and queries as vectors comprising average term term frequency inverse sentence frequency (TF-ISF) weights instead of representing them as vectors of term TF-IDF weight and two basic and effective similarity measures: Cosine and Jaccard were used. Using the MS MARCO dataset, this article analyzes and assesses the retrieval effectiveness of the TF-ISF weighting scheme. The result shows that the TF-ISF model with the Cosine similarity measure retrieves more relevant documents. The model was evaluated against the conventional TF-ISF technique and shows that it performs significantly better on MS MARCO data (Microsoft-curated data of Bing queries).
In this paper, we present multiple bit error correction coding scheme based on extended Hamming product code combined with type II HARQ using shared resources for on chip interconnect. The shared resources reduce the hardware complexity of the encoder and decoder compared to the existing three stages iterative decoding method for on chip interconnects. The proposed method of decoding achieves 20% and 28% reduction in area and power consumption respectively, with only small increase in decoder delay compared to the existing three stage iterative decoding scheme for multiple bit error correction. The proposed code also achieves excellent improvement in residual flit error rate and up to 58% of total power consumption compared to the other err
... Show MoreMixture experiments are response variables based on the proportions of component for this mixture. In our research we will compare the scheffʼe model with the kronecker model for the mixture experiments, especially when the experimental area is restricted.
Because of the experience of the mixture of high correlation problem and the problem of multicollinearity between the explanatory variables, which has an effect on the calculation of the Fisher information matrix of the regression model.
to estimate the parameters of the mixture model, we used the (generalized inverse ) And the Stepwise Regression procedure
... Show MoreCalculating the Inverse Kinematic (IK) equations is a complex problem due to the nonlinearity of these equations. Choosing the end effector orientation affects the reach of the target location. The Forward Kinematics (FK) of Humanoid Robotic Legs (HRL) is determined by using DenavitHartenberg (DH) method. The HRL has two legs with five Degrees of Freedom (DoF) each. The paper proposes using a Particle Swarm Optimization (PSO) algorithm to optimize the best orientation angle of the end effector of HRL. The selected orientation angle is used to solve the IK equations to reach the target location with minimum error. The performance of the proposed method is measured by six scenarios with different simulated positions of the legs. The proposed
... Show MoreAcceptable Bit Error rate can be maintained by adapting some of the design parameters such as modulation, symbol rate, constellation size, and transmit power according to the channel state.
An estimate of HF propagation effects can be used to design an adaptive data transmission system over HF link. The proposed system combines the well known Automatic Link Establishment (ALE) together with variable rate transmission system. The standard ALE is modified to suite the required goal of selecting the best carrier frequency (channel) for a given transmission. This is based on measuring SINAD (Signal plus Noise plus Distortion to Noise plus Distortion), RSL (Received Signal Level), multipath phase distortion and BER (Bit Error Rate) fo
... Show MoreWe explore the transform coefficients of fractal and exploit new method to improve the compression capabilities of these schemes. In most of the standard encoder/ decoder systems the quantization/ de-quantization managed as a separate step, here we introduce new way (method) to work (managed) simultaneously. Additional compression is achieved by this method with high image quality as you will see later.
The development of wireless sensor networks (WSNs) in the underwater environment leads to underwater WSN (UWSN). It has severe impact over the research field due to its extensive and real-time applications. However effective execution of underwater WSNs undergoes several problems. The main concern in the UWSN is sensor nodes’ energy depletion issue. Energy saving and maintaining quality of service (QoS) becomes highly essential for UWASN because of necessity of QoS application and confined sensor nodes (SNs). To overcome this problem, numerous prevailing methods like adaptive data forwarding techniques, QoS-based congestion control approaches, and various methods have been devised with maximum throughput and minimum network lifesp
... Show MoreThe development of wireless sensor networks (WSNs) in the underwater environment leads to underwater WSN (UWSN). It has severe impact over the research field due to its extensive and real-time applications. However effective execution of underwater WSNs undergoes several problems. The main concern in the UWSN is sensor nodes’ energy depletion issue. Energy saving and maintaining quality of service (QoS) becomes highly essential for UWASN because of necessity of QoS application and confined sensor nodes (SNs). To overcome this problem, numerous prevailing methods like adaptive data forwarding techniques, QoS-based congestion control approaches, and various methods have been devised with maximum throughput and minimum network lifesp
... Show More
abstract:
The aim of this study is to clarify the relationship between the concept of inconsistency and the measurement of the inverse relationship of overlap and contrast. The study will address the following points:
- Explaining the true nature of inconsistency and contrast, as understanding their relationship is essential for determining them.
- Examining the view of the scholars regarding their significance as approaches to understanding the relationship.
- Identifying the relationship between inconsistency and contrast in terms of overlap and contrast.
In this research , we study the inverse Gompertz distribution (IG) and estimate the survival function of the distribution , and the survival function was evaluated using three methods (the Maximum likelihood, least squares, and percentiles estimators) and choosing the best method estimation ,as it was found that the best method for estimating the survival function is the squares-least method because it has the lowest IMSE and for all sample sizes