The transmitting and receiving of data consume the most resources in Wireless Sensor Networks (WSNs). The energy supplied by the battery is the most important resource impacting WSN's lifespan in the sensor node. Therefore, because sensor nodes run from their limited battery, energy-saving is necessary. Data aggregation can be defined as a procedure applied for the elimination of redundant transmissions, and it provides fused information to the base stations, which in turn improves the energy effectiveness and increases the lifespan of energy-constrained WSNs. In this paper, a Perceptually Important Points Based Data Aggregation (PIP-DA) method for Wireless Sensor Networks is suggested to reduce redundant data before sending them to the sink. By utilizing Intel Berkeley Research Lab (IBRL) dataset, the efficiency of the proposed method was measured. The experimental findings illustrate the benefits of the proposed method as it reduces the overhead on the sensor node level up to 1.25% in remaining data and reduces the energy consumption up to 93% compared to prefix frequency filtering (PFF) and ATP protocols.
The ultimate goal of any sale contract is to maximize the combined returns of the parties, knowing that these returns are not realized (in long-term contracts) except in the final stages of the contract. Therefore, this requires the parties to the contract to leave some elements open, including the price, because the adoption of a fixed price and inflexible will not be appropriate to meet their desires when contracting, especially with ignorance of matters beyond their will and may affect the market conditions, and the possibility of modifying the fixed price through The elimination is very limited, especially when the parties to the contract are equally in terms of economic strength. Hence, in order to respond to market uncertainties, the
... Show MoreA series of liquid crystals comprising a heterocyclics dihydro pyrrole and 1,2,3-triazole rings [VII]-[X] were synthesized by many steps starting from a reaction of 3,3'-dimethyl-[1,1'-biphenyl]- 4,4'-diamine with chloroacetyl chloride in a mixture of solutions DMF and TEA to synthesise the compounds [I], then the compounds [I] reacted with malononitrile in 1,4-dioxane and TEA solutions to produce compounds [II], then the first step is repeated with compound [II] where it reacted with chloroacetyl chloride in mixture of DMF and TEA to give compound [III], this compound reacted with sodium azide in the presence of sodium chloride and DMF as solvent to produce the compound [IV], which reacted with acrylic acid by a 1.3 dipolar reaction in sol
... Show MoreAs a result of recent developments in highway research as well as the increased use of vehicles, there has been a significant interest paid to the most current, effective, and precise Intelligent Transportation System (ITS). In the field of computer vision or digital image processing, the identification of specific objects in an image plays a crucial role in the creation of a comprehensive image. There is a challenge associated with Vehicle License Plate Recognition (VLPR) because of the variation in viewpoints, multiple formats, and non-uniform lighting conditions at the time of acquisition of the image, shape, and color, in addition, the difficulties like poor image resolution, blurry image, poor lighting, and low contrast, these
... Show MoreThis work aims to develop a secure lightweight cipher algorithm for constrained devices. A secure communication among constrained devices is a critical issue during the data transmission from the client to the server devices. Lightweight cipher algorithms are defined as a secure solution for constrained devices that require low computational functions and small memory. In contrast, most lightweight algorithms suffer from the trade-off between complexity and speed in order to produce robust cipher algorithm. The PRESENT cipher has been successfully experimented on as a lightweight cryptography algorithm, which transcends other ciphers in terms of its computational processing that required low complexity operations. The mathematical model of
... Show MoreTremendous efforts have been exerted to understand first language acquisition to facilitate second language learning. The problem lies in the difficulty of mastering English language and adapting a theory that helps in overcoming the difficulties facing students. This study aims to apply Thomasello's theory of language mastery through usage. It assumes that adults can learn faster than children and can learn the language separately, and far from academic education. Tomasello (2003) studied the stages of language acquisition for children, and developed his theory accordingly. Some studies, such as: (Ghalebi and Sadighi, 2015, Arvidsson, 2019; Munoz, 2019; Verspoor and Hong, 2013) used this theory when examining language acquisition. Thus,
... Show MoreSemantic segmentation realization and understanding is a stringent task not just for computer vision but also in the researches of the sciences of earth, semantic segmentation decompose compound architectures in one elements, the most mutual object in a civil outside or inside senses must classified then reinforced with information meaning of all object, it’s a method for labeling and clustering point cloud automatically. Three dimensions natural scenes classification need a point cloud dataset to representation data format as input, many challenge appeared with working of 3d data like: little number, resolution and accurate of three Dimensional dataset . Deep learning now is the po
In this paper, a fast lossless image compression method is introduced for compressing medical images, it is based on splitting the image blocks according to its nature along with using the polynomial approximation to decompose image signal followed by applying run length coding on the residue part of the image, which represents the error caused by applying polynomial approximation. Then, Huffman coding is applied as a last stage to encode the polynomial coefficients and run length coding. The test results indicate that the suggested method can lead to promising performance.
Image pattern classification is considered a significant step for image and video processing.Although various image pattern algorithms have been proposed so far that achieved adequate classification,achieving higher accuracy while reducing the computation time remains challenging to date. A robust imagepattern classification method is essential to obtain the desired accuracy. This method can be accuratelyclassify image blocks into plain, edge, and texture (PET) using an efficient feature extraction mechanism.Moreover, to date, most of the existing studies are focused on evaluating their methods based on specificorthogonal moments, which limits the understanding of their potential application to various DiscreteOrthogonal Moments (DOMs). The
... Show More