In the field of data security, the critical challenge of preserving sensitive information during its transmission through public channels takes centre stage. Steganography, a method employed to conceal data within various carrier objects such as text, can be proposed to address these security challenges. Text, owing to its extensive usage and constrained bandwidth, stands out as an optimal medium for this purpose. Despite the richness of the Arabic language in its linguistic features, only a small number of studies have explored Arabic text steganography. Arabic text, characterized by its distinctive script and linguistic features, has gained notable attention as a promising domain for steganographic ventures. Arabic text steganography harn
... Show MoreBackground: Obesity tends to appear in modern societies and constitutes a significant public health problem with an increased risk of cardiovascular diseases.
Objective: This study aims to determine the agreement between actual and perceived body image in the general population.
Methods: A descriptive cross-sectional study design was conducted with a sample size of 300. The data were collected from eight major populated areas of Northern district of Karachi Sindh with a period of six months (10th January 2020 to 21st June 2020). The Figure rating questionnaire scale (FRS) was applied to collect the demographic data and perception about body weight. Body mass index (BMI) used for ass
... Show MoreThis work presents a symmetric cryptography coupled with Chaotic NN , the encryption algorithm process the data as a blocks and it consists of multilevel( coding of character, generates array of keys (weights),coding of text and chaotic NN ) , also the decryption process consists of multilevel (generates array of keys (weights),chaotic NN, decoding of text and decoding of character).Chaotic neural network is used as a part of the proposed system with modifying on it ,the keys that are used in chaotic sequence are formed by proposed key generation algorithm .The proposed algorithm appears efficiency during the execution time where it can encryption and decryption long messages by short time and small memory (chaotic NN offer capacity of m
... Show MoreText categorization refers to the process of grouping text or documents into classes or categories according to their content. Text categorization process consists of three phases which are: preprocessing, feature extraction and classification. In comparison to the English language, just few studies have been done to categorize and classify the Arabic language. For a variety of applications, such as text classification and clustering, Arabic text representation is a difficult task because Arabic language is noted for its richness, diversity, and complicated morphology. This paper presents a comprehensive analysis and a comparison for researchers in the last five years based on the dataset, year, algorithms and the accuracy th
... Show MoreThe main purpose of this work is to introduce some types of fuzzy convergence sequences of operators defined on a standard fuzzy normed space (SFN-spaces) and investigate some properties and relationships between these concepts. Firstly, the definition of weak fuzzy convergence sequence in terms of fuzzy bounded linear functional is given. Then the notions of weakly and strongly fuzzy convergence sequences of operators are introduced and essential theorems related to these concepts are proved. In particular, if ( ) is a strongly fuzzy convergent sequence with a limit where linear operator from complete standard fuzzy normed space into a standard fuzzy normed space then belongs to the set of all fuzzy bounded linear operators
The aim of this paper is to present a weak form of -light functions by using -open set which is -light function, and to offer new concepts of disconnected spaces and totally disconnected spaces. The relation between them have been studied. Also, a new form of -totally disconnected and inversely -totally disconnected function have been defined, some examples and facts was submitted.
The conventional procedures of clustering algorithms are incapable of overcoming the difficulty of managing and analyzing the rapid growth of generated data from different sources. Using the concept of parallel clustering is one of the robust solutions to this problem. Apache Hadoop architecture is one of the assortment ecosystems that provide the capability to store and process the data in a distributed and parallel fashion. In this paper, a parallel model is designed to process the k-means clustering algorithm in the Apache Hadoop ecosystem by connecting three nodes, one is for server (name) nodes and the other two are for clients (data) nodes. The aim is to speed up the time of managing the massive sc
... Show MoreThe unpredictable and huge data generation nowadays by smart computing devices like (Sensors, Actuators, Wi-Fi routers), to handle and maintain their computational processing power in real time environment by centralized cloud platform is difficult because of its limitations, issues and challenges, to overcome these, Cisco introduced the Fog computing paradigm as an alternative for cloud-based computing. This recent IT trend is taking the computing experience to the next level. It is an extended and advantageous extension of the centralized cloud computing technology. In this article, we tried to highlight the various issues that currently cloud computing is facing. Here
... Show MoreText categorization refers to the process of grouping text or documents into classes or categories according to their content. Text categorization process consists of three phases which are: preprocessing, feature extraction and classification. In comparison to the English language, just few studies have been done to categorize and classify the Arabic language. For a variety of applications, such as text classification and clustering, Arabic text representation is a difficult task because Arabic language is noted for its richness, diversity, and complicated morphology. This paper presents a comprehensive analysis and a comparison for researchers in the last five years based on the dataset, year, algorithms and the accu
... Show More