In the current paradigms of information technology, cloud computing is the most essential kind of computer service. It satisfies the need for high-volume customers, flexible computing capabilities for a range of applications like as database archiving and business analytics, and the requirement for extra computer resources to provide a financial value for cloud providers. The purpose of this investigation is to assess the viability of doing data audits remotely inside a cloud computing setting. There includes discussion of the theory behind cloud computing and distributed storage systems, as well as the method of remote data auditing. In this research, it is mentioned to safeguard the data that is outsourced and stored in cloud servers. There are four different techniques of remote data auditing procedures that are presented here for distributed cloud services. There are several difficulties associated with data audit methods; however, these difficulties may be overcome by using a variety of techniques, such as the Boneh-Lynn-Shacham signature or the automated blocker protocol. In addition to that, other difficulties associated with distributed-based remote data auditing solutions are discussed. In addition, a variety of approaches might be researched further for further examination in order to find answers to these impending problems.
The importance of economic and technical feasibility studies include all of the investor, the society, the banks, the project and the government. As it will assist in achieving the optimal use of the available economic incomes on the national level and arranging the projects in priority which participates in achieving the government benefit without wasting the amounts .This is the research goal So, the research covered its theoretical aspect, of SAIs and the economic and technical feasibility studies then the role of the auditors in auditing those studies Then a mechanism was proposed includes a guidebook that provides guidelines for organizing the process of preparing and auditi
... Show MoreThe rise of edge-cloud continuum computing is a result of the growing significance of edge computing, which has become a complementary or substitute option for traditional cloud services. The convergence of networking and computers presents a notable challenge due to their distinct historical development. Task scheduling is a major challenge in the context of edge-cloud continuum computing. The selection of the execution location of tasks, is crucial in meeting the quality-of-service (QoS) requirements of applications. An efficient scheduling strategy for distributing workloads among virtual machines in the edge-cloud continuum data center is mandatory to ensure the fulfilment of QoS requirements for both customer and service provider. E
... Show MoreThis work was influenced the separation and preconcentration steps were carried out to determination of metformin (MET) in pharmaceutical preparations and human serum samples. Complex formation method and cloud-point extraction (CPE) coupling with UV-Visible spectrophotometry were used to investigated of study target.The results has showed the best optical characteristic for calibration curve and statistical data which were obtained under optimum conditions. The first method is based on the reaction of MET with nickel (II) in alkaline medium an absorption maximum ?)max) at 434nm. ''Beer's low'' is obeyed in the concentration range (10-100µg.ml-1) with molar absorptivity of 3.9x103 L.mol-1.cm-1.The limit of detection and quantitation valu
... Show MoreTwitter data analysis is an emerging field of research that utilizes data collected from Twitter to address many issues such as disaster response, sentiment analysis, and demographic studies. The success of data analysis relies on collecting accurate and representative data of the studied group or phenomena to get the best results. Various twitter analysis applications rely on collecting the locations of the users sending the tweets, but this information is not always available. There are several attempts at estimating location based aspects of a tweet. However, there is a lack of attempts on investigating the data collection methods that are focused on location. In this paper, we investigate the two methods for obtaining location-based dat
... Show MoreThis paper addresses the nature of Spatial Data Infrastructure (SDI), considered as one of the most important concepts to ensure effective functioning in a modern society. It comprises a set of continually developing methods and procedures providing the geospatial base supporting a country’s governmental, environmental, economic, and social activities. In general, the SDI framework consists of the integration of various elements including standards, policies, networks, data, and end users and application areas. The transformation of previously paper-based map data into a digital format, the emergence of GIS, and the Internet and a host of online applications (e.g., environmental impact analysis, navigation, applications of VGI dat
... Show More<p>Vehicular ad-hoc networks (VANET) suffer from dynamic network environment and topological instability that caused by high mobility feature and varying vehicles density. Emerging 5G mobile technologies offer new opportunities to design improved VANET architecture for future intelligent transportation system. However, current software defined networking (SDN) based handover schemes face poor handover performance in VANET environment with notable issues in connection establishment and ongoing communication sessions. These poor connectivity and inflexibility challenges appear at high vehicles speed and high data rate services. Therefore, this paper proposes a flexible handover solution for VANET networks by integrating SDN and
... Show MoreSequence covering array (SCA) generation is an active research area in recent years. Unlike the sequence-less covering arrays (CA), the order of sequence varies in the test case generation process. This paper reviews the state-of-the-art of the SCA strategies, earlier works reported that finding a minimal size of a test suite is considered as an NP-Hard problem. In addition, most of the existing strategies for SCA generation have a high order of complexity due to the generation of all combinatorial interactions by adopting one-test-at-a-time fashion. Reducing the complexity by adopting one-parameter- at-a-time for SCA generation is a challenging process. In addition, this reduction facilitates the supporting for a higher strength of cove
... Show More