Big data analysis has important applications in many areas such as sensor networks and connected healthcare. High volume and velocity of big data bring many challenges to data analysis. One possible solution is to summarize the data and provides a manageable data structure to hold a scalable summarization of data for efficient and effective analysis. This research extends our previous work on developing an effective technique to create, organize, access, and maintain summarization of big data and develops algorithms for Bayes classification and entropy discretization of large data sets using the multi-resolution data summarization structure. Bayes classification and data discretization play essential roles in many learning algorithms such as decision tree and nearest neighbor search. The proposed method can handle streaming data efficiently and, for entropy discretization, provide su the optimal split value.
In this work, satellite images for Razaza Lake and the surrounding area
district in Karbala province are classified for years 1990,1999 and
2014 using two software programming (MATLAB 7.12 and ERDAS
imagine 2014). Proposed unsupervised and supervised method of
classification using MATLAB software have been used; these are
mean value and Singular Value Decomposition respectively. While
unsupervised (K-Means) and supervised (Maximum likelihood
Classifier) method are utilized using ERDAS imagine, in order to get
most accurate results and then compare these results of each method
and calculate the changes that taken place in years 1999 and 2014;
comparing with 1990. The results from classification indicated that
The primary objective of this paper is to introduce a new concept of fibrewise topological spaces on D is named fibrewise multi- topological spaces on D. Also, we entroduce the concepts of multi-proper, fibrewise multi-compact, fibrewise locally multi-compact spaces, Moreover, we study relationships between fibrewise multi-compact (resp., locally multi-compac) space and some fibrewise multi-separation axioms.
In this paper, we derived an estimators and parameters of Reliability and Hazard function of new mix distribution ( Rayleigh- Logarithmic) with two parameters and increasing failure rate using Bayes Method with Square Error Loss function and Jeffery and conditional probability random variable of observation. The main objective of this study is to find the efficiency of the derived of Bayesian estimator compared to the to the Maximum Likelihood of this function using Simulation technique by Monte Carlo method under different Rayleigh- Logarithmic parameter and sample sizes. The consequences have shown that Bayes estimator has been more efficient than the maximum likelihood estimator in all sample sizes with application
Nowadays, 3D content is becoming an essential part of multimedia applications, when the 3D content is not protected, hackers may attack and steal it. This paper introduces a proposed scheme that provides high protection for 3D content by implementing multiple levels of security with preserving the original size using weight factor (w). First level of security is implemented by encrypting the texture map based on a 2D Logistic chaotic map. Second level is implemented by shuffling vertices (confusion) based on a 1D Tent chaotic map. Third level is implemented by modifying the vertices values (diffusion) based on a 3D Lorenz chaotic map. Results illustrate that the proposed scheme is completely deform the entire 3D content accord
... Show MoreIn unpredicted industrial environment, being able to adapt quickly and effectively to the changing is key in gaining a competitive advantage in the global market. Agile manufacturing evolves new ways of running factories to react quickly and effectively to changing markets, driven by customized requirement. Agility in manufacturing can be successfully achieved via integration of information system, people, technologies, and business processes. This article presents the conceptual model of agility in three dimensions named: driving factor, enabling technologies and evaluation of agility in manufacturing system. The conceptual model was developed based on a review of the literature. Then, the paper demonstrates the agility
... Show MoreThe usage of remote sensing techniques in managing and monitoring the environmental areas is increasing due to the improvement of the sensors used in the observation satellites around the earth. Resolution merge process is used to combine high resolution one band image with another one that have low resolution multi bands image to produce one image that is high in both spatial and spectral resolution. In this work different merging methods were tested to evaluate their enhancement capabilities to extract different environmental areas; Principle component analysis (PCA), Brovey, modified (Intensity, Hue ,Saturation) method and High Pass Filter methods were tested and subjected to visual and statistical comparison for evaluation. Both visu
... Show MoreThe software-defined network (SDN) is a new technology that separates the control plane from data plane for the network devices. One of the most significant issues in the video surveillance system is the link failure. When the path failure occurs, the monitoring center cannot receive the video from the cameras. In this paper, two methods are proposed to solve this problem. The first method uses the Dijkstra algorithm to re-find the path at the source node switch. The second method uses the Dijkstra algorithm to re-find the path at the ingress node switch (or failed link).
... Show MoreThis paper presents a new and effective procedure to extract shadow regions of high- resolution color images. The method applies this process on modulation the equations of the band space a component of the C1-C2-C3 which represent RGB color, to discrimination the region of shadow, by using the detection equations in two ways, the first by applying Laplace filter, the second by using a Kernel Laplace filter, as well as make comparing the two results for these ways with each other's. The proposed method has been successfully tested on many images Google Earth Ikonos and Quickbird images acquired under different lighting conditions and covering both urban, roads. Experimental results show that this algorithm which is simple and effective t
... Show MoreDatabase is characterized as an arrangement of data that is sorted out and disseminated in a way that allows the client to get to the data being put away in a simple and more helpful way. However, in the era of big-data the traditional methods of data analytics may not be able to manage and process the large amount of data. In order to develop an efficient way of handling big-data, this work studies the use of Map-Reduce technique to handle big-data distributed on the cloud. This approach was evaluated using Hadoop server and applied on EEG Big-data as a case study. The proposed approach showed clear enhancement for managing and processing the EEG Big-data with average of 50% reduction on response time. The obtained results provide EEG r
... Show More