Results 31 to 40 of about 9,555 (241)
Past, Present and Future of Hadoop: A Survey [PDF]
In this paper, a technology for massive data storage and computing named Hadoop is surveyed. Hadoop consists of heterogeneous computing devices like regular PCs abstracting away the details of parallel processing and developers can just concentrate on their computational problem. A Hadoop cluster is made of two parts: HDFs and Mapreduce. Hadoop cluster
arxiv
As the need for large-scale data analysis is rapidly increasing, Hadoop, or the platform that realizes large-scale data processing, and MapReduce, or the internal computational model of Hadoop, are receiving great attention. This paper reviews the basic concepts of Hadoop and MapReduce necessary for data analysts who are familiar with statistical ...
Joong-Ho Won+3 more
openaire +2 more sources
With the rapid rate at which networking technologies are changing, there is a need to regularly update network activity datasets to accurately reflect the current state of network infrastructure/traffic.
Sikha S. Bagui+7 more
doaj +1 more source
Adaptive Dynamic Data Placement Algorithm for Hadoop in Heterogeneous Environments [PDF]
Hadoop MapReduce framework is an important distributed processing model for large-scale data intensive applications. The current Hadoop and the existing Hadoop distributed file system’s rack-aware data placement strategy in MapReduce in the homogeneous ...
Avishan Sharafi, Ali Rezaee
doaj
Large-Scale Encryption in the Hadoop Environment: Challenges and Solutions
Data is growing at an enormous rate in the present world. One of the finest and most popular technologies available for handling and processing that enormous amount of data is the Hadoop ecosystem.
Raj R. Parmar+4 more
doaj +1 more source
Hadoop-Benchmark: Rapid Prototyping and Evaluation of Self-Adaptive Behaviors in Hadoop Clusters [PDF]
Optimizing Hadoop executions has attracted a lot of research contributions in particular in the domain of self-adaptive software systems. However, these research efforts are often hindered by the complexity of Hadoop operation and the difficulty to reproduce experimental evaluations that makes it hard to compare different approaches to one another.
Romain Rouvoy+4 more
openaire +5 more sources
Coal mine safety management big data platform based on data mining
The safety management of coal mine enterprises has long been the top priority in coal mine management. Coal mining enterprises store and sort out all kinds of data by investing in environmental monitoring equipment, equipment management system ...
GAO Jing, ZHAO Liangjun, LYU Xuyang
doaj +1 more source
Hadoop and Risk Analytics [PDF]
This paper brings out the specific use case of Hadoop in risk analytics which forms an important part of every organization. Risk analytics is necessary because from business perspective, business leaders in any organization run into one or other kind of risk.
Pankesh Bamotra, J Saira Banu
openaire +1 more source
Big Data with Cloud Computing: Discussions and Challenges
With the recent advancements in computer technologies, the amount of data available is increasing day by day. However, excessive amounts of data create great challenges for users.
Amanpreet Kaur Sandhu
doaj +1 more source
Hadoop Configuration Tuning With Ensemble Modeling and Metaheuristic Optimization
MapReduce is a popular programming model for big data processing. Although the distributed processing framework Hadoop greatly reduced the development complexity of MapReduce applications, fine tuning of the Hadoop systems for optimal performance remains
Xingcheng Hua+2 more
doaj +1 more source