Results 221 to 230 of about 9,555 (241)
Some of the next articles are maybe not open access.
2019
Chapter 2 showed us how to use PolyBase to integrate SQL Server with Azure Blob Storage. In this chapter, we will integrate to the original external data source: Hadoop. In the first part of this chapter, we will take a peek at an already-built Hadoop cluster.
openaire +2 more sources
Chapter 2 showed us how to use PolyBase to integrate SQL Server with Azure Blob Storage. In this chapter, we will integrate to the original external data source: Hadoop. In the first part of this chapter, we will take a peek at an already-built Hadoop cluster.
openaire +2 more sources
The CPS with the Hadoop ecosystems
2013 7th International Conference on Application of Information and Communication Technologies, 2013The cyber physical systems (CPS) are timing sensitive with the real-time constraints. Under the industrial domain, the CPS generates a huge amount of real-time data, which gives essential patterns of system behavior and safety-critical logs. In this paper, we introduce system architecture for the CPS with the Hadoop ecosystem.
Eunmi Choi, Yoojin Lim
openaire +2 more sources
2014
Given that Hadoop-based Map Reduce programming is a relatively new skill, there is likely to be a shortage of highly skilled staff for some time, and those skills will come at a premium price. ETL (extract, transform, and load ) tools, like Pentaho and Talend, offer a visual, component-based method to create Map Reduce jobs, allowing ETL chains to be ...
openaire +2 more sources
Given that Hadoop-based Map Reduce programming is a relatively new skill, there is likely to be a shortage of highly skilled staff for some time, and those skills will come at a premium price. ETL (extract, transform, and load ) tools, like Pentaho and Talend, offer a visual, component-based method to create Map Reduce jobs, allowing ETL chains to be ...
openaire +2 more sources
2014
This chapter explores how Hadoop programs can be unit tested in your local Integrated Development Environment (IDE). Although unit-testing of Hadoop programs has come a long way since the early days of Hadoop, it continues to be challenging because Hadoop components such as Mappers and Reducers run in a distributed environment.
Madhu Siddalingaiah, Sameer Wadkar
openaire +2 more sources
This chapter explores how Hadoop programs can be unit tested in your local Integrated Development Environment (IDE). Although unit-testing of Hadoop programs has come a long way since the early days of Hadoop, it continues to be challenging because Hadoop components such as Mappers and Reducers run in a distributed environment.
Madhu Siddalingaiah, Sameer Wadkar
openaire +2 more sources
2014
Monitoring, as any system administrator will tell you, is ideal for getting to the root of performance issues. Monitoring can help you understand why a system is out of CPU or RAM resources, for example, and notify you when CPU or RAM usage nears a specified percent.
openaire +2 more sources
Monitoring, as any system administrator will tell you, is ideal for getting to the root of performance issues. Monitoring can help you understand why a system is out of CPU or RAM resources, for example, and notify you when CPU or RAM usage nears a specified percent.
openaire +2 more sources
Proceedings of the 2014 Symposium and Bootcamp on the Science of Security, 2014
Hadoop is a map-reduce implementation that rapidly processes data in parallel. Cloud provides reliability, flexibility, scalability, elasticity and cost saving to customers. Moving Hadoop into Cloud can be beneficial to Hadoop users. However, Hadoop has two vulnerabilities that can dramatically impact its security in a Cloud.
Mladen A. Vouk, Xianqing Yu, Peng Ning
openaire +2 more sources
Hadoop is a map-reduce implementation that rapidly processes data in parallel. Cloud provides reliability, flexibility, scalability, elasticity and cost saving to customers. Moving Hadoop into Cloud can be beneficial to Hadoop users. However, Hadoop has two vulnerabilities that can dramatically impact its security in a Cloud.
Mladen A. Vouk, Xianqing Yu, Peng Ning
openaire +2 more sources
Clustering Protein Structures with Hadoop
2016Machine learning is a widely used technique in structural biology, since the analysis of large conformational ensembles originated from single protein structures (e.g. derived from NMR experiments or molecular dynamics simulations) can be approached by partitioning the original dataset into sensible subsets, revealing important structural and dynamics ...
G Paschina+4 more
openaire +4 more sources
2014
Previous chapters outlined MapReduce concepts in detail and we started delving deeper into the way Hadoop is implemented at the end of Chapter 5. This chapter expands on that theme. First, compression schemes are explained, followed by a detailed discussion on Hadoop I/O. We address various types of files, such as Sequence and Avro files.
Sameer Wadkar, Madhu Siddalingaiah
openaire +2 more sources
Previous chapters outlined MapReduce concepts in detail and we started delving deeper into the way Hadoop is implemented at the end of Chapter 5. This chapter expands on that theme. First, compression schemes are explained, followed by a detailed discussion on Hadoop I/O. We address various types of files, such as Sequence and Avro files.
Sameer Wadkar, Madhu Siddalingaiah
openaire +2 more sources
2016
Couple of years back, I was taking a train home one evening and catching up on my round up of new Hadoop tools. A co-passenger was having a covert look at my laptop, and after a while his curiosity was just piqued too much to let it go. “What area do you work in?” he enquired.
openaire +2 more sources
Couple of years back, I was taking a train home one evening and catching up on my round up of new Hadoop tools. A co-passenger was having a covert look at my laptop, and after a while his curiosity was just piqued too much to let it go. “What area do you work in?” he enquired.
openaire +2 more sources
Proceedings of the 40th ACM technical symposium on Computer science education, 2009
The potential benefits of data-intensive scalable computing (DISC) in CS education are considered in the context of a small college with an active student-operated Beowulf cluster initiative. The map-reduce computational model, of great importance in industry, is reviewed, and the Hadoop implementation of that model
openaire +3 more sources
The potential benefits of data-intensive scalable computing (DISC) in CS education are considered in the context of a small college with an active student-operated Beowulf cluster initiative. The map-reduce computational model, of great importance in industry, is reviewed, and the Hadoop implementation of that model
openaire +3 more sources