Big Data

Ecosystem for Hadoop Hadoop’s HDFS is a highly fault-tolerant distributed file system and, like Hadoop in general, designed to be deployed on low-cost hardware. It provides high throughput access to application data and is suitable for applications that have large data sets. This course examines the Hadoop ecosystem by demonstrating all of the commonly used open source software components. You’ll explore a Big Data model to understand how these tools combine to create a supercomputing platform. You’ll also learn how the principles of supercomputing apply to Hadoop and how this yields an affordable supercomputing environment.
Big Data Opportunities and Challenges Big Data requires a holistic approach and a change to regular working practices. This course covers the way teams work in Big Data organizations, some projects and use cases for Big Data, and challenges and opportunities that Big Data presents.
Big Data Technology Wave A number of tools are available for working with Big Data. Many of the tools are open source and Linux distribution based. This course covers the fundamentals of Big Data, including positioning it in a historical IT context, the tools available for working with Big Data, the Big Data stack, and finally, an in-depth look at Apache Hadoop.