The rise of the Internet, social media and mobile technologies has dramatically increased our digital footprint. When companies like Google and Facebook were confronted with this, they started to look at massive amounts of data ("big data") in a completely different way. But you don't have to be Google or Facebook to benefit from these new possibilities.
Hadoop offers an open source solution based on the same technology that is used within Google. It allows you to store massive amounts of data, and to analyze these in a scalable way to gain new insights.
During this workshop, you have the chance to explore and understand this revolutionary technology and the concepts behind it.
Why should you attend this workshop about the Hadoop ecosystem ?
During this workshop, you will get real answers to these and other questions:
What is Hadoop and how does it fit in the big data (r)evolution ?
Which components do we find in the Hadoop ecosystem ?
How do you import data from a relational database to a Hadoop environment ?
In what ways can you store information in a Hadoop environment ?
How do you integrate existing scripts in MapReduce ?
How do you manage a Hadoop cluster ?
Who should attend this workshop ?
This workshop is aimed at everyone who wants to explore big data technology. This workshop is presented in English and a basic understanding of the following topics will make it easier to do the exercises:
Linux/Unix commands - We will occasionally use the basic commands like less, ls, cat, ...
SQL - We will execute simple queries to extract data from relational databases
The number of participants is limited to 16 to guarantee an optimal interaction and learning experience.
What do you need during this workshop ?
Exercises will be done using a virtual machine which needs to be installed on your own laptop.