Knowledge is knowledge

What is Hadoop?

 

Traditionally in computation and processing the data you have to bring data to the computer and you require a program for the processing of the data then you bring data to the program and the processing is done by the program. In a big data cluster Larry and Sergey Brin came up with an idea they sliced the data into small pieces and they distribute this data between thousands of computers first it was hundreds but now thousands and now it's ten thousand. And then they send the program to all the computers in the cluster. and each computer process the data and then sends the result back and then the result is combined and the data processing will be done in very little time. 



The first process is known as the map or mapper process and the second one is known as a reduction process. A fairly simple process but turns out very useful to process a large amount of data. Only twice the number of servers and you have twice the efficiency. So the bottleneck of all the major social media companies. 

Yahoo then got on board . Yahoo hired someone named Doug Cutting who had been working on a clone or a copy of the Google big data architecture and now that's called Hadoop. And now the Hadoop becomes a very famous one there are hundreds of thousands of companies out there that have some kind o footprint in the big data world.








Share:

No comments:

Post a Comment