Hadoop is fast becoming the preferred way to store and process big data. By T-System’s estimates, in five years, 80 percent of all new data will first land in Hadoop’s distributed file system (HDFS) or in alternative Object Storage architectures.
Yet with the excitement around this open source framework, enterprise users risk overlooking that all Hadoop flavors are not created equal. Choosing one implementation over another can mean veering off the path of genuine open source software and instead heading down the dead-end street of expensive vendor lock-in and stunted innovation.
A little history lesson
The enterprise tech world has been there before. Remember the Unix vs. Linux schism? The former started as a project at Bell Labs and UC Berkeley in the 1970s. Unix was acclaimed for its performance, stability and scalability. It was cutting-edge back then when it came to multi-user and multitasking capabilities, support of IP networks, tools…
View original post 832 more words