By Kaycee Lai | June 8, 2021
There are several models to collect, store and organize a company’s data, but one of the most talked about in the last decade has been the data lake, a huge repository that allows the input of massive amounts of any kind of data in raw form at high speed.
The concept of the data lake originated with a technology called Hadoop, in which a few engineers developed a data storage model designed for the immense input of data required by search platforms. It was eventually spun into an open source project run by the Apache Software foundation, which made Hadoop’s code available to the world for free.
To read the full article visit www.builtin.com.