article thumbnail

Brief History of Data Engineering

Jesse Anderson

They created MapReduce and GFS in 2004. Doug Cutting took those papers and created Apache Hadoop in 2005. They were the first companies to commercialize open source big data technologies and pushed the marketing and commercialization of Hadoop. Hadoop was hard to program, and Apache Hive came along in 2010 to add SQL.

article thumbnail

A Prequel to Data Mesh

Towards Data Science

Image by the author 2004 to 2010 — The elephant enters the room New wave of applications emerged — Social Media, Software observability, etc. Result: Hadoop & NoSQL frameworks emerged. Result: Companies started to sell pre-configured data warehouses as products. The concept of `Data Marts` was introduced.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Evolution of the Cloud Data Platform: From Google to Ascend

Ascend.io

Back in 2004, I got to work with MapReduce at Google years before Apache Hadoop was even released, using it on a nearly daily basis to analyze user activity on web search and analyze the efficacy of user experiments. I’ve had the good fortune to work at or start companies that were breaking new ground. Big data would be a big deal.

Cloud 52
article thumbnail

Evolution of the Cloud Data Platform: From Google to Ascend

Ascend.io

Back in 2004, I got to work with MapReduce at Google years before Apache Hadoop was even released, using it on a nearly daily basis to analyze user activity on web search and analyze the efficacy of user experiments. I’ve had the good fortune to work at or start companies that were breaking new ground. Big data would be a big deal.

Cloud 52
article thumbnail

Data Analysis with Spark

Zalando Engineering

For the sake of comparison, let’s recap the Hadoop way of working: Hadoop saves intermediate states to disk and communicates over a network. In fact, in a 2004 mapReduce research paper the designer states that key-value pairs is a key choice in designing mapReduce. Provides in memory storage for cached RDD’s.

article thumbnail

Industry Interview Series- How Big Data is Transforming Business Intelligence?

ProjectPro

Solocal has taken big data to the next stage of BI by designing a novel vision of BI with the open source distributed computing framework Hadoop. It replaced its traditional BI structure by integrating big data and Hadoop."-April In BI – there is a need to use ETL on top of Hadoop as there is not much scripting.

article thumbnail

How to Become a Data Engineer in 2024?

Knowledge Hut

Facebook It is a social media platform created originally by Mark Zuckerberg for college students in 2004. Most of the Data engineers working in the field enroll themselves in several other training programs to learn an outside skill, such as Hadoop or Big Data querying, alongside their Master's degree and PhDs.