Data generation is skyrocketing—traditional database systems fail to support “big data”

“Big data” encompass a wide range of the tremendous data generated from various sources such as mobile devices, digital repositories, and enterprise applications. The data can be structured as well as unstructured. It ranges from terabytes—10^12 bytes—to even exabytes—10^18 bytes. Working with “big data” is complex because of the five v’s associated with “big data.” Facebook (FB) gets ~10 million new photos uploaded every hour and Google (GOOG) processes over 24 petaBbytes of data every day. Twitter (TWTR) tweets ~400 million tweets per day. All this shows the magnificent volume, variety, value, and velocity of “big data.”

Why traditional database systems fail to support “big data”

The previous chart shows the tremendous growth “big data” is expected to experience in the near future. The “big data” market is expected to cross $50 billion by 2017.

RDBMS for data storage

The relational database management system (or RDBMS) had been the one solution for all database needs. Oracle, IBM (IBM), and Microsoft (MSFT) are the leading players of RDBMS. RDBMS uses structured query language (or SQL) to define, query, and update the database. However, the volume and velocity of business data has changed dramatically in the last couple of years. It’s skyrocketing every day.

Limitations of RDBMS to support “big data”

First, the data size has increased tremendously to the range of petabytes—one petabyte = 1,024 terabytes. RDBMS finds it challenging to handle such huge data volumes. To address this, RDBMS added more central processing units (or CPUs) or more memory to the database management system to scale up vertically.

Second, the majority of the data comes in a semi-structured or unstructured format from social media, audio, video, texts, and emails. However, the second problem related to unstructured data is outside the purview of RDBMS because relational databases just can’t categorize unstructured data. They’re designed and structured to accommodate structured data such as weblog sensor and financial data.

Also, “big data” is generated at a very high velocity. RDBMS lacks in high velocity because it’s designed for steady data retention rather than rapid growth. Even if RDBMS is used to handle and store “big data,”  it will turn out to be very expensive.
As a result, the inability of relational databases to handle “big data” led to the emergence of new technologies.

Latest articles

WeWork is gearing up for an IPO. On Wednesday, the company made its IPO filing with the SEC public and expects to garner $3.5 billion from its IPO.

After FCC Chairman Ajit Pai recommended the approval of the T-Mobile–Sprint merger, Representative David Cicilline urged the FCC to allow public comment.

Cresco Labs (CRLBF) is set to report its Q2 earnings on August 21 after the market closes. The company's stock fell 5.3% yesterday.

Cannabis stocks mostly traded in positive territory today. Supreme Cannabis (FIRE) and Aurora Cannabis (ACB) rose about 9.5% and 5.5%, respectively.

NVIDIA (NVDA) stock soared 6% in today’s trading session as its Q2 earnings for fiscal 2020 beat estimates. However, its guidance missed estimates.

Nio Inc. (NIO) has disrupted the automotive space since 2014 but only made waves in the market since its IPO. Investors have suffered numerous setbacks.