In the world of big data, the surrounding hype has spawned a brand-new lingo. Need a little clarity? Read on for a glossary of sorts highlighting some of the main data types you should understand.
The shining star in this constellation of terms is “fast data,” which is popping up with increasing frequency. It refers to “data whose utility is going to decline over time,” said Tony Baer, a principal analyst at Ovum who says he coined the term back in 2012.
It’s things like Twitter feeds and streaming data that need to be captured and analyzed in real time, enabling immediate decisions and responses. A capital markets trading firm may rely on it for conducting algorithmic or high-frequency trades.
“Fast data can refer to a few things: fast ingest, fast streaming, fast preparation, fast analytics, fast user response,” said Nik Rouda, a senior analyst with Enterprise Strategy Group. It’s “mostly marketing hype,” but it “shows the need for performance in a variety of ways.”
Increased bandwidth, commodity hardware, declining memory prices and real-time analytics have all contributed to the rise of fast data, Baer said.
At the opposite end of the spectrum is “slow data,” or data that might trickle in at a comparatively leisurely pace, warranting less-frequent analysis. Baer points to a device that monitors ocean tides as an example — for most purposes, real-time updates aren’t needed.
In general, this kind of data is better-suited for capture in a data lake and subsequent batch processing.
“Small data” is “anything that fits on one laptop,” said Gregory Piatetsky-Shapiro, president of analytics consultancy KDnuggets.