Big Data:
Knowledge Article
Posted By:
Big Data Framework
Posted On:
Share Post:

Big Data Characteristics

Big Data Characteristics – The Four V’s of Big Data

What is the difference between regular data analysis and when are we talking about “Big” data? Although the answer to this question cannot be universally determined, there are a number of characteristics that define Big Data. In this article, we will discuss the characteristics of Big Data are commonly referred to as the four Vs.

Volume of Big Data

The volume of data refers to the size of the data sets that need to be analyzed and processed, which are now frequently larger than terabytes and petabytes. The sheer volume of the data requires distinct and different processing technologies than traditional storage and processing capabilities. In other words, this means that the data sets in Big Data are too large to process with a regular laptop or desktop processor. An example of a high-volume data set would be all credit card transactions on a day within Europe.

card transaction

Figure 1: Example of card transaction

Velocity of Big Data

Velocity refers to the speed with which data is generated. High velocity data is generated with such a pace that it requires distinct (distributed) processing techniques. An example of a data that is generated with high velocity would be Twitter messages or Facebook posts.

Social Media

Figure 2: Example of social media channels such as Facebook, Instagram & WhatsApp

Variety of Big Data

Variety makes Big Data really big. Big Data comes from a great variety of sources and generally is one out of three types: structured, semi structured and unstructured (as discussed in the next section). The variety in data types frequently requires distinct processing capabilities and specialist algorithms. An example of high variety data sets would be the CCTV audio and video files that are generated at various locations in a city.


Figure 3: Example of CCTV footage

Veracity of Big Data

Veracity refers to the quality of the data that is being analyzed. High veracity data has many records that are valuable to analyze and that contribute in a meaningful way to the overall results. Low veracity data, on the other hand, contains a high percentage of meaningless data. The non-valuable in these data sets is referred to as noise. An example of a high veracity data set would be data from a medical experiment or trial.

Medical experiments

Figure 4: Example of medical experiment

Data that is high volume, high velocity and high variety must be processed with advanced tools (analytics and algorithms) to reveal meaningful information. Because of these characteristics of the data, the knowledge domain that deals with the storage, processing, and analysis of these data sets has been labeled Big Data.

To learn more about Big Data, visit our Big Data Knowledge Base. For more information, contact us at or drop us a message in the chatbox.

about author

Big Data Framework

Excepteur sint ocaecat cupidas proident sunt culpa quid officia desers mollit sed.

subscribe to newsletter

Receive more Big Data Knowledge article in your inbox: