This vendor-written piece has been edited by Executive Networks Media to eliminate product promotion, but readers should note it will likely favour the submitter's approach.
Data is the new natural resource of our time. Information that was previously thrown away is transforming every facet of business as organizations use technology to query and analyze data in real time to help them better understand their customers' behaviour and markets. Such knowledge and insight is helping businesses make better decisions and improve performance.
Yet for data to be worth anything, organizations must have the ability to store it. With more data being created every 20 minutes than is currently held by the Library of Congress, many organizations are overwhelmed by petabytes (one petabyte is the equivalent of 20 million filing cabinets of text). As a result, companies are moving to the cloud-not only because of its flexibility, cost efficiency and reliability, but because of the impact of "data gravity."
What is data gravity?
The concept of data gravity was first coined by Dave McCrory. It relates to the idea that as organizations adopt and migrate data to the cloud, data that remains outside of the cloud starts to gravitate to those applications running in the cloud.
In the past, it was difficult to move large chunks of data because it was time-consuming and expensive. Now, with active data replication a reality, the mass of data migrating to the cloud is increasing. As a result, it makes sense to have the applications that will access and manipulate that data in the same cloud in order to maximize efficiency.
McCrory predicted with data being moved to the cloud and applications being written for the cloud, the two activities mutually reinforce each other, increasing the gravitational pull: "As data accumulates, there is a greater likelihood that additional services and applications will be attracted to this data. This is the same effect gravity has on objects around a planet."
Nowhere can this gravitational pull be seen more than at Amazon Web Services. As the fastest growing enterprise technology company in history, it holds more data in its cloud than anyone else and has five times the utilized capacity of its nearest 14 competitors combined. This is not because of lock-in, but because of how data behaves: As applications and services are created to feed on the datasets, they create more data, which in turn gets analyzed by other applications that create their own data and so on, leading to a virtuous circle.
Data gravity the new KPI for cloud vendors
For Amazon and other tech companies, such as Google, IBM and Microsoft, data gravity is the new key performance indicator (KPI) as they compete to hold data by not only proposing better prices and better services, but in the ability to accumulate knowledge about their customers and deeply understand their needs.
Sign up for Computerworld eNewsletters.