![]() |
I have been discussing my experiences working with companies starting Big Data projects. Every customer I talk to today is working on or planning projects to leverage predictive analytics to improve customer product experience, improve delivery efficiency, and create new revenue opportunities. In my previous post I shared my thoughts on the people and process transformation that Big Data projects require. Big Data projects also require technology transformation. I will focus on three common technology transformations:
I am seeing more customers needing to incorporate near real time predictive analytics. This requires the ability to store and analyze data quickly using In Memory Data Grid (IMDG) data management solutions. At a high level, IMDG's leverage pools of memory from a cluster of servers as high speed, low latency storage. These IMDG periodically copy the data from the memory grid to HDFS for long term persistence and deep analytics. The combination of high speed, low latency storage (IMDG) with the scale, and
One trend I see emerging is customers looking to leverage cloud providers for disaster recovery and/or primary hosting of their cold core capacity. A cache of cold core capacity is kept on site and on application demand data is recalled to the local cold core cache. The cost advantage of leveraging a cloud core storage provider enables companies to keep more data, for longer periods of time. Customers are willing to absorb the performance impact to batch analytics for the extra cost effective capacity. I see this trend accelerating in 2015 as storage gateway technologies and the number of storage service provider's continue to increase. The next trend I have seen emerge is Big Data projects embracing enterprise infrastructure. Many customers start with a rack of servers, network, and direct attach storage for their initial Big Data project. As their infrastructure capacity need grows they feel the pain of infrastructure support. Customers start looking for ways to improve the reliability, manageability, and support of their infrastructure. The two most common solutions considered are:
Many customers are choosing to build their Big Data infrastructure on premise over public cloud for security, and flexibility. A Big Data practioner needs the ability to deploy the latest new analytics functionality in this quickly evolving market. The reliability and manageability of enterprise grade infrastructure is being achieved more and more through software. That allows the infrastructure to leverage generic server hardware. EMC recently introduced a storage software solution for Big Data, ViPR and we published a Haddop deployment guide a few months ago here. Interestingly many of our customers have asked for an option to buy our storage software bundled with hardware based on industry standard servers with enterprise break/fix support. Customers want the flexibility and low cost of commodity hardware combined with the predictability and support of an enterprise grade solution. Our new EMC Elastic Cloud Storage product is an example of this type solution. I believe we will see an acceleration of this type of infrastructure solution in 2015. It will likely include the addition of Big Data analytics software (i.e. Pivotal Big Data Suite) with the enterprise infrastructure.
|
![]() |
