HP’s Data Lake Meets Hadoop

Cameron Turner Cameron Turner
August 28, 2015 Big Data, Cloud & DevOps
Large enterprises usually have fragmented data deposits in department-centric silos or customer service systems. When cross-functional data are required for advanced analytics or for concurrent storage, the enterprise big data systems fail to deliver as they are not well integrated. A solution to that problem has been addressed by recent Apache Hadoop technology, where silo stored data are continuously moved to a central Hadoop system with a multi-server scenario for storing terabytes or petabytes of business data. Additionally, any MapR distribution with Hadoop can greatly enhance the data processing power of a wide variety of data types at high speed. In other words, Hadoop with MapR implementation truly characterizes the big data environment of volume, velocity, variety, and veracity.

For some years now, HP has been facing steep competition from hardware vendors, and while contemplating more efficient methods for big data analytics across their six data centers, or for enhanced customers experiences, they visualized the concept of a data lake to integrate all the data funnels from different silos into a single repository for advanced analytics. HP’s basic mission behind redesigning their enterprise data management included providing integrated big data analytics across the enterprise network, and enhancing the consumer engagement experience for increasing sales. HP’s choice of a suitable Hadoop vendor was narrowed down to MapR because they found this was the only Hadoop distribution that offered scalability, low downtime, high maintenance, and knowledgeable customer support. Keith Dornbusch, Manager of HP’s Big Data Services points out

The MapR technology was top-notch, particularly their performance and high availability features. Plus, our working relationship with the vendor was excellent. MapR’s responsiveness, service, and support, were second to none. And MapR addresses our long-term needs by being a reli­able Hadoop distribution geared towards the enterprise.

 

Hadoop with MapR implementation at HP

The enterprise data management architecture has been built around the concept of a data lake, which is nothing but a single, centralized platform to store data derived from different customer touch points or internal departments. Paul Westerman, Senior Director of IT Big Data Solutions at HP, says, The real value is break­ing down silos and bringing the data together. With Hadoop, you can drop files in and the data is there. It’s a different type of development environment. The HP Hadoop clusters are distributed over 320 servers with Dual Intel processors. The servers are capable of storing about 20 terabytes of data, while the processors provide 128 gigabytes of SD RM making the environment ideal for big data storage and processing.

 

The benefits of HP's Hadoop environment

Flexible file formats MapR ensures data of any type can be moved from silos to Hadoop for quick processing.

Scalability

With data growing at the rate of 50% every year, HP is well positioned to scale up their data storage due to Hadoop.

Zero downtime 

Built-in disaster recovery thresholds and fail-over between clusters ensure high up time.

Cost benefits 

Hadoop provides massive data storage at low cost; and MapR ensures quick processing, which again saves cost.

Monitoring customer activity through telemetry data

This machine-generated data can monitor customer experience with products and send back feedback to HP.
  • Experfy Insights

    Top articles, research, podcasts, webinars and more delivered to you monthly.

  • Cameron Turner

    Tags
    HadoopNoSQL & NewSQL
    © 2021, Experfy Inc. All rights reserved.
    Leave a Comment
    Next Post
    Does Time Matter? Modeling Temporal Dynamics for Better Predictions

    Does Time Matter? Modeling Temporal Dynamics for Better Predictions

    Leave a Reply Cancel reply

    Your email address will not be published. Required fields are marked *

    More in Big Data, Cloud & DevOps
    Big Data, Cloud & DevOps
    Cognitive Load Of Being On Call: 6 Tips To Address It

    If you’ve ever been on call, you’ve probably experienced the pain of being woken up at 4 a.m., unactionable alerts, alerts going to the wrong team, and other unfortunate events. But, there’s an aspect of being on call that is less talked about, but even more ubiquitous – the cognitive load. “Cognitive load” has perhaps

    5 MINUTES READ Continue Reading »
    Big Data, Cloud & DevOps
    How To Refine 360 Customer View With Next Generation Data Matching

    Knowing your customer in the digital age Want to know more about your customers? About their demographics, personal choices, and preferable buying journey? Who do you think is the best source for such insights? You’re right. The customer. But, in a fast-paced world, it is almost impossible to extract all relevant information about a customer

    4 MINUTES READ Continue Reading »
    Big Data, Cloud & DevOps
    3 Ways Businesses Can Use Cloud Computing To The Fullest

    Cloud computing is the anytime, anywhere delivery of IT services like compute, storage, networking, and application software over the internet to end-users. The underlying physical resources, as well as processes, are masked to the end-user, who accesses only the files and apps they want. Companies (usually) pay for only the cloud computing services they use,

    7 MINUTES READ Continue Reading »

    About Us

    Incubated in Harvard Innovation Lab, Experfy specializes in pipelining and deploying the world's best AI and engineering talent at breakneck speed, with exceptional focus on quality and compliance. Enterprises and governments also leverage our award-winning SaaS platform to build their own customized future of work solutions such as talent clouds.

    Join Us At

    Contact Us

    1700 West Park Drive, Suite 190
    Westborough, MA 01581

    Email: [email protected]

    Toll Free: (844) EXPERFY or
    (844) 397-3739

    © 2025, Experfy Inc. All rights reserved.