Hitachi Data Systems delivers next-generation hyper-converged, scale-out platform for Big Data, powered by Pentaho
PINPOINT Public Relations – Feb 04, 2016 11:00 SGT
New Turnkey HSP Appliance Delivers Native Integration With Pentaho for Robust Data Integration and Analytics; Simplifies Deployment, Operations and Scaling of Enterprise Big Data Projects
SINGAPORE – Hitachi Data Systems Corporation (HDS), a wholly owned subsidiary of Hitachi, Ltd. (TSE: 6501), today unveiled the next generation Hitachi Hyper Scale-Out Platform (HSP), which now offers native integration with the Pentaho Enterprise Platform to deliver a sophisticated, software-defined, hyper-converged platform for big data deployments. Combining compute, storage and virtualization capabilities, the HSP 400 series delivers seamless infrastructure to support big data blending, embedded business analytics and simplified data management.
Modern enterprises increasingly need to derive value from massive volumes of data being generated by information technology (IT), operational technology (OT), the Internet of Things (IoT) and machine-generated data in their environments. HSP offers a software-defined architecture to centralize and support easy storing and processing of these large datasets with high availability, simplified management and a pay-as-you-grow model. Delivered as a fully configured, turnkey appliance, HSP takes hours instead of months to install and support production workloads, and simplifies creation of an elastic data lake that helps customers easily integrate disparate datasets and run advanced analytic workloads.
HSP’s scale-out architecture provides simplified, scalable and enterprise-ready infrastructure for big data. The architecture also includes a centralized, easy-to-use user interface to automate the deployment and management of virtualized environments for leading open source big data frameworks, including Apache Hadoop, Apache Spark, and commercial open source stacks like the Hortonworks Data Platform (HDP).
“Many enterprises don’t possess the internal expertise to perform big data analytics at scale with complex data sources in production environments. Most want to avoid the pitfalls of experimentation with still-nascent technologies, seeking a clear path to deriving real value from their data without the risk and complexity,” said Nik Rouda, Senior Analyst at Enterprise Strategy Group (ESG). “Enterprise customers stand to benefit from turnkey systems like the Hitachi Hyper Scale-Out Platform, which address primary adoption barriers to big data deployments by delivering faster time to insight and value, accelerating the path to digital transformation.”
The next-generation HSP system now offers native integration with Pentaho Enterprise Platform to give customers complete control of the analytic data pipeline and enterprise-grade features such as big data lineage, lifecycle management and enhanced information security. The powerful combination of technologies in the next-generation HSP appliance was designed to accelerate time to business insight and deliver rapid return on investment (ROI), while simplifying the integration of information technology (IT) and operational technology (OT)—a strategic imperative for modern, data-driven enterprises.
“Modern enterprises must merge their IT and OT environments to extend the value of their investments. HSP is a perfect solution to accelerate and simplify IT/OT integration and increase the time to insight and business value of their big data deployments,” said James Dixon, chief technology officer at Pentaho. “The HSP-Pentaho appliance gives customers an affordable, enterprise-class option to unify all their disparate datasets and workloads—including legacy applications and data warehouses—via a modern, scalable and hyper-converged platform that eliminates complexity. We’re pleased to be working with HDS to deliver a simplified, all-in-the-box solution that combines compute, analytics and data management functions in a plug-and-play, future-ready architecture. The Hitachi Hyper Scale-Out Platform 400 is a great first-step in simplifying the entire analytic process.
”With HSP, Hitachi continues to deliver on the promise of the software-defined datacenter to simplify the delivery of IT services through greater abstraction of infrastructure, and improved data access and automation. While its initial focus is on big data analytics use cases, the company’s long-term direction for HSP is to deliver best-in-class total cost of ownership (TCO) for a variety of IT workloads. Hitachi will offer HSP in two configurations to support a broad range of enterprise applications and performance requirements: Serial Attached SCSI (SAS) disk drives, generally available now, and all-flash, expected to ship in mid-2016.
“We consistently hear from our enterprise customers that data silos and complexity are major pain points—and this only gets worse in their scale-out and big data deployments. We have solved these problems for our customers for years, but we are now applying that expertise in a new architecture with Hitachi Hyper Scale-Out Platform,” said Sean Moser, senior vice president, global portfolio and product management at Hitachi Data Systems. “Our HSP appliance gives them a cloud and IoT-ready infrastructure for big data deployments, and a pay-as-you-go model that scales with business growth. Seamless integration with the Pentaho Platform will help them put their IT and OT data to work—faster. This is only the first of many synergistic solutions you can expect to see from Hitachi and Pentaho. Together, we are making it easy for our enterprise customers to maximize the value of their IT and OT investments and accelerate their path to digital transformation.”
Learn More. Join the Conversation.