Cloud-scale analytics for the data centre

Combined offering delivers unified analytics and machine learning on a reliable platform.

  • 2 years ago Posted in

Vertica has announced integration with H3C ONEStor to deliver the benefits of cloud-native analytics to enterprise data centres. Together, Vertica and H3C empower analytically driven companies to elastically scale capacity and performance as data volumes grow and as machine learning initiatives become a business imperative – all from within hybrid environments.

“With this integration, data-driven leaders in the APAC region will benefit from a powerful combination of industry-leading platforms that accommodate any present and future strategic analytical and machine learning initiatives,” said Scott Richards, vice president and general manager of Vertica. “H3C has a solid presence in this region, enabling our joint customers to run Vertica’s cloud-optimised architecture with H3C’s ONEStor to meet the most demanding performance and financial requirements – from enterprise data centres or private clouds.”

Vertica with H3C ONEStor enables organisations to adopt cloud innovation for analytics wherever their data resides, even if the timing or costs of cloud migration is not feasible. Combining these two technologies offers fast analytics while simplifying data protection with easy backup and replication features. In addition, it provides 99.9999999 percent of reliability for storing mission-critical data as it leverages cloud technologies for on-premises deployments.

“We’re delighted to offer Vertica analytics and machine learning on top of our ONEStor from H3C. The analytical performance offered by Vertica combined with the data reliability of ONEStor offer ultra-large-scale and capacity, unmatched analytical performance, and high data reliability,” said Yili Liu, VP of Cloud and Intelligence Product Line from H3C.

The combined offering delivers high-performance analytics and machine learning with enterprise-grade object storage to enable organisations to:

 

•   Address scalability needs for now and in the future – Elastically scale-out to support terabytes to petabytes of data and thousands of users as your analytical and machine learning needs increase.

•   Leverage separation of compute and storage architecture – Administrators can scale compute and data storage resources separately to address varying dynamic workload requirements.

•   Simplify database operations – The solution offers excellent reliability by including many features for data protection. Data loss is extremely rare, offering up to nine 9’s of reliability.

•   Address all data consumer needs – Isolate analytical workloads to accommodate various data consumer needs – from business analysts to data scientists – without competing for resources.

Sophos has formed a strategic partnership with Tenable to provide Sophos Managed Risk, a worldwide...
Celonis and the BMW Group have significantly expanded their strategic partnership in order to...
Powerful combination of the AI-native CrowdStrike Falcon XDR platform and the Rubrik Security Cloud...
Cockroach Labs has formed a new partnership with Crayon, a global provider of software and cloud...
Research by Alteryx finds that data silos and quality pose (surmountable) challenges as IT teams...
With Splunk, Cisco becomes one of the largest software companies globally.
Proprietary AI engine identifies issues before an influx in help desk tickets.
Only 25% of data professionals believe their organisation’s strategic decisions are data-driven.