EMC adds unstructured big data analytics to Greenplum platform

EMC Corp.’s Greenplum subsidiary today announced a new capability to its Hadoop Data Computing Appliance (DCA) that allows users to mix and match unstructured and structured data analytics platforms.

EMC also announced its Greenplum Analytics Workbench, a 1,000-plus-node test bed for software integration tests of Apache Hadoop software.

The test bed provides the Hadoop open source community with the testing resources to quickly identify bugs, stabilize new releases and optimize hardware configurations in an effort to speed up the innovation of Hadoop. All testing and results will be given back to the Apache Software Foundation and the open source community. EMC’s testing will be planned in co-ordination with the Apache Hadoop project. Hadoop is an open source software platform, originally developed by Google, for analyzing large quantities of data.

On its Greenplum appliance product front, EMC introduced the Modular Data Computing Appliance, which allows users to combine a massively parallel processing relational database with enterprise-class Apache Hadoop in a single, unified appliance to achieve structured and unstructured data processing.

Greenplum introduced the Data Computing Appliance in October 2010. An updated version of the DCA that included a Hadoop appliance was released this past May .

The Greenplum HD (Hadoop) DCA is built on top of Intel X86 servers and uses both a structured database built by Greenplum, which EMC acquired last year , and the Apache open-source version of Hadoop. The older version of the appliance is based on Sun Fire x64-based servers.

According to Scott Yara, co-founder of Greenplum and vice-president of products of EMC ‘s data computing division, administrators can read and write files in parallel from Greenplum to HDFS (Hadoop File System), enabling rapid data sharing. Cross-platform analysis can be performed using Greenplum SQL and advanced analytic functions accessing data on HDFS.

The new Modular DCA adds high-performance computing modules in the form of SAS Institute’s in-memory analytics software, allowing it to serve up both structured data, such as databases, and unstructured file data, according to Yara.

“The main change is that it can perform parallel processing using server memory through the use of business analytics software [from SAS] ,” Yara said. “We wanted to offer a Lego building-block-type architecture.”

Through the use of the SAS software, structured and unstructured data can exist on multiple x86 hosts, the purpose of which is to allow users to perform computations in memory on each server node in a clustered configuration.

“The power of the appliance is that it can perform all these complex problems in parallel,” Yara said.

The new Modular DCA is undergoing product trials and is expected to be available by the end of this year, Yara said.

Would you recommend this article?

Share

Thanks for taking the time to let us know what you think of this article!
We'd love to hear your opinion about this or any other story you read in our publication.


Jim Love, Chief Content Officer, IT World Canada

Featured Download

Featured Articles

Cybersecurity in 2024: Priorities and challenges for Canadian organizations 

By Derek Manky As predictions for 2024 point to the continued expansion...

Survey shows generative AI is a top priority for Canadian corporate leaders.

Leaders are devoting significant budget to generative AI for 2024 Canadian corporate...

Related Tech News

Tech Jobs

Our experienced team of journalists and bloggers bring you engaging in-depth interviews, videos and content targeted to IT professionals and line-of-business executives.

Tech Companies Hiring Right Now