Sunday, October 27, 2013

Pivotal bulks up big data software portfolio


While EMC and VMware spinoff Pivotal prepares to launch its business intelligence PaaS (platform-as-a-service), the new company has also been busy building its portfolio of data mining and analysis software.


The company is releasing two programs to help data analysis. One is an in-memory data store for real-time analysis that works with the Hadoop data processing platform, and the other is a data discovery tool for business analysts.


[ Explore the current trends and solutions in BI with InfoWorld's interactive Business Intelligence iGuide. | Discover what's new in business applications with InfoWorld's Technology: Applications newsletter. ]


As part of an update to its Hadoop distribution, Pivotal HD 1.1, Pivotal will include GemFire HD, an in-memory transactional store that VMware acquired in 2010. Offering GemFire as part of its Hadoop package, Pivotal is hoping that organizations use the software as a base for building OLTP (online transaction processing) systems that can use Hadoop for long-term storage. This approach will provide organizations with the ability to analyze both current data, as it is being held in GemFire, and older transactional data that has been offloaded to Hadoop, said Abhishek Kashyap, Pivotal principal product manager. A GemFire instance is held entirely within the working memory of a server, which provides a speedy way to interrogate live operational data with SQL, useful for situational awareness and other forms of real-time analysis. The company also unveiled Pivotal Data Dispatch, a tool to help data workers find and prepare data sets for analysis.


Data Dispatch was originally developed by the New York Stock Exchange, whose data analysts have used the software since 2007 to better understand the effects of regulatory requirements.


The software allows analysts to pick, filter and combine the data sets from different sources they need for analysis. The resulting analyst-generated data sets are stored in a "sandbox" that is available to business intelligence tools, such as those offered by Oracle, IBM and SAP, said Todd Paoletti, Pivotal vice president of product marketing.


Using traditional BI tools, analysts typically would have to request the IT staff to combine data sets and deploy them to a data warehouse. This software eliminates the need to consult with IT to generate each new data set, Paoletti said.


Data Dispatch is one of a growing number of self-service BI tools--also offered by Tableau and MicroStrategy -- that provide an easy to use graphical database that data workers themselves can use.


The system administrator initially defines and tags the data sources that are then made available to the analyst. Source data files can be from database files, flat files, Hadoop files, Microsoft Project files or from other commonly-used formats. The resulting combined data sets can be stored either in Hadoop, or with Pivotal's Greenplum.


Pivotal plans to launch its cloud service Nov. 12.Pivotal HD 1.1 will be available Nov. 1.


Joab Jackson covers enterprise software and general technology breaking news for The IDG News Service. Follow Joab on Twitter at @Joab_Jackson. Joab's e-mail address is Joab_Jackson@idg.com


Source: http://akamai.infoworld.com/d/business-intelligence/pivotal-bulks-big-data-software-portfolio-229402?source=rss_business_intelligence
Similar Articles: Arsenio Hall   Darren Young  

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.