Big Data 2014: Powering Up the Curve

December 5, 2013

Last year, I predicted that 2013 would be the year big data analytics started to go into mainstream deployment and the research we recently commissioned with Enterprise Management Consultants indicates that’s happened. What really surprised me though is the extent to which the demand for data blending has powered up the curve and I believe this trend will accelerate big data growth in 2014.

Prediction one: The big data ‘power curve’ in 2014 will be shaped by business users’ demand for data blending
Customers like Andrew Robbins of Paytronix and Andrea Dommers-Nilgen of TravelTainment, who recently spoke about their Pentaho projects at events in NY and London, both come from the business side and are achieving specific goals for their companies by blending big and relational data. Business users like these are getting inspired by the potential to tap into blended data to gain new insights from a 360 degree customer view, including the ability to analyze customer behavior patterns and predict the likelihood that customers will take advantage of targeted offers.

Prediction two: big data needs to play well with others!
Historically, big data projects have largely sat in the IT departments because of the technical skills needed and the growing and bewildering array of technologies that can be combined to build reference architectures. Customers must choose from the various commercial and open source technologies including Hadoop distributions, NoSQL databases, high-speed databases, analytics platforms and many other tools and plug-ins. But they also need to consider existing infrastructure including relational data and data warehouses and how they’ll fit into the picture.

The plus side of all this choice and diversity is that after decades of tyranny and ‘lock-in’ imposed by enterprise software vendors, in 2014, even greater buying power will shift to customers. But there are also challenges. It can be cumbersome to manage this heterogeneous data environment involved with big data analytics. It also means that IT will be looking for Big Data tools to help deploy and manage these complex emerging reference architectures, and to simplify them.  It will be incumbent on the Big Data technology vendors to play well with each other and work towards compatibility. After all, it’s the ability to access and manage information from multiple sources that will add value to big data analytics.

Prediction three: you will see even more rapid innovation from the big data open source community
New open source projects like Hadoop 2.0 and YARN, as the next generation Hadoop resource manager, will make the Hadoop infrastructure more interactive. New open source projects like STORM, a streaming communications protocol, will enable more real-time, on-demand blending of information in the big data ecosystem.

Since we announced the industry’s first native Hadoop connectors in 2010, we’ve been on a mission to make the transition to big data architectures easier and less risky in the context of this expanding ecosystem. In 2013 we made some massive breakthroughs towards this, starting with our most fundamental resource, the adaptive big data layer. This enables IT departments to feel smarter, safer and more confident about their reference architectures and open up big data solutions to people in the business, whether they be data scientists, data analysts, marketing operations analysts or line of business managers.

Prediction four: you can’t prepare for tomorrow with yesterday’s tools
We’re continuing to refine our platform to support the future of analytics. In 2014, we’ll release new functionality, upgrades and plug-ins to make it even easier and faster to move, blend and analyze relational and big data sources. We’re planning to improve the capabilities of the adaptive data layer and make it more secure and easy for customers to manage data flow. On the analytics side, we’re working to simplify data discovery on the fly for all business users and make it easier to find patterns and catch anomalies. In Pentaho Labs, we’ll continue to work with early adopters to cook up new technologies to bring things like predictive, machine data and real-time analytics into mainstream production.

As people in the business continue to see what’s possible with blended big data, I believe we’re going to witness some really exciting breakthroughs and results. I hope you’re as excited as I am about 2014!

Quentin Gallivan, CEO, Pentaho

Big-Data-2014-Predictions-Blog-Graphic


Weka goes BIG

December 4, 2013

funny_science_nerd_cartoon_character_custom_flyer-rb4a8aff0894a4e25932056b8852f8b18_vgvyf_8byvr_512.jpgThe beakers are bubbling more violently than usual at Pentaho Labs and this time predictive analytics is the focus.  The lab coat, pocket-protector and taped glasses clad scientists have turned their attention to the Weka machine learning software.

Weka, a collection of machine learning algorithms for predictive analytics and data mining, has a number of useful applications. Examples include, scoring credit risk, predicting downtime of machines and analyzing sentiment in social feeds.  The technology can be used to facilitate automatic knowledge discovery by uncovering hidden patterns in complex datasets, or to develop accurate predictive models for forecasting.

Organizations have been building predictive models to aid decision making for a number of years, but the recent explosion in the volume of data being recorded (aka “Big Data”) provides unique challenges for data mining practitioners. Weka is efficient and fast when running against datasets that fit in main memory, but larger datasets often require sampling before processing. Sampling can be an effective mechanism when samples are representative of the underlying problem, but in some cases the loss of information can negatively impact predictive performance.

To combat information loss, and scale Weka’s wide selection of predictive algorithms to large data sets, the folks at Pentaho Labs developed a framework to run Weka in Hadoop. Now the sort of tasks commonly performed during the development of a predictive solution – such as model construction, tuning, evaluation and scoring – can be carried out on large datasets without resorting to down-sampling the data. Hadoop was targeted as the initial distributed platform for the system, but the Weka framework contains generic map-reduce building blocks that can be used to develop similar functionality in other distributed environments.

If you’re a predictive solution developer or a data scientist, the new Weka framework is a much faster path to solution development and deployment.  Just think of the questions you can ask at scale!

To learn more technical details about the Weka Hadoop framework I suggest to read the blog, Weka and Hadoop Part 1, by Mark Hall, Weka core developer at Pentaho.

Also, check out Pentaho Labs to learn more about Predictive Analytics from Pentaho, and to see some of the other cool things the team has brewing.

Chuck Yarbrough
Technical Solutions Marketing


Pentaho wins Red Herring Global 100 Award

November 22, 2013

RH_global100Following Pentaho’s success this summer in winning the Red Herring North America competition, which recognizes the most promising private technology companies, Pentaho was invited to participate in this week’s Red Herring 2013 Top 100 Global event and award competition. And we’re pleased to announce that Pentaho won!

We’re honored to be recognized as a Red Herring 2013 Global Top 100 company!

Red Herring Global culminates a full year of work scouring the globe by the Red Herring editorial team and venture capitalists reviewing thousands of privately held companies in regional competitions around the world. The world’s top technology companies are selected based on financial performance, technology innovation and a variety of other metrics that have made the Red Herring Global 100 a mark of distinction for identifying the most promising private companies and entrepreneurs.

Pentaho was joined by a list of impressive company finalists for the Global 100 competition including Pentaho customers Spreadshirt GmbH and Active Broadband Networks (both ended up on the top 100 list as well – our heartfelt congratulations go out to these innovators!). Top private companies from the Red Herring’s Regional competitions in Europe, North America and Asia flew to the Red Herring Global forum in Los Angeles for the final competition on November 19 and 20.

Pentaho CEO, Quentin Gallivan, was asked to present at the forum, where he shared our point of view about how the big data and analytics markets are transforming, particularly with the need to easily blend big data with other data sources for better business insights. Quentin provided insights from our customer front lines about how Pentaho’s Big Data pilot projects are transitioning into widespread deployments with real business impact—with the most powerful insights coming from blending relational and big data sources.

The Red Herring Global forum concluded with an awards gala, where the 2013 Red Herring Global 100 companies were announced.

photo.JPGThis is great validation for Pentaho, as well as our customers, partners and community. Together we’re driving the future of analytics, and the Global 100 award provides a very solid foundation to build upon as we push the boundaries of Analytics with Storm, YARN, and predictive analytics in 2014.

Rosanne

Rosanne Saccone
Chief Marketing Officer
Pentaho


Integration Solutions are Key in Unlocking Big Data Benefits

November 6, 2013

VR_HorizontalLogoWe’re tremendously excited that Ventana Research is launching a new benchmark research on Big Data Integration.  As it becomes clearer to the market that Big Data has huge potential to deliver business value, identifying the right tools & technologies to manage it in the context of specific organizational goals is crucial.  The new research aims to take a step forward in this realm, as it promises to take a hard look at current and planned Big Data deployments, honing in on the solutions for accessing and transforming Big Data in a variety of different technical architectures and business use cases. It should provide great insight into ways that organizations can “unlock” Big Data’s potential.  See the press release here – Ventana Research Launches Big Data Integration Benchmark Research

For further context, we recommend reading a recent blog post from Ventana’s VP & Research Director, Tony Consentino titled, Big Data and Analytics Helps Business Transform and Gain Competitive Advantage. Consentino highlights the broad applicability of Big Data solutions across industries to empower organizations in the face of accelerating changes in markets, operations, & customer preferences.  The blog also cites staffing and training as the two biggest challenges to Big Data analytics, which underlines the importance of the new benchmark research in helping businesses identify which integration approaches will help accelerate time to value through usability by the broadest base of team members.

If you are interested in learning more about or participating in this benchmark research visit ventanaresearch.com/BigDataIntegration. There are several incentives for qualified research participants such as a $5 Amazon.com gift certificate and a complementary report valued at $995 as well as access to a free educational webinar on best practices from the benchmark research.

Finally, the Ventana Research Innovation Summit is next week, November 11-12 in Chicago. Let us know if you will be attending as we would love to meet-up!

Ben Hopkins
Product Marketing


#PPSUMM13

October 15, 2013

More than forty partners from around the globe attended Pentaho Partner Summit 2013 (#PPSUMM13), October 3-4 in the beautiful village of Sintra, Portugal, famous for its 19th-century Romantic architecture and landscapes. The following video summarizes the two days in under two minutes:

Informative sessions

To further help partners to maximize and monetize big data opportunities, the Summit included sales and technical training sessions focusing on different aspects of the big data ecosystem. During these sessions Steve Jenkins, Vice President EMEA of Hadoop distributor MapR, and Luca Olivari, Director of Business Development EMEA of MongoDB, the NoSQL database, explained how Pentaho supports these important and popular big data technologies.

Partner of the Year Award

The Pentaho partners that were the most successful over the last year, delivering outstanding customer success, technical innovation and commercial results, were honored at a gala dinner with the Pentaho Partner of the Year Award. This year’s winners were Italy’s BNova (EMEA), Brazil’s OnCase Brazil (LATAM) and Japan’s KSK (APAC).

 New partner program

The Pentaho Partner Summit 2013 was the perfect occasion to share more details on our new enhanced Partner Program in EMEA, Asia Pacific and Latin America. The new program consists of three levels: Registered Member, Certified Partner and Premium Partner. Registered Member is a new entry level that makes it easier for new partners to sign up to the partner program. The main partnership level, now called ‘Certified Partner’ has been simplified. To reward commitment, Certified Partners that offer advanced services and sell successfully are eligible to earn the top ‘Premium’ status. These new levels replace the previous Bronze, Gold and Platinum and make it easier for customers to select the right partner for their needs.

Learn more about the Pentaho Partner Program visit pentaho.com/partners


Highlights From Splunk .conf2013 – Machine Data Meets Big Business

October 4, 2013
Eddie.jpg

Eddie White, EVP Business Development, Pentaho

This week, Pentaho was on site for Splunk .conf2013 in Las Vegas and the show was buzzing with excitement. Organizations big and small shared a range of new innovations leveraging machine data.

Eddie White, executive VP of business development at Pentaho, shares his first-hand impressions and insights on the biggest news and trends coming out of .conf2013.

Q: Eddie, what are your impressions of this year’s Splunk conference?

There’s a different feel at the show this year — bigger companies and more business users attended this year. What traditionally has been more of an “IT Show,” has evolved to showcase real business use cases, success stories and post-deployment analysis. It’s apparent that machine data has turned a corner. The industry is moving well beyond simply logging of machine data. Users integrate, analyze and leverage their vast resource of device data for business intelligence and competitive advantage.

For example, on the first day ADP shared how they leverage big data for real-time insights. Yahoo! shared details on a deployment of Splunk Enterprise at multi-terabyte scale that is helping to better monitor and manage website properties. Intuit spoke on leveraging Splunk for diagnostics, testing, performance tuning and more. And on the second day, StubHub, Harvard University, Credit Suisse, Sears and Wipro were all featuring compelling uses for Splunk.

What was most exciting to me was the 50+ end users I spoke with who wanted learn how Pentaho blends data with and in Splunk. Our booth traffic was steady and heavy. Pentaho’s enhanced visualization and reporting demos were a hit not only with the IT attendees, but with the business users who are searching for ways to harness the power of their Splunk data for deeper insights. 

Q: Does attendance indicate a bigger/growing appetite for analysis of machine data?

Splunk is helping to uncover new information and insights – tapping into the myriad of data types Splunk can support as a data platform. It’s clearly making an impact in the enterprise. Yet as all these organizations increasingly turn to Splunk to collect, index and harness their machine-generated big data…there is tremendous opportunity for organizations to turn to Pentaho , a Splunk Powered Technology Partner, to tap and combine Splunk data with any other data source for deeper insights.

Q: How is the market developing for machine data analytics?

We are seeing the market here change from being driven by the technologists, to being driven by the business user.  The technology has advanced and now has the scale, the flexibility and the models to make real business impacts for the enterprise.  The use cases are clearly defined now and the technology fits the customer needs.  The level of collaboration between the major players like Pentaho, Splunk and Hadoop vendors now presents CIOs with real value.

Q: You were invited this year to speak on a CXO Panel addressing Big Data challenges and opportunities. What were some of the highlights?

The CXO panel was fantastic. It was quite an honor to present and be on a panel with four founders and “rock stars” in Big Data: Matt Pfeil (DataStax), M.C. Srivas (MapR), Ari Zilka (Hortonworks) and Amr Awadallah (Cloudera).

Over a panel session that ran for 90 minutes, we tackled subjects on big data challenges. We heard that Splunk users are dealing with quite a few of the same questions and challenges.

Business users and IT professionals just getting started are struggling with what project to pick first and first steps. My advice is to pick a real business use case and push us vendors to do a proof-of-concept with you, your team and to show quantifiable results in 30 days.

We also heard a lot of questions about which vendor has the right answer to their individual use scenarios and challenges. It was great to see all of the panelists on the same page in their response. No one vendor has all the answers. As I mentioned on the panel, if any Big Data player tells you they can solve all your Big Data problems, you should disqualify them! Users need Splunk, they need Pentaho and they need Hadoop.

Q: Taking a high level view of the conference, what trends can you identify?

There were two major trends taking center stage. Business people were asking business questions, and almost everyone was looking to map adoption to real business use cases.  And again, there’s a clear awareness that no one vendor can answer all of their questions. They are all looking at how to best assemble Hadoop, along with Pentaho and extend their use of Splunk with those technologies.

Q: Pentaho and Splunk are demonstrating the new Pentaho Business Analytics and Splunk Enterprise offering, providing a first look to conference attendees. What kind of reaction are you getting from the demos?

The reaction from the audiences was tremendous. We had two sets of reactions. The end user customers took the time to go in-depth with technology demos and asked questions like… where Splunk ends and where Pentaho begins?  The demo we showed drew the business user in too. It was a very powerful visualization of how we can enable a Splunk enterprise to solve business problems.

The Splunk sales teams who visited the booth and saw the demo were able to clearly discuss how to position a total solution for their customer.

Learn more about Splunk and Pentaho.

 


Pentaho 5.0 blends right in!

September 12, 2013

Dear Pentaho friends,

Ever since a number of projects joined forces under the Pentaho umbrella (over 7 years ago) we have been looking for ways to create more synergy across this complete software stack.  That is why today I’m exceptionally happy to be able to announce, not just version 5.0 of Pentaho Data Integration but a new way to integrate Data Integration, Reporting, Analyses, Dashboarding and Data Mining through one single interface called Data Blending, available in Pentaho Business Analytics 5.0 Commercial Edition

Data Blending allows a data integration user to create a transformation capable of delivering data directly to our other Pentaho Business Analytics tools (and even non-Pentaho tools).  Traditionally data is delivered to these tools through a relational database. However, there are cases where that can be inconvenient, for example when the volume of data is just too high or when you can’t wait until the database tables are updated.  This for example leads to a new kind of big data architecture with many moving parts:

Evolving Big Data Architectures

Evolving Big Data Architectures

From what we can see in use at major deployments with our customers, mixing Big Data, NoSQL and classical RDBS technologies is more the rule than the exception.

So, how did we solve this puzzle?

The main problem we faced early on was that the default language used under the covers, in just about any business intelligence user facing tool, is SQL.  At first glance it seems that the worlds of data integration and SQL are not compatible.  In DI we read from a multitude of data sources, such as databases, spreadsheets, NoSQL and Big Data sources, XML and JSON files, web services and much more.  However, SQL itself is a mini-ETL environment on its own as it selects, filters, counts and aggregates data.  So we figured that it might be easiest if we would translate the SQL used by the various BI tools into Pentaho Data Integration transformations. This way, Pentaho Data Integration is doing what it does best, not directed by manually designed transformations but by SQL.  This is at the heart of the Pentaho Data Blending solution.

MattCasters_Blog_graphic

The internals of Data Blending

In other words: we made it possible for you to create a virtual “database” with “tables” where the data actually comes from a transformation step.

To ensure that the “automatic” part of the data chain doesn’t become an impossible to figure out “black box”, we made once more good use of existing PDI technologies.  We’re logging all executed queries on the Data Integration server (or Carte server) so you have a full view of all the work being done:

Data Blending Transparency

Data Blending Transparency

In addition to this, the statistics from the queries can be logged and viewed in the operations data mart giving you insights into which data is queried and how often.

We sincerely hope that you like these new powerful options for Pentaho Business Analytics 5.0!

Enjoy!

Matt

–If you want to learn more about the new features in this 5.0 release, Pentaho is hosting a webinar and demonstration on September 24th – Two options to register:  EMEA & North America time zones.

Matt Casters
Chief Data Integration, Kettle founder, Author of Pentaho Kettle Solutions (Wiley)


Rackspace brings ETL to the Cloud with Pentaho: Hadoop Summit Q&A

June 27, 2013

This week Pentaho has been meeting with the movers and shakers of the Apache Hadoop community in San Jose, at the 6th annual Hadoop Summit. Pentaho and Rackspace are drawing attention on this final day of the show with the announcement of a partnership that brings ETL to the cloud. We’re introducing Rackspace Big Data, a powerful enterprise grade Hadoop as a Service solution. As the industry leader in cost effective data integration for Hadoop, Pentaho is proud to team with Rackspace, the industry leader in enterprise IAAS, to deliver this new era of big data in the cloud.

photo.JPG

L) Eddie White, EVP business development, Pentaho | R) Sean Anderson product marketing manager for cloud big data solutions, Rackspace Hosting

To learn more about the news, we’re talking today with Pentaho’s Eddie White, executive vice president of business development.

Give us a quick overview of this Rackspace news, and how Pentaho is involved.

Rackspace Big Data is an exciting Hadoop as a Service offering with full enterprise features. This is the next evolution in the big data ecosystem, delivering the ongoing structure to allow enterprise customers to choose a variety of consumption models over time. Customers can choose managed dedicated servers, and public, private or hybrid cloud options. Pentaho was chosen as the only Hadoop ETL / Data integration partner for this Cloud Tools Hadoop offering.

So is this a solution for enterprise customers looking to grow their big data operations?

Yes, absolutely. Hadoop as a Service is an attractive alternative for customers that need enterprise-level infrastructure support. Pentaho gives Rackspace a partner with the skills and talent on-board to deliver big data for production environments, along with the support and stability that Rackspace customers demand from their service-level agreements. Enterprises are looking for a Cloud partner with an enterprise-grade infrastructure to support running their business; not just test and development efforts.

What makes up this Hadoop as a Service model?

Together, Rackspace, Hortonworks and Pentaho have jointly delivered an offering that facilitates ease of use and ease of adoption of Hadoop as a Service. Rackspace Big Data includes the HortonWorks Data Platform for Hadoop; Pentaho Business Analytics as the ETL / Big Data Integration partner; and Karmasphere providing Hadoop analytics.

Rackspace excels at the enterprise IaaS model, and now they’ve partnered with Hortonworks and Pentaho to introduce an easy-to-use, consume-as-you-scale Hadoop as a Service offering – so customers can get started today, confident their solution will scale along with their big data needs. Rackspace chose to partner with Pentaho because it is the industry-leading Hadoop ETL and Big Data Analytics platform. Rackspace Big Data offers a range of models to meet any organization’s changing needs, from dedicated to hybrid, and for private and public clouds. And the offering ensures the ability to bi-directionally move data in and out of enterprise clusters, with minimal technical effort and cost.

What does Pentaho Data Integration bring to Rackspace Big Data?

Rather than speak for our partner, I’ll let Sean Anderson, Rackspace Hosting’s product marketing manager for cloud big data solutions, answer that. He sums up what Pentaho brings to the partnership nicely:

“Pentaho Data Integration is all about easing adoption and enhancing utilization of Rackspace big data platforms, with native, easy-to-use data integration. Pentaho is leading the innovation of Hadoop Integration and Analytics, and the upcoming cloud offering with Rackspace reduces the barriers to instant success with Hadoop, so customers can adopt and deploy quickly, delivering faster ROI,” said Anderson.

“Pentaho’s powerful data integration engine serves as a platform, enabling delivery of that content right into an enterprise’s pre-existing business intelligence and analytics tools,” continued Anderson. “Rackspace Big Data customers who require multiple data stores can leverage the ease of operation inherent in their visual ETL tool Pentaho provides. Customers will be able to complement their platform offering by adding the validated Pentaho tool via the Cloud Tools Marketplace.”

A key takeaway is that Rackspace Big Data customers may choose to bridge to the Pentaho Business Analytics platform. As an example, Pentaho’s full suite can be used where a Rackspace customer wants to use both Hortonworks and ObjectRocket. We bring the data in both of these databases to life for the Rackspace customer.

Why is Pentaho excited about this announcement?

This is exciting news because it is Pentaho’s first strategic cloud partnership. As the big data market has matured, it’s now time for production workloads to be moved over to Big Data Service offerings. Rackspace is the recognized leader providing the enterprise with IaaS, with an enterprise-grade support model. We see Rackspace and a natural partner for us to make our move into this space. We are market leaders in our respective categories with proven experience that enterprises trust for service, reliability, scalability and support. As the market for Hadoop and Big Data is developing and maturing, we see Rackspace as the natural strategic partner for Pentaho to begin providing Big Data / Hadoop as a Service.

MarketplaceHow can organizations buy Rackspace Big Data?

For anyone looking to leverage Hadoop as a Service, Rackspace Big Data is available directly from Rackspace. For more information and pricing visit: www.rackspace.com/big-data. Pentaho will also be in the Rackspace Cloud Tools marketplace.


“There is nothing more constant than change”—Heraclitus 535BC

June 26, 2013

13-090 Pentaho Labs logo v3

Change and more change. It’s been incredible watching the evolution of and innovation in the big data market.  A few years ago we were helping customers understand Hadoop and the value it could bring in analyzing large volumes of unstructured data. Flash-forward to today as we attend our third Hadoop Summit in San Jose and we see the advances customers have made in adopting these technologies in their production big data environments..

It’s the value of a continuum of innovation. As the market matures we are only limited by what we don’t leave ourselves open to.  Think for a minute about the next “big data,” because there will be one. We can’t anticipate what it look like, where it will come from or how much of it will be of value.  In the same way we couldn’t predict the advent of Facebook or Twitter.

We do know that innovation is a constant. Today’s big data will be tomorrow’s “traditional” data.

Pentaho’s announcement today of an adaptive big data layer and Pentaho Labs are in anticipation of just this type of change.  We’ve simplified for Pentaho and our customers the ability to leverage current and new big data technologies like Hadoop, NoSQL and specialized big data stores.

In the spirit of innovation (which stems from our open source history) we’ve established Pentaho Labs – our place for free thinking innovation that leads to new capabilities in our platform in areas like real time and predictive analytics.

Being a leader at the forefront of a disruptive and ever-changing market means embracing change and the innovation. That’s the future of analytics.

Donna Prlich
Senior Director, Product Marketing, Pentaho


Informatica jumps on the Pentaho bandwagon

June 12, 2013

Big-Data_web.jpgYou know that a technology megatrend has truly arrived when the large vendors start to jump on the bandwagon. Informatica recently announced Informatica Vibe™ – its new virtual data machine (VDM), an embeddable data management engine that allows developers to “Map Once, Deploy Anywhere,” including into Hadoop, without generating or writing code. According to Informatica, developers can instantly become Hadoop developers without having to acquire new skills. Sound familiar?

I applaud Informatica’s efforts – but not for innovating or changing the landscape in data integration.  What I applaud them for is recognizing that the landscape for data integration has indeed changed, and it was time for them to join the party. “Vibe” itself may be new, but it is not a new concept, nor unique to the industry.  In fact, Pentaho recognized the need for a modern, agile, adaptive approach to data integration for OEMs and customers. We pioneered the Kettle “design once, run anywhere” embeddable virtual data engine back in 2005. And let’s set the record straight – Pentaho extended its lightweight data integration capabilities to Hadoop over three years ago as noted in this 2010 press release.

Over the past three years, Pentaho has delivered on Big Data Integration with many successful Hadoop customers, such as BeachMint, MobileThink, TravelTainment and Travian Games and continued our innovation — with not only Hadoop but also NoSQL, Analytical Engines, and other specialized Big Data stores. We have added test, deploy and real time monitoring functionality.  The Pentaho engine is embedded in multiple SaaS, Cloud, and customer applications today such as Marketo, Paytronix, Sharable Ink and Soliditet, with many more on the horizon. Our VDM is completely customer extensible and open. We insulate customers from changes in their data volumes, types, sources, computing platforms, and user types.  In fact, what Informatica states as intention and direction with Vibe, Pentaho Data Integration delivers today, and we continue to lead in this new landscape.

VisualDataManagement

The Data Integration market has changed– the old, heavyweight, proprietary infrastructure players must adapt to current market demands. Agile, extensible, open, embeddable engines with pluggable infrastructures are the base, but it doesn’t end there. Companies of all sizes and verticals are requiring shorter development cycles, broad and deep big data ecosystem support, attractive price points and rich functionality, and all without vendor lock-in.  Informatica is adapting to play in the big data integration world by rebranding its products and signaling new direction.  Tony Baer, principal analyst at Ovum, summarizes this adaptation in his blog, “Informatica aims to get its vibe back.”

The game is on and Pentaho is at the forefront. We have very exciting big data integration news in store for you at the Hadoop Summit in Santa Clara on June 26-27 that unfortunately I have to keep the lid on for now. Stay tuned!

Richard

Richard Daley

Co-founder and chief strategy officer


Follow

Get every new post delivered to your Inbox.

Join 88 other followers