edo optimizes data warehouse, increases loyalty and targets new customers

February 10, 2014

edo

What do you do when you need to track, store, blend, and analyze over 6 billion financial data transactions with the outlook of daily growth by the millions? edo Interactive, inc is a digital advertising company that leverages payment networks to connect brands with consumers. Their legacy data integration and analysis system took more than 27 hours to run, meaning that meeting daily Service Level Agreements was nearly impossible. However, after only a few weeks of implementing a data distribution on Hadoop, with Pentaho for data integration, edo Interactive was able to reduce its processing time to less than 8 hours, often as little as 2 hours.

Minimum timesaving’s of 70% quickly precipitated cost savings. With an optimized data warehouse, edo and its clients also spend less time navigating IT barriers. Pentaho’s graphical user interface, removes cumbersome coding of batch process jobs, enabling sophisticated and simplified conversion of data from PostgreSQL to Hadoop, Hive and HBase. edo and its clients quickly gain insights to customer preferences, refine marketing strategies and provide their customers with improved experience and satisfaction.

Edo Interactive successfully navigated many of the obstacles faced when implementing a big data environment and created a lasting and scalable solution. Their vision to provide end-users a better view of their customers has helped shape a new data architecture and embedded analytics capabilities.

To learn more about edo’s Big Data vision and success, read their customer success overview and case study on Pentaho.com. We are excited to announce that Tim Garnto, SVP of Product Engineering at edo, will share his story live when he presents at O’Reilly + Strata on Thursday, February 13th in Santa Clara (11:30AM, Ballroom G).

Strata Santa Clara is already sold out! If you are interested to learn more about edo’s Big Data deployment, leave your questions in the comments section below and we will ask Tim during his speaking session at Strata.

Ben Mayer
Customer Marketing
Pentaho


Bring your Big Data to Life With Pentaho at Strata Santa Clara

January 30, 2014

horizon_bigdatalife

If you are like most Enterprise IT decision makers, there’s a 50/50 chance you are already knee deep into Big Data or on a path to figuring out how to get started. One of the “must attend” conferences for anyone involved in Big Data is the O’Reilly Strata Conference (Santa Clara, February 11-15, 2014).

Join Us!

Pentaho is excited to return as a sponsor this year and we have a number of ways you can learn more about getting the most out of your Big Data initiatives.

The Pentaho team of executives, technologist and data scientist will be on hand to share the latest big data innovations from Pentaho Labs such as integration with Apache Hadoop YARN and Storm. Come get answers to your all of your big data integration and analytics question. Let us help you bring your Big Data to life!

Below is a list of all activities for Pentaho in and around the conference. Register with code Pentaho20 and receive 20% off registration.

Exhibit booth

You will find the Pentaho team in the Sponsor Pavilion at booth 710 (located near the O’reilly Media booth). Learn all about how Pentaho can help bring your Big Data to life! Don’t forget to get your Pentaho t-shirt and enter for the chance to win a Go Pro camera.

Meetups

Big Data Science Meet-up at Strata Conference

  • Monday, 2/10 at 5:30-9:30 in Ballroom E
  • Nick Gonzalez, Data Scientist at Pentaho will speak about Real World Big Data Prescriptive Analytics
  • Today’s large and convoluted data landscape coupled with the abundance of available computing resources presents unique opportunities for data scientists around the world. To remain competitive in this landscape, we must go beyond generating predictions to generating solutions from big data that are driven by actions derived from data driven predictions. And we have to do this as fast as possible.  This is the real world of big data prescriptive analytics. This talk will address each one of these challenges and present technical solutions and algorithms to address them.  By the end of this presentation each individual solution will come together in a symphony of code and hardware to form a unified automated process that is the backbone of a successful big data prescriptive analytics solution.

Breakout Sessions

Getting There from Here: Moving Data Science into the Boardroom

  • Rosanne Saccone (Pentaho), Scott Chastain (SAS), Chris Selland (HP Vertica)
  • Tuesday, 2/11 at 11:15 on the Data Driven Business Track, Ballroom CD
  • Pundits and analysts agree—the data-driven enterprise is here to stay. But how will companies balance analysis with action? Will optimization of the current model leave firms more vulnerable than ever to disruption by what’s new and unpredictable? And how do we balance legacy investments in data warehousing and business intelligence with emerging technologies for massive, real-time data processing? Join Scott Chastain, Roseanne Saccone, Chris Selland, and Strata Chair Alistair Croll for a look at the practical concerns facing tomorrow’s data-driven business.

Lessons from the Trenches: edo Interactive Leverages Hadoop to Build Customer Loyalty

  • Thursday, 2/13 at 11:30am, Ballroom G
  • Tim Garnto (edo) & Rob Rosen (Pentaho)
  • Hadoop presents as an enabling technology to better understand customer preferences and behaviors, but organizations often struggle with time-consuming data preparation and analytics processes. edo Interactive – a leader in providing card-linked offers to financial services and retailers – shares how they drive agile, improved decision-making by complementing native Hadoop technologies with analytical databases and ETL optimization and data visualization solutions from vendors such as Pentaho.

We hope to see you soon at Strata in Santa Clara. If you would prefer a private meeting with Pentaho at the conference send us a message via our contacts page or direct message us on twitter @Pentaho.


Analyze 10 years of Chicago Crime with Pentaho, Cloudera Search and Impala

December 23, 2013

Hadoop is a complex technology stack and many people getting started with Hadoop spend an inordinate amount of time focusing on operational aspects – getting the cluster up and running, obtaining foundational training, and ingesting data. Consequently it can be difficult to get a good picture of the true value that Hadoop provides, namely unlocking insight across multiple data streams that add valuable context to the transactional history comprising most of the core data in the enterprise.

At Strata Hadoop World in October, Pentaho’s Lord of 1’s and 0’s or CTO, James Dixon, unveiled a powerful demonstration of the true value that Hadoop – combined with enabling technology from Pentaho and our partner Cloudera – can provide. He took a publicly available data set provided by the City of Chicago and built a demo around it that enables nontechnical end-users to understand how crime patterns have changed over time in Chicago, unlocking insight into the type of crimes being committed in different areas of the city – not only historically but also broken down by time of day and day of week. As a result, citizenry as well as law enforcement have a much better sense of what to expect on the streets of Chicago from the insight the demonstration provides.

In the demo, end-users start with a dashboard that provides a high-level understanding of the mix of crimes historically committed on the streets of Chicago over the last ten years. Watch the demo here:

This kind of top-to-bottom understanding of (in this case) crime patterns is uniquely enabled by the capability Pentaho delivers to the market, combining dashboarding, analytics and data integration into one easily-embedded platform that leverages blending across multiple data sets.

The deep understanding that Pentaho’s solution delivers to end-users is enabled by two key technologies from Cloudera: Cloudera Search and Impala. The original data set provided by the City of Chicago was loaded into a Cloudera Hadoop cluster using Pentaho’s data integration tool, Pentaho Data Integration (“PDI”). End-user drilldown is powered by Cloudera Search, which executes a faceted search on behalf of Pentaho’s dashboard. Once an area of interest has been located, Cloudera’s Impala executes low-latency performance of SQL on the raw data stored in the Hadoop cluster to bring up individual crime records.

Although Hadoop is often perceived as a geek’s playground, the power of Pentaho’s business-friendly interface is readily apparent when engaging this demo. Unlocking the power of Hadoop can be as simple as engaging Pentaho’s integrated approach to analytics together with Cloudera’s foundational platform to deliver an integrated solution whose value is apparent to nontechnical executives wondering whether Hadoop is the right choice for a key initiative.

Rob Rosen
Field Big Data Lead
Pentaho


Big Data, Big Revenue for Marketers

December 12, 2013

Why might Big Data mean millions for marketing?  Because it has the potential to create a more complete picture of the buyer, thereby empowering marketers to more effectively deliver the right message to the right individual at the right time – and ultimately increase sales.  In the following brief video from DMA 2013, Marketo VP/Co-founder Jon Miller and Pentaho CMO Rosanne Saccone provide a crash course on what Big Data means for marketers.  It covers:

  • The defining characteristics of Big Data – Velocity, Variety, & Volume
  • How marketers can leverage Big Data to blend operational information (CRM, ERP) and online data (web activity, social networking interactions) for new insights
  • Sample Big Data use cases that organizations are green-lighting today to optimize customer interactions and drive marketing’s contribution to revenue

Note that this is an excerpt from a larger presentation – for the full video please click here.

We’d also recommend this blog post by Jon Miller for more context on Big Data in marketing.

For additional compelling use cases that leverage Big Data for marketing and other functions, see here.

Ben Hopkins
Product Marketing
Pentaho


Big Data 2014: Powering Up the Curve

December 5, 2013

Last year, I predicted that 2013 would be the year big data analytics started to go into mainstream deployment and the research we recently commissioned with Enterprise Management Consultants indicates that’s happened. What really surprised me though is the extent to which the demand for data blending has powered up the curve and I believe this trend will accelerate big data growth in 2014.

Prediction one: The big data ‘power curve’ in 2014 will be shaped by business users’ demand for data blending
Customers like Andrew Robbins of Paytronix and Andrea Dommers-Nilgen of TravelTainment, who recently spoke about their Pentaho projects at events in NY and London, both come from the business side and are achieving specific goals for their companies by blending big and relational data. Business users like these are getting inspired by the potential to tap into blended data to gain new insights from a 360 degree customer view, including the ability to analyze customer behavior patterns and predict the likelihood that customers will take advantage of targeted offers.

Prediction two: big data needs to play well with others!
Historically, big data projects have largely sat in the IT departments because of the technical skills needed and the growing and bewildering array of technologies that can be combined to build reference architectures. Customers must choose from the various commercial and open source technologies including Hadoop distributions, NoSQL databases, high-speed databases, analytics platforms and many other tools and plug-ins. But they also need to consider existing infrastructure including relational data and data warehouses and how they’ll fit into the picture.

The plus side of all this choice and diversity is that after decades of tyranny and ‘lock-in’ imposed by enterprise software vendors, in 2014, even greater buying power will shift to customers. But there are also challenges. It can be cumbersome to manage this heterogeneous data environment involved with big data analytics. It also means that IT will be looking for Big Data tools to help deploy and manage these complex emerging reference architectures, and to simplify them.  It will be incumbent on the Big Data technology vendors to play well with each other and work towards compatibility. After all, it’s the ability to access and manage information from multiple sources that will add value to big data analytics.

Prediction three: you will see even more rapid innovation from the big data open source community
New open source projects like Hadoop 2.0 and YARN, as the next generation Hadoop resource manager, will make the Hadoop infrastructure more interactive. New open source projects like STORM, a streaming communications protocol, will enable more real-time, on-demand blending of information in the big data ecosystem.

Since we announced the industry’s first native Hadoop connectors in 2010, we’ve been on a mission to make the transition to big data architectures easier and less risky in the context of this expanding ecosystem. In 2013 we made some massive breakthroughs towards this, starting with our most fundamental resource, the adaptive big data layer. This enables IT departments to feel smarter, safer and more confident about their reference architectures and open up big data solutions to people in the business, whether they be data scientists, data analysts, marketing operations analysts or line of business managers.

Prediction four: you can’t prepare for tomorrow with yesterday’s tools
We’re continuing to refine our platform to support the future of analytics. In 2014, we’ll release new functionality, upgrades and plug-ins to make it even easier and faster to move, blend and analyze relational and big data sources. We’re planning to improve the capabilities of the adaptive data layer and make it more secure and easy for customers to manage data flow. On the analytics side, we’re working to simplify data discovery on the fly for all business users and make it easier to find patterns and catch anomalies. In Pentaho Labs, we’ll continue to work with early adopters to cook up new technologies to bring things like predictive, machine data and real-time analytics into mainstream production.

As people in the business continue to see what’s possible with blended big data, I believe we’re going to witness some really exciting breakthroughs and results. I hope you’re as excited as I am about 2014!

Quentin Gallivan, CEO, Pentaho

Big-Data-2014-Predictions-Blog-Graphic


Weka goes BIG

December 4, 2013

funny_science_nerd_cartoon_character_custom_flyer-rb4a8aff0894a4e25932056b8852f8b18_vgvyf_8byvr_512.jpgThe beakers are bubbling more violently than usual at Pentaho Labs and this time predictive analytics is the focus.  The lab coat, pocket-protector and taped glasses clad scientists have turned their attention to the Weka machine learning software.

Weka, a collection of machine learning algorithms for predictive analytics and data mining, has a number of useful applications. Examples include, scoring credit risk, predicting downtime of machines and analyzing sentiment in social feeds.  The technology can be used to facilitate automatic knowledge discovery by uncovering hidden patterns in complex datasets, or to develop accurate predictive models for forecasting.

Organizations have been building predictive models to aid decision making for a number of years, but the recent explosion in the volume of data being recorded (aka “Big Data”) provides unique challenges for data mining practitioners. Weka is efficient and fast when running against datasets that fit in main memory, but larger datasets often require sampling before processing. Sampling can be an effective mechanism when samples are representative of the underlying problem, but in some cases the loss of information can negatively impact predictive performance.

To combat information loss, and scale Weka’s wide selection of predictive algorithms to large data sets, the folks at Pentaho Labs developed a framework to run Weka in Hadoop. Now the sort of tasks commonly performed during the development of a predictive solution – such as model construction, tuning, evaluation and scoring – can be carried out on large datasets without resorting to down-sampling the data. Hadoop was targeted as the initial distributed platform for the system, but the Weka framework contains generic map-reduce building blocks that can be used to develop similar functionality in other distributed environments.

If you’re a predictive solution developer or a data scientist, the new Weka framework is a much faster path to solution development and deployment.  Just think of the questions you can ask at scale!

To learn more technical details about the Weka Hadoop framework I suggest to read the blog, Weka and Hadoop Part 1, by Mark Hall, Weka core developer at Pentaho.

Also, check out Pentaho Labs to learn more about Predictive Analytics from Pentaho, and to see some of the other cool things the team has brewing.

Chuck Yarbrough
Technical Solutions Marketing


Pentaho wins Red Herring Global 100 Award

November 22, 2013

RH_global100Following Pentaho’s success this summer in winning the Red Herring North America competition, which recognizes the most promising private technology companies, Pentaho was invited to participate in this week’s Red Herring 2013 Top 100 Global event and award competition. And we’re pleased to announce that Pentaho won!

We’re honored to be recognized as a Red Herring 2013 Global Top 100 company!

Red Herring Global culminates a full year of work scouring the globe by the Red Herring editorial team and venture capitalists reviewing thousands of privately held companies in regional competitions around the world. The world’s top technology companies are selected based on financial performance, technology innovation and a variety of other metrics that have made the Red Herring Global 100 a mark of distinction for identifying the most promising private companies and entrepreneurs.

Pentaho was joined by a list of impressive company finalists for the Global 100 competition including Pentaho customers Spreadshirt GmbH and Active Broadband Networks (both ended up on the top 100 list as well – our heartfelt congratulations go out to these innovators!). Top private companies from the Red Herring’s Regional competitions in Europe, North America and Asia flew to the Red Herring Global forum in Los Angeles for the final competition on November 19 and 20.

Pentaho CEO, Quentin Gallivan, was asked to present at the forum, where he shared our point of view about how the big data and analytics markets are transforming, particularly with the need to easily blend big data with other data sources for better business insights. Quentin provided insights from our customer front lines about how Pentaho’s Big Data pilot projects are transitioning into widespread deployments with real business impact—with the most powerful insights coming from blending relational and big data sources.

The Red Herring Global forum concluded with an awards gala, where the 2013 Red Herring Global 100 companies were announced.

photo.JPGThis is great validation for Pentaho, as well as our customers, partners and community. Together we’re driving the future of analytics, and the Global 100 award provides a very solid foundation to build upon as we push the boundaries of Analytics with Storm, YARN, and predictive analytics in 2014.

Rosanne

Rosanne Saccone
Chief Marketing Officer
Pentaho


Integration Solutions are Key in Unlocking Big Data Benefits

November 6, 2013

VR_HorizontalLogoWe’re tremendously excited that Ventana Research is launching a new benchmark research on Big Data Integration.  As it becomes clearer to the market that Big Data has huge potential to deliver business value, identifying the right tools & technologies to manage it in the context of specific organizational goals is crucial.  The new research aims to take a step forward in this realm, as it promises to take a hard look at current and planned Big Data deployments, honing in on the solutions for accessing and transforming Big Data in a variety of different technical architectures and business use cases. It should provide great insight into ways that organizations can “unlock” Big Data’s potential.  See the press release here – Ventana Research Launches Big Data Integration Benchmark Research

For further context, we recommend reading a recent blog post from Ventana’s VP & Research Director, Tony Consentino titled, Big Data and Analytics Helps Business Transform and Gain Competitive Advantage. Consentino highlights the broad applicability of Big Data solutions across industries to empower organizations in the face of accelerating changes in markets, operations, & customer preferences.  The blog also cites staffing and training as the two biggest challenges to Big Data analytics, which underlines the importance of the new benchmark research in helping businesses identify which integration approaches will help accelerate time to value through usability by the broadest base of team members.

If you are interested in learning more about or participating in this benchmark research visit ventanaresearch.com/BigDataIntegration. There are several incentives for qualified research participants such as a $5 Amazon.com gift certificate and a complementary report valued at $995 as well as access to a free educational webinar on best practices from the benchmark research.

Finally, the Ventana Research Innovation Summit is next week, November 11-12 in Chicago. Let us know if you will be attending as we would love to meet-up!

Ben Hopkins
Product Marketing


#PPSUMM13

October 15, 2013

More than forty partners from around the globe attended Pentaho Partner Summit 2013 (#PPSUMM13), October 3-4 in the beautiful village of Sintra, Portugal, famous for its 19th-century Romantic architecture and landscapes. The following video summarizes the two days in under two minutes:

Informative sessions

To further help partners to maximize and monetize big data opportunities, the Summit included sales and technical training sessions focusing on different aspects of the big data ecosystem. During these sessions Steve Jenkins, Vice President EMEA of Hadoop distributor MapR, and Luca Olivari, Director of Business Development EMEA of MongoDB, the NoSQL database, explained how Pentaho supports these important and popular big data technologies.

Partner of the Year Award

The Pentaho partners that were the most successful over the last year, delivering outstanding customer success, technical innovation and commercial results, were honored at a gala dinner with the Pentaho Partner of the Year Award. This year’s winners were Italy’s BNova (EMEA), Brazil’s OnCase Brazil (LATAM) and Japan’s KSK (APAC).

 New partner program

The Pentaho Partner Summit 2013 was the perfect occasion to share more details on our new enhanced Partner Program in EMEA, Asia Pacific and Latin America. The new program consists of three levels: Registered Member, Certified Partner and Premium Partner. Registered Member is a new entry level that makes it easier for new partners to sign up to the partner program. The main partnership level, now called ‘Certified Partner’ has been simplified. To reward commitment, Certified Partners that offer advanced services and sell successfully are eligible to earn the top ‘Premium’ status. These new levels replace the previous Bronze, Gold and Platinum and make it easier for customers to select the right partner for their needs.

Learn more about the Pentaho Partner Program visit pentaho.com/partners


Highlights From Splunk .conf2013 – Machine Data Meets Big Business

October 4, 2013
Eddie.jpg

Eddie White, EVP Business Development, Pentaho

This week, Pentaho was on site for Splunk .conf2013 in Las Vegas and the show was buzzing with excitement. Organizations big and small shared a range of new innovations leveraging machine data.

Eddie White, executive VP of business development at Pentaho, shares his first-hand impressions and insights on the biggest news and trends coming out of .conf2013.

Q: Eddie, what are your impressions of this year’s Splunk conference?

There’s a different feel at the show this year — bigger companies and more business users attended this year. What traditionally has been more of an “IT Show,” has evolved to showcase real business use cases, success stories and post-deployment analysis. It’s apparent that machine data has turned a corner. The industry is moving well beyond simply logging of machine data. Users integrate, analyze and leverage their vast resource of device data for business intelligence and competitive advantage.

For example, on the first day ADP shared how they leverage big data for real-time insights. Yahoo! shared details on a deployment of Splunk Enterprise at multi-terabyte scale that is helping to better monitor and manage website properties. Intuit spoke on leveraging Splunk for diagnostics, testing, performance tuning and more. And on the second day, StubHub, Harvard University, Credit Suisse, Sears and Wipro were all featuring compelling uses for Splunk.

What was most exciting to me was the 50+ end users I spoke with who wanted learn how Pentaho blends data with and in Splunk. Our booth traffic was steady and heavy. Pentaho’s enhanced visualization and reporting demos were a hit not only with the IT attendees, but with the business users who are searching for ways to harness the power of their Splunk data for deeper insights. 

Q: Does attendance indicate a bigger/growing appetite for analysis of machine data?

Splunk is helping to uncover new information and insights – tapping into the myriad of data types Splunk can support as a data platform. It’s clearly making an impact in the enterprise. Yet as all these organizations increasingly turn to Splunk to collect, index and harness their machine-generated big data…there is tremendous opportunity for organizations to turn to Pentaho , a Splunk Powered Technology Partner, to tap and combine Splunk data with any other data source for deeper insights.

Q: How is the market developing for machine data analytics?

We are seeing the market here change from being driven by the technologists, to being driven by the business user.  The technology has advanced and now has the scale, the flexibility and the models to make real business impacts for the enterprise.  The use cases are clearly defined now and the technology fits the customer needs.  The level of collaboration between the major players like Pentaho, Splunk and Hadoop vendors now presents CIOs with real value.

Q: You were invited this year to speak on a CXO Panel addressing Big Data challenges and opportunities. What were some of the highlights?

The CXO panel was fantastic. It was quite an honor to present and be on a panel with four founders and “rock stars” in Big Data: Matt Pfeil (DataStax), M.C. Srivas (MapR), Ari Zilka (Hortonworks) and Amr Awadallah (Cloudera).

Over a panel session that ran for 90 minutes, we tackled subjects on big data challenges. We heard that Splunk users are dealing with quite a few of the same questions and challenges.

Business users and IT professionals just getting started are struggling with what project to pick first and first steps. My advice is to pick a real business use case and push us vendors to do a proof-of-concept with you, your team and to show quantifiable results in 30 days.

We also heard a lot of questions about which vendor has the right answer to their individual use scenarios and challenges. It was great to see all of the panelists on the same page in their response. No one vendor has all the answers. As I mentioned on the panel, if any Big Data player tells you they can solve all your Big Data problems, you should disqualify them! Users need Splunk, they need Pentaho and they need Hadoop.

Q: Taking a high level view of the conference, what trends can you identify?

There were two major trends taking center stage. Business people were asking business questions, and almost everyone was looking to map adoption to real business use cases.  And again, there’s a clear awareness that no one vendor can answer all of their questions. They are all looking at how to best assemble Hadoop, along with Pentaho and extend their use of Splunk with those technologies.

Q: Pentaho and Splunk are demonstrating the new Pentaho Business Analytics and Splunk Enterprise offering, providing a first look to conference attendees. What kind of reaction are you getting from the demos?

The reaction from the audiences was tremendous. We had two sets of reactions. The end user customers took the time to go in-depth with technology demos and asked questions like… where Splunk ends and where Pentaho begins?  The demo we showed drew the business user in too. It was a very powerful visualization of how we can enable a Splunk enterprise to solve business problems.

The Splunk sales teams who visited the booth and saw the demo were able to clearly discuss how to position a total solution for their customer.

Learn more about Splunk and Pentaho.

 


Follow

Get every new post delivered to your Inbox.

Join 96 other followers