10 Great Reasons to Attend PentahoWorld

August 20, 2014
Reason #10

Reason #10

It’s hard to pick only ten reasons why PentahoWorld is going to be the must-attend event of the year for the data community, but we’ve given it a go! For full information on the event, sessions, speakers, celebrations, and registration please visit www.pentahoworld.com.

1: Stay on top of industry trends – If you are interested in the future of big data analytics, Orlando is the place to be on 8 – 10 October. Hear about Pentaho’s vision for tomorrow from the Pentaho product and technology leadership team.

2: Meet the Experts – We plan to have our own version of the Genius Bar at PentahoWorld. This is your opportunity to talk to the experts and get hands on help with your questions, ideas or just to meet face to face! Tap into their expertise for information that will help you and your organization excel when it comes to big data.

3: Test-drive product training – Would you like to enhance your skillset with Hadoop, NoSQL and Weka? We have picked our most popular training classes and top trainers to create specialized classes to help you build a solid set of skills in using these cutting edge technologies with Pentaho.

4: Ecosystem access – Pentaho believes that big data must play well with others. As the big data analytics ecosystem continues to expand and evolve, we have invited our technology partners to talk about integration, emerging technologies, and showcase solutions that are changing the way leading businesses are run.

5: It’s our 10th birthday – Of course there will be a party! Celebrations for our friends, partners, users, and community to say THANK YOU to everyone that has been with us for the journey.

6: Pentaho Excellence Awards – Could you be the Pentaho User of the Year? Get some well-deserved recognition for the business value you have driven with Pentaho. During the keynote, we’ll be celebrating our customers success awarding 6 category winners for the Pentaho Excellence Awards. Submissions have now closed and winners will be announced in early September – good luck to everyone that has entered!

7: Impressive agenda – Highlights include keynotes from industry leaders, interactive breakout sessions, topic tracks and hands-on labs. All this complemented with countless opportunities to learn, network and experience product development and innovation news! Whether you are new to Pentaho or a seasoned expert, there will be sessions to boost your skills and knowledge in analytics and big data.

8: Bring your data to life with valuable insights on data integration, big data, and embedded and advanced analytics. Your company has a plethora of data just waiting to drive business benefits. Find out how to get more, do more, and succeed more with your data projects.

9: Networking – Learn, share, discover. This is your chance to connect with Pentaho leaders, technology experts, fellow Pentaho users, community members, partners, and more.

10: Location, location, location – The Hilton Bonnet Creek hotel is located in a beautiful nature preserve and surrounded by Disney Theme Parks. The hotel has a globally ranked golf course, health and fitness spa, and its famous “lazy river” pool was recently ranked #1 by TripAdvisor’s Top 10 Fantastic Pools. Grab your inner tubes, we’ll see you on the river!

Early bird pricing ends soon on August 30th. Sign up now and pay just $695 – regular price $995

Register and find out more at www.pentahoworld.com

 


What’s new at Pentaho – Q2 2014 review

August 8, 2014

We are a very busy bunch at Pentaho. We are makers and doers, firmly focused on the future of analytics.

Sometimes it’s good to stop and reflect back on customer success stories, product developments and recent achievements. It is an exciting time for Pentaho with the first PentahoWorld worldwide users conference taking place in October, the recent release of Pentaho 5.1 and our users being recognized in national awards for the work they have done using Pentaho analytics.

Below is a summary of what’s new at Pentaho from Q2 2014: news, content, press mentions and blogs.

Press releases

  1. Pentaho Equips Companies to Easily Scale Big Data Operations, Regardless of IT Resources
  2. Introducing the Pentaho Excellence Awards
  3. Pentaho Data Science Pack Operationalizes Use of R and Weka
  4. Pentaho to Host First Worldwide Users’ Conference
  5. Bywaters’ Customers Save Money and Co2 with Pentaho Embedded Analytics
  6. Pentaho Business Analytics Certified on Cloudera 5 for Apache Hadoop in the Enterprise
  7. Head of IBM Analytics Product Strategy Joins Pentaho as Chief Product Officer

New Content added to Pentaho.com

eBooks

Pentaho Authored Whitepapers

Third Party Content

Recorded Webinars

Videos

Sound Clips

Press highlights

Pentaho 5.1 highlights

Data blending highlights

Blogs

Thought Leadership

  1. Dinosaurs Have Had Their Day
  2. Hadoop Summit 2014 – Big Data Keeps Getting Bigger
  3. Horses for courses
  4. Cloudera Stamp of Approval

Product or Campaign Focused

  1. Spark on Fire! Integrating Pentaho and Spark
  2. Introducing Pentaho 5.1 – Powering Big Data Analytics at Scale
  3. World Cup, Twitter sentiment and equity prices…any correlation?
  4. Good news, your data scientist just got a personal assistant
  5. I’ll Take Some Science with My Art

Company

  1. Recognizing and Rewarding Your Work: The Pentaho Excellence Awards
  2. Award time at Pentaho
  3. Why are WE excited about PentahoWorld?
  4. PentahoWorld 2014 – save the date
  5. Pentaho and MongoDB Rendezvous in SF

 


Preview of PentahoWorld Speakers – Cloudera, Forrester, NASDAQ

July 25, 2014
PentahoWorld

PentahoWorld

We’ve been busy curating compelling keynotes and general sessions for PentahoWorld that will help you develop a roadmap for scale, growth and success. Before the big reveal of all of our speakers, we wanted to give you a preview of some of the big data leaders and mavericks that will be speaking at PentahoWorld in Orlando, October 8-10.

We’ve selected keynotes that are big data innovators and leaders who’ve had the vision to see the potential of big data well ahead of the curve, while our general sessions will focus on more tactical information as you work to leverage the latest technologies, solve your biggest data challenges, and get the most out of your Pentaho implementation.

Preview of the First Wave of Speakers
Quentin GallivanChairman and Chief Executive Officer, Pentaho
Mike OlsonChief Strategy Officer, Cloudera
Christopher DziekanChief Product Officer, Pentaho
Mike GualtieriPrincipal Analyst, Forrester Research
Michael WeissSr Software Engineer, NASDAQ
John DinhLead Product Manager, NASDAQ

Next steps for you:

  1. Check the Agenda page often as we continue to add to the agenda
  2. Make sure to register today! Early bird ends August 31st (30% savings!)
  3. Are you using Pentaho in an interesting or innovative way? Apply for the Pentaho Excellence Awards – Deadline to submit is July 31st.

 


Attention Retail Banks, Its Time for Change!

July 18, 2014

Walhalla_(1896)_by_Max_Brückner

Retail banks, which have been wracked by scandals relating to PPI fraud, LIBOR rigging, unpopular bonus schemes and IT failures, need to think beyond upselling and cross-selling and consider how big data analytics can repair trust and improve the whole customer experience. In the article, Montetising Big Data in Retail Banks Starts with a Better Customer Experience, Davy Nys, VP of EMEA & APAC at Pentaho shares how retail banks can achieve  the ‘Valhalla’ of customer value pricing (CVP), or maximising the total value of a customer to a bank throughout all interactions and transactions. He explains how big data integration and analytics supports CVP in five ways:

  1. Supporting a two-way, 360-degree view
  2. Lower costs
  3. Smarter offers
  4. Customer friendly fraud detection
  5. Measuring customer sentiment

Learn more about how to achieve the ‘Valhalla’ of CVP read the full article here and register/attend the live webinar featuring Forrester Analyst Martha Bennett on the topic: Making the Most of your Data in the Financial Sector on July 22nd at 11am GMT


Pentaho 5.1 in LEGO

July 16, 2014

Two weeks ago we launched Pentaho Business Analytics 5.1. The new capabilities in Pentaho 5.1 support our ongoing strategy to make the hardest aspects of big data analytics faster, easier and more accessible to all. In honor of our Chief Architect, Will Gorman (also a LEGO Master Builder), we decided to have some fun with LEGO and now present to you the LEGO explanation of new features and functionality in Pentaho 5.1:

Lego_5.1

Direct Analytics on MongoDB – Unleash the value of MongoDB analytics for IT and Business Analysts with no coding required.

MongoDB5.1_2

Data Science Pack – Operationalize predictive models for R and Weka, drastically reducing data preparation time and effort.

Lego_RWeka

Full YARN Support – Reduce complexity for big data developers while leveraging the full power of Hadoop

YARN_5.1

Visit the 5.1 landing page to learn more about this release and access resources such as videos, data sheets, customer profiles and download.

 


World Cup Dashboard 2014 – in 15 minutes

July 3, 2014

dashboard fifa

Are you caught in the World Cup craze? Two of my passions are English football and analytics (hence I’m a SE at Pentaho based in London). So when it came time for this years’ World Cup, naturally I combined my passions to analyse who is going to win and what makes a winning team?

It turns out that a Big Data Analytics team in Germany tried to predict the winners based on massive data sets. Thus far three of their top five predicted teams have faltered. So what went wrong? Is Big Data not accurate? Are analytics not the answer?

Fret not. Exploring their methodology, their analysis was based from only one source of data. At Pentaho, we believe that the strongest insights come from blended data. We don’t just connect to large data sets; we make connecting to all data easy, regardless of format or location.

So why is my little World Copa Dashboard different (pictured above)? Its simple, intuitive and quick to create using Pentaho to access and organize the information I wanted — taking only 15 minutes. Yes, raw data to dashboard in 15 minutes! Here is how I did it:

I took 2010 FIFA World Cup data in CSV format and turned it into three panels of information in a format I wanted to see. It was my view into the action. I wanted more information beyond the basic winners and loosers, rather the interest to look behind the scenes (such as how and why) about one of the most spectacular sporting events on the planet. With analytics, I wanted to explore the following questions:

  1. How did strikers compare in the final outcome?
  2. Across a team, how hard did they work?
  3. What was the strategic approach of the winning teams versus the most industrious ones?

Using Pentaho’s built in mapping features I could visualise the countries with the ‘busiest’ forwards. This told me something about the work rate of the formations that country’s manager/coach was using. I could see the efficiency of the forward position players.

leading forward players

With the pivot table I could conditionally format the table to give me an at-a-glance view of the teams with goalkeeping performance, individual defending and attacking maneuvers.

player stats

 

Embedding the Official FIFA website into my dashboard also gave me access to FIFA’s own ‘trending’ players of this tournament. I could see historic performance versus the latest information.

official FIFA

One of the most vital aspects of analytics is offering the information in the right context for the user. With the rich visualisations of Pentaho Analyser it was simple for me to divide up the teams into a scatter plot showing me which ones worked the hardest for the duration they were in the tournament and which balance took the successful teams furthest.

teamstats

In summary, I got a behind the scenes look at my World Cup data in four different visualisations all combined into one dashboard easily, quickly and efficiently using Pentaho User Console and I was able to prepare the data into a format of my choosing rapidly using Pentaho Data Integration.

Turn your data into your own personal viewpoint with the power of Pentaho Business Analytics 5.1.

GO ENGLAND…for the next World Cup, perhaps? ;)

Zaf Khan
Presales Engineer
Pentaho


Spark on Fire! Integrating Pentaho and Spark

June 30, 2014

One of Pentaho’s great passions is to empower organizations to take advantage of amazing innovations in Big Data to solve new challenges using the existing skill sets they have in their organizations today.  Our Pentaho Labs’ innovations around natively integrating data engineering and analytics with Big Data platforms like Hadoop and Storm have already led dozens of customers to deploy next-generation Big Data solutions. Examples of these solutions include optimizing data warehousing architectures, leveraging Hadoop as a cost effective data refinery, and performing advanced analytics on diverse data sources to achieve a broader 360-degree view of customers.

Not since the early days of Hadoop have we seen so much excitement around a new Big Data technology as we see right now with Apache Spark.  Spark is a Hadoop-compatible computing system that makes big data analysis drastically faster, through in-memory computation, and simpler to write, through easy APIs in Java, Scala and Python.  With the second annual Spark Summit taking place this week in San Francisco, I wanted to share some of the early work Pentaho Labs and our partners over at Databricks are collaborating on to deeply integrate Pentaho and Spark for delivering high performance, Big Data Analytics solutions.

Big Data Integration on Spark

Big Data Integration on SparkAt the core of Pentaho Data Integration (PDI) is a portable ‘data machine’ for ETL which today can be deployed as a stand-alone Pentaho cluster or inside your Hadoop cluster though MapReduce and YARN.  The Pentaho Labs team is now taking this same concept and working on the ability to deploy inside Spark for even faster Big Data ETL processing.  The benefit for ETL designers is the ability to design, test and tune ETL jobs in PDI’s easy-to-use graphical design environment, and then run them at scale on Spark.  This dramatically lowers the skill sets required, increases productivity, and reduces maintenance costs when to taking advantage of Spark for Big Data Integration.

Advanced Analytics on Spark

Last year Pentaho Labs introduced a distributed version of Weka, Pentaho’s machine learning and data mining platform. The goal was to develop a platform-independent approach to using Weka with very large data sets by taking advantage of distributed environments like Hadoop and Spark. Our first implementation proved out this architecture by enabling parallel, in-cluster model training with Hadoop.

Advanced Analytics on Spark

We are now working on a similar level of integration with Spark that includes data profiling and evaluating classification and regression algorithms in Spark.  The early feedback from Pentaho Labs confirms that developing solutions on Spark is faster and easier than with MapReduce. In just a couple weeks of development, we have demonstrated the ability to perform in-cluster Canopy clustering and are very close to having k-means++ working in Spark as well!

Next up: Exploring Data Science Pack Integration with MLlib

MLlib is already one of the most popular technologies for performing advanced analytics on Big Data.  By integrating Pentaho Data Integration with Spark and MLlib, Data Scientists will benefit by having an easy-to-use environment (PDI) to prepare data for use in MLlib-based solutions.  Furthermore, this integration will make it easier for IT to operationalize the work of the Data Science team by orchestrating the entire end-to-end flow from data acquisition, to data preparation, to execution of MLlib-based jobs to sharing the results, all in one simple PDI Job flow.  To get a sense for how this integration might work, I encourage you to look at a similar integration with R we recently launched as part of the Data Science Pack for Pentaho Business Analytics 5.1.

Experiment Today with Pentaho and Spark!

You can experiment with Pentaho and Spark today for both ETL and Reporting.  In conjunction with our partners at Databricks, we recently certified for the following use cases combining Pentaho and Spark:

  • Reading data from Spark as part of an ETL workflow by using Pentaho Data Integration’s Table Input step with Apache Shark (Hive SQL layer runs on Spark)
  • Reporting on Spark data using Pentaho Reporting against Apache Shark

We are excited about this first step in what we both hope to be a collaborative journey towards deeper integration.

Jake Cornelius
Sr. Vice President, Product Management
Pentaho

 


Follow

Get every new post delivered to your Inbox.

Join 101 other followers