Tuesday, 29 January 2013

OTN Developer Day - Free Oracle Big Data Workshop

If you want to get hands-on time with Oracle's big data technology stack then we are rolling out, in collaboration with OTN, a new complementary one-day hands-on workshop - yes it really is FREE! The date will is Wednesday, February 20, 2013 at Oracle's HQ in Redwood Shores, California.

The workshop will be based around the MoviePlex demonstration application that was first shown at last year's OpenWorld and you will get hands-on time with the source code and we will show you how to develop some of the key features of this application.

What is Oracle MoviePlex?
The Oracle MoviePlex is a fictional on-line movie streaming company and like many other on-line stores, they needed a cost effective approach to tackle their “big data” challenges. They recently implemented Oracle’s Big Data Platform to better manage their business, identify key opportunities and enhance customer satisfaction. The key challenge for Oracle Movieplex is how to manage and process massive volumes of unstructured data flowing in to their environment.

The business users at Oracle MoviePlex want to use their big data platform to help them explore some new business opportunities:
  • Make the right movie offers at the right time?
  • Better understand the viewing trends of various customer segments?
  • Optimize marketing spend by targeting customers with optimal promotional offers?
  • Minimize infrastructure spend by understanding bandwidth usage over time?
Below is a slideshow that will help provide some background and more general information about this application: (Thanks to Marty Gubar, Director of Product Management Big Data, for overseeing the development of the MoviePlex application, putting together the workshop and for creating the slides below)

What will you learn at this workshop?
The overall objective for this workshop is to show how you can develop a low latency, personalized recommendations environment that leverages Oracle's advanced analytic capabilities. During the workshop you will learn from technical experts who will demonstrate the following core technical topics:
  • Write MapReduce on Oracle’s Big Data Platform
  • Manage a Big Data environment
  • Access Oracle NoSQL Database
  • Manage Oracle NoSQL DB Cluster
  • Use data from a Hadoop Cluster with Oracle
  • Develop analytics on big data
Don't delay - register today to learn these key big data skills which you can immediately put to use within your organization.Space is limited so do not delay clicking on this link [here] to register.

Friday, 25 January 2013

2013 - The year of In-Database Analytics!

Many of our customers spent 2012 kick-starting big data projects. Based on the number of analyst reports, news articles and general chatter on the web I think it is fair to say that 2012 was the Year of Big Data. To help our customers plan and coordinate their big data projects we developed a very simple high level five step workflow for big data (Stream-Acquire-Organize-Analyze-Decide) as shown here:

Many of the projects that our customers stated in 2012 have at least completed the first three steps (with some having completed stages 4 and 5 and now looping back to stage 1 to kick-start the next round of data enrichment and knowledge discovery):
1) STREAM - all the relevant data streams have been identified and the APIs coded to collect the data on a regular basis
2) ACQUIRE - these data streams are increasing landing on our Big Data Appliance and that is great news!
3) ORGANIZE - many project teams have completed the first round of light-touch transformations and made their new data sets available for analysis. Typically this is being done using R, Java and our Big Data Connectors
We can see that during 2012 an awful lot was achieved. For 2013 we are going to see everyone focusing on the ANALYZE phase of the workflow and my prediction for 2013 is that this will be the year of "Big Data Analytics". The good news for Oracle customers is that you do not have to wait until 2015 to analyse your Hadoop data.
Screen Shot 2013 01 25 at 12 17 01
The most important thing is that customers must learn the key lesson from those old days of data warehousing There are two ways to manage ANALYZE the analyse phase:
  1. Use subject-specific specialised analytic engines and take the data to the analysis
  2. Use an analytically rich database and take the analysis to the data
1) Taking the data to the analysis
The danger for many organizations is that in delivering this type solution to the business they simply create “analytic silos” that are designed to resolve specific business problems. These analytic silos create unnecessary cost, they increase complexity and cause increased levels of data movement across the network as data is pushed in and results pulled out from each silo. This continuous movement of data creates time delays in being able to view results because business questions usually require multiple levels of analysis using lots of different types of analytical functions. The longer it takes to arrive at an answer the greater the chance you will make the wrong business decisions or completely miss out on an significant opportunity.
As before each of these specialised platforms have their own proprietary engine, tools and languages and this makes it difficult to kick-start and grow a project. Broadening the analysis can be complex and in many cases impossible. There has to be a better way to analyse data!
2) Taking the analysis to the data with In-Database Analytics
The end game for all big data projects (whether they realise it or not when the project first starts out) is to deliver an environment that offers the following: a broad range of analytical tools that can analyse all types of data and be accessed using existing skills and tools.  What you need is a single place to run your analysis against all your data so your business users can easily apply layer after layer of analysis. This ensures that the process of transforming data into insight delivers the right data at the right time, to the right person and on the right device.
To get the most from your data streams it is important that your analysis has the ability to incorporate the broadest range of analytical functions and that these can be applied to all your data steams with results delivered in a real-time. The ANALZYZE phase in our workflow leverages the Oracle Database. The Oracle Database started with a rich set of built-in SQL analytic features that allowed developers to process data directly inside the database using standard SQL syntax rather than having to move data to a separate platform, process it using a proprietary language and then return the results to the database.  Oracle's approach to analytics is to take the "analysis to the data", i.e. Oracle provides in-database analytics. In-database analytics offers some important advantages over the three other options outlined above:
  • Reduced latency - data can be analyzed in-place
  • Reduced risk – a single set of data security policies can be applied across all types of analysis
  • Increased reusability – all data types are accessible via SQL making it very easy to re-use analysis and analytical workflows across many tools: ETL, business intelligence reports and dashboards and operational applications
With each release of the database the types of data supported by the in-database analytics and the types of SQL analytics available to process that data continues to expand. The latest release of the Oracle Database covers a wide variety of data types and structures, including: numerical, text, images, videos, voice, XML, network, spatial, semantic and multi-dimensional. There is a wide a range of analytical features to support these data types that allow users to explore and layer their analysis. The picture below provides an overview of the rich analytical ecosystem within the Oracle Database:
InDB Analytics
Why is in-database analytics so important?
By making all the data available from inside the data warehouse and moving the analytics to the data Oracle is able to provide a wide range of analytical functions and features and ensure data governance and security is enforced at every stage of the analytic process, while at the same time providing timely delivery of analytical results. Oracle’s single integrated analytic platform offers many advantages.
To help explain those advantages  I am busy writing a whitepaper on precisely this topic and it should be released shortly. The whitepaper will explore the points I have highlighted above and will explain how Oracle can help you transform all your data into real actionable insight. When you change the way you analyse your data by moving the algorithms to the data rather than the traditional approach of extracting the data and moving it to the algorithms for analysis, it CHANGES EVERYTHING, including your business. You can know more about your products, your operations and your customers by having that insight delivered at the right time, to the right person and on the right device.
As part of the paper I will outline how some of Oracle’s industry leading customers are already using our in-database analytics today to help them understand more about their products, operations and customers and how many of them are layering different types of analysis one on top of the other, something that we believe is only possible with Oracle: if you want to enhance spatial analytics with data mining or drive multi-dimensional analysis using data mining or use semantics to help manage your data warehouse metadata then Oracle is the answer. There is a better way to do analysis in 2013

Screen Shot 2013 01 25 at 12 17 04
Which in-database analytical features do you use?
I have started a poll on LinkedIn to capture information about which in-database features people use the most. If you want to vote in the poll then click [here] - if you are not a member of the group then click the "request to join" button and I will approve you as soon as I get the email alert from LinkedIn
 As soon as the whitepaper is ready I will post a note on my blog.