Oracle BI Architecture Masterclass
Yesterday I’m fortunate enough to attend Oracle BI
Architecture masterclass for partners. The class largely covers the new Oracle
Reference Architecture version which you can read in great detail from thispdf file (please note that this file is actually related to BI Architecture version 2, the version 3 is not released yet but they're enough to understand the architecture).
If you are familiar with the previous Oracle reference
architecture, you will find that it has the similarity of the component namely
Staging, Foundation and Access and Performance layers. However it now adopts
the most talk about BI buzzword ‘Big Data’ into its component and Data
Discovery/Knowledge Discover that supports agile BI which has become a norm
this day has a special place in the architecture.
During the introduction, the instructors showing us an
Interesting radar graph that shows how the big data is not a replacement of the
old good RDBMS technology but a complement to each other. Here is the graph that also can be found on page 24 of the above pdf file:
Even Facebook the biggest Hadoop user realises this paradigm
(http://www.informationweek.com/software/information-management/facebook-exec-databases-hadoop-belong-together/d/d-id/1112126?)
and has used RDMS as their analytic technology alongside Hadoop. In line with
the Oracle suggestion, the Facebook exec says that they keep the low-level,
most granular, detailed data in Hadoop but move the transformed and aggregated
data into RDBMS for slicing and dicing.
There are a couple of surprising suggestions that they
mentioned for the implementation of the architecture. One revelation is the
extraction of the sources data should be done as a whole and as quickly as
possible (there are a hot and interesting discussion on this issue as one of
the partner objected that this is not always possible). In doing so, the data
warehouse is more agile and any future changes that required new data items
from the sources do not need new extractions or requests from the source
owners. The other thing that came out of the discussion is whether Master Data
Management (MDM) should be kept in data warehouse environment (or be
part of data warehouse project) or should be kept out of the data warehouse. It
seems reasonably enough that the suggestion is to keep it out as the data
warehouse is complex enough for another one to part of it.
Everyone in the room seems to see a rapid growth of interest
and implementation of big data in the past few years. The instructors which
have several years of working with the clients also confirm it and they seem
very optimist of the advance of the Oracle technology to support big data
solutions. They also think that they can give clearer guidance on how to answer
the not so clear ‘big data ‘ requirement from clients.
A few interesting points that came out of the class:
- Try to have small number components as possible. The architecture is not rigid so be pragmatic.
- Try to have small number components as possible i.e. be pragmatic
- Don’t prevent for the users to access their data on every layer
- Don’t attempt to build MDM on the data warehouse
- Don’t wait to ingest data into Raw data layer
- Consider to implement in-memory solution
- Data literate users should be involved early on by giving them access to raw data layer
- Get Finance analytics (EPM) fed back to source system
- Don’t be scare to use COTS BI to integrate with Oracle Reference Architecture as they normally fit to it
Comments
Post a Comment