Skip to Content
Author's profile photo Dan Grew

Federated Business Intelligence at AstraZeneca

I’ve had a really busy year so far but I thought it was about time I should share some of the things I’ve been working on with you. My role as BI Architect at AstraZeneca means I get involved in some interesting projects and challenges and this year has certainly been no exception. We have a major programme running in the commercial space to rationalise toolsets and introduce a more centralised approach to BI.

Traditionally, our marketing companies across the globe have been autonomous and free to choose their own IS solutions and have been in control of their own IT budgets. That model has now changed somewhat and a centralised approach has been adopted. So, last year we set off to review the BI landscape out in the markets and we were astonished to discover a total of 21 BI tools from different vendors, in other words we had pretty much every BI tool out there! In order to realise significant savings from a licencing perspective a programme was given the go ahead to rationalise the landscape by adopting a standardised set of BI solutions; we called it “BI in a box”. The solutions would use the SAP BusinessObjects product set which has already been a standard within central IS for some years. Standard reporting packs for Finance and CRM reporting were rolled out based around WebIntelligence, Crystal Reports, Xcelsius and Dashboard manager. Power users are trained in Webintelligence for ad-hoc reporting and some more mature marketing companies are being given the capability to manage their own environments. We have established a BusinessObjects shared service in each of our three world class data centres and using a delegated administration model a marketing company can have their own development, test and production environments. In addition, Predictive Workbench is being rolled out to the larger marketing companies to enable analysts to mine their data and populate their own data marts for reporting purposes.

So, a lot is going on as you can imagine and the rollout will continue over the next couple of years to cover every corner of the globe. But what’s the bit about federation in the title of this blog all about? Well, we found that some of the larger marketing companies that I mentioned earlier already had quite a mature BI capability and whilst our standard solutions and tools gave them the core functionality to run certain aspects of their business there were some areas that were quite specific to an individual market. For example, marketing data pertinent to the local market is often purchased from external providers and legal requirements for individual countries may require additional data to be maintained and reported upon. A good example is Spain who already had a well established data warehouse based around Microsoft SQL Server and Analytics Services. They were keen to adopt the standard solutions and the SAP BusinessObjects toolset as the benefits were obvious but we quickly realised that we couldn’t replace everything that they currently had. We wanted to keep the central Enterprise Data Warehouse as generic as possibly so were reluctant to start loading local country specific data as this would bring a whole new element of data governance into the equation. We also wanted the marketing companies to retain a certain element of independence as we were already forcing a lot of change onto them. So, we needed some way to integrate the data on the fly. We considered a number of options like merging data at report level but that would mean having to repeat the data integration in every report and with added complexities like master data inconsistencies (a whole other story!) it just added to the problems. The obvious solution to these issues it seemed was SAP BusinessObjects Data Federator.

Data Federator enables you combine data from multiple sources on the fly. It handles a whole host of different database platforms using JDBC and other sources such as flat files, XML or data services. Data Federator works out the most appropriate execution path and order in which to select data from the source systems. In other words, it doesn’t just pull the result sets from each source and then try to join them – it works out which is the smaller of the two data sets, retrieves that first and then joins it to the second source to reduce the result set appropriately. The result is a “virtual” relational database against which standard ANSI SQL can be executed.

We started a proof of concept with the Spanish data and setup a sandbox environment in our European Data Centre. With the help of SAP consultancy services we built some projects and mappings that pulled together central finance data from the Oracle based Enterprise Data Warehouse with some local data from the Spanish SQL Server data warehouse. The Designer tool enabled us to write rules to provide consistency around the master data mismatches and through a series of ETL-like transformations we were able to produce a consistent set of integrated output data. We then built a Universe on top of the virtual database in the same way that you would build any other Universe against a relational database. WebIntelligence reports were then written to query the data and the response times were impressive considering what was going on behind the scenes. The PoC was considered a success and we are now planning to roll out a production system with Spain being the first marketing company to use the platform. There is demand from other marketing companies for similar requirements  and the US have expressed an interest in using Data Federator as a way of integrating internal CRM data with external web services. I’ll tell you more about that when we have proved the tools capabilities in that area.

One thing to bear in mind before you rush out to buy a Data Federator licence is that some of the functionality for linking data sources will be provided in Information Designer (the replacement for Universe Designer) In SAP BusinessObjects 4.0 which will be available early 2011.

Assigned Tags

      7 Comments
      You must be Logged on to comment or reply to a post.
      Author's profile photo Former Member
      Former Member
      I think this is a great example of how to really use Data Federator the right way. Would BI4 have been enough to make this work out?

      Great job, Dan!

      Author's profile photo Dan Grew
      Dan Grew
      Blog Post Author
      Thanks Jamie, good question about whether BI4 would have been enough - I'm not sure is the answer. I haven't been able to find out exactly what DF functionality is included in Information Designer other than the ability to include multiple data sources.
      Author's profile photo Dan Grew
      Dan Grew
      Blog Post Author
      I met with Matt Shaw from SAP today and he gave me a good overview of Information Designer. From a multiple data source perspective it will do everything that DF does in terms of doing the joins and working out the most efficient execution path, etc but it doesn't let you create the mappings that can be used to manipulate the data. So, in our example we'd need to use the full-blown DF tool which will still be available as a separate package in BI4. If you're doing straight joins from one data source to another then you'd be good to go with Information Designer.
      Author's profile photo Witalij Rudnicki
      Witalij Rudnicki
      ... thank you for sharing!!
      As a technical mind I am always looking not only for pros but as well for cons of the solutions. I am curious to know what (technical 🙂 problems you faced with both: centralized BI tools roll-out as well as application of BO Data Federator.
      Best regards,
      -Vitaliy
      Author's profile photo Dan Grew
      Dan Grew
      Blog Post Author
      We've had many challenges along the way with our BI tools roll-out, it's hard to know where to start. What sort of things are you particularly interested in - bugs/limitations with the software, design issues, geographic challenges, etc? With the Data Federator PoC we had technical problems with JDBC drivers and various setup issues but all were resolved fairly easily.
      Author's profile photo Witalij Rudnicki
      Witalij Rudnicki
      I am specifically interested in limitations and performance results. Thank you.
      Author's profile photo Dan Grew
      Dan Grew
      Blog Post Author
      We didn't experience any limitations or performance problems with the PoC but there were two reasons for that:
      1. The data volumes were quite small and the data sets did not pose any problems that the tool couldn't deal with.
      2. Both database servers were hosted in the same data centre.
      We're still not live the solution as yet but I'll keep you posted with regards to any limitations we come across.