Contents


Data Transfer Process (DTP) BI 7.3

  1. Use

Key Benefits of using a DTP over conventional IP loading

Data Transfer Process (DTP) contains three tabs

  1. Extraction
  2. Update
  3. Execute

Detail Explanation Each Tab

  1. EXTRACTION
  1. Only get delta once:
  2. Get all new data request by request
  3. Delta Init without Data: (new option in BI 7.3)
  4. Optimal Package Size
  5. Parallel Extraction new option in BI 7.3)
  6. Semantic Group
  7. Filter
  8. Delta Init. Extraction from
  1. UPDATE TAB
  1. Error Handling
  2. DTP Settings to Increase the Loading Performance
  3. Handle Duplicate Records Key
  1. EXECUTE TAB
  1. Automatically Repeat Red Request in Process Chain  new option in BI 7.3)

Data Transfer Process (DTP) BI 7.3


DTP determines the process for transfer of data between two persistent objects within BI.

SAP Net Weaver 7.3, Info Package loads data from a Source System only up to PSA. It is DTP that determines the further loading of data thereafter.


Use

Loading data from PSA to Info Provider(s).

Transfer of data from one Info Provider to another within BI.

Data distribution to a target outside the BI system; e.g. Open HUBs, etc.


Key Benefits of using a DTP over conventional IP loading

DTP follows one to one mechanism between a source and a Target i.e. one DTP sources data to only one data target whereas, IP loads data to all data targets at once. This is one of the major advantages over the Info Package method as it helps in achieving a lot of other benefits.

Isolation of Data loading from Source to BI system (PSA) and within BI system. This helps in scheduling data loads to Info Provider at any time after loading data from the source.

Better Error handling mechanism with the use of Temporary storage area, Semantic Keys and Error Stack.


Data Transfer Process (DTP) contains three tabs


1. Extraction

2. Update

3. Execute


Detail Explanation Each Tab


1. Extraction

There are two types of Extraction modes for a DTP – FULL and DELTA.

When the source of data is any one from the below Info Provider, FULL and DELTA Extraction Mode is available.

  • Data Store
  • Info Cube
  • Multi Provider
  • Info Set
  • Info object :Attribute
  • Info object :Text
  • Info object :Hierarchies
  • Info object :Hierarchies (One segment)
  • Semantically partitioned  objects
  • Info object :Attribute

Note: Now in SAP BW 7.3, we can extract data from Multi Provider and Info set which is not possible in earlier version BI 7.0 and BW 3.5

Please check the below screen.

    

In the process of transferring data within BI, the Transformations define mapping and logic of data updating to the data targets whereas, the Extraction mode and Update mode are determined using a DTP. DTP is used to load data within BI system only.

If you selected transfer mode Delta, you can define further parameters:

 

A. Only get delta once:

            It can select this option where the most recent data required in data target. In case delete overlapping request from data target have to select this option and use delete overlapping request process type in process chain. If used these setting then from the second loads

it will delete the overlapping request from the data target and keeps only the last loaded request in data target.

B. Get all new data request by request:

            If don’t select this option then the DTP will load all new requests from source into a single request. Have to select this option when the number of new requests is more in source and the amount of data volume is more. If selected this option then the DTP will load request by request from source and keep the same request in target.

C. Delta Init without Data: (new option in BI 7.3)

            This option specifies whether the first request of delta DTP should be transferred without any data. This is same as “Initialization without data transfer” in Delta Info Package.

If we select this option, the source data is only flagged as extracted and is not moved to the target (first request of delta DTP will get successful with one record). This option will sets the data mart status for the requests in source and we cannot extract this requests using delta DTP anymore.

We will have this option in Extraction tab of Data transfer Process and can be seen in the below screen as well.

Note: If we are running any DTP for testing purpose, then it is not recommended to use this option. Instead, on Execute tab page, set the processing type to “Mark Source Data as Retrieved”.


D. Optimal Package Size:


            In the standard setting in the data transfer process, the size of a data package is set to 50,000 data records, on the assumption that a data record has a width of 1,000 bytes. To improve performance, you can set the value for the data package size, depending on the size of the main memory.

Enter this value under Package Size on the Extraction tab in the DTP maintenance transaction.


E. Parallel Extraction  new option in BI 7.3)

            System uses Parallel Extraction to execute the DTP when the following conditions are reached.

The Source of the DTP supports Parallel Extraction (Ex: Data Source).

Error Handling is De-activated.

The list for creating semantic groups in DTP’s and transformations is empty.

The Parallel Extraction field is selected.

This field allows us to control the processing type of the Data Transfer Process and switch between the “Parallel Extraction and Processing” and “Serial Extraction, Immediate Parallel Processing” processing types.

How it works: If the system is currently using the processing type “Parallel Extraction and Processing”, we can select this option and change the processing type to “Serial Extraction, Immediate Parallel Processing”.

If the system is currently using the processing type “Serial Extraction, Immediate Parallel Processing”, we can select the Parallel Extraction field and change the processing type to “Parallel Extraction and Processing” (provided that error handling is deactivated and the list of grouping fields is empty).

This field will be selected automatically if source of the DTP support parallel extraction.


F. Semantic Group


We chose Semantic keys to specify how we want to build the data packages that are read from the source. Based on the key fields we define in semantic keys, records with the same key are combined in a single data package.

This setting is only relevant for Data Store objects with data fields that are overwritten. This setting also defines the key fields for the error stack. By defining the key for the error stack, you ensure that the data can be updated in the target in the correct order.


G. Filter


            Instead of transferring large amount data at a time, we can divide the data by defining filters. The filter thus restricts the amount of data to be copied and works like the selections in the Info Package. We can specify single values, multiple selections, intervals, selections based on variables, or routines.


 

H. Delta Init. Extraction from


Active Table (with Archive)

The data is read from the DSO active table and from the archived data.

Active Table (Without Archive)
The data is only read from the active table of a DSO. If there is data in the archive or in near-line storage at the time of extraction, this data is not extracted.

Archive (Full Extraction Only)
The data is only read from the archive data store. Data is not extracted from the active table.

Change Log
the data is read from the change log and not the active table of the DSO.


2. UPDATE TAB

 

A. Error Handling

  • Deactivated:

            Using this option error stack is not enabled at all. Hence for any failed records no data is written to the error stack. Thus if the data load fails, all the data needs to be reloaded again.

  • No update, no reporting:

            If there is erroneous/incorrect record and we have this option enabled in the DTP, the load stops therewith no data written to the error stack. Also this request will not be available for reporting. Correction would mean reloading the entire data again.

  • Valid Records Update, No reporting (Request Red):

            Using this option all correct data is loaded to the cubes and incorrect data to the error stack. The data will not be available for reporting until the erroneous records are updated and QM status is manually set to green. The erroneous records can be updated using the error DTP.

  • Valid Records Updated, Reporting Possible (Request Green):

            Using this option all correct data is loaded to the cubes and incorrect data to the error stack. The data will be available for reporting and process chains continue with the next steps. The erroneous record scan be updated using the error DTP.

B. DTP Settings to Increase the Loading Performance

Number of Parallel Process:

We can define the number of processes to be used in the DTP

Open DTP and Go to Menu

Select settings for Batch Manager

 

Now appear below screen

Here defined 3, hence 3 data packages are processed in parallel.


If we want we can change the job priority also. We have three types of priorities for background jobs, they are

i) A –     High Priority.

ii) B-     Medium Priority

iii) C-    Low Priority

C. Handle Duplicate Records Key:

In case load to DSO, we can eliminate duplicate records by selecting option “Unique Data Records”. If loading to master data it can be handled by selecting “handling duplicate record keys” option in DTP. If you select this option then It will overwrite the master data record in case it time independent and will create multiple entries in case dime dependent master data.

3. EXECUTE TAB

A. Automatically Repeat Red Request in Process Chain new option in BI 7.3)

 

Sometimes DTP load fails with one off error messages like RFC connection error, missing packets, database connection issues etc. Usually these are resolved by repeating the failed DTP load. 

In BW 7.3 there is an option under execute tab which if you check will automatically repeat red requests in process chains. I think this is a cool feature and saves us time and effort in monitoring data loads.

Automatically Repeat Red Requests in Process Chains.

If a data transfer process (DTP) has produced a request containing errors during the previous run of a periodically scheduled process chain, this indicator is evaluated the next time the process chain is started.

How it works: If we select this option, the previous request that contains errors is automatically deleted and a new one is started.

If the indicator is set, the previous request that contains errors is automatically deleted and a new one is started.

If the indicator is not set, the DTP terminates and an error message appears explaining that no new request can be started until the previous request is either repaired or deleted.

Hopes it is helps….


To report this post you need to login first.

19 Comments

You must be Logged on to comment or reply to a post.

  1. Benedict Venmani Felix

    Hi Kodandapani,

    This document will be helpful for someone who is new to BW.

    I have got a suggestion. Instead of listing out the options available in ech tab, you can add a small note on what each option does. Like for example, choosing between the three priorities for Bg jobs or the paralle extraction new option in BW 7.3. This will be really useful.

    Is it BW 7.3 or BI 7.3

     

    Benedict

    (0) 
  2. Sourabh Jain

    Hi Pani.

     

    Really nice document, i have one question if you can please help.

     

    On Execute tab, option “Automatically Repeat Red Request in Process Chain“, is there any flip side of it in terms of performance, i have read F1 help for this option which mentions it, but if you can help me with some examples i.e. what kind of performance issue can come with this.

     

    Rgds

    Sourabh

    (0) 
    1. Phani KV Post author

      Hi,

       

      this feature available in 7.3  – if you check  the this option at DTP it will delete the RED request form the target then it will load new request.

      it not for the performance indicator.

      If the indicator is set, the previous request that contains errors is automatically deleted and a new one is started.

      (0) 
      1. Sourabh Jain

        Hi Pani,

         

        Thanks for you reply, what i am talking about is;

         

        Dependencies

         

        Note that if a delta update is carried out, the repaired
        request contains not only the data of the terminated request, but all data that
        has been added to the source since then. This can lead to both performance and
        consistency problems.

         

        Example of a Consistency Problem:

         

        A transfer that has the prerequisite that the data of only one day is
        transferred at a time, and that produces incorrect results, leads to consistency
        problems if, during the repair, the data of more than one day is to be
        transported at the same time.

         

        Can you please help me understand this dependency and example.

         

        This is from F1 help of the check box “Automatically Repeat Red Request in Process Chain“.

         

        Rgds

        Sourabh

        (0) 

Leave a Reply