Skip to Content
Author's profile photo Leigh Mason

SAP Business Objects Data Quality Management ( DQM ) Enhancements for more fields other than address data for Duplicate Check.

Relevant Systems

These enhancements were successfully implemented in the following systems:

  1. Customer Relationship Management ( CRM ) 7.02 SP 4
  2. SAP Business Objects Data Quality Management, version for SAP Solutions 4.0 SP2

This blog is relevant for:

  • ABAP Developers who have been asked to include additional fields ( such as Date of Birth ) for duplicate matching with DQM
  • ABAP Developers who have been asked to encapsulate the DQM functionality into a re-usable function / object.

This blog does not cover:

  • Details of the Postal Validation, we are only concerned with Duplicate Check

Reference Material

User Guide:

Brian Kuzmanoski’s Blog

SAP BusinessObjects Data Quality Management (DQM): Enhancements for more fields other than address data for Duplicate Check (Part 2)


DQM is an add-on to your ERP / CRM system to provide address validation ( SAP calls it PV for Postal Validation ) and duplicate checking for Business Partners.

Technical details in this blog are relevant for a CRM 7.02 system as this is where I’ve implemented the enhancements explained later successfully however they ( enhancements ) should be relevant for any system in accordance with the DQM  install guide.

In this blog entry I’ll only cover off items within a CRM system, I’m trying to get my colleague Brian Kuzmanoski to write a separate blog that deals with the DQM side of things for completeness of the solution overview.

Keep in mind I’m not going to explore every nook and cranny about the DQM add-on and this blog is just one answer to a couple of questions I’ve seen SCN, this is the first time I’ve touched DQM so have probably gone through the same learning curve as yourselves.

This is a great tool to provide address validation and formatting plus Business Partner duplicate checks in conjunction with an address directory, in my case the Australian address directory which is added to the DQM server ( every address in Australia ).

Installing the add-on out of the box with CRM and performing the initial setup, you get address and duplicate checking via T-CODE “BP” and automatic integration into the Account BSP Components.

The one catch I’ve found with DQM is that the duplicate checking is address focused which leads nicely into our problem definition.   But first a brief look at how the duplicate check solution works:

SPRO Activities

When you’ve successfully installed DQM on your CRM system, you can find all your SPRO activities under:

SAP NetWeaver->Application Server->Basis Services->Address Management->SAP BusinessObjects Data Quality Management

SAP NetWeaver->Application Server->Basis Services->Address Management->Duplicate Check


Typical Scenario

You need to create a new customer account record in CRM, you go to your BP Account BSP or transaction BP in the backend CRM system and start entering their details.  You then click Save and trigger the duplicate check.  The following process is triggered in CRM:


  1. A “Driver” record is constructed which is the record you want to create in the SAP system ( only in memory, not in the DB yet ).
  2. A selection of “Passenger” records are selected based on the Match Code from the CRM Database
  3. The “Driver” and “Passenger” records are passed to DQM to perform the matching algorithms
  4. A result set is determined with percentile weightings for the matching routines and takes into consideration your configured “Threshold” ( configured in SPRO ).

The Match Code – Table /FLDQ/AD_MTCCODE

As part of the setup of DQM you would have run three reports ( refer to the user guide ) one of these reports went through all your Business Partners and generated a Match Code and stored it in table /FLDQ/AD_MTCCODE.

The Match Code is just an index that reduces the number of potential records with which DQM needs to run it’s matching routines.

When a Driver record is created, DQM knows how to generate it’s Match Code. When the “Candidate Selection” occurs ( selecting all the records with the same Match Code in CRM ) you are taking a slice of possible data rather than comparing every Business Partner in your database.  This is obviously for efficiency however there are some pitfalls, we’ll cover these later.

The Results

If there are any possible matches then you may see several screens, one for the Address Validation and one if there are any possible duplicate records.

Presto, you’ve stopped a potential duplicate record entering your CRM system or you’ve validated and formatted the new customer’s address.

Notice from the screenshots there is a Similarity %, this is what we’re interested in, which begs the questions:

  • How can DQM put more weight on Date of Birth and Gender rather than the address details
  • How do I pass the additional fields across to DQM
  • Do I need to go through a stack of enhancements and potentially break the standard SAP solution?
  • What if I have my own services to create customers in CRM, can I encapsulate the Duplicate Check functionality and re-use it in my own code?

Problem Definition

You’ve installed DQM and taken advantage of the standard functionality provided, but you realise that you have requirements to match on more than address data such as Date of Birth, Middle Name or Gender.

You also want to create your own matching algorithms based on these additional fields on the DQM side which are not so address focused and want to put percentile weightings on the additional fields over address fields.

Remember what we talked about before, DQM is address so it doesn’t support all the Business Partner fields in the Driver and Passenger records when they are passed to DQM.

If you think of a typical scenario you probably want to match Date of Birth or Middle Name or Gender in combination with address details.  Out of the box you can’t use these three fields.

You may have custom services that create your customers and therefore want to plug in the DQM Duplicate Check ( outside of the BAdi framework ) such as a web service

For this blog we’ll focus on passing across the following additional fields to DQM:

  • Date of Birth
  • Middle Name
  • Gender
  • Country of Birth

Technical Implementation

We took a path of least resistance approach to these enhancements. What we didn’t want to do was to completely re-write the duplicate checking framework just to accommodate a few additional fields from the Business Partner.

Yes , our solution can be considered a hack, but what it doesn’t do is force you down a path where you completely own the solution ( insert your favourite statement about TCO here ) where you have enhancement after enhancement all through standard SAP code which is a nightmare to maintain in the long run.

First let’s define some requirements:

  • I want to use Date of Birth, Middle Name, Gender, Country of Birth in my driver and passenger records so DQM can match on these fields as well
  • I want a class/function that encapsulates the call to DQM and I can re-use it elsewhere in my SAP System.
  • I want to provide different % weightings to the other fields.

Encapsulating Badi ADDRESS_SEARCH

This is the Badi that is called to perform the duplicate check or address search.  If you dive into the implementation you will find:



The method ADDRESS_SEARCH is where all the magic happens and SAP makes an RFC call into DQM for processing of matching algorithms and returns some results.

Create a Class

We created a class with a global variable go_fldq_duplicate_check which is an instantiation of class CL_EX_ADDRESS_SEARCH during the constructor.

We added a method called FIND_DUPLICATES where the driver record is constructed.  To make it simple we added two structures as importing attributes for capturing the customer details:

  • IS_BUILHEADER – BOL Entity Structure for BuilHeader
  • IS_ADDRESS –  BOL Entity Structure for BuilAddress

When constructing the Driver record, all you are doing is adding lines to lt_search_params such as:

CLEAR ls_search_param.
ls_search_param-tablename = 'ADRC'.
ls_search_param-fieldname = 'CITY1'.
ls_search_param-content = is_address-city.
APPEND ls_search_param TO lt_search_params.

Passing Your Own Attributes – Hijack Other Fields

This is where we need to pass in our other fields Date of Birth, Middle Name, Gender and Country of Birth.  We do this by hijacking a couple of available fields from the supported field list. NB: We also pass in all other address information as well just to adhere to the standard functionality.


You add these field values as search parameters:

CLEAR: ls_search_param.
ls_search_param-tablename = 'ADRC'.
ls_search_param-fieldname = 'NAME1'.
CONCATENATE is_builheader-birthdate is_builheader-middlename INTO ls_search_param-content SEPARATED BY space.
CONDENSE ls_search_param-content NO-GAPS.
APPEND ls_search_param TO lt_search_params.
CLEAR: ls_search_param.
ls_search_param-tablename = 'ADRC'.
ls_search_param-fieldname = 'NAME2'.
CONCATENATE is_builheader-*** is_builheader-countryorigin INTO ls_search_param-content SEPARATED BY space.
CONDENSE ls_search_param-content NO-GAPS
APPEND ls_search_param TO lt_search_params

Pitfalls Here

Yes, hijacking NAME1 and NAME2 fields is not optimal however it represents an opportunity to pass in additional fields  so you can unpack them n the DQM side.

There is obviously some limitations here, i.e. the field lengths and if you have a lot of additional fields to pass in, how do you separate them logically so you know how to unpack them in DQM correctly every time.

Supported Field List – /FLDQ/AD_MTCCODE

This structure is used to construct the passenger records sent to DQM. Start here and also cross check on the DQM side the structure that is present there.

Call to DQM

Once the Driver Record is constructed i.e. your lt_search_params are populated the call is constructed like this:


Sample Code is provided below, we’ll just hard code a few items for demo purposes.

If IF_EX_ADDRESS_SEARCH~ADDRESS_SEARCH is executed successfully you will hopefully have some results in the importing parameter EX_T_SEARCH_RESULTS that contains the % weightings for your potential duplicates.

From here you can build your own return structure with these and other customer details.

So there you have it, you should have successfully encapsulated the DQM functionality in a class where you can now re-use it in other parts of your CRM system if you need to.

Pitfalls Here

When you encapsulate this functionality, you are constructing the Driver record manually which ensures your custom fields are passed in correctly.

In the typical scenario where you are creating a Business Partner via transaction BP or through the CRM Web UI, the Driver record is constructed using only available address fields plus First Name and Last Name.

You will need to make a small enhancement ( discussed next  ) in order to ensure the scenario’s where you are using SAP GUI or CRM Web UI also pass the desired fields for the DQM matching algorithms to ensure consistent results across your business processes.

Driver and Passenger Records Enhancement

Before the physical RFC call is made, a selection of candidate records is chosen based on the Match Code.

Your Passenger records will not have all the custom fields you want sent across to DQM to perform the matching algorithms.

What we need to do here is create a small implicit enhancement where we can populate the other details of the candidate records with the same hijacking idea as we saw above.


This is where the candidate selection occurs.  The selections are stored in it_cand_matches.  This has a table structure /FLDQ/AD_MTCCODE.

If you look at the table structure, you will see the available fields that make up the Passenger Records, as you can see there is mainly address fields in here including NAME1 and NAME2.

This is where we need to make a small enhancement in order to populate Date of Birth, Gender, Middle Name and Country of Birth as we did above when you made your class.  Only this time, we’re dealing with actual Business Partners in the database so we need to read their records and populate the candidate selection record.

Make an implicit enhancement at the end of FORM GET_CANDIDATES, you can use the following code as a guide:


The ZCL_DQM_PARTNER class is a simple class in order to read BUT000 and BUT020 without having to do direct SELECT statements.

When the code in the enhancement is executed, you are populating Date of Birth, Gender, Middle Name into the passenger records so on arrival to DQM your matching algorithms can use them.

By performing this enhancement you are doing two things

  1. All your additional fields are populated in the candidate selection so DQM can actual use those fields in the matching routines when comparing to the Driver record
  2. When a standard scenario is executed, such as creating a Business Partner in SAP GUI or CRM Web UI, these details are correctly passed to DQM for matching.

You are obviously placing a little more load on the system here by retrieving more details about the Business Partner.  Try and be as efficient as possible in this code to reduce any performance issues.

In reality, your Match Code should have reduced the number of candidates so any impact here should be minimal.

Populating the Driver Record when creating a Business Partner via transaction BP or via CRM Web UI

As mentioned a few times earlier, during the scenario when you create a Business Partner via SAP GUI ( transaction BP ) or CRM Web UI, these scenario’s you don’t have control over the Driver record as you did in your class with method FIND_DUPLICATES.

There is a relatively simple solution for this that doesn’t involve any more enhancements to standard SAP code.

The DQM add-on in CRM allows you to implement your own code to determine the Driver record details and plug your code in via some settings in SPRO.

Under activity “Maintain Operating Parameters” you have two places where you can maintain your own function modules for some scenarios. You can refer to the User Guide for these.

The one we’re interested in is AD_ADDRESS_SEARCH_FUNC_NM.


Follow the user guide and implement your own Function Module as per those guidelines.

You place the name of your Function Module in the Parameter Value field above.

What your function module is going to do is essentially the same as what you did before in FIND_DUPLICATES except this time you need to pull the Business Partner context out of memory and populate the Driver record with any Business Partner details ( such as Date of Birth ) that have been entered by the user.

Here is a sample implementation, feel free to make improvements etc:

FUNCTION zdqm_match_code_duplicate_chk.
 *"*"Local Interface:

 DATA: lt_but000         TYPE TABLE OF   bus000___i,

 wa_match_codes    TYPE            /fldq/ad_mtccode,
 ls_but000         LIKE LINE OF    lt_but000,

 lv_gender         TYPE            char1,
 lv_records_in_mem TYPE            i.

 LOOP AT im_driver INTO wa_match_codes.

 i_xwa          = 'X'
 t_but000       = lt_but000.


 DESCRIBE TABLE lt_but000 LINES lv_records_in_mem.

 CHECK lv_records_in_mem > 0.

 IF lv_records_in_mem  = 1.

 READ TABLE lt_but000 INTO ls_but000 INDEX 1.


 READ TABLE lt_but000 INTO ls_but000 WITH KEY name_first = wa_match_codes-name_first
 name_last  = wa_match_codes-name_last.


 CONCATENATE ls_but000-birthdt ls_but000-namemiddle INTO wa_match_codes-name1.
 CONDENSE wa_match_codes-name1 NO-GAPS.

 IF ls_but000-xsexf = abap_true.
 lv_gender = '1'.
 ELSEIF ls_but000-xsexu = abap_true.
 lv_gender = '0'.
 ELSEIF ls_but000-xsexm = abap_true.
 lv_gender = '2'.

 CONCATENATE lv_gender ls_but000-cndsc INTO wa_match_codes-name2.
 CONDENSE wa_match_codes-name2 NO-GAPS.
 MODIFY im_driver FROM wa_match_codes INDEX sy-tabix.
 CLEAR: ls_but000, lt_but000.



By implementing this Function Module you can now ensure that your additional fields are passed to DQM in the Driver Record when using the scenario of creating a Business Partner via SAP GUI or CRM Web UI.

Check Point – Is it working so far?

If you have:

  • Installed DQM correctly
  • Performed all necessary post install configuration and setup
  • Successfully performed the “Hand Shake” procedure with DQM
  • Setup your matching routines in DQM
  • Built your own class in CRM which can be tested in isolation
  • Created your enhancement point for Passenger records ( candidate selection )
  • Created a Function Module to implement the Driver record population during standard Business Partner create scenario’s

You should be able to get some results back from testing your class. Just ensure you’ve actually setup some test data that makes sense in order to get results back from DQM.

Some Gotcha’s and Other Thoughts

With anything you work on for the first time, there are almost always downstream impacts that you didn’t consider.

Here are a couple:


If you use this BAPI anywhere and you have installed DQM, then you will probably start to get failures if you’re trying to create a Business Partner that is a possible duplicate or the address is invalid if you have Postal Validation switched on. This is a problem if you have any interfaces that consume this function module.


Apart from your normal unit testing you will need to go through some regression tests because you’ve created an implicit enhancement point that affects some standard Business Partner creation scenarios.

Ensure the following behaves as expected:

  • Create a business partner in SAP GUI via Transaction BP
  • Create a Business Partner in CRM Web UI
  • Ensure any interfaces that consume BAPI_BUPA_CREATE_FROM_DATA are working correctly now that DQM is switched on

Switching off PV and Duplicate Check

There are a few options for you to switch off the Postal Validation and Duplicate check if you need to.  Although this doesn’t really tie in to the theme of this blog, we discovered some limitations here.

  1. You can activate / de-activate the Postal Validation and Duplicate check in a SPRO activity.  This completely turns off these checks.  If you do this, you will need to re-run the match-code generation when you turn it back on in case new Business Partners were created whilst DQM was switched off.
  2. You can assign a program or transaction code ( see screen shot below ).  Here you can suppress several functions, Validation (PV) Search (Duplicate Check) Suggestions ( Postal address suggestions ).


The technical implementation of this check happens in Function Module /FLDQ/AD_CHECK_TCODE_VALIDITY.

This is where it gets interesting.  The code behind this looks for the program name in SY-CPROG at runtime. Just beware that when this executes in a non-dialog process such as an RFC call or Web Service call, the value in SY-CPROG will be the overarching framework program.  In the case of RFC calls , program SAPMSSY1. What this tells me is that “Maintain Exception Table of Transaction Codes” activity is only meant for dialog processes ( e.g. creating a business partner via transaction BP ).


Summary and Final Thoughts

The DQM add-on is really good piece of kit, we had the scenarios in this blog up and running in a couple of days including the DQM side which hopefully my colleague Brian Kuzmanoski will blog about very soon.

There are obviously a couple of limitations that we discovered, but remember the product is address focused, we’ve just demonstrated that you can include additional fields for matching even though our solution is not entirely optimal however we’ve avoided major enhancements which was the goal.

Just keep in mind that if you’re trying to switch off DQM via the “Exception” list, there is a limitation here for non-dialog processes.

Finally, DQM should be implemented in conjunction with an overall master data governance strategy, this is just an enabling tool for that strategy but by no means will solve all your master data problems.

There are further things to explore here such as how to connect DQM to a MDG system or even MDM, where you would effectively be cleansing your master data before it’s even created in an SAP system such as CRM.

Hope you enjoyed this blog. Please look out for Brian Kuzmanoski’s blog that covers the matching algorithms on the DQM side..

Assigned Tags

      You must be Logged on to comment or reply to a post.
      Author's profile photo Former Member
      Former Member

      Very nice written blog about DQM.

      I hope the product is more mature now. πŸ˜†

      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      Thanks Tommy, it's quite a cool piece of kit for a quick win.  I just posted a link to Brian's blog which covers off on the actual DQM side which is great, so you should have a more rounded view on what we did at a customer.


      Author's profile photo Former Member
      Former Member

      Good one. Leigh. I need to get some information from you on DQM. Can you please drop in your mail id to

      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      Hey mate, thanks for the comment.  Let's discuss the details on SCN, maybe others can benefit from it is well.  If you have specific customer requirements, please omit your customer's details from the discussion for privacy.  πŸ™‚

      Author's profile photo Former Member
      Former Member

      Ok Sure. I have a requirement to cleanse the address of customer data from CRM using DQM?. The installation is complete i have a set of standard jobs on the data services which were in once the installation was complete. Is it going to be like i have to manually configure the standard jobs on the data services or i don want to worry on that end?.Just stuck on where to start with can you help me out?.

      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      What have you done so far? I’m assuming at this point you have:

      • Installed the DQM / RFC Server, maybe you’ve installed a postal address file and geo-coding module?
      • Installed the add-on to CRM ( what CRM version by the way? )
      • Performed post installation tasks required in CRM ( follow the install guide )
      • Started the services on the DQM side and performed any post install tasks
      • Does your “handshake” occur, this is just a TCP/IP connection test for FL_GDQC_SERVER, you can do this via SPRO ( or transaction SM59 ).
      • Are you using the standard break-key for duplicate check or have you configured your own?

      Now you will be able to run the initialisation programs that generate the break-keys and cleanse the current records in your CRM database. 

      Please ensure you read the install guide carefully when running the initialisation programs in CRM, there are several warnings, for example if you are using Duplicate Check - Error Tolerant search

      Before running the initial setup program, you must run Quarterly

      Adjustments to perform postal validation on the data that already exists in

      your SAP system. This process is required even if you are upgrading this

      product from a previous version because having correct address data

      increases the effectiveness of finding duplicate data.”


      If execution errors occur when you run the report /FLDQ/RSADRQU2, you

      must not run the report RSADRQU3 next. All three of the Quarterly

      Adjustment reports must be run successfully in sequence. Running

      RSADRQU3 on data resulting from a failed run of

      /FLDQ/RSADRQU2 can compromise the integrity of the data written back

      to the SAP system.”

      Author's profile photo Former Member
      Former Member

      Thanks a lot for the detailed explanation.

      • Once after doing all this do we need to customize anything for the address check to work or just running the quarterly adjustments reports would be enough?.
      • And should anything else be done on the data services side i see that there are some 10 to 15 jobs ( 2 in batch and the rest in real time) that are in once the DQM installation was complete. Should i do anything to them for address check to work or they will automatically pull the source data in CRM and do the needful?.
      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      The product will work out of the box in CRM.

      I’ll assume that you have followed the install guide and you are using (1) Postal Validation and (2) Duplicate Check.

      In the case of (2) you should have run the initialisation programs that create the break keys for the address records in the CRM database, this is absolutely necessary for the duplicate check to work.

      As Brian and I mentioned in our blogs, this is just a tool that enables your data governance strategy, so keep in mind this a multi-step process:

      Ongoing Data Quality

      The functionality allows you to validate and pick up possible duplicate records ongoing.   The guides refer to this as “interactive cleansing”, that is when an end user is creating a customer record in CRM.

      When you create a business partner in GUI or Web UI, you will automatically be presented with the pop-ups showing address validation and possible duplicates.  The duplicates will be matched on the standard break key that is setup with DQM on install ( defined in DQM ).

      Post Install Cleanse and Ongoing Data Quality

      The install guide instructs you to run the initialise programs that will provide a report of what addresses in your CRM database require attention and will allow you to correct those address records in batch, this is discussed in “Quarterly Adjustment Reports”.

      Obviously, after install you want to correct the data you have, hence why you run the QAR after install. I would suggest that you test this in your development or quality system so you are comfortable with the results and process of cleansing before doing this in the Production environment.  Yes, when you run them, it will match address records in CRM against the “postal file” in DQM and you can commit the address changes in batch to the CRM database.

      I can see where you are going with batch jobs ( SM37 right? ).  Assuming here you want to just setup some batch jobs to periodically run the QAR’s?

      Here are some additional thoughts:

      • Your data quality strategy should be process driven, you need to pick a point(s) in time you want to cleanse duplicates or address data, someone has to own the process and relevant stakeholders engaged to handle errors, manual work and to investigate why ( if any ) you are getting duplicate records or bad addresses into your database.
      • Are you sure you want to just run batch jobs to clean up potential address records?  My experience with these things is that unless there is a rigid process in place, your batch jobs will just run without anyone knowing and somewhere down the line someone will say “what is this batch job doing?  And switch it off possibly.
      • I would recommend starting with a manual process for executing the QAR’s, this way your customer can monitor the amount of “poor” data ( if any ) is being created in your CRM and can then work towards why and how the “poor” data is slipping through the gaps.

      Hope this helps, keep the conversation going if you still need clarification.  I will try and get Brian Kuzmanoski to drop some comments in from his side as well.



      Author's profile photo Former Member
      Former Member

      Thanks a lot Leigh for providing so many valuable inputs. I tried running the RSADRQU1 and the index were created after which i ran the /FLDQ/RSADRQU2 which both the options checked on and it takes forever to run it keeps running with a please wait message at the bottom πŸ™

      But i did a check with few postal codes and i recieved the error code T002 : which says exception system failure when i checked the /Fldq/RSADRQU2 source code i saw that the problem was in ADR_VALIDATE_QTR in the firstlogic call when i click on ADR_validate_qtr i get a message stating that the function module doesnot exist πŸ™

      Is it something that comes aspart of BAPI installation.How to fix this?

      Also the i get the following errors and warnings for the Data services jobs

      |Session Job_Realtime_DQ_SAP_US_DPV_Address_Cleanse_Suggestions

        Transform <Global_AddressCleanse>:Option Error(Option: SUGGESTION_LIST/PRIMARY_ADDRESS_COMPONENTS/LOCALITY1): The option


      |SESSION Job_Realtime_DQ_SAP_US_DPV_Address_Cleanse_Suggestions|DATAFLOW DF_Realtime_DQ_SAP_US_DPV_Address_Cleanse_Suggestions|STATEMENT <GUID::'c908e3d7-46d3-4531-ba9b-0ce837f09b95' TRANSFORM SDKTransform OUTPUT(Global_AddressCleanse)>

                                                               Warning: The Run as separate process flag in transform <SDKTransform> is ignored because the transform contains a column

                                                               representing a nested table.

      Please throw some light πŸ™‚


      Sudharsan C

      Author's profile photo Former Member
      Former Member

      Hi Sudharsan,

      As Leign mentioned, the solution should work out of the box for address cleansing at point of entry and also for the Quarterly Adjustment Reports with the standard jobs delivered in Data Services (there is a realtime job and a batch job, respectively).

      You can of course go in and adjust the Dataflows to customise to your requirements -- i.e. tweak the GAC transforms for example -- but this is optional.

      Have you attempted to customise the Data Services jobs at all? If not, you could redeploy the standard ATLs to get them back to stock which should eliminate the errors/warnings you get on that side.

      Regarding the error T002 on the back-end system: according to the DQM users guide this is "Exception SYSTEM FAILURE raised during RFC call." You may need to go back through the User's Guide to make sure you've done all of the installation/configuration steps correctly.


      Author's profile photo Former Member
      Former Member

      Thanks a lot Brain. I have one quick question. The current CRM has like 1.4 million records running a quarterly adjustment report for the same takes so much time that it does not return back any value after the /FLDQ/RSADRQU2 so i was thinking to run it by portions can you please let me know what needs to be done inorder to run the report by portions?.


      Author's profile photo Former Member
      Former Member


           If i wish to unsintall the DQM component and reinstall it again would it just uninstall the DQM component or the complete data services needs to be uninstalled?. Can you let me know the impact?.


      Sudharsan C

      Author's profile photo Former Member
      Former Member

      Hi, no, the components work together but are otherwise independent. You will just need to ensure the connectivity to the correct DS DQM environemnt is reestablished.

      Author's profile photo ankit verma
      ankit verma

      Hi Leigh/Brian,

      Thanks for sharing wonderful articles.Request if you can please answer a query of mine on SAP CRM & DQM integration.We are tying to achieve Account and Contact name cleansing run time when created/edited on UI based on certain naming rules.Want to know is it really possbile with DQM ,because all that I have read about DQM it seems to cleanse address data and also duplicacy is checked on the same.

      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      Hi Ankit,

      Sorry for the late reply.  What Brian and I demonstrated was clearly not "out of the box" functionality, it was intended to show exactly what you described, the ability to cleanse based on rules for names and how DQM can accommodate those rules.

      Have you managed to play around with it?  If you're still stuck, please reply to the thread and I'll go find someone from our product team to discuss any new features that may be available or in the pipeline πŸ™‚



      Author's profile photo ankit verma
      ankit verma

      Hi Leigh,

      Thanks for your reply.Yes I managed the requirement.Basicaly in our case along with duplicate check we required cleansing of filds like if user entered mr.ankit in firstname of UI so only Ankit remains in firstname and Mr. goes in title other rules like this.Since this cannot be done in Duplicate check std process we created a web service in DQM and did it through a proxy call from CRM in ADDRESS_CHECK BADI to get a cleansed response and populate data.This way I was able to fulfill this requirement and would like to thank you for your detailed useful articles.

      Kind Regards,


      Author's profile photo Former Member
      Former Member

      Hello Leigh,

      I know this is outside the scope of your blog, but I'd like to ask you more about the limitations with disabling the PV and Search features.  We learned about this issue the hard way.  We implemented DQM on our CRM system late last year; and DQM itself isn't the problem. 

      The issue we have is that customer addresses are being modified on our ECC system, and due to middleware configuration, synchronized back to CRM.  With no PV check on ECC, sometimes the address are entered incorrectly, which get corrected in CRM (via DQM), and sent back to ECC -- this causes what we've called the "echoing effect" and records are being changed unwantedly.  Until we get DQM up and running on ECC as well, and better understand the E2E process, we wanted to turn DQM off; which we did via SPRO (and went as far as removing the RFC connections between CRM and DQM, just in case). 

      However, we're still seeing some validation checking (kind of) when records come over from ECC to CRM.  When we check the change logs for these cusomer records, we see that the changes are being made by CRM.  Do you have any more information on how to completely disable this functionality?

      Thanks in advance,


      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      Hi Jose,

      Sorry for the late reply here..  Interesting question, if you've cut the RFC cord to the DQM server ( Microsoft server? ) then you shouldn't be getting PV / Duplicate checks from there.

      Have you switched off the BADI's, ADDRESS_UPDATE, or ADDRESS_SEARCH.  Not sure if these are applicable in your case, but double check what the impact is if you attempt to switch them off.



      Author's profile photo Former Member
      Former Member

      Thanks for the reply, Leigh.  Yes, its a Microsoft server in our case.  I tried deactivating the functionality via SPRO, I'm assuming this will switched off the BADIs, but unfortunately this didn't help.  We currently have a high-priority message open with SAP on this issue.  Things have gotten worse, and as a result its causing data inconsistencies between CRM and ECC.   Hopefully SAP can provide us with a work around.

      Thanks again.


      Author's profile photo Leigh Mason
      Leigh Mason
      Blog Post Author

      Can you try one thing.  Go directly to transaction SE18, choose BADI name ADDRESS_UPDATE, goto the top menu bar and choose Implementation Overview.  Can you check if any of those implementation are being called?

      Author's profile photo Former Member
      Former Member

      Hello Leigh,

      we have successfully implemented DQM, now for BP address validation and duplicate check validation is happening in GUI but in WEB UI only address validation is happening, Duplicate validation popup is not all coming, any clue about this.

      And also we need perform address validation for other master data's say example IBASE. Could you please guide us how this can be performed using DQM.

      Thanks in advance.



      Author's profile photo Former Member
      Former Member

      Hi Leigh,

      Thanks for the information.

      Its really helpful and informative.

      I am facing a similar issue which Tejtaj is facing.

      The Duplicate Check functionality via DQM is successfully working fine in CRM GUI, butΒ in WEB UI only address validation is happening, Duplicate validation popup is not all coming.

      Can you please help us what might be causing this issue?



      Author's profile photo Alienok Carter
      Alienok Carter

      It's really great! I like this resource, but sometimes I also use to get some educational help.