Experimenting with Universal Data Cleanse

I am working with Universal Data cleanse and custom dictionaries for a project I am working on.
The goal is to clean the leagacy firm information with standards. The issue I am having is using the dicitonary, how to I replace words with the standard word.
For example, the text contains the word: CLEANME
I want to have the dictionary replace that with the standard: CLEAN ME
IBoth were added as dictionary entries, one at a time, and the other as the Alias or the Standard in that dictionary entry. I get no results.
Reading the documentation, it seems I should be setting it up like this:
Dictionary Entry: CleanMe
Classification: Firm_Name, Firm_Name_Only
Gender: Unassigned
Standard: Clean Me
Thanks for all help!

There seem to be a couple of things going on here:
1. If you are using your custom dictionaries, then you have to map your input to MULTILINE1 and enable the custom parsers - just something to be aware of
2. You mentioned that you made some changes to the existing dictionary and you are not seeing any changes. To be clear, do you have different TEST and PRODUCTION environments? Or is it the same environment except that you have a local DS repository and for the dictionary you are pointing to another repository (using Dictionary --> Manage Connections)?
Having the dictionaries on a different repository should not make any difference as long as you point to them in your designer using the Dictionary --> Manage Connection option.
So I think there may be some issue with your job setup and/or dictionary values need to be looked at. You can start by adding another output field named "EXTRA" to see whether or not your data is getting parsed at all. Also, make sure the entry "CLEANME" is classified as FIRM_NAME_ALONE in the dictionary and that you are selecting the correct dictionary name in the Datacleanse Transform options.

Similar Messages

  • Problem with universal data cleanse

    I have problem with universal data cleanse.
    I am using DS 3.2.x (12.2.2).
    I created:
    Dictionary: TEST
    Classification: TEST_CL
    Custom outputs: category: TEST_CAT
    Dictionary Entry:
    Primary: BO
    Classification: TEST_CL
    Gender: Unassigned
    Secondary information:
    When used as: TEST_CAT
    Standard: Business Objects
    Rule file:
    DataCleanse Rule File v2.0;
    TEST_CAT = TEST_CL;
    action = TEST_CAT;
    TEST_CAT = 1 : TEST_CAT : 1;
    end_action
    Data Cleanse:
    Input: Multiline1
    Options:
    Parsing Dictionary: TEST
    Rule file: .../test_rules.dat
    Break On Whitespace Only: Yes
    Parser Sequences Multiline1: TEST_CAT
    Output:
    Parent_component: TEST_CAT1
    Generated_field_name: TEST_CAT and RULE_LABEL
    Generated_field_class: parsed/standardized
    Contenet_type: none
    and EXTRA field
    I wanted to replace word u201CBOu201D with the standard words u201CBusiness Objectu201D, but I have result: there is "BO" in the extra field,
    there are "null" in others fields.
    What am I doing wrong?
    Thanks for all help!
    P.S. I don't have cleansing packages installer.

    There seem to be a couple of things going on here:
    1. If you are using your custom dictionaries, then you have to map your input to MULTILINE1 and enable the custom parsers - just something to be aware of
    2. You mentioned that you made some changes to the existing dictionary and you are not seeing any changes. To be clear, do you have different TEST and PRODUCTION environments? Or is it the same environment except that you have a local DS repository and for the dictionary you are pointing to another repository (using Dictionary --> Manage Connections)?
    Having the dictionaries on a different repository should not make any difference as long as you point to them in your designer using the Dictionary --> Manage Connection option.
    So I think there may be some issue with your job setup and/or dictionary values need to be looked at. You can start by adding another output field named "EXTRA" to see whether or not your data is getting parsed at all. Also, make sure the entry "CLEANME" is classified as FIRM_NAME_ALONE in the dictionary and that you are selecting the correct dictionary name in the Datacleanse Transform options.

  • Crystal XI with universe data sources - persistent behavior

    We have been working with Crystal XI reports against universes and are faced with challenges of fully understanding how the universe/report information is kept synchronized after a report is published in BOE repository.
    For example, I have a Crystal XI report that is built against a universe and has prompts against universe objects (not necessarily pre-defined universe conditions) which have associated LOVs. These universe based prompts are automatically created as static prompts in Crystal. When I run this report and run an SQL trace on the database, here are my observations:
    -Report will not submit the underlying LOV SQL query to the database and use the static prompt values stored in the report at design time.
    -If the underlying LOV SQL now returns values that are different than the ones stored in the report, these do not automattically get reflected in the report.
    -If a universe change is made and then exported back to the BOE repository, from that point on, all refreshes of the report re-execute the LOV queries and thus get the updated prompt values which are then reflected on the report. There is a slight twist to this, the LOV queries get submitted twice, once when a report is refreshed before the prompts page is displayed and once when OK button is clicked on the prompts page along with the main report queries.
    -It seems like BOE architecture compares the timestamp of the underlying universe against the time when the report was last modified and if the report was modified after the universe, the LOV queries are not executed and static prompt values saved within the report are used.
    Can someone please confirm/explain Crystal persistent behavior with respect to universe objects, specially for LOV based prompts or point to technical resources that explain in detail the workings of Crystal XI reports against universe objects?
    Thanks.

    Hi Farhan,
    The description which you have posted for your issue is bit confusing.
    Can you please a break into the issue.
    As the LOV's are created in Universe and we will get the refreshed data into your report.
    We cannot create any static parameters at the report level while using universe as DataSource .
    Let me know if i am wrong.
    Thanks for understanding.
    Regards,
    Naveen.

  • Opening an Excel 2003 file in Office 2010 Universal Data Connection error when opening the file from a domain.

    Hi.
    One of my users are having  problem with old accounting files in Excel 2003 when opening Excel 2010.
    The user experience that linked information is not being updated properly.
    And suspects that it something do with Universal Data Connection .
    How can an Excel 2003 spread sheet be converted to Excel 2010 spread sheet and still contain active links?
    When the user is opening and converting the file as mention in your suggestion, on the common share on a common domain, the links "breaks".
    we are not using SharePoint on our domain.
    But when doing it on a local machine, the links does not "breaks".
    Looking forward to get solved to this problem.
    Thank you in advance.
    Rgds

    it sounds link, the links changed from relative paths to absolute path. You need to use absolute path in original file.
    KR

  • Formula Created in BI Query appears in Universe as Measure with No Data

    Hi,
    I have created the universe on top of SAP BI Query(Which is built on Infoset).
    There are some formulas created in BI Query as mentioned below:
    Eg: Status1=
                        If Completion Code = null then Status1 = 0 
                        If Completion Code = =10,11,18 then Status1 = 1
          Status2=
                         If Status1= 1 and Field Completion Date <= Regulatory Due Date then Status 2 = 3
                         If Status1= 0 and Report Date <= Regulatory Due Date then Status 2 = 4
    In the Universe I get Status1 and Status2 as Measures.
    When I use these Measure Objects in WebI report, I donu2019t see any Data for these objects in WebI Report. Both the columns for Status1 and Status2 appear Blank with no data in it, although I get the data in SAP BI Query for both Status1 and status2.
    Is there any issue with the formulas to be used in SAP BI Query?
    Are Formulas supported in Business Objects from SAP BI Query?
    regards,
    Nisha

    Hi Ingo,
    I tried running the standard test MDX in MDXTEST and I got the data for those calculations.
    But I wonder why there is no data in WebI for those formulas(Key Figures)?
    In Standard test MDX the MDX Query is as Follows:
    SELECT
    [Measures].MEMBERS ON COLUMNS,
    NON EMPTY [Z_WM_IS01___F98].[LEVEL01].MEMBERS ON ROWS
    FROM ZWM_M02/Z_ZWM_M02_Q001 SAP VARIABLES
    [!V000001] INCLUDING [Z_WM_IS01___F15].[3]
    Based on above Query I created the WebI report which includes the objects as described below:
    one Dimension Object Notification Number which is equivalent to [Z_WM_IS01___F98].[LEVEL01].MEMBERS  from above query.
    Selected All the measures objects available in query which refers to [Measures].MEMBERS
    And  one Prompt on Region which is equivalent to [!V000001] INCLUDING [Z_WM_IS01___F15].[3]
    Why there is no data for Calculation columns (Key Figures Status1 and Status2) in WebI Report???

  • 64bit CR-Enterprise wont talk with 32bit data source, 32bit IDT wont use 64bit connection for universe. How do I connect them.

    Hi.
    I installed the full suite of tools as I am involved with a company intended to become a partner and reseller so we are trying to get everything working.
    SAP Crystal Server 2013 SP1
    BusinessObjects Business Intelligence platform
    SAP Crystal Reports for Enterprise
    SAP HANA
    SAP Crystal Reports 2013 SP1
    SAP BusinessObjects Explorer
    SAP Crystal Dashboard Design 2013 SP1
    Crystal Server 2013 SP1 Client Tools
    SAP BusinessObjects Live Office 4.1 SP1
    .NET SDK
    I found out that BI was needed only after I installed all the others but I installed it without problem.
    My issue is that the Information Design Tool (IDT) which creates my universes successfully, and even lets me open and see the dada no problem. Is using a 32bit (ODBC connected to SQL Server 2008) data source.
    However, I am unable to load the .UNX in crystal reports (CR) 2011, so I used CR Enterprise (CRE) to connect to my universe. Now when I try to open the universe I start getting the following error:
    [Microsoft][ODBC Driver Manager] The specified DSN contains an architecture mismatch between the Driver and Application
    When I do searches online I get very generic information relating to setting up data sources. While I believe the problem is indeed with the data source, I don't believe it is setup wrong.
    When I access the universe using the IDT (which uses the 32bit version of the data source to build its connection, it wont let me use a 64bit) I can load a table under the "result objects for query #1" press refresh, I get a list of data.
    When I access the same universe using CRE (which "Seems" to use a 64bit data source, but I am not sure), and follow the same process as above. I get the above error.
    If I cancel the process, when CRE lists the fields for the report I can later manually map these fields to an ODBC connection. But if I have to do this what is the point of using the universes at all as the end user has full access to all the various DB's available in the data source.
    Thanks in advance for any help you can provide.

    On the server where Crystal Reports Server is installed, create a 64-bit ODBC connection with the same name as the 32-bit connection.  CRS will then automatically use the 64-bit version of the connection.
    -Dell

  • Current date in BO queries with Universe built on Infocube

    Hi dear experts,
    We are facing the following issue :
    - we built a universe on top of a multiprovider (with 2 infocubes) since we don't want to manage BW Queries.
    - we need to restrict data on WebI queries using current date or system date (to compare with a date from the universe)
    No options to do that seem to be available through MDX in the universe.
    Has anyone already solved such an issue and how do you proceed?
    Thanks for your help.
    André

    Hi Andre,
    I assume you are aware that you are losing items when connecting to the cubes directly, such as restricted keyfigures, calculated keyfigures, variables, structures, navigational attributes, and the option to use authorization variables.
    In the universe right now you don't have an option to reference something like a system data - that is done with an EXIT variable in the query.
    Ingo

  • Data and Cleansing export TO SQL table with Melissa Data appended fails

    I am using Data Quality services with Melissa Data Address Check as reference data.  Everything works fine until I take the option to export Data and Cleansing Info which will give me my cleansed data plus additional data points such as geocodes from
    Melissa.  When I do it fails with the error below.
    (Failed to create a new table geocode in database DQS_STAGING_DATA. Check whether the table already exists  and have the database administrator make sure the DQS Service has CREATE TABLE rights in the destination database and can INSERT to the destination
    table.)
    This error makes no sense as the table does not exist and I do have proper rights. I can export Data and Cleansing data if Melissa Data is not involved  ,  when I dig further it seems to be complaining about column header lengths.   
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_CBSADivisionCod' is too long. Maximum length is 128.;
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_DeliveryPointCo' is too long. Maximum length is 128.;
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_ResponseRecordI' is too long. Maximum length is 128.;
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_DeliveryPointCh' is too long. Maximum length is 128.;
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_CBSADivisionLev' is too long. Maximum length is 128.;
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_CongressionalDi' is too long. Maximum length is 128.;
    The identifier that starts with 'Address Validation_Melissa Data Corporation - Address Check - Verify, Correct, Geocode US and Canadian Addresses_CBSADivisionTit' is too long. Maximum length is 128.;
    I can see no option to control these column headers in DQS.  Has anyone else experienced this ?  Does anyone know of a workaround ? 
    I have already reported to Melissa data and they agreed the problem was the column header length but said they also had no control of that.

    Hello,
    You can create an SR with a based outbound filter. All object that match the filter will be provisioning to CS SQL (if you do not define filter, all objects will be provisioning).
    Or you can create an MVextension rules
    Regards,
    Sylvain

  • Data Cleansing in CRM Web UI

    Dear Experts,
    I have activated Data Cleansing in CRM 7.0 and it is working.
    Can anyone advise on the setting or configuration required to get the CRM Data cleansing functionality to copy communication data stored at the relationship from the Source record to the Master record.
    Example:
    Company A has a relationship with Contact person A - this relationship has a mobile number and email address maintained.
    When merging the account of Company A (source) with a duplicate company A1(master), we need to copy the relationship (already occurs) between Contact Person A and Company A and the relationship level mobile number and email address to the master account which will be company A1 (this does not occur).
    Currently the relationship level communication data is not copied.
    Please provide step by step config settings to enable this. I already have the SAP Notes on Data Cleansing and have covered all WIKI's on the web. What i need are specific steps to overcome this issue.
    Thanks!!!

    In order to transfer communication data maitained at the Contact Person relationship level, you have to maintain the required node in transaction BUSWU02: BUP115     Contact's Work Address and BUP120 Contact's Communication Data
    If data is still not transferred, SAP has released notes: 1243559, 1491950, and 1493240.  to resolve.  Check the relevance before implementing.
    FK

  • Error when refreshing WEBI report with Universe Connection Type "SSO"

    Hi Experts:
    We are trying to refresh the Webi report in Infoview with Universe Connection set as "Use Single Sign On when refreshing the report at view time", so that we can leverage SAP OLAP authorization variable from Bex Query which the Universe is built on.
    However got the error of "incomplete logon data" after all the configurations done following below blogs:
    SNC Part 1
    /people/ingo.hilgefort/blog/2009/07/03/businessobjects-enterprise-and-client-side-snc-part-1-of-2
    SNC Part 2
    /people/ingo.hilgefort/blog/2009/07/03/businessobjects-enterprise-and-client-side-snc-part-2-of-2
    We already have Win AD SSO to SAP setup, and in BO CMC, Win AD user is mapped to SAP user ID.
    The SNC settings are:
    - AD Account: service.test.bobj (all lower-letters)
    - 32-bit gsslib on the BO server, and 64 bit on the BW server side.
    - SNC0: p:service.test.bobj at DOMAIN
    - SU01 --> BO_Service ; SNC: p:service.test.bobj at DOMAIN
    - Entitlement system tab --> username: BO_Service
    SNC Name: p:service.test.bobj at DOMAIN
    - SNC settings tab:
    SNC Lib: c:\winnt\gsskrb5.dll
    Mutual Authentication settings: p:SAPServiceBP0 at DOMAIN
    In CMC, the role can be imported if "RFC activated" option unchecked in SNC0.
    I found a few threads on the same topic, but they are all not answered:
    SNC Client side configuration error
    SNC Configuration Error: Incomplete logon Data
    Can you please provide details of the solution if you have impleted a same scenario successsfully, or any thoughts to help the investigation?
    Thanks in advance!
    Regards,
    Jonathan

    Hi Ingo,
    Sorry for taking so long to reply, we are trying to set up server side trust and enable SSO; but we still couldn't success.
    What we did is:
    1. We followed installation guide chapter 6, generate certificate and PSE, etc. All looks good.
    2. Then we still have the "incomplete logon data" error when refreshing webi report after logon using Windows AD user ID.
    3. Then we trace the PFC connection, the log is as below. We checked several BO notes, e.g. 1500150, 1461247.. The part bothers us is that we even don't have URI displayed in the log when system trying to use SNC, and we couldn't get more info on this which make us very difficult to diagnosis.
    Can you please help? Thanks a lot!
    Thu Mar 31 10:54:46.857 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Authentication model for SAP connectivity is SSO
    Thu Mar 31 10:54:46.857 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Determining if we can connect using SNC. Calling CanAuthenticate...
    Thu Mar 31 10:54:46.919 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Unable to authenticate using SNC because the URI does not meet the minimum connection requirements.
    Thu Mar 31 10:54:46.919 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Determining if we can connect using SSO. Calling CanAuthenticate...
    Thu Mar 31 10:54:46.919 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Authentication model for SAP connectivity is SSO
    Thu Mar 31 10:54:47.013 ThreadID<1980> SAPMODULE : SAPAuthenticationService: The SAP SSO authentication process will fail because the SAP secondary credential are not properly updated and the password is blank.
    Thu Mar 31 10:54:47.013 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Trying to connect to SAP using this URI : occa:sap://;PROVIDER=sapbw_bapi,R3NAME=PB0,GROUP=BI_Group1,MSHOST=sapaupdb04,LANG=en,CLIENT=100,CATALOG="ZSPUM602",CUBE="ZSPUM602/ZSPUM602_Q50"
    Thu Mar 31 10:54:47.013 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Calling m_pRfcWrapper->RfcOpenEx() ...
    Thu Mar 31 10:54:47.154 ThreadID<1980> SAPMODULE : SAPAuthenticationService: RfcOpenEx(...) returned 0
    Thu Mar 31 10:54:47.154 ThreadID<1980> SAPMODULE : SAPAuthenticationService: Call to m_pRfcWrapper->RfcOpenEx() took 0.141 seconds
    Thu Mar 31 10:54:47.154 ThreadID<1980> SAPMODULE : SAPAuthenticationService: SAPAuthenticationService::~SAPAuthenticationService

  • BP_TASK Data Cleansing

    Happy New Year Experts,
    I have a question or three for you on Data Cleansing in the Web IC.  I will explain what I have done and what I need answers to.
    The setup of the Data Cleansing Cases and Account search is fine.  We can search for Business Partners and Merge Now or Merge Later and then search for cases if we chose to Merge Later.
    When we go into the Case to process it most of the functionality is fine.  What isn't ok is this BP_TASK config .
    I understand that in order to execute the 'START' button in the Case Processing screens you need to have the Task config setup.  I have done this to an extent as described below:
    1) Setup Number Ranges - IMG->CRM->Master Data->BP->DQA-> Maintain Number Ranges [Create line 01-0000000001-9999999999]
    2) In Task (1003) add the action profile BP_TASK - IMG->CRM->Transactions->Basic Settings->Define transaction types
    3) Create job CRM_BUPA_REALIGNMENT (periodic job 5-10 min)
    Point 3 is what I am coming unstuck with.  I cannot create a periodic job without assiging an ABAP Program to run etc.  There is nothing anywhere that says 'use this program or method' when creating the job step.
    Secondly, when I select the 'START' button in the Case Processing screen - after confirming the changes, the case errors with a vague message saying the case cannot be saved.  However, if I actually hit the 'SAVE' button the case saves and the changes I confirmed are processed between the various accounts.  So the whole process is about 95% great and 5% annoying.
    Before the questions, all authorisation settings are good as well.
    The questions are then:
    1)  What parameters are required above what I described for the periodic job?
    2)  Does the 'Task' transaction type need to be in the Business Transaction profile for the specific business role the Data Cleansing functionality is assigned within?
    3)  Once a task is created, I guess that the job will process these and that a user does not necessarily need to manual process these tasks?
    4)  Should I change the Action Definition and Condition at all over and above the standard setup that it is currently in?
    Any help and guidance would be great - I'm afraid I have 'Wood for Trees' syndrome now
    Many Thanks for any assistance.
    Regards,
    Mat.

    Hi Gregor,
    I have read your interesting blog. We have a similar kind of data cleansing activity. But when I tried to implement the same, we get a message that the Data Cleansing option is not available for the Insurance industry specific settings. Can you pls help us here ? When I analysed the same the FM 'FKK_CLEANSING_ALLOWED' has a hard coded code element as follows:
    *4.71: Event 7500 does not exist yet*
      IF 1 = 2.
         REFRESH: t_fbstab[].
         CALL FUNCTION 'FKK_FUNC_MODULE_DETERMINE'
           EXPORTING
              i_fbeve            = gc_event_7500
              i_only_application = gc_x
              i_applk            = applk
           TABLES
              t_fbstab           = t_fbstab[].
         LOOP AT t_fbstab INTO fbstab.
           CALL FUNCTION fbstab-funcc
              CHANGING
                c_cleansing_allowed = cleansing_allowed.
         ENDLOOP.
         IF 1 = 2.
    *     für Verwendungsnachweis
           CALL FUNCTION 'FKK_SAMPLE_7500'
              CHANGING
                c_cleansing_allowed = cleansing_allowed.
         ENDIF.
      ENDIF.
    Our SAP version is as folows:
    SAP_BASIS      620         0056 SAPKB62056             SAP Basis Component            
    INSURANCE     472           0010     SAPKIPIN10     INSURANCE 472 : Add-On Installation
    Any help is much appreciated.
    Thanks.
    I Peter

  • How to update the Dataset of a report running with universe?

    I am facing some issue in cross-tab.
    My requirement is , on click of any cell in cross-tab i want to convert  it into editable cell.
    Converting into editable cell is something which I achieved using below code:
    function onCellClick(e) {
      var text=this.innerHTML;
      var text = $.trim(this.innerHTML);
             $('<input />').attr({ type: 'text', name: 'text' }).appendTo($(this)).val(text).select().blur(
            function () {
                var newText = $(this).val();
                $(this).parent().text(newText).find('input').remove();
                var rIndex=$(this).closest('tr').index();
                data.data[rIndex]=newText;
                that.firePropertiesChanged(["data"]);
                //that.firePropertiesChanged(["visSelection"]);
        that.fireEvent("onSelect");
    I just modified the sample code.
    My report is  running with universe.
    Now I want to update the Dataset with this updated value.
    Can anyone provide any help on the same?
    Thanks

    Hi Michael,
    You got it right.
    Let me tell you the whole story:
    I have a weird requirement of creating editable grid. And the values which I edit into grid get saved into database.
    By using javascript I am able to edit grid cell item into HTML.
    Now after this I have two hitches:
    1. I am not able to get the updated cell values in Design Studio. I think this issues goes to SDK side. I try to create a external script variable and use this into Design Studio.But somehow it always throw blank.
    2.  I am not able to update dataset. I know updating the dataset permanently is not possible as its getting created from universe. But I just want to update the dataset so that any change in measure values also update total of that column.
    I start this thing this week only. So might be I am asking few stupid question. Please bear with me.
    Thanks
    Amit

  • Information Steward 4.2.1 not installing Data Cleansing Advisor (DCA) database

    Our situation is as follows:
    We have a SAP BusinessObjects BI 4.1 Sp3 system on Windows Server 2012.
    We are installing the following components onto a separate, new Microsoft Windows Server 2012 platform (in this order):
    - Information Platform Services (IPS) 4.1 Sp3
    - Data Services 4.2 Sp1 Patch 3
    - BusinessObjects BI 4.1 Sp3 Client Tools (for the BOE Metadata Integrator)
    - Information Steward  4.2 Sp1 Patch 3
    We verified the cross compatibility between IPS, DS and IS with SAP Note 1740516.
    SAP Note 1740516, as referenced in the PAM, lists the following: "BI 4.1 SP3 was officially certified with DS/IS 4.2 SP1 Patch 3".
    Our repository databases are located onto a separate Microsoft SQL Server 2008 Enterprise instance and the domain "DS/IS" service account under which IPS, DS and IS are installed, and will operate the Tomcat, DS and SIA services, has full access to these databases.
    All required repository databases, including the IS repository, were created at forehand.
    IPS and Data Services install fine without any problems. IS also installs fine, without the "Addnode" or Metadata Integrator installation issues that I have seen with previous versions. There were no errors of any kind.
    The "Information Steward" and "Data Services" nodes are available in the IPS CMC and can be accessed.
    However - the EIM APS is shown as "Running with errors"!
    Trouble shooting the EIM APS reveals that the Data Cleansing Advisor (DCA) Service is the culprit.
    Selecting the properties of the EIM APS also shows that the DCA Service could not be started and is marked in red. All other entries seem fine.
    Once the DCA Server is removed from the EIM APS, the EIM APS will return to a green "Running" status.
    This seems to indicate that the DCA Service is faulty. Checking the ..\BusinessObjects Enterprise XI 4.0\Data\DCA-SA\Data directory shows further evidence of this issue, as this directory is EMTPY. The expected DCA database file and the DCA log files are not there!
    We have tried a number of things:
    1. Removing the DCA service from the EIM APS, restarting the SIA/IPS, adding the DCA service to the EIM APS again.
    2. Uninstalling IS 4.2 Sp1 Patch 3, rebooting the server and repeating the installation again.
    2. Uninstalling IS 4.2 Sp1 Patch 3, rebooting the server and repeating the installation with IS 4.2 Sp1 Patch 0 (GA release)
    3. Upgrading the IS 4.2 Sp1 Patch 0 (GA release) with the IS 4.2 Sp1 Patch 3 version.
    None of these have made any difference whatsoever.
    I have found 1 post on SCN with the exact same problem:  http://scn.sap.com/thread/3475164
    Unfortunately, the listed solution of reinstalling has clearly not worked for us.
    Has anybody seen this problem before? Any ideas what else we can do to resolve this issue? Our next port of call is opening a SAP support call but we really want to get IS up and running ASAP  - so any suggestions would be most welcome!

    Hi Erik and James,
    This has been fixed in 4.2 SP2 and will be fixed in an upcoming patch for 4.2 SP 1, as well.
    Below are steps you can take to work around the issue you are seeing:
    1. Copy the files that are in <Install Directory>\InstallCache\im.lac.icc.cdq.cognitivedq.pjs.services.backend-4.0-core-nu\14.2.1.<your build number>\actions\plugins\14.1.2
    To
    <Install Directory>\SAP BusinessObjects\SAP BusinessObjects Enterprise XI 4.0\java\pjs\services\ICC.DCAService\lib
    2. Restart the EIMAdaptiveProcessingServer.
    Thank you,
    Lynne

  • Interface and conversion testing of SAP environments with Master Data

    Hi guy's
    Please let me know if some one of you know about SAP Conversion Project. Below you have more description:
    - testing of Interfaces from Legacy Systems
    - testing of conversion programs used in the conversion or transposition of data from legacy systems
    - data cleansing activities associated with conversion
    - identify and populate various SAP environments with Master Data necessary for both conversion and interface testing
    Any detail info in that and what kind of knowledge you need to have in some of project will be useful.
    Thanks in advance
    Adeel

    Hi Yannick,
    I am trying to do the exact same thing. Have you gotten any further on this issue?

  • How to fill a prompt in webi with current date?

    A webi report is being generated on BEx Query.
    IN BEx query variables are defined for some date fields to prefill them with current date.
    Foe ex. there is license start date. For this field variable is defined so that in variable screen License start date is automatically filled with greater than or equal to current date(which is system date). WHich is a mandatory variable.
    On BO side, a webi report is being developed on a Universe (based on this BEx query).
    When the webi report is run, the license start date(mandatory variable in BEx) is automatically prompted, but date is not getting prefiiled with current date.
    How to prefill this prompt with current date?
    the prompt should be like
    LIcense start date >= <current date>

    Hi,
    Yes, it isn't possible to reference anything like 'Current Date' in the universe (like you would with a SQL data source).
    The solution I have seen recently goes like this, I don't know the exact details (so I might have some terminology wrong), but someone with ABAP skills should get what I mean.
    Assuming you want to filter on 0CALDAY.
    In ABAP, create a customer EXIT that calculates the current date as a number (so today might be 40651). Do the same for 0CALDAY.
    In the BEx create a formula variable that references each of those, then a Formula Key Figure (I forget the exact name of them) that's just the current day number minus the 0CALDAY number. This gives you the 'age' of that row of data.
    Your Webi report can then have a built in filter (or prompt with a default value) of, say Age <= 30. The report can be scheduled and will always run with the last 30 day's data.
    We actually did the customer exists to show the 'age' in months.
    Sorry I can't explain it any better, but it's working great for us.
    Hope that helps.

Maybe you are looking for

  • Could not save because the ICC profile description is invalid

    I'm getting this error whenever trying to do a "save" or "save as" from Photoshop CC. I've tried converting to different color profiles but that doesn't seem to make a difference. Even if I create a new blank document and try to save it, I get the sa

  • Calling ucomm in ALV_Grid

    Hi experts, I am trying to do some thing but I have some difficulties. I would appreciate if some one tells me how to achieve this. I am trying to invoke the ucomm and CALL SELECTION-SCREEN 500 STARTING AT 10 10 WHEN the field 'CITYFROM' is pressed.

  • Anyway to force a user to logout?

    Hi, I wonder whether there is a way for root to force a user out of the system, if the user is currently in it. I get two reasons to do it: 1) sometime after a remote telnet PC crash, I find the user is reported to be still in the host from "who". I

  • Hide budget button in portal

    Dear expert Please let's me know how  to hide budget button in 'Additional Specifications tab'  at the portal. my manager want some user not allowed to see that budget information. thx

  • Cannot find trash in iPhoto '11

    I am having trouble finding the Trash folder in iPhoto '11. In looking around at other discussions, it seems it should be beneath the "Recent" label. However, I can only find my Photo Stream, Last Import, Last 12 Months, and Flagged Folders/Albums be