Re: Data Source: 0CRM_PROD_OBJ_LIST

Dear All
We are using CRM data source:0CRM_PROD_OBJ_LIST(Delta) and ODS 0CRM_OL.
Issue: Data is loaded in to PSA correctly but when it is staged to ODS it is skipping some of the records.
Between this PSA and ODS their is a Start routine in Update rule.
Update rule name; 4ETV81OKAKCXHUFPSRYXMIVAT
Code in Start routine is attached.
1. Can Some one please check whether it is standard or Z? Because I had debugged the program I found no problem in it.
Sorry to attach the code here it is just to verify.
YPES: BEGIN OF new_data_structure,
         item_guid(32) TYPE c,
         item_changed_ts(8) TYPE p,
         recordmode TYPE c,
       END OF new_data_structure.
    This global table contains information about relevant update
    records. This table is filled when the first data package is
    processed
DATA:  gt_new_data TYPE SORTED TABLE OF new_data_structure
       WITH NON-UNIQUE KEY item_guid item_changed_ts recordmode.
DATA: IN    TYPE F,
      OUT   TYPE F,
      DENOM TYPE F,
      NUMER TYPE F.
data: rate like tcurr-UKURS.
constants: c_msgty_e value 'E'.
$$ end of global - insert your declaration only before this line   -
The follow definition is new in the BW3.x
TYPES:
  BEGIN OF DATA_PACKAGE_STRUCTURE.
     INCLUDE STRUCTURE /BIC/CS0CRM_PROD_OBJ_LIST.
TYPES:
     RECNO   LIKE sy-tabix,
  END OF DATA_PACKAGE_STRUCTURE.
DATA:
  DATA_PACKAGE TYPE STANDARD TABLE OF DATA_PACKAGE_STRUCTURE
       WITH HEADER LINE
       WITH NON-UNIQUE DEFAULT KEY INITIAL SIZE 0.
FORM startup
  TABLES   MONITOR STRUCTURE RSMONITOR "user defined monitoring
           MONITOR_RECNO STRUCTURE RSMONITORS " monitoring with record n
           DATA_PACKAGE STRUCTURE DATA_PACKAGE
  USING    RECORD_ALL LIKE SY-TABIX
           SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS
  CHANGING ABORT LIKE SY-SUBRC. "set ABORT <> 0 to cancel update
$$ begin of routine - insert your code only below this line        -
fill the internal tables "MONITOR" and/or "MONITOR_RECNO",
to make monitor entries
**declaration
DATA: lt_active_data TYPE TABLE OF /bi0/acrm_ol00.
DATA: ls_active_data TYPE /bi0/acrm_ol00.
  DATA: lt_active_data TYPE TABLE OF DATA_PACKAGE_structure.
  DATA: ls_active_data TYPE DATA_PACKAGE_structure.
  DATA: lv_fsname(60) TYPE c.
  DATA: lv_psa_name TYPE rsodstech.
  DATA: lv_dso_tabname TYPE tabname.
  DATA: ls_new_data TYPE new_data_structure.
  DATA: ls_new_data_pre TYPE new_data_structure.
  DATA: ls_DATA_PACKAGE TYPE DATA_PACKAGE_structure.
  DATA: help_index LIKE sy-tabix.
  FIELD-SYMBOLS: <fs_minfo_requnr> TYPE ANY.
  FIELD-SYMBOLS: <fs_minfo_datapakid> TYPE ANY.
  FIELD-SYMBOLS: <fs_minfo_updmode> TYPE ANY.
  FIELD-SYMBOLS: <fs_minfo_logsys> TYPE ANY.
**constants
  CONSTANTS:  lc_dso_name TYPE rsdodsobject VALUE '0CRM_OL'.
**end of declaration
**fetch request number
  lv_fsname = 'g_s_minfo-requnr'.
  ASSIGN (lv_fsname) TO <fs_minfo_requnr>.
**fetch data package number
  lv_fsname = 'g_s_minfo-datapakid'.
  ASSIGN (lv_fsname) TO <fs_minfo_datapakid>.
**fetch update mode
  lv_fsname = 'g_s_minfo-updmode'.
  ASSIGN (lv_fsname) TO <fs_minfo_updmode>.
**fetch source system
  lv_fsname = 'g_s_minfo-logsys'.
  ASSIGN (lv_fsname) TO <fs_minfo_logsys>.
**delete all data_package entries that...
**...are not object list elements
**...are dummy deletion records
**...are erroneous
  DELETE DATA_PACKAGE WHERE flag NE 'O'.
OR
                           recordmode EQ 'D'.
**delta handling only processed for update modes 'delta' and 'repeat'
  IF <fs_minfo_updmode> = 'D' OR <fs_minfo_updmode> = 'R'.
data selection and deletion handling only for the first data package
    IF <fs_minfo_datapakid> = '000001'.
   fetch technical name of actual PSA table
   with I_SELTYPE = 'D' (Date)
      CALL FUNCTION 'RSAR_ODS_NAME_GET'
        EXPORTING
          i_logsys                = <fs_minfo_logsys>
          i_isource               = '0CRM_PROD_OBJ_LIST'
          i_istype                = 'D'
          i_date                  = sy-datum
        I_VERSION               =
          i_seltype               = 'D'
        I_SEGMENT_ID            =
        IMPORTING
        E_ODSNAME               =
          e_odsname_db            = lv_psa_name
        E_ODS_TABTYPE           =
        E_S_ODS                 =
        E_T_ODSFIELDS           =
        E_PARTITIONED           =
        EXCEPTIONS
          parameter_failure       = 1
          no_ods_found            = 2
          no_fields_to_ods        = 3
          OTHERS                  = 4
      IF sy-subrc = 0.
     select new data from PSA Table
        SELECT * FROM (lv_psa_name)
        INTO CORRESPONDING FIELDS OF TABLE gt_new_data
     only select data from actual request
     ...do not select erroneous records
     ...do not select object list records
        WHERE request      = <fs_minfo_requnr> AND
              bwsttecsys2  NE '10' AND
              flag_pl_ol   NE 'P'.
        IF sy-subrc = 0.
       delete adjacent duplicates in gt_new_data
          DELETE ADJACENT DUPLICATES FROM gt_new_data
          COMPARING ALL FIELDS.
       get technical name of active DSO table
          CALL METHOD cl_rsd_odso=>get_tablnm
            EXPORTING
              i_odsobject   = lc_dso_name
              i_tabt        = rsdod_c_tabt-active
            IMPORTING
              e_tablnm      = lv_dso_tabname
           E_TTYPENAME   =
           E_VIEWNM      =
           E_CHNGLOGNM   =
           E_INFOSOURCE  =
           E_DATASOURCE  =
            EXCEPTIONS
              OTHERS        = 1.
          IF sy-subrc <> 0.
            lv_dso_tabname = '/bi0/acrm_ol00'.
          ENDIF.
       select active data from ODS A-Table for relevant items
          SELECT * FROM (lv_dso_tabname)
          INTO CORRESPONDING FIELDS OF TABLE lt_active_data
          FOR ALL ENTRIES IN gt_new_data
          WHERE crm_itmgui = gt_new_data-item_guid.
       delete older after images, only keep the latest one
          LOOP AT gt_new_data INTO ls_new_data.
            IF ls_new_data-item_guid EQ ls_new_data_pre-item_guid AND
               ( ls_new_data-item_changed_ts NE
  ls_new_data_pre-item_changed_ts OR
            for dummy deletion records on item level there could be
            several entries with same value of item_changed_ts
            because the changing date is not updated in case of
            deletion of an item
            in this case we only keep the dummy deletion record
            itself
               ls_new_data-recordmode = 'D' ).
              help_index = sy-tabix - 1.
              DELETE gt_new_data INDEX help_index.
            ENDIF.
            MOVE ls_new_data TO ls_new_data_pre.
          ENDLOOP.
       Now we delete all dummy deletion records in gt_new_data
          DELETE gt_new_data WHERE recordmode = 'D'.
       modify data of gt_active_data for field 'RECORDMODE' = 'D'
       deletion records are written to the data package (Index 1)
          LOOP AT lt_active_data INTO ls_active_data.
            ls_active_data-recordmode = 'D'.
            MOVE-CORRESPONDING ls_active_data TO ls_DATA_PACKAGE.
            INSERT ls_DATA_PACKAGE INTO DATA_PACKAGE INDEX 1.
          ENDLOOP.
       delete records of older after images
       do not consider deletion records
          LOOP AT DATA_PACKAGE INTO ls_DATA_PACKAGE
               WHERE recordmode NE 'D'.
            help_index = sy-tabix.
         relevant for update?
            READ TABLE gt_new_data WITH KEY
              item_guid = ls_DATA_PACKAGE-crm_itmgui
              item_changed_ts = ls_DATA_PACKAGE-crm_itchts
              BINARY SEARCH
              TRANSPORTING NO FIELDS.
         if not relevant, delete it!
            IF sy-subrc <> 0.
              DELETE DATA_PACKAGE INDEX help_index.
            ENDIF.
            CLEAR ls_DATA_PACKAGE.
            CLEAR help_index.
          ENDLOOP.
        ENDIF.
      ENDIF.
for data packages > 000001
    ELSE.
   delete records of older after images
      LOOP AT DATA_PACKAGE INTO ls_DATA_PACKAGE.
        help_index = sy-tabix.
     relevant for update?
        READ TABLE gt_new_data WITH KEY
          item_guid = ls_DATA_PACKAGE-crm_itmgui
          item_changed_ts = ls_DATA_PACKAGE-crm_itchts
          BINARY SEARCH
          TRANSPORTING NO FIELDS.
     if not relevant, delete it!
        IF sy-subrc <> 0.
          DELETE DATA_PACKAGE INDEX help_index.
        ENDIF.
        CLEAR ls_DATA_PACKAGE.
        CLEAR help_index.
      ENDLOOP.
    ENDIF.
  ENDIF.
if abort is not equal zero, the update process will be canceled
  ABORT = 0.
$$ end of routine - insert your code only before this line         -
ENDFORM.
Analysis on it:
   1. I read in the SAP help that u201CBefore you perform a delta update for this ODS object, ensure that data previously loaded into the ODS object is activated. Otherwise the delta update could lead to inconsistencies.u201D  (http://help.sap.com/saphelp_tm80/helpdata/en/ba/21423fc657e07fe10000000a114084/frameset.htm). But I had seen in the ODS
(0CRM_OL) manage that number of requests are activated at once so I had deleted the ODS requests and reloaded one by one by activating them sequentially.
Till the request which has got problem that is SID of request 90949 (PSA) dated 06/09/2011. After loading the said request the contract number (0000132672) was still missing in the ODS. Then I had debugged the 9 data package of 90949 in which the contract number 0000132672 was present. I found that after the code for Start routine is completed the contract number was still existing in the internal table. But that is not updated into the New(Queue)  Table of ODS.
Unable to debug the background job which is generated while updating the data from PSA to ODS, as no real data found at that time. 
Code Process in Start Routine (It will use both the active table of ODS and PSA data (of relevant request )):
Conclusion: There is no problem in Start routine. (Seems it is standard routine (need to be confirmed)).
2.       Found in Sap help that u201CWhen you use this InfoSource to load data into the ODS objects Product Lists (0CRM_PL) and Object Lists (0CRM_OL), you must set the processing method Only PSA with the option Update Subsequently in Data Targets in the corresponding InfoPackage. The update rules for the above ODS objects use a global table in their start routine to enable delta updates for this extractor. Once the data to be loaded has been distributed to multiple data packages, the contents of this global table are only retained if you have set the above processing method. If you set any other processing method, the contents of this global table are deleted between processing of data packagesu201D (http://help.sap.com/saphelp_bw33/helpdata/en/05/bb423fd9575003e10000000a114084/frameset.htm).
For this to implement need to delete data from ODS and from PSA so as to load the data with new info package (Process type Only PSA update sub sequential).
Edited by: sreekanth goud on Oct 25, 2011 8:02 AM

Hi Sreekanth,
First check whether you ur DSO had the right combination of key fields,once u r sure about it,check the start routine,please send the code. it's heavy to see.
Thanks & Regards
Rams Thota

Similar Messages

  • Error in Creation of BW data source

    Hi All,
    While creating a data source, i am selecting the Origin type Business ware house cube & in the  RFC destination field i can able to select the BW syste. The problem is when i go to the next filed BW report and try to get the relevant BW report from F4 the system shows a log error as "System error in System" Message no. CRM_MKTTG_DS_MISC033.
    In second case
    I have taken an existing BW Data source and tried to recreate it manually filling all the fields with existing data source, RFC destination, BW report, Business partner and Function module.
    In this case the system doesnu2019t throws any error, it is working fine if I create manually.
    I am in a dilema why the system behaving like this. could you please throw some light on this issue.
    Thanks in advance,
    Vijay.

    Hi,
      See the link.
    http://help.sap.com/saphelp_crm50/helpdata/en/00/dc54384ac9a81be10000009b38f8cf/frameset.htm
    regards
    Srinu

  • Unable to capture the Data Source into a Transport Request

    Hi All,
    We have a product hierarchy and we are using the data source :4R_PRODH_D_LGEN_HIER for the hierarchy.
    Now we need to transport this structure to the quality environment but we were not able to capture the datasource:4R_PRODH_D_LGEN_HIER into a transport request.
    When ever we activate the data source:4R_PRODH_D_LGEN_HIER it is asking for the Package and the Transport Request Number.If we give these details and save it, data source is not getting captured in the request, only the "bject Directory Entry" is getting captured.
    Can someone please guide me on how to capture the datasource under "Data Sources in BW" in a transport request.
    Regards,
    Sachin Dehey.

    Hi Sachin,
    Hierarachy datasource is not captured as Attributes and Text Datasource. So what ever you have done is correct.
    What ever is captured in Object Directory Entry is correct. So go ahead with your transports, once transport is done check the Hierarchy Infopackage with Available OLTP hierarchies and load the data.
    Most important thing first see that the all Master & Transactional Datasources are transported in R/3 Dev to QA to PRD
    In BW, datasources are not transported, only their replica is transported.
    Transportation of Datasource is done in R/3. Only their replica is transported in BW.
    So wht ever you have done till now is correct. So go ahead.
    While attaching Hierarchy Datasource it is captured only in "Object Directory Entry"
    Regards,
    Vishnu.

  • Using a SQL data source and XML data source in the same template

    I am trying to develop a template for the Request for Quote report generated in Apps 11.5.10. I have loaded the data from the XML output into the template, but I am missing one field - I need the org_id from the po_headers table. Is it possible to use a sql data source (i.e., "select org_id from po_headers_all where po_header_id = [insert header_id from xml data]...") in addition to the xml data source to populate the template at runtime? When you use the Insert > SQL functionality is it static at the time the template is created, or does it call to the database at runtime? I've looked through all the docs I could find, but this isn't clear.
    Thanks for any help or suggestions you may have.
    Rhonda

    Hi Pablo
    Thats a tough one ... if you go custom with a data template you will at least get support on the data template functionality ie you have a problem when you try and build one. You will not get support on the query inside the data template as you might have gotten with the Oracle Report, well you could at least log a bug against development for a bad query.
    Eventually that Oracle Report will be converted by development anyway, theres an R12 project going on right now to switch the shipped OReports to data templates. AT this point you'll be fully supported again but:
    1. You have to have R12 and
    2. You'll need to wait for the patch
    On reflection, if you are confident enough in the query then Oracle will support you on its implementation within a data template. Going forward you may be able to swap out your DT and out in the Oracle one without too much effort.
    Regards, Tim

  • "Open sql" report no data sourced defined in windows

    when I want to import data from sql database, the Open sql window report error
    "There are no data sourced defined. Please create one to continue",
    but I have defined the odbc data source using windows 's odbc administrator, why the error?
    chuliang

    A couple of possibilities.
    First are you licensed for Sql interface? If not, you might get that message.
    Second, where did you create the ODBC connection on your computer or the server. It has to be on the server.
    Third, did you create the odbc as a system driver? it needs to be so
    Fourth, what operating system? If it's unix or AIX there is a hole bunch more you need to do to get odbc connections working. If it's windows, it's pretty easy.
    Fifth, did you test the connection before trying it in Essbase?

  • Excel-2007 cannot connect to Oracle ODBC data source, Control Panel can.

    <p>
    I cannot make an ODBC connection from Exce-2007 to Oracle work. I am an expert Excel and VBA user (since 1994) and I have frequently used Excel to access ODBC databases, including Oracle (I have done this with Excel-2003 both with worksheet queries and have written VBA ADO-connection routines). And even though in Excel-2007 a worksheet ODBC query is supposed to be easier to create than in previous version of Excel, my connection fails. Any suggestions and all help are welcome and much appreciated.
    DETAILS
    </p>
    <p>
    1) <strong>What is my system?</strong> I am using Excel-2007 on Windows Vista x64 and Oracle server v.11g on my computer (all this is on my computer, no network issues).
    2)<strong> Why use Excel with Oracle at all?</strong> I use Excel-2007 to access Oracle rather than Access-2007 (or any other application like TOAD, etc.) because I do engineering calculations with the data stored in Oracle. These calculations are easier done in Excel (I suppose that one alternative to this could be to use some sql or Access to get the data from the database, then store it as plain vanilla CSV file, then open this file in Excel, then do the math (the math involves complex optimisation algorithms), then save the results as CSV, then use some sql or Access to put the data back into the database. Howwever this does not strike me as a quick or neat solution. And after all Excel has been designed to access ODBC databases, so why not use it?)
    3) <strong>What do I do in Excel-2007 that won't work?</strong> I create an ODBC link to Oracle that does not work. In Excel-2007 this is straightforward:
    </p>
    <ul>
         <li>define an ODBC connection (Data tab --&gt; From other sources --&gt; From data connection wizard);</li>
         <li>define a query on the worksheet -- that's it, this is all!</li>
    </ul>
    <p>
    I start with creating an ODBC connection:
    a) I choose an ODBC data source type: <strong><em>ODBC DSN</em></strong>
    b) Excel-2007 displays the list of the available ODBC data sources. I see in it <strong><em>my Oracle database name</em></strong> and I select it.
    c) Excel-2007 displays the Data Link Properties:
    - the "Provider" has a list of the OLE DB drivers with preselected "<strong><em>Microsoft OLE DB Provider for ODBC Drivers</em></strong>". I keep this default selection.
    - the "Connection" tab has a connection string "<strong><em>DSN=&lt;my database name&gt;</em></strong>" which I keep, it also has fields for the <strong><em>user name</em></strong> and the <strong><em>password</em></strong>, which I fill with the correct credentials.
    - Finally there is a button "Test Connection", which when I click produces the following error message:
    <strong><font color="#ff0000">"Test connection failed because of an error in initializing provider.</font></strong><strong><br />
    </strong><strong><font color="#ff0000">Unespecified error"</font></strong><strong>
    </strong>
    4) <strong>Additional food for thought:</strong>
    a) In the above walk-through the only data, which I type, are the user name and password, everything else is selected from lists offered by Excel-2007, hence any possibility of typos being the cause of the problem can safely be discarded.
    b) I can test the ODBC driver in the Control Panel and it shows that it can connect to the Oracle database:
    - in <strong><em>Control Panel --&gt; Admin Tools --&gt; Data Sources (ODBC)</em></strong> on the "User DSN" tab I can see the list of the available ODBC data sources (same list as in Excel-2007, point 3b above) with the name of my database in it;
    - selecting the name of my database from the list of the sources and clicking "Configure" button opens a tab with <strong><em>Data Source Name</em></strong> (same as in Excel-2007), TNS Service Name and User ID. I enter <strong><em>&lt;user name&gt;/&lt;password&gt;</em></strong> and click "Test Connection" button. A message "<strong><em>Connection successful</em></strong>" appears (just for a test I enter <u>incorrect user credentials</u> and "<strong><em>Unable to connect</em></strong>" message appears)
    BOTTOM LINE
    </p>
    <p>
    The procedure for using an ODBC connection from Excel is very simple, in the past I have created and used such connections numerous times with Excel-2003 and earlier on Win-XP and earlier. But now on Excel-2007 and Vista-x64 I cannot make it work.
    Also, testing an ODBC connection driver is really easy and simple to be done in the Control Panel. There testing the same ODBC connection, which fails in Excel-2007, results in success.
    I am frustrated by the simplicity of the problem and yet the persistant error. I have lost now two full days in failed attempts to make the simple procedure work and in searching the internet for answers.
    All help is highly appreciated <img class="emoticon" src="images/emoticons/happy.gif" border="0" alt="" width="16" height="16" />,
    Plamen
    </p>

    Did you find the solution to your problem?
    If not, I think I may know.
    Excel 2007 is a 32-bit application.
    When installing 32-bit applications in a 64-bit environment, the "default" location is:
    C:\Program files (x86)\...
    Excel then launches from this location.
    However, when it connects to Oracle and passes the name of the calling program, Oracle attempts to "interpret" the value of (x86) as if
    the value within the parenthesis are being passed as a reference. Of course, iOracle doesn't find anything, so the result is (),
    and then it cannot return the connection info back to the calling application.
    I corrected it by installing Excel in C:\Program Files\, and once launched from that location, it works the same as on the 32-bit machines.
    However, at my location, they are FORCING Excel to be installed in the(x86) location.
    What I'm trying to discover now is:
    Is it possible to flag Oracle to NOT process the embedded variables?
    Or, is it possible to assign a variable in Oracle such that x86 = "(x86)", so that the end result is viable?
    Have you had any luck with your installation?
    thanks,
    Paul

  • Refreshing the Data Source View in Analysis Services

    I have added columns to the SQL Database table that is used as a dimension in an Analysis Services Cube.  The new columns will be used as additional Property Fields for the dimension.  When I attempted to refresh the Data Source view so that the additional columns are present, I am given the following error:
    System.Data
    Property not accessible because 'Parent Columns and Child Columns don't have type-matching columns'
    I have done nothing to the columns used for the parent of child and the error message provides nothing to gon on. Does anyone have any ideas on this?
    Gary

    Olga,
    Thanks for your response.  I will try and answer your questions
    1) I have not tried removing the columns yet.  I will try that this afternoon but have limited hope.  The two columns I added are simple text columns that will be used as attributes in the dimension.  I have made no change to the parent or child columns.
    2) The table I modified is the source table for a parent-child dimension.
    3) The reference to the "check list" does not take me to any kind of check list.
    4) The parent-child dimensions I am trying to modify have been in use for months and the parent and child columns do have the dame data types.
    5) I have also check the data types between the dimension table and the fact table.  they use the same data types (small int).
    6) I have not made a collection for the parent key, it is a single column. The remainder of your last paragraph is not clear to me. Can you give me an example.
    I am fairly inexperienced with Analysis Services, please talk slow and use small words  :-)
    Thanks again for your help!
    Gary

  • OSB: Cannot acquire data source error while using JCA DBAdapter in OSB

    Hi All,
    I've entered 'Cannot acquire data source' error while using JCA DBAdapter in OSB.
    Error infor are as follows:
    The invocation resulted in an error: Invoke JCA outbound service failed with application error, exception: com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/DBAdapter1/RetrievePersonService [ RetrievePersonService_ptt::RetrievePersonServiceSelect(RetrievePersonServiceSelect_inputParameters,PersonTCollection) ] - WSIF JCA Execute of operation 'RetrievePersonServiceSelect' failed due to: Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/soademoDatabase].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.soademoDatabase'. Resolved 'jdbc'; remaining name 'soademoDatabase'.
    ; nested exception is:
    BINDING.JCA-11622
    Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    JNDI Name for the Database pool: eis/DB/soademoDatabase
    JNDI Name for the Data source: jdbc/soademoDatabase
    I created a basic DBAdapter in JDeveloper, got the xsd file, wsdl file, .jca file and the topLink mapping file imported them into OSB project.
    Then I used the .jca file to generate a business service, and tested, then the error occurs as described above.
    Login info in RetrievePersonService-or-mappings.xml
    <login xsi:type="database-login">
    <platform-class>org.eclipse.persistence.platform.database.oracle.Oracle9Platform</platform-class>
    <user-name></user-name>
    <connection-url></connection-url>
    </login>
    jca file content are as follows:
    <adapter-config name="RetrievePersonService" adapter="Database Adapter" wsdlLocation="RetrievePersonService.wsdl" xmlns="http://platform.integration.oracle/blocks/adapter/fw/metadata">
    <connection-factory location="eis/DB/soademoDatabase" UIConnectionName="Connection1" adapterRef=""/>
    <endpoint-interaction portType="RetrievePersonService_ptt" operation="RetrievePersonServiceSelect">
    <interaction-spec className="oracle.tip.adapter.db.DBReadInteractionSpec">
    <property name="DescriptorName" value="RetrievePersonService.PersonT"/>
    <property name="QueryName" value="RetrievePersonServiceSelect"/>
    <property name="MappingsMetaDataURL" value="RetrievePersonService-or-mappings.xml"/>
    <property name="ReturnSingleResultSet" value="false"/>
    <property name="GetActiveUnitOfWork" value="false"/>
    </interaction-spec>
    </endpoint-interaction>
    </adapter-config>
    RetrievePersonService_db.wsdl are as follows:
    <?xml version="1.0" encoding="UTF-8"?>
    <WL5G3N0:definitions name="RetrievePersonService-concrete" targetNamespace="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService" xmlns:WL5G3N0="http://schemas.xmlsoap.org/wsdl/" xmlns:WL5G3N1="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService" xmlns:WL5G3N2="http://schemas.xmlsoap.org/wsdl/soap/">
    <WL5G3N0:import location="RetrievePersonService.wsdl" namespace="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService"/>
    <WL5G3N0:binding name="RetrievePersonService_ptt-binding" type="WL5G3N1:RetrievePersonService_ptt">
    <WL5G3N2:binding style="document" transport="http://www.bea.com/transport/2007/05/jca"/>
    <WL5G3N0:operation name="RetrievePersonServiceSelect">
    <WL5G3N2:operation soapAction="RetrievePersonServiceSelect"/>
    <WL5G3N0:input>
    <WL5G3N2:body use="literal"/>
    </WL5G3N0:input>
    <WL5G3N0:output>
    <WL5G3N2:body use="literal"/>
    </WL5G3N0:output>
    </WL5G3N0:operation>
    </WL5G3N0:binding>
    <WL5G3N0:service name="RetrievePersonService_ptt-bindingQSService">
    <WL5G3N0:port binding="WL5G3N1:RetrievePersonService_ptt-binding" name="RetrievePersonService_ptt-bindingQSPort">
    <WL5G3N2:address location="jca://eis/DB/soademoDatabase"/>
    </WL5G3N0:port>
    </WL5G3N0:service>
    </WL5G3N0:definitions>
    Any suggestion is appricated .
    Thanks in advance!
    Edited by: user11262117 on Jan 26, 2011 5:28 PM

    Hi Anuj,
    Thanks for your reply!
    I found that the data source is registered on server soa_server1 as follows:
    Binding Name: jdbc.soademoDatabase
    Class: weblogic.jdbc.common.internal.RmiDataSource_1033_WLStub
    Hash Code: 80328036
    toString Results: ClusterableRemoteRef(8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1 [8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1/291])/291
    Binding Name: jdbc.SOADataSource
    Class: weblogic.jdbc.common.internal.RmiDataSource_1033_WLStub
    Hash Code: 92966755
    toString Results: ClusterableRemoteRef(8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1 [8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1/285])/285
    I don't know how to determine which server the DBAdapter is targetted to.
    But I found the following information:
    Under Deoloyment->DBAdapter->Monitoring->Outbound Connection Pools
    Outbound Connection Pool Server State Current Connections Created Connections
    eis/DB/SOADemo AdminServer Running 1 1
    eis/DB/SOADemo soa_server1 Running 1 1
    eis/DB/soademoDatabase AdminServer Running 1 1
    eis/DB/soademoDatabase soa_server1 Running 1 1
    The DbAdapter is related to the following files:
    C:\ Oracle\ Middleware\ home_11gR1\ Oracle_SOA1\ soa\ connectors\ DbAdapter. rar
    C:\ Oracle\ Middleware\ home_11gR1\ Oracle_SOA1\ soa\ DBPlan\ Plan. xml
    I unzipped DbAdapter.rar, opened weblogic-ra.xml and found that there's only one data source is registered:
    <?xml version="1.0"?>
    <weblogic-connector xmlns="http://www.bea.com/ns/weblogic/90">
    <enable-global-access-to-classes>true</enable-global-access-to-classes>
    <outbound-resource-adapter>
    <default-connection-properties>
    <pool-params>
    <initial-capacity>1</initial-capacity>
    <max-capacity>1000</max-capacity>
    </pool-params>
    <properties>
    <property>
    <name>usesNativeSequencing</name>
    <value>true</value>
    </property>
    <property>
    <name>sequencePreallocationSize</name>
    <value>50</value>
    </property>
    <property>
    <name>defaultNChar</name>
    <value>false</value>
    </property>
    <property>
    <name>usesBatchWriting</name>
    <value>true</value>
    </property>
    <property>
    <name>usesSkipLocking</name>
    <value>true</value>
    </property>
    </properties>
              </default-connection-properties>
    <connection-definition-group>
    <connection-factory-interface>javax.resource.cci.ConnectionFactory</connection-factory-interface>
    <connection-instance>
    <jndi-name>eis/DB/SOADemo</jndi-name>
              <connection-properties>
                   <properties>
                   <property>
                   <name>xADataSourceName</name>
                   <value>jdbc/SOADataSource</value>
                   </property>
                   <property>
                   <name>dataSourceName</name>
                   <value></value>
                   </property>
                   <property>
                   <name>platformClassName</name>
                   <value>org.eclipse.persistence.platform.database.Oracle10Platform</value>
                   </property>
                   </properties>
              </connection-properties>
    </connection-instance>
    </connection-definition-group>
    </outbound-resource-adapter>
    </weblogic-connector>
    Then I decided to use eis/DB/SOADemo for testing.
    For JDeveloper project, after I deployed to weblogic server, it works fine.
    But for OSB project referencing wsdl, jca and mapping file from JDeveloper project, still got the same error as follows:
    BEA-380001: Invoke JCA outbound service failed with application error, exception:
    com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/DBAdapterTest/DBReader [ DBReader_ptt::DBReaderSelect(DBReaderSelect_inputParameters,PersonTCollection) ] - WSIF JCA Execute of operation 'DBReaderSelect' failed due to: Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/SOADataSource].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.SOADataSource'. Resolved 'jdbc'; remaining name 'SOADataSource'.
    ; nested exception is:
    BINDING.JCA-11622
    Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/SOADataSource].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.SOADataSource'. Resolved 'jdbc'; remaining name 'SOADataSource'.
    You may need to configure the connection settings in the deployment descriptor (i.e. DbAdapter.rar#META-INF/weblogic-ra.xml) and restart the server. This exception is considered not retriable, likely due to a modelling mistake.
    It almost drive me crazy!!:-(
    What's the purpose of 'weblogic-ra.xml' under the folder of 'C:\Oracle\Middleware\home_11gR1\Oracle_OSB1\lib\external\adapters\META-INF'?
    ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

  • Error while activating the data source.

    Dear all,
    We are using the BI version 7.00 - i have a problem i.e., while creating the infopackage for the data source
    0PM_OM_OPA_2 the following error is displayed.
    "No active transfer rule for DS 0PM_OM_OPA_2 and ....source system CLNTDEV400; create first..."
    Moreover the data source is found to be in version "M" - while activating the same when in click on the display/change icon , the following error is popped up.
    "You can only edit 3.x DataSources using transfer rule
    maintenance Do you want to create the transfer rules now?"
    What could be the reason for this error. What are the things that are to be done to replicate the data and pull data?
    Valid experts comment is expected in this regard.
    Regards,
    M.M

    Hi Magesh,
    Replicate your data source in BW and then right click onit and choose Assign Info source, it may take you to BC wherein you may need to install the Transfer Rules and related info objects etc.... Once you maintain this data source in the source system then you can create an info pack withput any extra effort.
    Hope it helps..

  • Error while creating a new report data source in Sharepoint Document library

    I installed SSRS(sql server 2012 SP1) in my sharepoint 2013 farm(single server installation).
    I have installed SSRS addin for report also. 
    I created a document library and included the report data source content type, report builder report and report builder model content types. On click of any of these content types i get the below error. 
    "new Document requires a Microsoft SharePoint Foundation-compatible application and web browser. To add document to document library, click on the 'Upload' button."
    I am using windows 7 64-bit operation system with Google chrome and IE 8 64 bit.
    Any help would be appreciated. 

    Rakesh, is this an SSRS question?
    This doesn't look like Power View.
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • Getting 401 error while creating a Report Data Source with MOSS 2010 Foundation

    I have setup SQL Server 2008 R2 Reporting Services with SharePoint 2010 Foundation in SharePoint integrated mode. SharePoint Foundation is in machine 1 whereas SQL Server 2008 R2 and SSRS Report Server are in machine 2. While configuring Reporting
    Services - Sharepoint integration, I have used Authentication Mode as "Windows Authentication" (I need to use Kerberos).
    My objective is to setup a Data Connection Library, a Report Model Library, and a Reports Library so that I can upload a Report Data Source, some SMDLs, and a few Reports onto their respective libraries.
    While creating the top level site, "Business Intelligence Center" was not available for template selection since SharePoint Foundation is being used. I therefore selected "Blank Site" as the template.
    While creating a SharePoint Site under the top level site, for template selection I again had to select "Blank Site".
    I then proceeded to create a library for the data connection. Towards this, I created a new document library and selected "Basic page" as the document template. I then went to Library Settings for this newly created library and clicked on
    Advanced Settings. In the Advanced Settings page, for "Allow management of content types?" I selected "Yes". Then I clicked on "Add from existing content types" and selected "Report Data Source". I deleted the existing
    "Document" content type for this library.
    Now I wanted to created a Data Connection in the above Data Connection library. For this when I clicked on "New Document" under "Documents" of "Library Tools" and selected "Report Data Source", I got the error "The
    request failed with HTTP status
    401: Unauthorized.".
    Can anybody tell me why I am getting this error?
    Note: I have created the site and the library using SharePoint Admin account.

    Hi,
    Thank you for your detailed description. According to the description, I noticed that the report server was not part of the
    SharePoint farm. Add the report server to the
    SharePoint farm and see how it works.
    To join a report server to a SharePoint farm, the report server must be installed on a computer that has an instance of a SharePoint product or technology. You can install the report server before or after installing the SharePoint product
    or technology instance.
    More information, see
    http://msdn.microsoft.com/en-us/library/bb283190.aspx
    Thanks.
    Tracy Cai
    TechNet Community Support

  • Error while creating data source in bi for uc connect

    Hi Experts,
    we have using ud connect to pull the data from ms sql server using jdbc,we have maintained all the settings in the j2ee engine and also we performed the smoke test which showed the tables content in the ms sql data base.
    and also we have maintained the rfc connection between the j2ee and sap bi,and performed the test connection which excecuted great.
    but while creating the data source in sap bi administration work bench(rsa1)>sourcesytem>udconnect>create
    we have maintained the following parameters
    rfc destination----
    we have maintained the rfc destination,which we could see in sm59 for the sap bi client
    logical system name----
    we have mentioned a generic logical system name(here we are not sure)
    type of connector----
    jdbc
    name of the connector------ sdk_jdbc
    sourcesystemname----
    it has taken the logical system name description when we enter the logical system name.
    once after maintaining the setting.when we performed the check.
    it showed us the error
    check failed rfc connection:check destination

    Hi,
    Check the gateway settings for
    Gateway Host
    Gateway service
    Also check the UDC connection information.
    Check these links:
    1 - http://help.sap.com/saphelp_nw04/helpdata/en/78/ef1441a509064abee6ffd6f38278fd/content.htm
    2 - http://help.sap.com/saphelp_srm40/helpdata/es/54/ddfb3ff778ed28e10000000a1550b0/content.htm
    3 - https://encrypted.google.com/url?sa=t&source=web&cd=7&ved=0CCwQFjAG&url=http%3A%2F%2Fwww.bi-  
    expertonline.com%2Fdownloads%2FSchroeder0604.doc&rct=j&q=ud%20connection%20in%20sap%20bi&ei=WMDnTdm_E4ruvQOEzZWcDg&usg=AFQjCNGvqKuoTq9kypNMNGhis8WrZ-gFCg&cad=rja
    Let us know in case the issue still persists.
    Debanshu

  • Error while deleting the Data Source

    Hi gurus,
    I am getting an error while deleting a Data Source - "Source system XXXXX  not found RSAR205".
    but that source system no more available.
    Is there any way to delete the Data Source.
    Thanks in advance.
    CK.

    Hi,
    That source system deleted.
    Using RSDS also same error.
    Thanks,
    CK
    Edited by: CK on Oct 3, 2008 11:37 AM

  • Simple creation of Update Rule from BW Data Source

    Hi guys,
       Pertaining standard SAP Business Content extractors
       I am referring to <b>InfoCube : 0PA_C01(Headcount and Personnel Actions)</b>
       I am attempting to<u> create </u>an <b>Update Rule</b> from <b>Info Source : 0HR_PA_PA_1(Headcount)</b>
       This <b>Info Source : 0HR_PA_PA_1(Headcount)</b> is connected to <b>BW Data Source</b>(Not R/3!) 0HR_PA_PA_1
       I have created an Info Package for this Info Source and managed to get 15 records{In Contrast to my 68800 Records from Info Source : 0HR_PA_0(Employee)}
       So, when I create an Update Rule to Connect <b>Info Cube: 0PA_C01(Headcount and Personnel Actions)</b> to <b>Info Source to Info Source : 0HR_PA_PA_1(Headcount)</b>, I get the following error
    ERROR : <b>IC=0PA_C01 IS=0HR_PA_PA_1 error when checking the update rules</b>
      Could you please also advice, why do I only get 15 records for Data Source 0HR_PA_PA_1 ?
      P/S : I am on BW 3.5

    Hey Rohini,
       This <b>Data Source: 0HR_PA_PA_1(Headcount)</b> is tricky to me because it`s a BW Data Source.
       Exact Error Message is as follows :-
      "<b>Error Message : RSAU461
        IC=0PA_C01 IS=0HR_PA_PA_1 error when checking the update rules</b> "
       My Exact Problem is that I don`t see any values for the following fields in my Info Cube : 0PA_C01(Headcount and Personnel Actions)
      Country;
      Country Code;
      Gender;
      Nationality;
      Language;
      Postal Code;
      Region;
      Position;
      Job;
      Payroll Area;
      Payroll Group;
      Pay Scale's;
      Pay Grade's
      This is because, this information is supplied by InfoSource : 0HR_PA_PA_1
      But I don`t have an Update Rule for this InfoSource in my InfoCube : 0PA_C01
      So, that's why I am attempting to create this additional Update Rule
    <i>  And also, could someone enlighten me why would SAP not include such a standard Update Rule when they have already idenfied those needed fields in a Cube ? This is suppose to be a STANDARD workable Business Content right ?</i>
      P/S: I have applied Note : 336229

  • Creation of data sources by Ant

    Hi all
    Please help me..
    when i create data source by ant(copying data-sources.xml already stored in my hard disk with all valid entries.)and then deploy my EAR file and get the following error:
    10g-deploy:
    [exec] ADMN-300075
    [exec] Nested exception
    [exec] Resolution:
    [exec] Base Exception:
    [exec] javax.naming.NamingException
    [exec] Lookup error: javax.naming.AuthenticationException: RMIConnection Disconnected; nested exception is:
    [exec] javax.naming.AuthenticationException: RMIConnection Disconnected
    [exec] javax.naming.NamingException: Lookup error: javax.naming.AuthenticationException: RMIConnection Disconnected; nested exception is:
    [exec] javax.naming.AuthenticationException: RMIConnection Disconnected [Root exception is javax.naming.AuthenticationException: RMIConnection Disconnected]
    [exec] at com.evermind.server.rmi.RMIContext.lookup(RMIContext.java:168)
    [exec] at oracle.ias.sysmgmt.utility.Oc4jUtil.getOC4JApplicationServerAdministrator(Unknown Source)
    [exec] at oracle.ias.sysmgmt.deployment.j2ee.runtime.RuntimeIf.getOC4JApplicationServerAdministrator(Unknown Source)
    [exec] at oracle.ias.sysmgmt.deployment.j2ee.runtime.LocalDeploy.deployOnSingleInstance(Unknown Source)
    [exec] at oracle.ias.sysmgmt.deployment.j2ee.runtime.LocalDeploy.doExecute(Unknown Source)
    [exec] at oracle.ias.sysmgmt.deployment.j2ee.runtime.RuntimeIf.execute(Unknown Source)
    [exec] at oracle.ias.sysmgmt.deployment.j2ee.adapter.DeploymentAdapter.doEvaluateDeploy(Unknown Source)
    [exec] at oracle.ias.sysmgmt.deployment.j2ee.adapter.DeploymentAdapter.evaluate(Unknown Source)
    [exec] at oracle.ias.sysmgmt.task.TaskMaster.sync_evaluate(Unknown Source)
    [exec] at oracle.ias.sysmgmt.task.TaskMaster.internal_evaluate(Unknown Source)
    [exec] at oracle.ias.sysmgmt.task.RemoteEvaluate.execCommand(Unknown Source)
    [exec] at oracle.ias.sysmgmt.task.DaemonWorker.run(Unknown Source)
    [exec] Caused by: javax.naming.AuthenticationException: RMIConnection Disconnected
    [exec] at com.evermind.server.rmi.RMIConnection.connect(RMIConnection.java:2523)
    [exec] at com.evermind.server.rmi.RMIConnection.connect(RMIConnection.java:2339)
    [exec] at com.evermind.server.rmi.RMIConnection.lookup(RMIConnection.java:1781)
    [exec] at com.evermind.server.rmi.RMIServer.lookup(RMIServer.java:721)
    [exec] at com.evermind.server.rmi.RMIContext.lookup(RMIContext.java:149)
    [exec] ... 11 more
    BUILD SUCCESSFUL
    Total time: 4 minutes 58 seconds
    But when i crete data source using Oracle enterprise manager,I m able to deploy my EAR file successfully.
    Can nebody please tell me how can i create data source manually by ant script ant then deploy my appln. successfully
    The entries done by me in data-sources.xml are:
    <data-source class="com.evermind.sql.orionCMTDataSource" location="jdbc/los" connection-driver="oracle.jdbc.driver.OracleDriver" username="dbprod_test" password="dbprod_test" url="jdbc:oracle:thin:@//10.0.52.160:1521/orcl" name="jdbc/los"><description>LOSDB</description></data-source
    Thanks in advance..

    user534925,
    Do you want to create a global datasource or one for a specific J2EE application?
    For a global datasource, you can use the "admin.jar" file.
    Type the following command for more details:
    java -jar admin.jar -helpOne of the options is "-installDataSource".
    You should be able to launch "admin.jar" from an "ant" script (I think -- I don't use "ant").
    Sorry, I don't remember how to set up a datasource for a specific application.
    (And I'm too lazy to search for it for you, now :-)
    Good Luck,
    Avi.

Maybe you are looking for

  • How do i have one account on two computers

    I am fixing to leave for college and my phone and ipod is synched onto my family desktop. I am wanting to change that too my laptop but still have all of my music using my mom's itunes account. PLEASE HELP!

  • How to change the Application flow in ISA 7.0

    Hi All, I am new to ISA 7.0 and working on B2B application. We have made changes in the branding of custom application and now we want some custom functionality to be included in the application flow of our B2B application. Suppose, I have p1,p2,p3,p

  • How to implement request/response domain in JMS

    hi friends, I need help regarding implementing request/response domain in jms.please help me.

  • Cleaning my mac book.

    When browsing I am getting windows coming up I have not searched for including mac keeper and thesmartsearch.net. How do I clean it up? My Imac is fine at work just at home. My daughter and wife have been using it but think it might just be no real f

  • Smart Albums

    I'm having a problem with smart albums...When i try to create a smart album the Keyword option is grayed out, meaning i can't create a new smart album based on the keywords! Anyone know what's the problem?