FIM - Data is not getting extracted from BPC

Hello,
I have created FIM job to load data from BPC NW to PCM. The FIM job is getting executed successfully. But there is no data loaded into PCM. I can see data in BPC.
I have checked data services job log. It also shows successful completion. But there is no record in the target table of extraction dataflow.
Can anyone please help in finding the issue and resolving.
FIM version is 10.
Thank a lot in advance!
Regards,
Rajesh

Hi,
An error log is found at BPC side corresponding to the FIM job execution. Log is given below.
Can anyone help me to understand the issue here and how to resolve?
C Tue Feb 17 16:13:47 2015
C  *** ERROR =>   execute() of C_0792, internal error : no result set
[dbsdbsql.cpp 1579]
C  {root-id=001B21585E281EE4ADD4D9C1E11853C4}_{conn-id=001B21585E281EE4ADD4D9C1E11873C4}_1
C  *** ERROR => EXECUTE C_0792 on connection 0, rc=-9300
[dbslsdb.cpp  9682]
C  SQLCODE    : -9300
C  SQLERRTEXT : Internal error : no result set
C  sc_p=7f0b4afe8150,no=792,idc_p=(nil),con=0,act=1,slen=171084,smax=171264,#vars=0,stmt=7f06a0ddd3b0,table=
C  MDX SELECT
C    NON EMPTY {
many lines with actual BPC data
C  NTH1].&[TRGT_20_22_PPMD] } } ON 0
C  FROM [eSight02_CORPORATE] CELL PROPERTIES VALUE  ;
B  ***LOG BY2=> sql error -9300 performing OPC [dbds         437]
B  ***LOG BY0=> Internal error : no result set [dbds         437]
Thanks and Regards,
Rajesh

Similar Messages

  • Validity end date is not getting copied from inquiry to quotation.

    Hi Friends,
    While creating quotation the validity end date is not getting copied from inquiry to quotation.
    What setting needs to be done, pl suggest.
    Regards,
    Hari.

    Sreehari,
    i dont find any logic of having Same validity date for Quotation and Inquiry....
    the simple reason is explained by example: Say you have created Inquiry on 1st november and end date is 15th novembr. Customer comes to you on 14th november and asked for Quotation, and if you refer quotation on that day that means its only applicable for one day, which generally wont be the case.
    There is no option for copying the validity dates from Inquiry to Quotation as per standard. If you want, create routine, place it in copy control.
    Thanks,
    raja

  • Data not getting transffered from SRM system to BW

    Hi All,
    We have an issue where in the data is not getting transferred from SRM to BW. If i see in RSA 3 of SRM system, the datasource is pulling around 3000 records and in WE02, the idoc is generated and passed without any errors.
    Whereas in BW the idoc is coming with code 8 and no data is getting transferred to BW.
    What can be the probable reasons for this mismatch ??
    Thanks
    Abhi

    Hi Abhi,
    After an upgrade, i would go for following checks
    1. Check the Data Source on CRM side and check if it is Active.
    2. With Upgrade, usually DataSources become inactive and also, pls check if any required fields are hidden in the data source before you activate.
    3. After unhiding required fields, i would activate the Data Source and Replicate the Data Source on BI Side.
    4. Now I would delete the failed request by forcing its status to RED in RSMO and then go for the load.
    Let me know if it worked.

  • Baseline date not getting populated from SRM invoices posted to ECC 6.0

    Hi,
    We are currently in ECC 6.0. The SRM system is in 4.0. We have this particular problem that the baseline date is not getting populated from the Invoices posted from the SRM. The baseline date is blank and the document is saved.
    Did anyone face the same problem??? Can you please refer SAP note 541934 and expand........
    Thanks
    Aravind
    Edited by: Aitipamula Aravind on Aug 10, 2009 5:13 AM

    Hello,
    Yes, you could check the attached note 541934 and regard the implementation of the field ZFBDT in the invoicing document.
    Note 541934 provides a solution how to achieve the MM-customizing for payment baseline dates getting effective for invoices with SRM-origin. This note needs to be applied at your ERP-MM-side in the former plug-in-layer. As described in the note, this is a modification. It's recommended to apply in non-productive system first.
    Hope the above infor.are helps for you!
    Best Regards,
    Gladys xing

  • Billing Plan Data not getting copied from Contract to SO

    Hi ,
    I have a Business requirement in which i ham supposed to create a Contract and then release/sales orders with reference to this contract. Whle creating the SO, the header billing plan dates are not getting copied from COntract to Sales Order.
    Please guide me on how can this be done.
    Regards,
    Madhav

    Hi,
    plz check your copy controls VTAA from contract to release order ,under the contract item category there is Field-FPLA check what is the routine that is assigned.
    you can either use routine-401 or 402

  • Characterestics is not getting displayed from ODS in Multiprovider

    I have Join ODS and remote cube in Multiprovider.And data is not getting displayed from ODS fields neither char not KF getting displyed in Multiprovider.I have checked the identification in Char that Char is presend in both the Cube and ODS,But that Char has data in ODS and not in Cube so I checked Char from ODS in Identification.I am still not getting data from ODS.
    Can anyone help me in it?

    How many requests do u have in the ODS?
    If u have only 1 then just right click on DSO and say delete data it will be faster than deleting by request wise....
    What is the message u r getting when u r trying to delete the request?
    Khaja

  • Delta  reocrds are not getting extracted . Problem with delta Quaue

    Hello friends .
    Could you please help me in this scenario ?
    It’s related to delta load from application 12 ,  - 2LIS_12_VCITM.
    Some how data is not getting transferred from LBWQ to RSA7 . There are more than 200000 entries in LBWQ , but the job is not able to transfer any entries to RSA7.
    For rest other application , data is going correctly .
    The job status showing is correct , like XXXXX LUW are transferred … there is no error in Job log , it technically correct .
    I have found one notes , in SAP service market place , but still it doesn’t match with my job status . My job status is NOTESID  ..where as per the notes it should WKEDPI*** ( something like this ) .
    I have taken below steps to solve this problem but still it doesn’t work .
    1.     Removed job from schedule ( the job for LBWQ to RSA7 ) and Executed the job manually .
    2.     Executed Program RMSBW12.
    3.     Deleted the Delta Q from RSA for application 12 and regenerated the same once again by running the infopackage , early delta initialization with out data transfer , . Once Delta Q is generated , run the job once again .
    I hope , I tried with all correct way ..but still I could not able to transfer the data .
    But with all above steps , still data is not transferred . Daily infopackage is running as per the schedule , and bringing only 0 record . If it still doesn’t work then I have to do Re-initilization .  and this is big cost for us . as we have to lock the sys for users and so on .
    Could you please give me some tips , that how can I transferr the data from LBWQ to delta Q ( RSA7) . If it works then definitely , I will save my lot of time .
    Please suggest me , how can I proceed . Please save me from this situation.
    Many many thanks in advance.
    Regards,

    Hi Akshay,
    Go to SMQ1 and check MCEX12 Queue. If your delta records are piling up in SMQ1 and not being transfered to RSA7, you need to check your back gound job control option once again.
    The back ground collector job will collect the recrords from SMQ1 and push to RSA7.
    Try to execute RMBWV312 program in SE38 and manully push all the recrods to RSA7. Once all the records are available in RSA7 you will find the MCEX12 queue will be cleared in SMQ1.
    Cheers
    Praveen

  • Data is not getting displayed in the report from an Infoset.

    Hi All,
    I am having a report  based on an infoset. This report is displaying the data in the Dev. envmt. When it is transported to the QA, it is not displaying the data in the BEx as well as RSRT, in the QA envmt. The patch levels of both the Dev. and QA are the same. The Queries are same in the Dev and QA also.
    While trying to display the data from the infoset (rt.click- display data), i am able to view the data, in the QA.
    Could anyone please suggest why the data is not getting displayed in the query designer.
    Thanks & Regards,
    A.V.N.Rao

    Hi Ashish,
    I ran the "ZPS/!ZPS" in RSRT where ZPS is the infoset name. In Dev, it displayed the values. In QA, it displayed the below messages:
    ECharacteristic 0TCAKYFNM does not exist. Check authorizations
    WThere are calculated elements. These results are bracketed [  ]
    and below that, it displayed the values for Number of records. But, it has not displayed the values for the other figures.
    Does this has any impact in QA.
    Thanks & Regards,
    AVN Rao.

  • Data not getting fetched from Quotation to Contract

    Hi,Myself new to TM. I have a ticket mentioning data not getting fetched from quotation to contract. How to solve it? Please help...
    Message was edited by: Michael Appleby

    Please add the version of the TM product and which SPs have been installed.  Also more information on the quotation and contract.
    What do you mean by ticket?  It you mean an error message, from where are you reading the message?
    Regards, Mike
    SAP Customer Experience Group - CEG

  • Profit Center not getting copied from WBS master data to settlement rule

    Hello,
    I am facing issues during settlement to G/L. From FI side we have set up some validations that the  FI postings shld always carry a Profit Ctr. Now when I am doing settlement to B/S account, the profit ctr from WBS master data gets copied to settlement details and hence on FI posting (Credit line).
    However when I am trying to settle to Expenxse/P&L account this is not happening. The profit Ctr is not getting copied from WBS master data to Settlement details and hence the settlement is not going thru. The profit derived is a dummy profit ctr and not the one on WBS master data.
    I have posted this query on FI forum too but no satisfactory answers.
    Pls let me know, what is the system behaviour....will the prof ctr on WBS master data gets copied on settlement rule details automatically ?
    Sarang

    Which exit do I use in OPSN? any thts or some exit in FI?
    I am on ECC-6.0...
    One thing to mention, the prof ctr is getting copied to settlement rules for Balance sheet account but not for P&L.This is more confusing...
    Sarang

  • Data is not getting posted to R/3 from CRM

    Hi to All,
    When i am creating sales order in CRM data is not getting replicated to R/3. System is showing that B Doc is created , posted and validated. But here in CRM system outbound queue is generated but in R/3 there is no data in inbound queue.
    Can anybody tell me the reason behind this?
    Regards,
    Anurag

    Hi,
    Could you please check the if the Site-R/3 in SMOEAC has the subscription All Business Transactions (MESG)? If not please create a subscription to the R/3 site.
    If this doesn't work please update the post.
    Hope this helps.
    Thanks,
    Karuna.

  • Data is not Getting refresh in Dashboard

    Hi I have implemented Xcelsius on the top of SAP BI System.
    I have the following architecture -
    SAP BI Cube --> BO XI 3.1 Universe Based on Cube --> QWaaS on the Universe --> Xcelsius Dashboard on the QWaas.
    When i created a dashboard and run it , it worked fine and showed me data , which is lying in Cube.
    But When the Data get updated in Cube and i again refresh the Dashboard , It still showing me the old data , the Data is not getting refresh in Dashboard .
    what could be the possible casue ?
    Is QWaas always fetch the dat at run time ?
    Thanks

    QaaWS web service call needs to be invoked to get the data updates from the datasource. Xcelsius provides several options to command how frequently web services connections are calle/refreshed ('refresh on load', 'refresh every X seconds' or trigger cells...).
    You should also keep in mind, that, since QaaWS hits datasource each time web service is called, this can bring some performance issues (web service calls WebI server to run corresponding query, which performs a refresh). In order to mitigate this, since XI 3.0, QaaWS features a cache mechanism, that is used to store data results from each query refresh (corresponding to a web service call), in order to serve it again with better efficiency, in case same request is performed during a specific time interval (which typically corresponds to dashboard interactions).
    Cache sessions are sorted according to the user names and prompt values used, cache is emptied after a defauilt duration without any interaction (request from the same session).
    Cache timeout duration is set for each QaaWS query, and can be tuned from QaaWS Designer when modifying/creating the query : go to Advanced... button on first step of query edition wizard, cache lifetime corresponds to timeout value (in seconds) displayed on Advanced parameter panel (default value being set to 60 seconds).
    Cache lifetime may also be an explanation why you do not see data refreshed (if you are with QaaWS XI 3.0 or later).
    Hope that helps,
    David.

  • Data is not getting replicating to the destination db.

    I has set up streams replication on 2 databases running Oracle 10.1.0.2 on windows.
    Steps for setting up one-way replication between two ORACLE databases using streams at schema level followed by the metalink doc
    I entered a few few records in the source db, and the data is not getting replication to the destination db. Could you please guide me as to how do i analyse this problem to reach to the solution
    setps for configuration _ steps followed by metalink doc.
    ==================
    Set up ARCHIVELOG mode.
    Set up the Streams administrator.
    Set initialization parameters.
    Create a database link.
    Set up source and destination queues.
    Set up supplemental logging at the source database.
    Configure the capture process at the source database.
    Configure the propagation process.
    Create the destination table.
    Grant object privileges.
    Set the instantiation system change number (SCN).
    Configure the apply process at the destination database.
    Start the capture and apply processes.
    Section 2 : Create user and grant privileges on both Source and Target
    2.1 Create Streams Administrator :
    connect SYS/password as SYSDBA
    create user STRMADMIN identified by STRMADMIN;
    2.2 Grant the necessary privileges to the Streams Administrator :
    GRANT CONNECT, RESOURCE, AQ_ADMINISTRATOR_ROLE,DBA to STRMADMIN;
    In 10g :
    GRANT CONNECT, RESOURCE, AQ_ADMINISTRATOR_ROLE,DBA to STRMADMIN;
    execute DBMS_STREAMS_AUTH.GRANT_ADMIN_PRIVILEGE('STRMADMIN');
    2.3 Create streams queue :
    connect STRMADMIN/STRMADMIN
    BEGIN
    DBMS_STREAMS_ADM.SET_UP_QUEUE(
    queue_table => 'STREAMS_QUEUE_TABLE',
    queue_name => 'STREAMS_QUEUE',
    queue_user => 'STRMADMIN');
    END;
    Section 3 : Steps to be carried out at the Destination Database PLUTO
    3.1 Add apply rules for the Schema at the destination database :
    BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_RULES(
    schema_name => 'SCOTT',
    streams_type => 'APPLY ',
    streams_name => 'STRMADMIN_APPLY',
    queue_name => 'STRMADMIN.STREAMS_QUEUE',
    include_dml => true,
    include_ddl => true,
    source_database => 'REP2');
    END;
    3.2 Specify an 'APPLY USER' at the destination database:
    This is the user who would apply all DML statements and DDL statements.
    The user specified in the APPLY_USER parameter must have the necessary
    privileges to perform DML and DDL changes on the apply objects.
    BEGIN
    DBMS_APPLY_ADM.ALTER_APPLY(
    apply_name => 'STRMADMIN_APPLY',
    apply_user => 'SCOTT');
    END;
    3.3 Start the Apply process :
    DECLARE
    v_started number;
    BEGIN
    SELECT decode(status, 'ENABLED', 1, 0) INTO v_started
    FROM DBA_APPLY WHERE APPLY_NAME = 'STRMADMIN_APPLY';
    if (v_started = 0) then
    DBMS_APPLY_ADM.START_APPLY(apply_name => 'STRMADMIN_APPLY');
    end if;
    END;
    Section 4 :Steps to be carried out at the Source Database REP2
    4.1 Move LogMiner tables from SYSTEM tablespace:
    By default, all LogMiner tables are created in the SYSTEM tablespace.
    It is a good practice to create an alternate tablespace for the LogMiner
    tables.
    CREATE TABLESPACE LOGMNRTS DATAFILE 'logmnrts.dbf' SIZE 25M AUTOEXTEND ON
    MAXSIZE UNLIMITED;
    BEGIN
    DBMS_LOGMNR_D.SET_TABLESPACE('LOGMNRTS');
    END;
    4.2 Turn on supplemental logging for DEPT and EMPLOYEES table :
    connect SYS/password as SYSDBA
    ALTER TABLE scott.dept ADD SUPPLEMENTAL LOG GROUP dept_pk(deptno) ALWAYS;
    ALTER TABLE scott.EMPLOYEES ADD SUPPLEMENTAL LOG GROUP dep_pk(empno) ALWAYS;
    Note: If the number of tables are more the supplemental logging can be
    set at database level .
    4.3 Create a database link to the destination database :
    connect STRMADMIN/STRMADMIN
    CREATE DATABASE LINK PLUTO connect to
    STRMADMIN identified by STRMADMIN using 'PLUTO';
    Test the database link to be working properly by querying against the
    destination database.
    Eg : select * from global_name@PLUTO;
    4.4 Add capture rules for the schema SCOTT at the source database:
    BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_RULES(
    schema_name => 'SCOTT',
    streams_type => 'CAPTURE',
    streams_name => 'STREAM_CAPTURE',
    queue_name => 'STRMADMIN.STREAMS_QUEUE',
    include_dml => true,
    include_ddl => true,
    source_database => 'REP2');
    END;
    4.5 Add propagation rules for the schema SCOTT at the source database.
    This step will also create a propagation job to the destination database.
    BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_PROPAGATION_RULES(
    schema_name => 'SCOTT',
    streams_name => 'STREAM_PROPAGATE',
    source_queue_name => 'STRMADMIN.STREAMS_QUEUE',
    destination_queue_name => 'STRMADMIN.STREAMS_QUEUE@PLUTO',
    include_dml => true,
    include_ddl => true,
    source_database => 'REP2');
    END;
    Section 5 : Export, import and instantiation of tables from
    Source to Destination Database
    5.1 If the objects are not present in the destination database, perform
    an export of the objects from the source database and import them
    into the destination database
    Export from the Source Database:
    Specify the OBJECT_CONSISTENT=Y clause on the export command.
    By doing this, an export is performed that is consistent for each
    individual object at a particular system change number (SCN).
    exp USERID=SYSTEM/manager@rep2 OWNER=SCOTT FILE=scott.dmp
    LOG=exportTables.log OBJECT_CONSISTENT=Y STATISTICS = NONE
    Import into the Destination Database:
    Specify STREAMS_INSTANTIATION=Y clause in the import command.
    By doing this, the streams metadata is updated with the appropriate
    information in the destination database corresponding to the SCN that
    is recorded in the export file.
    imp USERID=SYSTEM@pluto FULL=Y CONSTRAINTS=Y FILE=scott.dmp IGNORE=Y
    COMMIT=Y LOG=importTables.log STREAMS_INSTANTIATION=Y
    5.2 If the objects are already present in the desination database, there
    are two ways of instanitating the objects at the destination site.
    1. By means of Metadata-only export/import :
    Specify ROWS=N during Export
    Specify IGNORE=Y during Import along with above import parameters.
    2. By Manaually instantiating the objects
    Get the Instantiation SCN at the source database:
    connect STRMADMIN/STRMADMIN@source
    set serveroutput on
    DECLARE
    iscn NUMBER; -- Variable to hold instantiation SCN value
    BEGIN
    iscn := DBMS_FLASHBACK.GET_SYSTEM_CHANGE_NUMBER();
    DBMS_OUTPUT.PUT_LINE ('Instantiation SCN is: ' || iscn);
    END;
    Instantiate the objects at the destination database with
    this SCN value. The SET_TABLE_INSTANTIATION_SCN procedure
    controls which LCRs for a table are to be applied by the
    apply process. If the commit SCN of an LCR from the source
    database is less than or equal to this instantiation SCN,
    then the apply process discards the LCR. Else, the apply
    process applies the LCR.
    connect STRMADMIN/STRMADMIN@destination
    BEGIN
    DBMS_APPLY_ADM.SET_SCHEMA_INSTANTIATION_SCN(
    SOURCE_SCHEMA_NAME => 'SCOTT',
    source_database_name => 'REP2',
    instantiation_scn => &iscn );
    END;
    Enter value for iscn:
    <Provide the value of SCN that you got from the source database>
    Note:In 9i, you must instantiate each table individually.
    In 10g recursive=true parameter of DBMS_APPLY_ADM.SET_SCHEMA_INSTANTIATION_SCN
    is used for instantiation...
    Section 6 : Start the Capture process
    begin
    DBMS_CAPTURE_ADM.START_CAPTURE(capture_name => 'STREAM_CAPTURE');
    end;
    /

    same problem, data not replicated.
    its captured,propagated from source,but not applied.
    also no apply errors in DBA_APPLY_ERROR. Looks like the problem is that LCRs propagated from source db do not reach target queue.can i get any help on this?
    queried results are as under:
    1.at source(capture process)
    Capture Session Total
    Process Session Serial Redo Entries LCRs
    Number ID Number State Scanned Enqueued
    CP01 16 7 CAPTURING CHANGES 1010143 72
    2. data propagated from source
    Total Time Executing
    in Seconds Total Events Propagated Total Bytes Propagated
    7 13 6731
    3. Apply at target(nothing is applied)
    Coordinator Session Total Total Total
    Process Session Serial Trans Trans Apply
    Name ID Number State Received Applied Errors
    A001 154 33 APPLYING 0 0 0
    4. At target:(nothing in buffer)
    Total Captured LCRs
    Queue Owner Queue Name LCRs in Memory Spilled LCRs in Buffered Queue
    STRMADMIN STREAMS_QUEUE 0 0 0

  • Data is not getting updated in DB table

    hi all
    i am doing IDOC to jdbc scenario
    i am triggering idoc from R/3 and the data is going into DB table
    sender side:       ZVendorIdoc
    receiver side:
    DT_testVendor
      Table
        tblVendor
          action       UPDATE_INSERT
          access      1:unbounded
            cVendorName 1
            cVendorCode 1
         fromdate    1
         todate      1
          Key
            cVendorName  1
    if i trigger idoc for example vendor 2005,2006 and 2010 data is getting updated in the table
    but again if i trigger idoc for same vendor nos data does not get updated in DB table while message is successfull in moni and RWB both
    plz suggest if any change need to be done to update the data
    Regards
    sandeep sharma

    Hi Ravi
    you are right, vendor no is my key field . problem is when i send data again then it should Update the data but it's not updating the data again
    i did on exp with this : i deleted all the record from the table and then  triggered idoc for vendor 2005 , 2006,2010 after this data is updated in the table i deleted the rows for vendor no 2006 and 2010 and kept the row for vendor 2005
    then i again trigered the idoc for vendor no 2005,2006 and 2010 now this should update and it should insert rows for vendor no 2006 and 2010 but i am surprised its not updating the data
    Thanks
    sandeep

  • Data is not getting posted in ABAP Proxy.

    Hi,
    I am working on File to ABAPProxy scenario. The data is not going to proxy.
    In PI sxmb_moni is status is successfull and in the R3 sxmb_moni the status is successfull. But the data is not getting posted in the tables(ABAPProxy). I had checked the inbound and oubound queues. And with the same input data, abapers cheked their code in Abapproxy, then the table is getting updated. When we trigger from PI, the tables are not getting updated.Please help me.
    Thanks,
    Pragathi.

    Hi,
    The problem may be
    Case 1: Cache is not getting refreshed(Check SXI_CACHE)
    Case 2: The Queue Is blocked (Check SMQ1 & SMQ2)
    Regards,
    Sainath

Maybe you are looking for

  • ABAP Report Text ID FKK0 language EN not found

    I'm having a problem with my program. I added the following code and Iu2019m getting the below error from u2018FKK0u2019 can anyone help with my problem? *& Form DISPLAY text --> p1 text <-- p2 text form DISPLAY . CALL FUNCTION 'READ_TEXT' EXPORTING

  • Financial Reporting batch runs without being scheduled when services restart

    Hi, hopefully someone may have come across this before: I set up a scheduled batch yesterday (workspace 11.1.1.1.3)  pointing at a batch of 2 large reports. As part of the testing I ran it at a scheduled time during the day and it worked perfectly. H

  • Developing a solution for a client with SAP

    Hi there, We're looking to work with a client who is running an SAP based system, and in advance of our next meeting in a month we have pledged to research the SAP platforms and how to interact with them. The client will be upgrading to SAP 6 in a co

  • Server backup connections

    Hi, We have a setup involving servers connected to the core switch. there is an ASA acting as a designated dmz zone for subsidaries to access the network. A dedicated backup vlan is required. The criteria being kept is this backup vlan will not talk

  • HT1349 husband gave me his old ipad (business); how do I make it mine?

    How do I change the apple ID to my name?  The ipad 1 was my husband's business one.  He got a new iPad II.