No data from datasource 0CFM_INIT_POSITIONS

Hello everyone,
Issue 1)
I have a bit of an issue extracting data via the dataSource 0CFM_INIT_POSITIONS. To test why no date is being transfered in my infopackage, I have used RSA3 to check the extraction. When running the check with update mode " I " (or "Transfer of an opening balance for non-cumulative values") and the selection values TR_ACC_CODE "1000 , VALUATION_AREA "002" I get the message that 0 data objects were selected. I have looked through the positions in transaction TPM26 in the source system, and there is data for this selection.
What could be the reason that no data is selected in the extraction checker?
(The extraction is based on the logical database FTI_TR_PERIODS.)
Issue 2)
Also: once the extraction check has been run with the same selections as above,  and I then subsequently try to re-run the check I get the error message:
"No initialization with InfoSource 0CFM_INIT_POSITIONS and selection TR_ACC_CODE 1000/VALUATION_AREA 002".
,and that:
"The position initialization (InfoSource 0CFM_INIT_POSITIONS) has already been run with the selection TR_ACC_CODE 1000/VALUATION_AREA 002 ."
It is suggested that I "reset the position initialization for the selection TR_ACC_CODE 1000/VALUATION_AREA 002."
But how do I do that?
Your help is much appreciated.
Thank you!
Roger

True,
That is exactly what I am trying to figure out. The datasource is based on a logical database. So there is no clear table from which it is extracting. What are the steps to identify the tables, without debugging the logical database code.
Also, would anyone happen to now how to reset the extractions checker once it hs been run with specific selection values?
much appreciated
Roger

Similar Messages

  • Error while loading data from DataSource to InfoCube/DSO through DTP

    Hi Friends,
    I am trying to load data from DataSource to InfoCube/DSO through DTP and getting the following error:
    Exception in substep: Extraction completed
    Processing terminated
    Data package 1 / 30.04.2009 22:38:30 / Status 'Processed with Errors'
    When i see the detail message it says:
    Syntax error in routine "convert_to_applstru", line 25 of generated program"
    I checked all my transformation and looks ok. I was able to load data earlier and i getting this error all of sudden. Does some one know what is this routine "convert_to_applstru" ?
    Thanks,
    Amit

    Hi Arun,
    Where do i see this generated program in RSRV. I dont see this in RSRV. Please guide me.
    Thanks,
    Amit

  • Trying to extract data from Datasource, getting an S:RSSDK:300 error

    "While pre-viewing data using UDC DataSource - we get ""S:RSSDK:300 Query execution failed: ERROR:  Relation 'ZV1' doe s not exist  Message no. RSSDK300"" error "
    Any help would be greatly appreciated

    Hi Raj,
    The issue we were facing was an Netezza issue. When import table into the Netezza Data Warehouse we import table into the wrong structure. We simple moved the table over to the right structure within the Netezza Data Warehouse and we were able to preview data from Datasource.

  • Extract data from datasource 0APO_MSP_CHECK_ORDER error

    Hi,expert. Happy new year. I met a problem today,when I use TCODE RSA3 to extract data of datasource ,0APO_MSP_CHECK_ORDER,I got an error.
    1. Follow-up error
        Follow-up error
        Message no. /SAPAPO/OM_ERROR184
    2. om_other_failed
    ( om_other_failed )
    Message no. /SAPAPO/OM012
    So, I can't extract data from this datasource.
    The datasource 0APO_MSP_SLOT had the same error too.
    PLZ help me. Thank you very much.
    Happy new year.

    It has been solved.

  • FM to extract data from Datasource

    Hi,
    Can any one of you suggest me an idea or let me know where to start and how to start ?
    I want to write a ABAP code/ create a Function Modules which extracts data from a DATASOURCE be it any on condition(say date) and the output given should be filtered on some fields
    Does a standard datasource be extracted using a custom FM?
    Are there any standard FM to do this?

    Hi,
    go to the TX- SE80 and create the function group.
    after take the FM - RSAX_BIW_GET_DATA_SIMPLE
    please look at the below doc it will help.
    Generic Extraction based on Function Module
    Thanks,
    Phani.

  • Error while extracting data from datasource 0GLACCEXT_T011_HIER in RSA3

    Hi Experts,
    While trying to extract the data for the datasource 0GLACCEXT_T011_HIER in RSA3, am getting the below Dump.
    Runtime Errors         SAPSQL_ARRAY_INSERT_DUPREC
    Except.                CX_SY_OPEN_SQL_DB
    Short text
        The ABAP/4 Open SQL array insert results in duplicate database records.
    What happened?
        Error in the ABAP Application Program
        The current ABAP program "SAPLFAGL_FSV" had to be terminated because it has
        come across a statement that unfortunately cannot be executed.
    Error analysis
        An exception occurred that is explained in detail below.
        The exception, which is assigned to class 'CX_SY_OPEN_SQL_DB', was not caught
         in
        procedure "FSV_TEXTS_CONVERT" "(FORM)", nor was it propagated by a RAISING
         clause.
        Since the caller of the procedure could not have anticipated that the
        exception would occur, the current program is terminated.
        The reason for the exception is:
        If you use an ABAP/4 Open SQL array insert to insert a record in
        the database and that record already exists with the same key,
        this results in a termination.
        (With an ABAP/4 Open SQL single record insert in the same error
        situation, processing does not terminate, but SY-SUBRC is set to 4.)
    Kindly help me out in resolving this error.

    Hi,
    This dump occurs if the storage of original files is not happening in a Content Server, which is recommended by SAP. You can check this easily if you go to transaction DC10 and check if the flag 'Use KPRO' is set for the used document type.
    SAP do not recommend storing such larger files in SAP DB. If you try to store any original in the SAP DB, the file will be stored in the table DRAO. That means you are consuming table space memory. For one original we can have maximum of 9999 rows in the table. To store larger originals in SAP DB more than 9999 rows in the table are needed and this is not possible. That's the reason behind this dump. Storing such huge files inthe SAP DB will also affect the performance of the system while accessing the original
    Refer the thread
    "http://wiki.sdn.sap.com/wiki/display/PLM/SAPSQL_ARRAY_INSERT_DUPRECwhilesaving+original"
    Thanks,

  • DTP load data from DSO to CUBE can't debug

    Hi,all expert.
    I met a problem yestoday.
    When I load data from datasource to DSO,and debug the start routine,it was ok.But when I load data from DSO to CUBE,and debug the routine,it got an error like this:
    inconsistent input parameter(parameter:<unknow>,value <unknown>)
    message no: RS_EXCEPTION 101.
    I don't know why,PLZ help me,thank you very much.

    Hi,
    An alternative way to do this is to have a small code for infinite loop in your start routine. And when you start the load goto SM50 and debug the process which is executing the load, from there you can close the infinite loop and carry on with the debugging. Please let me know if you require further information.
    Thanks,
    Arminder

  • Error reading the data from the remote cube

    Hi all,
    When we try to get the data for remote cube from LISTCUBE, we are getting the following msg.
    1) Messages for DataSource 9AUPA_DP_HK_01 from source system AD1CLNT100
    2) 224(/SAPAPO/TSM): No planning version selected
    3) Errors have occurred while extracting data from DataSource 9AUPA_DP_HK_01
    4) Error reading the data of InfoProvider UICHKRMTC.
    Any Inputs?

    Hi,
    Check whether the sourcesystem is responding.
    And also in the error mesg: 224(/SAPAPO/TSM): No planning version selected
    It seems you have not selected any planning version. Give any planning version in the listcube selection screen and execute.
    Regards,
    Ravi Kanth

  • Aggregating Data from multple sources with ESB

    Hi
    I want to aggregate data from multiple data-sources with an esb service and after that call a bpel-process with a request build from this data.
    1. read data from datasource A (dbadapter-select-call)
    2. read data from data-source B (dbadapter-select-call)
    3. put the data together in xsl-transofrmation
    4. call bpel
    Is this possible? How can I get the data from the first call together with data from the second call for transformation? If I get data from the second call, the data from the first call seems to be lost.
    Any ideas?
    Gregor

    Gregor,
    It seems this aggregation of data is not possible in ESB.This can be done in BPEL only that too using assigns but not using transformations.I have tried using transformations by giving the third argument to the ora:processXSLT function.But couldnot achieve the desired result.
    For more info on passing a second variable(of another schema) as a parameter to xslt pls refer to the post
    http://blogs.oracle.com/rammenon/2007/05/
    and the sectiion "Passing BPEL Variable contents into XSLT as Parameters".
    Hope this helps you.
    Thanks,Venkat.

  • Runtime error while extracting data from 0C0_PC_*

    Hi,
    When I am trying  to extract data from DataSources 0CO_PC_01 and 0CO_PC_02, I am getting an
    error. it says  "DO_WHILE_VARY_NOT_IN_RANGE"
    Any Comments?
    Thanks,
    Pritam

    Hi,
    Please searcn SDN, you will get the solutions straight away:
    https://www.sdn.sap.com/irj/sdn/advancedsearch?query=do_while_vary_not_in_range&cat=sdn_all
    It seems that there is a note for this, Please check SAP Note: 650725
    Hope this solves your concern...
    Cheers,
    Habeeb

  • Generic datasource by function module to fetch data from multiple tables?

    I'm writing a function module to fetch price, for generic datasource.
    At first, extract test is OK. But InfoPackage never stop  when loading data to PSA in BW.
    And I find the example codes:
         OPEN CURSOR WITH HOLD S_CURSOR FOR
          SELECT (S_S_IF-T_FIELDS) FROM SFLIGHT
                                   WHERE CARRID  IN L_R_CARRID AND
                                         CONNID  IN L_R_CONNID.
        ENDIF.                             "First data package ?
    * Fetch records into interface table.
    *   named E_T_'Name of extract structure'.
        FETCH NEXT CURSOR S_CURSOR
                   APPENDING CORRESPONDING FIELDS
                   OF TABLE E_T_DATA
                   PACKAGE SIZE S_S_IF-MAXSIZE.
        IF SY-SUBRC <> 0.
          CLOSE CURSOR S_CURSOR.
          RAISE NO_MORE_DATA.
        ENDIF.
        S_COUNTER_DATAPAKID = S_COUNTER_DATAPAKID + 1.
      ENDIF.
    There using Cursor to fetch data package by package, and raise exception NO_MORE_DATA to stop the loading process.
    Now I fetch data from multiple tables, I don't think I can use Cursor.
    Then How can I handle this?  
    Thanks a lot.

    Thanks
    IF IT_999[] IS INITIAL.
        SELECT A~KNUMH A~MATNR A~KSCHL VKORG VTWEG A~DATBI A~DATAB KBETR KMEIN KPEIN C~MTART APPENDING CORRESPONDING FIELDS OF
          TABLE TP_DATA
            FROM A999 AS A
              INNER JOIN KONP AS B
                  ON A~KNUMH = B~KNUMH
              INNER JOIN MARA AS C
                 ON A~MATNR = C~MATNR
    *          FOR ALL ENTRIES IN IT_999
                    WHERE
    *      A~KNUMH = IT_999-KNUMH  AND
           ( ( A~KSCHL = 'ZPRC' AND VKORG = 'Z000' AND VTWEG = 'Z1' ) OR
                          ( A~KSCHL = 'ZPRD' AND VKORG = 'A000' AND VTWEG = 'Y3' ) ) AND
    *                      A~DATBI >= SY-DATUM AND
                          LOEVM_KO = ''.
        SELECT A~KNUMH A~MATNR A~KSCHL VKORG VTWEG A~DATBI A~DATAB KBETR AS KHETR  KMEIN KPEIN C~MTART APPENDING CORRESPONDING FIELDS OF
          TABLE TP_DATA
            FROM A999 AS A
              INNER JOIN KONP AS B
                  ON A~KNUMH = B~KNUMH
              INNER JOIN MARA AS C
                 ON A~MATNR = C~MATNR
    *          FOR ALL ENTRIES IN IT_999
                    WHERE
    *      A~KNUMH = IT_999-KNUMH AND
          A~KSCHL = 'ZPR3' AND A~VKORG = 'I000' AND
    *                      DATBI >= SY-DATUM AND
                          LOEVM_KO = ''.
      ENDIF.
      IF IT_997[] IS INITIAL.
        SELECT A~KNUMH A~MATNR A~KSCHL VTWEG A~DATBI A~DATAB KBETR AS KHETR KMEIN KPEIN C~MTART APPENDING CORRESPONDING FIELDS OF
          TABLE TP_DATA
            FROM A997 AS A
              INNER JOIN KONP AS B
                  ON A~KNUMH = B~KNUMH
              INNER JOIN MARA AS C
                 ON A~MATNR = C~MATNR
    *          FOR ALL ENTRIES IN IT_997
                    WHERE
    *      A~KNUMH = IT_997-KNUMH      AND
          A~KSCHL = 'ZPRA' AND VTWEG = 'Y1' AND
    *                      DATBI >= SY-DATUM AND
                      LOEVM_KO = ''.
      ENDIF.
      IF IT_996[] IS INITIAL.
        SELECT A~KNUMH A~MATNR A~KSCHL A~DATBI A~DATAB KBETR AS KHETR KMEIN KPEIN C~MTART APPENDING CORRESPONDING FIELDS OF
           TABLE TP_DATA
             FROM A996 AS A
               INNER JOIN KONP AS B
                   ON A~KNUMH = B~KNUMH
               INNER JOIN MARA AS C
                  ON A~MATNR = C~MATNR
    *          FOR ALL ENTRIES IN IT_996
                    WHERE
    *      A~KNUMH = IT_996-KNUMH AND
          A~KSCHL = 'ZPRB' AND
    *                       DATBI >= SY-DATUM AND
          LOEVM_KO = ''.
      ENDIF.
      SELECT   MATNR     "u7269u6599u53F7u7801
               MEINH     "u4ED3u50A8u5355u4F4Du7684u5907u7528u8BA1u91CFu5355u4F4D
               UMREZ     "u57FAu672Cu8BA1u91CFu5355u4F4Du8F6Cu6362u5206u5B50
               UMREN     "u8F6Cu6362u4E3Au57FAu672Cu8BA1u91CFu5355u4F4Du7684u5206u6BCD
          FROM MARM
          INTO CORRESPONDING FIELDS OF TABLE IT_MARM
           FOR ALL ENTRIES IN TP_DATA
         WHERE MATNR = TP_DATA-MATNR AND  MEINH = TP_DATA-KMEIN.
      LOOP AT TP_DATA.
        IF TP_DATA-KPEIN NE 0.
          TP_DATA-KBETR =  TP_DATA-KBETR / TP_DATA-KPEIN.
          TP_DATA-KHETR =  TP_DATA-KHETR / TP_DATA-KPEIN.
        ENDIF.
        IF TP_DATA-KSCHL = 'ZPRA'.
    *       TP_DATA-MEINH = 'ZI'.
    *      TP_DATA-KSCHL = 'B4'.
          IF TP_DATA-KMEIN = 'ZI'.
            TP_DATA-KBETR = TP_DATA-KHETR / '1.17'.
          ELSE.
            READ TABLE IT_MARM INTO WA_MARM1 WITH KEY MATNR = TP_DATA-MATNR MEINH = TP_DATA-KMEIN.
    *           READ TABLE IT_MARM INTO WA_MARM2 WITH KEY MATNR = TP_DATA-MATNR MEINH = 'CT'.
            TP_DATA-KHETR = TP_DATA-KHETR * WA_MARM1-UMREN / WA_MARM1-UMREZ.
    *           * WA_MARM2-UMREZ / WA_MARM2-UMREN.
            TP_DATA-KBETR = TP_DATA-KHETR / '1.17'.
          ENDIF.
        ELSEIF TP_DATA-KSCHL = 'ZPRB'.
    *      TP_DATA-KSCHL = 'L0'.
    *       TP_DATA-MEINH = 'ZI'.
          IF TP_DATA-KMEIN = 'ZI'.
            TP_DATA-KBETR = TP_DATA-KHETR / '1.17'.
          ELSE.
            READ TABLE IT_MARM INTO WA_MARM1 WITH KEY MATNR = TP_DATA-MATNR MEINH = TP_DATA-KMEIN.
    *           READ TABLE IT_MARM INTO WA_MARM2 WITH KEY MATNR = TP_DATA-MATNR MEINH = 'BAG'.
            TP_DATA-KHETR = TP_DATA-KHETR * WA_MARM1-UMREN / WA_MARM1-UMREZ.
    *           * WA_MARM2-UMREZ / WA_MARM2-UMREN.
            TP_DATA-KBETR = TP_DATA-KHETR / '1.17'.
          ENDIF.
        ELSEIF TP_DATA-KSCHL = 'ZPRC' OR TP_DATA-KSCHL = 'ZPRD'.
    *       TP_DATA-MEINH = 'ZI'.
          IF TP_DATA-KMEIN = 'ZI'.
            TP_DATA-KHETR = TP_DATA-KBETR * '1.17'.
          ELSE.
            READ TABLE IT_MARM INTO WA_MARM1 WITH KEY MATNR = TP_DATA-MATNR MEINH = TP_DATA-KMEIN.
    *           READ TABLE IT_MARM INTO WA_MARM2 WITH KEY MATNR = TP_DATA-MATNR MEINH = 'WZI'.
            TP_DATA-KBETR = TP_DATA-KBETR * WA_MARM1-UMREN / WA_MARM1-UMREZ.
    *           * WA_MARM2-UMREZ / WA_MARM2-UMREN.
            TP_DATA-KHETR = TP_DATA-KBETR * '1.17'.
          ENDIF.
        ELSEIF TP_DATA-KSCHL = 'ZPR3'.
    *      TP_DATA-KSCHL = 'B2'.
          IF TP_DATA-KMEIN = 'ZI'.
            TP_DATA-KBETR = TP_DATA-KHETR / '1.17'.
          ELSE.
            READ TABLE IT_MARM INTO WA_MARM1 WITH KEY MATNR = TP_DATA-MATNR MEINH = TP_DATA-KMEIN.
    *           READ TABLE IT_MARM INTO WA_MARM2 WITH KEY MATNR = TP_DATA-MATNR MEINH = 'BAG'.
            TP_DATA-KHETR = TP_DATA-KHETR * WA_MARM1-UMREN / WA_MARM1-UMREZ.
    *           * WA_MARM2-UMREZ / WA_MARM2-UMREN.
            TP_DATA-KBETR = TP_DATA-KHETR / '1.17'.
          ENDIF.
        ENDIF.
        TP_DATA-MEINH = '01'.
        MODIFY TP_DATA.
    E_T_DATA-MATNR =   TP_DATA-MATNR.
    E_T_DATA-KSCHL =   TP_DATA-KSCHL.
    E_T_DATA-KHETR =   TP_DATA-KHETR.
    E_T_DATA-KBETR =   TP_DATA-KBETR.
    E_T_DATA-KMEIN =   TP_DATA-KMEIN.
    E_T_DATA-DATAB =   TP_DATA-DATAB.
    E_T_DATA-DATBI =   TP_DATA-DATBI.
    APPEND E_T_DATA.
        CLEAR WA_MARM1.
        CLEAR WA_MARM2.
      ENDLOOP.
    Edited by: Shen Peng on Oct 20, 2010 10:09 AM

  • Get the Data from a DataSource

    Hi,
    I start with JMF. I use a processor to get DataSource from an audio file. Until there, all runs very well.
    But after that, I would like to get the raw data of my DataSource to aplly a filter on these data and after that, to create a new DataSource with filters data and play these data with my processor.
    I searched in the API documentation of JMF but I did'nt find the solution to get the raw data of my DataSource !
    Someone could help me ? and give me an idea of how to proceed ?
    Thanks for yours answers.
    Sylvain.

    Hi,
    It seems no one has replied to this post. I have a sample code that can help getting raw data from a DataSource. Here it is:
              PushBufferDataSource source = null;
              source = (PushBufferDataSource)processor.getDataOutput();
              /* Now we can retrieve the PushBufferStreams that will enable us to
              * access the data from the camera
              PushBufferStream[] streams = source.getStreams();
              camStream = null;
              for (int i = 0; i < streams.length; i++) {
              /* Use the first Stream that is RGBFormat (there should be only one */
              if (streams.getFormat() instanceof RGBFormat) {
                   camStream = streams[i];
                   RGBFormat rgbf = (RGBFormat)streams[i].getFormat();
                   converter = new BufferToImage(rgbf);
                   break;
    Please let me know whether it helped solving your problem..
    Suresh

  • Error while extracting data from Generic datasource

    Hello Gurus,
    I have encountered an error "Errors in source system" with status red while extracting data from generic datasource into DSO.
    I have done below things:-
    1. Checked the jobs in source system, the job is successfully completed
    2. The generic datasource is active and supporting delta with calender day  in source system.
    3. No TRfc error and no idocs stucked.
    4. No short dumps in source system.
    5. The delta queue(RSA7) is showing 0 with status as green.
    6. Replicated the datasource On BW side and also activated the transfer rules by RS_TRANSU_ACTIVATE_ALL
    The problem has not yet solved. Could you guys please suggest me what might be the reason for this error......
    Thanks,
    Sonu

    Hello,
    I have used the generic delta using calender day.
    Safety upper limit : - 1
    Safety lowerr limit : - 0
    I have made the entries in va01 transaction, but the delta queus is not getting updated which should show 1.
    Do Generic extraction require any job to transfer data from Database table to delta queue.
    We are not able to get the delta records on BW side. Please suggest.
    Please suggest.
    Thanks,
    Sonu

  • Is there any datasource available to pull data from NAST table?

    The field I am trying to pull out of NAST (Message Status) table is NACHA. Please provide me hints if any Business Content Datasource uses this table or we have to create a custom extractor for the same?
    Regards,
    Joy

    I dont think there is a standard business content datasource to extract data from NAST table. If you are going to use this data in some other LO modules based on the requirement enhance the datasource and append fields and write code to populate the message types.
    If you are going to use this as stand alone data, then you can go ahead and create a generic datasource based on table.

  • Migrate tables and data from ODBC datasource to Oracle 11.2

    Hiya
    I have a database file in Dynamics NAV 5.0 native format - about 200-300 tables totalling maybe 10 GB worth of data.
    My task is to migrate those tables with their data into a schema in our Oracle 11.2 database.
    No tools support the NAV format directly :-) But I have a windows ODBC driver for that format that works.
    The Oracle database runs on Linux and I have no ODBC transparent gateway and no linux ODBC driver that would let me access the NAV datafile directly from within Oracle.
    I tried migration wizard in SQL Developer, but it seems it cannot use a "generic ODBC" datasource (even with JDBC-ODBC bridge), it must be a supported database format?
    Then I have tried trial of product Altova MapForce, but it can't seem to read the schema (table/columns) from the ODBC datasource for some reason.
    I know reading the schema from the ODBC datasource is possible - I have used third party product "Database Browser" that reads the ODBC with no problems.
    And a colleague has read the schema of the ODBC in C# also.
    I am primarily a SQL developer. My colleague said he could write in C# a program that looped over the tables in the ODBC, read the columns, executed "create table" in Oracle, read data from the ODBC, inserted data in Oracle.
    But he is going on vacation right now :-( And I keep thinking that someone else must have written such a migration program ;-)
    I just can't seem to find it by googling? Altova Mapforce was my best bet, but it just seems to have problems with this specific ODBC datasource.
    Any ideas for a migration tool that can accept a generic ODBC datasource and not specifically SQL Server, MySQL, DB2, etc.?
    Thanks for any hints that might help ;-)
    Regards
    Kim Berg Hansen

    Hi Kim,
       The earlier 9i and 10g generic connectivity option - HSODBC - could only run as part of the RDBMS but the 11g DG4ODBC can be installed and run standalone without an RDBMS.  You then connect using database links from Linux using an entry in the RDBMS tnsnames.ora as you have said.
    See these notes depending if Windows is 32-bit or 64-bit -
    How to Configure DG4ODBC (Oracle Database Gateway for ODBC) on Windows 32bit to Connect to Non-Oracle Databases Post Install (Doc ID 466225.1)
    How to Configure DG4ODBC (Oracle Database Gateway for ODBC) on 64bit Windows Operating Systems to Connect to Non-Oracle Databases Post Install (Doc ID 1266572.1)
    The latest 11.2 version is available from My Oracle Support as -
    Patch 13390677: 11.2.0.4.0 PATCH SET FOR ORACLE DATABASE SERVER
    and for the gateways you only need to download -
    p13390677_112040_WINNT_4of6.zip
    for 32-bit but
    p13390677_112040_MSWIN-x86-64_5of7.zip
    for 64-bit.
    The downloads have everything necessary for a gateway standalone install.
    Regards,
    Mike

Maybe you are looking for