Reverse Data Flow

Hello Experts,
I have a InfoCube, and i am backtracking it's data all the way to R/3. I checked the data flow to this cube and it showed that it's been loaded from an Infosource ( there's no ODS in between the Infosource and the Infocube) - When i click on Infosource, i can see the Communication Structure ( Which i suppose is the structure in which the data is stored in an Infosource - Correct me if i am wrong anywhere)  and the Transfer Structure ( Which i suppose is the structure in which the data is stored in the Datasource) and where i also can see the Datasource assigned to this Infosource and this datasource would tell me whether this is a delta load or init load.
Coming to the questions ?
How is data loaded into a datasource - I know a datasource is something which is present both at R/3 and BW end. So how is data loaded into a Datasource at R/3 end and how is it loaded into BW. If someone can provide how to do in BW, i mean the T Codes or the procedure - that would help me a lot in understing or getting the bigger picture.
Where does PSA come into picture in this whole process ?
All answers would be really appreciated and duly rewarded with points.
Thanks,
Nandita

Hi Nandita,
Data permanently stored in tables only.
Data source is structure  and it provide the data to other objects at run time
data source can be crated by views, tabels, structure and feeding programme, sap queryetc below transactions are useful for managing data sources
RS02,RSA6,RSA5,RSA3,RSA7,BWA1,BWA7 and RSA2
Here is data flow.
data source(R3) - Transfer structure (data source ) (BW) (PSA) -Transfer rules (BW)Infosource (CS) (BW)-Update rules (BW)---- data target (BW)
BW side tranfer structure is just replication of r/3 side data source to move date from data source through infosource to data target.
Update rules and transfer rules are joining infocube , info source and infosource to data source.
PSA is the  persisting staging area, whatever data is there in r/3 same data available in BW as PSA, it entry point of data to BW.
i hope it will give some idea.
Thnaks
Thota

Similar Messages

  • AR DATA FLOW (TABLE LEVEL) - ON ACCOUNT CREDITS & CASH RECEIPTS편

    제품 : FIN_AR
    작성날짜 : 2003-09-16
    AR DATA FLOW (TABLE LEVEL) - ON ACCOUNT CREDITS & CASH RECEIPTS편
    ================================================================
    PURPOSE
    이 문서에서는 AR table들에 대해 어떻게 data가 들어가는지에 대해
    설명한다.
    On Account와 Cash Receipt을 기준으로 설명한다.
    Explanation
    1. On Account Credits
    특정 Customer에 대한 Credit정보를 입력했으나, 아직 특정 Invoice에는 Apply되지 않은 정보를 "On Account" credit이라고 한다.
    Credit정보가 들어있기 때문에, Credit Memo와 유사하게 table에 저장된다.
    Credit Memo와 다른 점은 아래와 같다.
    o When first entered the payment schedule will be fully remaining.
    o When first entered no records are inserted into AR_RECEIVABLE_APPLI
    CATIONS_ALL.
    o The line records will have NULL values in
    PREVIOUS_CUSTOMER_TRX_ID
    PREVIOUS_CUSTOMER_TRX_LINE_ID
    o The distribution lines have to be typed in as there is no invoice
    to copy them from.
    "On Account"는 같은 Supplier상에서는 어떠한 invoice에 대해서 적용이 가능하다는 점에서,
    특정 invoice에만 apply가 가능한 Credit Memo와는 다르다.
    2. Cash Receipts
    Recipt정보가 입력되면, 아래의 table들에 insert된다.
    o AR_CASH_RECEIPTS_ALL
    o AR_CASH_RECEIPT_HISTORY_ALL
    o AR_PAYMENT_SCHEDULES_ALL
    o AR_RECEIVABLE_APPLICATIONS_ALL
    입력된 Receipt정보가 특정 invoice와 match되면, 추가 record가
    AR_RECEIVABLE_APPLICATIONS_ALL에 insert되고, AR_PAYMENT_SCHEDULE_ALL에는 update된다.
    | | | |
    | RECEIPT |------------------| PAYMENT |
    | | | SCHEDULE |
    | |
    | |
    ^ ^
    /|\ /|\
    | | | |
    | RECEIPT | |APPLICATIONS|
    | HISTORY | | |
    2.1 AR_CASH_RECEIPTS_ALL
    Receipt에 대한 기본정보가 insert된다. 한 receipt당 한줄이 insert된다.
    Key = CASH_RECEIPT_ID (from sequence AR_CASH_RECEIPTS_S)
    Important Fields
    AMOUNT - Value of receipt in entered currency
    RECEIPT_NUMBER - Payment Number entered by user.
    STATUS - (APP)lied, (UNAPP)lied, (REV)ersed Payment,
    (STOP) payment, (NSF) insufficient funds.
    It will only change to APP once the whole
    amount of the receipt is applied.
    REVERSAL_DATE - NULL unless receipt reversed
    PAY_FROM_CUSTOMER - Contains CUSTOMER_ID for RA_CUSTOMERS
    2.2 AR_CASH_RECEIPT_HISTORY_ALL
    Receipt하나당 한줄의 data가 insert되고, GL로 Posting된 정보가 들어간다.
    Receipt이 reverse되면, 새로운 row가 insert된다.
    Key = CASH_RECEIPT_HISTORY_ID (from sequence)
    Important Fields
    CASH_RECEIPT_ID - Foreign key to AR_CASH_RECEIPTS record.
    STATUS - CLEARED for manually input receipts.
    GL_DATE - Accounting date
    ACCOUNT_CODE_COMBINATION_ID - Key to GL_CODE_COMBINATIONS
    POSTING_CONTROL_ID - -3 if unposted
    REVERSAL_POSTING_CONTROL_ID - NULL unless payment reversed
    CURRENT_RECORD_FLAG - Y if this is latest record
    PRV_STAT_CASH_RECEIPT_HIST_ID - Key to previous receipt history record.
    2.3 AR_PAYMENT_SCHEDULES_ALL
    Invoice에 apply된 Total 금액정보가 저장된다.
    Key = PAYMENT_SCHEDULE_ID (from sequence)
    Important Fields
    CUSTOMER_TRX_ID - NULL
    CASH_RECEIPT_ID - Foreign key to AR_CASH_RECEIPTS record.
    AMOUNT_DUE_ORIGINAL - Total amount of receipt (usually negative)
    AMOUNT_DUE_REMAINING - Unapplied amount of receipt.
    AMOUNT_APPLIED - How much of this receipt is applied .
    STATUS - (OP)en or (CL)osed. Will only be closed if
    AMOUNT_DUE_REMAINING is zero.
    All of the fields holding LINE, TAX and FREIGHT amounts are NULL.
    2.4 AR_RECEIVABLE_APPLICATIONS
    Receipt이 처음 생성될때, 한줄의 data가 이 table에 insert되고,
    invoice에 대해 Apply혹은 Unapply가발생하면, 두줄씩 새롭게 생성된다.
    예를들면, 아래와 같다.
    Record 1 UNAPP 700
    { Record 2      UNAPP       -200
    { Record 3      APP          200      cross referenced to the Invoice
    { Record 4      UNAPP       -500
    { Record 5      APP          500      cross referenced to 2nd Invoice
    The sum of the amounts on records that have a particuar status should add up
    to the running totals on the payment schedulesi, but with the opposite sign.
    i.e. In the example above
    AR_PAYMENT_SCHEDULES.AMOUNT_DUE_ORIGINAL = -700
    AR_PAYMENT_SCHEDULES.AMOUNT_DUE_REMAINING = 0
    AR_PAYMENT_SCHEDULES.AMOUNT_APPLIED = -700
    UNAPP = 700 -200 -500 = 0
    APP = 200 + 500 = 700
    Statuses of these records can be:-
    UNAPP - Unapplied
    APP - Applied
    ACC - On Account
    UNID - Unidentified (Customer Not known)
    이러한 record내역은 invoice/credit/receipt에 있는 Transaction History form에서 확인할 수 있다.
    2.5 AR_PAYMENT_SCHEDULE (Invoice)
    Receipt이 특정 invoice에 apply되면, invoice에 대한 Payment Schedule record가 update된다.
    remaining amount field 값은 Payment금액만큼 줄어들게 된다.
    만약, remaining amount가 "0"가 되면, invoice Payment schedule은 closed상태가 된다.
    예를들어, Receipt금액 "200"이 "1175" invoice금액(Tax금액 175가 포함된)에 apply되었다면, Invoice의 Payment Schedule은 아래와 같이 조정된다.
    Before After
    AMOUNT_DUE_REMAINING 1175.00 975.00
    AMOUNT_LINE_ITEMS_REMAINING 1000.00 800.00
    TAX_REMAINING 175.00 175.00
    FREIGHT_REMAINING 0.00 0.00
    Note that receipts are applied in a fixed sequence:-
    1. Line Amounts
    2. Tax Amounts
    3. Freight Amounts
    ie The TAX_REMAINING figure will only start to decrease when the
    AMOUNT_LINE_ITEMS_REMAINING is zero.
    Reference Documents
    Note : 29277.1 & 29278.1

    Hi,
    This query works fine for me:
    SELECT CR.CASH_RECEIPT_ID,
                CR.RECEIPT_NUMBER,
                CR.RECEIPT_DATE,
                CR.CURRENCY_CODE,
                DECODE ( CR.TYPE, 'MISC', NULL, NVL (SUM (DECODE (RA.STATUS, 'ACC', NVL (RA.AMOUNT_APPLIED, 0), 0)), 0)) ON_ACCOUNT_AMOUNT
             FROM AR_RECEIVABLE_APPLICATIONS_ALL RA,
                AR_CASH_RECEIPTS_ALL CR,
                AR_RECEIPT_METHODS RM
          WHERE RA.CASH_RECEIPT_ID = CR.CASH_RECEIPT_ID
                AND CR.RECEIPT_METHOD_ID = RM.RECEIPT_METHOD_ID
                AND CR.ORG_ID = <org_id>
          GROUP BY CR.CASH_RECEIPT_ID,
                CR.RECEIPT_DATE,
                CR.RECEIPT_NUMBER,
                RM.NAME,
                CR.CURRENCY_CODE,
                CR.TYPE order by receipt_date desc
    Let me know if it worked.
    Octavio

  • COMM_SRTUCTURE is uknown when migrating data flow bw 3.x to 7.4

    Dear ALL,
    2LIS_13_VDHDR data flow migrating 3.x to 7.x , ABAP syntax error COMM_STRUCTURE is unknown infosoure transformation level, present we are using 7.4 sp5 . after migration ABAP code
    TYPES:
          BEGIN OF _ty_s_TG_1_full,
    *      InfoObject: 0CHNGID Change Run ID.
            CHNGID           TYPE /BI0/OICHNGID,
    *      InfoObject: 0RECORDTP Record type.
            RECORDTP           TYPE /BI0/OIRECORDTP,
    *      InfoObject: 0REQUID Request ID.
            REQUID           TYPE /BI0/OIREQUID,
    *      InfoObject: 0CALDAY Calendar Day.
            CALDAY           TYPE /BI0/OICALDAY,
    *      InfoObject: 0CALMONTH Calendar Year/Month.
            CALMONTH           TYPE /BI0/OICALMONTH,
    *      InfoObject: 0CALWEEK Calendar year / week.
            CALWEEK           TYPE /BI0/OICALWEEK,
    *      InfoObject: 0FISCPER Fiscal year / period.
            FISCPER           TYPE /BI0/OIFISCPER,
    *      InfoObject: 0FISCVARNT Fiscal year variant.
            FISCVARNT           TYPE /BI0/OIFISCVARNT,
    *      InfoObject: 0BILLTOPRTY Bill-to party.
            BILLTOPRTY           TYPE /BI0/OIBILLTOPRTY,
    *      InfoObject: 0COMP_CODE Company code.
            COMP_CODE           TYPE /BI0/OICOMP_CODE,
    *      InfoObject: 0DISTR_CHAN Distribution Channel.
            DISTR_CHAN           TYPE /BI0/OIDISTR_CHAN,
    *      InfoObject: 0DOC_CATEG Sales Document Category.
            DOC_CATEG           TYPE /BI0/OIDOC_CATEG,
    *      InfoObject: 0PLANT Plant.
            PLANT           TYPE /BI0/OIPLANT,
    *      InfoObject: 0SALESORG Sales Organization.
            SALESORG           TYPE /BI0/OISALESORG,
    *      InfoObject: 0SALES_GRP Sales group.
            SALES_GRP           TYPE /BI0/OISALES_GRP,
    *      InfoObject: 0SALES_OFF Sales Office.
            SALES_OFF           TYPE /BI0/OISALES_OFF,
    *      InfoObject: 0SHIP_TO Ship-To Party.
            SHIP_TO           TYPE /BI0/OISHIP_TO,
    *      InfoObject: 0SOLD_TO Sold-to party.
            SOLD_TO           TYPE /BI0/OISOLD_TO,
    *      InfoObject: 0VERSION Version.
            VERSION           TYPE /BI0/OIVERSION,
    *      InfoObject: 0VTYPE Value Type for Reporting.
            VTYPE           TYPE /BI0/OIVTYPE,
    *      InfoObject: 0DIVISION Division.
            DIVISION           TYPE /BI0/OIDIVISION,
    *      InfoObject: 0MATERIAL Material.
            MATERIAL           TYPE /BI0/OIMATERIAL,
    *      InfoObject: 0SHIP_POINT Shipping point.
            SHIP_POINT           TYPE /BI0/OISHIP_POINT,
    *      InfoObject: 0PAYER Payer.
            PAYER           TYPE /BI0/OIPAYER,
    *      InfoObject: 0DOC_CLASS Document category /Quotation/Order/Deliver
    *y/Invoice.
            DOC_CLASS           TYPE /BI0/OIDOC_CLASS,
    *      InfoObject: 0DEB_CRED Credit/debit posting (C/D).
            DEB_CRED           TYPE /BI0/OIDEB_CRED,
    *      InfoObject: 0SALESEMPLY Sales Representative.
            SALESEMPLY           TYPE /BI0/OISALESEMPLY,
    *      InfoObject: 0SUBTOT_1S Subtotal 1 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_1S           TYPE /BI0/OISUBTOT_1S,
    *      InfoObject: 0SUBTOT_2S Subtotal 2 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_2S           TYPE /BI0/OISUBTOT_2S,
    *      InfoObject: 0SUBTOT_3S Subtotal 3 from pricing proced.for conditi
    *on in stat. curr..
            SUBTOT_3S           TYPE /BI0/OISUBTOT_3S,
    *      InfoObject: 0SUBTOT_4S Subtotal 4 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_4S           TYPE /BI0/OISUBTOT_4S,
    *      InfoObject: 0SUBTOT_5S Subtotal 5 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_5S           TYPE /BI0/OISUBTOT_5S,
    *      InfoObject: 0SUBTOT_6S Subtotal 6 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_6S           TYPE /BI0/OISUBTOT_6S,
    *      InfoObject: 0OPORDQTYBM Open orders quantity in base unit of meas
    *ure.
            OPORDQTYBM           TYPE /BI0/OIOPORDQTYBM,
    *      InfoObject: 0OPORDVALSC Net value of open orders in statistics cu
    *rrency.
            OPORDVALSC           TYPE /BI0/OIOPORDVALSC,
    *      InfoObject: 0QUANT_B Quantity in base units of measure.
            QUANT_B           TYPE /BI0/OIQUANT_B,
    *      InfoObject: 0DOCUMENTS No. of docs.
            DOCUMENTS           TYPE /BI0/OIDOCUMENTS,
    *      InfoObject: 0DOC_ITEMS Number of Document Items.
            DOC_ITEMS           TYPE /BI0/OIDOC_ITEMS,
    *      InfoObject: 0NET_VAL_S Net value in statistics currency.
            NET_VAL_S           TYPE /BI0/OINET_VAL_S,
    *      InfoObject: 0COST_VAL_S Cost in statistics currency.
            COST_VAL_S           TYPE /BI0/OICOST_VAL_S,
    *      InfoObject: 0GR_WT_KG Gross weight in kilograms.
            GR_WT_KG           TYPE /BI0/OIGR_WT_KG,
    *      InfoObject: 0NT_WT_KG Net weight in kilograms.
            NT_WT_KG           TYPE /BI0/OINT_WT_KG,
    *      InfoObject: 0VOLUME_CDM Volume in cubic decimeters.
            VOLUME_CDM           TYPE /BI0/OIVOLUME_CDM,
    *      InfoObject: 0HDCNT_LAST Number of Employees.
            HDCNT_LAST           TYPE /BI0/OIHDCNT_LAST,
    *      InfoObject: 0CRM_PROD Product.
            CRM_PROD           TYPE /BI0/OICRM_PROD,
    *      InfoObject: 0CP_CATEG Category.
            CP_CATEG           TYPE /BI0/OICP_CATEG,
    *      InfoObject: 0FISCYEAR Fiscal year.
            FISCYEAR           TYPE /BI0/OIFISCYEAR,
    *      InfoObject: 0BP_GRP BP: Business Partner Group (from Hierarchy).
            BP_GRP           TYPE /BI0/OIBP_GRP,
    *      InfoObject: 0STAT_CURR Statistics Currency.
            STAT_CURR           TYPE /BI0/OISTAT_CURR,
    *      InfoObject: 0BASE_UOM Base Unit of Measure.
            BASE_UOM           TYPE /BI0/OIBASE_UOM,
    *      InfoObject: 0PROD_CATEG Product Category.
            PROD_CATEG           TYPE /BI0/OIPROD_CATEG,
    *      InfoObject: 0VOLUME Volume.
            VOLUME           TYPE /BI0/OIVOLUME,
    *      InfoObject: 0VOLUMEUNIT Volume unit.
            VOLUMEUNIT           TYPE /BI0/OIVOLUMEUNIT,
    *      InfoObject: 0FISCPER3 Posting period.
            FISCPER3           TYPE /BI0/OIFISCPER3,
    *      InfoObject: 0SALES_DIST Sales District.
            SALES_DIST           TYPE /BI0/OISALES_DIST,
    *      InfoObject: 0BILL_TYPE Billing type.
            BILL_TYPE           TYPE /BI0/OIBILL_TYPE,
    *      InfoObject: 0MOVE_PLANT Receiving Plant/Issuing Plant.
            MOVE_PLANT           TYPE /BI0/OIMOVE_PLANT,
    *      InfoObject: 0SHIP_COND Shipping conditions.
            SHIP_COND           TYPE /BI0/OISHIP_COND,
    *      InfoObject: 0AB_RFBSK Status for Transfer to Accounting.
            AB_RFBSK           TYPE /BI0/OIAB_RFBSK,
    *      InfoObject: 0AB_FKSTO Indicator: Document Is Cancelled.
            AB_FKSTO           TYPE /BI0/OIAB_FKSTO,
    *      InfoObject: 0CUST_GRP5 Customer Group 5.
            CUST_GRP5           TYPE /BI0/OICUST_GRP5,
    *      InfoObject: ZCU_COND1 Constomer Condition Group 1.
            /BIC/ZCU_COND1           TYPE /BIC/OIZCU_COND1,
    *      InfoObject: ZCU_COND2 Customer Condition Group 2.
            /BIC/ZCU_COND2           TYPE /BIC/OIZCU_COND2,
    *      InfoObject: ZBATCHCD Batch Code.
            /BIC/ZBATCHCD           TYPE /BIC/OIZBATCHCD,
    *      InfoObject: 0BATCH Batch number.
            BATCH           TYPE /BI0/OIBATCH,
    *      InfoObject: ZBATCH Batch number.
            /BIC/ZBATCH           TYPE /BIC/OIZBATCH,
    *      Field: RECORD Data record number.
            RECORD           TYPE RSARECORD,
          END   OF _ty_s_TG_1_full.
    * Additional declaration for update rule interface
      DATA:
        MONITOR       type standard table of rsmonitor  WITH HEADER LINE,
        MONITOR_RECNO type standard table of rsmonitors WITH HEADER LINE,
        RECORD_NO     LIKE SY-TABIX,
        RECORD_ALL    LIKE SY-TABIX,
        SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS.
    * global definitions from update rules
    * TABLES: ...
    DATA: IN    TYPE F,
          OUT   TYPE F,
          DENOM TYPE F,
          NUMER TYPE F.
    * Def. of 'credit-documents': following doc.categ. are 'credit docs'
    *   reversal invoice (N)
    *   credit memo  (O)
    *   internal credit memo (6)
    * Credit-documents are delivered with negative sign. Sign is switched
    * to positive to provide positive key-figures in the cube.
    * The combination of characteristics DE_CRED and DOC-CLASS provides
    * a comfortable way to distinguisch e.g. positive incoming orders or
    * order returns.
    * Def. der 'Soll-Dokumente': folgende Belegtypen sind 'Soll-Belege'
    *   Storno Rechnung (N)
    *   Gutschrift (O)
    *   Interne Verrechn. Gutschr. (6)
    * Soll-Dokumente werden mit negativem Vorzeichen geliefert. Um die Kenn-
    * zahlen positiv in den Cube zu schreiben, wird das Vorzeich. gedreht
    * Die Kombination der Merkmale DEB_CRED und DOC-CLASS gibt Ihnen die
    * Möglichkeit schnell z.B. zwischen Auftrags-Eingang oder Retouren zu
    * unterscheiden.
    DATA: DEB_CRED(3) TYPE C VALUE 'NO6'.
    FORM routine_0002
      TABLES
       P_MONITOR         structure rsmonitor
      CHANGING
        RESULT         TYPE _ty_s_TG_1_full-DOCUMENTS
        RETURNCODE     LIKE sy-subrc
        ABORT          LIKE sy-subrc
      RAISING
        cx_sy_arithmetic_error
        cx_sy_conversion_error.
    * init variables
    * fill the internal table "MONITOR", to make monitor entries
    CLEAR RESULT.
    RESULT = COMM_STRUCTURE-NO_INV.
    IF COMM_STRUCTURE-DOC_CATEG CA DEB_CRED.
       RESULT = RESULT * ( -1 ).
    ENDIF.
    RETURNCODE = 0.
      p_monitor[] = MONITOR[].
      CLEAR:
        MONITOR[].
    ENDFORM.                    "routine_0002
    FORM routine_0003
      TABLES
       P_MONITOR         structure rsmonitor
      CHANGING
        RESULT         TYPE _ty_s_TG_1_full-DEB_CRED
        RETURNCODE     LIKE sy-subrc
        ABORT          LIKE sy-subrc
      RAISING
        cx_sy_arithmetic_error
        cx_sy_conversion_error.
    * init variables
    * fill the internal table "MONITOR", to make monitor entries
      IF COMM_STRUCTURE-DOC_CATEG CA DEB_CRED.
        RESULT = 'C'.
      ELSE.
        RESULT = 'D'.
      ENDIF.
      RETURNCODE = 0.
      p_monitor[] = MONITOR[].
      CLEAR:
        MONITOR[].
    ENDFORM.   
    Error:
    E:Field "COMM_STRUCTURE-NO_INV" is unknown. It is neither in one of the
    specified tables nor defined by a "DATA" statement. "DATA" statement.
    communication structure chaged to sours fields but no uses , please suggest how can i proceed , Thanks in Advance immediate replay
    Thanks & Regards
    Ramesh G

    Hi Gareth,
    You have two options:
    1. Transport from BW 3.1 to BI 7.0. You'll need to create a transport route between both systems. This may cause you some troubles in the future when you want to modify the objects you transported.
    2. As there are few objects, you can use XML export utility from Transport connection. There, you make an XML file with the objects you need to transport. One thing that you may take care of, in this option, is that the business content objects you are exporting need to be activated in the destination system. Another problem is that querys are not exported.
    Since it's only a cube, maybe you can create the objects manually. Look that in BI 7.0 there are several new functionalities, i don't know how transport or xml export would work.
    Hope this helps.
    Regards,
    Diego

  • Reversing a flow in TPM10 when the posting period is closed

    Hi,
    I have an interest rate instrument. I want to reverse one of the flows that got posted. The flow was posted in November 2010 (which is closed now) using TBB1 with creating a payment request. The payment request is still open, did not get paid through F111.
    I went to FTR_EDIT, marked the flow as Reverse.
    Went to TPM10, I can see the flow. When I attempt to reverse the flow, I get the error that period 11, 2010 is closed.
    In the selection screen of TPM10, I enter a posting date in the current open period. I get the error that reason 01 allows posting only in date 03.11.2010.
    There is no option to change the FI reversal reason to 02. How can I reverse a flow in TPM10 in a closed period?

    Hi,
    In customizing, Transaction manager -- General Settings -- Accounting -- Define Reasons for Reversal.  In this check how reversal reason 01 is defined.    This is the reversal reason you specify in TPM10 and this is in addition to the FI reversal reason.  Here you can specify the FI reversal reason as reversal in current period and enter a current date for reversal or give a reversal reason as reversal in closed period.
    Regards,
    Ravi

  • Transport sequence of the data flow

    HI Gurus,
    I have created data model in DEV which I have to transport .
    My data flow is DS- TRANFORMATION -DSO-TRANSFORMATION -DSO.
    I want sequenceof the objects  to transported .
    The same Newly infobjects are present in the both dso . Can transport Both DSO in same request as both contain same newly created infoobject or in differnet request?
    What about the DTPS and Infopackage swhen I should transport them ?
    Please guide me.
    Regards
    sriram
    Edited by: sriram1983 on Nov 18, 2011 11:05 AM

    Hi Sriram,
    -bundle objects into request.
    -release objectsusing se01
    -import objects using stms
    -u can keep both DSO in one request.
    -you have to transport infopackage and DTP also.
    - First transport DSOs & transport IP and DTP.
    http://sap.ittoolbox.com/groups/technical-functional/sap-bw/transporting-bw-queries-1635374
    Reverse transport of  BI objects from quality  to Development ?
    /message/10741618#10741618 [original link is broken]
    /thread/2060078 [original link is broken]
    Regards,
    rvc

  • Using asynchronous timer for data flow control

    Hi all,
      I am using system sleep to control the data flow (some digital lines and analog output). The pseudo code is something like this
    Sleep(150);
    // the following sections are exectuted in parallel
      #pragma omp parallel sections
        #pragma omp section
          DAQmxWriteDigitalLines(...); // output TTL to one digitla line
        #pragma omp section
          DAQmxWriteDigitalLines(...); // output TTL to another digitla line     
        #pragma omp section
          Sleep(2); // sleep 2ms
    // the following sections are exectuted in parallel
      #pragma omp parallel sections
        #pragma omp section
          DAQmxWriteDigitalLines(...); // output TTL to one digitla line
        #pragma omp section
          DAQmxWriteAnalogScalarF64(...); // analog output to one channel
        #pragma omp section
          Sleep(1); // delay 1ms
    // the following sections are exectuted in parallel
      #pragma omp parallel sections
        #pragma omp section
          DAQmxWriteDigitalLines(...); // output TTL to one digitla line
        #pragma omp section
          DAQmxWriteAnalogScalarF64(...); // analog output to one channel
    #pragma omp section
          DAQmxWriteAnalogScalarF64(...); // analog output to another channel
        #pragma omp section
          Sleep(11); // delay 11ms
    // ... other stuffs
    I am running windows XP and I know it is not possible to get realtime control but  I want a as precise timing as possible. Above code is not perfect but it works 95% of times. I just read an article about using the asynchronous timer to control the time delay. I try that idea with the following code frame
    int CVICALLBACK ATCallback(int reserved, int timerId, int event, void *callbackData, int eventData1, int eventData2)
    if (event==EVENT_TIMER_TICK)
    int *nextdelay = (int *)callbackData;
    SuspendAsyncTimerCallbacks();
    if (timerId>=0)
    double time;
    if (*nextdelay==0) time=2.0;
    else if (*nextdelay==1) time=1.0;
    else time=12.0;
    SetAsyncTimerAttribute(timerId, ASYNC_ATTR_INTERVAL, time);
    if (*nextdelay==0)
    #pragma omp parallel sections
    #pragma omp section
    DAQmxWriteDigitalLines(...); // output TTL to one digitla line
    #pragma omp section
    DAQmxWriteDigitalLines(...); // output TTL to another digitla line
    *nextdelay++;
    else if (*nextdelay==2)
    #pragma omp parallel sections
    #pragma omp section
    DAQmxWriteDigitalLines(...); // output TTL to one digitla line
    #pragma omp section
    DAQmxWriteAnalogScalarF64(...); // analog output to one channel
    *nextdelay++;
    else if (*nextdelay==3)
    #pragma omp parallel sections
    #pragma omp section
    DAQmxWriteDigitalLines(...); // output TTL to one digitla line
    #pragma omp section
    DAQmxWriteAnalogScalarF64(...); // analog output to one channel
    #pragma omp section
    DAQmxWriteAnalogScalarF64(...); // analog output to another channel
    *nextdelay++;
    ResumeAsyncTimerCallbacks();
    return 0;
    void main(void)
    int n = 0;
    int timeid;
    timeid = NewAsyncTimer(120.0/1000.0, 3, 1, ATCallback, &n);
    But it doesn't work. There is no compilation and runtime error but the timing just not right. I wonder do I have to suspend the timer in the callback function when I reset the delay for next call? If I do so, I am worry if it will apply too much delay (since I suspend and resume the timer in the delay) so it will cause even worse timing. But if I don't suspend the timer before I reset the time, what happen if the code running in the callback function not finished before the next callback arrive. It is quite confusing how to use asynchronous timer in this case. Thanks.

    Yeah, unfortunately the 6711 doesn't have clocked digital I/O.  There are only two counters anyway so even if you could use them to generate your signals you wouldn't have enough (*maybe* something with the 4 AO channels and a counter depending on what your output signals need to look like?  The AO channels can output "digital" as well if you write 0V or 5V only).
    A PCI DAQ card which does support clocked digital I/O and has 2 analog outputs is the 6221 (or if you could use PCIe the 6321 is a more updated version with two extra counters and some additional functionality).
    If there isn't a way to implement clocked outputs afterall, one thing you could do to make your code a little more efficient is to consolidate the writes.  You can put your digital lines into a single task and write them at ocne, and you can put your analog channels into a single task and write them at once as well.
    I'm not sure about the callback issue, you might find some more help in the CVI forum.  I don't think it's going to solve your underlying problem though as ultimately the execution timing of your software calls is at the mercy of your OS.
    Best Regards,
    John Passiak

  • Display Data Flow - Short Dump

    Hi all,
    When i select display data flow of any cube...it is going for a short dump.
    I have searched for the answer in previous Forum questions. I could find only for previous BW versions but not for for BI7.
    Could you please let me know the solution for this issue.
    Thanks & Regards,
    Eswari

    Hi All,
    Thank you very much for all of your responces.....
    I am working on Support Package 10.
    Here is the detailed description of the short dump.
    Short text
        The current application triggered a termination with a short dump.
    What happened?
        The current application program detected a situation which really
        should not occur. Therefore, a termination with a short dump was
        triggered on purpose by the key word MESSAGE (type X).
    What can you do?
        Note down which actions and inputs caused the error.
        To process the problem further, contact you SAP system
        administrator.
        Using Transaction ST22 for ABAP Dump Analysis, you can look
        at and manage termination messages, and you can also
        keep them for a long time.
    Error analysis
        Short text of error message:
        GP: Control Framework returned an error; contact system administrator
        Long text of error message:
         Diagnosis
             The Graphical Framework is based on the basis technology known as
             the Control Framework. A method in the Control Framework returned
             an error.
         Procedure
             It probably involves a programming error. You should contact your
             system administrator.
         Procedure for System Administration
             Check the programming of the graphics proxy especially for the
             parameters that were sent and, if necessary, correct your program.
        Technical information about the message:
        Message class....... "APPLG"
        Number.............. 229
        Variable 1.......... " "
        Variable 2.......... " "
        Variable 3.......... " "
        Variable 4.......... " "
    How to correct the error
        Probably the only way to eliminate the error is to correct the program.
        If the error occures in a non-modified SAP program, you may be able to
        find an interim solution in an SAP Note.
        If you have access to SAP Notes, carry out a search with the following
        keywords:
        "MESSAGE_TYPE_X" " "
        "CL_AWB_OBJECT_NET_SAPGUI======CP" or "CL_AWB_OBJECT_NET_SAPGUI======CM005"
        "PBO"
        If you cannot solve the problem yourself and want to send an error
        notification to SAP, include the following information:
        1. The description of the current problem (short dump)
           To save the description, choose "System->List->Save->Local File
        (Unconverted)".
        2. Corresponding system log
           Display the system log by calling transaction SM21.
           Restrict the time interval to 10 minutes before and five minutes
        after the short dump. Then choose "System->List->Save->Local File
        (Unconverted)".
        3. If the problem occurs in a problem of your own or a modified SAP
        program: The source code of the program
           In the editor, choose "Utilities->More
        Utilities->Upload/Download->Download".
       4. Details about the conditions under which the error occurred or which
       actions and input led to the error.
    Thanks,
    Eswari.

  • SRM 7.0 data flow frm SC to PO

    Hi,
    As I am new to SRM, I wish to know how the data flow from SC to PO.
    Here is my scenarion, we are in extended classic scenarion, we have 10 custom fields in SC and the same 10 fields in PO. The moment a SC is ordered, approved and created our PO will be created . I want to know how the data for the custom fields and the standard fields flow from SC to PO.
    I found some BADi's, and I learnt that all those badis are related to BAckend PO,
    BBP_CREATE_BE_PO_NEW - For classic scenario in newer versions of SRM
    BBP_ECS_PO_OUT_BADI     - For extended classic scenario
    BBP_EXTLOCALPO_BADI     - For controlling the Extended classic scenario on Local PO
    Please correct me if I am wrong. Let me the exact data flow.
    Information on this highly appreciated and will be higly rewared. Thanks in advance.
    Krishna Chaitanya

    Solved

  • R/3 data flow is timing out in Data Services

    I have created an R/3 data flow to pull some AP data in from SAP into Data Services.  This data flow outputs to a query object to select columns and then outputs to a table in the repository.  However the connection to SAP is not working correctly.  When I try to process the data flow it just idles for an hour until the SAP timeout throws an error.  Here is the error:
    R/3 CallReceive error <Function Z_AW_RFC_ABAP_INSTALL_AND_RUN: connection closed without message (CM_NO_DATA_RECEIVED)
    I have tested authorizations by adding SAP_ALL to the service account I'm using and the problem persists.
    Also, the transports have all been loaded correctly.
    My thought is that it is related to the setting that controls the method of generating and executing the ABAP code for the data flow, but I can't find any good documentation that describes this, and my trial and error method so far has not produced results.
    Any help is greatly appreciated.
    Thanks,
    Matt

    You can't find any good documentation??? I am working my butt off just.......just kiddin'
    I'd suggest we divide the question into two parts:
    My dataflow takes a very long time, how can I prevent the timeout after an hour? Answer:
    Edit the datastore, there is a flag called "execute in background" to be enabled. With that the abap is submitted as a background spool job, hence does not have the dialog-mode timeout. Another advantage is, you can watch it running by brwosing the spool jobs from the SAP GUI.
    The other question seems to be, why does it take that long even? Answer:
    Either the ABAP takes that long because of the data volume.
    Or the ABAP is not performing well, e.g. join via ABAP loops with the wrong table as inner.
    Another typical reason is to use direct_download as transfer method. This is fine for testing but it takes a very long time to download data via the GUI_DOWNLOAD ABAP function. And the download time would be part of the ABAP execution.
    So my first set of questions would be
    a) How complex is the dataflow, is it just source - query - data_transfer or are there joins, lookups etc?
    b) What is the volume of the table(s)?
    c) What is your transfer method?
    d) Have you had a look at the generated abap? (in the R/3 dataflow open the menu Validation -> Generate ABAP)
    btw, some docs: https://wiki.sdn.sap.com:443/wiki/display/BOBJ/ConnectingtoSAP

  • Error on Data Flow Task MSSQL 2012 Clustered "Description: The version of Lookup is not compatible with this version of the DataFlow. "

    We have an SSIS package that runs on clustered MSSQL 2012 Enterprise Nodes that is failing.  We use a job to executer the package.
    Environmental information:
    Product - Microsoft SQL Server Enterprise: Core-based Licensing (64-bit)
    Operating System - Microsoft Windows NT 6.1 (7601)
    Patform - NT x64
    Version - MSSQL Version 11.0.3349.0
    Package is set to 32 -bit.  All permissions verified.  Runs in lower environments, same MSSQL version.  All environments are clustered.  In the failing environment, all nodes are at the same service pack.  I have not verified if all
    nodes in the failing environment have SSIS installed.  Data access is installed.  We have other simpler packages that run in this environment, just not this one.  Time to ask the community for help!
    Error:
    Source: Data Flow Task - Data Flow Task (SSIS.Pipeline)     Description: The version of Lookup is not compatible with this version of the DataFlow.  End Error  Error:  Code: 0xC0048020    
    Description: Component "Conditional Split, clsid {7F88F654-4E20-4D14-84F4-AF9C925D3087}" could not be created and returned error code 0x80070005 "Access is denied.". Make sure that the component is registered correctly.  End Error 
    Description: The component is missing, not registered, not upgradeable, or missing required interfaces. The contact information for this component is "Conditional Split;Microsoft Corporation; Microsoft SQL Server; (C) Microsoft Corporation; All Rights
    Reserved; http://www.microsoft.com/sql/support;0".  End Error 
    (Left out shop specific information.  This is the first error in the errors returns by the job history for this package. )
    Thanks in advance.

    Hi DeveloperMax,
    According to your description, the error occurs when you execute the package with Agent job on clustered MSSQL 2012 Enterprise Nodes.
    As per my understanding, I think this issue can be caused by you use SQL Server Agent to schedule a SQL Server Integration Services package in a 64-bit environment. And the SSIS package is referencing some 32-Bit DLL or 32-Bit drivers which are available
    only in 32-bit versions, so the job failed.
    To fix this issue, we should use the 32-bit version of the DTExec.exe utility to schedule the 64-bit SQL Server Agent to run a package. To run a package in 32-bit mode from a 64-bit version of SQL Server Agent, we can go to the Job Step dialog box, then
    select “32 bit runtime” in the Advanced tab.
    Besides, we should make sure that SQL Server Integration Services is installed on the failing environment.
    If there are any other questions, please feel free to ask.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Data flows are getting started but not completing successfully while extracting/loading of the data

    Hello People,
    We are facing a abnormal behavior with the dataflows in the data services job.
    Scenario:
    We are extracting the data from CRM end in parallel. Please refer the build:
    a. We have 5 main workflows flows i.e :
       => Main WF1 has 6 more sub Wf's in it, in which each sub Wf has 1/2 DF's associated in parallel.
       => Main WF2 has 21 DF's and 1 WFa->with a DF & a WFb. WFb has 1 DF in parallel.
       => Main WF3 has 1 DF in parallel.
       => Main WF4 has 3 DF in parallel.
       => Main WF5 has 1 WF & a DF in sequence.
    b. Regularly the job works perfectly fine but, sometimes it gets stuck at the DF’s without any error logs.
    c. Job doesn’t stuck at a specific dataflow or on a specific day, many a times it strucks at different DF’s.
    d. Observations in the Monitor Log:
    Dataflow---------------------- State----------------RowCnt------LT-------AT------ 
    +DF1/ZABAPDF
    PROCEED
    234000
    8.113      394.164
    /DF1/Query
    PROCEED
    234000
    8.159      394.242
    -DF1/Query_2
    PROCEED
    234000
    8.159      394.242
    Where LT: Lapse Time and AT: Absolute time
    If you check the monitor log, the State of the Dataflow DF1 remains PROCEED till the end, ideally it should complete.
    In successful jobs, the status for DF1  is STOP . This DF takes approx. 2 min to execute.
    The row count for DF1 extraction is 234204 but, it got stuck at  234000.
    Then we terminate the job after sometime,but for surprise it gets executed successfully on next day.
    e. As per analysis over all the failed jobs, same things were observed over the different data flows that got stuck during the execution.Logic related to the data flows is perfectly fine.
    Observations in the Trace log:
    DATAFLOW: Process to execute data flow <DF1> is started.
    DATAFLOW: Data flow <DF1> is started.
    ABAP: ABAP flow <ZABAPDF> is started.
    ABAP: ABAP flow <ZABAPDF> is completed.
    Cache statistics determined that data flow <DF1>
    uses <0>caches with a total size of <0> bytes. This is less than(or equal to) the virtual memory <1609564160> bytes available for caches.
    Statistics is switching the cache type to IN MEMORY.
    DATAFLOW: Data flow <DF1> using IN MEMORY Cache.
    DATAFLOW: <DF1> is completed successfully.
    The highlighted text in the trace log is not appearing in the unsuccessful job but, it appears for the successful one.
    Note: The cache type is pageable cache, DS ver is 3.2.
    Please suggest.
    Regards,
    Santosh

    Hi Santosh,
    just a wild guess.
    Would you be able to replicate all the DF\WF , delete original DF\WF, rename replicated objects to original to DF\WF names(for your convenience)   and excute it.
    Some time reference does not work.
    Hope this should work.
    Regards,
    Shiva Sahu

  • Data flow fails on packed decimal field moving iSeries DB2 data from one iSeries DB to another

    I' trying to use SSIS to move table content from one iSeries DB2 database to another.  I'm using the .Net providers for OleDb\IBM DB2 for i5/OS IBMDA400 OLE DB Provider in the connection managers for the source and destination and the test connection
    works fine.  When I try to run the data flow task however it fails on the first packed decimal field it encounters with the exceptions ...
    [select from hydro520 hydroweb2 blpmstr [16]] Error: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR.  The "component "select from hydro520 hydroweb2 blpmstr" (16)" failed because error code 0x80004002 occurred, and the error
    row disposition on "output column "MSPRIB" (55)" specifies failure on error. An error occurred on the specified object of the specified component.  There may be error messages posted before this with more information about the failure.
    [select from hydro520 hydroweb2 blpmstr [16]] Error: The component "select from hydro520 hydroweb2 blpmstr" (16) was unable to process the data. Pipeline component has returned HRESULT error code 0xC0209029 from a method call.
    [SSIS.Pipeline] Error: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "select from hydro520 hydroweb2 blpmstr" (16) returned error code 0xC02090F5.  The component returned a failure code when the pipeline
    engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    ...in the progress tab.  Can someone kindly tell me what I need to do to get the connection manager to work with DB2 packed decimal fields?  Or is it a different issue all together?  Thanks tonnes for any help, Roscoe

    Hi rpfinn,
    From the Data Types mapping rules between SSIS and DB2, we can see that both the NUMERIC and DECIMAL data types in DB2 are mapped to DT_NUMBERIC data type in SSIS. Now that the source data in your DB2 database is NUMERIC data type, changing the DT_NUMBRIC
    data type to DT_DECIMAL is invalid. Besides, if we check the data types of the target External column and Output column from the Advanced Editor for ADO NET Source adapter, the data type should be defined as DT_NUMERIC with Precision as 9 and Scale as 2. I
    am not clear where you see the DT_NUMBERIC(9,0) e.g. DT_NUMERIC with Precision as 9 and Scale as 0, but it may be the cause of the issue. You need to make sure the DT_NUMERIC data type also has Scale 2 instead of 0.
    If you don’t know how to modify the data type, please elaborate the Data Flow Task of the package so that we can make further analysis. Besides, the error messages you posted are incomplete, it will be helpful if you post the complete error message.
    Regards,
    Mike Yin
    If you have any feedback on our support, please click
    here
    Mike Yin
    TechNet Community Support

  • How to derive month/year from date in SAP BW 3.5 data flow

    Hi
    How we can derive cal year/month and fiscal month/year from date in SAP BW 3.5 data flow (we're using transfer and update rule)..
    Thanks,
    PK

    Hi,
    if you have any date filed in source side you can just map to any time char system will automatically convert to target objects.
    please look at the screen shot for understanding. (not 3.x it is 7.x)
    Thanks,
    Phani.

  • Problem with context mapping and data flow in a FPM application

    Hi All,
    I am trying to develop an ESS application using FPM. For the same, the requirement is to see the history of an employee in the second view.
    The first view has got just the overview information and the second one has got the detail. So, the records or the fields are the same on both the views.
    As per the FPM guidelines, the Model is residing in the Fc component and the respective Vc components are using the model data accordingly.
    I am executing the model in the Fc component calling the executable method in the interfaceController of the first view and then trying to display the output data of the BAPI in the first view which provides the overview information.This is working fine.
    But when i am trying to map the same output node to the Table UI for the second view, the record size is coming zero and thus no information is available.
    For the above issue, I am again executing the RFC in the InterfaceController of the second view to populate the records, which is incorrect as it is already executed and the data is available for the first view.
    I request you to let me know the correct approach to Context mapping and data flow when using FPM-roadmap. Is their any standard method or approach available to deal with such requirements? Please let me know.
    Thanks in advance.
    Regards
    DK

    Hi Idhaya,
    I model node is available in Fc and the Fc interface controller is being used in the first Vc and the second Vc.
    So the idea is, as the executable method is generated in the Fc, so i have created a custom method to call the executable method in Fc, where the input parameter is getting passed and this custom method is finally getting called is the first Vc.
    So , now my first Vc is ready to call the custom method in Fc and execute the RFC. Once the RFC is executed, the nodes in the Fc should get populated which is the ideal case.
    And as the Fc is used as a component in the second Vc, the same node is available to the UI elements.
    But, when I check the record size for the output node, it is always zero, for the second Vc.
    Regards
    DK

  • Data Flow terminated due to error 120307

    Hi.
    I get this error when executed  project.
    Source system: SyBase IQ.
    Target system: SAP HANA.
    Part of tables copied successfully, but job terminated anyway.
    I attached screenshots with Progress screen and Monitoring.
    Error log is empty.
    In trace log I see errors from SUBJ.
    Also I have another one strange message in trace log:
    Cache statistics determined that data flow <SYBASE_IQ_2_HOD_DBA_FACT_FINAL> uses 0 caches with a total size of 0 bytes, which is less than (or equal to) 3757047808 bytes available for caches in virtual memory. Data flow will use IN MEMORY cache type.

    I executed this job from data services designer and get another error.
    main Bufman: An error was detected on a database page. You may have a damaged index. For additional information, please check your IQ message file or run sp_iqcheckdb
    Trying to find issue in goggle

Maybe you are looking for

  • Wireless Printing To A Windows Computer

    Hello I have 2 windows pcs with a usb switch to switch my printer between the two, without having to use a print server. The 2 machines are connected to my home network with a linksys wireless router. I just got a new Macbook and i can't figure out h

  • Re:Module pool programming

    HI,      i have developed one interface  , when i click on submit the values which i type in the text box i/o fields should get inserted into database table( before this i would like to store these values in the internal table n would display whether

  • Zoom in on video clips???

    anyone know of any plugins that allow zooming in on video clips? sort of like the ken burns effect for still photos? thanks

  • All shot in 24p except 1 scene.  Please help, thank you.

    I shot the whole movie in 24p but captured and used a 29.97 timeline...this is fine. There is one scene we didn't film in 24p, but i want to give it the same look as the rest of the footage. I was thinking to convert to 23.98 fps, but it will get scr

  • Access to hardware and software inventory?

    It seems that as soon as I make someone an administrator in the ZCM console, they get access to all the detailed hardware and software inventory of all devices. Is this true and if so, is there a way to deny it? What I want to be able to do is allow