When to run RSMIGRATE (Data Flow Conversion)?

Hi Guys
Situation: Are about to upgrade (with DMO) a BW 7.01/DB2 system to BW 7.40 SP7/HANA.
We have a discussion if RSMIGRATE (Data Flow Conversion) must be done BEFORE the upgrade or AFTER ?
Transaction RSMIGRATE does not exists on source release BW 7.01, but can of cause be implemented by a SAP note. However - is this required or can we do the upgrade and DB migration and then run this transaction after this on release 7.40/HANA ?
Best regards

Hi Nick,
Master data is usually stored in the ODS which has a overwriting property so data will not duplicate. Hence there is no need to delete the master data before running the chain .
And it is usually loaded daily.
Hope it helps.
Regards
Hemant Khemani

Similar Messages

  • Data flow fails on packed decimal field moving iSeries DB2 data from one iSeries DB to another

    I' trying to use SSIS to move table content from one iSeries DB2 database to another.  I'm using the .Net providers for OleDb\IBM DB2 for i5/OS IBMDA400 OLE DB Provider in the connection managers for the source and destination and the test connection
    works fine.  When I try to run the data flow task however it fails on the first packed decimal field it encounters with the exceptions ...
    [select from hydro520 hydroweb2 blpmstr [16]] Error: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR.  The "component "select from hydro520 hydroweb2 blpmstr" (16)" failed because error code 0x80004002 occurred, and the error
    row disposition on "output column "MSPRIB" (55)" specifies failure on error. An error occurred on the specified object of the specified component.  There may be error messages posted before this with more information about the failure.
    [select from hydro520 hydroweb2 blpmstr [16]] Error: The component "select from hydro520 hydroweb2 blpmstr" (16) was unable to process the data. Pipeline component has returned HRESULT error code 0xC0209029 from a method call.
    [SSIS.Pipeline] Error: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "select from hydro520 hydroweb2 blpmstr" (16) returned error code 0xC02090F5.  The component returned a failure code when the pipeline
    engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    ...in the progress tab.  Can someone kindly tell me what I need to do to get the connection manager to work with DB2 packed decimal fields?  Or is it a different issue all together?  Thanks tonnes for any help, Roscoe

    Hi rpfinn,
    From the Data Types mapping rules between SSIS and DB2, we can see that both the NUMERIC and DECIMAL data types in DB2 are mapped to DT_NUMBERIC data type in SSIS. Now that the source data in your DB2 database is NUMERIC data type, changing the DT_NUMBRIC
    data type to DT_DECIMAL is invalid. Besides, if we check the data types of the target External column and Output column from the Advanced Editor for ADO NET Source adapter, the data type should be defined as DT_NUMERIC with Precision as 9 and Scale as 2. I
    am not clear where you see the DT_NUMBERIC(9,0) e.g. DT_NUMERIC with Precision as 9 and Scale as 0, but it may be the cause of the issue. You need to make sure the DT_NUMERIC data type also has Scale 2 instead of 0.
    If you don’t know how to modify the data type, please elaborate the Data Flow Task of the package so that we can make further analysis. Besides, the error messages you posted are incomplete, it will be helpful if you post the complete error message.
    Regards,
    Mike Yin
    If you have any feedback on our support, please click
    here
    Mike Yin
    TechNet Community Support

  • Display Data Flow - Short Dump

    Hi all,
    When i select display data flow of any cube...it is going for a short dump.
    I have searched for the answer in previous Forum questions. I could find only for previous BW versions but not for for BI7.
    Could you please let me know the solution for this issue.
    Thanks & Regards,
    Eswari

    Hi All,
    Thank you very much for all of your responces.....
    I am working on Support Package 10.
    Here is the detailed description of the short dump.
    Short text
        The current application triggered a termination with a short dump.
    What happened?
        The current application program detected a situation which really
        should not occur. Therefore, a termination with a short dump was
        triggered on purpose by the key word MESSAGE (type X).
    What can you do?
        Note down which actions and inputs caused the error.
        To process the problem further, contact you SAP system
        administrator.
        Using Transaction ST22 for ABAP Dump Analysis, you can look
        at and manage termination messages, and you can also
        keep them for a long time.
    Error analysis
        Short text of error message:
        GP: Control Framework returned an error; contact system administrator
        Long text of error message:
         Diagnosis
             The Graphical Framework is based on the basis technology known as
             the Control Framework. A method in the Control Framework returned
             an error.
         Procedure
             It probably involves a programming error. You should contact your
             system administrator.
         Procedure for System Administration
             Check the programming of the graphics proxy especially for the
             parameters that were sent and, if necessary, correct your program.
        Technical information about the message:
        Message class....... "APPLG"
        Number.............. 229
        Variable 1.......... " "
        Variable 2.......... " "
        Variable 3.......... " "
        Variable 4.......... " "
    How to correct the error
        Probably the only way to eliminate the error is to correct the program.
        If the error occures in a non-modified SAP program, you may be able to
        find an interim solution in an SAP Note.
        If you have access to SAP Notes, carry out a search with the following
        keywords:
        "MESSAGE_TYPE_X" " "
        "CL_AWB_OBJECT_NET_SAPGUI======CP" or "CL_AWB_OBJECT_NET_SAPGUI======CM005"
        "PBO"
        If you cannot solve the problem yourself and want to send an error
        notification to SAP, include the following information:
        1. The description of the current problem (short dump)
           To save the description, choose "System->List->Save->Local File
        (Unconverted)".
        2. Corresponding system log
           Display the system log by calling transaction SM21.
           Restrict the time interval to 10 minutes before and five minutes
        after the short dump. Then choose "System->List->Save->Local File
        (Unconverted)".
        3. If the problem occurs in a problem of your own or a modified SAP
        program: The source code of the program
           In the editor, choose "Utilities->More
        Utilities->Upload/Download->Download".
       4. Details about the conditions under which the error occurred or which
       actions and input led to the error.
    Thanks,
    Eswari.

  • How data flow when SSIS packages are run on a different server than the DB server

    The scenario is that i have a dedicated SQL SErver 2014 SSIS machine that executes the packages.
    The database server is a separate machine with SQL Server 2008R2.
    1) Running SSIS packages that transfer data within SQL Server 2008R2 (same machine)
    2) Running SSIS packages that transfer data between 2 separate SQL Server servers.
    How the data flow in these two cases and what resource is being used where? (cpu,disk,ram,network)
    Elias

    When you have a dedicated SSIS server, all data read flows to that server, is processed using the resources of that ETL server and then sent back over the network to the destination server.
    It doesn't matter if source and destination are the same server. If you use a data flow, all data flows over the network twice.
    The only exception is when you don't use a data flow, but only SQL statements. In that case, data flows only between source and destination.
    MCSE SQL Server 2012 - Please mark posts as answered where appropriate.

  • DS 4.2 get ECC CDHDR deltas in ABAP data flow using last run log table

    I have a DS 4.2 batch job where I'm trying to get ECC CDHDR deltas inside an ABAP data flow.  My SQL Server log table has an ECC CDHDR last_run_date_time (e.g. '6/6/2014 10:10:00') where I select it at the start of the DS 4.2 batch job run and then update it to the last run date/time at the end of the DS 4.2 batch job run.
    The problem is that CDHDR has the date (UDATE) and time (UTIME) in separate fields and inside an ABAP data flow there are limited DS functions.  For example, outside of the ABAP data flow I could use the DS function concat_date_time for UDATE and UTIME so that I could have a where clause of 'concat
    _date_time(UDATE, UTIME) > last_run_date_time and concat_date_time(UDATE, UTIME) <= current_run_date_time'.  However, inside the ABAP data flow the DS function concat_date_time is not available.  Is there some way to concatenate UDATE + UTIME inside an ABAP data flow?
    Any help is appreciated.
    Thanks,
    Brad

    Michael,
    I'm trying to concatenate date and time and here's my ABAP data flow where clause:
    CDHDR.OBJECTCLAS in ('DEBI', 'KRED', 'MATERIAL')
    and ((CDHDR.UDATE || ' ' || CDHDR.UTIME) > $CDHDR_Last_Run_Date_Time)
    and ((CDHDR.UDATE || ' ' || CDHDR.UTIME) <= $Run_Date_Time)
    Here are DS print statements showing my global variable values:
    $Run_Date_Time is 2014.06.09 14:14:35
    $CDHDR_Last_Run_Date_Time is 1900.01.01 00:00:01
    The issue is I just created a CDHDR record with a UDATE of '06/09/2014' and UTIME of '10:48:27' and it's not being pulled in the ABAP data flow.  Here's selected contents of the generated ABAP file (*.aba):
    PARAMETER $PARAM1 TYPE D.
    PARAMETER $PARAM2 TYPE D.
    concatenate CDHDR-UDATE ' ' into ALTMP1.
    concatenate ALTMP1 CDHDR-UTIME into ALTMP2.
    concatenate CDHDR-UDATE ' ' into ALTMP3.
    concatenate ALTMP3 CDHDR-UTIME into ALTMP4.
    IF ( ( ALTMP4 <= $PARAM2 )
    AND ( ALTMP2 > $PARAM1 ) ).
    So $PARAM1 corresponds to $CDHDR_Last_Run_Date_Time ('1900.01.01 00:00:01') and $PARAM2 corresponds to $Run_Date_Time ('2014.06.09 14:14:35').  But from my understanding ABAP data type D is for date only (YYYYMMDD) and doesn't include time, so is my time somehow being defaulted to '00:00:00' when it gets to DS?  I ask this as a CDHDR record I created on 6/6 wasn't pulled during my 6/6 testing but this 6/6 CDHDR record was pulled today.
    I can get  last_run_date_time and current_run_date_time into separate date and time fields but I'm not sure how to build the where clause using separate date and time fields.  Do you have any recommendations or is there a better way for me to pull CDHDR deltas in an ABAP data flow using something different than a last run log table?
    Thanks,
    Brad

  • COMM_SRTUCTURE is uknown when migrating data flow bw 3.x to 7.4

    Dear ALL,
    2LIS_13_VDHDR data flow migrating 3.x to 7.x , ABAP syntax error COMM_STRUCTURE is unknown infosoure transformation level, present we are using 7.4 sp5 . after migration ABAP code
    TYPES:
          BEGIN OF _ty_s_TG_1_full,
    *      InfoObject: 0CHNGID Change Run ID.
            CHNGID           TYPE /BI0/OICHNGID,
    *      InfoObject: 0RECORDTP Record type.
            RECORDTP           TYPE /BI0/OIRECORDTP,
    *      InfoObject: 0REQUID Request ID.
            REQUID           TYPE /BI0/OIREQUID,
    *      InfoObject: 0CALDAY Calendar Day.
            CALDAY           TYPE /BI0/OICALDAY,
    *      InfoObject: 0CALMONTH Calendar Year/Month.
            CALMONTH           TYPE /BI0/OICALMONTH,
    *      InfoObject: 0CALWEEK Calendar year / week.
            CALWEEK           TYPE /BI0/OICALWEEK,
    *      InfoObject: 0FISCPER Fiscal year / period.
            FISCPER           TYPE /BI0/OIFISCPER,
    *      InfoObject: 0FISCVARNT Fiscal year variant.
            FISCVARNT           TYPE /BI0/OIFISCVARNT,
    *      InfoObject: 0BILLTOPRTY Bill-to party.
            BILLTOPRTY           TYPE /BI0/OIBILLTOPRTY,
    *      InfoObject: 0COMP_CODE Company code.
            COMP_CODE           TYPE /BI0/OICOMP_CODE,
    *      InfoObject: 0DISTR_CHAN Distribution Channel.
            DISTR_CHAN           TYPE /BI0/OIDISTR_CHAN,
    *      InfoObject: 0DOC_CATEG Sales Document Category.
            DOC_CATEG           TYPE /BI0/OIDOC_CATEG,
    *      InfoObject: 0PLANT Plant.
            PLANT           TYPE /BI0/OIPLANT,
    *      InfoObject: 0SALESORG Sales Organization.
            SALESORG           TYPE /BI0/OISALESORG,
    *      InfoObject: 0SALES_GRP Sales group.
            SALES_GRP           TYPE /BI0/OISALES_GRP,
    *      InfoObject: 0SALES_OFF Sales Office.
            SALES_OFF           TYPE /BI0/OISALES_OFF,
    *      InfoObject: 0SHIP_TO Ship-To Party.
            SHIP_TO           TYPE /BI0/OISHIP_TO,
    *      InfoObject: 0SOLD_TO Sold-to party.
            SOLD_TO           TYPE /BI0/OISOLD_TO,
    *      InfoObject: 0VERSION Version.
            VERSION           TYPE /BI0/OIVERSION,
    *      InfoObject: 0VTYPE Value Type for Reporting.
            VTYPE           TYPE /BI0/OIVTYPE,
    *      InfoObject: 0DIVISION Division.
            DIVISION           TYPE /BI0/OIDIVISION,
    *      InfoObject: 0MATERIAL Material.
            MATERIAL           TYPE /BI0/OIMATERIAL,
    *      InfoObject: 0SHIP_POINT Shipping point.
            SHIP_POINT           TYPE /BI0/OISHIP_POINT,
    *      InfoObject: 0PAYER Payer.
            PAYER           TYPE /BI0/OIPAYER,
    *      InfoObject: 0DOC_CLASS Document category /Quotation/Order/Deliver
    *y/Invoice.
            DOC_CLASS           TYPE /BI0/OIDOC_CLASS,
    *      InfoObject: 0DEB_CRED Credit/debit posting (C/D).
            DEB_CRED           TYPE /BI0/OIDEB_CRED,
    *      InfoObject: 0SALESEMPLY Sales Representative.
            SALESEMPLY           TYPE /BI0/OISALESEMPLY,
    *      InfoObject: 0SUBTOT_1S Subtotal 1 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_1S           TYPE /BI0/OISUBTOT_1S,
    *      InfoObject: 0SUBTOT_2S Subtotal 2 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_2S           TYPE /BI0/OISUBTOT_2S,
    *      InfoObject: 0SUBTOT_3S Subtotal 3 from pricing proced.for conditi
    *on in stat. curr..
            SUBTOT_3S           TYPE /BI0/OISUBTOT_3S,
    *      InfoObject: 0SUBTOT_4S Subtotal 4 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_4S           TYPE /BI0/OISUBTOT_4S,
    *      InfoObject: 0SUBTOT_5S Subtotal 5 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_5S           TYPE /BI0/OISUBTOT_5S,
    *      InfoObject: 0SUBTOT_6S Subtotal 6 from pricing proced. for condit
    *ion in stat. curr..
            SUBTOT_6S           TYPE /BI0/OISUBTOT_6S,
    *      InfoObject: 0OPORDQTYBM Open orders quantity in base unit of meas
    *ure.
            OPORDQTYBM           TYPE /BI0/OIOPORDQTYBM,
    *      InfoObject: 0OPORDVALSC Net value of open orders in statistics cu
    *rrency.
            OPORDVALSC           TYPE /BI0/OIOPORDVALSC,
    *      InfoObject: 0QUANT_B Quantity in base units of measure.
            QUANT_B           TYPE /BI0/OIQUANT_B,
    *      InfoObject: 0DOCUMENTS No. of docs.
            DOCUMENTS           TYPE /BI0/OIDOCUMENTS,
    *      InfoObject: 0DOC_ITEMS Number of Document Items.
            DOC_ITEMS           TYPE /BI0/OIDOC_ITEMS,
    *      InfoObject: 0NET_VAL_S Net value in statistics currency.
            NET_VAL_S           TYPE /BI0/OINET_VAL_S,
    *      InfoObject: 0COST_VAL_S Cost in statistics currency.
            COST_VAL_S           TYPE /BI0/OICOST_VAL_S,
    *      InfoObject: 0GR_WT_KG Gross weight in kilograms.
            GR_WT_KG           TYPE /BI0/OIGR_WT_KG,
    *      InfoObject: 0NT_WT_KG Net weight in kilograms.
            NT_WT_KG           TYPE /BI0/OINT_WT_KG,
    *      InfoObject: 0VOLUME_CDM Volume in cubic decimeters.
            VOLUME_CDM           TYPE /BI0/OIVOLUME_CDM,
    *      InfoObject: 0HDCNT_LAST Number of Employees.
            HDCNT_LAST           TYPE /BI0/OIHDCNT_LAST,
    *      InfoObject: 0CRM_PROD Product.
            CRM_PROD           TYPE /BI0/OICRM_PROD,
    *      InfoObject: 0CP_CATEG Category.
            CP_CATEG           TYPE /BI0/OICP_CATEG,
    *      InfoObject: 0FISCYEAR Fiscal year.
            FISCYEAR           TYPE /BI0/OIFISCYEAR,
    *      InfoObject: 0BP_GRP BP: Business Partner Group (from Hierarchy).
            BP_GRP           TYPE /BI0/OIBP_GRP,
    *      InfoObject: 0STAT_CURR Statistics Currency.
            STAT_CURR           TYPE /BI0/OISTAT_CURR,
    *      InfoObject: 0BASE_UOM Base Unit of Measure.
            BASE_UOM           TYPE /BI0/OIBASE_UOM,
    *      InfoObject: 0PROD_CATEG Product Category.
            PROD_CATEG           TYPE /BI0/OIPROD_CATEG,
    *      InfoObject: 0VOLUME Volume.
            VOLUME           TYPE /BI0/OIVOLUME,
    *      InfoObject: 0VOLUMEUNIT Volume unit.
            VOLUMEUNIT           TYPE /BI0/OIVOLUMEUNIT,
    *      InfoObject: 0FISCPER3 Posting period.
            FISCPER3           TYPE /BI0/OIFISCPER3,
    *      InfoObject: 0SALES_DIST Sales District.
            SALES_DIST           TYPE /BI0/OISALES_DIST,
    *      InfoObject: 0BILL_TYPE Billing type.
            BILL_TYPE           TYPE /BI0/OIBILL_TYPE,
    *      InfoObject: 0MOVE_PLANT Receiving Plant/Issuing Plant.
            MOVE_PLANT           TYPE /BI0/OIMOVE_PLANT,
    *      InfoObject: 0SHIP_COND Shipping conditions.
            SHIP_COND           TYPE /BI0/OISHIP_COND,
    *      InfoObject: 0AB_RFBSK Status for Transfer to Accounting.
            AB_RFBSK           TYPE /BI0/OIAB_RFBSK,
    *      InfoObject: 0AB_FKSTO Indicator: Document Is Cancelled.
            AB_FKSTO           TYPE /BI0/OIAB_FKSTO,
    *      InfoObject: 0CUST_GRP5 Customer Group 5.
            CUST_GRP5           TYPE /BI0/OICUST_GRP5,
    *      InfoObject: ZCU_COND1 Constomer Condition Group 1.
            /BIC/ZCU_COND1           TYPE /BIC/OIZCU_COND1,
    *      InfoObject: ZCU_COND2 Customer Condition Group 2.
            /BIC/ZCU_COND2           TYPE /BIC/OIZCU_COND2,
    *      InfoObject: ZBATCHCD Batch Code.
            /BIC/ZBATCHCD           TYPE /BIC/OIZBATCHCD,
    *      InfoObject: 0BATCH Batch number.
            BATCH           TYPE /BI0/OIBATCH,
    *      InfoObject: ZBATCH Batch number.
            /BIC/ZBATCH           TYPE /BIC/OIZBATCH,
    *      Field: RECORD Data record number.
            RECORD           TYPE RSARECORD,
          END   OF _ty_s_TG_1_full.
    * Additional declaration for update rule interface
      DATA:
        MONITOR       type standard table of rsmonitor  WITH HEADER LINE,
        MONITOR_RECNO type standard table of rsmonitors WITH HEADER LINE,
        RECORD_NO     LIKE SY-TABIX,
        RECORD_ALL    LIKE SY-TABIX,
        SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS.
    * global definitions from update rules
    * TABLES: ...
    DATA: IN    TYPE F,
          OUT   TYPE F,
          DENOM TYPE F,
          NUMER TYPE F.
    * Def. of 'credit-documents': following doc.categ. are 'credit docs'
    *   reversal invoice (N)
    *   credit memo  (O)
    *   internal credit memo (6)
    * Credit-documents are delivered with negative sign. Sign is switched
    * to positive to provide positive key-figures in the cube.
    * The combination of characteristics DE_CRED and DOC-CLASS provides
    * a comfortable way to distinguisch e.g. positive incoming orders or
    * order returns.
    * Def. der 'Soll-Dokumente': folgende Belegtypen sind 'Soll-Belege'
    *   Storno Rechnung (N)
    *   Gutschrift (O)
    *   Interne Verrechn. Gutschr. (6)
    * Soll-Dokumente werden mit negativem Vorzeichen geliefert. Um die Kenn-
    * zahlen positiv in den Cube zu schreiben, wird das Vorzeich. gedreht
    * Die Kombination der Merkmale DEB_CRED und DOC-CLASS gibt Ihnen die
    * Möglichkeit schnell z.B. zwischen Auftrags-Eingang oder Retouren zu
    * unterscheiden.
    DATA: DEB_CRED(3) TYPE C VALUE 'NO6'.
    FORM routine_0002
      TABLES
       P_MONITOR         structure rsmonitor
      CHANGING
        RESULT         TYPE _ty_s_TG_1_full-DOCUMENTS
        RETURNCODE     LIKE sy-subrc
        ABORT          LIKE sy-subrc
      RAISING
        cx_sy_arithmetic_error
        cx_sy_conversion_error.
    * init variables
    * fill the internal table "MONITOR", to make monitor entries
    CLEAR RESULT.
    RESULT = COMM_STRUCTURE-NO_INV.
    IF COMM_STRUCTURE-DOC_CATEG CA DEB_CRED.
       RESULT = RESULT * ( -1 ).
    ENDIF.
    RETURNCODE = 0.
      p_monitor[] = MONITOR[].
      CLEAR:
        MONITOR[].
    ENDFORM.                    "routine_0002
    FORM routine_0003
      TABLES
       P_MONITOR         structure rsmonitor
      CHANGING
        RESULT         TYPE _ty_s_TG_1_full-DEB_CRED
        RETURNCODE     LIKE sy-subrc
        ABORT          LIKE sy-subrc
      RAISING
        cx_sy_arithmetic_error
        cx_sy_conversion_error.
    * init variables
    * fill the internal table "MONITOR", to make monitor entries
      IF COMM_STRUCTURE-DOC_CATEG CA DEB_CRED.
        RESULT = 'C'.
      ELSE.
        RESULT = 'D'.
      ENDIF.
      RETURNCODE = 0.
      p_monitor[] = MONITOR[].
      CLEAR:
        MONITOR[].
    ENDFORM.   
    Error:
    E:Field "COMM_STRUCTURE-NO_INV" is unknown. It is neither in one of the
    specified tables nor defined by a "DATA" statement. "DATA" statement.
    communication structure chaged to sours fields but no uses , please suggest how can i proceed , Thanks in Advance immediate replay
    Thanks & Regards
    Ramesh G

    Hi Gareth,
    You have two options:
    1. Transport from BW 3.1 to BI 7.0. You'll need to create a transport route between both systems. This may cause you some troubles in the future when you want to modify the objects you transported.
    2. As there are few objects, you can use XML export utility from Transport connection. There, you make an XML file with the objects you need to transport. One thing that you may take care of, in this option, is that the business content objects you are exporting need to be activated in the destination system. Another problem is that querys are not exported.
    Since it's only a cube, maybe you can create the objects manually. Look that in BI 7.0 there are several new functionalities, i don't know how transport or xml export would work.
    Hope this helps.
    Regards,
    Diego

  • How to run just the data flow

    We are in the process of converting from one ETL tool to Data Services 3.0.  I have a workflow with several data flows underneath it and have to make a change to the data flow.  Is there a way to run just the dataflow without executing the whole job?  For example, I just want to run DF4 below.
    WF --> DF1 --> DF2 --> DF3 --> DF4 --> DF5
    Thanks,
    Dan

    Although this topic is solved, just a few clarifications:
    When you create a new job and drag 'n drop the workflow/dataflow into it, you are not creating a second object (class), you are calling the object a second time (instantiate). It is like a BASIC program. I created a sub-procedure and called in DF1. In my filesystem I can see that sub-procedure as a seperate file in the folder object_library -> dataflow_procedures.
    Then I have two main programs, both just calling that sub-procedure with "gosub DF1".
    In the job you see the calls, when you drill into DF1 you see its definition and what object it calls.
    Using a conditional is okay but not adviced in my opinion. You are losing too much, the handling is complex and development takes a few seconds longer. my two cents only.
    https://wiki.sdn.sap.com/wiki/display/BOBJ/Testing

  • Change of sign when running the data management package

    Hi All,
    I have written a script logic for copying my Budget member under category dimension to say another member say "XYZ" under category dimension and this is for my Member Sales under ACcount dimension having acctype as inc.
    When i run the logic thru default i am getting the correct value. however if i run the same thru data manager package- my sign changes.
    Eg. 100000 is copied as -100000 and -50000 is copied as 50000.
    Please help me out
    Regards
    Navin

    Hi Navin,
    CREDITPOSITIVE= YES | NO
    Default: YES
    If No, all amounts referring to an ACCOUNT type (LEQ, INC) will have their signs reversed.
    Since you are referring to the member account sales as INC the default logic considers the sign as reversed.
    Please look at this link which helps you to understand
    Significance of Account Type EXP,LEQ,INC,AST in BPC- Need some valid reason
    Hope this helps.
    Regards,
    Sanjeev

  • My performance is very slow when I run graphs. How do I increase the speed at which I can do other things while the data is being updated and displayed on the graphs?

    I am doing an an aquisition and displaying the data on graphs. When I run the program it is slow. I think because I have the number of scans to read associated with my scan rate. It takes the number of seconds I want to display on the chart times the scan rate and feeds that into the number of samples to read at a time from the AI read. The problem is that it stalls until the data points are aquired and displayed so I cannot click or change values on the front panel until the updates occur on the graph. What can I do to be able to help this?

    On Fri, 15 Aug 2003 11:55:03 -0500 (CDT), HAL wrote:
    >My performance is very slow when I run graphs. How do I increase the
    >speed at which I can do other things while the data is being updated
    >and displayed on the graphs?
    >
    >I am doing an an aquisition and displaying the data on graphs. When I
    >run the program it is slow. I think because I have the number of
    >scans to read associated with my scan rate. It takes the number of
    >seconds I want to display on the chart times the scan rate and feeds
    >that into the number of samples to read at a time from the AI read.
    >The problem is that it stalls until the data points are aquired and
    >displayed so I cannot click or change values on the front panel until
    >the updates occur on the graph. What can I do to be a
    ble to help
    >this?
    It may also be your graphics card. LabVIEW can max the CPU and you
    screen may not be refreshing very fast.
    --Ray
    "There are very few problems that cannot be solved by
    orders ending with 'or die.' " -Alistair J.R Young

  • Hi am trying to save Data into a write to measurement file vi using a NI PXI 1042Q with a real time mode but it is not working but when i run it with uploading it into the PXI it save in to the file

    Hi am trying to save Data into a write to measurement file vi using a NI PXI 1042Q and DAQ NI PXI-6229 with a real time mode but it is not working but when i run it without uploading it into the PXI it save in to the file please find attached my vi
    Attachments:
    PWMs.vi ‏130 KB

     other problem is that the channel DAQmx only works at real time mode not on stand alone vi using Labview 8.2 and Real time 8.2

  • How can I save my data and the date,the time into the same file when I run this VI at different times?

    I use a translation stage for the experiment.For each user in the lab the stage position (to start an experiment) is different.I defined one end of the stage as zero. I want to save the position , date and time of the stage with respect to zero.I want all these in one file, nd everytime I run it it should save to the same file, like this:
    2/12/03 16:04 13567
    2/13/03 10:15 35678
    So I will track the position from day to day.If naybody helps, I appreciate it.Thanks.

    evolution wrote in message news:<[email protected]>...
    > How can I save my data and the date,the time into the same file when
    > I run this VI at different times?
    >
    > I use a translation stage for the experiment.For each user in the lab
    > the stage position (to start an experiment) is different.I defined one
    > end of the stage as zero. I want to save the position , date and time
    > of the stage with respect to zero.I want all these in one file, nd
    > everytime I run it it should save to the same file, like this:
    > 2/12/03 16:04 13567
    > 2/13/03 10:15 35678
    >
    > So I will track the position from day to day.If naybody helps, I
    > appreciate it.Thanks.
    Hi,
    I know the function "write to spreadsheet file.vi"
    can append the data
    to file. You can use the "concatenate strings" to display the date,
    time as well as data... Hope this help.
    Regards,
    celery

  • How can I do to acquire and save date in the same time and in the same file when I run continual my VI without interrupti​on.

    I've attached a VI that I am using to acquire amplitude from Spectrum analyzerse. I tried to connect amplitude ouput to the VI Write Characters To File.vi and Write to Spreadsheet File.vi. Unfortunately when I run continual this VI without interruption, labview ask me many time to enter a new file name to save a new value.
    So, How can I do to aquire and save date in the same time and in the same file when I run continual my VI for example during 10 min.
    Thank you in advance.
    Regards,
    Attachments:
    HP8563E_Query_Amplitude.vi ‏37 KB

    Hi,
    Your VI does work perfectly. Unfortunately this not what I want to do. I've made error in my last comment. I am so sorry for this.
    So I explain to you again what I want to do exactly. I want to acquire amplitude along road by my vehicle. I want to use wheel signal coming from vehicle to measure distance along road. Then I acquire 1 amplitude each 60 inches from spectrum analyzer.
    I acquire from PC parallel port a coded wheel signal coming from vehicle (each period of the signal corresponds to 12 Inches). Figure attached shows the numeric signal coming from vehicle, and the corresponding values “120” and “88” that I can read from In Port vi.
    So I want to acquire 1 time amplitude from spectrum analyser each 5
    period of the signal that I am acquiring from parallel port.
    So fist I have to find how can I count the number of period from reading the values “120” and “88” that I am acquiring from In Port (I don’t know the way to count a number of period from reading values “120” and “88”).
    Here is a new algorithm.
    1) i=0 (counter: number of period)
    2) I read value from In Port
    3) If I acquire a period
    i= i+1 (another period)
    4) If i is multiple of 5 (If I read 5 period)
    acquire 1 time amplitude and write to the same
    file this amplitude and the corresponding distance
    Distance = 12*i). Remember each period of signal
    Corresponds to 12 Inches).i has to take these
    values: 5,10,15,20,25,35,40,45,50,55,60............
    5) Back to 2 if not stop.
    Thank you very much for helping me.
    Regards,
    Attachments:
    Acquire_Amplitude_00.vi ‏59 KB
    Figure_Algorithm.doc ‏26 KB

  • Getting - ORA-01403: no data found - When I Run expdp

    Oracle database version 11.1.0.7.0 on Windows 32-bit.
    When I run:
    expdp system schemas=vms dumpfile=data_pump_dir:vms_bjw%u.dmp logfile=data_pump_dir:vms_bjw.log parallel=2
    Connected to: Oracle Database 11g Enterprise Edition Release *11.1.0.7.0* - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-31626: job does not exist
    ORA-31637: cannot create job SYS_EXPORT_SCHEMA_01 for user SYSTEM
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 736
    ORA-39080: failed to create queues "" and "" for Data Pump job
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPC$QUE_INT", line 1665
    ORA-01403: no data found
    I turned on tracing like this:
    1.) Set the Error Stack 1403 to know exactly which SQL is failing:-
    SQL> alter system set events '1403 trace name errorstack level 3';
    Re-ran the datapump export to reproduce the error then disable the events.
    SQL> alter system set events '1403 trace name errorstack off';
    The trace file, oddly enough, shows that this is the failing SQL:
    Error Stack Dump
    ORA-01403: no data found
    Current SQL Statement for this session (sql_id=3n58uzvnuw2hj) -----
    SELECT VALUE FROM V$NLS_PARAMETERS WHERE PARAMETER = 'NLS_CHARACTERSET'
    Under that, I get the call stack.
    I ran this SQL manually and I get:
    SQL> SELECT VALUE FROM V$NLS_PARAMETERS WHERE PARAMETER = 'NLS_CHARACTERSET';
    VALUE
    AL32UTF8
    Any ideas as to why I'm getting the ORA-01403 when I run expdp?
    Thank you

    Hello,
    Your error looks like what is descibed into the Metalink Note 345198.1.
    But this note apply to 10.1.0.4 not to 11.1.0.7.
    So, you should open an issue to My Oracle Support.
    Best regards,
    Jean-Valentin

  • HT1338 Mac forgets time and date when battery runs flat

    My Mac Book Pro forgets time, date and network password when battery runs flat. My software is up to date. The computer is about three years, but surely it shouldn't be necessary to change battery every three years? I have seen others with problem, but no answer yet.

    Resetting your Mac's PRAM and NVRAM
    If resetting pram does resolve your issue - Apple/About This Mac/More Info/Hardware/Power/Battery Information - copy and paste the information here. Do not include any serial numbers.

  • Date format becomes incorrect when you run check option on portal

    Hi,
    We have a report which has Date variable in selection screen. When we enter selection as interval and then when we click on Check option before executing the report, the date format becomes garbage.
    Here is an example:
    This is what I enter in Date variable :
    Date : 01/01/2001-01/01/2010
    After that when I click on CHECK option, the system automatically converts the first date into garbage value i.e. 01/01/2001 will be converted to 1//20/01/0
    This only happens when we run report in Portal 7.0 When we try to run similar report using default web template of BW 3.5, it works fine.
    Has anybody faced similar problem ? We are on Portal Java SP13 with Patch 6.
    Thanks,
    Parin Gandhi.

    Thanks Frank for a quick reply.
    I have used the SimpleDateFormat for formatting with the pattern as, "MM/dd/yyyy" and it is working fine if user enters 4 digit year.
    However, for a entry with 2 digits for a year e.g. 04/10/07 when i parse the date using SimpleDateFormat,it is giving me date as 04/10/0007.
    Is there a way by means of which i could force the user to enter 4 digits for the year?
    I have tried using dateStyle attribute of <af:convertDateTime> tag with 'shortish' as its value, it converts a 2 digit year into 4digit on the browser side & displays on the screen e.g. if user enters 04/10/07 it will appear as 04/10/2007 on the screen however, in value change listener method, i could see date as 04/10/0007.
    Any help in this regard is highly appreciated.
    Thanks in advance,
    Shriniwas

Maybe you are looking for

  • Some SAP R/3 Tables are Missing

    Hi Friends, I made a connection between SAP R/3 and Crystal Reports.  When I am trying to see the tables of SAP R/3 I am not able to see all the tables only few tables are visible.  For example in Standard table I am able to see only LIKP table, but

  • XML payload in RESTful webservice

    Hi all, Can anyone please suggest on how to call the RESTful webservice (which is used by our client) by passing an XML payload from a PL SQL stored procedure? It is for the authentication purpose. And the username and password has to be passed as th

  • HT4623 how do i restore to a previous ( before iso9 ) backup

    how do i restore to a previous ( previous iso8 ) back up. My ipad is not working and gets stuck on apps.  I DO NOT like new update

  • Ipod disabled and forgot passcode

    My granddaughter forgot her passcode on her ipod.  It is now disabled and says to connect to iTunes.  However, we don't have the itunes it was originally set up on.  After installing itunes on a new computer, it says that we need the passcode for the

  • Problem after run report with spreadsheet

    hi evry one ther an a problem after run report with spreadsheet . after runing iexplorar open and close fast no thing else when i check the job from appserver job success with spreadsheet dis format. what shall i do. thanks.