How to control data load in Info-Package using ABAP?

Hello Gurus:
I am trying to resolve couple of issues. 
1.  I need to load data DAILY Full Load into a Planning cube.  I have the 0NETDUEDATE   for selection in the
     Info-package.  I am getting data from anothe base A/R cube.  I want to Load Daily ONLY those records
     for which the "0NETDUEDATE" is Greater than "System Date".  I think the logic would be IF 0NETDUEDAT
     .GT. SYS-DATUM, bruing the record. What should be the statement in 'Result' for keeping the record? 
     (I don't know ABAP, so need help here...!)
2.  Similalry, before loading the data above, I want to delete all the existing records from Planning cube where
     0NETDUEDATE is GT System date (records loaded the previosu day need to be deleted as the amounts may
     have changed).  How can I achieve this selective deletion automatically  in PC??
Appreciate your feedback very much.
Thanks... SMaa

Hi Shruti,
1) If i understand your requirement correctly you need to load only data which has 0NETDATE greater then sydatum.
So you can do this in infopackage .Infopackage>Data selection->0NETDATE(field)-->type(colum).here you can selecy ABAP type(6).Now you create ABAP code in this.
Here in ABAP code you need to specify high and low range ,Also the relation character like GT or BT(between).
i guess some how ur code will look like this
$$ begin of global - insert your declaration only below this line  -
TABLES: ...
DATA:   ...
$$ end of global - insert your declaration only before this line   -
    InfoObject      = 0NETDUEDATE
    Fieldname       = NETDUEDATE
    data type       = NUMC
    length          = 0000010
    convexit        = PERI6
form compute_NETDUEDATE
  tables l_t_range structure rssdlrange
  changing p_subrc like sy-subrc.
      Insert source code to current selection field
$$ begin of routine - insert your code only below this line        -
data: l_idx like sy-tabix.
          l_idx = sy-tabix.
          l_t_range-low = sy-datum.
          l_t_range-option = 'GT'.
          modify l_t_range index l_idx.
          p_subrc = 0.
$$ end of routine - insert your code only before this line         -
endform.
i am not sure about coding but you can surely do it it is like specifying the ranges dynamically(you need to try it out)
(also as mentioned by others you can do this in traansformation routines)but infopackage routine you can directly write in PROD system.
2) And regarding the Selective deletion you donu2019t have nay process type in Process chain which does this
But there is function module u201C rsdri_infoprov_delete " which can be used to develop a program( I guess you must be having a existing ZProgram in ur system already available because this is used commonly.
Other wise you need to develop a ZProgram for this.
Thanks and Regards
Arun

Similar Messages

  • Data selection in info package using ABAP routine.

    Hi,
    when we are  scheduling the info package(Infp package name is like XYZ),we want to load past 6 months data from the current date(INFO OBJECT(0CREATEDON) BETWEEN (Sy-Datum - 6 months TO  Sy-Datum),in the selection tab in schedule i want to write the abap routine (type 6).please can any one have this type of code please sedn to me asap.
    this is very urgent.delivery is tomorrow.

    hi Lekha,
    try following code.
    seems you post the same question twice, one posting is sufficient. and please don't forget to reward helpful answers.
    data: l_idx like sy-tabix.
    read table l_t_range with key
    iobjnm = '0CREATEDON'.
    l_idx = sy-tabix.
    DELETE l_t_range
    WHERE iobjnm = '0CREATEDON'.
    L_t_RANGE-SIGN = 'I'.
    L_t_RANGE-OPTION = 'EQ'.
    if not work, try
    L_t_RANGE-OPTION = 'BT'.
    BT = between
    last 6 month
    L_t_RANGE-LOW = sy-datum - 180.
    current date
    L_t_RANGE-HIGH = sy-datum.
    append l_t_range.
    modify l_t_range index l_idx.
    p_subrc = 0.
    $$ end of routine - insert your code only before this line -
    endform.

  • How to set request number of info package using start routine

    Hi All,
    I have a specific requirement in which I need to upload only selective request nos from PSA to DSO.
    Any one can suggest how to set request number of info package using start routine or any other mode so that only selective request should proceed using DTP?
    Thanks
    Sangita

    You'll probably want to do this in a start routine.  It is probably not advisable to hard-code the technical name of the InfoPak.  They seem to remain the same upon transport, unlike DTP technical names which are locally generated.
    I would do a SQL select like this in a start routine, and then filter based on the results of the SQL.  You'll probably want to sort the results by date and time, both descending.  Or do a MAX in the SQL.
    Please let me know if this isn't clear.
    select q~rnr q~logdpid q~tdatum q~tuzeit
    from rsreqdone as q
      inner join rsldpio as p
        on p~logdpid = q~logdpid
    where p~objvers = 'A'
    and p~logdpid like 'ZPAK%'
    and p~source = 'your data source'

  • How to pass data between two internal sessions using ABAP memory?

    Hi,
    How to pass data between two internal sessions using ABAP memory?
    It would be fine if you could explain with an example.
    And also let me clear about the data passing between two main sessions and two external sessions with specific examples.
    Thanks.

    Hi ,
      check the example.
    Reading Data Objects from Memory
    To read data objects from ABAP memory into an ABAP program, use the following statement:
    Syntax
    IMPORT <f1> [TO <g 1>] <f 2> [TO <g 2>] ... FROM MEMORY ID <key>.
    This statement reads the data objects specified in the list from a cluster in memory. If you do not use the TO <g i > option, the data object <f i > in memory is assigned to the data object in the program with the same name. If you do use the option, the data object <f i > is read from memory into the field <g i >. The name <key> identifies the cluster in memory. It may be up to 32 characters long.
    You do not have to read all of the objects stored under a particular name <key>. You can restrict the number of objects by specifying their names. If the memory does not contain any objects under the name <key>, SY-SUBRC is set to 4. If, on the other hand, there is a data cluster in memory with the name <key>, SY-SUBRC is always 0, regardless of whether it contained the data object <f i >. If the cluster does not contain the data object <f i >, the target field remains unchanged.
    In this statement, the system does not check whether the structure of the object in memory is compatible with the structure into which you are reading it. The data is transported bit by bit. If the structures are incompatible, the data in the target field may be incorrect.
    PROGRAM SAPMZTS1.
    DATA TEXT1(10) VALUE 'Exporting'.
    DATA ITAB LIKE SBOOK OCCURS 10 WITH HEADER LINE.
    DO 5 TIMES.
      ITAB-BOOKID = 100 + SY-INDEX.
      APPEND ITAB.
    ENDDO.
    EXPORT TEXT1
           TEXT2 FROM 'Literal'
      TO MEMORY ID 'text'.
    EXPORT ITAB
      TO MEMORY ID 'table'.
    SUBMIT SAPMZTS2 AND RETURN.
    SUBMIT SAPMZTS3.
    The first part of this program is the same as the example in the section Saving Data Objects in Memory. In the example, the programs SAPMZTS1 and SAPMZTS2 are called using SUBMIT. You can create and maintain the programs called using the SUBMIT statement by double-clicking their names in the statement. For further information about the SUBMIT statement, refer to Calling Executable Programs (Reports)
    Example for SAPMZTS2:
    PROGRAM SAPMZTS2.
    DATA: TEXT1(10),
          TEXT3 LIKE TEXT1 VALUE 'Initial'.
    IMPORT TEXT3 FROM MEMORY ID 'text'.
    WRITE: / SY-SUBRC, TEXT3.
    IMPORT TEXT2 TO TEXT1 FROM MEMORY ID 'text'.
    WRITE: / SY-SUBRC, TEXT1.
    Example for SAPMZTS3:
    PROGRAM SAPMZTS3.
    DATA JTAB LIKE SBOOK OCCURS 10 WITH HEADER LINE.
    IMPORT ITAB TO JTAB FROM MEMORY ID 'table'.
    LOOP AT JTAB.
      WRITE / JTAB-BOOKID.
    ENDLOOP.
    The output is displayed on two successive screens. It looks like this:
    and
    The program SAPMZTS2 attempts to read a data object TEXT3 from the data cluster "text", which does not exist. TEXT3 therefore remains unchanged. The existing data object TEXT2 is placed in TEXT1. In both cases, SY-SUBRC is 0, since the cluster "text" contains data.
    The program SAPMZTS3 reads the internal table ITAB from the cluster "table" into the internal table JTAB. Both tables have the same structure, namely that of the ABAP Dictionary table SBOOK.
    Pls. reward if useful.....

  • How to tune data loading time in BSO using 14 rules files ?

    Hello there,
    I'm using Hyperion-Essbase-Admin-Services v11.1.1.2 and the BSO Option.
    In a nightly process using MAXL i load new data into one Essbase-cube.
    In this nightly update process 14 account-members are updated by running 14 rules files one after another.
    These rules files connect 14 times by sql-connection to the same oracle database and the same table.
    I use this procedure because i cannot load 2 or more data fields using one rules file.
    It takes a long time to load up 14 accounts one after other.
    Now my Question: How can I minimise this data loading time ?
    This is what I found on Oracle Homepage:
    What's New
    Oracle Essbase V.11.1.1 Release Highlights
    Parallel SQL Data Loads- Supports up to 8 rules files via temporary load buffers.
    In an Older Thread John said:
    As it is version 11 why not use parallel sql loading, you can specify up to 8 load rules to load data in parallel.
    Example:
    import database AsoSamp.Sample data
    connect as TBC identified by 'password'
    using multiple rules_file 'rule1','rule2'
    to load_buffer_block starting with buffer_id 100
    on error write to "error.txt";
    But this is for ASO Option only.
    Can I use it in my MAXL also for BSO ?? Is there a sample ?
    What else is possible to tune up nightly update time ??
    Thanks in advance for every tip,
    Zeljko

    Thanks a lot for your support. I’m just a little confused.
    I will use an example to illustrate my problem a bit more clearly.
    This is the basic table, in my case a view, which is queried by all 14 rules files:
    column1 --- column2 --- column3 --- column4 --- ... ---column n
    dim 1 --- dim 2 --- dim 3 --- data1 --- data2 --- data3 --- ... --- data 14
    Region -- ID --- Product --- sales --- cogs ---- discounts --- ... --- amount
    West --- D1 --- Coffee --- 11001 --- 1,322 --- 10789 --- ... --- 548
    West --- D2 --- Tea10 --- 12011 --- 1,325 --- 10548 --- ... --- 589
    West --- S1 --- Tea10 --- 14115 --- 1,699 --- 10145 --- ... --- 852
    West --- C3 --- Tea10 --- 21053 --- 1,588 --- 10998 --- ... --- 981
    East ---- S2 --- Coffee --- 15563 --- 1,458 --- 10991 --- ... --- 876
    East ---- D1 --- Tea10 --- 15894 --- 1,664 --- 11615 --- ... --- 156
    East ---- D3 --- Coffee --- 19689 --- 1,989 --- 15615 --- ... --- 986
    East ---- C1 --- Coffee --- 18897 --- 1,988 --- 11898 --- ... --- 256
    East ---- C3 --- Tea10 --- 11699 --- 1,328 --- 12156 --- ... --- 9896
    Following 3 out of 14 (load-) rules files to load the data columns into the cube:
    Rules File1:
    dim 1 --- dim 2 --- dim 3 --- sales --- ignore --- ignore --- ... --- ignore
    Rules File2:
    dim 1 --- dim 2 --- dim 3 --- ignore --- cogs --- ignore --- ... --- ignore
    Rules File14:
    dim 1 --- dim 2 --- dim 3 --- ignore --- ignore --- ignore --- ... --- amount
    Is the upper table design what GlennS mentioned as a "Data" column concept which only allows a single numeric data value ?
    In this case I cant tag two or more columns as “Data fields”. I just can tag one column as “Data field”. Other data fields I have to tag as “ignore fields during data load”. Otherwise, when I validate the rules file, an Error occurs “only one field can contain the Data Field attribute”.
    Or may I skip this error massage and just try to tag all 14 fields as “Data fields” and “load data” ?
    Please advise.
    Am I right that the other way is to reconstruct the table/view (and the rules files) like follows to load all of the data in one pass:
    dim 0 --- dim 1 --- dim 2 --- dim 3 --- data
    Account --- Region -- ID --- Product --- data
    sales --- West --- D1 --- Coffee --- 11001
    sales --- West --- D2 --- Tea10 --- 12011
    sales --- West --- S1 --- Tea10 --- 14115
    sales --- West --- C3 --- Tea10 --- 21053
    sales --- East ---- S2 --- Coffee --- 15563
    sales --- East ---- D1 --- Tea10 --- 15894
    sales --- East ---- D3 --- Coffee --- 19689
    sales --- East ---- C1 --- Coffee --- 18897
    sales --- East ---- C3 --- Tea10 --- 11699
    cogs --- West --- D1 --- Coffee --- 1,322
    cogs --- West --- D2 --- Tea10 --- 1,325
    cogs --- West --- S1 --- Tea10 --- 1,699
    cogs --- West --- C3 --- Tea10 --- 1,588
    cogs --- East ---- S2 --- Coffee --- 1,458
    cogs --- East ---- D1 --- Tea10 --- 1,664
    cogs --- East ---- D3 --- Coffee --- 1,989
    cogs --- East ---- C1 --- Coffee --- 1,988
    cogs --- East ---- C3 --- Tea10 --- 1,328
    discounts --- West --- D1 --- Coffee --- 10789
    discounts --- West --- D2 --- Tea10 --- 10548
    discounts --- West --- S1 --- Tea10 --- 10145
    discounts --- West --- C3 --- Tea10 --- 10998
    discounts --- East ---- S2 --- Coffee --- 10991
    discounts --- East ---- D1 --- Tea10 --- 11615
    discounts --- East ---- D3 --- Coffee --- 15615
    discounts --- East ---- C1 --- Coffee --- 11898
    discounts --- East ---- C3 --- Tea10 --- 12156
    amount --- West --- D1 --- Coffee --- 548
    amount --- West --- D2 --- Tea10 --- 589
    amount --- West --- S1 --- Tea10 --- 852
    amount --- West --- C3 --- Tea10 --- 981
    amount --- East ---- S2 --- Coffee --- 876
    amount --- East ---- D1 --- Tea10 --- 156
    amount --- East ---- D3 --- Coffee --- 986
    amount --- East ---- C1 --- Coffee --- 256
    amount --- East ---- C3 --- Tea10 --- 9896
    And the third way is to adjust the essbase.cfg parameters DLTHREADSPREPARE and DLTHREADSWRITE (and DLSINGLETHREADPERSTAGE)
    I just want to be sure that I understand your suggestions.
    Many thanks for awesome help,
    Zeljko

  • Initialize Delta Load in Info Package

    Hi,
    I want to create 2 info packages for my load from datasource.
    One with setting " Initialize Delta load with data transfer " in info package for my first load.
    Second info package should be with " Delta Lad " which i am not finding in the second info package i am creating. Can you please tell me how can i get this setting as i am getting only Full or Intialize delta load only.
    Regards
    Syed

    Make sure the following:
    1) DataSource is Delta Enabled
    2) An Init (or an Early Init) has already been done successfully. You should see the entry in RSA7 of the Source ERP
    You should then be able to get the Delta option in the InfoPackage

  • Example Code for Routine at Data Target in Info Package.

    Hi,
    Can any one send the example code for writing the routines at Data Target of Info Package for Deleting the Request from Info Cube after Update.
    Thanks in Advance,
    Regards,
    srinivas

    Hi All,
    I want to know some examples for guidance in writing the Abap Routine at Data Target of Info Package for Deleting the Request from Info Cube after Update.
    In the Data Targets of Info Package for Info Cubes we can delete the previous request based on the Conditions, exceptions and also based on the Abap Routine. I just want to know how do we write the Routine.
    This option is adjacent to the Maintain to Data Target in Data Targets of Info Package and will available for Info cube only.
    Thanks in Advance,
    Regards,
    srinivas

  • How to control data

    Hi,
    When 2 SKUs are removed how to control data.
    Regards,
    krishna.

    Hi Krishna,
    The following is the process for controlling data during
    SKU deletions
    1) Once the SKU deletion confirmation is received, the
    corresponding statistical/concensus forecast from DP
    has to be deleted
    2) The CVC's has to be deleted and time series needs to
    be updated
    3) The previous forecast if released to SNP has to be
    deleted using transaction /sapapo/rlcdel
    4) The demand generated in SNP/PPDS generated based
    on previous planning run needs to be deleted
    5) Integration models needs to be blocked for those SKUs
    6) If there are any confirmed/firmed production orders, those
    has to be deleted
    7) Communication to be given to concerned plants not to
    make that product
    Regards
    R. Senthil Mareeswaran.

  • Problem while scheduling info package using HR datasource.

    hi experts,
    while scheduling the info package i am getting the following error.
    Delete init. request
    REQU_D340A67OSEQ9A6Q2B20X9SIOM before running
    init. again with same selection
    I think by deleting the setup it will work...
    am i right or is there any other solution for this problem..
    please reply soon.
    its too urgent.
    thanks n regards.
    uday shankar.

    Hi,
    You are correct there is no setup table concept in HR extractors.
    But there is a cache. Usually HR extractions will be full load. But it also supports delta. There is no setup table concept in HR.Once you perform your init (with or without Data transfer your delta path is set and delta will follow else can go for Full loads alone. It will extract data from the relevant tables like generic extractors. The data will be filtered based on the selection parameters.
    Except for payroll and time management rest all sub-functional areas supports only full load.
    In case of full loads, old data needs to be deleted to avoid duplicated records in the target.
    Auditing Information Technical name: 0HR_PY_PP_1
    http://help.sap.com/saphelp_nw70/helpdata/EN/c7/532d7835db9a41b744aea58ed699ed/frameset.htm
    Delta Update
    This DataSource supports delta updates. The source system fills table HRMS_BIW_PPDELTA with the date and time of the first data transfer (initialization) and the date and time of all subsequent data transfers. This information enables the system to perform a delta transfer. It only selects posting runs with the above statuses whose last changed on date is after the date in table HRMS_BIW_PPDELTA.
    If you use DataSources 0HR_PY_PP_2 (Posting Documents) and 0HR_PY_PP_1 (Auditing Information), you must transfer data to BW from both DataSources using the same selection criteria.
    Also
    http://help.sap.com/saphelp_nw04/helpdata/en/2a/77eb3cad744026e10000000a11405a/frameset.htm
    https://websmp201.sap-ag.de/~sapdownload/011000358700004478862001E
    /people/swapna.gollakota/blog/2008/01/14/one-stage-stop-to-know-all-about-bw-extractors-part1
    step by step
    https://websmp201.sap-ag.de/~sapdownload/011000358700004478862001E
    HR extraction:
    The HR module enables customers to effectively manage information about the people in their organization, and to integrate that information with other SAP modules and external systems
    HR broadly has the following modules:
    PA (Personnel Administration) and Organization Management
    Personnel Development
    Payroll Accounting
    Time Management
    Compensation
    Benefits
    Training and Events
    The Personnel Administration (PA) sub module helps employers to track employee master data, work schedules, salary and benefits information. Personnel Development (PD) functionality focuses on employees' skills, qualifications and career plans. Finally, the Time Evaluation and Payroll sub modules process attendance and absences, gross salary and tax calculations, and payments to employees and third-party vendors
    HR delivers a rich set of business content objects that covers all HR sub-functional areas.
    How the data extraction happens:
    Before getting into how the data gets populated into HR info cube
    Let's understand the term info type
    "An info type is a collection of logical and/or business-related characteristics of an object or person"
    Here the data will be extracted from an info type (PA, PD, time management etc) and for few other applications it is from the cluster tables (Payroll, compensation etc.)
    HR is basically master data centric because it is always related to people related Info Objects, such as Employee, Person. In most of the cases HR master data is defined as Time Dependent to enable historical evaluation. HR R/3 system records a specific period of validity for each Info type.
    Procedure to extract the HR data:
    Activate Data Sources in the Source system (R/3)
    Replicate Data Sources in BW system:
    Activate business contents in BW.
    Populate HR cubes with data by scheduling info packages.
    Note: Master Data should be loaded first
    Except for payroll and time management rest all sub-functional areas supports only full load.
    In case of full loads, old data needs to be deleted to avoid duplicated records in the target.
    Check the below links for more information
    Technical Considerations in Global SAP BW/HR Implementations
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/505351fe-ec8c-2910-c5b5-a43bbf53f6fc
    BW/HR Implementation
    https://websmp201.sap-ag.de/~sapdownload/011000358700004478862001E
    Hope this helps.
    Thanks,
    JituK

  • How to define data-sources-alias.xml to use Oracle data source

    Hi,
    I created Oracle jdbc Data Source named "MYDS" in Visual Admin.  I added the alias = MYDS_ALIAS.  I set the Initial Connection to 2.  On the Monitor tab, I see a green line running across the screen.  The datasource MYDS is connected to the Oracle.
    Now, I would like to define "data-sources-alias.xml" to use MYDS.
    When creating data-source-aliases.xml it creates the alias by default on ${com.sap.datasource.default} which is the default DB.
    I followed the link --> How to point data-source-aliases.xml to another Datasource
    and my data-sources-alias.xml is now:
    <data-source-aliases>
         <aliases>
              <data-source-name>$</data-source-name>
              <alias>BRANCH50DS</alias>
         </aliases>
    </data-source-aliases>
    The "EAR generatation has finished successfully", but when "Deploy to J2EE Engine", I got the error below:
    1) How to define data-sources-alias.xml to use Oracle data source?
    2) I found the link (SDA Deployment error) talking about "sda-dd.xml".  How to create "sda-dd.xml" and make it use substitution variables "MYDS"
    Dec 20, 2005 3:53:41 PM /userOut/deploy (com.sap.ide.eclipse.sdm.threading.DeployThreadManager) [Thread[Deploy Thread,5,main]] INFO:
    [003]Additional log information about the deployment
    <!LOGHEADER[START]/>
    <!HELP[Manual modification of the header may cause parsing problem!]/>
    <!LOGGINGVERSION[1.5.3.7181 - 630_SP]/>
    <!NAME[C:\usr\sap\J2E\JC00\SDM\program\log\sdmcl20051220205339.log]/>
    <!PATTERN[sdmcl20051220205339.log]/>
    <!FORMATTER[com.sap.tc.logging.TraceFormatter(%24d %s: %m)]/>
    <!ENCODING[Cp1252]/>
    <!LOGHEADER[END]/>
    Dec 20, 2005 3:53:39 PM  Info: -
    Starting deployment -
    Dec 20, 2005 3:53:39 PM  Info: Loading selected archives...
    Dec 20, 2005 3:53:39 PM  Info: Loading archive 'C:\usr\sap\J2E\JC00\SDM\program\temp\temp856850022Ear.ear'
    Dec 20, 2005 3:53:40 PM  Info: Selected archives successfully loaded.
    Dec 20, 2005 3:53:40 PM  Info: Actions per selected component:
    Dec 20, 2005 3:53:40 PM  Info: Update: Selected development component '50022Ear'/'sap.com'/'localhost'/'2005.12.20.15.53.28' updates currently deployed development component '50022Ear'/'sap.com'/'localhost'/'2005.12.19.18.39.12'.
    Dec 20, 2005 3:53:40 PM  Info: Saved current Engine state.
    Dec 20, 2005 3:53:40 PM  Info: Error handling strategy: OnErrorStop
    Dec 20, 2005 3:53:40 PM  Info: Update strategy: UpdateAllVersions
    Dec 20, 2005 3:53:40 PM  Info: Starting: Update: Selected development component '50022Ear'/'sap.com'/'localhost'/'2005.12.20.15.53.28' updates currently deployed development component '50022Ear'/'sap.com'/'localhost'/'2005.12.19.18.39.12'.
    Dec 20, 2005 3:53:40 PM  Info: SDA to be deployed: C:\usr\sap\J2E\JC00\SDM\root\origin\sap.com\50022Ear\localhost\2005.12.20.15.53.28\temp856850022Ear.ear
    Dec 20, 2005 3:53:40 PM  Info: Software type of SDA: J2EE
    Dec 20, 2005 3:53:40 PM  Info: ***** Begin of SAP J2EE Engine Deployment (J2EE Application) *****
    Dec 20, 2005 3:53:41 PM  Info: Begin of log messages of the target system:
    05/12/20 15:53:40 -  ***********************************************************
    05/12/20 15:53:41 -  Start updating EAR file...
    05/12/20 15:53:41 -  start-up mode is lazy
    05/12/20 15:53:41 -  com.sap.engine.deploy.manager.MissingSubstitutionException: Missing substitution value for variable [MYDS].
                              at com.sap.engine.deploy.manager.DeployManagerImpl.makeTempEar(DeployManagerImpl.java:3727)
                              at com.sap.engine.deploy.manager.DeployManagerImpl.makeNewEar1(DeployManagerImpl.java:3695)
                              at com.sap.engine.deploy.manager.DeployManagerImpl.deployUpdateAction(DeployManagerImpl.java:523)
                              at com.sap.engine.deploy.manager.DeployManagerImpl.update(DeployManagerImpl.java:512)
                              at com.sap.sdm.serverext.servertype.inqmy.extern.EngineApplOnlineDeployerImpl.performDeployment(EngineApplOnlineDeployerImpl.java:196)
                              at com.sap.sdm.serverext.servertype.inqmy.extern.EngineDeployerImpl.deploy(EngineDeployerImpl.java:96)
                              at com.sap.sdm.serverext.servertype.inqmy.EngineProcessor.executeAction(EngineProcessor.java:224)
                              at com.sap.sdm.app.proc.deployment.impl.PhysicalDeploymentActionExecutor.execute(PhysicalDeploymentActionExecutor.java:60)
                              at com.sap.sdm.app.proc.deployment.impl.DeploymentActionImpl.execute(DeploymentActionImpl.java:186)
                              at com.sap.sdm.app.proc.deployment.controllers.internal.impl.DeploymentExecutorImpl.execute(DeploymentExecutorImpl.java:46)
                              at com.sap.sdm.app.proc.deployment.states.eventhandler.ExecuteDeploymentHandler.executeAction(ExecuteDeploymentHandler.java:83)
                              at com.sap.sdm.app.proc.deployment.states.eventhandler.ExecuteDeploymentHandler.handleEvent(ExecuteDeploymentHandler.java:60)
                              at com.sap.sdm.app.proc.deployment.states.StateBeforeNextDeployment.processEvent(StateBeforeNextDeployment.java:127)
                              at com.sap.sdm.app.proc.deployment.states.InstContext.processEventServerSide(InstContext.java:73)
                              at com.sap.sdm.app.proc.deployment.states.InstContext.processEvent(InstContext.java:59)
                              at com.sap.sdm.app.sequential.deployment.impl.DeployerImpl.doPhysicalDeployment(DeployerImpl.java:127)
                              at com.sap.sdm.app.sequential.deployment.impl.DeployerImpl.deploy(DeployerImpl.java:96)
                              at com.sap.sdm.apiimpl.local.DeployProcessorImpl.deploy(DeployProcessorImpl.java:67)
                              at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                              at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
                              at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
                              at java.lang.reflect.Method.invoke(Method.java:324)
                              at com.sap.sdm.is.cs.remoteproxy.server.impl.RemoteProxyServerImpl.requestRemoteCall(RemoteProxyServerImpl.java:127)
                              at com.sap.sdm.is.cs.remoteproxy.server.impl.RemoteProxyServerImpl.process(RemoteProxyServerImpl.java:38)
                              at com.sap.sdm.apiimpl.remote.server.ApiClientRoleCmdProcessor.process(ApiClientRoleCmdProcessor.java:81)
                              at com.sap.sdm.is.cs.session.server.SessionCmdProcessor.process(SessionCmdProcessor.java:67)
                              at com.sap.sdm.is.cs.cmd.server.CmdServer.execCommand(CmdServer.java:76)
                              at com.sap.sdm.client_server.launch.ServerLauncher$ConnectionHandlerImpl.handle(ServerLauncher.java:280)
                              at com.sap.sdm.is.cs.ncserver.NetCommServer.serve(NetCommServer.java:43)
                              at com.sap.sdm.is.cs.ncwrapper.impl.ServiceWrapper.serve(ServiceWrapper.java:39)
                              at com.sap.bc.cts.tp.net.Worker.run(Worker.java:50)
                              at java.lang.Thread.run(Thread.java:534)
    05/12/20 15:53:41 -  ***********************************************************
    Dec 20, 2005 3:53:41 PM  Info: End of log messages of the target system.
    Dec 20, 2005 3:53:41 PM  Info: ***** End of SAP J2EE Engine Deployment (J2EE Application) *****
    Dec 20, 2005 3:53:41 PM  Error: Aborted: development component '50022Ear'/'sap.com'/'localhost'/'2005.12.20.15.53.28':
    Caught exception during application deployment from SAP J2EE Engine's deploy API:
    com.sap.engine.deploy.manager.MissingSubstitutionException: Missing substitution value for variable [MYDS].
    (message ID: com.sap.sdm.serverext.servertype.inqmy.extern.EngineApplOnlineDeployerImpl.performAction(DeploymentActionTypes).DMEXC)
    Dec 20, 2005 3:53:41 PM  Info: J2EE Engine is in same state (online/offline) as it has been before this deployment process.
    Dec 20, 2005 3:53:41 PM  Error: -
    At least one of the Deployments failed -

    Hi,
    If you remove the $ and braces (as shown below) it will work.
    <data-source-aliases>
    <aliases>
    <data-source-name>MYDS</data-source-name>
    <alias>BRANCH50DS</alias>
    </aliases>
    </data-source-aliases>
    Regards,
    S.Divakar

  • How to exctract data from a report to use it in a process

    Hi! I would like to know how to exctract data from a report to use it in a process. Let's say that I have a report which have following fields: id_person, name_person. And then, I want to make a process which introduce the id_person (from report) into an associative tabel. How can I reffer to the fields from my report?
    I was thinking about something like that( this should be the code from my process)
    insert into associative_table(id_person)
    values(id_person) or
    insert into associative_table(id_person)
    values(Report_name.id_person) but it doesn't work....
    Thank you very much.

    Another option, use collections.
    Create an on load process that loads all the rows you would have in your report into a collection:
    begin
    HTMLDB_COLLECTION.CREATE_OR_TRUNCATE_COLLECTION(
    p_collection_name => 'ASSIGNED_CALENDAR_CONTACTS' );
    for x in (SELECT CONTACT_ID, CONTACT_FNAME, CONTACT_LNAME
    FROM CONTACT
    WHERE InStr( pkg_CalendarMgr.f_FetchContactIDsForDay( TO_DATE( :P3_CALENDAR_DT, 'YYYYMMDD' ), 'OF' ), CONTACT_ID ) > 0 )
    loop
    htmldb_collection.add_member(p_collection_name => 'ASSIGNED_CALENDAR_CONTACTS',
    p_c001 => x.CONTACT_ID,
    p_c002 => x.CONTACT_FNAME,
    p_c003 => x.CONTACT_LNAME);
    end loop;
    end;
    Then, instead of your report selecting from the table, have it select from the collection, this way your display will match what's in the collection:
    SELECT c001, c002 || Chr(32) || c003 CONTACT_NAME, 'Remove' remove
    FROM htmldb_collections
    WHERE collection_name = 'ASSIGNED_CALENDAR_CONTACTS'
    ORDER BY c003, c002
    Then you can have a button on the form that steps through the collection, one element at a time and does whatever you want it to.

  • How to read data from a router by using labview

    I am a  beginner labview. How to read data from a router by using labview ? 

    What kind of data are you trying to read?
    Does the router behave like a webserver that you log into?  If so, search the forums for threads discussing HTML.

  • How can we post data from CRM to SAP using ABAP proxies???

    Hi ,
      Can anyone hep me to create interface for the following scenario How can we post data from CRM to SAP using ABAP proxies???, can I find any document ???????
    Thanks in advance
    Andy

    Andy,
    Please look at these weblogs.
    /people/ravikumar.allampallam/blog/2005/03/14/abap-proxies-in-xiclient-proxy
    /people/ravikumar.allampallam/blog/2005/03/03/creating-purchase-order-idoc-through-xi
    /people/siva.maranani/blog/2005/04/03/abap-server-proxies
    /people/vijaya.kumari2/blog/2006/01/26/how-do-you-activate-abap-proxies
    These should give a good idea about implementing Proxies.
    Regards,
    Ravi

  • Could someone tell me how to control a small dc motor by using labeled and a NI 6008 daq. The motor is adjustable by using a pot that ranges between 0-V if that is of any additional benefit

     Could someone tell me how to control a small dc motor by using labeled and a NI 6008 daq.  The motor is adjustable by using a pot  that ranges between 0-V if that is of any additional benefit.

    Read the second thread you have a link to. While both of these will produce PWM signals, the first one is a couple of orders of magnitude too slow to do anything useful in control the speed of a motor, what you would see would be the motor run at the full speed, then stop for a period, then run at full speed, rather than at the "average of on and off"  as the Pulse Widths will be very long relative to the motor's response. The second one has a better chance, IF all you want to do is run the motor. If there is any other calculations ...   Look at both and try and get an understanding of what they are doing. For PWM speed control of a motor you really need a pretty fast pulse width, and probably some filtering to smooth out the resulting "harsh" waveform, giving you the "average" voltage resulting from the PWM.
    Putnam
    Certified LabVIEW Developer
    Senior Test Engineer
    Currently using LV 6.1-LabVIEW 2012, RT8.5
    LabVIEW Champion

  • HOW TO GENARATE DATE LIKE 12 DEC 2007 USING xp20:format-dateTime(dateTime a

    HOW TO GENARATE DATE LIKE 12 DEC 2007 USING xp20:format-dateTime(dateTime as string, format as string). PLEASE HELP.

    If i look at the w3 org i would say :
    xp20:format-dateTime('yourfield','[D]-[[MN,*-3]-[Y]')
    or just MN will do too.

Maybe you are looking for

  • How do I make my Airport Extreme Base Station stop asking to update?

    For awhile my Airport base station has been periodically asking to update its firmware.  But since everything is working fine and I've seen reports of windows computers having trouble connecting after stepping up to that version I'd rather leave it b

  • How to convert RTF or XSL-FO template to PDF template

    We have a "satisfied" RTF template used in 11.15.9. But it no longer works in 11.5.10.2. We would like to try what metalink note#305307.1 suggested. Has anyone done template conversion from RTF or XSL-FO to PDF? Thanks.

  • Where is the value field in antique photos?

    I'm working through the PS Touch for iPad tutorial on antique photos. Step Two requires changing the intensity.  The "Tip" says that the intensity can be changed by tapingnthenvaluenfieldmto,open the keypad editor. Where is the value field ? How do I

  • Linksys WRT150N Crashes all the time...

    Hey, I have a problem concerning my WRT150N router.  About 1-4 times a day the **bleep** thing just stops everything. When I look at the status leds everything still looks ok and it seems to still be sending/receiving stuff from the internet. I use i

  • Help finishing up my first flash site - using navigation to display different sections of the site?

    Thanks for looking! i am making my first flash site, and im not sure what to do next in regards to navigating through the site. I have different sections, home, info, hardware, equipment, etc, but im not sure how to make each section appear when you