Getting error while scheduling load from one cube to another

Hi,
I have create one cube ,which has got loaded from two data sources and load into the cube went successful.I have created copy cube and schedule the load into copy cube from the base cube,where i am getting error message syaing that"New work Area started,terminated the system" and "Core_dump","Rolled-Out session terminated" all the above messages are getting and so many short dumps.I am getting all the above messages.
Even i have change data packet size from 20000 to 5000 and increares dialup process from 2 to 5.But still i am getting same kind of messages.
Please some one can help me out from this issue.

Hi,
In which version you are.
If you in 7.0 then have you checked in DTP.
Just logooff from your server, and again relog, and try to load once again.
Let us know status ....
Reg
Pra

Similar Messages

  • Getting error , while passing parameters from one page to another page

    Hello friends,
    i am getting error, while passing parameters from one page to another page, below code i wrote.
    public void processFormRequest(OAPageContext pageContext, OAWebBean webBean)
    super.processFormRequest(pageContext, webBean);
    ArrayList arl=new ArrayList();
    EresFrameworkAMImpl am=(EresFrameworkAMImpl)pageContext.getApplicationModule(webBean);
    ERecordImpl ERecordObj=new ERecordImpl();
    HashMap hMap = new HashMap();
    hMap.put("1",ERecordObj.getTransactionName());
    hMap.put("2",ERecordObj.getTransactionKey());
    hMap.put("3",ERecordObj.getDeferredMode());
    hMap.put("4",ERecordObj.getUserKeyLabel());
    hMap.put("5",ERecordObj.getUserKeyValue());
    hMap.put("6",ERecordObj.getTransactionAuditId());
    hMap.put("7",ERecordObj.getRequester());
    hMap.put("8",ERecordObj.getSourceApplication());
    hMap.put("9",ERecordObj.getPostOpAPI());
    hMap.put("10",ERecordObj.getPayload());
    // hMap.put(EresConstants.ERES_PROCESS_ID,
    if(pageContext.getParameter("item1")!=null)
    pageContext.forwardImmediately(EresConstants.EINITIALS_FUNCTION,
    OAWebBeanConstants.KEEP_MENU_CONTEXT,
    null,
    hMap,
    true,
    OAWebBeanConstants.ADD_BREAD_CRUMB_YES
    Error(71,2): method forwardImmediately(java.lang.String, byte, null, java.util.HashMap, boolean, java.lang.String) not found in interface oracle.apps.fnd.framework.webui.OAPageContext
    Thanks
    krishna.

    Hi,
    You have imported the wrong class for HashMap.
    Import
    com.sun.java.util.collections.HashMap; instead of java.util.HashMap
    Thanks,
    Gaurav

  • Load from one cube to another

    hi all,
    can any one gimme  step by step procedure to load data from one cube to another cube in BI7
    thanx in advance

    Hi Srikanth,
    Cube to Cube Load 
    You need to move some data from one cube to another.
    The steps involved are :-
    First You need to  create 'Export Data Source' from original cube (right-click on the InfoCube and select Generate Export Data Source). 
    Then, assign the new datasource to new cube. (you may click on 'Source system' and select your BW server and click 'Replicate'). 
    Then, Create DTP and execute. 
    Lastly, you are ready to load. 
    Regards
    Suresh B.G.

  • Getting error in file transfer from one system to another system

    Hi All
    One of my client (User) wants to transfer his file from  SAP R/3 System SAP XI system. He is using a customize t-code ZFIR0325 (Trial balance report transfer to Hyperian System). When user execute the  program he is getting error "File Transfer Failed due to Network chk with Admin"
    I checked with Network team but they said it is okay from their side. We also check for authorization but there is no issue regarding authorization.
    Name of File Path of Aplication Server, XI destinaion and XI directory entered by user is correct.
    We also put authorization trace in XI System to RFC User (In file transfer from one system to another system this RFC User is geeing Used) but no authorization issue was found.
    Please help me to resolve this problem. I any detail require I will provide the same.

    Guruprasad Wrote:
    He is using a customize t-code ZFIR0325 (Trial balance report transfer to Hyperian System).
    We also put authorization trace in XI System to RFC User (In file transfer from one system to another system this RFC User is geeing Used) but no authorization issue was found.
    Firstly, no one in the community can answer on how a custom transcation code and a program associated with it behaves.
    If there are no authorization issues in the system, you have to look at the exact error message. Look if your XI system is allowed to receive files from the SAP system.
    Unfortunately, you should do some more home work before putting your question here. If everything is fine and configured correctly, then the issue might be with the program.
    Regards,
    Raghu

  • Getting error while creating table from one database to other.

    Hi,
    We are getting below error while creating the table from one database to other.
    SQL> create table fnd_lobs parallel compress as select * from [email protected];
    create table fnd_lobs parallel compress as select * from [email protected]
    ERROR at line 1:
    ORA-01555: snapshot too old: rollback segment number 28 with name "_SYSSMU28$"
    too small
    ORA-02063: preceding line from EEXIT2TEST
    ORA-01555: snapshot too old: rollback segment number 28 with name "_SYSSMU28$"
    too small
    ORA-02063: preceding line from EEXIT2TEST
    ORA-01555: snapshot too old: rollback segment number 28 with name "_SYSSMU28$"
    too small
    ORA-02063: preceding line from EEXIT2TEST
    ORA-01555: snapshot too old: rollback segment number 28 with name "_SYSSMU28$"
    too small
    Regards,
    Bhatia

    hi
    what are the apps version local and remote database???
    Snapshot too old errors occur because Oracle can 't reconstruct a consistent
    image of a block for the purposes of a consistent read.
    I feel at remote database, you are using UNDO, it will be rather easy to iincrease the undo retention time or increase the undo tablespace size.. if you are dealing with roll back segments, you may have rollback segments whose optimal values are too small...
    increase roll back segments size and select again then
    the following metalink notes might be helpful
    ORA-01555 "Snapshot too old" - Detailed Explanation Doc ID: 40689.1
    How To Avoid ORA-01555: Snapshot Too Old When Running PAAPIMP Doc ID: 603259.1
    OERR: ORA 1555 "snapshot too old (rollback segment too small)" Doc ID: 18954.1

  • Fatal error while transferring data from one bb to another

    I was trying to transfer data from one bb to a different bb using the bb desktop manager tool.  It collected all the data from the first bb fine and then while transferring data to the new one it said a fatal error occured to try again.  That bb will not even turn on now.  No white screen or anything.  Its like there is no operating system.  I don't know what to do.  I tried to connect it and just update the software but that isnt working either.  Says it is not responding during initialization! I think I killed it.... any ideas what to do???

    You can try to reload the OS in one of the following ways
    http://www.blackberry.com/btsc/KB03485
    http://blackberryfaq.net/index.php/How_do_I_wipe_the_BlackBerry_using_Jl_Cmder%3F
    If someone has been helpful please consider giving them kudos by clicking the star to the left of their post.
    Remember to resolve your thread by clicking Accepted Solution.

  • Error while duplicate database from one server to another server

    Hi there,
    I am practicing duplicate database from one single database server to another server. Hardware and driver path on 2 machines is the same. Here are my steps:
    --------------------- Clone site:
    1. modify listener.ora with content:
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1521))
    (ADDRESS = (PROTOCOL = TCP)(HOST = localhost.localdomain)(PORT = 1521))
    ADR_BASE_LISTENER = /u01/app/oracle
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (GLOBAL_DBNAME = orcl)
    (ORACLE_HOME = /u01/app/oracle/product/11.2.0/db_1)
    (SID_NAME= orcl)
    ENABLE_GLOBAL_DYNAMIC_ENDPOINT_LISTENER = ON
    2. stop and start listener again, message show like :
    Service “orcl” has 1 instance(s).
    Instance “orcl”, status UNKNOWN, has 1 handler(s) for this service…
    The command completed successfully
    3. create pfile from spfile.
    4. shutdown and startup nomount with pfile is new pfile create from step 3.
    --------------------------------------- Source site
    1. modify tnsnames.ora with content like:
    VVCOREDB =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = localhost.localdomain)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = vvcoredb)
    ORCL =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP) (HOST = 10.254.1.235) (PORT=1521))
    (CONNECT_DATA =
         (SERVER = DEDICATED)
         (SERVICE_NAME = orcl)
    2. run command from rman like these:
    rman> connect target sys@vvcoredb
    rman> target database Password:
    rman> connect auxiliary sys@orcl
    rman> auxiliary database Password:
    rman> DUPLICATE TARGET DATABASE to orcl from active database spfile nofilenamecheck;
    Then the result show below:
    Starting Duplicate Db at 24-SEP-12
    using target database control file instead of recovery catalog
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=63 device type=DISK
    contents of Memory Script:
    backup as copy reuse
    targetfile '/u01/app/oracle/product/11.2.0/db_1/dbs/spfilevvcoredb.ora' auxiliary format
    '/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora' ;
    sql clone "alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''";
    executing Memory Script
    Starting backup at 24-SEP-12
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=74 device type=DISK
    Finished backup at 24-SEP-12
    sql statement: alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''
    contents of Memory Script:
    sql clone "alter system set db_name =
    ''ORCL'' comment=
    ''duplicate'' scope=spfile";
    shutdown clone immediate;
    startup clone nomount;
    executing Memory Script
    sql statement: alter system set db_name = ''ORCL'' comment= ''duplicate'' scope=spfile
    Oracle instance shut down
    connected to auxiliary database (not started)
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 09/24/2012 17:44:12
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-04014: startup failed: ORA-01261: Parameter db_recovery_file_dest destination string cannot be translated
    ORA-01262: Stat failed on a file destination directory
    Linux-x86_64 Error: 2: No such file or directory
    I don't know what wrong with my pfile? I specify correct my pfile on clone site like this:
    orcl.__db_cache_size=1476395008
    orcl.__java_pool_size=16777216
    orcl.__large_pool_size=16777216
    orcl.__oracle_base='/u01/app/oracle'#ORACLE_BASE set from environment
    orcl.__pga_aggregate_target=1342177280
    orcl.__sga_target=1979711488
    orcl.__shared_io_pool_size=0
    orcl.__shared_pool_size=436207616
    orcl.__streams_pool_size=0
    *.audit_file_dest='/u01/app/oracle/admin/orcl/adump'
    *.audit_trail='db'
    *.compatible='11.2.0.0.0'
    *.control_files='/u01/app/oracle/oradata/orcl/control01.ctl','/u01/app/oracle/flash_recovery_area/orcl/control02.ctl'
    *.db_block_size=8192
    *.db_domain=''
    *.db_name='orcl'
    *.db_recovery_file_dest='/u01/app/oracle/flash_recovery_area/vvcoredb'
    *.db_recovery_file_dest_size=4070572032
    *.diagnostic_dest='/u01/app/oracle'
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=orclXDB)'
    *.memory_target=3313500160
    *.open_cursors=300
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.undo_tablespace='UNDOTBS1'
    Please suggest me do i miss some steps when try to duplicate database?
    Any suggestions are welcome.
    Thanks in advance!
    P/s: Sorry for my bad English.

    Oh god, it worked. Thank you. But i got another problem like this:
    duplicate target database to orcl from active database spfile set "db_recovery_file_dest" = "/u01/app/oracle/flash_recovery_area/vvcoredb" nofilenamecheck;
    Starting Duplicate Db at 24-SEP-12
    using target database control file instead of recovery catalog
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=63 device type=DISK
    contents of Memory Script:
    backup as copy reuse
    targetfile '/u01/app/oracle/product/11.2.0/db_1/dbs/spfilevvcoredb.ora' auxiliary format
    '/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora' ;
    sql clone "alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''";
    executing Memory Script
    Starting backup at 24-SEP-12
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=74 device type=DISK
    Finished backup at 24-SEP-12
    sql statement: alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''
    contents of Memory Script:
    sql clone "alter system set db_name =
    ''ORCL'' comment=
    ''duplicate'' scope=spfile";
    sql clone "alter system set db_recovery_file_dest =
    ''/u01/app/oracle/flash_recovery_area/vvcoredb'' comment=
    '''' scope=spfile";
    shutdown clone immediate;
    startup clone nomount;
    executing Memory Script
    sql statement: alter system set db_name = ''ORCL'' comment= ''duplicate'' scope=spfile
    sql statement: alter system set db_recovery_file_dest = ''/u01/app/oracle/flash_recovery_area/vvcoredb'' comment= '''' scope=spfile
    Oracle instance shut down
    connected to auxiliary database (not started)
    Oracle instance started
    Total System Global Area 3307048960 bytes
    Fixed Size 2217872 bytes
    Variable Size 1811941488 bytes
    Database Buffers 1476395008 bytes
    Redo Buffers 16494592 bytes
    contents of Memory Script:
    sql clone "alter system set db_name =
    ''VVCOREDB'' comment=
    ''Modified by RMAN duplicate'' scope=spfile";
    sql clone "alter system set db_unique_name =
    ''ORCL'' comment=
    ''Modified by RMAN duplicate'' scope=spfile";
    shutdown clone immediate;
    startup clone force nomount
    backup as copy current controlfile auxiliary format '/u01/app/oracle/oradata/vvcoredb/control01.ctl';
    restore clone controlfile to '/u01/app/oracle/recovery_area/vvcoredb/control02.ctl' from
    '/u01/app/oracle/oradata/vvcoredb/control01.ctl';
    alter clone database mount;
    executing Memory Script
    sql statement: alter system set db_name = ''VVCOREDB'' comment= ''Modified by RMAN duplicate'' scope=spfile
    sql statement: alter system set db_unique_name = ''ORCL'' comment= ''Modified by RMAN duplicate'' scope=spfile
    Oracle instance shut down
    Oracle instance started
    Total System Global Area 3307048960 bytes
    Fixed Size 2217872 bytes
    Variable Size 1811941488 bytes
    Database Buffers 1476395008 bytes
    Redo Buffers 16494592 bytes
    Starting backup at 24-SEP-12
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting datafile copy
    copying current control file
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 09/24/2012 18:42:09
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 09/24/2012 18:42:09
    ORA-17628: Oracle error 19505 returned by remote Oracle server
    I have a comment that source site doesn't have flash_recovery_area, instead it's name is recovery_area, is it problem?
    Edited by: 788420 on Sep 24, 2012 4:44 AM

  • Error while data shifting from one table to another table

    i have two tables.
    in one table i dont have constraints.in anothe rtable i have constraints.
    my requiremnt is i have to shift data from unconstrained table to constrained table
    while shifting data any constaint violation is there system hasto raise error by specifying the error
    in below example it should be ename,esal values not null.
    i tried with save exception but it gives only error.i want error by mentiong that column names also ie ename,esal is not null
    if ename value only null then it should be give ename is null
    create table tablea (eno number(10), ename varchar2(20) not null, esal number(5) not null)
    create table tabl (eno number(10), ename varchar2(20) , esal number(5) )
    DECLARE
    TYPE tabl_tab IS TABLE OF tabl%rowTYPE;
    tab tabl_tab;
    -- create an exception handler for ORA-24381
    ERRORS NUMBER;
    dml_errors EXCEPTION;
    PRAGMA EXCEPTION_INIT (dml_errors, -24381);
    BEGIN
    SELECT *
    BULK COLLECT INTO tab
    FROM tabl
    WHERE rownum<=2;
    -- add '_SR' to the job_id of the most senior employees
    FORALL i IN tab.FIRST .. tab.LAST SAVE EXCEPTIONS
    insert into tablea values tab(i);
    EXCEPTION
    WHEN dml_errors
    THEN
    rollback ; --roll backs all the good records updation
    -- Now we figure out what failed and why.
    ERRORS := SQL%BULK_EXCEPTIONS.COUNT;
    DBMS_OUTPUT.put_line ('Number of statements that failed: ' || ERRORS);
    FOR i IN 1 .. ERRORS
    LOOP
    DBMS_OUTPUT.put_line ( 'Error #'
    || i
    || ' occurred during '
    || 'iteration #'
    || SQL%BULK_EXCEPTIONS (i).ERROR_INDEX
    DBMS_OUTPUT.put_line ( 'Error message is '
    || SQLERRM
    (-SQL%BULK_EXCEPTIONS (i).ERROR_CODE)
    END LOOP;
    END;
    how to implement that.

    it should give error as ename column is not null

  • Error while migrating reports from one instance to another(no torch seen)

    Hi,
    I am facing this problem when I am trying to migrate reports from UAT to Production. After migration when I try to run the report from Discoverer plus, I cannot see the torch symbol which is general if we have any parameters for that report. So, I again migrate the same .eex file and now I see the torch.
    Can any one throw some thoughts why is this happening, why I cannot see the torch symbol in the first time itself?
    Thanks,
    Srini.

    Hi Srini
    The list of values is not being associated with the item the first time around. The order in which the items are selected is important so for example if lists of values were imported first without the items upon which they depend they will be ignored. Running the import a second time will cause the LOVs to recognize their associated item and thus the association will be made.
    Generally I do two exports. I will export my folders and workbooks first and then export my item classes.
    Hope this helps explain what you are seeing.
    Best wishes
    Michael

  • Error while transporting objects from one package to another

    Hi Everyone,
    I am using a FM 'HR_INFOTYPE_OPERATION' to update a value into a table pa0584 table.
    If i try to upload all the values i am getting a return value like " No of Entries exceeds more than 20".
    So please help me out.
    Thanks
    Vijay

    Hi,
    You can merge your different requests into a single request. goto se09  select any one of your request and in menu,
    Utilities -> Reorganize -> Merge Requests. or press CTRL + F12.
    hope this helps u.,
    Thanks & Regards,
    Kiran

  • Inventory data load from Inventory Cube to another Copy Cube

    Hello Experts,
    I am trying to load Inventory data from the Inventory cube(say YCINV) to a copy cube  say YCOPY_CINV(copy of Inventory cube), but the results appear inconsistant when I compare the reports on these 2 cubes. I am trying to populate a fiield in copy cube so that I can populate the same data back to the original cube with the new field in it, I am doing this reload back and forth for historical data purpose only.
    I have seen lot of posts as how to run the setups for Inventory data, but my case does not need to perform set up runs.
    Does the note 1426533 solve the issue of loading from one cube to another ? we are on SAP BI 7.01 with SP 06 ,but the note specifies SP 07 ?
    I have tried note 375098 to see if it works, but I do not see the options as mentioned from  step "Using DTP (BW 7.x)" in BI to perform this note
    Please advise on whether to go with implementing note 1426533 or is there any other way to load inventory data from one cube to other.
    Regards,
    JB

    Hi Luis,
    Thanks for your reply,
    I do not see any setting like Intial stock in DTP except "initial non-cumulative for non- cumulative". I did try using the option "initial non-cumulative for non- cumulative" ,but the results still do not match with the inventory cube data.I do not see the check box for marker in the copy cube (under roll up tab). Please let me know if we can really implement this solution ,i.e. copying from inventory cube to copy cube and then re-loading it back to inventory cube for historical data. Currenlty, I am comparing the queries on these 2 cubes, if the data matches then I can go  ahead and implement it in Production, other wise it would not be wise to do so .
    Regards,
    JB

  • HT202157 New apple is not asking for the WIFI password.Home sharing is not working,i am getting Error while streaming videos from Youtube.

    I got new Apple TV 10 days back,its not asking for the WIFI password.Home sharing is not working,i am getting Error while streaming videos from Youtube.
    I can only see pics and video saved in my iphone. No other options are working.The worst part is that its not asking wifi passoword.
    Not able to connect istores and itunes.
    I cannot do software update. Can anyone help in this content?

    Thanks Brian- I am not connecting it with iphone teethering. I am using it with my DSL wifi.
    I am using it with ethernet cable then its working fine.If i am trying to use it with WIFI then its giving error.
    I tried using Netwrok testing but its giving same error.

  • Move data from one cube to another cube

    Hi,
    I am on BW 3.5 and I have moved the data from one cube to another cube and found that the number of records in the original cube does not match to the newly created cube. for eg. if the original cube contains 8,549 records then the back up cube contains 7,379 records.
    Please help me on what I need to look on and if in case the records are getting aggregated then how do I check the aggregating record.
    Regards,
    Tyson

    Dear tyson m ,
    check with any update rules in ur transfer.If so check in it.
    Just go through these methods for making transfer from one cube to another cube fully without missing data.
    Update rules method
    if it's updated from ods, you can create update rules for cube2 and update from ods
    or you can try datamart scenario
    cube1 right click 'generate export datasource'
    create update rules for cube2, assign with cube1
    rsa1->source system->bw myself, right click 'replicate datasource'
    rsa1-> infosource -> search 8cube1 name
    (if not get, try right click root note 'infosource'->insert lost node(s)
    from that infosource, you will find assigned with datasource, right click and 'create infopackage', schedule and run.
    Copy from
    While creating the new cube give the cube name in the "Copy from" section. It would copy all the characteristics and Key figures. It would even copy the dimensions and Navigational attributes
    Another option is:
    The steps for copying the contents of one cube to another:
    1. Go to Manage -> Recontruct of the new cube.
    2. Select the "selection button"(red , yellow, blue diamond button).
    3.In the selection screen you can give the technical name of the old cube, requests ids you want to load, from & to date.
    4.Execute and the new cube would be loaded.
    Its all that easy!!!!!!
    Refer this link:
    Copying the structure of an Infocube
    Reward if helpful,
    Regards
    Bala

  • ABAP Function Module Example to move data from one Cube into Another

    Hi experts,
    Can any please help out in this ..?
    A Simple ABAP Function Module Example to move data from one Cube into Another Cube
    (How do i send the data from one client to another client using Function moduel).
    Thanks
    -Upen.
    Moderator message: too vague, help not possible, please describe problems in all technical detail when posting again, BI related? ("cube"), also search for information before asking.
    Edited by: Thomas Zloch on Oct 29, 2010 1:19 PM

    This is the start routine to duplicate records in two currencies.
    DATA: datew   TYPE /bi0/oidateto,
          datew2  TYPE rsgeneral-chavl,
          fweek   TYPE rsgeneral-chavl,
          prodhier TYPE /bi0/oiprod_hier,
          market  TYPE /bic/oima_seg,
          segment TYPE /bic/oizsegment.
    DATA: BEGIN OF S_DATA_PACK OCCURS 0.
            INCLUDE STRUCTURE /BIC/CS8ZSDREV.
    DATA: END OF S_DATA_PACK.
    S_DATA_PACK[] = DATA_PACKAGE[].
      REFRESH DATA_PACKAGE.
      LOOP AT S_DATA_PACK.
        move-corresponding s_data_pack to DATA_PACKAGE.
        if DATA_PACKAGE-loc_currcy = 'EUR'.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalloc.
          DATA_PACKAGE-CURRENCY = 'EUR'.
          APPEND DATA_PACKAGE.
        else.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
        endif.
      ENDLOOP.
    This is to load Quantity field
    RESULT = COMM_STRUCTURE-BILL_QTY.
    This is to load Value field
    RESULT = COMM_STRUCTURE-NETVAL_INV.
    UNIT = COMM_STRUCTURE-currency.

  • Partitioning two Years Of Data From One Cube Yo Another

    Hi Experts,
    Can any one please help me on the following issue.
    I got the requirement where I need to copy the data from one cube to another having unequal number of dimensions.
    I have achieved this using Transparent partitioning, But I could be able to copy only one year data.
    In cube1 I have the months like Jan, Feb...Dec and in cube2 I have the structure like Jan-11..Dec-11, Jan-12...Dec-12.
    While creating the partition how can I map the Jan with Jan-11 and Jan-12.
    Thanks In Advance,
    Ram

    You can map manually. it is possible.

Maybe you are looking for