Procedure for loading data through 0CO_OM_CCA_9 without taking downtime .

Dear Friends ,
We have already maintained the flow from 0CO_OM_CCA_9 to Infocube 0COOM_C02 . Now i enhanced two fields in the datasource which i want to load into the cube . The infocube already contains data from 2008 onwards . Now i know that for controlling datasources SAP ECC/R3 downtime is not needed . Requirement is to load the historical data again for the two enhanced fields also from 2008 onwards .
Kindly let me know the steps for doing as the delta mechanism is already activated for this flow ?
Thanks & Regards ,
Ashutosh Singh

Hi All,
I am able to create this in client 100 but not in 120. I created this program by copying the source code from sap note and run in test mode in client 100 of development server. I am getting the message that "there is no inconsistency in table BWMO2_TIMEST".
1. I am using client for loading data 120 not 100.
Please let me know the further steps.
Thanks,
Harry

Similar Messages

  • Procedure for loading data into gl_interface

    hi all
    iam new to oracle applications.....i just want to know how to load data into gl_interface table using pl/sql (that is...first loading data into a temporary table and then using pl/sql procedure to load into gl_interface). can anybody help me out with this by providing the pl/sql structure for it??
    thanx in advance

    Asuming you have data in a datafile and file is camma delimited. I asume table has two columns you can add more columns also.
    CREATE OR REPLACE PROCEDURE p10
          IS
          lv_filehandle                                UTL_FILE.FILE_TYPE;
          lv_iicc_premium                              ref_cursor;
          lv_newline                                   VARCHAR2(2000); -- Input line
          lv_header_line                               VARCHAR2(20);
          lv_trailer_line                              VARCHAR2(20);
          lv_file_dir                                  VARCHAR2(100);
          lv_file_name                                 VARCHAR2(100);
          lv_col1                                      VARCHAR2(10);
          lv_col2                                      VARCHAR2(50);
          lv_comma                                     VARCHAR2(1) := ',';
        BEGIN
           gv_PrFnName := '[pr_iicc_premium]';
           DELETE FROM temp_table;
            lv_file_dir  := 'p:\temp';
            lv_file_name  := 'test.dat';
            lv_filehandle := UTL_FILE.FOPEN (lv_file_dir, lv_file_name, 'r', 32766);
            UTL_FILE.PUT_LINE (lv_filehandle, lv_header_line);
            LOOP
               FETCH lv_iicc_premium INTO lv_newline;
               EXIT WHEN lv_iicc_premium%NOTFOUND;
               UTL_FILE.PUT_LINE (lv_filehandle, lv_newline);
               lv_col1 := substr(lv_newline, 1, instr(lv_newline, ',', 1)-1);
               lv_col2 := substr(lv_newline, instr(lv_newline, ',', 1)+1, instr(lv_newline, ',', 2)-1);
               INSERT INTO temp_table VALUES (lv_col1, lv_col2);
               COMMIT;
            END LOOP;
            INSERT INTO your_production_tables VALUES ( SELECT * FROM temp_table );
            COMMIT;
            UTL_FILE.FFLUSH (lv_filehandle);
            UTL_FILE.FCLOSE (lv_filehandle);
        EXCEPTION
           WHEN UTL_FILE.INVALID_PATH THEN
             RAISE_APPLICATION_ERROR(-20100,'Invalid Path');
           WHEN UTL_FILE.INVALID_MODE THEN
             RAISE_APPLICATION_ERROR(-20101,'Invalid Mode');
           WHEN UTL_FILE.INVALID_OPERATION then
             RAISE_APPLICATION_ERROR(-20102,'Invalid Operation');
           WHEN UTL_FILE.INVALID_FILEHANDLE then
             RAISE_APPLICATION_ERROR(-20103,'Invalid Filehandle');
           WHEN UTL_FILE.WRITE_ERROR then
             NULL;
           WHEN UTL_FILE.READ_ERROR then
             RAISE_APPLICATION_ERROR(-20105,'Read Error');
           WHEN UTL_FILE.INTERNAL_ERROR then
             RAISE_APPLICATION_ERROR(-20106,'Internal Error');
           WHEN OTHERS THEN
           UTL_FILE.FCLOSE(lv_filehandle);
        END p10;
    /Code is not tested.
    Hope this helps
    Ghulam

  • Taking much time for loading data

    Dear All
    While loading data in <b>PRD system</b> (for master data and transaction data), it's taking much time. For ex., 2LIS_05_QOITM, I am loading delta data only in <b>PSA</b> from R/3. Some times (yesterday) it's taking 2 mins and some times (today) it's taking 5 hrs not yet completed (it's yellow). Actually yesterday, we went to SM58 in R/3 side and executed that LUW for some other data sources. Now also we can do but we don't want to like that. Because we are expecting permanent soulution. Could u please advice me. I am getting below message in status tab
    Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the source system.
    I am loading data through Process chain and user is <b>BWREMOTE (authorized user).</b>
    Please help me.
    Thanks a lot in advance
    Raja

    Dear Karthik
    No I couldn't resolve till now
    But Everything is fine.
    Now status is yellow only (209 from 209). Now what i want to do.
    Im getting below message <b>in status tab</b>
    Missing data packages for PSA Table
    Diagnosis
    Data packets are missing from PSA Table . BI processing does not return any errors. The data transport from the source system to BI was probably incorrect.
    Procedure
    Check the tRFC overview in the source system.
    You access this log using the wizard or following the menu path "Environment -> Transact. RFC -> Source System".
    Error handling:
    If the tRFC is incorrect, resolve the errors listed there.
    Check that the source system is connected properly to BI. In particular, check the remote user authorizations in BI.
    <b>In detail tab</b>, I am getting below message
    Info IDoc 2 : sent, not arrived ; IDoc ready for dispatch (ALE service)
    Thanks in advance
    Raja

  • FDMEE Import error "No periods were identified for loading data into table 'AIF_EBS_GL_BALANCES_STG'

    Hi,
    We are having trouble while importing one ledger 'GERMANY EUR GGAAP'. It works for Dec 2014 but while trying to import data for 2015 it gives an error.
    Import error shows " RuntimeError: No periods were identified for loading data into table 'AIF_EBS_GL_BALANCES_STG'."
    I tried all Knowledge docs from Oracle support but no luck. Please help us resolving this issue as its occurring in our Production system.
    I also checked all period settings under Data Management> Setup> Integration Setup > Global Mapping and Source Mapping and they all look correct.
    Also its only happening to one ledger rest all ledgers are working fine without any issues.
    Thanks

    Hi,
    there are some Support documents related to this issue.
    I would suggest you have a look to them.
    Regards

  • Unable to load data through outline load utility

    I am unable to load data by using outline load utility.
    Assigned Data load dimension as: Account
    Driver Dimension as : Period
    Driver member as : Jan
    Login file:
    Account,Jan,Point-of-View,Data Load Cube Name
    Investment,1234,"India,Current,Drat,FY13",Plan1
    Command at Outline utility
    OutlineLoad /A:pract /U:admin /I:C:\test1.csv /D:Account /L:C:\lg.log /X:C:\ex.exc
    Exception file
    [Thu Mar 28 18:05:39 GMT+05:30 2013] Error occurred loading data record 1: Investments,1234,"""India,Current,Draft,FY14""",Plan1
    [Thu Mar 28 18:05:39 GMT+05:30 2013] com.hyperion.planning.InvalidMemberException: The member India,Current,Draft,FY14 does not exist or you do not have access to it.
    [Thu Mar 28 18:05:40 GMT+05:30 2013]Planning Outline data store load process finished with exceptions: exceptions occured, examine the exception file for more information. 1 data record was read, 1 data record was processed, 0 were successfully loaded, 1 was rejected.
    Logfile:
    Successfully logged into "pract" application, Release 11.113, Adapter Interface Version 5, Workforce supported and not enabled, CapEx not supported and not enabled, CSS Version 3
    [Thu Mar 28 18:05:38 GMT+05:30 2013]Successfully located and opened input file "C:\load.csv".
    [Thu Mar 28 18:05:38 GMT+05:30 2013]Header record fields: Account, Jan, Point-of-View, Data Load Cube Name
    [Thu Mar 28 18:05:38 GMT+05:30 2013]Located and using "Account" dimension for loading data in "pract" application.
    [Thu Mar 28 18:05:40 GMT+05:30 2013]Load dimension "Account" has been unlocked successfully.
    [Thu Mar 28 18:05:40 GMT+05:30 2013]A cube refresh operation will not be performed.
    [Thu Mar 28 18:05:40 GMT+05:30 2013]Create security filters operation will not be performed.
    [Thu Mar 28 18:05:40 GMT+05:30 2013]Examine the Essbase log files for status if Essbase data was loaded.
    [Thu Mar 28 18:05:40 GMT+05:30 2013]Planning Outline data store load process finished with exceptions: exceptions occured, examine the exception file for more information. 1 data record was read, 1 data record was processed, 0 were successfully loaded, 1 was rejected.
    Infact those members exist in the Outline.
    Any help would be appreciated.

    Check for the double quotes (the one John mentioned). Also I can see that you are taking the approach of Data Load dimension. Is Data load dimension and driver members defined correctly in Planning?
    Regards
    Celvin
    http://www.orahyplabs.com
    Please mark the answers as helpful/correct if applicable

  • Error while Loading data through .csv file

    Hi,
    I am getting below date error when loading data through into Olap tables through .csv file.
    Data stored in .csv is 20071113121100.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TE_7007 Transformation Evaluation Error [<<Expression Error>> [TO_DATE]: invalid string for converting to Date
    ... t:TO_DATE(u:'2.00711E+13',u:'YYYYMMDDHH24MISS')]
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TT_11132 Transformation [Exp_FILE_CHNL_TYPE] had an error evaluating output column [CREATED_ON_DT_OUT]. Error message is [<<Expression Error>> [TO_DATE]: invalid string for converting to Date
    ... t:TO_DATE(u:'2.00711E+13',u:'YYYYMMDDHH24MISS')].
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TT_11019 There is an error in the port [CREATED_ON_DT_OUT]: The default value for the port is set to: ERROR(<<Expression Error>> [ERROR]: transformation error
    ... nl:ERROR(u:'transformation error')).
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TT_11021 An error occurred moving data from the transformation Exp_FILE_CHNL_TYPE: to the transformation W_CHNL_TYPE_DS.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> CMN_1086 Exp_FILE_CHNL_TYPE: Number of errors exceeded threshold [1].
    Any help is greatly appreciated.
    Thanks,
    Poojak

    1) Wrong format, you wont get much support loading OLAP cubes in here I think.
    2) Has your CSV file been anywhere near excel by chance the conversion of 20071113121100 to 2.00711E+13 looks very much like what I see when Excel has an invalid number mask / precision specified for a cell.
    *** We are using EBS. File was already generated by consultants through SQL query. Getting error while loading though Informatica. Target table is setup as date datatype and source is String(19).
    Expression in Informatica is setup as below.
    IIF(ISNULL(CREATED_ON_DT), NULL, TO_DATE(CREATED_ON_DT, 'YYYYMMDDHH24MISS'))
    Thanks,
    Poojak

  • Process chains for loading data to target is not functioning

    Hi SAPians,
    Recently, we have upgraded the firmware on IBM P590 with the help of IBM on Saturday i.e. 06/12/2008 (The firmware of P5-590 was SF235_209. It was upgraded to SF240_XXX) and since then the process chains for loading data to targets are not functioning properly.  We have stopped all the SAP services, database services & OS services from our end and services have been rebooted after firmware upgrade.
    However, the problem with the process chains that load transaction data and hierarchies is solved but the chains that load master data are still not working as scheduled.
    We tried to load the master data manually, by running DTP to load data to an object analysis code (attributes) the request remained YELLOW. After refreshing it several times, the request turned into RED. The error message was - PROCESSING TERMINATED (screenshot 1 attached).
    To mitigate this we tried deleting the bad request and it prompted with the below message:
    "Cannot lock request DTPR_4C37TCZGDUDX1PXL38LVD9CLJ of table ZANLYSIS for deletion (Message no. RSMPC178)" (screenshot 2 attached)
    Please advise us how the bad request should be deleted?
    Regards,
    Soujanya

    Hi Sowjanya,
    Follow the below procedure to make yellow request to RED:
    Use SE37, to execute the function module RSBM_GUI_CHANGE_USTATE
    From the next screen, for I_REQUID enter that request ID and execute.
    From the next screen, select 'Status Erroneous' radiobutton and continue.
    This Function Module, change the status of request from Green / Yellow to RED.
    Once it is RED, you can manually delete that request...
    Releasing LocK
    Gott Transaction Code SM12 and release the lock.
    Hope It helps you.
    Regardss,
    Nagaraju.V

  • Steps for loading data into the infocube in BI7, with dso in between

    Dear All,
    I am loading data into the infocube in BI7, with dso in between. By data flow looks like....
    Top to bottom:
    InfoCube (Customized)
    Transformation
    DSO (Customized)
    Transformation
    DataSource (Customized).
    The mapping and everything else looks fine and data is also seen in the cube on the FULL Load.
    But due to some minor error (i guess), i am unable to see the DELTA data in DSO, although it is loaded in DataSource through process chains.
    Kindly advise me, where did i miss.
    OR .. Step by step instructions for loading data into the infocube in BI7, with dso in between would be really helpful.
    Regards,

    Hi,
    my first impulse would be to check if the DSO is set to "direct update". In this case there is no Delta possible, because the Change log is not maintained.
    My second thought would be to check the DTP moving data between the DSO and the target cube. If this is set to full, you will not get a delta. It is only possible to create one DTP. So if you created one in FULL mode you can't switch to Delta. Just create the DTP in Delta mode.
    Hope this helps.
    Kind regards,
    Jürgen

  • Loading data through rule files in Essbase

    Hi everyone,
    I really need help because something that I don't understand is happening when I am loading data through a rule files that I have created in Essbase. In the "Field Properties" of that rules files, I added few accounts in the "Global Properties" that I wanted to be replaced by others when I perform the load since I had troubles with the hierarchy of my accounts. My problem is that I have 3 accounts that I cannot add when I am loading my data and I don't understand why. By example, in my data extract file, I want to add the account : 601820SN60005 but since it doesn't exist in my Dimension Library I changed my rule file. I add the following propertie : Replace 601820SN60005 with 601820, so the system should put the data in that account instead. But it's not what'S happening. When I perform my load, the log tells me that the member 601820SN60005 does not exist in the database which is true but it seems that the propertie I had is not working. The load should work property. Is it possible that my rules files is corrupted? Or anything else?
    Please give me a clue of what's happening here because I really don't understand!
    Thanks a lot!

    I will just pipe in with the mostly worthless comment that doing any kind of ETL within a load rule is really not the world's best idea. Could you do your transformations in the source? There is this fantastic language called "SQL" that allows all kinds of cool data manipulations. Oracle even sells this product called ODI that I hear is just dandy for ETL work. :)
    A slightly more useful suggestion -- it's really tough to view all of the various transformations in a load rule, as you have found to your sorrow. Did you know you can print the load rule and get all of the transformations? I use it all the time when a client hasn't listened to my whine about not doing ETL in a load rule.
    Regards,
    Cameron Lackpour

  • Loading data through pl/sql

    Hi frnds,
    i want to load .mdb table into oracle 9i through pl/sql procedure.
    i am doing this through sql* loader
    my script is :
    LOAD DATA INFILE '//empwrpub1/Data_Wrangler/Owens/Original Data/037 Refresh 10-06/SAP/SRM_INVOICE.txt'
    INTO TABLE SRM_INVOICE_200610
    FIELDS TERMINATED BY ',' ENCLOSED BY '"'
    LINES TERMINATED BY '\n'
    IGNORE 1 LINES
    now i want this through pl/sql procedure
    thanks in advance

    Hi,
    The two different questions have arised during this Forum.
    1. Load data through SQL loader from PL/SQL.
    2. Use of many DDL and DML statements in the process of Data Loading.
    I will start from the second Question :
    this is great question I myself has faced this problem like first create temporary tables load data, update spaces with null or date formatting etc.
    This is easily achieved when you load the data by using External Tables. There you need to write a script mentioning your datafile, location, column and line termination etc. In this same script you use your DDL statements first and then data loading script and there after DML if any.
    This block would look like something PL/SQL block.
    So, First questioin has also been answered with the condition that you have to use External tables and then load data by writing PL/SQL. I am not sure as I have never tried to call SQL loader commands from PL/SQL. So it Looks impossible. If at all you want to carry out the same activity from PL/SQL block then write a SQL loader Script, CTL file and then write a SHELL Program and then call it from PL/SQL Block.
    Bye

  • Can't load data through smart view (ad hoc analysis)

    Hi,
    There is EPM application where I want to give ability to planners to load data through smart view (ad hoc analysis). In Shared Services there are four options in
    EssbaseCluster-1: Administrator, Create/Delete Application, Server Access, Provisioning Manager. Only Administrator can submit data in smart view (ad-hoc analysis). But I don't want to grant Essbase administrator to planners, I'm just interested to give them ability to load data through ad-hoc analysis. Please suggest!

    I take that you refreshed the Planning security, If not refresh the security of those users. Managing Security Filters
    Check in EAS whether those filters are created with "Write" permissions.
    Regards
    Celvin
    http://www.orahyplabs.com

  • Problem in data sources for transaction data through flat file

    Hello Friends,
    While creating the data sources for transaction data through flat file, I am getting the following error "Error 'The argument '1519,05' cannot be interpreted as anumber' while assigning character to application structure" Message no. RSDS016
    If any one come across this issue, please provide me the solution.
    Thanks in Advance.
    Regards
    Ravi

    Hallo,
    just for information.
    I had the same problem.
    Have changed the field type from CURR to DEC and have set external instead of internal.
    Then, the import with flatfile worked fine.
    Thank you.

  • Submitting data through API without exposing DB credentials

    Submitting transaction / master data through API without exposing DB credentials
    I am developing a scenario where I will submit transaction / master data like sales order / master item to Oracle apps (EBS) through MS Excel via API / interface table but like to provide only apps (EBS) user credentials rather then DB credentials so DB password could not expose to normal end user. Please suggest something on same.
    [email protected]
    Thanks & regards,
    Sachin Gupta

    Sachin, you need a valid userid and password to connect to the database.
    Since these APIs are defined under apps, you need apps password to execute them.
    The workaround is that you define a new d/b userid.
    Grant execute on the API (and other objects as necessary) to this user.
    And then you call the api using the new userid and password.
    This way you are not exposing apps password. But since this means you have to grant a number of accesses to this new user.
    If you can use the interface tables, it becomes a little easier.
    All you have to do is grant insert access on a handful of interface to this new userid. And then you schedule a concurrent program that will process the interface records.
    If you absolutely can not expose ANY database password, then you can communicate with EBS using flat files. Your system creates a flat file on the server and a scheduled program in EBS reads the files and calls appropriate api.
    Hope this helps,
    Sandeep Gandhi

  • If I buy a 32GB IPhone 4S, will i be able to store my current ITunes library on my phone through ICloud without taking up the 32GB built into my phone?

    If I buy a 32GB IPhone 4S, will I be able to store my current ITunes library on my phone through ICloud without taking up the 32GB built into my phone?

    iCloud does not store music except in two ways.  First, you can redownload music previously purchased from the iTunes store should you ever need to (as long as the songs are still available); which is referred to as "iTunes in the Cloud" (but has nothing to do with your personal iCloud account).  Secondly, you can store your music library in iCloud (up to 25,000 songs) if you subscribe to a fee-based service called iTunes Match (see http://www.apple.com/itunes/itunes-match/).  If you subscribe to iTunes Match you can keep your music in iCloud, but (at least at the current time) when you listen to this music it is simultaneously downloaded to your phone so it ends up using some of you phone's capacity anyway.

  • Need to load the data through excel for plan data through IP?

    Hi all,
    The user needs to load the data through the excel and click on the save button so that the data is saved in the cube through integrated planning.
    The objects used are
    customer, material, plant, calendarmonth sales qty, gross billing!!
    Can anyone pls let me know how to do this?
    Thanks

    Hi,
    You can use the following How to guide to load data from file into transactional cube.
    If using IP :
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/90cef248-29e0-2910-11bb-df8dece7e22c?quicklink=index&overridelayout=true
    Can also Refer to Marc's Blog:
    /people/marc.bernard/blog/2007/11/25/how-to-load-a-file-into-sap-netweaver-bi-integrated-planning-part-1
    If you are using bPS you can get the details in Wiki also for the same.
    You can then create the hyperlink in Excel ,which user can use to upload data which can then be saved to cube using save button.

Maybe you are looking for

  • How do I create a hyperlink in my iCloud email signature?

    I have created an email signature in iCloud and would like to create hyperlinks for my websites. How do I do this in my email signature? I know I can click the www button on each email, but would like it to appear in the signature every time.

  • Digital surround sound not working

    i have just now taken the time to play some movies through the apple tv and the center channel sound is non existant.  i change sources and its fine, but only pn apple tv does the center channel drop out.  then it will suddenly reappear. any ideas? a

  • Error 503 in app manager or online

    When I run adobe app manager it errors out with a 503 error and says it can't get list. I tried the solution on forum renamed folders and reinstall application manager and still same. I logged in to adobe.com and tried to bring up list from the websi

  • Inexpensive, easy to use, reliable RAID USB/FW enclosure

    Does anyone have any experience with RAID enclosures ? For example, there is one by NewerTech that takes two drives and uses USB to connect to the Mac. The description claims it is compatible with OS X. How do these things work with OS X Leopard ? Do

  • SXMB_MONI missing XML steps

    Hi I am working on XI7.0. In transaction SXMB_MONI, once I select individual message, it takes you to next screen where we can see details of each processing in sequence. We normally suppose to see on left hand side with following details in sequence