Help required in data loads

Hi all,
We have actually installed some patch in R/3 system. so now we are testing BW3.5 system  whether the full and delta loads are working or not.
Full loads are not having any problem but few delta loads are having problems . what we did is -
1) Re- initialize the delta loads .
2) Put/ change  some records in R/3 system
3) Schedule delta loads in BW .
But 0 records are coming in BW  though in R/3 the records are present.
Kindly help.

Hi
You mentioned that no recorde are moved to BW.
1)  Did you  check in R/3 side weather the data is in RSA7 or not?.
2)  If not check in SM13 (V3) and lBWQ(Qued Delta).Activate the data source.
3)  Replicate the datasource in BW and
4)  Run the Program RS_TRANSTRU_ACTIVATE_ALL in SE38.
5)  Perform data Loading.
You can get data in this way.
I think this will help you.
Regards,
Siva

Similar Messages

  • Urgent help require-ASO data loading

    I am working on 9.3.0 and want to do incremental data loading so that it wont affect my past data.I am still not sure that it is doable or not .Now the question is
    do I need to design a new aggragation after loading data.
    Thanks in advance

    Hi,
    The ASO cube will clear off all the data , if you make any structural changes to the cube ( i.e if you change your outline ,and you can find out what exactly can clear off hte data ,for example , if you add a new member ,then ,it clears off , and in other case ,if you just add a simple formula to the existing member , it might not clear off the data ).
    If you dont want to effect the past data ,and yet you want to incrementally load , then ensure that all the members are already in the otl( take care of time dimension , let all the time members be there in place) and then you can load in ,using the option of 'add to existing values' while you load .
    But remember , you can only do ,without any structural changes. Else , you need to load all together again
    Its good if you design aggregation , as it helps in retrieval performance, but its not mandatory.
    Sandeep Reddy Enti
    HCC

  • Need Help about Spatial Data Load - Mapviewer

    Hello everybody,
    I need an immediate help about spatial data load. I installed Oracle mapviwer quick start and try to work on it. However, I could not pass the load step. My questions are;
    1- Where can I find and download my country's data set (spatial data)
    2- With mapviwer, how can I load spatial data to my tables on database (Oracle). Those tables have sdo_geometry columns and I want to query location data, but could not load
    Regards,
    Dilek

    For Mapviewer questions, please post in the following forum:
    MapViewer
    Thanks

  • Coding Required for Data Load...

    Hi All,
    I am stuck up in an issue where I know the logic but unable to write a code. I have the requirement as follows :-
    I have the records in Excel as given below:-
    Fiscal Year          2004
    Key                     004
    Capitalization       A
    Norm Debts         B=0.7*A
    Rate of Interest    C
    Repayment          D=B/10
    Op. Balance         D=B/10
    Cl. Balance          E = B-D
    Total Interest        G=E+F/2*C
    The above records are the data which I am loading first time and only one record. The cube shud be populated with the data till Closing Balance becomes zero.
    So my logic will be
    First record by excel and second onwards is :-
    Fiscal Year         2004 +1 (It shud add 1 to every record till closing balance becomes zero)
    Key                    004 (Constant till Closing balance becomes Zero)
    Capitalization      A (Constant till Closing balance becomes Zero)
    Norm Debts        B (Constant till Closing balance becomes Zero)
    Rate of Interest    C (Constant till Closing balance becomes Zero)
    Repayment         D (Constant till Closing balance becomes Zero)
    Opening Balance   (Closing Balance will become Opening Balance here) = B-D
    Closing Balance     (Closing balance will now again will be debited by repayment that is B-2D)
    Total interest        = (Opening Balance + Closing Balance)/ 2 * Rate of Interest
    So next record will be opening balance = B-2D and Closing Balance will be B-3D and the record shud be automatically fetched till Closing Balance becomes zero.
    Any help will be assigned with points.
    Thanks and Regards,
    Sangini Mathur.

    Hi Sangini,
    I think in ur case u have to LOOP aginst source_package
    in the start routine only then you can get ur requirement.
    Data: w_temp1 like source_package-cl.balance.
    while w_temp1 eq 0.
    Put all your calculations here
    endwhile.
    Hope it helps
    Bhaskar
    Edited by: shanthi bhaskar on Jun 3, 2008 2:34 PM

  • Help required in SQL Loader

    Is it possible to load a single target file from multiple flat file sequentially using SQL Loader?
    Example -
    Table structure -
    <Emp_id><Emp_Name><Dept_id><Total_sales><Commission>
    <Emp_id><Emp_Name><Dept_id> column will be load from Emp_master flat file
    <Total_sales> column will be load from sales flat file - there will be two fields emp_id and total_sales
    <Commission> column will be load from commission flat file - there will be two fields emp_id and commission.
    I have tried this to merge these three flat file and create a single flat file, but in my real requirement i found it is quite tedious.
    Is there any other approach?
    Thanks in Advance

    You can load into 3 temporary tables and insert into main table using a query
    OR
    You can use shell script (I hope UNIX environment) to merge the files
    There is no direct way to load the data for same row from multiple files using SQL * Loader

  • Need to help in initial data loading from ISU into CRM

    One our client has requirement as
    All ISU data applicable (BP, BA, the appropriate technical data, Contracts, Products, Product Configuration and their correspoinding price Keys and Price Amount) to CRM  should be loaded into CRM as the part of the intial load.
    Eventhough ECRM_GENERATE_EVERH  in ISU but its documentation is not available.
    Is any provision like Report or RFC or function module present in SAP.
    I would appreciate  for you all quick reply with positive and appropriate solution. mailto [email protected]

    Got my answer.We can clear the data using MDX.

  • PLEASE HELP-PROBLEM IN DATA LOADING

    Hi,
    I had made a repair load for 5 records from a source ods to target ods.After Repair load I found 3 records are correct but 2 still missing . I made again repair load but data was again missing. All the 5 records share the same conditions...Plz tell me what can be the reasons that 3 are correctly loaded and 2 could not. What Can be the solution for them???
    Please reply!!1

    hi,
    have you chked in source ods active table for the presence of 5 records which you require.
    if so do the full repair upto PSA and chk its presence in the PSA.
    then load to ods and chk in new data table then activate the ods and chk
    so you could findthe records where it is missing.
    Ramesh

  • Need help in BW data load

    Hi gurus,
                 I am just practicing in SAP BW. I have created Info Catalog, InfoObjects, InfoAreas, InfoProviders. All I did and went throught correct. But when I check data in Infocube. When I want to see the InfoCube Contents, it is taking all infoobjects but it is not show ing data values. Please help me how to see data while seeing with infocube->manage->data browser: Table /bic/fic_demoBC: selecttion screen.
                  your help will be appriciate
    Thanks
    Happy

    hi ,
       Considering that u r loadinf from the Flat file .. follow the steps ...
    1.create Source System : RSA13 --- Create Source System.
    2.Create InfoArea : RSA11 /// RSA 14
       Create info object catalogues and the Info Objects .. Key figs n chars .
    3.Create Infosource if transaction data using flexible update n if master data using direct update.
    4.Do mapping n propoose transfer rules. n see the communication structure.
       n activate transfer rules.
    5.create Info cube n according the starschema concept . i.e. Tcode RSA11 give the IC name n desc .. give the infosource n create dimensions n assign the chars logically n give time chars .
    6.Now context menu on the IC n create update rules to update the data from the Infosource to the data target.
    7. Now schedule the data from Source to the DT using the Info package .
    8.in the scheduler screen u can give the selection n the data target  n the processing n finally star tloading .
    now u can check the data in the monitor screen by selcting the dispaly contents n selecting the chars n Kfs to be dispalyed .
    let us know if any help requuired .....
    regards .. Kumar Sarvepalli ..........,,,

  • Clarifications required on Data load update types...

    Hi Friends,
    I have seen many previus threads in the forum regarding Init, delta  Full update and Full Update with Repair.
    Can some one expain me clearly the following?
    1: What is Repair with Full update. When do we use this ?Give me some example.
    2: If a delta failed to ODS, what should I do?
    3: If delta fails from ODS to Cube ?
    4: When we use Full update ? Every time when we Full update wether the records get duplicated or not?
    Advance thanks for your answers.
    Sudha

    What is Repair with Full update. When do we use this ?Give me some example.
    You can go in for a repair full request when you have missed any delta loads or there are data corruption issues. By doing a full repair load, you can ensure that your data is correct and has good integrity.
    With this option you can continue to use your existing delta and not worry about resetting the delta.
    Re: BIW
    2: If a delta failed to ODS, what should I do?
    Chk whether the data has come to PSA for all the package then u can update the data from PSA. else if extraction part is not over then u need to find out what is the error. force the request to red in ODS and delete it from target. rectify teh error and reload it
    3: If delta fails from ODS to Cube ?
    Need to force the request to red in Cube. reset the datamart(tick in ODS) then identify what is the error rectify it then load it(In this case too if data is there in PSA tehn no need to rest the data mart u can pull it from intermediate PSA itself)
    4: When we use Full update ? Every time when we Full update wether the records get duplicated or not?
    Full will pull records from setup dates. It depends on ur datatarget settings(Duplication will occur). In many cases the previous full request will be deleted and a new full will be loaded.
    Regards,
    B

  • R3 Table Required for data load status

    Hi all,
    I am in 3.x version so rsstatmanpart(fast table only available in bi7) wont work.
    I want the no of records added and transferred for a specific cube on a specific date.
    Thanks in advance.

    Check Tables RSMONICTAB, RSMONFACT, RSMONICDP
    Hope this helps..
    /pradeep

  • Help Required for date display & mutating trigger

    Hai all..
    I need clarification for 2 problems.
    Problem 1
    Is there a possibility to display the values with a date range.
    I mean if there is no data for a particular date within the given range. it should display some null value.
    Eg.
    table name : sample
    Sno Date1 Name desgn
    1 1-may-2004 xyz engr
    1 3-may-2004 xyz engr-1
    1 4-may-2004 xyz engr-2
    The query is
    select * from sample where date1 between '01-may-2004' and '04-may-2004'.
    The result of query will be
    Sno Date1 Name desgn
    1 1-may-2004 xyz engr
    1 3-may-2004 xyz engr-1
    1 4-may-2004 xyz engr-2
    But i need the result to be displayed as
    Sno Date1 Name desgn
    1 1-may-2004 xyz engr
    nil 2-may-2004 nil nil
    1 3-may-2004 xyz engr-1
    1 4-may-2004 xyz engr-2
    Pls let me know how to achieve this.
    Problem 2:
    While i am updating a value using a trigger.
    I am getting "mutating" error.
    Is there a possibility to avoid the mutation.
    The two triggers i am using is
    Trigger 1
    CREATE OR REPLACE TRIGGER TRIG_DURATION_UPDATED
    AFTER UPDATE ON
    TIMESHEET_DATA
    FOR EACH ROW
    DECLARE
    ID_COUNT DATE;
    A date;
    NN NUMBER;
    AA NUMBER;
    NN1 NUMBER;
    BEGIN
    b_dur :=0;
    IF UPDATING('PROGRESS') THEN
    SELECT COUNT(1) INTO AA FROM TIMESHEET_DATA_PAUSE WHERE DATA_sNO = :OLD.DATA_SNO AND EMP_ID = :OLD.EMP_ID AND
    ACNUM = :OLD.ACNUM AND DURATION IS NULL;
    IF AA > 0 THEN
    SELECT to_date(START_TIME,'DD-MON-YYYY HH:MI AM') into a FROM TIMESHEET_dATA_PAUSE WHERE
    EMP_ID = :NEW.EMP_ID AND ACNUM = :new.acnum AND DATA_SNO = :new.data_sno AND pause_TIME IS NULL
    AND DURATION IS NULL;
    select to_date(sysdate,'DD-MON-YYYY HH:MI AM') into id_count from dual;
    SELECT A - ID_COUNT INTO NN FROM DUAL;
    SELECT round(TRUNC(NN*24),2) INTO NN1 FROM DUAL;
    UPDATE TIMESHEET_dATA_PAUSE SET DURATION = NN1,PAUSE_TIME = SYSDATE WHERE
    EMP_ID = :NEW.EMP_ID AND ACNUM = :new.acnum
    AND DATA_SNO = :new.data_sno AND pause_TIME IS NULL AND DURATION IS NULL;
    END IF;
    END IF;
    END;
    Trigger 2
    CREATE OR REPLACE TRIGGER TRIG_DURATION_UPDATED_PAUSE
    AFTER UPDATE ON
    TIMESHEET_DATA_PAUSE
    FOR EACH ROW
    DECLARE
    ID_COUNT DATE;
    A_COUNT NUMBER(10);
    A date;
    NN NUMBER;
    AA NUMBER;
    NN1 NUMBER(10,2);
    BEGIN
    IF UPDATING('DURATION') THEN
    SELECT COUNT(1) INTO A_COUNT FROM TIMESHEET_DATA WHERE DATA_sNO = :new.data_sno AND EMP_ID =:new.EMP_ID
    AND ACNUM = 'ACNUM' AND DURATION IS NULL;
    IF A_COUNT >0 THEN
    SELECT NVL(SUM(DURATION),0) INTO NN1 FROM TIMESHEET_DATA_PAUSE WHERE DATA_sNO = :new.data_sno AND EMP_ID =:new.EMP_ID AND ACNUM = :NEW.ACNUM;
    UPDATE TIMESHEET_DATA SET DURATION = NN1 WHERE DATA_sNO = :new.data_sno AND EMP_ID =:new.EMP_ID AND ACNUM = :NEW.ACNUM AND DURATION IS NULL;
    END IF;
    END IF;
    EXCEPTION WHEN DUP_VAL_ON_INDEX THEN
    DBMS_OUTPUT.PUT_LINE ('CANNOT INSERT');
    END;
    Pls let me know the solutions for the two..
    Thanks in Advance
    Gaya3

    In future please start two threads when you have two questions to ask. Othewise it makes things difficult to follow and may well lead to some of your questions not getting an answer. Unlike some other sites there is no limit to the number of posts we can have :)
    To answer your first question:
    If you have Oracle9i you can do this using a PIPELINED function to act as a pivot table (in earlier versions you will have to have a table of some description - ALL_OBJECTS will do in a pinch).
    CREATE OR REPLACE TYPE t_dates AS TABLE OF DATE
    CREATE OR REPLACE FUNCTION f_dates
      (st IN DATE, cnt IN INTEGER) RETURN t_dates
      PIPELINED
    AS
      dd date;
    BEGIN
      FOR a IN 1..cnt LOOP
        SELECT  st + (a-1)
          INTO  dd
          FROM  dual;
        PIPE ROW (dd);
      END LOOP;
      RETURN;
    END f_dates;
    /So, given this data...
    SQL> select * from t57
      2  /
          COL1 COL2
             3 05-MAY-04
             4 06-MAY-04
             5 07-MAY-04
             6 12-MAY-04
             8 15-MAY-04
    5 rows selected.
    SQL>...we can run this query...
    SQL> SELECT nvl(to_char(t.col1), 'Nil'), dd.ndate
      2  FROM   t57 t
      3         , ( select column_value AS ndate from table(cast(f_dates(trunc(SYSDATE), 11) AS t_dates))) dd
      4  WHERE trunc(dd.ndate) = trunc(t.col2(+))
      5  /
    NVL(TO_CHAR(T.COL1),'NIL')               NDATE
    3                                        05-MAY-04
    4                                        06-MAY-04
    5                                        07-MAY-04
    Nil                                      08-MAY-04
    Nil                                      09-MAY-04
    Nil                                      10-MAY-04
    Nil                                      11-MAY-04
    6                                        12-MAY-04
    Nil                                      13-MAY-04
    Nil                                      14-MAY-04
    8                                        15-MAY-04
    11 rows selected.
    SQL> Obviously, you will want to tidy up the function to suit your needs.
    Cheers, APC

  • Help with data load model

    Hi,
    I need help with a data load model. First, i'm doing delta extraction from R/3, we load data with a InfoSource to InfoCube A and InfoCube B.
    I'm doing master data validation on the load, so if a load fails for InfoCube A, it fails for InfoCube B too (this is because i can have only 1 InfoPackage for the 2 infocubes, because of the delta update).
    So i propose a new model in wich:
    - The delta load is taked first to an ODS.
    - ODS is cleaned before the delta update.
    - Then i create 2 InfoPackages for full load from ODS to  Infocube A, and from ODS to InfoCube B.
    With this solution i can have 2 infopackages from ODS because i'm not doing a delta load from here to the cubes, and with 2 infopackages i can have independent validations for each cube so if one of them fails, the other can still be loaded sucessfully.
    The solution fails because if i load delta from R/3 to the ODS i can't clean it first. The initialization and the old updates needs to be previuslly loaded on the ODS. Then i can't do full load to the cubes and neither have 2 infopackages.
    Please help me to solve this issue.
    thanks a lot

    Hi jeremy,
    what about this simple solution:
    load data by delta from R/3 in your ODS. You can also have an ODS/cube for the historical data which is more space-saving than holding all the old data in PSA. Then you load your historical data from PSA into the historical ODS/cube.
    From your ODS with the actual data you update your requests by full from ODS into the cubes with 2 different full infopackages. But because you load by full you have to use deletion selections in the infopackages to avoid duplicate data!
    regards,
    Jürgen

  • How to configure once data load then trigerd or run ibot?

    Hi Experts,
    I have a one requirement,
    1) Every day run one workflow( means data load into data warehouse)
    2) After, ibot should be run and delivery to users.
    3) We scheduled the workflows in DAC for every day morning.
    Requirement:
    Once data loaded, then IBot should be run and send to users dynamically (without scheduling).
    If workflow failed, IBot won’t be delivered.
    How to find out or configure once data load then trigerd or run ibot?
    I am using obi 10g and informatica 8 and os xp.
    Advance thanks..
    Thanks,
    Raja

    Hi,
    Below are the details for automating the OBIEE Scheduler.
    Create a batch file or Sh file with following command
    D:\OracleBI\server\Bin\saschinvoke -u Administrator/udrbiee007 -j 8
    -u is username/Password for Scheduler (username/password that u given while configuration)
    -j is a job id number when u create a ibot it will assign the new job number that can be identified from"Jobmanager"
    Refer the below thread for more information.
    iBot scheduling after ETL load
    Or ,
    What you the above also it will work but problem is we need specify the time like every day 6.30 am .
    Note: The condition report is true then the report will be delivered at 6.30 pm only but the condition is false the report will not triggered.
    I also implemented this but that is little bit different .
    Hope this help's
    Thanks
    Satya
    Edited by: Satya Ranki Reddy on Jul 13, 2012 12:05 PM

  • Data load failed while loading data from one DSO to another DSO..

    Hi,
    On SID generation data load failed while loading data  from Source DSO to Target DSO.
    Following are the error which is occuuring--
    Value "External Ref # 2421-0625511EXP  " (HEX 450078007400650072006E0061006C0020005200650066
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0BBP
    So, i'm  not getting  WHY in one DSO i.e Source  it got successful but in another DSO i.e. Target its got failed??
    While analyzing all i check that SIDs Generation upon Activation is ckecked in source DSO but not in Target DSO..so it is reason its got failed??
    Please explain..
    Thanks,
    Sneha

    Hi,
    I hope your data flow has been designed in such a way where the 1st DSO as a staging Device and all transformation rules and routine are maintained in between 1st to 2nd dso and sid generation upon activation maintained in 2nd DSO.  By doing so you will be getting your data 1st DSO same as your source system data since you are not doing any transformation rules and routine etc.. which helps to avoid data load failure.  
    Please analyze the following
    Have you loaded masterdata before transaction data ... if no please do it first
    go to the property of first dso and check whether there maintained sid generation up on activation (it may not be maintained I guess)
    Goto the property of 2nd Dso and check whether there maintained sid generation up on activation (It may be maintained I hope)
    this may be the reason.
    Also check whether there is any special char involvement in your transaction data (even lower case letter)
    Regards
    BVR

  • Data load performance using infoset Vs View.

    Hi Guru,
    I am performing generic extraction in that i am loading data to cube but my Data source is based on Infoset in R/3.
    it is taking 30 MIn. to load 10,00000 Lakh (Ten Lakh) records ideally it has to take 10 to 15 min. rit ?
    can anybody suggest me where i need to check for increase in performance or shall i create datasource over a view and try to load data will it help me in data load performance ?
    thanks,
    ganesh.

    hi Ganesh,
    Primary Index ->
    When you create a database table in the ABAP Dictionary, you must specify the combination of fields that enable an entry within the table to be clearly identified. The key fields must be specified at the top of the table field list, and define them as key fields. A minimum of 1 key field and a maximum of 16 key fields can be defined.
    When the table is activated, an index formed from all key fields is created on the database (with Oracle, Informix, DB2), in addition to the table itself. This index is called the primary index The primary index is unique by definition.
    In addition to the primary index you can define one or more secondary indexes for a table in the ABAP Dictionary, and create them on the database. Secondary indexes can be unique or non-unique.
    If you dispatch an SQL statement from an ABAP program to the database, the program searches for the data records requested either in the database table itself (full table scan) or by using an index ( index unique scan or index range scan). If all fields requested are found in the index using an index scan, the table records do not need to be accessed.
    The index records are stored in the index tree and sorted according to index field. This enables accelerated access using the index The table records in the table blocks are not sorted.
    An index should not consist of too many fields. Having a few very selective fields increases the chance of reusability, and reduces the chance of the database optimizer selecting an unsuitable access path.
    To create Index ->
    Yo have to use trx SE11 into Dev system.
    Enter the database table name and press
    Display -> Indexes -> Create
    Enter index name.
    Choose Maintain logon language.
    Enter short description and index fields.
    Then press save and create the request to transport the index to QA and PRD. Then press activate.
    Hope this helps,
    VA
    Edited by: Vishwa  Anand on Sep 29, 2010 12:50 PM

Maybe you are looking for