Selective load into setup table

Hello,
Is there any way we can load multiple selections into setup table in OLI7BW. Currently I am only able to load range of values. But i have a list of almost 2000 records and they are not in sequence. And loading one by one requires too much effort. Loading ranges would load millions of records.
Please advice.
Regards,
Tints

Hi Tinu,
Pick all 200 0 orders and check in VBAK table for CREATED ON field.
Down load all 2000 orders into excel and put filter on created on value....then note down all the created on dates.
Now you can load orders based on created on date. make sure that you are loading the data to DSO as it avoids duplication
Regards
SujanR

Similar Messages

  • Mainframe data loaded into Oracle tables - Test for low values using PL/SQL

    Mainframe legacy data has been copied straight from the legacy tables into mirrored tables in Oracle. Some columns from the mainframe data had 'low values' in them. These columns were defined on the Oracle tables as varchar2 types. In looking at the data, some of these columns appear to have data that looks like little square boxes, not sure but maybe that is the way Oracle interprets the 'low values' in the original data into varchar. When I run a select to find all rows where this column is not null, it selects these columns. In the results of the select statement, the columns appear to be blank, however, in looking at the data in the column using SQL Developer, I can see the odd 'square boxes'. My guess is that the select statement is detecting that something exists in this column. Long story short, some how I am going to have to test this legacy data on the Oracle table using Pl/Sql to test for 'low values'. Does anyone have any suggestions on how I could do this????? Help! The mainframe data we are loading into these tables is loaded with columns with low values.
    I am using Oracle 11i.
    Thanks
    Edited by: ncsthbell on Nov 2, 2009 8:38 AM

    ncsthbell wrote:
    Mainframe legacy data has been copied straight from the legacy tables into mirrored tables in Oracle. Not a wise thing to do. Mainframe operating systems typically use EBCDIC and Unix and Windows servers use ASCII. The endian is also different (big endian vs little endian).
    Does anyone have any suggestions on how I could do this????? As suggested, use the SQL function called DUMP() to see the actual contents (in hex) of these columns.

  • Load from Setup tables breaking for high volume

    Hello Friends,
    There are around 50,000 records in the setup table for 03 related appl area. The load is breaking for 50,000 records in BW. But it works if I further select only one day data in the BW scheduler.
    Why is this?
    Immediate help is appreciated.
    Regards,
    Simmi

    Hello gurus,
    So, I tried reducing the packet size to 20,000 and no of Idocs to 10 per Infopckage. But this is working for only 40K records which a just a weeks data for the 2LIS_03_BF extractor.
    If I run 30 days worth of load form setup tables for the same etxractor, the load is breaking.
    The same is happening for a customer built Zextractor. For this I actually tried with 10,000 records and 5 IDOCS per package.
    Does, this mean for sure that there is a memory issue?
    Also, I see that all the BGD WPs have been occupied by other jobs. Is this an issue?
    But we have deltas running every night from R3, but t hese are less volumes(<10,000)and we have not seen this issue.
    However within BW there are heavy loads between data targets and they don't break.
    Is this a issue only for full loads from R3?
    Appreciate your immediate response.
    Thanks
    Simmi
    Edited by: simmi on Jun 20, 2008 1:39 AM

  • 2LIS_03_UM - FULL LOAD VIA SETUP TABLE NOT EQUAL TO DELTA

    Hello,
    I try to load stock data with 2LIS_03_UM.
    When I extract data using setup table and full load I don't get the same records as I get with delta load. Although the key figures values are the same, the documents numbers are different. Most documents from delta load are ML (material ledger) documents while documents from full load (via setup table) are mainly FI documents.
    What can be the reason? How can I fix it?
    Thank you,
    Hadar

    Hi,
    Delta load extracts data from Delta Queue based on your update mode selected in LBWE.
    Full Load or Init load extracts data from Setup tables, here if you are extracting data from setup tables ? did you fill the setup tables now or not after deleting the existing data of the setup tables?
    are you getting same values when compare to R/3 and BW data for some document numbers?

  • Is it possible to automate the load of setup tables for LIS datasources?

    Hello,
    For LIS datasources, eg. 04 (shop floor control), once the production confirmation is done, we have to manually run the transaction OLI4BW to load the setup tables before extracting data to BW. Is there a way to automate the load of setup tables?
    Regards,
    Suraj

    Hi,
    Yes we can !
    Create variants for small ranges and scheduled all of them at a time. So that, loading will be completed soon.
    Thanks,
    Saru

  • Extracting explanations from planning app and loading into Oracle table

    Hi All,
    I had a requirement where I had to extract data from a planning application through ODI 11g and load it into Oracle RDBMS.
    I used essbase as my source in technology (since planning data is stored on essbase side) and oracle as my target.
    Now the data is getting extracted from essbase side and is getting loaded into Oracle table through ODI.
    Now the client requires that they want to extract the explanations or text values also from planning application and load them into Oracle table.
    How this can be achieved?Is there a table on the sql side(since sql database is being used at back end for planning app) which stores the explanations,if yes please let me know which table it is.
    Kindly help me with this requirement.

    Hi,
    IKM SQL Control Append is perfect if you don't need incremental updates. If you need it, go for IKM Oracle Incremental Update (MERGE) or something like that.
    Regards,
    JeromeFr

  • Select Insert into staging table does not load varchar2

    I have a select insert query that loads a prebuilt table. I am running this query through Heterogenous services which points to a MySql database.
    If I run just a select query in SQL PLUS then it returns all rows just fine.
    However, when I put the select insert into a stored procedure, compile it, then run it. It does not load any VARCHAR2 datatypes but it does load DATE and NUMBER datatypes.
    I also have another table with VARCHAR2 and it loads fine.
    How can the database only load some of the items in a row and not others (VARCHAR2)
    For instance, the query will load TRANSACTION_ID, ORER_NUMBER but will not load any data into BILLFIRST_NAME, and their is data to load.
    CREATE TABLE "TRAN_STG"
    (     "TRANSACTION_ID" NUMBER(11,0),
         "ORDER_NUMBER" NUMBER(20,0),
         "BILL_FIRST_NAME" VARCHAR2(100 BYTE),
    Any suggestions would be appreciated.

    can it be that there are differences in characterset? Can it be that there are differences in datatypes? Can it be that there are differences in field length?

  • Selective Deletion and Data Load from Setup Tables for LIS Extractor

    Hi All,
    We came across a situation where one of the delta in PSA was missed to load in DSO. This DSO updates another cube in the flow. Since it has been many days since this miss come in our knowledge we need to selectively delete data for those documents from DSO & Cube and then take a full load for the Documents filling the setup table.
    Now what will be the right approach to load this data from setup table > DSO > cube. There is change log present for those documents and a few KPI's in DSO are in summation mode.
    Regards
    Jitendra

    thanks Ajeet!!!!
    This is Sales Order extractor, the data got loaded to ODS just fine, but since the data is coming to ODS from different extractor. Everything is fine in ODS, but not in the cube. Will Full repair request and Full load would it make difference when the data is going to cube? I thought that it would matter only if I am loading to ODS.
    what do you mean "Even if you do a full load without any selections you should do a full repair ".
    thanks.
    W

  • Multiple parallel loads into one table with TABLE LOCK option

    Hi everyone:
    We have several SSIS packages where each package has the basic design of having one sequence container. Inside each sequence container can be anywhere from 2 - 9 data flow tasks where for each data is selected from a different table but all 2-9 tasks do
    an OLEDB fast load (with table lock option checked) into the same single destination table.
    The number of rows were pulling from the various sources might be anywhere from 5 up to 100,000.
    Right now this doesn't seem to be causing any issues, but I wanted to check to see if this set up (since we're doing a review) could potentially cause problems down the road?
    We seem to think that each parallel task will acquire its data as normal, and just won't be able to insert until one of the other parallel tasks completes its fast load. To us, that's no big deal as we're at least able to acquire the data in parallel.
    What are everyone's thoughts?
    Thanks!

    >>>We seem to think that each parallel task will acquire its data as normal, and just won't be able to insert
    until one >>>of the other parallel tasks completes its fast load. 
    That is correct observation.
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • Error displaying a jpg file loaded into a table with blob field

    This may not be the correct forum for this question, but if it isn't could someone direct me to the correct one.
    I have created a table with a blob field in which I have loaded a jpg image. This appeared to work correctly, but when I try to display the image in internet explorer it comes back and tells me that it doesn't recognize the file type. Enclosed is the table create, load, and display pl/sql code. Can anyone tell me what I am doing wrong. Thanks. For the mime/header I used owa_util.mime_header('images/jpg') because my image is a jpg file.
    The database is 10g
    -- Create table
    create table PHOTOS
    IMAGEID NUMBER(10),
    IMAGE BLOB,
    IMAGE_NAME VARCHAR2(50)
    load image
    CREATE OR REPLACE PROCEDURE load_file ( p_id number, p_photo_name in varchar2) IS
    src_file BFILE;
    dst_file BLOB;
    lgh_file BINARY_INTEGER;
    BEGIN
    src_file := bfilename('SNAPUNCH', p_photo_name);
    -- insert a NULL record to lock
    INSERT INTO photos (imageid, image_name, image)
    VALUES (p_id , p_photo_name, EMPTY_BLOB())
    RETURNING image INTO dst_file;
    -- lock record
    SELECT image
    INTO dst_file
    FROM photos
    WHERE imageid = p_id AND image_name = p_photo_name
    FOR UPDATE;
    -- open the file
    dbms_lob.fileopen(src_file, dbms_lob.file_readonly);
    -- determine length
    lgh_file := dbms_lob.getlength(src_file);
    -- read the file
    dbms_lob.loadfromfile(dst_file, src_file, lgh_file);
    -- update the blob field
    UPDATE photos
    SET image = dst_file
    WHERE imageid = p_id
    AND image_name = p_photo_name;
    -- close file
    dbms_lob.fileclose(src_file);
    END load_file;
    display image
    PROCEDURE display_image(p_id NUMBER) IS
    Photo BLOB;
    v_amt NUMBER DEFAULT 4096;
    v_off NUMBER DEFAULT 1;
    v_raw RAW(4096);
    BEGIN
    -- Get the blob image
    SELECT image
    INTO Photo
    FROM PHOTOS
    WHERE IMAGEID = p_id;
    owa_util.mime_header('images/jpg');
    BEGIN
    LOOP
    -- Read the BLOB
    dbms_lob.READ(Photo, v_amt, v_off, v_raw);
    -- Display image
    htp.prn(utl_raw.cast_to_varchar2(v_raw));
    v_off := v_off + v_amt;
    v_amt := 4096;
    END LOOP;
    dbms_lob.CLOSE(Photo);
    EXCEPTION
    WHEN NO_DATA_FOUND THEN
    NULL;
    END;
    END;
    The url I enter is: http://webdev:7777/tisinfo/tis.tiss0011.Display_Image?p_id=1

    Just a little more information. When I enter owa_util.mime_header('image/jpeg') I can't display the file. It just shows up with a red x for the file.
    When I enter owa_util.mime_header('image/jpg') it displays the file, but in the format
    ¿¿¿¿JFIF¿¿-Intel(R) JPEG Library, version [2.0.16.48]¿¿C
    This is the way I would expect it to look if I opened it with Notepad, or an application that doesn't recognize jpg files. Can anyone tell me what I am doing wrong?? Thanks.

  • SQL*Loader loading into different tables

    Hi
    I have a data file as below with column names Emp ID,Emp name ,Dept id,Dept name
    Emp ID,Emp name ,Dept id,Dept name
    10,name 1,100,Dept 1
    11,name 2,101,Dept 2
    12,name 3,102,Dept 3
    I want to load this into two tables employee and depatment as below
    Employee table
    Emp_id Emp_name
    10 name 1
    11 name 2
    12 name 3
    Department table
    Dept_id Department_name
    100 Dept 1
    101 Dept2
    102 Dept 3
    Please help

    What have you tried so far ? If you use delimiters to define your columns (as in Example 2), you should be able to load your data.
    SQL> create user srini identified by xxxxx default tablespace INTERIM temporary tablespace temp;
    User created.
    SQL> grant dba to srini;
    Grant succeeded.
    SQL> connect srini/xxxxx
    Connected.
    SQL> create table emp (emp_id number(3), emp_name varchar2(50));
    Table created.
    SQL> create table dept (dept_id number(3), dept_name  varchar2(50));
    Table created.
    SQL> select * from emp;
    no rows selected
    sqlldr srini/xxxxx CONTROL=input.ctl LOG=input.log
    SQL*Loader: Release 10.2.0.3.0 - Production on Thu Jan 20 11:44:01 2011
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Control File:   input.ctl
    Data File:      input.dat
      Bad File:     input.bad
      Discard File:  none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array:     64 rows, maximum of 256000 bytes
    Continuation:    none specified
    Path used:      Conventional
    Table EMP, loaded from every logical record.
    Insert option in effect for this table: INSERT
       Column Name                  Position   Len  Term Encl Datatype
    EMP_ID                              FIRST     *   ,       CHARACTER
    EMP_NAME                             NEXT     *   ,       CHARACTER
    Table DEPT, loaded from every logical record.
    Insert option in effect for this table: INSERT
       Column Name                  Position   Len  Term Encl Datatype
    DEPT_ID                              NEXT     *   ,       CHARACTER
    DEPT_NAME                            NEXT     *  WHT      CHARACTER
    Table EMP:
      3 Rows successfully loaded.
      0 Rows not loaded due to data errors.
      0 Rows not loaded because all WHEN clauses were failed.
      0 Rows not loaded because all fields were null.
    Table DEPT:
      3 Rows successfully loaded.
      0 Rows not loaded due to data errors.
      0 Rows not loaded because all WHEN clauses were failed.
      0 Rows not loaded because all fields were null.
    Space allocated for bind array:                  66048 bytes(64 rows)
    Read   buffer bytes: 1048576
    Total logical records skipped:          0
    Total logical records read:             3
    Total logical records rejected:         0
    Total logical records discarded:        0
    SQL> select * from emp;
        EMP_ID
    EMP_NAME
            10
    name 1
            11
    name 2
            12
    name 3
    SQL> select * from dept;
       DEPT_ID
    DEPT_NAME
           100
    Dept
           101
    Dept
           102
    Dept
    $ more *dat
    10,name 1,100,Dept 1
    11,name 2,101,Dept 2
    12,name 3,102,Dept 3When you try, you succeed ;-)
    Srini

  • How to select data in setup tables based on date

    hello gurus,
    I have a question . I want to fill the setup tables but not complete historical data. I need to fill it based on some date selection.Can anyone help me the process. I  went to oli9bw but i cannot see any option like date range for data selection...... i.e., I need to extract a subset of data from ecc to bw.
    Thanks a lot

    Hi Sandeep,
    While filling Setup tables we won't have any selection(It loads all historical data), when you run the IP to load data to BW.follow the below 2 cases.
    1)If you require further delta loads on this, run the full load based on your date selections.then create another IP for delta,run "init without data transfer" as option in Update tab of IP.
    2)If you don't require any delta further, simply run Full load with date selections.
    Hope this helps.
    Regards,
    Venkatesh.

  • Design approach for XML load into DB table

    The question is more related to find a design approach.
    1. I have a few XML files to load data from. There will be more files which will be loaded into the DB periodically.
    2. Each file populates 4 tables.
    3. After loading all the data each file has to be moved into an archived folder.
    4. We have a STAGE & PROD setup in the DB. The usual approach is to create 2 packages for EACH table.
    One to load data into STAGE
    2nd to move data from STAGE to PROD
    Now, I'm not sure if I should create 4 tables * 2 = 8 packages or just 2 packages (1 for stage + 1 for prod).
    Please suggest!! I know the experts here have already come across such a scenarios.
    I wanted to know pros and cons of each approach. As I'm doing something like this for the first time, I can't imagine which will be the best solution.
    Thanks!!!

    I'd create one package.
    I do not see why de-coupling the loading to staging from prod needs to be so.
    When it works as one unit you can rollback the entire load as one as you may not want to end up with some data partially loaded.
    Arthur
    MyBlog
    Twitter

  • How can I make an easy *.CSV file to load into database table

    Hi All,
    I have a huge excel sheet having columns item#, description and qty. The description column sometimes maybe one word name, two word name separated with space or may be , spearated name. I want to write and PL/SQl code which will read this file and load it into database table. Now the *.CSV file is either comma delimited or tab text delmited which both do not solve my issue. Is there any better solution with anyone which can prevent the manual editing to the *.CSV file and I can easily load it to table.
    Your help is appreciated,
    Thanks
    Zahir

    SQL*Loader is probably the fastest method, but since you specifically asked for a PL/SQL method:
    http://asktom.oracle.com/pls/ask/f?p=4950:8:::::F4950_P8_DISPLAYID:464420312302

  • How to get all values in the range of select option into internal table?

    Hi,
    I need to capture all entries coming in the range of select option into one internal table.
    How to do get that?
    For E.g
    select-options: matnr for mara-matnr.(select option)
    IF I enter G0100013507892 as lower value of matnr and G0100014873947 as higher value
    and if there are 10,000 materials in the above range, then I want to capture all theses 10000 materails in one internal table. How to do that?
    Regards,
    Mrunal

    Hello Mrunal Mhaskar  ,
    What i understand you can do one thing  go in debug mode
    Try this code : -
    LOOP AT s_matnr_ex.
      IF s_matnr_ex-low IS NOT INITIAL.
        i_matnr-matnr = s_matnr_ex-low.
        i_matnr-option = s_matnr_ex-option.
        APPEND i_matnr.
        CLEAR : i_matnr.
      ENDIF.
    ENDLOOP.
    LOOP AT s_matnr_ex.
      IF s_matnr_ex-high IS NOT INITIAL.
        i_matnr-matnr = s_matnr_ex-high.
        i_matnr-option = s_matnr_ex-option.
        APPEND i_matnr.
        CLEAR : i_matnr.
      ENDIF.
    ENDLOOP.
    In the i_matnr table high and low values are there.
    Regards,
    Vandana.

Maybe you are looking for