Data pulling from MARD, MARC etc. - Which table first I hv to use?

Hi Experts,
User enters the Matnr, plant, storage location inputs in the selections.
It is need to get the corresponding data from MARD, MARC, MDBS, VBBE.
So,
1) Is it possible that, suppose MARD having 10 matnrs, where as MARC is having 15, where as MDBS have only 7? or every table does hv 10 definately?
2) as I need to pull the data from these tables, pls. suggest me, in which order I hv to pull, I mean, first from Which table i hv to pull, and as second, wht is next table? for more reliability.
thanq

HI
select matnr
          LBSTF
   from mard
into table itab1
where matnr = p_matnr
condition.
if not itab[] is initial.
select EISBE
     from marc
  into table itab1
for all entries in itab
where matnr = itab-matnr
and any other condition
select MENGE
      from MDBS
  into table itab2
for all entries in itab1
where matnr = itab1-matnr condition
endif.
write like this
you will get data if availble

Similar Messages

  • BAPI - DATA TRANSFER FROM .XLS FILE TO DB TABLE

    I am upload data from xls file to database table.
    first i have load xls file data to internal table (it_data) that is working.
    then from internal table to database table (zskv_rate) by function module (ZBAPI_RATE_SAVEDATA) which is not working.
    ex.
        LOOP AT IT_DATA.
    Header
        ZSKV_RATE-DIVN = IT_DATA-DIVN.
        ZSKV_RATE-REGN = IT_DATA-REGN.
        ZSKV_RATE-PLANT = IT_DATA-PLANT.
        ZSKV_RATE-RATE = IT_DATA-RATE.
       APPEND IT_DATA.
        CALL FUNCTION 'ZBAPI_RATE_SAVEDATA'
          EXPORTING
            HEADDATA = ZSKV_RATE
          IMPORTING
            RETURN = BAPI_RETURN
          TABLES
            DATA = IT_DATA
        IF BAPI_RETURN-TYPE = 'E'.
            WRITE:/ 'Error:' ,BAPI_RETURN-MESSAGE ,'for material:' ,IT_DATA-DIVN.
        ELSEIF BAPI_RETURN-TYPE = 'S'.
            WRITE: 'Successfully created material' ,IT_DATA-DIVN.
        ELSE.
            WRITE:/ 'NONE' ,IT_DATA-DIVN.
        ENDIF.
      ENDLOOP.
    how to transfer data from internal table to database table.

    Hi Sanjeev,
    as u hv uploaded the excel file into  the internal table,
    u can update the ztable from internal table..
    modify <dbtab>from itab.
    regards.
    Mdi.Deeba

  • Data pulling from R3 to BW server in BI7.0

    Hi all,
    I am very new in BI7.0 but know the concept of  pulling the data in BW3.5 .
    I create a generic data source and replicate it in to BI7.0 .Now i am confuse in BI server how to i pull the data from R3 server Bcz each and every screen is differ from BW3.5 can anybody tell me the way for data pulling in BI 7.0 .or give me some idea abt BI7.0 data pulling steps.
    Thanks and Regards.
    Ankit modi.

    Hi Ankit
      As of BI 7.0 there were several changes in the data loading process.
      First of all, you need to replicate the DataSource (here you can choose between use a 3.5 or 7.0 replication. I recomend the last one)
      Now you are not obligue to create an InfoSource.
      Then you have to connect your InfoProvider (DSO, InfoCube, etc) to the DataSource. This is made with a Transformation (the symbol is similar to the previous Update Rules but without an small square/dot in the graphic). Once you have created the Transformation you connect the fields of the DataSource with the InfoObject of your DataProvider.
      Finally you'll create an InfoPackage to load data from your R/3 system to PSA and then you'll create a DTP to load the data from the PSA to the InfoProvider.
    Hope it helps.
    Kindly regards.
    Germán.-

  • Data Load from XML file to Oracle Table

    Hi,
    I am trying to load data from XML file to Oracle table using DBMS_XMLStore utility.I have performed the prerequisites like creating the directory from APPS user, grant read/write to directory, placing the data file on folder on apps tier, created a procedure ‘insertXML’ to load the data based on metalink note (Note ID: 396573.1 How to Insert XML by passing a file Instead of using Embedded XML). I am running the procedure thru below anonymous block  to insert the data in the table.
    Anonymous block
    declare
      begin
      insertXML('XMLDIR', 'results.xml', 'employee_results');
      end;
    I am getting below error after running the anonymous block.
    Error :     ORA-22288: file or LOB operation FILEOPEN failed”
    Cause :   The operation attempted on the file or LOB failed.
    Action:   See the next error message in the error stack for more detailed
               information.  Also, verify that the file or LOB exists and that
               the necessary privileges are set for the specified operation. If
               the error still persists, report the error to the DBA.
    I searched this error on metalink and found DOC ID 1556652.1 . I Ran the script provided in the document. PFA the script.
    Also, attaching a document that list down the steps that I have followed.
    Please check and let me know if I am missing something in the process. Please help to get this resolve.
    Regards,
    Sankalp

    Thanks Bashar for your prompt response.
    I ran the insert statement but encountered error,below are the error details. statement.
    Error report -
    SQL Error: ORA-22288: file or LOB operation FILEOPEN failed
    No such file or directory
    ORA-06512: at "SYS.XMLTYPE", line 296
    ORA-06512: at line 1
    22288. 00000 -  "file or LOB operation %s failed\n%s"
    *Cause:    The operation attempted on the file or LOB failed.
    *Action:   See the next error message in the error stack for more detailed
               information.  Also, verify that the file or LOB exists and that
               the necessary privileges are set for the specified operation. If
               the error still persists, report the error to the DBA.
    INSERT statement I ran
    INSERT INTO employee_results (USERNAME,FIRSTNAME,LASTNAME,STATUS)
        SELECT *
        FROM XMLTABLE('/Results/Users/User'
               PASSING XMLTYPE(BFILENAME('XMLDIR', 'results.xml'),
               NLS_CHARSET_ID('CHAR_CS'))
               COLUMNS USERNAME  NUMBER(4)    PATH 'USERNAME',
                       FIRSTNAME  VARCHAR2(10) PATH 'FIRSTNAME',
                       LASTNAME    NUMBER(7,2)  PATH 'LASTNAME',
                       STATUS  VARCHAR2(14) PATH 'STATUS'
    Regards,
    Sankalp

  • Saving data retrieved from text file to JDBC table

    Hi,
    I would like to save data, loaded from a text file to a dataset, into an existing JDBC table.
    The documentation I found on this subject is very brief, and no matter how I tried, I couldn't succeed.
    I would appreciate any help.
    Thank you in advance

    Hi there,
    Here is an example of reading a String and a date from a text file. Once you have what you need in the Strings (or whatever) it should be easy to insert them into and SQL statement. Let me know if you need an example of this as well and I'll post it in the morning. Right now I'm off to bed :o)
    Cheers
    Mark
    import java.io.*;
    import java.util.*;
    import java.text.*;
    public class WriteTest {
    public static void main(String [] arg) throws Exception {          
       Date today = new Date();          
       String name = "";          
       SimpleDateFormat formatter = new SimpleDateFormat("ddMMyyyy");          
       File myfile = new File("hello.txt");               
       // how to input     
       BufferedReader in = new BufferedReader(new FileReader(myfile));          
       String name="",
       date="";          
       boolean eof = false;          
       do {
          name = in.readLine();               
          date = in.readLine();               
          if(!(eof = (name==null || date==null)))                    
          System.out.println(name+" "+formatter.parse(date));
       while(!eof);     
    }

  • Loading Date fields from flat file to Oracle tables

    Hi,
    I have a flat file with a few date columns. I have given the format for the date field while creating the data store for the flat file. The format I used is 'YYYY-MM-DD'.
    But I get an error when I execute it after associating the module with LKM & IKM.
    The error message is as follows:
    org.apache.bsf.BSFException: exception from Jython: Traceback (innermost last):
    File "<string>", line 3, in ?
    OS command has signalled errors
    Can anyone help me out.
    Thanks

    At the time of DataStore creation. You take that date field as string.
    After this, in interface, when you map this date field with table field. You should use CONVERT(<field name>,DATE) in expression editor.

  • SQL Server 2005 data pull from Oracle

    Hi, I am learning how to use Oracle and I ran into this problem.
    My database is running on SQL Server 2005.
    I have an ODBC connection to Oracle. I also have a VB .Net script that queries the Oracle table that the data resides in.
    When the job is running, and if it stalls, I do not get a timeout error. It locks down my database and no other source systems can feed data to me; these are back logged in the queue.
    Is there an easier way for me to make a connection that will pull data on a consistent basis? Please help ...
    Student Learner

    Learning how to use Oracle does not include taking data from Oracle and putting it into SQL Server.
    From your description of what is happening I agree with BluShadow. This is not an Oracle issue. This is a coding in VB .NET issue.
    Posting the statements you are executing against some unknown version of Oracle would be a good starting point.

  • Convert/Cast string data type from file to DT_GUID into table

    Hi to all
    is there anyone who help me please?!?
    I'm not able to resolve the following problem.
    I have a .csv file with a string field A and its values are GUID.
    I must export this .csv file into SQL Server Table which have a uniqueidentifier (GUID) field B.
    I must map string field A to GUID field B of table.
    But I get an error like this: "Error at Data Flow Task [Derived Column [xxx]]: Error code 0x800401F3 occurred attempting to convert from data type DT_WSTR to data type DT_GUID.".
    I read other posts like this in this forum and i also tried the solution suggested in those posts.
    The proposed solution had suggested of insert in the expression field of "Derived Column" Transformation this expression (DT_GUID) ("{" + [YourGuid] + "}") in place of (DT_GUID) [YourGuid].
    But i get always the same error.
    is there anyone who help me please to resolve this issue?!?

    The derived column expression should work if the yourGuid column contains a real guid.
    Does the source looks something like: 25892e17-80f6-415f-9c65-7395632f0223
    And are all rows filled with a correct value or are there also empty strings or null values?
    Please mark the post as answered if it answers your question | My SSIS Blog:
    http://microsoft-ssis.blogspot.com |
    Twitter

  • Data transfer from ALV-Grid to internal table

    Hello,
    I use an editable alv-grid in my program. But the data changes in the alv-grid are not transferred back to the internal table which stores the data of the alv-grid. Is there a function I have to call to transfer the data or is there a property in the alv-grid I have to set? Thanks.
    Regards, Lars.

    Hi
    The problem is if you back from grid, the system doesn't trigger the PAI of grid, so your changing aren't moved to internal table.
    You should press some button of grid bar to update your internal table. So it should be better you insert a SAVE button in your status to update the changing.
    Max
    Message was edited by: max bianchi
    Message was edited by: max bianchi

  • How do I get dates pulled from database to display in Spanish?

    I'm pulling a date from a MySQL database and need to have it
    display in Spanish. If it were in English, these are the two
    statements I would be using:
    #DateFormat(event_date,"mmmm")#
    #DateFormat(event_date,"mmm d")#
    I tried using <cfset SetLocale("Spanish (Standard)")>
    in my Application.cfm , but it didn't seem to work. I'm not really
    sure how it is supposed to be used or even if it would work if I
    were using it correctly.
    Any help would be appreciated.
    Thanks -CaddyX

    Bagger Vance wrote:
    > setLocale("es_ES"); // or whatever locale you're using
    for "Spanish"
    > // all the subsequent LS functions will use this locale
    > writeoutput("#lsDateFormat(eventDate,"FULL")#");
    >
    >
    >
    > Paul, Thanks for responding. I'm still not sure how to
    make it work. Do I just
    > put <cfset SetLocale("es_ES")> somewhere in the
    document? That didn't seem to
    > change the result in the following code:
    you can put it at the top of pages that need to be localized
    or better yet in
    your application cfm/cfc.
    >
    <cfoutput>#DateFormat(get_event.event_date,"mmmm")#</cfoutput>
    yes it won't change that, the regular cf functions all use
    en_US locale. you
    want the "LS" functions (re-read the code snippet above).

  • Data Migration from CSV file to Database Table.

    Hi,
    I have checked few answered threads on Data Migration Issues but didn't find the solution yet.
    I am reading data from an CSV file to internal table but the complete row is coming under one field of the internal table.
    How i can get values in different fields of the internal table from the CSV file ?
    Thanks & Regards.
    Raman Khurana.

    Hi,
    If you are using GUI_UPLOAD, you might have missed to make has_field_separator  as 'X'.
      EXPORTING
    CALL FUNCTION 'GUI_UPLOAD'
      EXPORTING
       filename                      = 'C:\File.csv'
       filetype                      = 'ASC'
       has_field_separator           = ' X'  "<= Set this as 'X'.
      TABLES
        data_tab                      = itab.
    Regards,
    Manoj Kumar P

  • Migrate Data only from MS SQL to existing tables in Oracle

    Hi,
    I just downloaded Sql developer and I just want to copy data from sql server tables to oracle. I already have existing tables in Oracle and I just want to import the data there. How could I possible do that?
    Thanks in advance :)

    try this it must be helpful
    how can i connect sql server in sql developer?
    Regards

  • Max data pull from Virtual Cube - is this a setting?

    We have a user doing a query against a Remote cube in our BW system, and they're hitting a "maximum data" limit of data from this remote cube.  Is this a setting for this cube or globals, and can you modify it?
    Thanks,
    Ken Little
    RJ Reynolds Tobacco

    Hi,
    MAXSIZE = Maximum size of an individual data packet in KB.
    The individual records are sent in packages of varying sizes in the data transfer to the Business In-formation Warehouse. Using these parameters you determine the maximum size of such a package and therefore how much of the main memory may be used for the creation of the data package. SAP recommends a data package size between 10 and 50 MB.
    https://www.sdn.sap.com/irj/sdn/directforumsearch?threadid=&q=cube+size&objid=c4&daterange=all&numresults=15
    MAXLINES = Upper-limit for the number of records per data packet
    The default setting is 'Max. lines' = 100000
    The maximum main memory space requirement per data packet is around
    memory requirement = 2 * 'Max. lines' * 1000 Byte,
    meaning 200 MByte with the default setting
    3 THE FORMULA FOR CALCULATING NUMBER OF RECORDS
    The formula for calculating the number of records in a Data Packet is:
    packet size = MAXSIZE * 1000 / transfer structure size (ABAP Length)
    but not more than MAXLINES.
    eg. if MAXLINES < than the result of the formula, then MAXLINES size is transferred into BW.
    The size of the Data Packet is the lowest of MAXSIZE * 1000 / transfer structure size (ABAP Length) or MAXLINES
    Goto RSCUSTV6 tcode and set it.
    Go to your Infopackage, from tool bar, scheduler, data packet settings, here you can specify your data packet size
    Go to R/3, Transaction SBIW --> General settings --> Maintain Control Parameters for Data Transfer.
    Here you can set the maximum number. But the same can be reduced in BW...
    Info package>Scheduler>data’s default data transfer-->here you can give the size, but can reduce the size given in R/3 side, you can’t increase here...
    In RSCUSTV6 you can set the package size...press F1 on it to have more info and take a look to OSS Notes 409641 'Examples of packet size dependency on ROIDOCPRMS' and 417307 'Extractor Package Size Collective Note'...
    Also Check SAP Note 919694.
    This applies irrelevant of source system meaning applicable for all the DS:
    Go To SBIW-> General Settings -> Maintain Control Parameters for Data Transfer -> Enter the entries in table
    If you want to change at DS level then:
    IS->IP -> Scheduler Menu -> Data’s. Default Data Transfer and change the values.
    Before changing the values keep in mind the SAP recommended params.
    Hope this helps u..
    Best Regards,
    VVenkat..

  • Arranging Random Data Pieces From SLX/Smart Catalogue Into Table Format

    Hi,
    We have a performance data grid that needs to appear on our quotation report.  The data is organized currently into rows in a 2-table linked format and manually entered as plain txt by the user.  We are now migrating towards designing a report in Crystal Reports to try to recreate this table layout but still needing to mine data for other purposes.  Therefore, the data pieces need to be distinct and separate.  Further, the data pieces required for the table can vary and can be organized in up to 9 different ways depending upon the given situation.  Is there an effective way to aesthetically arrange random data pieces in a table format?  The data pieces are coming from SalesLogix / SmartCatalog.  Has any other user of Crystal Reports attempted to do this - i.e. create a data table on a Crystal report?
    Sincerely,
    Sarah Imig
    International Sales Administrator/Database Administrator
    HayssenSandiacre

    Hi,
    We have a performance data grid that needs to appear on our quotation report.  The data is organized currently into rows in a 2-table linked format and manually entered as plain txt by the user.  We are now migrating towards designing a report in Crystal Reports to try to recreate this table layout but still needing to mine data for other purposes.  Therefore, the data pieces need to be distinct and separate.  Further, the data pieces required for the table can vary and can be organized in up to 9 different ways depending upon the given situation.  Is there an effective way to aesthetically arrange random data pieces in a table format?  The data pieces are coming from SalesLogix / SmartCatalog.  Has any other user of Crystal Reports attempted to do this - i.e. create a data table on a Crystal report?
    Sincerely,
    Sarah Imig
    International Sales Administrator/Database Administrator
    HayssenSandiacre

  • Simple Max date pull from several rows with the same employee numbers.

    Ok so I'm hoping this is flying over my head because it's almost 3am, but I haven't been able to figure this query out for the last 2hrs, but I like to figure stuff out on my own. I'm wondering why the following statement works fine:
    select emp_num, to_char(max(end_date)) as newestdate
    from bank_history
    where emp_num in ('22964', '21667', '20758', '12739', '12731', '20929', '22795', '20594', '23077', '12588', '21294', '20618', '21204', '22952', '19990', '20632', '03093', '19991', '22951', '07779', '20014', '11981', '06149', '20364', '21103')
    and bank_type = 'P' group by emp_num
    BUT! When I start to add more columns that I need to see it starts adding the multiple employee numbers again.
    select emp_num, bank_type, earn_hrs, used_paid_hrs, used_paid_hrs, to_char(max(end_date)) as newestdate
    from bank_history
    where emp_num in ('22964', '21667', '20758', '12739', '12731', '20929', '22795', '20594', '23077', '12588', '21294', '20618', '21204', '22952', '19990', '20632', '03093', '19991', '22951', '07779', '20014', '11981', '06149', '20364', '21103')
    and bank_type = 'P' group by emp_num, bank_type, earn_hrs, used_paid_hrs, used_unpaid_hrs
    The orginal table looks something like this:
    EMP_NUM BANK_TYPE EARN_HRS USED_PAID_HRS END_DATE
    60393 P 0.58 0 3-Aug-2004
    60394 P 7.48 1 28-Oct-2003
    60394 P 40 40 28-Oct-2004
    60394 P 40 12.4 28-Oct-2005
    60395 P 40 40 21-Oct-2004
    60395 P 0 0 21-Oct-2003
    60395 P 40 40 21-Oct-2005
    60397 P 40 39.85 21-Oct-2004
    60397 P 0.97 0.97 21-Oct-2003
    Much thanks for any direction you can give to help guide me on my way to a solution. If you can help it please don't give me the answer, but try to give me the reason it's not working. I'll never learn if you give it to me right up front lol. Thanks again.
    Luke
    Message was edited by: trying to fix the table formating sorry it's so unreadable
    Luke22

    I just give you a prompt:
    SQL> with t as (select 1 emp_num, 'A' bank, date '2007-08-01' dt from dual union all
      2             select 1 emp_num, 'B' bank, date '2007-08-02' dt from dual union all
      3             select 1 emp_num, 'C' bank, date '2007-08-03' dt from dual union all
      4             select 2 emp_num, 'A' bank, date '2007-08-05' dt from dual union all
      5             select 2 emp_num, 'B' bank, date '2007-08-04' dt from dual union all
      6             select 3 emp_num, 'A' bank, date '2007-08-07' dt from dual union all
      7             select 3 emp_num, 'C' bank, date '2007-08-08' dt from dual)
      8  --
      9             select emp_num, max(dt) from t
    10             group by emp_num
    11  /
       EMP_NUM MAX(DT)
             1 03.08.2007
             2 05.08.2007
             3 08.08.2007
    SQL>
    SQL> with t as (select 1 emp_num, 'A' bank, date '2007-08-01' dt from dual union all
      2             select 1 emp_num, 'B' bank, date '2007-08-02' dt from dual union all
      3             select 1 emp_num, 'C' bank, date '2007-08-03' dt from dual union all
      4             select 2 emp_num, 'A' bank, date '2007-08-05' dt from dual union all
      5             select 2 emp_num, 'B' bank, date '2007-08-04' dt from dual union all
      6             select 3 emp_num, 'A' bank, date '2007-08-07' dt from dual union all
      7             select 3 emp_num, 'C' bank, date '2007-08-08' dt from dual)
      8  --
      9             select emp_num, bank, max(dt) from t
    10             group by emp_num, bank
    11  /
       EMP_NUM BANK MAX(DT)
             1 A    01.08.2007
             1 B    02.08.2007
             1 C    03.08.2007
             3 A    07.08.2007
             2 A    05.08.2007
             2 B    04.08.2007
             3 C    08.08.2007
    7 rows selected
    SQL>
    SQL> with t as (select 1 emp_num, 'A' bank, date '2007-08-01' dt from dual union all
      2             select 1 emp_num, 'B' bank, date '2007-08-02' dt from dual union all
      3             select 1 emp_num, 'C' bank, date '2007-08-03' dt from dual union all
      4             select 2 emp_num, 'A' bank, date '2007-08-05' dt from dual union all
      5             select 2 emp_num, 'B' bank, date '2007-08-04' dt from dual union all
      6             select 3 emp_num, 'A' bank, date '2007-08-07' dt from dual union all
      7             select 3 emp_num, 'C' bank, date '2007-08-08' dt from dual)
      8  --
      9             select emp_num, max(bank) keep (dense_rank last order by dt) bank, max(dt) from t
    10             group by emp_num
    11  /
       EMP_NUM BANK MAX(DT)
             1 C    03.08.2007
             2 A    05.08.2007
             3 C    08.08.2007

Maybe you are looking for