Help Required -- Can we use SQL Query to READ data from SAP MDM Tables

Hi All,
Please help.........
Can we use SQL Query to READ(No Creation/Updation/Deletion  just Read) Data from SAP MDM tables directly, without using MDM Syndicator.
Or direct SQL access to SAP MDM tables is not possible. Only through MDM Syndicator can we export data.
Thanks in Advance
Regards

All the tables you create in Repository comes under A2i_CM_Tables in Database named as your repository name. So the tables names are fields of table A2i_CM_Tables. Now i tried it but cant make it.
Now, I dont think its possible to extract all fields in tables and there values using select query. May be pure sql guy can do that or not.
But there is no relation of data extraction and syndicator. Data is viewed in Data Manager. and you can also store data in a file from DM also.
BR,
Alok

Similar Messages

  • Select query to read data from a view

    Hi friends,
    We have been using a query to read data from a custom view which used to work perfectly. Now the program sits at that select query forever. We are able to extract same data from se16. Not sure what could be the problem.
    Thanks in advance.

    Dev
    Have a look at the Table Index for the tables involved in the View... I think there is some change in the Indexes.. (Add / Remove / Change)
    Thanks
    Amol Lohade

  • How can I use Automator to extract specific Data from a text file?

    I have several hundred text files that contain a bunch of information. I only need six values from each file and ideally I need them as columns in an excel file.
    How can I use Automator to extract specific Data from the text files and either create a new text file or excel file with the info? I have looked all over but can't find a solution. If anyone could please help I would be eternally grateful!!! If there is another, better solution than automator, please let me know!
    Example of File Contents:
    Link Time =
    DD/MMM/YYYY
    Random
    Text
    161 179
    bytes of CODE    memory (+                68 range fill )
    16 789
    bytes of DATA    memory (+    59 absolute )
    1 875
    bytes of XDATA   memory (+ 1 855 absolute )
    90 783
    bytes of FARCODE memory
    What I would like to have as a final file:
    EXCEL COLUMN1
    Column 2
    Column3
    Column4
    Column5
    Column6
    MM/DD/YYYY
    filename1
    161179
    16789
    1875
    90783
    MM/DD/YYYY
    filename2
    xxxxxx
    xxxxx
    xxxx
    xxxxx
    MM/DD/YYYY
    filename3
    xxxxxx
    xxxxx
    xxxx
    xxxxx
    Is this possible? I can't imagine having to go through each and every file one by one. Please help!!!

    Hello
    You may try the following AppleScript script. It will ask you to choose a root folder where to start searching for *.map files and then create a CSV file named "out.csv" on desktop which you may import to Excel.
    set f to (choose folder with prompt "Choose the root folder to start searching")'s POSIX path
    if f ends with "/" then set f to f's text 1 thru -2
    do shell script "/usr/bin/perl -CSDA -w <<'EOF' - " & f's quoted form & " > ~/Desktop/out.csv
    use strict;
    use open IN => ':crlf';
    chdir $ARGV[0] or die qq($!);
    local $/ = qq(\\0);
    my @ff = map {chomp; $_} qx(find . -type f -iname '*.map' -print0);
    local $/ = qq(\\n);
    #     CSV spec
    #     - record separator is CRLF
    #     - field separator is comma
    #     - every field is quoted
    #     - text encoding is UTF-8
    local $\\ = qq(\\015\\012);    # CRLF
    local $, = qq(,);            # COMMA
    # print column header row
    my @dd = ('column 1', 'column 2', 'column 3', 'column 4', 'column 5', 'column 6');
    print map { s/\"/\"\"/og; qq(\").$_.qq(\"); } @dd;
    # print data row per each file
    while (@ff) {
        my $f = shift @ff;    # file path
        if ( ! open(IN, '<', $f) ) {
            warn qq(Failed to open $f: $!);
            next;
        $f =~ s%^.*/%%og;    # file name
        @dd = ('', $f, '', '', '', '');
        while (<IN>) {
            chomp;
            $dd[0] = \"$2/$1/$3\" if m%Link Time\\s+=\\s+([0-9]{2})/([0-9]{2})/([0-9]{4})%o;
            ($dd[2] = $1) =~ s/ //g if m/([0-9 ]+)\\s+bytes of CODE\\s/o;
            ($dd[3] = $1) =~ s/ //g if m/([0-9 ]+)\\s+bytes of DATA\\s/o;
            ($dd[4] = $1) =~ s/ //g if m/([0-9 ]+)\\s+bytes of XDATA\\s/o;
            ($dd[5] = $1) =~ s/ //g if m/([0-9 ]+)\\s+bytes of FARCODE\\s/o;
            last unless grep { /^$/ } @dd;
        close IN;
        print map { s/\"/\"\"/og; qq(\").$_.qq(\"); } @dd;
    EOF
    Hope this may help,
    H

  • How will write SQL query to fetch data from  each Sub-partition..

    Hi All,
    Anyone does have any idea about How to write SQL query to fetch data from Sub-partition.
    Actually i have one table having composite paritition(Range+list)
    Now if i want to fetch data from main partition(Range) the query will be
    SELECT * FROM emp PARTITION(q1_2005);
    Now i want to fetch data at sub-partition level(List) .But i am not able to get any SQL query for that.
    Pls help me to sort out.
    Thanks in Advance.
    Anwar

    SELECT * FROM emp SUBPARTITION(sp1);

  • Can i use one interface to load data into 2 different tables

    Hi Folks,
    Can i use one interface to load data into 2 different tables(same schema or different schemas) from one source table with same structure ?
    Please give me advice
    Thanks
    Raj
    Edited by: user11410176 on Oct 21, 2009 9:55 AM

    Hi Lucky,
    Thanks for your reply,
    What iam trying is ...Iam trying to load the data from legacy tables(3) into oracle staging tables.But i need to load the same source data into two staging tables(these staging tables are in two different schemas)
    can i load this source data into two staging tables by using single standard interface(some business logic is there)
    If i can then give me some suggestion how to do that
    Thanks in advance
    Raj

  • Can we use impdp to import the data from an normal exp dump?

    Hi All,
    I have a export dump taken from a 9i database. Can i use impdp to import the data into 10g database from that 9i exp dump?
    Please suggest
    thanks and Regards
    Arun

    Hi,
    I have a export dump taken from a 9i database. Can i use impdp to import the data into 10g database from that 9i exp dump?Yes, it can be.
    Refer:
    http://wiki.oracle.com/thread/3734722/can+a+9i+dump+file+be+imported+into+10%3F
    thanks,
    X A H E E R

  • Regarding Native SQL to access Oracle Data from SAP

    Dear Gurus,
    This query is regarding Native SQL.
    Our database is Oracle and the client's database is also Oracle.
    To access the data directly from the client's database(Oracle), can I use Native SQL commands?
    Is there any disadvantage to use Native SQL?
    If Native SQL can be used, can any one send me the code to access the data from Oracle?
    Points will be rewarded.
    Thanks & Regards,
    Neeraj K.

    The problem is always data integrity. Doing things this way you are bypassing any business logic. So for example if you are connecting to a SAP Oracle database. SAP will not support your SAP system as you are doing direct updates to the database. I would suppose that this would be the same for any other product using the Oracle database. If you are only reading data from the database then it should be fine. I would however always using API's to access/update data.
    Regards

  • What all sofware is required to develop a userinterface.vi to read data from a data logger and save in a file

    hi,
    i am working on creating a GUI.vi to read data from the datalogger and the same should be used to set the date ,time,etc.i am thinking of using VISA to read data through serial port(COM1),that interface is available.but when i tried to read data from the logger i am getting timeout error.what should i do ?i am using the example given with labview 7.0 .can any one suggest me what all things i need to do if i want to read datalogger to a file.Is there any one who workrd on this.please reply me as early as possible,
    bye
    vbkraju

    You're welcome--I'm happy to help.
    The LabVIEW <-> Serial.vi is a good example that allows for writing commands to an instrument and reading back from it using a VISA Session. One of the inputs on this VI is the Resource Name. You need to ensure that the name of your serial port is in this drop down menu, so that the program is sending and receiving data from the correct place.
    The VI is setup so that you can send a command to the external device, and then receive data back. I am not sure about the Spectrum 4000, but most devices have commands that they receive to tell the device which data to send back over the serial port. The VI has a default string to write of *IDN?, which is a common command that tells the instrument to return an ident
    ity string. The fact that the program is timing out for you indicates that the program never received any data from the device. Either the device is not configured to automatically send data, or the correct command is not being sent to the device. Your datalogger manual may be able to help you determine what serial commands it needs to receive to send the data out you expect.
    I hope this will get you further along in your application.
    John M

  • How to use Power Query to load data from a single source into multiple tables

    Hi all,
    I have a requirement to load my data into three different data models using Power Query. Is that possible?
    My source is a SharePoint survey list, with similar questions like:
    1) Course lecturer - John Doe
    1a) The course was useful (rate 1 to 5)
    1b) The lecturer displayed good knowledge of topic (rate 1 to 5)
    2) Course Lecturer - Mary Brown
    2a) The course was useful (rate 1 to 5)
    2b) The lecturer displayed good knowledge of topic (rate 1 to 5)
    I would like to split the data into separate data models (one for John Doe; another for Mary Brown), so that I can compare the different lecturers. Other than running separate surveys for each of them, I thought of using Power Query to transform the data.
    Is it possible?
    Thanks.
    Regards
    GM

    Yes, this is possible.
    Start with a single query that returns you the data for all lecturers.
    Right-click on the "all lecturers" query in the queries pane, and choose Reference, once for each lecturer. This will create a query for each lecturer.
    Open the query for each lecturer and filter the data so that only that lecturer's results are visible.
    Click "Close & Load To..." for each lecturer's query  to load the data into the data model. This will create a data model table for each lecturer.
    If your question is more about how to transform such a survey list into a table that can be easily filtered, please provide an example of how the list shows up in Power Query.
    Ehren

  • What Function Module can be used to create vendor master data in SAP R/3?

    Hello -
    I've been trying to find a standard SAP delivered BAPI that can be used to create vendor master records in ECC 6.0. Up to this point I have not been able to do so. With SAP MDM solutions and SRM solutions it seems logical that SAP would deliver a standard BAPI that can be used to create vendor master records in the ERP/R3 environment (similar to the BAPI_MATERIAL_SAVEREPLICA for creating material master data records).
    I have been able to find BAPI_VENDOR_CREATE but this only calls the online transaction and does not offer much additional functionality beyond what transaction code XK02 already offers.
    The next best alternative appears to be IDOC_INPUT_CREDITOR but our preference would be to not use IDOCs for the project we are working on.
    Can anyone offer any insight or personal experience on how vendor master records have been successfully interfaced with SAP R/3? Did you have to use IDOCs or create a custom RFC call? Thanks for the help!
    Rich Wortmann
    << Personal information removed >>
    Edited by: Rob Burbank on May 14, 2009 10:43 AM

    Quite a gap in the BAPI arsenal. I think the options are IDoc, which you don't want, good old program RFBIKR00 (requires flat file for input, can be used via LSMW) or DIY (very complex).
    Thomas

  • Can i use OER to store the data from a excel sheet that pertains to the entire integrations ?

    hi ,
    i know that the OER can be used to view the entire landscape for the SOA composites and how they are related with each other. However we have a requirement wherein we need to store the data for the entire integration end systems e.g. ebiz, mfgpro, siebel etc ( that is we need to store data one level up).
    Also,  we want a customized form in OER wherein user can manually add data regarding the integrations.
    So is it possible to do so ? If yes, how ?

    ok, let me put it this way,
    I tried using the import/export utility available under OER admin tab.
    When i export my existing assets it generates the .xml file for each asset. Now as i have hundreds on new assets to be created for each asset type i want to be able to create a new .xml file and want it to create the new asset once i import it using import utility. I tried creating a new .xml file giving it a new UUID but it does not seem to work. Has anyone tried this earlier who can give the steps to do this ?

  • How to use SQL*LOADER to read data in |SMITH|ALFRED| format

    the data I need to upload to table using SQL*LOADER is in format as below:
    |AD |Argentina  |
    |CN |China       |
    |US |America            |
    |GB |England        |so how should I write my control file to read the data into two columns table?
    I googled and people say that use Enclosed fields—delimiter (|), and I tried with below, but not working:
    LOAD DATA
    INFILE *
    insert
    INTO TABLE tmp_country_mapping
    FIELDS ENCLOSED BY '|'
    (country_id, country_name)
    BEGINDATA
    |AD |Argentina  |
    |CN |China       |
    |US |America            |
    |GB |England        |Anyone could help?
    BTW, I don`t want the spaces behind the value of the second column.
    Thanks
    Edited by: PhoenixBai on Dec 28, 2010 2:50 PM

    Problem solved, by removing the NLS_LANG=ENGLISH.
    >
    before the load, data is as below:
    |AD |安道尔共和国         |
    |AE |阿联酋            |
    |AF |阿富汗            |
    |AG |安提瓜和巴布达        |
    |AI |安圭拉岛           |
    |AL |阿尔巴尼亚          |after the load, data displays as below:
    AD      °2μà??12oí1ú
    AE      °¢áa??
    AF      °¢??o1
    AG      °2ìá1?oí°í2?′?
    AI      °21?à-μo
    AL      °¢??°í?á??And my database supports Chinese words and it displays correctly for other tables, except this one.
    What could be the problem?
    >
    Edited by: PhoenixBai on Dec 28, 2010 5:05 PM

  • Can I use flash drives to transfer data from one MBP to another at a differ

    I've looked and cannot find any info on the use of flash drives. Can I use a flash drive to transfer data from one MBP to another one at a different location? (ie; school Apple and home MBP) Grade tests and papers at home, enter data into home computer, then put it on flash drive and carry flash drive to school, transferring data to school desk top? Are certain flash drives preferrable over others, name brands, size, etc?

    Yep, you sure can. Just about any drive will work as mentioned above. But I'm partial to SanDisk and PNY. I've used both and they work great! Currently I carry both a 2GB SanDisk Micro Cruzer and a retractable 2GB SanDisk Cruzer, both really nice. Replaced my oldie but goodie 256MB PNY drive that I've been using since they first came out.

  • Sql query to bind data from grid and print total count and amount total when date changes

    SELECT SLHD.VOUCH_CODE,SLHD.VOUCH_DATE,SLHD.VOUCH_NUM,SUM(SLTXN.CALC_NET_AMT) AS AMT,ACT.ACT_NAME,SUM(SLTXN.TOT_QTY) AS QTY
    FROM SL_HEAD20132014 AS SLHD,ACCOUNTS AS ACT,SL_TXN20132014 AS SLTXN
    WHERE SLHD.ACT_CODE=ACT.ACT_CODE AND SLTXN.VOUCH_CODE=SLHD.VOUCH_CODE
    GROUP BY SLHD.VOUCH_CODE,SLHD.VOUCH_DATE,SLHD.VOUCH_NUM,ACT.ACT_NAME
    ORDER BY SLHD.VOUCH_DATE 
    i want to print total quatity and total sale in grid when data changes
    like
    date amount quantity
    01/02/2013 1200 1
    01/02/2013  200 1
    01/02/2013  1400 2 // date changes here 
    02/03/2013 100 1 
    02/03/2013 50 4
    02/03/2013 150 5 // date changes and so on

    this query only print all the data from table i want total quantity and total amount of daily sale in same grid when ever date changes
    You may add the date filter to Visakh's query:
    SELECT SLHD.VOUCH_DATE,SUM(SLTXN.CALC_NET_AMT) AS AMT,SUM(SLTXN.TOT_QTY) AS QTY
    FROM SL_HEAD20132014 AS SLHD,ACCOUNTS AS ACT,SL_TXN20132014 AS SLTXN
    WHERE SLHD.ACT_CODE=ACT.ACT_CODE AND SLTXN.VOUCH_CODE=SLHD.VOUCH_CODEand SLHD.VOUCH_DATE = @yourdate --passed from the front end application
    GROUP BY SLHD.VOUCH_DATE
    WITH CUBE
    ORDER BY SLHD.VOUCH_DATE
    Having said, each time when you select the date, you query the table would be expensive method. May be you can filter the date within your dataset already populated if you have entire data in the dataset.

  • HT204350 Can I use Migration Assistant to transfer data from macbook running 10.7.5 To a macbook air running Mavericks? On first attempt destination machine says that os versions are not compatible for Migration Assistant.

    I am trying to transfer contents from old macbook running os 10.7.5 to a macbook air running os 10.9.1 using migration assistant but the destination macobook air says that migration cannot proceed (yellow triangle) because the os on the macbook is not up to date (assuming that eans 10.9.1).  However old macbook cannot be updated to 10.9.1.
    Can anyone suggest a way around this please?

    Try looking at this section in the Migration Assistant article.
    Migrating using FireWire or ThunderBolt 

Maybe you are looking for