FDM Data load to External Database

I am looking to export data and mapping data from FDM to external database, kind of looking to create a custom adapter. Has anyone tried this ?
Appreciate any pointers.

Yes because you are able to write your own extract and external data load code. The pull adaptor does have an existing data extract and output file but this can be adapted to your own needs including writing the necessary SQL to connect and load to an external database.

Similar Messages

  • Best Practices on Routine Data Load.

    Can someone please tell me what are the best practices on routine data load from one database to another?
    We have PeopleSoft system where new employees' records are created; however, these new employees are required to take new employee tests that is being tracked by an application outside Peoplesoft on an Oracle db. Therefore, we need to populate the Oracle db with the new employee's information - on a daily basis or as needed. The data we will need to track are new employees or rehires, changes on existing employees - position, title, etc, terminated employees - date of termination, etc.
    What is the best practice to get the employee's information to the Oracle db?
    Any suggestions are appreciated.
    -andy

    Depends on your source and your database version which you didn't mention. What is the easiest way to get them out of your source database?
    Perhaps a database link though that might be a security violation.
    Perhaps as a delimited ASCII file loaded using SQL*Loader or an external table.
    Can you provide more information and database version numbers?

  • How to extract R/3 (4.7) data to an external data warehouse server

    Hi,
    What are the methods or steps of extracting R/3(4.7 - no BW module) data to an external database (or flat file) for generating reports in an external BI system?
    Can I use ABAP to do ETL and how?
    Thank you,
    Bruce

    Hi Harsha,
    You cannot extract the multiple values using classification data sources, this data can be extracted by creating a generic extractor based on a view (with join on tables AUSP and INOB). this extractor brings the data in the following format.
    OBJEK | ATINN | ATZHL | ATWRT
    1 | HOBBY | 1 | CRICKET
    1 | HOBBY | 2 | PAINTING
    1 | HOBBY | 3 | READING
    2 | HOBBY | 1 | CRICKET
    2 | HOBBY | 2 | PAINTING
    3 | HOBBY | 1 | READING
    i.e. for field Hobby these are values assigned, if you want you can create 3 InfoObjects in BW and flatten this data by setting flags as below.
    OBJEK | HOBBY_CRICKET | HOBBY_PAINTING | HOBBY_READING
    1 | X | X | X |
    2 | X | X |    |
    3 |    |    | X |
    Let me know if you need more details.
    Regards,
    Rk

  • How to connect external database with SAP

    Hi,
    How to get the data from an external database into SAP by directly linking the SAP Program to the external database.
    Regards
    Guhapriyan

    Hi,
    http://help.sap.com/saphelp_nw04/helpdata/en/58/54f9c1562d104c9465dabd816f3f24/content.htm
    This might help,
    Regards,
    Gaurav

  • Need to generate multiple error files with rule file names during parallel data load

    Hi,
    Is there a way that MAXL could generate multiple error files during parallel data load?
    import database AsoSamp.Sample data
      connect as TBC identified by 'password'
      using multiple rules_file 'rule1' , 'rule2'
      to load_buffer_block starting with buffer_id 100
      on error write to "error.txt";
    I want to get error files as this -  rule1.err, rule2.err (Error files with rule file name included). Is this possible in MAXL? 
    I even faced a situation , If i hard code the error file name like above, its giving me error file names as error1.err and error2.err. Is there any solution for this?
    Thanks,
    DS

    Are you saying that if you specify the error file as "error.txt" Essbase actually produces multiple error files and appends a number?
    Tim. 
    Yes its appending the way i said.
    Out of interest, though - why do you want to do this?  The load rules must be set up to select different 'chunks' of input data; is it impossible to tell which rule an error record came from if they are all in the same file?
    I have like 6 - 7 rule files using which the data will be pulled from SQL and loaded into Essbase. I dont say its impossible to track the error record.
    Regardless, the only way I can think of to have total control of the error file name is to use the 'manual' parallel load approach.  Set up a script to call multiple instances of MaxL, each performing a single load to a different buffer.  Then commit them all together.  This gives you most of the parallel load benefit, albeit with more complex scripting.
    Even i had the same thought of calling multiple instances of a Maxl using a shell script.  Could you please elaborate on this process? What sort of complexity is involved in this approach.? Did anyone tried it before?
    Thanks,
    DS

  • Error while loading the data from excel to database.

    Hi,
    I am using PL/SQL developer to load the data from excel to database. I will set the data source in the control panel and will proceed through ODBC importer in pl/sql developer to import the data.
    What exactly the error is when i click the filename to view the result preview it shows an error as:
    The field is too small to accept the amount of data you attempted to add. Try inserting or pasting less data.
    Kindly help with solution.
    Thanks/Regards
    Sakthivarman J.

    Hello;
    That error message comes from Microsoft, so something in you Excel sheet is the cause.
    Its a pain but I would check properties of each column in case Excel decided to add something, a comma for example.
    Do you have a column over 255 characters? Look there first. If any length is greater than 255 it will crash and burn.
    Or convert it to a CSV and create an external table.
    Best Regards
    mseberg
    Might also throw an 3163 as an error where you cannot see it.
    Edited by: mseberg on Sep 9, 2011 7:34 AM

  • Comparison of Data Loading techniques - Sql Loader & External Tables

    Below are 2 techniques using which the data can be loaded from Flat files to oracle tables.
    1)     SQL Loader:
    a.     Place the flat file( .txt or .csv) on the desired Location.
    b.     Create a control file
    Load Data
    Infile "Mytextfile.txt" (-- file containing table data , specify paths correctly, it could be .csv as well)
    Append or Truncate (-- based on requirement) into oracle tablename
    Separated by "," (or the delimiter we use in input file) optionally enclosed by
    (Field1, field2, field3 etc)
    c.     Now run sqlldr utility of oracle on sql command prompt as
    sqlldr username/password .CTL filename
    d.     The data can be verified by selecting the data from the table.
    Select * from oracle_table;
    2)     External Table:
    a.     Place the flat file (.txt or .csv) on the desired location.
    abc.csv
    1,one,first
    2,two,second
    3,three,third
    4,four,fourth
    b.     Create a directory
    create or replace directory ext_dir as '/home/rene/ext_dir'; -- path where the source file is kept
    c.     After granting appropriate permissions to the user, we can create external table like below.
    create table ext_table_csv (
    i Number,
    n Varchar2(20),
    m Varchar2(20)
    organization external (
    type oracle_loader
    default directory ext_dir
    access parameters (
    records delimited by newline
    fields terminated by ','
    missing field values are null
    location ('file.csv')
    reject limit unlimited;
    d.     Verify data by selecting it from the external table now
    select * from ext_table_csv;
    External tables feature is a complement to existing SQL*Loader functionality.
    It allows you to –
    •     Access data in external sources as if it were in a table in the database.
    •     Merge a flat file with an existing table in one statement.
    •     Sort a flat file on the way into a table you want compressed nicely
    •     Do a parallel direct path load -- without splitting up the input file, writing
    Shortcomings:
    •     External tables are read-only.
    •     No data manipulation language (DML) operations or index creation is allowed on an external table.
    Using Sql Loader You can –
    •     Load the data from a stored procedure or trigger (insert is not sqlldr)
    •     Do multi-table inserts
    •     Flow the data through a pipelined plsql function for cleansing/transformation
    Comparison for data loading
    To make the loading operation faster, the degree of parallelism can be set to any number, e.g 4
    So, when you created the external table, the database will divide the file to be read by four processes running in parallel. This parallelism happens automatically, with no additional effort on your part, and is really quite convenient. To parallelize this load using SQL*Loader, you would have had to manually divide your input file into multiple smaller files.
    Conclusion:
    SQL*Loader may be the better choice in data loading situations that require additional indexing of the staging table. However, we can always copy the data from external tables to Oracle Tables using DB links.

    Please let me know your views on this.

  • FDM event scripts firing twice during data loads

    Here's an interesting one. I have added the following to three different event scripts (one at a time, ensuring only one of these exists at any one time), to clear data before loading to Essbase:
    Event Script content:
    ' Declare local variables
    Dim objShell
    Dim strCMD
    ' Call MaxL script to run data clear calculation.
    Set objShell = CreateObject("WScript.Shell")
    strCMD = "D:\Oracle\Middleware\EPMSystem11R1\products\Essbase\EssbaseClient\bin\startMAXL.cmd D:\Test.mxl"
    API.DataWindow.Utilities.mShellAndWait strCMD, 0
    MaxL Script:
    login ******* identified by ******* on *******;
    execute calculation 'FIX("Member1","Member2") CLEARDATA "Member3"; ENDFIX' on *******.*******;
    exit;
    However, it appears that the clear is carried out twice, both before and after the data has been loaded to Essbase. This has been verified at each step by checking the Essbase application log:
    No event script:
    - No Essbase data clear in application log
    Adding above to "BefExportToDat" event script:
    - Script is executed once after clicking on Export in FDM Web Client (before the "Target System Load" modal popup is displayed). Entries are visible in Essbase Application log.
    - Script is then executed a second time when clicking on the OK button in the "Target System Load" modal popup. Entries are visible in Essbase Application log.
    Adding above to "AftExportToDat" event script:
    - Script is executed once after clicking on Export in FDM Web Client (before the "Target System Load" modal popup is displayed). Entries are visible in Essbase Application log.
    - Script is then executed a second time when clicking on the OK button in the "Target System Load" modal popup. Entries are visible in Essbase Application log.
    Adding above to "BefLoad" event script:
    - Script is NOT executed after clicking on Export in FDM Web Client (before the "Target System Load" modal popup is displayed).
    - Script is executed AFTER the data load to Essbase when clicking on the OK button in the "Target System Load" modal popup. Entries are visible in Essbase Application log.
    Some notes on the above:
    1. "BefExportToDat" and "AftExportToDat" are both executed twice, before and after the "Target System Load" modal popup. :-(
    2. "BefLoad" is executed AFTER the data is loaded to Essbase! :-( :-(
    Does anyone please have any idea how we might execute an Essbase database clear before loading data, and not after we have loaded fresh data? And perhaps on why the above event scripts appear to be firing twice?! There does not appear to be any logic to this!
    BefExportToDat - Essbase Application Log entries:
    +[Wed May 16 16:19:51 2012]Local/Monthly/Monthly/admin@Native Directory/140095859451648/Info(1013091)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:19:51 2012]Local/Monthly/Monthly/admin@Native Directory/140095859451648/Info(1013162)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:19:51 2012]Local/Monthly/Monthly/admin@Native Directory/140095859451648/Info(1012555)+
    +Clearing data from [Member3] partition with fixed members [Period(Member1); Scenario(Member2)]+
    +...+
    +[Wed May 16 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info(1003037)+
    Data Load Updated [98] cells
    +[Wed May 16 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info(1003024)+
    Data Load Elapsed Time : [0.52] seconds
    +...+
    +[Wed May 16 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info(1013091)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info(1013162)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:20:12 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info(1012555)+
    +Clearing data from [Member3] partition with fixed members [Period(Member1); Scenario(Member2)]+
    AftExportToDat - Essbase Application Log entries:
    +[Wed May 16 16:21:32 2012]Local/Monthly/Monthly/admin@Native Directory/140095933069056/Info(1013091)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:21:32 2012]Local/Monthly/Monthly/admin@Native Directory/140095933069056/Info(1013162)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:21:32 2012]Local/Monthly/Monthly/admin@Native Directory/140095933069056/Info(1012555)+
    +Clearing data from [Member3] partition with fixed members [Period(Member1); Scenario(Member2)]+
    +...+
    +[Wed May 16 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info(1003037)+
    Data Load Updated [98] cells
    +[Wed May 16 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095930963712/Info(1003024)+
    Data Load Elapsed Time : [0.52] seconds
    +...+
    +[Wed May 16 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095928858368/Info(1013091)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095928858368/Info(1013162)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:21:47 2012]Local/Monthly/Monthly/admin@Native Directory/140095928858368/Info(1012555)+
    +Clearing data from [Member3] partition with fixed members [Period(Member1); Scenario(Member2)]+
    BefLoad - Essbase Application Log entries:
    +[Wed May 16 16:23:43 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info(1013091)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:23:43 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info(1013162)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:23:43 2012]Local/Monthly/Monthly/admin@Native Directory/140095932016384/Info(1012555)+
    +Clearing data from [Member3] partition with fixed members [Period(Member1); Scenario(Member2)]+
    +...+
    +[Wed May 16 16:23:44 2012]Local/Monthly/Monthly/admin@Native Directory/140095929911040/Info(1003037)+
    Data Load Updated [98] cells
    +[Wed May 16 16:23:44 2012]Local/Monthly/Monthly/admin@Native Directory/140095929911040/Info(1003024)+
    Data Load Elapsed Time : [0.52] seconds
    +...+
    +[Wed May 16 16:23:45 2012]Local/Monthly/Monthly/admin@Native Directory/140095860504320/Info(1013091)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:23:45 2012]Local/Monthly/Monthly/admin@Native Directory/140095860504320/Info(1013162)+
    +Received Command [Calculate] from user [admin@Native Directory]+
    +[Wed May 16 16:23:45 2012]Local/Monthly/Monthly/admin@Native Directory/140095860504320/Info(1012555)+
    +Clearing data from [Member3] partition with fixed members [Period(Member1); Scenario(Member2)]+

    Hi Larry,
    As mentioned, our exports do not appear to be generating the "-B.Dat" and "-C.Dat" files at present. However, you are correct with the Export and Load event scripts firing twice (once for the main TB file and again for the journal file). Does this also mean it could continue to fire an additional two times for the "-B.Dat" and "-C.Dat" files?
    On the last run, the output was as follows with the modified scripts:
    After clicking on Export in Workflow, the Target System Load modal popup is displayed, and the first two files have been generated:
    14.24.15.0527_BefExportToDat.txt
    14.24.17.0617_AftExportToDat.txt
    After clicking on OK in the Target System Load modal popup, the actual load to Essbase takes place. A further six files are generated:
    14.24.21.0289_BefLoad.txt
    14.24.22.0117_AftLoad.txt
    *14.24.22.0152_BefExportToDat-A.txt*
    *14.24.22.0414_AftExportToDat-A.txt*
    *14.24.22.0433_BefLoad-A.txt*
    *14.24.22.0652_AftLoad-A.txt*
    This makes a lot more sense, since one can see that the event scripts are being run a second time against the journal files during the data load. Many thanks, this solves my problem as I can now place my script where I want in the process chain. It's just a shame that there are not separate event scripts to distinguish between the various .Dat exports/loads, which are clearly occuring at separate times in the process chain.
    Many thanks! :-)
    P.S. Updated script below if anyone wishes to use it:
    Sub BefExportToDat(strLoc, strCat, strPer, strTCat, strTPer, strFile)
    Dim strF, fso, tf, t, temp, m, miliseconds, strSuf
    t = Timer
    temp = Int(t)
    m = Int((t-temp)*1000)
    miliseconds = String(4 - Len(m), "0") & m
    strF = "D:\TEST\" & Replace(Time, ":", ".") & "." & miliseconds & "_BefExportToDat"
    strSuf = UCase(Left(Right(strFile,6),2))
    If strSuf = "-A" Or strSuf = "-B" Or strSuf = "-C" Then
    strF = strF & UCase(strSuf) & ".txt"
    Else
    strF = strF & ".txt"
    End If
    Set fso = CreateObject("Scripting.FileSystemObject")
    Set tf = fso.CreateTextFile(strF, True)
    tf.WriteLine(strFile)
    tf.Close
    Set fso = Nothing
    End Sub

  • Unable to load data into any application.database on server.

    I have a rather odd problem that's been vexing my for a few days.
    I am unable to do a data import into any cubes within Essbase, its as if the cube is in read only mode, though everything seems OK. I'm not running any sort of archiving, and to manually check, within essmsh, I did an "alter database end archive". I've tried with our existing and unchanged data load script as well as from within the EAS ( Right-click Data Load ). If I check the processes running on the server ( LINUX server ) the ESSSVR process is the top process eating 100% CPU. I can do outline builds OK.
    From within the application log file, the last few entries are:
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021044)
    Starting to execute query
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021045)
    Finished executing query, and started to fetch records
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021000)
    Connection With SQL Database Server is Established
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1003040)
    Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
    [Wed Apr  3 09:45:19 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021047)
    Finished fetching data
    Nothing has changed on this server for a few weeks so I'm somewhat flummoxed and no other errors show in any other logs ( nohup.out ).
    Essbase 11.1.1.2
    Redhat LINUX_x64 2.6

    Yes, I tried rebooting the whole box, server has 9Gb free space, same story when I load just 1 row via a text file in the EAS, it seems to hang and the ESSSVR process for that cube goes to 100%.
    I've tried a few different cubes, and they all have the same problem, so I suspect its something to do with Essbase itself rather than the specific cube I'm having problems with.
    This is our test server I'm experiencing the problem with and I tried migrating 1 app from UAT back to test which still didnt work.
    Stumped!

  • RFC XI JDBC to external database.... Get data in multiple rows

    Hi,
    We have been really struggling with this scenario. We have created a RFC that has 1 Import parameter (Order_ID) and table parameter (3 Columns: Order_ID, Partner_type, Partner_no).
    When we call this RFC in SAP, it should make a connection to external database via JDBC and get multiple row data for every Order_ID. We are not using Oracle Stored Procedure, instead we are using simple Select statement. Did all the mapping and configured correctly. However, we are getting short-dump when we execute RFC. When we checked in XI, it says " Check xml instance is valid for source xsd and target-field mapping fulfills requirements of target xsd at "
    Has anyone tried getting multiple row information from external database(Oracle) using Select statement in XI ? IF so, can you pl let us know if we need to something different in Mapping ?
    Niranjan

    Niranjan,
    Check this blog of mine to see how the datatype for Synchronous Select should be constructed,
    https://weblogs.sdn.sap.com/pub/wlg/3928. [original link is broken] [original link is broken] [original link is broken] [original link is broken]
    If this looks good, a few checks
    1. Check the request inbound payload , i.e the RFC request in MONI. Is this as expected?
    2. Check the output of the request mapping and check if this is as expected.
    3. Check the response from JDBC, is this as per the datatype defined.
    4. Check the mapping from JDBC response to RFC response.
    To me this looks like an issue with the JDBC response structure.
    Meanwhile to understand how to test mapping, check blog of michal,
    /people/michal.krawczyk2/blog/2005/09/16/xi-how-to-test-your-mapping-in-real-life-scenarios
    Regards
    Bhavesh

  • CDC Error (JOURNALIZED DATA is not loading into tartet database)

    HI,
    I have enabled source database for CDC and got green mark on source database model tables.
    while inserting data into source, J$ tables updating JRN_SUBSCRIBER and other filed.
    when I run the package/ interface JOURNALIZED DATA is not loading into target database.
    i have implemented cdc for 7 source table.
    and using JKM MSSQL Simple
    and enable JOURNALIZED DATA in interface level.
    and
    source database is : MSSQL Server
    Target Database : Oralce 11g.
    please advice me.
    thanks in advance.
    Zakeer Hussain

    Zakeer look into this link , -http://odiexperts.com/?p=1096 . Hope this helps.
    also before running Can you right click on the Source Datastore and click on Journal Data and can you see the data ? and if still the data is not passing through ,in that case make temporary objects t- yes in LKM , IKM and debug and see at which step data is not flowing through and look if there is any filter or condition which is stopping it .
    Still not able to figure out please tell us which step the data is not flowing through we will try to guide you.

  • Exracting data from two tables existing in external database

    Hi SDN friends,
    I have 2 tables in external database system. I cannot create a view as I am using Netweaver Java Dictionary. I therefore would like to get the data on BI side & perform join there. TABLE1 & TABLE2 has many to many relationship.
    <b>TABLE1</b> - FieldA, FieldB
    <b>TABLE2</b> - FieldB, FieldC(texts)
    Please help me find the best approach to populate the data into the BI infoobjects.
    I assure I will give full points to anyone who replies.
    Thanks,
    Sunil

    Hi
    If you are having data into 2 infoobjects, then its as good as having data nto tables. So now the question comes, whts the scenario?
    You can use these twe infoobjects in ODS ..create update rule and so on....
    Sunil, I m not able to understand further whts the problem you are facing exactly....
    You want help with ODS or wht?
    If you want to have join on infobjects then we can go ahead with infosets...
    Plz expain problem so tht we can work ahead
    Regards
    Sorabh
    Dont forgot to assign points in SDN -****
    Message was edited by: Sorabh Arora

  • Data transfer from oracle database(external database) to sap database

    Hi Abapers,
    I have a requirment like excel data upload from oracle database(external database) to SAP database .Plz any one can help me how to create a bdc program for this requarment.
    Advance thanks you.
    Jnana

    HI
    use function module 'ASLM_EXCEL_TO_INTERNAL_TABLE' TO UPLOAD DATA FROM EXCEL SHEET TO INTERNAL TABLE. FROM THERE U CAN SAVE IT TO DATABASE USING BDC FUNCTION MODULES.

  • Loading xml data to a oracle database problem in style sheet

    Hi,
    I Have one small Problem While loading xml data to a oracle database.
    In the XML i have the Columns and Data Like this.
    <Data>
    <TRAN>
    <Type_Trs>A</Type_Trs>
    <T1>2</T1>
    <T2>3</T2>
    </TRAN>
    <TRAN>
    <Type_Trs>B</Type_Trs>
    <T1>2</T1>
    <T2>3</T2>
    </TRAN>
    </Data>
    I have TRAN Table having Field Like this.
    TRAN(Type_Trs,T1,F)
    Note:- The XML Column T2 is Not matching with TRAN Table Column F
    I want to fetch T2 data into F column.
    When I am fetching the xml data into oracle table only
    one row is fetching.
    Can You help me
    To load the the all the record of XML data into the Oracle Table
    What will be my XLS Style Sheet File Please
    suggest me.
    Regards
    MBR
    Thnks in Advance

    Hi MBR
    You would be better posting your question to the general XML forums:
    http://forums.oracle.com/forums/category.jspa?categoryID=51
    This is purely for XML/BI Publisher where we get the data back out of the db :o)
    Tim

  • Oracle Database Table data Load it into Excel

    Hello All,
    Please I need your help for this problem:
    I need to load Oracle database Table data and load it into Excel and saved with xls format.
    Example -Select * from Slase data load it into the Excel.
    I appreciate ans sample code to help me do that, Please help me out. This is very urgent.
    Thanks alot and best regards,
    anbu

    >
    I need to load Oracle database Table data and load it into Excel and saved with xls format.
    Example -Select * from Slase data load it into the Excel.
    I appreciate ans sample code to help me do that, Please help me out. This is very urgent.
    >
    Nothing in these forums is 'urgent'. If you have an urgent problem you should contact Oracle support or hire a consultant.
    You have proven over and over again that you are not a good steward of the forums. You continue to post questions that you say are 'urgent' but rarely take the time to mark your questions ANSWERED when they have been.
    Total Questions: 90 (78 unresolved)
    Are you willing to make a commitment to to revisit your 78 unresolved questions and mark them ANSWERED if they have been?
    The easiest way to export Oracle data to Excel is to use sql developer. It is a free download and this article by Jeff Smith shows how easy it is
    http://www.thatjeffsmith.com/archive/2012/09/oracle-sql-developer-v3-2-1-now-available/
    >
    And One Last Thing
    Speaking of export, sometimes I want to send data to Excel. And sometimes I want to send multiple objects to Excel – to a single Excel file that is. In version 3.2.1 you can now do that. Let’s export the bulk of the HR schema to Excel, with each table going to it’s own workbook in the same worksheet.
    >
    And you have previously been ask to read the FAQ at the top of the thread list. If you had done that you would have seen that there is a FAQ for links that have many ways, with code, to export data to Excel.
    5. How do I read or write an Excel file?
    SQL and PL/SQL FAQ

Maybe you are looking for

  • ITunes crashes after startup XP

    What I'm seeing after a successful re-install and startup of iTunes 9.0.3 in the *Windows Event Viewer (Application)*. *Application Error –* Faulting application itunes.exe, version 9.0.3.15, faulting module quicktimeaudiosupport.qtx, version 7.9.0.0

  • Xf86-video-ati (6.13.2-2 - 6.14.0-1) makes xorg unusable *workaround*

    Yesterday, I have upgraded the system by running pacman -Suy. The following packages were updated: [2011-02-07 13:20] Running 'pacman -Suy' [2011-02-07 13:20] synchronizing package lists [2011-02-07 13:20] starting full system upgrade [2011-02-07 13:

  • How do I remove a space before the first word in a sentence using find and replace or GREP?

    Right, so I am laying out a document in inDesign from a word document and it has a space before every first word in a sentence. It is really annoying and I need a quick way to get rid of them. I wasn't too sure how to do it with find/change but I am

  • Why must this method invocation be caught?!

    Hi there. I have this following piece of code: public class Dog extends Animal {      public static void main(String[] args) {           Animal animal = new Dog();           System.out.println(animal.getAge(args[0]));      int getAge(String arg) {   

  • Unable to find Update rules in BI content

    Hi All, We are unable to find the update rules for the datasource 2LIS_03_BX 2LIS_03_BM 2LIS_03_UF in the Business Content for the cube 0IC_C03 But we can find the update rules for the datasource 2LIS_40_S278 for the cube 0IC_C03 The BI content was o