Weekend IT recovery saves customer from SQL failure

Plan B Disaster Recovery saved a customer from IT downtime at the weekend when they experienced a SQL cluster failure on Saturday afternoon.The recruitment company…
Read More »The post Weekend IT recovery saves customer from SQL failure appeared first on Disaster Recovery Solutions | IT Disaster Recovery | Plan B Better IT protection for less.
Read More
This topic first appeared in the Spiceworks Community

Yeah: what dj_paige said, or try copying catalog to a writable storage location, like your local hard drive.

Similar Messages

  • Save output from sql

    Is it possible to save the output from a sql query which returns 260,000 rows? I can't save it with sqldeveloper. Five thousand lines were saved as a LST file in sqlplus and I cannot open it. How can I get the output from a sqlplus query saved as a csv file? What is the best method for saving the output?

    Spool it to a file
    http://www.oracle.com/pls/db102/search?remark=quick_search&word=spool&tab_id=&format=ranked
    or
    UTL_FILE
    http://www.oracle.com/pls/db102/search?word=UTL_FILE&partno=
    or
    APEX
    Edited by: hoek on Jun 12, 2009 7:50 PM

  • Save files from SQL DB and Sharepoint 2003

    I am needing to save a bunch of files from our Sharepoint 2003 site. I have figured out that they are stored in an SQL database on the server. The only problem is, I honestly have no clue how to retrieve them. Help? Please?
    This topic first appeared in the Spiceworks Community

    Nope, you can patch in either order. Should not be any conflicts.
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Save results from sql worksheet

    Hi,
    Is there a way to save the results of worksheet to a tab-delimited text file which has the first line the column names?
    I can copy and paste the results now, but they don't include the column names.
    On text file save would be best to include all rows, not only the ones displayed in worksheet.
    Thanks

    Should performance improve if the second statement uses bind variables for the date columns? Probably, it depends on how expense the parse is compared to the rest of teh query execution. But generally, if you've got hundreds of statements which are the same except for literals then it's usually a good idea to replace them with a statement that uses bind variables. The best thing to do is benchmark it.
    Cheers, APC

  • Utl_http package from our custom pl/sql package.

    We have a requirement to invoke a thrid party Url that uses HTTPS. For this we are using the call to utl_http package from our custom pl/sql package. When we invoke the custom package form Oracle Forms it works fine. But when we try to invoke the same from ApplicationModule Class in our custom OA Framework form we get the following error.
    ORA-29273: HTTP request failed
    ORA-06512: at "SYS.UTL_HTTP", line 1577
    ORA-28862: SSL connection failed

    Duplicate post.
    ApplicationModule class in a custom OA Framework
    ApplicationModule class in a custom OA Framework

  • Failure to connect to 11G XE from Sql Developer

    Hi Gurus,
    I have been trying for more than a week to make this work. Basically I am trying to install XE database 11.2.0 on my windows 7 64 bit computer.
    I have had problems with the getting the listener up , finally managed to get it up by deleting the Oracle_home and reinstalling 11g XE.
    Changed the listener.ora file to include the sid XE - restarted th listener so now its working.
    -- I am able to start the instance from command prompt as
    SQL> conn / as sysdba
    Connected.
    SQL> startup
    ORACLE instance started.
    Total System Global Area 1071333376 bytes
    Fixed Size 1388352 bytes
    Variable Size 620757184 bytes
    Database Buffers 444596224 bytes
    Redo Buffers 4591616 bytes
    ORA-00205: error in identifying control file, check alert log for more info**
    When I checked the log files here are the details
    from C:\Oracle11XE\app\oracle\diag\rdbms\xe\xe\alert
    <txt>ORA-00210: cannot open the specified control file
    ORA-00202: control file: &apos;C:\ORACLE11XE\APP\ORACLE\ORADATA\XE\CONTROL.DBF&apos;
    ORA-27041: unable to open file
    OSD-04002: unable to open file
    O/S-Error: (OS 2) The system cannot find the file specified.
    </txt>
    Please let me know how I can resolved this issue. The eventual goal is to get the database up and running and be able to conenct from Sql Developer installed on the same machine.
    any help will be greatly appreciated.
    Here are some details
    tnsnames.ora
    XE =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = <Cumputername>)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SID = XE)
    EXTPROC_CONNECTION_DATA =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    (CONNECT_DATA =
    (SID = PLSExtProc)
    (PRESENTATION = RO)
    ORACLR_CONNECTION_DATA =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    (CONNECT_DATA =
    (SID = CLRExtProc)
    (PRESENTATION = RO)
    listener.ora
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = C:\Oracle11XE\app\oracle\product\11.2.0\server)
    (PROGRAM = extproc)
    (SID_DESC =
    (SID_NAME = CLRExtProc)
    (ORACLE_HOME = C:\Oracle11XE\app\oracle\product\11.2.0\server)
    (PROGRAM = extproc)
    (SID_DESC =
    (SID_NAME = XE)
    (ORACLE_HOME = C:\Oracle11XE\app\oracle\product\11.2.0\server)
    (PROGRAM = extproc)
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    (ADDRESS = (PROTOCOL = TCP)(HOST = <ComputerName>)(PORT = 1521))
    DEFAULT_SERVICE_LISTENER = (XE)
    lsnrctl status result
    LSNRCTL> status
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=EXTPROC1)))
    STATUS of the LISTENER
    Alias LISTENER
    Version TNSLSNR for 32-bit Windows: Version 11.2.0.2.0 - Produ
    ction
    Start Date 30-JAN-2013 11:35:47
    Uptime 0 days 0 hr. 0 min. 13 sec
    Trace Level off
    Security ON: Local OS Authentication
    SNMP OFF
    Default Service XE
    Listener Parameter File C:\Oracle11XE\app\oracle\product\11.2.0\server\network
    \admin\listener.ora
    Listener Log File C:\Oracle11XE\app\oracle\diag\tnslsnr\ZaidHP\listener\
    alert\log.xml
    Listening Endpoints Summary...
    (DESCRIPTION=(ADDRESS=(PROTOCOL=ipc)(PIPENAME=\\.\pipe\EXTPROC1ipc)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=ZaidHP)(PORT=1521)))
    Services Summary...
    Service "CLRExtProc" has 1 instance(s).
    Instance "CLRExtProc", status UNKNOWN, has 1 handler(s) for this service...
    Service "PLSExtProc" has 1 instance(s).
    Instance "PLSExtProc", status UNKNOWN, has 1 handler(s) for this service...
    Service "XE" has 1 instance(s).
    Instance "XE", status UNKNOWN, has 1 handler(s) for this service...
    The command completed successfully
    Thanks
    Edited by: user12026382 on Jan 30, 2013 10:57 AM

    Hi clcarter
    Thanks for your input. I actually did a complete de-install and re-install.
    The two things I did differently this time was
    1. Delete any ORACLE_HOME or ORACLE_SID variables that were existing and just to make sure check from cmd> set ORACLE_HOME - should return not defined.
    (This I found at a different forum)
    2. Run the setup (as administrator) this time.
    And you guessed it.. It worked. I was able to login to the database. Also could go to the database home page.
    Thanks for your inputs.
    On a side note.. since I am new to forums, I wanted to know how you quote things in the grey area as you did when you replied 'ORA-00205: error in identifying control file, check alert log for more info'.

  • How to get changes of a custom PL/SQL package updated in ISG?

    Hi All,
    I need some help on how to change a custom PL/SQL package in ISG integration repository:
    I already uploaded my custom PL/SQL package to the integration repositoy some time ago. Now I have made some changes to PL/SQL record/table types defined in the package (added some fields and removed some others). These changed record/table types are used by a procedure I already uploaded to the integration repository. When regenerating the WSDL for my PL/SQL package in integration repository I get an error that the wrapper for respective procedure can't be found. When checking up the SQL types that were automatically generated (by JPublisher) for corresponding PL/SQL types, these generated SQL types are invalid...
    So it seems that in one case JPublisher generated an empty SQL type (with no attributes). In another case JPublisher generated the SQL type with old attribute list, so it didn't adopt the changes of corresponding PL/SQL type in the package...
    Can anybody tell me how to get the changes of a cusotm PL/SQL package updated in ISG? Or is there any possibility to first remove a whole PL/SQL package from the ISG integration repository (and getting rid of all old automatically generated SQL types), so that I can newly upload my PL/SQL package afterwards?
    Thanks
    Konrad
    Edited by: Konrad on 24.10.2011 10:51

    Thanks Daniel, for your reply...
    (-> for info: I'm a colleague of Konrad...)
    Yes, we used a higher version number - and the upload of iLDT file works fine...
    But, the problem occured when generating the WSDL...
    However, I think we have found an approach of how to get the changes of a custom PL/SQL package updated correctly in ISG:
    1. Save a copy of PL/SQL package and then delete it in the database
    2. Regenerate the WSDL for PL/SQL package in ISG (seems that all atutomatically generated JPub wrappers and SQL types are also deleted)
    3. Restart all processes
    4. Create the changed version of PL/SQL package in the database again
    5. upload of iLDT
    6. Regenerate the WSDL for PL/SQL package in ISG
    Best Regards
    Carolin

  • How do I save files from my HD onto Time Machine?

    If I delete files on my hard drive to free up space (i.e., movie files that I want to save in an external harddrive), time machine will not keep these files for me on my external harddrive, since time machine just backs up what is in Macintosh HD. Right? How do I make it so Time Machine does NOT delete these files from it's archive?

    beccasanchezmartin
    How do I make it so Time Machine does NOT delete these files from it's archive?
    You need to seriously consider options OTHER than time machine for important data.  TM premise is storing a copy of your machine, its design and purpose is not for archiving important data youve spent years working on, despite countless people who in fact, and in error, use it that way.
    since time machine just backs up what is in Macintosh HD. Right?
    Correct, time machine is an entry level backup methodology, you need to consider something other than time machine for that.   You NEVER want to work off the premise of needing to delete things.
    Buy an external HD and store vital files on there, .....Time Machine is a sytem backup, NOT a data archive, nor a primary redundancy.
    Methodology to protect your data. Backups vs. Archives. Long-term data protection
    #1. Time Machine / Time Capsule
    Drawbacks:
    1. Time Machine is not bootable, if your internal drive fails, you cannot access files or boot from TM directly from the dead computer.
    2. Time machine is controlled by complex software, and while you can delve into the TM backup database for specific file(s) extraction, this is not ideal or desirable.
    3. Time machine can and does have the potential for many error codes in which data corruption can occur and your important backup files may not be saved correctly, at all, or even damaged. This extra link of failure in placing software between your data and its recovery is a point of risk and failure. A HD clone is not subject to these errors.
    4. Time machine mirrors your internal HD, in which cases of data corruption, this corruption can immediately spread to the backup as the two are linked. TM is perpetually connected (or often) to your computer, and corruption spread to corruption, without isolation, which TM lacks (usually), migrating errors or corruption is either automatic or extremely easy to unwittingly do.
    5. Time Machine does not keep endless copies of changed or deleted data, and you are often not notified when it deletes them; likewise you may accidently delete files off your computer and this accident is mirrored on TM.
    6. Restoring from TM is quite time intensive.
    7. TM is a backup and not a data archive, and therefore by definition a low-level security of vital/important data.
    8. TM working premise is a “black box” backup of OS, APPS, settings, and vital data that nearly 100% of users never verify until an emergency hits or their computers internal SSD or HD that is corrupt or dead and this is an extremely bad working premise on vital data.
    9. Given that data created and stored is growing exponentially, the fact that TM operates as a “store-it-all” backup nexus makes TM inherently incapable to easily backup massive amounts of data, nor is doing so a good idea.
    10. TM working premise is a backup of a users system and active working data, and NOT massive amounts of static data, yet most users never take this into consideration, making TM a high-risk locus of data “bloat”.
    11. In the case of Time Capsule, wifi data storage is a less than ideal premise given possible wireless data corruption.
    12. TM like all HD-based data is subject to ferromagnetic and mechanical failure.
    13. *Level-1 security of your vital data.
    Advantages:
    1. TM is very easy to use either in automatic mode or in 1-click backups.
    2. TM is a perfect novice level simplex backup single-layer security save against internal HD failure or corruption.
    3. TM can easily provide a seamless no-gap policy of active data that is often not easily capable in HD clones or HD archives (only if the user is lazy is making data saves).
    #2. HD archives
    Drawbacks:
    1. Like all HD-based data is subject to ferromagnetic and mechanical failure.
    2. Unless the user ritually copies working active data to HD external archives, then there is a time-gap of potential missing data; as such users must be proactive in archiving data that is being worked on or recently saved or created.
    Advantages:
    1. Fills the gap left in a week or 2-week-old HD clone, as an example.
    2. Simplex no-software data storage that is isolated and autonomous from the computer (in most cases).
    3. HD archives are the best idealized storage source for storing huge and multi-terabytes of data.
    4. Best-idealized 1st platform redundancy for data protection.
    5. *Perfect primary tier and level-2 security of your vital data.
    #3. HD clones (see below for full advantages / drawbacks)
    Drawbacks:
    1. HD clones can be incrementally updated to hourly or daily, however this is time consuming and HD clones are, often, a week or more old, in which case data between today and the most fresh HD clone can and would be lost (however this gap is filled by use of HD archives listed above or by a TM backup).
    2. Like all HD-based data is subject to ferromagnetic and mechanical failure.
    Advantages:
    1. HD clones are the best, quickest way to get back to 100% full operation in mere seconds.
    2. Once a HD clone is created, the creation software (Carbon Copy Cloner or SuperDuper) is no longer needed whatsoever, and unlike TM, which requires complex software for its operational transference of data, a HD clone is its own bootable entity.
    3. HD clones are unconnected and isolated from recent corruption.
    4. HD clones allow a “portable copy” of your computer that you can likewise connect to another same Mac and have all your APPS and data at hand, which is extremely useful.
    5. Rather than, as many users do, thinking of a HD clone as a “complimentary backup” to the use of TM, a HD clone is superior to TM both in ease of returning to 100% quickly, and its autonomous nature; while each has its place, TM can and does fill the gap in, say, a 2 week old clone. As an analogy, the HD clone itself is the brick wall of protection, whereas TM can be thought of as the mortar, which will fill any cracks in data on a week, 2-week, or 1-month old HD clone.
    6. Best-idealized 2nd platform redundancy for data protection, and 1st level for system restore of your computers internal HD. (Time machine being 2nd level for system restore of the computer’s internal HD).
    7. *Level-2 security of your vital data.
    HD cloning software options:
    1. SuperDuper HD cloning software APP (free)
    2. Carbon Copy Cloner APP (will copy the recovery partition as well)
    3. Disk utility HD bootable clone.
    #4. Online archives
    Drawbacks:
    1. Subject to server failure or due to non-payment of your hosting account, it can be suspended.
    2. Subject, due to lack of security on your part, to being attacked and hacked/erased.
    Advantages:
    1. In case of house fire, etc. your data is safe.
    2. In travels, and propagating files to friends and likewise, a mere link by email is all that is needed and no large media needs to be sent across the net.
    3. Online archives are the perfect and best-idealized 3rd platform redundancy for data protection.
    4. Supremely useful in data isolation from backups and local archives in being online and offsite for long-distance security in isolation.
    5. *Level-1.5 security of your vital data.
    #5. DVD professional archival media
    Drawbacks:
    1. DVD single-layer disks are limited to 4.7Gigabytes of data.
    2. DVD media are, given rough handling, prone to scratches and light-degradation if not stored correctly.
    Advantages:
    1. Archival DVD professional blank media is rated for in excess of 100+ years.
    2. DVD is not subject to mechanical breakdown.
    3. DVD archival media is not subject to ferromagnetic degradation.
    4. DVD archival media correctly sleeved and stored is currently a supreme storage method of archiving vital data.
    5. DVD media is once written and therefore free of data corruption if the write is correct.
    6. DVD media is the perfect ideal for “freezing” and isolating old copies of data for reference in case newer generations of data become corrupted and an older copy is needed to revert to.
    7. Best-idealized 4th platform redundancy for data protection.
    8. *Level-3 (highest) security of your vital data. 
    [*Level-4 data security under development as once-written metallic plates and synthetic sapphire and likewise ultra-long-term data storage]
    #6. Cloud based storage
    Drawbacks:
    1. Cloud storage can only be quasi-possessed.
    2. No genuine true security and privacy of data.
    3. Should never be considered for vital data storage or especially long-term.
    4. *Level-0 security of your vital data. 
    Advantages:
    1. Quick, easy and cheap storage location for simplex files for transfer to keep on hand and yet off the computer.
    2. Easy source for small-file data sharing.
    Make a (some) Redundancy of data
    Secondary data on a single external HD or RAID array is not a redundancy, but a PRIMARY data copy.          First Redundancy takes place on a Secondary storage medium/locus. (both of these external of course)
    2nd level protection in any redundancy necessitates it "be unconnected".           Data ON the actual computer is not considered, ergo a single copy OFF a computer is just that, a backup / copy.  Redunancy begins as secondary to a FIRST COPY

  • Chart from SQL Colour

    I am creating reports in information publisher and wish to include charts from SQL to summarise the data. As some of these reports will include job execution status and similar status fields it would be useful to show Failures in Red and Successes in Green like the "All Targets Status" region on the Enterprise Manager Grid Control home page.
    Is there any way I can influence the colour that items are plotted in Pie and other Charts? I am using release Oracle Enterprise Manager 10g Grid Control Release 10.2.0.1.0.
    Thanks in advance for any help you can offer.
    Regards,
    Matthew

    Hi Siegwin,
    Although the data section of your generated chart XML contains references to styles, such as CadetBlue2, yellow, etc, you haven't actually defined those styles within the <period_styles> tags of the XML. Therefore, the chart will apply the default blue colour to each timeline of the chart. You'll see that there's a default style called defaultStyle already generated for you in the XML i.e. <period_style name="defaultStyle"> What you need to do is create your own styles, mirroring the defaultStyle example, in order to apply other colours to your gantt chart. I've updated your XML to include a new yellow style, so you can see that one of the elements of the chart now picks up that colour.
    My sample chart, Resource Gantt with Custom Styles, should hopefully help to explain what you need to do in order to apply custom colours to your chart: http://apex.oracle.com/pls/apex/f?p=36648:60
    I hope this helps.
    Regards,
    Hilary

  • Error happened during the migration from sql server 2000 to oracle 8.1.7

    /*[SPCONV-ERR(18)]:CREATE TABLE TT_CHANP1 statement passed to ddl file*/
    DELETE FROM TT_CHANP1;
    error as above.
    note:
    chanp1 is a temporary table under sql 2000
    syntax:
    create table #chanp1 (chanpid varchar(50))
    we try the codes:
    CREATE TEMP TABLE chanp1
    Error: PLS-00103: "CREATE"
    begin declare end exception
    exit for goto if loop mod null pragma raise return select update while
    <an identifier><a double-quoted delimited-identifier>
    <a bind variable><<close current delete fetch lock insert open
    roll back save point set sql execute commit for all
    <a single-quoted SQL string>
    Line: 63
    we try " CREATE GLOBAL TEMPORARY TABLE chanp1" again.
    but error .

    it seems like you are trying to perform a DDL from within the PL/SQL code, which is not allowed directly.
    You could use EXECUTE IMMEDIATE within your pl/sql code to perform a DDL statement.

  • How to open and save pictures from 1 folder at the same time?

    Is there a way to open and save pictures from 1 folder at the same time?
    I'm creating a vi that resizes the picture but I want to do it with 1 folder at a time... Is this possible?
    greetings!

    Hi Kway,
    Here you are. I wrote a file a customer times ago & It did what you want!
    Just see the attachements and keep me inform.
    Regards,
    Julien Roland - District Sales Manager
    NI Belgium - Technical Support
    Don't forget to rate a good answer
    Attachments:
    Test existing Files.vi ‏23 KB
    Delete files from array of files (SubVI).vi ‏16 KB

  • How to load and save custom workspaces in PE7

    I've written a little utility which allows PE7 users to load and save custom workspaces. It's attached to post #1 of this thread.
    System requirements
    The utility has been tested with Windows XP Pro SP3 only. It might or might not run under Vista. It will only work with PE7 installations where the executable files are in the default places.
    Installation
    When you have downloaded the utility (it's only 199KB) you can run it from any convenient place in your system - it doesn't need to be installed as such.
    Usage
    It's pretty self-explanatory in use but here's the manual!
    Run the downloaded utility "PE7WS.exe".
    A file selection dialog appears. Choose the workspace layout file you wish to use. "LastUsed.layout" is the last used workspace (that's all you will be offered the first time you run the program). If you press Cancel at this point, the program will exit and nothing will happen.
    Once you have selected the workspace, click "Open" and Premiere Elements 7 will run, using the workspace layout you chose.
    After you exit from Premiere Elements 7, a workspace layout file save dialog will appear. If you made changes to the workspace during your Premiere Elements 7 session, you can now give the revised workspace a name and save it.
    Do not save the workspace in a different directory from that which the dialog offers!
    If you have changed an existing custom workspace, and you want to update the version previously stored, just select the name and over-write the saved workspace layout with the new one.
    If you don't want to save the workspace at all, just click "Cancel" - next time you run Premiere Elements, you can choose the "LastUsed.layout" workspace to use that last used unsaved workspace in any event.
    If you use the PE7WS utility to run Premiere Elements 7, and then you use the menu option "Window > Restore Workspace", then exit, you can either cancel the workspace saving stage, (which will mean that "LastUsed.layout" will contain that default workspace), or you could give that default workspace a name like "PE7 Default" so you can start a future Premiere Elements 7 session with the default workspace, even though you last used a custom workspace.
    If you run Premiere Elements 7 in the usual way without using PE7WS.exe, the workspace that appears will be the last used one.
    The usual disclaimer...
    The utility should be entirely safe in use, but I can't accept responsibility for any loss or damage it might cause. However, the only file manipulation is does is to workspace files in the directory that Premiere Elements 7 uses purely for the purpose - the utility doesn't go anywhere near your precious project files.

    I'm finding it very handy on my particular twin 22" monitor setup to sometimes use a one-screen layout and sometimes a two screen layout. I 'discovered' that having the timeline on the second monitor enables me to tweak edits there while having the playback full size after pressing the "full screen" button on the first monitor. Where the timeline would normally be I have the history, mixer, and other windows always open. As the second monitor is also used by another PC (the one I have the net and email etc on) using "Maxivista" monitor sharing software, I revert to single screen layout when doing stuff like exporting to DVD so I can see what is going on with the render while typing stuff like this.
    Selectable workspaces are dead handy!

  • Query based Report (Reports From SQL Query)

    Hi All,
    I am facing a problem with a report. I need your help.
    I am creating a Report From SQL Query (Portal) with some arguments passed at runtime. I am able to view the output, if the query returns few rows ( arount 1000 rows). But for some inputs it needs to generate >15000 records, at this point the page is getting time out (i think!) and showing error page. I am able to execute query from the SQL Plus console ot using TOAD editor. Here the query is not taking more that 2 mins time to show the result.
    If i am executing from Portal i observed that, once i give the appropriate input and hit submit button a new oracle process is getting created for the query on UNIX (I am usign "TOP" command to check processes). The browser page will be shown error page after 5 minutes (i am assuming session time out!) , but on the backend the process will be executed for more than 30 mins.
    Can i increase a timeout of portal page, how i can do this?
    The data returned as a result of the query is sized more than 10 MB. Is caching this much data is possible by the browser page? is the returned data is creating any problem here.
    Please help me to find appropriate reasone for the failure?

    Can you post the URL that is giving you the page not found error ? I am not sure if I can be of much help though.
    In order to increase your chances to get a reply you should also post your question in the Portal Forum - Oracle Application Server Portal
    Edited by: Rodolfo Ferrari on Jun 16, 2009 3:49 PM

  • Need help in Report From SQL Query

    Hi All,
    I am facing a problem with a report. I need your help.
    I am creating a Report From SQL Query (Portal) with some arguments passed at runtime. I am able to view the output, if the query returns few rows ( arount 1000 rows). But for some inputs it needs to generate >15000 records, at this point the page is getting time out (i think!) and showing error page. I am able to execute query from the SQL Plus console ot using TOAD editor. Here the query is not taking more that 2 mins time to show the result.
    If i am executing from Portal i observed that, once i give the appropriate input and hit submit button a new oracle process is getting created for the query on UNIX (I am usign "TOP" command to check processes). The browser page will be shown error page after 5 minutes (i am assuming session time out!) , but on the backend the process will be executed for more than 30 mins.
    I tried also increase the page time out in httpd.conf, but no use.
    The data returned as a result of the query is sized more than 10 MB. Is caching this much data is possible by the browser page? is the returned data is creating any problem here.
    Please help me to find appropriate reasone for the failure?

    user602513 wrote:
    Hi All,
    I am facing a problem with a report. I need your help.
    I am creating a Report From SQL Query (Portal) with some arguments passed at runtime. I am able to view the output, if the query returns few rows ( arount 1000 rows). But for some inputs it needs to generate >15000 records, at this point the page is getting time out (i think!) and showing error page. I am able to execute query from the SQL Plus console ot using TOAD editor. Here the query is not taking more that 2 mins time to show the result.
    If i am executing from Portal i observed that, once i give the appropriate input and hit submit button a new oracle process is getting created for the query on UNIX (I am usign "TOP" command to check processes). The browser page will be shown error page after 5 minutes (i am assuming session time out!) , but on the backend the process will be executed for more than 30 mins.
    I tried also increase the page time out in httpd.conf, but no use.
    The data returned as a result of the query is sized more than 10 MB. Is caching this much data is possible by the browser page? is the returned data is creating any problem here.
    Please help me to find appropriate reasone for the failure?Do you get any errors or warnings or it is just the slow speed which is the issue?
    There could be a variety of reasons for the delayed processing of this report. That includes parameter settings for that page, cache settings, network configurations, etc.
    - explore best optimization for your query;
    - evaluate portal for best performance configuration; you may follow this note (Doc ID: *438794.1* ) for ideas;
    - third: for that particular page carrying that report, you can use caching wisely. browser cache is neither decent for large files, nor practical. instead, explore the page cache settings that portal provides.
    - also look for various log files (application.log and apache logs) if you are getting any warnings reflecting on some kind of processing halt.
    - and last but not the least: if you happen to bring up a portal report with more than 10000 rows for display then think about the usage of the report. Evaluate whether that report is good/useful for anything?
    HTH
    AMN

  • Using a custom PL/SQL to populate the primary key in a tabular form

    I want to use a Custom PL/SQL Function to populate the primary key when I insert a new record into a tabular form. I want to get the value from a hidden page Item. The code I am using for the primary key source is:
    BEGIN
    INSERT INTO TEAM_MEMBERS(TEAM_ID)
    VALUES(:P75_TEAM_ID);
    END;
    When I try to insert a new record I get the following error:
    Error      ERR-1904 Unable to compute item default: type = Function Body computation_type= BEGIN INSERT INTO TEAM_MEMBERS(TEAM_ID) VALUES(:P75_TEAM_ID); END; .
    ORA-06550: line 5, column 2: PLS-00103: Encountered the symbol ";" when expecting one of the following: begin case declare end exception exit for goto if loop mod null pragma raise return select update while with << close current delete fetch lock insert open rollback savepoint set sql execute commit forall merge pipe The symbol "exit" was substit
    Any ideas what I am doing wrong?
    Thanks!

    Brian - Sometimes whitespace at the end of the block causes this. Be sure to trim everything after the last semicolon including tabs and newlines.
    Scott

Maybe you are looking for