EXP|ANDING MY G4

Hi Forum,
Looking for a wee bit of advice on best way to upgrade my G4 - it's kinda running pretty slow. Not looking for to much just to surf the net & play some music really.
Really just a wee bit more speed and storage. Where is the best place to shop for these parts and is there any specific requirements for the RAM and HD for this machine. New to Mac.
Thanks.
My system profile looks like:
Hardware Overview:
Machine Model: Power Mac G4 (AGP graphics)
CPU Type: PowerPC G4 (2.9)
Number Of CPUs: 1
CPU Speed: 400 MHz
L2 Cache (per CPU): 1 MB
Memory: 256 MB
Bus Speed: 100 MHz
Boot ROM Version: 4.2.8f1
20G hd storage
Eithernet Network Card
AGP Card
ATY,Rage128Pro: AGP Slot-A 16MB
Memory:
DIMM0/J21:Empty
DIMM1/J22:Empty
DIMM2/J23:Empty
DIMM3/J24:
Size: 256 MB
Type: SDRAM
Speed: PC133-333

While OWC is a good place to shop for Mac upgrades here in the US, perhaps some of the Forum regulars in the UK will post some sources that are a bit closer to you. You can use the same PC-100 (or backward-compatible PC-133) memory that's used in older PCs, as long as it has quality components and isn't registered or ECC SDRAM. Your G4 can use a standard 3.5" IDE/EIDE hard drive, parallel ATA-100 or ATA-133 (a SATA drive requires a Mac SATA controller PCI card). The G4 AGP-400 model has a maximum capacity recognition limit of 128 GBs or less, when the drive is connected to the motherboard's IDE controller. This effectively makes a 120 GB drive the ideal choice, because anything larger will have wasted/unused storage capacity, unless you connect it to a Mac-compatible ATA-133 controller PCI card. Once installed, the new hard drive must be formatted with Apple's "Drive Setup" (pre-OS X) or "Disk Utility" (OS X), before the OS software can be installed on it. Faster G4 processor cards are available, but you really have to decide in advance, how much you're willing to invest in a 7 year-old computer. It's very easy to spend a lot of money to improve an older Mac, and have a computer that's still worth less than the sum of its upgraded parts, simply because of its inherent hardware limitations, like a slow 100 MHz system bus speed.

Similar Messages

  • Error while converting class file to exp and jca file

    error while converting *.class file to *.exp and *.jca file
    =====================================================================================================================
    linux-y60u:/home/admin/java_card_kit-2_2_1/samples/src # converter -exportpath "/home/admin/java_card_kit-2_2_1/lib/" com/sun/javacard/samples/HelloWorld 0x00:0x01:0x02:0x03:0x04:0x05:0x06:0x07:0x0b 1.0 -v -applet 0x00:0x01:0x02:0x03:0x04:0x05:0x06:0x07:0x0b:0x01 Identity
    Java Card 2.2.1 Class File Converter, Version 1.3
    Copyright 2003 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms.
    parsing /home/admin/java_card_kit-2_2_1/samples/src/com/sun/javacard/samples/HelloWorld/HelloWorld.class
    parsing /home/admin/java_card_kit-2_2_1/samples/src/com/sun/javacard/samples/HelloWorld/Identity.class
    error: com.sun.javacard.samples.HelloWorld.HelloWorld: unsupported class file format of version 50.0.
    error: com.sun.javacard.samples.HelloWorld.Identity: unsupported class file format of version 50.0.
    conversion completed with 2 errors and 0 warnings.
    =====================================================================================================================

    i compile a file javacard use this command:
    ===
    javac -source 1.3 -target 1.1 -g -classpath ./classes:../lib/api.jar:../lib/installer.jar src/com/sun/javacard/samples/Identity/Identity.java
    ===
    and try to convert this class use this command
    ===
    /home/xnuxerx/admin/java_card_kit-2_2_1/bin/converter -exportpath "/home/xnuxerx/admin/java_card_kit-2_2_1/lib/" com/sun/javacard/samples/Identity 0x00:0x01:0x02:0x03:0x04:0x05:0x06:0x07:0x0b 1.0 -v -applet 0x00:0x01:0x02:0x03:0x04:0x05:0x06:0x07:0x0b:0x01 Identity
    ===
    result convert:
    ===
    Java Card 2.2.1 Class File Converter, Version 1.3
    Copyright 2003 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms.
    parsing /home/xnuxerx/admin/java_card_kit-2_2_1/samples/classes/com/sun/javacard/samples/Identity/Identity.class
    converting com.sun.javacard.samples.Identity.Identity
    error: export file framework.exp of package javacard.framework not found.
    conversion completed with 1 errors and 0 warnings.
    ===
    why ??
    please your comment for this problem.
    thank 4 all.

  • Hoping for a quick response : EXP and Archived REDO log files

    I apologize in advance if this question has been asked and answered 100 times. I admit I didn't search, I don't have time. I'm leaving on vacation tomorrow, and I need to know if I'm correct about something to do with backup / restore.
    we have 10g R2 running a single instance on a single server. The application vendor has "embedded" oracle with their application. The vendor's backup is a batch file using EXP - thus:
    exp system/xpwdxx@db full=y file=D:\Orant\admin\db\EXP\db_full.dmp log=D:\Orant\admin\db\EXP\db_full.txt direct=y compress=y
    This command is executed nightly at midnight. The files are then backed up by our nightly backup to offsite storage media.
    Te database is running in autoarchive mode. The problem is, the archived redo files filled the drive they were being stored on, and it is the drive the database is on. I used OS commands to move 136G of archived redo logs onto other storage media to free the drive.
    My question: Since the EXP runs at midnight, when there is likely NO activity, do I need to run in AutoArchive Mode? From what I have read, you cannot even apply archived redo log files to this type of backup strategy (IMP) Is that true? We are ok losing changes since our last EXP. I have read a lot of stuff about restoring consistent vs. inconsistent, and just need to know: If my disk fails, and I have to start with a clean install of Oracle and nothing else, can I IMP this EXP and get back up and running as of the last EXP? Or do I need the autoarchived redo log files back to July 2009 (136G of them).
    Hoping for a quick response
    Best Regards, and thanks in advance
    Bruce Davis

    Bruce Davis wrote:
    Amardeep Sidhu
    Thank you for your quick reply. I am reading in the other responses that since I am using EXP without consistent=y, I might not even have a backup. The application vendor said that with this dmp file they can restore us to the most recent backup. I don't really care for this strategy as it is untested. I asked them to verify that they could restore us and they said they tested the dmp file and it was OK.
    Thank you for taking the time to reply.
    Best Regards
    BruceThe dump file is probably ok in the sense it is not corrupted and can be used in an imp operation. That doesn't mean the data in it is transactionally consistent. And to use it at all, you have to have a database up and running. If the database is physically corrupted, you'll have to rebuild a new database from scratch before you can even think about using your dmp file.
    Vendors never understand databases. I once had a vendor tell me that Oracle's performance would be intolerable if there were more than 5 concurrent connections. Well, maybe in HIS product ..... Discussions terminated quickly after he made that statement.

  • Scripts for exp and imp in linux

    please give me some links for exp and imp tables peridical bactch script in oracle 10 R2 over RHEL 4.5.

    user13653962 wrote:
    please give me some links for exp and imp tables peridical bactch script in oracle 10 R2 over RHEL 4.5.try this.. change the same script for your environment
    $ crontab -l
    00 22 * * * sh /u01/db/scripts/testing_user1.sh
    $ cat /u01/db/scripts/testing_user1.sh
    . /home/oracle/TEST1.env
    export ORACLE_SID=TEST1
    export ORA_USER=dbdump
    export ORA_PASSWORD=dbdump
    export TNS_ALIAS=TEST1
    expdp $ORA_USER/$ORA_PASSWORD@$TNS_ALIAS tables=YOUR_tables_NAMEs  dumpfile=TEST_`date +'%Y-%m-%d`.dmp directory=YOUR_DUMPDIR logfile=TEST_`date +'%Y-%m-%d`.logif you need for original export change the exp command,
    exp username/password tables=table_name1,table_name2,etc... file=your_file_name log=your_log_file_name

  • Using expdp and impdp instead of exp and imp

    Hi All,
    I am trying to use expdp and impdp instead of exp and imp.
    I am facing few issues while using expdb. I have a Job which exports data from one DB server and then its imported so another DB server. Both DB servers are run on separate machines. Job runs on various clients machine and not on any of DB server.
    For using expdp we have to create DIRECTORY and as I understand it has to be created on DB server. Problem here is Job can not access DB Server or files on DB server. Also dump file created is moved by Job to other machines based on requirement( Usually it goes to multiple DB server).
    I need way to create dump files on server where job runs.
    If I am not using expdp correctly please guide. I am new to expdp/impdp and imp/exp.
    Regards,

    Thanks for quick reply ..
    Job executing expdb/impdp runs on Red Hat Enterprise Linux Server release 5.6 (Tikanga)
    ORACLE server Release 11.2.0.2.0
    Job can not access the ORACLE server as it does not have privileges (In fact there is no user / password to access ORACLE server machines). Creating dump on oracle server and moving is not an option for this JOB. It has to keep dump with itself.
    Regards,

  • Conversions between character sets when using exp and imp utilities

    I use EE8ISO8859P2 character set on my server,
    when exporting database with NLS_LANG not set
    then conversion should be done between
    EE8ISO8859P2 and US7ASCII charsets, so some
    characters not present in US7ASCII should not be
    successfully converted.
    But when I import such a dump, all characters not
    present in US7ASCII charset are imported to the database.
    I thought that some characters should be lost when
    doing such a conversions, can someone tell me why is it not so?

    Not exactly. If the import is done with the same DB character set, then no matter how it has been exported. Conversion (corruption) may happen if the destination DB has a different character set. See this example :
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:01 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> create table test(col1 varchar2(1));
    Table created.
    TEST@db102 SQL> insert into test values(chr(166));
    1 row created.
    TEST@db102 SQL> select * from test;
    C
    ¦
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:55 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ©
    Typ=1 Len=1: 166
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ exp test/test file=test.dmp tables=test
    Export: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:47 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P15 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                           TEST          1 rows exported
    Export terminated successfully without warnings.
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:56 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> drop table test purge;
    Table dropped.
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ imp test/test file=test.dmp
    Import: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:15 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P15 character set (possible charset conversion)
    . importing TEST's objects into TEST
    . importing TEST's objects into TEST
    . . importing table                         "TEST"          1 rows imported
    Import terminated successfully without warnings.
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:34 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ¦
    Typ=1 Len=1: 166
    TEST@db102 SQL>

  • Conversions between character sets when using 'exp' and 'expdp' utilities

    When I make export with exp utility with NLS_LANG environment
    variable not set, export is done with US7ASCII character set,
    but when server uses some other character set e.g. EE8ISO8859P2,
    there can be some national specific characters lost with such a conversion.
    But when I use expdp utility and have not NLS_LANG set,
    in log file there is not mentioned that export is done in other character set
    than that on the server, does it mean that it uses character set on server?

    When I make export with exp utility with NLS_LANG environment
    variable not set, export is done with US7ASCII character set,
    but when server uses some other character set e.g. EE8ISO8859P2,
    there can be some national specific characters lost with such a conversion.
    But when I use expdp utility and have not NLS_LANG set,
    in log file there is not mentioned that export is done in other character set
    than that on the server, does it mean that it uses character set on server?

  • Exporting using exp and TRIGGERS=N doesn't  work

    Exporting using exp on version Export: Release 8.1.6.0.0 - with the option TRIGGERS=N stills exports with triggers. is it a bug ?
    Any sugesttions ?

    This is actually a known behaviour.
    As of table level export, all the dependent objects are exported and is expected namely, indexes, constraints, triggers... and even restricting with other parameters set to n will not disable their export.
    As of schema level export, the whole schema is exported and with this we can pose our restrictions like no constraints, no rows, no indexes, no triggers etc.. which will work because we have control.
    For table level export we dont have control on individual objects. 10g solves your problem.

  • Exp and import

    Hi,
    I have two databases
    1. 10g on RHEL 4 (64 bit)
    2. 9i on RHEL (32 bit)
    I have a requirement to export from 10g and import it in 9i. Is that possible?

    Table 19-7 Using Different Releases of Export and Import
    Export from->Import to Use Export Release Use Import Release
    9.0.2 -> 10.1.0 9.0.2 10.1.0
    10.1.0 -> 9.0.2 9.0.2 9.0.2 [http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/exp_imp.htm#i1013164]
    Use 9i exp release to export from 10g and import into 9i database.

  • Data EXP and IMP

    Hi,
    I expoted(EXP) data from database by using following command in Ms-Dos.
    C:\ImpExp>exp userid=scott/tiger@orcl tables=emp
    Now EXPDAT.DMP at C:\ImpExp.
    How to import this file this into database table.
    pls guide me.
    Regards,
    Venkat

    Hi,
    U forget about previous error message. Manuall mistake in command.
    This is the error.
    D:\ImpExp>imp scott/tiger@kzt0sy file=EXPDAT.DMP log=EXPDAT.DMP.log tables=emp3
    My target table is emp3. This is empty need to load data.
    Import: Release 10.2.0.1.0 - Production on Fri Jun 20 15:52:04 2008 Copyright (c) 1982, 2005, Oracle. All rights reserved. Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production With the Partitioning, OLAP and Data Mining options Export file created by EXPORT:V10.02.01 via conventional path import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set . importing SCOTT's objects into SCOTT IMP-00033: Warning: Table "EMP3" not found in export file Import terminated successfully with warnings.

  • Database merge with the exp and imp...

    Hi All,
    I am very new to these dba activities. I have two databases with the same schema and has some data in both of it. Now I would like to merge both the databases and come up with one database. In this process I don't wanted to loose any data.
    Does the oracle exp/imp helps in this scenario ? If not any other tools helps us in doing this?
    What are the best practices to follow when we are doing this kind of work?
    What kind of verifications we need to do pre/post merge ?
    Any help would really be appreciated... Thank you in adavance...
    K.

    NewBPELUser wrote:
    Hi All,
    I am very new to these dba activities. I have two databases with the same schema and has some data in both of it. Now I would like to merge both the databases and come up with one database. In this process I don't wanted to loose any data.
    Does the oracle exp/imp helps in this scenario ? If not any other tools helps us in doing this?
    What are the best practices to follow when we are doing this kind of work?
    What kind of verifications we need to do pre/post merge ?
    Any help would really be appreciated... Thank you in adavance...
    K.What do you mean with "merging" data?
    How many objects are in both schemas?
    Are both schemas need to be merged in mutual way or in one way?
    Do you have option to use MERGE command for each object in the schema?
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14200/statements_9016.htm#SQLRF01606
    P.S. If the version of the database is greater than 9i, then use Data Pump (expdp/impdp) instead of deprecated exp/imp utitlies
    Kamran Agayev A.
    Oracle ACE
    My Oracle Video Tutorials - http://kamranagayev.wordpress.com/oracle-video-tutorials/

  • Migrating DATABASE : Exp and LOGMINER

    Hello,
    I have to migrate a 9i DATABASE from one platform (HP-UX V1 PARISC) to another platform (HP-UX V3 Itanium) and to upgrade it to 11Gr2.
    I know that multiple scenario exist (Ex:Exp/Imp,DataGuard,GoldenGate,Dblink/Direct Load,etc) but i'am interested in another solution and i'am curious to konw if it work.
    Scenario:
    1-Activate force logging and suplemental logging in the source database.
    2-Export the source database SCHEMA in consistent mode and import them in the 11gr2 database.
    3-Record the scn of le last user transaction applied before the export.
    4-Use LOGMINER to capture all the transaction applied after the export.
    5-Stop the source database, applie the captured transaction by logminer and then switch the user to the new database.
    I thougt that the down time will be minimal, it will take only the time of reapplieng the comited transaction captured by logminer.
    It's feasible to do that or logminer have some limitation ?
    It's recomended or difficult to do that ?
    Thanks for you help,

    I'am not familiar with logminer
    I don'think it would be so easy... In such situations, and whenever possible, I use RMAN and upgrade, as per MOS Doc ID 790559.1 RMAN Restore of Backups as Part of a Database Upgrade .

  • Compress=Y in Exp and Expdp

    I have been using oracle export/import as well as datapump export/import on a regular basis in building our database environments. I noticed the difference between the normal export and the datapump export.
    Export(exp) has the compress=Y/N option whereas i didnt see the option in Datapump.
    I was looking at the Documentation and found that
    http://download.oracle.com/docs/cd/B14117_01/server.101/b10825/dp_export.htm#i1005864
    "A parameter comparable to COMPRESS is not needed. " --This is what described in the documentation.
    Does Datapump export does automatic compresssion?
    I have tried exporting a table(size-14G) from our Database with "norows" option using normal export and tried Datapump export using Metadata_only option and generated the Sql for the metadata definitions.
    The script from the normal export shows the initial extent size of 2g whereas SQLfile from the datapump import shows 260M.
    Now,which is the best approach for creating objects while Data movement across Databases.
    Please clarify.

    using compress=y in export means:
    exp determines the total size of the segment, uses that as the size of the new initial extent, and sets the next extent to 10 percent of the initial extent.
    It is a remnant from the V6 days when there were only dictionary managed tablespaces, and many DBAs (including yours truly) where micromanaging table size and free space size.
    That said, even Oracle admits it was not their brightest moment, and you shouldn't use it, ever.
    Obviously introducing expdp, in a LMT world, was the moment to get rid of it.
    Sybrand Bakker
    Senior Oracle DBA

  • Strange behavior when searching a phrase using reg exp and dynamic sql

    Hi,
    I have a strange issue while using dynamic sql for an apex page. I have a requirement to search a string in the database column which is entered by user on a page item. The search process should search the whole phrase only.
    I have a query generated dynamically in the back end and use it in a cursor in the stored procedure
      SELECT t.group_cn , t.group_desc, t.group_type, t.partner_organization_id, t.partner_organization
      FROM vr_idm_group t WHERE regexp_like(t.group_desc,'(^|\W)HR Selection & Assignment(\W|$)', 'i')The pl sql code with the dynamic sql statements are below.
       IF p_search_process NOT in ('PARTNER') THEN
          OPEN v_cursor FOR v_sql;
       ELSE
          OPEN v_cursor FOR v_sql USING p_search_id;
       END IF;
       LOOP
          FETCH v_cursor INTo v_obj.group_cn, v_obj.group_desc, v_obj.group_type, v_obj.partner_organization_id,
             v_obj.partner_organization, v_obj.match_count;
          EXIT WHEN v_cursor%NOTFOUND ;
          v_search_array.extend;
          v_search_array(v_search_array.last) := v_obj;
          dbms_output.put_line(v_sql);
       END LOOP;The search works fine if the search string does not contain any special character like &,- etc.
    However, if the search string contains any special character, it does not return any thing. This strange issue happens only if I call the procedure from the apex page and the search string contains a special character. (please note that the procedure works fine even from apex if the string does not have a special character). When I debugged this, found that, the cursor does not fetch any rows (it is supposed to fetch two rows) for unknown reason. When I run the query separately, it returns the two rows (in which the column group_desc contains the search string "HR Selection & Assignment") as desired. Also, when I test the procedure in the back end (PLSQL developer), it works fine.
    Any idea, what is causing this strange behaviour?
    Advance thanks.
    Regards,
    Natarajan

    i don't see anything about a dataProvider.  you're assigning a source for a scrollpane.  scrollpane's don't have a dataProvider property.
    anyway, other than arrayRun always being false when that last if-statement executes, what's the problem?  doesn't that movieclip display when that 2nd branch of the last if-statement executes (assuming instance is defined correctly etc)?

  • Tables for previous emp exp and present.

    Hi experts,
    I created a view based on 2 tables i.e PA0023 - HR Master Record: Infotype 0023 (Other/Previous
    Employers) and PA0000 - HR Master Record: Infotype 0000 (Actions).
    But I am unable to get the records of employees who are freshers.
    Bcz PA0023 does not hold freshers data.
    Is there any table which holds both the Previous employers data as well as freshers data.
    Regards,
    Nishuv.

    hi
    depends what freshers data you are wanting to update, there is no separate table for freshers employment data (as they are freshers) you can use the same infotype to store their apprentice ship details etc.......if you do not want the same to be counted as experience in some report that you may have developed for experience you may have to put a custom field for maintaining the same or configure the work/contract type here.
    hope this helps.....or give more details
    best regards

  • Datapump exp and imp using API method

    Good Day All,
    I want to know what is the best way of error handling of datapump export and Import using API. I need to implement in my current project as there lot of limitations and the only way to see the process worked is writing the code with error handling method using exceptions. I have seen some examples on the web but if there are practicle examples or good links with examples that will work sure way, I would like to know and explore. I have never used API method so I am not sure of it.
    Thanks a lot for your time.
    Maggie.

    I wrote the procedure with error handling but it does not out put any information of the statuses while kicking off the expdp process. I have put dbms_output.put_line as per oracle docs example but it doesnt display any messages, just kicks off and created dumpfiles. As a happy path its ok but I need to track if something goes wrong. I even stated set serveroutput on sqlplus. It doesnt even display if job started. Please help me where I made a mistake to display the status . Do I need to modify or add anything. Help!!
    CREATE OR REPLACE PROCEDURE SCHEMAS_EXPORT_TEST AS
    --Using Exception Handling During a Simple Schema Export
    --This Proceedure shows a simple schema export using the Data Pump API.
    --It extends to show how to use exception handling to catch the SUCCESS_WITH_INFO case,
    --and how to use the GET_STATUS procedure to retrieve additional information about errors.
    --If you want to get status up to the current point, but a handle has not yet been obtained,
    --you can use NULL for DBMS_DATAPUMP.GET_STATUS.http://docs.oracle.com/cd/B19306_01/server.102/b14215/dp_api.htm
    h1 number; -- Data Pump job handle
    l_handle number;
    ind NUMBER; -- Loop index
    spos NUMBER; -- String starting position
    slen NUMBER; -- String length for output
    percent_done NUMBER; -- Percentage of job complete
    job_state VARCHAR2(30); -- To keep track of job state
    sts ku$_Status; -- The status object returned by get_status
    le ku$_LogEntry; -- For WIP and error messages
    js ku$_JobStatus; -- The job status from get_status
    jd ku$_JobDesc; -- The job description from get_status
    BEGIN
    h1 := dbms_datapump.open (operation => 'EXPORT',job_mode => 'SCHEMA');
    dbms_datapump.add_file (handle => h1,filename => 'SCHEMA_BKP_%U.DMP',directory => 'BKP_SCHEMA_EXPIMP',filetype => DBMS_DATAPUMP.KU$_FILE_TYPE_DUMP_FILE);
    dbms_datapump.add_file (handle => h1,directory => 'BKP_SCHEMA_EXPIMP',filename => 'SCHEMA_BKP_EX.log',filetype => DBMS_DATAPUMP.KU$_FILE_TYPE_LOG_FILE);
    ---- A metadata filter is used to specify the schema that will be exported.
    dbms_datapump.metadata_filter (handle => h1, name => 'SCHEMA_LIST',value => q'|'XXXXXXXXXX'|');
    dbms_datapump.set_parallel( handle => h1, degree => 4);
    -- Start the job. An exception will be returned if something is not set up
    -- properly.One possible exception that will be handled differently is the
    -- success_with_info exception. success_with_info means the job started
    -- successfully, but more information is available through get_status about
    -- conditions around the start_job that the user might want to be aware of.
    begin
    dbms_datapump.start_job (handle => h1);
    dbms_output.put_line('Data Pump job started successfully');
    exception
    when others then
    if sqlcode = dbms_datapump.success_with_info_num
    then
    dbms_output.put_line('Data Pump job started with info available:');
    dbms_datapump.get_status(h1,
    dbms_datapump.ku$_status_job_error,0,
    job_state,sts);
    if (bitand(sts.mask,dbms_datapump.ku$_status_job_error) != 0)
    then
    le := sts.error;
    if le is not null
    then
    ind := le.FIRST;
    while ind is not null loop
    dbms_output.put_line(le(ind).LogText);
    ind := le.NEXT(ind);
    end loop;
    end if;
    end if;
    else
    raise;
    end if;
    end;
    -- The export job should now be running. In the following loop, we will monitor the job until it completes.
    -- In the meantime, progress information is displayed.
    percent_done := 0;
    job_state := 'UNDEFINED';
    while (job_state != 'COMPLETED') and (job_state != 'STOPPED') loop
    dbms_datapump.get_status(h1,
    dbms_datapump.ku$_status_job_error +
    dbms_datapump.ku$_status_job_status +
    dbms_datapump.ku$_status_wip,-1,job_state,sts);
    js := sts.job_status;
    -- If the percentage done changed, display the new value.
    if js.percent_done != percent_done
    then
    dbms_output.put_line('*** Job percent done = ' ||to_char(js.percent_done));
    percent_done := js.percent_done;
    end if;
    -- Display any work-in-progress (WIP) or error messages that were received for
    -- the job.
    if (bitand(sts.mask,dbms_datapump.ku$_status_wip) != 0)
    then
    le := sts.wip;
    else
    if (bitand(sts.mask,dbms_datapump.ku$_status_job_error) != 0)
    then
    le := sts.error;
    else
    le := null;
    end if;
    end if;
    if le is not null
    then
    ind := le.FIRST;
    while ind is not null loop
    dbms_output.put_line(le(ind).LogText);
    ind := le.NEXT(ind);
    end loop;
    end if;
    end loop;
    -- Indicate that the job finished and detach from it.
    dbms_output.put_line('Job has completed');
    dbms_output.put_line('Final job state = ' || job_state);
    dbms_datapump.detach (handle => h1);
    -- Any exceptions that propagated to this point will be captured. The
    -- details will be retrieved from get_status and displayed.
    Exception
    when others then
    dbms_output.put_line('Exception in Data Pump job');
    dbms_datapump.get_status(h1,dbms_datapump.ku$_status_job_error,0, job_state,sts);
    if (bitand(sts.mask,dbms_datapump.ku$_status_job_error) != 0)
    then
    le := sts.error;
    if le is not null
    then
    ind := le.FIRST;
    while ind is not null loop
    spos := 1;
    slen := length(le(ind).LogText);
    if slen > 255
    then
    slen := 255;
    end if;
    while slen > 0 loop
    dbms_output.put_line(substr(le(ind).LogText,spos,slen));
    spos := spos + 255;
    slen := length(le(ind).LogText) + 1 - spos;
    end loop;
    ind := le.NEXT(ind);
    end loop;
    end if;
    end if;
    END SCHEMAS_EXPORT_TEST;

Maybe you are looking for

  • Error when trying to sign in on mobile phone

    Dear Admin, I want to browse the ORACLE ADF FORUMS in my mobile phone (nokia x2-01 with UC and Opera mini Browser), the forum page opens correctly but when i try to signin it gives me error. ORACLE ACCESS MANAGER ERROR: System error. Please re-try yo

  • In Business One, is the transaction time stored anywhere?

    Working with SAP Business One 2005 SP1 PL:29 The accounting team would like to know what time each transaction happened.  So far all I can see is the created date and update date on the transactions(OJDT) and in the Change Log (ADJT). Primarily this

  • What is the best way to add text to an image

    I used the draw ap div to add text to an image.  will that cause any problems?  If so, what is the best way to add text to an image?  Thanks!

  • Having multiple sessions open

    I find myself working in many clients at the same time, Development, QA, PRD, Staging and so on. With multiple sessions open in each cleint (upto 6 at a time) it can get difficult trying to figure out what window belongs to what client. In the task b

  • Why only +String concatation for Operator Overloading ?

    Java does not support operator overloading other than the + operator for addition of two Strings or a number and a String.Why only the + operator and Strings, even when we have a concat() for addition of Strings ? Wondering if anyone could provide me