Import Indexes only using datapump

Hi Gurus,
I have 2 identical schemas like same table structures but row number and data is different so I want to import the indexes from one schema to other. Can you please advise how can I do that using datapump.
Thanks in advance.

It doesn't matter what data is in the tables, as long as the columns the indexes are on exist. The indexes get created and built using a normal
create index ...
statement. So, if the original table had 100 rows and the new table has 10, the index will be built using the 10 rows. All that gets stored in the dumpfile is the ddl (in xml format). No index data is stored in the dumpfile. (If the op was using transportable, that would be a different story.)
Dean

Similar Messages

  • Importing indexes only

    Here is the scenario:
    Export A contains tables, data, no indexes
    Export B (same database, different machine, contains tables, no data, indexes.
    Is there a way to import indexes only from export B into Database A without including tables?

    Hi Germando
    You can let IMP create a file with the CREATE INDEX statements with the option INDEXFILE, e.g.:
    imp <user>/<pass> indexfile=<out.sql> file=<in.dmp> full=y
    Then you can run it with SQL*Plus.
    Chris

  • Check for import errors programmatically using datapump api

    Hi,
    I have pl/sql that imports some tables over DBLink using dbms_datapump. Is there a way to programmatically check that the import job had no errors at all, before proceeding to the next step after the import is completed?
    I saw several samples for showing the progress of the import job using while loops and whether it completed or not.
    But none explaining how to check whether any of the import steps failed, or if all steps completed successfully.
    Thanks,
    JGP

    check this
    [Datapump Demos|http://www.morganslibrary.org/reference/dbms_datapump.html]

  • INDEX failed to create while  Import using datapump

    Hi DBA's
    database 11.2.01
    OS : Solaris 10
    I am trying to Import the data using the following command.
    impdp system/******* SCHEMAS=EIMS DIRECTORY=expdp_dir DUMPFILE=eimsexp.dmp*
    but it fails import index and error out, but the tables and data imported.
    I checked in the metalink and followed the Note id (During Data Pump Import Index Creation Fails With: ORA-39083 and ORA-14102 errors OR ORA-39083 ORA-02158 and ORA-39112 errors [ID 1066635.1]) , but no luck.
    Kindly advice me.
    ERROR
    ORA-39083: Object type INDEX failed to create with error:
    ORA-14102: only one LOGGING or NOLOGGING clause may be specified
    Failing sql is:
    CREATE UNIQUE INDEX "EIMS"."PK_MDS_PSTN_LEVEL" ON "EIMS"."MDS_PSTN_LEVEL" ("REGION_CODE", "AREA_CODE", "LEVEL_NO", "EXCHANGE_CODE") PCTFREE 10 INITRANS 2 MAXTRANS 255 NOLOGGING NOCOMPRESS LOGGING STORAGE( INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1) TABLESPACE "EIMS_DATA" PARALLEL 1
    ORA-39083: Object type INDEX failed to create with error:
    ORA-14102: only one LOGGING or NOLOGGING clause may be specified
    Best Regards,
    SG

    Dear Hussein,
    Yes i applied the patch, still i am getting the same error and again i followed one more note (Impdp Failed With Error Ora-39083, 0ra-01403 On Index_statistics (Doc ID 755253.1))
    in that note i given option "exclude=index", then i got in different error.
    We are in the process of doing testing on Tablespace Encryption.
    We are having two instances one is dev and uat instance.
    we are creating Encryption tablespace on uat instance and trying to import the data from dev instance only from particular user.
    We completed the followings.
    1. Created tablespace Encryption.
    2. Created user and assigned the tablespace to the user EIMS1.
    3. Took export from the dev instance using the following command.
    expdp system/******** SCHEMAS=EIMS DIRECTORY=expdp_dir DUMPFILE=eimsexp.dmp
    4. Importing data from
    impdp system/******** REMAP_SCHEMA=EIMS:EIMS1 DIRECTORY=expdp_dir exclude=index DUMPFILE=eimsexp.dmp
    Error
    ====
    ORA-39083: Object type CONSTRAINT failed to create with error:
    ORA-14102: only one LOGGING or NOLOGGING clause may be specified
    Failing sql is:
    ALTER TABLE "EIMS1"."DOCS" ADD PRIMARY KEY ("ID") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 NOLOGGING NOCOMPRESS LOGGING STORAGE( INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1) TABLESPACE "EIMS_DATA" ENABLE
    ORA-39083: Object type CONSTRAINT failed to create with error:
    ORA-14102: only one LOGGING or NOLOGGING clause may be specified
    before i used to get the bellow errors
    ==========================
    ALTER TABLE "EIMS1"."TST1" ADD CONSTRAINT "ADMACHINEDEPLOYMENTSTATUS_PK" PRIMARY KEY ("MID") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 NOLOGGING NOCOMPRESS LOGGING STORAGE( INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1) TABLESPACE "EIMS_DATA" ENABLE
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39083: Object type INDEX_STATISTICS failed to create with error:
    ORA-01403: no data found
    ORA-01403: no data found
    Failing sql is:
    DECLARE I_N VARCHAR2(60); I_O VARCHAR2(60); c DBMS_METADATA.T_VAR_COLL; df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS'; BEGIN DELETE FROM "SYS"."IMPDP_STATS"; c(1) := 'ID'; DBMS_METADATA.GET_STAT_INDNAME('EIMS1','DOCS',c,1,i_o,i_n); INSERT INTO "SYS"."IMPDP_STATS" (type,version,flags,c1,c2,c3,c5,n1,n2,n3,n4,n5,n6,n7,n8,n9,n10,n11,n12,d1,cl1) VALUES ('I',5,0,I_N,NUL
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MDS_PSTN_LEVEL" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MD_AMT_REQUEST_PRIORITY" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MD_AMT_REQUEST_QUEUE" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MD_BH02_REQUEST_PRIORITY" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MD_BH03_REQUEST_PRIORITY" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MD_BH03_REQUEST_QUEUE" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"EIMS1"."PK_MD_BH04_REQUEST_PRIORITY" creation failed
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "SYSTEM"."SYS_IMPORT_FULL_01" completed with 1049 error(s) at 20:43:02
    Thanks
    SG

  • Only import constraint from the full backup dump using datapump !

    Dear Firends ,
    I am using Oracle 10g in my production server . I firstly export the full database backup using expdp . Now I want to import just the constraints of my production server to a test server . Is it possible to import constraints only to a test server from a full dump backup ?
    plz inform ... ...

    Yet another 'I refuse to read the documentation, unless some volunteer spoon feeds me' question.
    What happened to the DBA community. Is there some special virus spreading all over the globe?
    Or is the virus called 'OTN' and is OTN making DBAs permanently lazy, trying constantly how to do a little as possible for their money.
    You can easily construct the answer yourself from this overview.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_import.htm#sthref389
    And please re-read the Forums Etiquette post again, or read it for the first time, stating you need to peruse the documentation prior to asking a question.
    Your assistance in diminishing the boot load of RTFM questions is appreciated.
    Sybrand Bakker
    Senior Oracle DBA

  • Importing the METADATA ONLY using DBMS_DATAPUMP

    Hi DBAs,
    Using the DBMS_DATAPUMP , how can I import the metadata only for a particular table . Also I dont want to import any associated INDEXES and TRIGGERS. I have the following codes but it is trying to import every thing. Also if the table exist, it is not importing the METADATA rather erroring out.
    handle1 := DBMS_DATAPUMP.OPEN('IMPORT','SCHEMA', 'QAXDB.WORLD');
    DBMS_DATAPUMP.METADATA_FILTER(handle1, 'SCHEMA_EXPR', 'IN (''HR'')');
    DBMS_DATAPUMP.SET_PARAMETER(handle1, 'INCLUDE_METADATA', 1);
    Thanks
    -Samar-

    See the below link,
    http://download.oracle.com/docs/cd/B19306_01/appdev.102/b14258/d_datpmp.htm
    Hope this helps,
    Regards,
    http://www.oracleracexpert.com
    Click here for [Cross platform Transportable tablespace using Datapump|http://www.oracleracexpert.com/2009/08/transportable-tablespace-export-import.html]
    Click here to learn [Oracle data pump export/import with examples.|http://www.oracleracexpert.com/2009/08/oracle-data-pump-exportimport.html]

  • Error in import table data using oracle datapump

    i am trying to import table data using oracle datapump
    CREATE TABLE emp_xt (
    ID NUMBER,
    NAME VARCHAR2(30)
    ORGANIZATION EXTERNAL (
    TYPE ORACLE_DATAPUMP
    DEFAULT DIRECTORY backup
    LOCATION ('a.dmp')
    it return the following error
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04084: The ORACLE_DATAPUMP access driver does not support the ROWID column.
    ORA-06512: at "SYS.ORACLE_DATAPUMP", line 19
    please help me

    .dmp file generated from exp command file not from oracle_datapump

  • Stats Not getting imported while importing partition using DATAPUMP API

    Hi,
    I am using datapump API for exporting a partition of a table and the importing the partition into table on another DB.
    The data is getting imported but i could the partition stats are not getting imported. I am using TABLE_EXISTS_ACTION = APPEND.
    any idea on how to achieve the stats also to be imported into the partition.
    Thanks!

    Hi,
    I am using datapump API for exporting a partition of a table and the importing the partition into table on another DB.
    The data is getting imported but i could the partition stats are not getting imported. I am using TABLE_EXISTS_ACTION = APPEND.
    any idea on how to achieve the stats also to be imported into the partition.
    Thanks!

  • Can i import indexes and constraints only?

    i exported all the objects from one schema and now i want to import this file in two steps. first i want to import tables and data into a particular tablespace. after importing this i'll change the default tablespcae of the user and then want to import only indexes and constraints.
    i tried but nothing happen... is this possible?

    Thanks for your reply but i want to create a script to load all the objects once, but split into diff tablespaces..
    i have 40GB data and i don't want to move objects manually from one tablespace to another tablespace....
    the idea was something like that....
    1) set a default tablespace to import tables .
    2) then change the default tablespace to import indexes...
    but it doesn't work....
    if there is no way then i have to do something same you suggested me..
    Thanks

  • Total number of records of partition tables exported using datapump

    Hi All,
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    OS: RHEL
    I exported a table with partitions using datapump and I would like to verify the total number of all the records exported of all these partition tables.The export has a query on it with WHERE ITEMDATE< TO_DATE (1-JAN-2010). I need it to compare with the exact amount of records in the actual table if it's the same.
    Below is the log file of the exported table. It does not show the total number of rows exported but only individually via partition.
    Starting "SYS"."SYS_EXPORT_TABLE_05": '/******** AS SYSDBA' dumpfile=data_pump_dir:GSDBA_APPROVED_TL.dmp nologfile=y tables=GSDBA.APPROVED_TL query=GSDBA.APPROVED_TL:"
    WHERE ITEMDATE< TO_DATE(\'1-JAN-2010\'\)"
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 517.6 MB
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q3" 35.02 MB 1361311 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q4" 33.23 MB 1292051 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q4" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q1" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q3" 30.53 MB 1186974 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q3" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q1" 30.44 MB 1183811 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q2" 30.29 MB 1177468 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q4" 30.09 MB 1170470 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q2" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q2" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q1" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q3" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q4" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q1" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2006_Q3" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2006_Q4" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q1" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q2" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q3" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q4" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q1" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q2" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q2" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q3" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q4" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_MAXVALUE" 0 KB 0 rows
    Master table "SYS"."SYS_EXPORT_TABLE_05" successfully loaded/unloaded
    Dump file set for SYS.SYS_EXPORT_TABLE_05 is:
    /u01/export/GSDBA_APPROVED_TL.dmp
    Job "SYS"."SYS_EXPORT_TABLE_05" successfully completed at 12:00:36
    Edited by: 831134 on Jan 25, 2012 6:42 AM
    Edited by: 831134 on Jan 25, 2012 6:43 AM
    Edited by: 831134 on Jan 25, 2012 6:43 AM

    I assume you want this so you can run a script to check the count? If not and this is being done manually, then just add up the individual rows. I'm not very good at writing scripts, but I would think that someone here could come up with a script that would sum up the row count for the partitions of the tables from your log file. This is not something that Data Pump writes to the log file.
    Dean

  • Importing a table and its index statistics, cannot import index stats

    Hi,
    Oracle 10.2.0.4 on Solaris.
    I have DBA to imprt tables and index statistics for a table from prod into QA for further analysis. The stats for this table are locked in prod
    DBA has used the following command for export and import table statistics
    exec dbms_stats.export_table_stats('SCHEMA','TABLE',null,'TABLE_20130225',null,true,'INV');
    exec dbms_stats.import_table_stats('SCHEMA','TABLE',null,'TABLE_20130225',null,true,'SCHEMA');Although cascade is set to true above, this resulted in "only table stats" to be imported no index stats. So we have imported the prod table level stats but no index stats! FYI, the indexes in prod have stats (last_analyzed) set.
    Next DBA tried the export and import using export_index_stats and import_index_stats but no luck. DBA advising me that the only option we have is to import the table itself from prod to QA. It seems that import with cascade does not work.
    Is this a bug in 10g or there is another way around to get index statistics as well?
    Thanks
    Edited by: 902986 on 25-Feb-2013 06:22

    902986 wrote:
    Hi,
    Oracle 10.2.0.4 on Solaris.
    I have DBA to imprt tables and index statistics for a table from prod into QA for further analysis. The stats for this table are locked in prod
    DBA has used the following command for export and import table statistics
    exec dbms_stats.export_table_stats('SCHEMA','TABLE',null,'TABLE_20130225',null,true,'INV');
    exec dbms_stats.import_table_stats('SCHEMA','TABLE',null,'TABLE_20130225',null,true,'SCHEMA');Although cascade is set to true above, this resulted in "only table stats" to be imported no index stats. Problem Exists Between Keyboard And Chair
    "Gather statistics on the indexes for this table. Index statistics gathering is not parallelized. Using this option is equivalent to running the GATHER_INDEX_STATS Procedure on each of the table's indexes. Use the constant DBMS_STATS.AUTO_CASCADE to have Oracle determine whether index statistics to be collected or not. This is the default. The default value can be changed using theSET_PARAM Procedure."
    Handle:     902986
    Status Level:     Newbie
    Registered:     Dec 17, 2011
    Total Posts:     69
    Total Questions:     18 (12 unresolved)
    Why so MANY unanswered questions?

  • IMPDP INDEX only

    Hi
    i have schema level export DUMP backup, and i need to to just import the INDEX only as All tables are already imported.
    so is there any command to do this plz

    refer thread:Re: IMPORT using impdp for indexes

  • Index not used on view when table stats exist

    Hello,
    I would be grateful if someone comes with ideas on the following problem I'm currently facing.
    I have a table with XMLTYPE data type column:
    sql-->desc ACFBNK_STMT008
    RECID     NOT NULL     VARCHAR2(200)
    XMLRECORD XMLTYPE
    I have a view V_ACFBNK_STMT008 on that table, in which the view columns are defined as extracted tags values from the XMLTYPE field, e.g. for the view field N_BOOKING_DATE:
    numcast(extractValue(xmlrecord,'/row/c25')) "N_BOOKING_DATE"
    (note: numcast is just a simple function that returns TO_NUMBER of its input argument)
    I have also a function-based index on this field of the table:
    CREATE INDEX train4.NIX_ACFBNK_STMT008_C25
    ON train4.ACFBNK_STMT008("TRAIN4"."NUMCAST"(extractValue(xmlrecord,'/row/c25')))
    And so, I'm executing on the view the following SQL statement:
    SELECT RECID FROM V_ACFBNK_STMT008 WHERE (N_BOOKING_DATE > TO_NUMBER('20070725'));
    Now, the problem comes: when statistics exist on the view base table (that is ACFBNK_STMT008) then the above statement is not using the index and is making a "table access full". When I delete the statistics for the table then the SQL runs fast with an "index range scan".
    Which is further strange - when I change the ">" operand with a "=" the SQL statement correctly captures the index regardless of whether or not statistics exist.
    I've tried to manually rewrite the SQL and include the "numcast" function in it:
    SELECT RECID FROM TRAIN4.V_ACFBNK_STMT008 WHERE ( N_BOOKING_DATE>train4.numcast(TO_NUMBER( '20010725' ) ));
    And in this way the index is used OK even with statistics existing!
    But regretfully I don't have a way to change the application and the SQL, so the only things I can change is the view and/or the index.
    Thank you in advance,
    Evgeni
    P.S.
    I've tried gathering statistics in both the following ways but still the problem persists:
    sql-->analyze table train4.ACFBNK_STMT008 compute statistics;
    sql-->exec dbms_stats.gather_table_stats(ownname=>'TRAIN4', tabname=>'ACFBNK_STMT008', CASCADE=>TRUE, partname=>NULL);

    Oh, and I forgot to mention: I cannot change the view definition as well (for example, to remove the "numcast"), since every now and then the application would recreate it automatically with the same code. :(

  • Finder says Home Disk  is full, utilities say I've only used 93Gb of 232Gb

    Earlier this evening I all of a sudden got an alert that my disk was full, 232Gb. 0Gb remaining.
    It said I needed to renove some files. I tried. I removed about 10 Gb, but it didn't give me any more space.
    I booted from my Clone to to see if I could run dik utility repair. It came up saying the disk was fine.
    Repaired Permissions. Everything came up OK there too.
    I ran Disk Warrior from it;s disk and it came up stating that I was only using 93Gb of 232Gb. (Which is what I remembered and was about how much my recent Carbon Copy Clone added up to)
    I tried WhatSize and it said the same. I'm currently waiting for a license for Tager Cache Cleaner to see if it will help. None of the utilities is finding any hidden large files.
    I've also tried backing up the mail folder and removing Envelope Index, but, of course, that made Mail crash since it's being told there's no room on the hard drive. I'm totally stumped.
    I imagine there's some corrupt file somewhere, but I'm not getting a fix on it.
    Recent changes to the drive - I upgraded to 10.4.9 a few ays ago. Had to install a third party driver for my HP PSC 2175 - hpijs-foomatic-2.0.2.
    Printer stopped working earlier today and I had to reinstall CUPS and then the driver again before printer would work. Sortly thereafter that I got the error message about disk full.
    System info:
    Machine Name: Power Mac G5
    Number Of CPUs: 1
    CPU Speed: 1.6 GHz
    L2 Cache (per CPU): 512 KB
    Memory: 2 GB
    Bus Speed: 800 MHz
    Boot ROM Version: 5.1.5f2
    I have several different firewire drives attached as well as a USB "jump" drive.
    Any advice would be most appreciated.
    G5 1.6Gz   Mac OS X (10.4.8)  

    I just found this discussion at the MacFixIt forums. You might read it carefully, and also take a look at the MacOSXHints article mentioned:
    http://www.macosxhints.com/article.php?story=20070319143243249
    I have a batch of those job history files as well so thought I would try removing them and see what happened. I first got myself a root shell:
    NoobiX:~ francine$ sudo -s
    Hit return and entered my password, hit return again, then changed directory:
    NoobiX:~ root# cd /private/var/spool/cups
    Hit return and ran a list command:
    NoobiX:/private/var/spool/cups root# ls -al
    total 304
    drwx--x--- 41 root lp 1394 Mar 26 21:59 .
    drwxr-xr-x 7 root wheel 238 Aug 3 2006 ..
    -rw------- 1 root lp 3571 Aug 24 2006 c00001
    -rw------- 1 root lp 3574 Aug 24 2006 c00002
    -rw------- 1 root lp 3505 Dec 8 00:24 c00003
    ....and so on.... down to
    -rw------- 1 root lp 3531 Mar 26 21:59 c00038
    drwxrwx--T 2 root lp 68 Dec 21 22:37 tmp
    Which verified I was in the right place. I tried removing just one of the files:
    NoobiX:/private/var/spool/cups root# rm c00001
    which worked fine. I then decided doing this would be tedious, so tried this:
    NoobiX:/private/var/spool/cups root# rm -r *
    There is a space between the rm and -r and after -r before the * (a wildcard). That was probably a mistake since it not only nuked the history files but also the tmp directory. I should have probably used something like rm -r "c*" (since all the history files begin with c that should have got them and spared the tmp folder). I tried recreating the tmp folder with this:
    NoobiX:/private/var/spool/cups root# mkdir tmp
    Which seemed to work:
    NoobiX:/private/var/spool/cups root# ls -al
    total 0
    drwx--x--- 3 root lp 102 Jun 7 12:00 .
    drwxr-xr-x 7 root wheel 238 Aug 3 2006 ..
    drwxr-xr-x 2 root lp 68 Jun 7 12:00 tmp
    But being a suspicious sort I ran Disk Utility to Repair Permissions and it did indeed report an error on the /private/var/spool/cups/tmp folder, but it did repair the permissions on the folder successfully, so I'm assuming all is now OK. Of course, I haven't actually tried printing anything yet...
    If you choose to try this after you finish type exit and hit return to kill the root shell, then quit Terminal.
    Francine
    Francine
    Schwieder

  • How can i detach the video from a clip so i can only use the audio when using imovie?

    how can i detach the video from a clip so i can only use the audio when using imovie?

    First, go to iMovie/Preferences and enable the Advanced Tools.
    Import the video clip into an iMovie Event. When you get ready to use it in an iMovie Project, select the clip in the Event and drag & drop it on top of a video or photo clip in the project at the place where you want it to begin. When the pop-up menu appears, choose Audio Only.

Maybe you are looking for