DB Import is very slow on ESX

Hi All,
We are performing OS/DB migration from Sun/Oracle to Win/SQL server. Export took 9 hours when we perform import it is taking more than 25 hours and we are using ESX ( VMWARE), import system configuration is
OS : Windows 2003
DB : SQL Server
8 CPUS
32 GB RAM
1.5 TB HD ( for databases)
500GB for log files 1
500GB for log file 2
Importing is taking very slow and we are split export and import method.  Please help me reason for the slow performance and let know any specific configuration we have to use for sap on vmware specific to SQL Server DB.
Regards,
Kanthi.

> - How big is your queued disk I/O? (use performance monitor to check)
This number indicates whether you have an I/O problem (using too many R3loads in parallel). If you want to trace down performance problems you should get familiar with the Windows tools, you may need that later during production also.
> - How many parallel processes are you running?
>            
>                       20
And you make sure using migration monitor that no packages of the same table are loaded simultaneously if you have split the talbles?
> Network team is complaining about some ESX issues are experiencing, is this effect to the performance of the DB-Import.
I can't say since I don't know if you load the data from a network drive or locally or if the system loading the data into the database is local or remote.
I suggest you open an OSS call, they may have a look on the system and may give you some idea. Just from a description without having actual I/O key numbers it's impossible to give any hint where to look at to speed things up.
Markus

Similar Messages

  • Import is very slow...Ki9ndly suggest

    friend...
    When I am running the import statement with 5gb of dmp file..It is running very slow even thou i have given the buffer as 5000000..
    kindly suggest where i am going wrong...
    it very urgent...
    Srini

    Here is some thing from the Oracle Utilities guide.
    Processing Restrictions
    The following restrictions apply when you process data with the Export and Import utilities:
    Java classes, resources, and procedures that are created using Enterprise JavaBeans (EJB) are not placed in the export file.
    Constraints that have been altered using the RELY keyword lose the RELY attribute when they are exported.
    When a type definition has evolved and then data referencing that evolved type is exported, the type definition on the import system must have evolved in the same manner.
    The table compression attribute of tables and partitions is preserved during export and import. However, the import process does not use the direct path API, hence the data will not be stored in the compressed format when imported.

  • Importing music very slow 0.5x and iTunes freezes during every backup

    Everytime I plug my iPod into my computer it freezes and I need to end app also while importing new CDs they are taking about 2-3hrs and it's only transferring at 0.5x. My pc is a windows xp sp3 2.68GHz 760 RAM

    I have had some help and have tried many things to resolve this. Bottom line, it was the CD reader. I tried to load the same "slow loading" CD into someone else's computer and it went in faster than I've ever seen (9.5X - 13X). I borrowed his CD reader and put in in my machine and it worked fine for the 5 CD of the 100 I had left to load because they loaded so slow. This time, there was one of these 5 that read in at 1.1X. Again, I borrowed a different CD reader, put it in my machine, and it read fine. This problem was NOT program/system related: I tried cleaning files, running Adware, Antivirus, defrag, etc.) I read/burn lots of CDs and must wear them out in my laptop.

  • Import goes VERY SLOW

    On Windows Vista 64. Based on a quick read of other subjects, this problem may be due to using the new (8.1?) version of iTunes. Was importing some CD's last night (3/31) and the first would go just fine. Then, pop in a 2nd CD and it would burn about 1 song at normal speed (say, 10-14X), but then would slow down dramatically to 0.2X or something (basically, not working). I don't think its the CD R/W drive (as it works for other things, and Vista says the device is functioning properly). Reboot, and exact same thing happens (1st CD just fine, then it craps out). Anybody else have this happen? Anyone have any ideas? Thanks!
    Message was edited by: avest

    wow - just saw a thread on page 2 w/ the exact same problem - sorry for the repeat!

  • Import running Very Slow!!!

    Hi Guys,
    This is in sequence with my first two posts...i'lll explain the whole scenario....
    I have a 44 GB table on an Oracle DB 9.2.0.8. I first tried to truncate it but it was taking huge amount of time...
    So, i truncated the table with:-
    SQL> Truncate table TABLENB REUSE STORAGE;
    After this i tried to Drop the table and it took approx 16-18 hrs to drop it. Now, i have started a import of the table using following command:-
    imp schema/passwd
    file=/oracle/XX/97/imp/TABLENB.1.exp,/oracle/XX/98/imp/TABLENB.2.exp,/oracle/XX/99/imp/TABLENB.3.exp log=/oracle/XX/97/exp/TABLENB.imp.log tables=TABLENB feedback=10000 indexes=N ignore=Y
    It has taken 16 hrs till now and when i run the following query the output is as follows:-
    SQL> select extents from dba_segments where segment_name='TABLENB';
    EXTENTS
    12450
    Initially the table had about 1 million extents....
    Can somebody tell, is dere any way i can speed up the import process or is dere any lock on the table which is making the import slow????
    Regards,
    Nick.
    Message was edited by:
    Nick-- wud b DBA

    Yeah Justin, u've recalled correctly.........
    Unfortunately, as per the Pre-plan (Assigned to us by
    someone else), we are importing back into the
    Dictionary Managed Tablespace....... :-(I'd suggest talking to that "someone else" and figuring out why this is being done. Dropping a 1 million extent table from a DMT in 9.2 only to import that same table back into a DMT without even fixing the underlying extent allocation issue is a rather unfortunate plan. At some point, the right answer has to be to fix the plan rather than continuing to slog forward with a poorly conceived plan.
    Do u have some suggestion for us so that the import
    can be made faster if possible... or simply we have
    to sit and wait......Quoting from my earlier reply-- "It would almost certainly be more efficient to pre-create the table with more appropriate extent sizes and then to import the data into that existing table"
    Justin

  • Import is very Slow with xmltype column in oracle 9.2.0.7

    Hi
    We are using oracle 9.2.0.7 on Solaris
    The import without XMLTYPE column is faster the import of the table with XMLTYPE column.Also when we do the export in direct path it is still going for Conventional(I suppose this is correct for XMLTYPE).
    Is this the general behaviour of XMLTYPE...if so how to improve the performance of the import.
    Can anybody help us out in resolving the issue.

    here is the output
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    CREATE TABLE "MMSCBMC"."PAYMENTS_AUDIT"
    ( "PMTID" NUMBER NOT NULL ENABLE,
    "PMTSTATUS" VARCHAR2(32) NOT NULL ENABLE,
    "PMTSTATUS_CODE" VARCHAR2(10) NOT NULL ENABLE,
    "CSREFID" VARCHAR2(32),
    "CURAMT" NUMBER(15,2) NOT NULL ENABLE,
    "CUSTOMER_KEY" NUMBER NOT NULL ENABLE,
    "ACCTID" VARCHAR2(32) NOT NULL ENABLE,
    "BANKREFID" VARCHAR2(34) NOT NULL ENABLE,
    "PAYEEID" NUMBER,
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    "PAYMENTS_AUDIT_KEY" NUMBER NOT NULL ENABLE,
    "PROCDATE" DATE NOT NULL ENABLE,
    "DUEDATE" DATE NOT NULL ENABLE,
    "IMMEDIATE_GENERATION" NUMBER DEFAULT 0 NOT NULL ENABLE,
    "RECURRINGPMTID" NUMBER,
    "BATCHID" NUMBER,
    "REMIT_INFO" VARCHAR2(200),
    "BILLING_ACCT" VARCHAR2(32),
    "PAYMENT_METHOD" VARCHAR2(20) NOT NULL ENABLE,
    "IS_ACCTNUMBER" NUMBER,
    "SPREFID" VARCHAR2(36),
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    "FIDEBITTRCNUM" VARCHAR2(32),
    "FEE" NUMBER(12,2),
    "FEE_TYPE" VARCHAR2(10) DEFAULT NULL ,
    "DATE_ADDED" DATE NOT NULL ENABLE,
    "CLIENTIP" VARCHAR2(32) NOT NULL ENABLE,
    "LAST_UPDATED" DATE NOT NULL ENABLE,
    "ACTION" VARCHAR2(32),
    "MODIFIEDBY" VARCHAR2(32) NOT NULL ENABLE,
    "ISCSR" VARCHAR2(1) NOT NULL ENABLE,
    "SUBUSER_KEY" NUMBER,
    "TOKEN" NUMBER DEFAULT 0 NOT NULL ENABLE,
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    "FUNDINGSTATUS_CODE" NUMBER DEFAULT 0 NOT NULL ENABLE,
    "FUNDINGSTATUS" VARCHAR2(32) DEFAULT 'Funds Pending' NOT NULL ENABLE,
    "PRODUCT_CODE" VARCHAR2(10),
    "SUB_PRODUCT_CODE" VARCHAR2(10),
    "ACCT_TYPE" VARCHAR2(3),
    "DEST_PRODUCT_CODE" VARCHAR2(10),
    "DEST_SUB_PRODUCT_CODE" VARCHAR2(10),
    "DEST_BANKREFID" VARCHAR2(34),
    "DEST_ACCT_TYPE" VARCHAR2(3),
    "ISDEBIT" NUMBER NOT NULL ENABLE,
    "PAYMENTFROM_NICKNAME" VARCHAR2(96),
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    "PAYMENTFROM_ACCTID" VARCHAR2(32),
    "PAYMENTTO_NICKNAME" VARCHAR2(96),
    "PAYMENTTO_ACCTID" VARCHAR2(32),
    "DEST_ISBUSINESSACCT" NUMBER,
    "APPROVE_DATE" DATE,
    "APPROVE_KEY" NUMBER,
    "LAST_MODIFIER_KEY" NUMBER,
    "CHECK_DESC" VARCHAR2(200),
    "FAILURE_REASON" VARCHAR2(250),
    "PMTFUNDRETRY_COUNT" NUMBER DEFAULT 0 ,
    "MQ_TRANSACTIONID" NUMBER,
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    "FREQUENCY" NUMBER,
    "NUMBEROFREMPAYMENTS" NUMBER,
    "OPENENDED" NUMBER,
    "ORIGINATOR_KEY" NUMBER NOT NULL ENABLE,
    "EXTERNALREFERENCE" "SYS"."XMLTYPE" ,
    "EXTERNALINFO" "SYS"."XMLTYPE" ,
    "PRINCIPAL_AMT" NUMBER(15,2),
    "ESCROW_AMT" NUMBER(15,2),
    "LATEFEE_AMT" NUMBER(15,2),
    "OTHERFEE_AMT" NUMBER(15,2),
    "REGULARPMT_AMT" NUMBER(15,2),
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    CONSTRAINT "PK_PAYMENTSADT_KEY" PRIMARY KEY ("LAST_UPDATED", "PAYMENTS_AUDIT_K
    EY")
    USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "MMS_DATA_TBS" ENABLE,
    CONSTRAINT "FK01_PAYMENTS_AUDIT" FOREIGN KEY ("APPROVE_KEY")
    REFERENCES "MMSCBMC"."CUSTOMER" ("CUSTOMER_KEY") ENABLE,
    CONSTRAINT "FK02_PAYMENTS_AUDIT" FOREIGN KEY ("LAST_MODIFIER_KEY")
    REFERENCES "MMSCBMC"."CUSTOMER" ("CUSTOMER_KEY") ENABLE,
    CONSTRAINT "FK06_PAYMENTS_AUDIT" FOREIGN KEY ("ORIGINATOR_KEY")
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    REFERENCES "MMSCBMC"."ORIGINATOR" ("ORIGINATOR_KEY") ENABLE
    ) PCTFREE 2 PCTUSED 40 INITRANS 1 MAXTRANS 255 LOGGING
    STORAGE(
    BUFFER_POOL DEFAULT)
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS"
    PARTITION BY RANGE ("LAST_UPDATED")
    (PARTITION "M1_2005" VALUES LESS THAN (TO_DATE(' 2005-02-01 00:00:00', 'SYYYY-
    MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
    PCTFREE 2 PCTUSED 40 INITRANS 1 MAXTRANS 255
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS"
    LOB ("SYS_NC00059$") STORE AS (
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSI
    ON 10
    NOCACHE
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT))
    LOB ("SYS_NC00061$") STORE AS (
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSI
    ON 10
    NOCACHE
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)) NOCOMPRESS ,
    PARTITION "M2_2005" VALUES LESS THAN (TO_DATE(' 2005-03-01 00:00:00', 'SYYYY-M
    M-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
    PCTFREE 2 PCTUSED 40 INITRANS 1 MAXTRANS 255
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS"
    LOB ("SYS_NC00059$") STORE AS (
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSI
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    ON 10
    NOCACHE
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT))
    LOB ("SYS_NC00061$") STORE AS (
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSI
    ON 10
    NOCACHE
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)) NOCOMPRESS ,
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    PARTITION "MAX_VALUE" VALUES LESS THAN (MAXVALUE)
    PCTFREE 2 PCTUSED 40 INITRANS 1 MAXTRANS 255
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS"
    LOB ("SYS_NC00059$") STORE AS (
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSI
    ON 10
    NOCACHE
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT))
    DBMS_METADATA.GET_DDL('TABLE','PAYMENTS_AUDIT')
    LOB ("SYS_NC00061$") STORE AS (
    TABLESPACE "PAYMENTS_AUDIT_DATA_TBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSI
    ON 10
    NOCACHE
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)) NOCOMPRESS )

  • Lightroom 5.5 Import from Catalog *very* slow performance.

    Importing from a second catalog from a shoot is *very* slow with the last version or couple of versions of Lightroom.  It's been about 20 minutes and the import is moving glacially slow and is only about 20% done.
    Here are as many stats as I have:
    iMac 27" 3.5Ghz Core i7, 32G of ram, 3T fusion drive (the late 2013 model maxed out)
    Master catalog has 117,000 images in it.
    The catalog I'm importing has 948 images in it.
    CPU use during the import is 100% used by lightroom (note that the multiple CPUs can go to more than 100% so this means I think that one CPU is maxed out, though LR can sometimes take 500% of my CPUs during an export)
    Lightroom is using 4.2G of RAM (10G free in activity monitor)
    Hard drive is fairly full, but still has 366G free on it (so about 10%)
    Lightroom Mobile sync is enabled (not on this folder I don't think) but says "service unavailable" at this point.
    Lightroom is fully up to date (5.5) on both iMac and laptop (also a mac in case that matters).  Both are up to date with the latest OS software.
    How I did it was:
    Export the shoot as a new catalog on my iMac
    Transfer it to my laptop (smart previews only, no masters)
    Edit images, metadata, etc
    Transfer the entire folder back (.lrcat, previews, smart previews) to iMac
    In my master catalog do an import from catalog
    Waited for ages for checking for dupes (which I can understand with such a big catalog)
    Selected to import all, changing metadata and develop edits, and selecting to put new images in the folder (not sure why it asked as there weren't any new images)
    The last time I did an import I started it and then went out somewhere so I didn't notice how long it took, but this seems completely crazy.  In the time it's taken to write this up it's now about 20 minutes later and the progress bar is maybe at the 33% range.  It's still working away, it's just crazy crazy slow.
    Anything to help out or debug would be greatly appreciated.  I suspect it's pure size of the catalog, but 120k isn't outside the realm of what I've heard people say works fine, or maybe the Fusion drive (maybe the OS is trying to move files around behind the scenes)? 
    Either way this didn't seem to be an issue until the last couple of versions (I haven't gone and re-installed 5.4 or 5.3 to check) and it's starting to really frustrate me

    Hi Jim,
    Great suggestion! It opens perfectly fine on its own. I even went as far as re-exporting it once it loaded but am still experiencing the same issue. That same catalog loads perfectly fine on my assistant's iMac.
    I can't really think about anything that changed in my environment with the exception of a new Bootcamp partition to run windows on an external Thunderbolt HD but I doubt this is related.
    I'm open to any suggestions while I am trying this again... 10 minutes of 'Checking for changed and duplicate photos'.

  • Import ABAP running very slow

    guys, I am installing IDES EHP4 on a solaris box with oracle 10g. The import is running very slow. in 20 hours, it has only imported 10 out of 105 jobs. Top comand shows i have 26 GB free physical memory (out of 48GB total) and also 123 GB free swap out of 150 GB total. CPU load is also only 3 %. I have already configured the project file acording to sap recommedation. MY oracle init file configuration looks like this. Also, you will see my project and user_attr files below. Please advise what i can change tomake things go faster.
    # (c)Copyright SAP AG, Walldorf # # @(#) $Id: //bc/701-1_REL/src/ins/SAPINST/impl/tpls/ora/ind/INITSIDBASE.ORA#1 $ ########################################################### ########################################################### # PART I, TUNING PARAMETERS # ########################################################### #### SYSTEM GLOBAL AREA BUFFERS # Total System Global Area = sum(bytes) of all caches # Variable Size = shared_pool_size (approx.) # Database Buffers = db_block_buffers * db_block_size # Redo Buffers = log_buffer # # unit of shared_pool_size: bytes # unit of db_block_buffers: number of cached blocks # unit of log_buffer: bytes shared_pool_size = 2264924160 # 10% of shared_pool_size shared_pool_reserved_size = 226492416 log_buffer = 1048576 #### MANAGING SORTS # A small sort requires sort_area_retained_size in # memory. Larger sorts are allocating segments in # PSAPTEMP. They are using sort_area_size in memory. sort_area_retained_size = 0 sort_area_size = 2097152 #sort_spacemap_size = 512 #### PGA-PARAMETERS open_cursors = 800 #cursor_space_for_time = true #### LOCKS # dml_locks = processes * 50 dml_locks = 4000 #enqueue_resources = 8000 #### CPU-PARAMETERS #spin_count = 200 #_cpu_count = 4 #### TUNING ARCHIVING #log_archive_buffer_size 127 #log_archive_buffers 4 #### USE VECTOR READ #use_readv = true #### OTHER TUNING PARAMETERS #sequence_cache_entries = 100 #sequence_cache_hash_buckets = 89 ########################################################### # PART II, CHANGEABLE PARAMETERS # ########################################################### #### CONTROL-FILES # the controlfiles should be mirrored to every new # database mount point @SAPDATA_HOME@/sapdata<#> #control_files = XXXXX #### AUTOMATICLY STARTED BACKGROUND PROCESSES #### MANAGING LOG SWITCHES # checkpoints occur only when switching logs, if # log_checkpoint_interval * size of OS blocks > size of # the actual redo log log_checkpoint_interval = 0 #log_checkpoint_timeout = 0 #### PROCESS-PARAMETERS # The number of allocated semaphores is equal to the # number of processes # sessions = 1.2 * processes processes = 500 sessions = 600 #### AUDITING AND STATISTICS # sql_trace=TRUE # audit_trail = true # db_block_lru_extended_statistics = 1000 # db_block_lru_statistics = true ########################################################### # PART III, STATIC PARAMETERS # ########################################################### #### DB-NAME db_name = XXX #### DB-BLOCKSIZE db_block_size = 8192 #### DB-FILES db_files = 254 #### OPTIMIZER MODE #optimizer_mode = choose #optimizer_search_limit = 3 #### PATHS / DESTINATIONS / TRACES # /xxx\ : trace files of the background # processes # xxxx /saptrace/usertrace: trace files of the user processes # log_archive_dest is a destination, not a path. # The archivefiles get the name # XXXXXXX background_dump_dest = xxxx user_dump_dest = xxxx core_dump_dest = /xxxxx log_archive_dest = xxxxxx #log_archive_format = %t_%s #### OTHERx # reduce_alarm only supported on HP #reduce_alarm = TRUE #### ORACLE OPS PARAMETER remote_os_authent = true # retention time for RMAN backup information in control file control_file_record_keep_time = 30 #see SAP note 124361 #db_file_multiblock_read_count = 8 log_checkpoints_to_alert = true #### OPTIMIZER MODE #optimizer_index_cost_adj = 10 #### AUDITING AND STATISTICS #timed_statistics = true compatible = 10.2.0 parallel_execution_message_size = 16384 query_rewrite_enabled = false replication_dependency_tracking = false star_transformation_enabled = true undo_retention = 43200 undo_tablespace = PSAPUNDO undo_management = AUTO pga_aggregate_target = 3019898880 sga_max_size = 84529848320 db_cache_size = 4264924160 job_queue_processes = 1 remote_login_passwordfile = exclusive recyclebin = off event = "10191 trace name context forever, level 1" FILESYSTEMIO_OPTIONS = setall OPTIMPEEK_USER_BINDS = FALSE BTREE_BITMAP_PLANS = FALSE INDEXJOIN_ENABLED = FALSE INMEMORY_UNDO = FALSE OPTIMIZERMJC_ENABLED = FALSE SORTELIMINATION_COST_RATIO = 10 TABLELOOKUP_PREFETCH_SIZE = 0
    USER_ATTR FILE --- I have replaced my sid with SID for privacy reasons ************************************
    adm::::profiles=Log Management lp::::profiles=Printer Management postgres::::type=role;profiles=Postgres Administration,All root::::project=MYSID;auths=solaris.*,solaris.grant;profiles=Web Console Management,All;lock_after_retries=no;min_label=admin_low;clearance=admin_high SIDADM::::project=SID oraSID::::project=SID
    ************************************** PROJECT FILE --- my sid replaced with 'SID'*********************************
    system:0:::: user.root:1::::process.max-file-descriptor=(basic,65536,deny),(priv,65536,deny);process.max-sem-nsems=(priv,512,deny);project.max-sem-ids=(priv,512,deny);project.max-shm-ids=(priv,256,deny);project.max-shm-memory=(priv,18446744073709551615,deny) noproject:2:::: default:3::::process.max-file-descriptor=(basic,65536,deny),(priv,65536,deny);process.max-sem-nsems=(priv,512,deny);project.max-sem-ids=(priv,512,deny);project.max-shm-ids=(priv,256,deny);project.max-shm-memory=(priv,18446744073709551615,deny) group.staff:10:::: SID:400:SAP System SID:SIDadm,root,oraDBSID::process.max-file-descriptor=(basic,65536,deny),(priv,65536,deny);process.max-sem-nsems=(priv,512,deny);project.max-sem-ids=(priv,512,deny);project.max-shm-ids=(priv,256,deny);project.max-shm-memory=(priv,18446744073709551615,deny) **************************************************************************************************************************************
    what can i do to make the import faster ? my system definitely has enough resources but i am not able to utilizethem all. Thanks in advance

    You posted this in two different forums and it is not allowed.

  • Very Slow preview rendering for lightroom 5 on import

    I am seeing very very slow rendering of preview images only when using the import function in lightroom 5. By slow, I mean for a shoot of 60 shots in raw, it could take 30 minutes before I can preview them all in the pane to see which ones I want to actually import. This was not an issue with lightroom 4. It is not an issue when using lightroom. Previews are fast there, even for images I've never pulled up before.
    My hardware is a PC, with an intell i7 chip, and 16 gig of memory. My nvidia graphics card has the latest drivers.
    I've seen a few others reporting this issue, so I'm trying to log it as a bug. Even though the dial up support text on the adobe lightroom says available 24/7, it now plays a voice recording that says only available m-f.
    Brad

    I'm having the same problem using Windows 7 (64b). I just took a shoot over the weekend for a significant birthday event. Now I can't load the photos. The work-around doesn't work as the program freezes before I get the chance to select the camera under "files".
    I recently upgraded from 3.6 to 5 as I was also noticing a slowdown in the import function. That turned out to be a bad move I think. I also deliberately cleaned up the CF card and removed all my video recordings, as this seems to be reported as part of the problem, but to no avail.
    This seems to be a pretty major bug to have in a major release that's been around for a while now, that I'm expected to pay for, and no sign of a resolution to date!
    Cam. 

  • Aperture 3.1.3 Very slow importing raw files

    Aperture 3.1.3 Very slow importing raw files please help
    Pro Photographer. I have a Mac Pro Quad core 1 yr old with 12GB Ram. I import via 2 Firewire 800 card readers - A Lexar & a UDMR Reader.Only using them singly. System has 4 1TB hard drives 1. Applications, 2. Aperture Library, 3. Aperture Vault, 4 Time Machine
    On a photo shoot, first CF card may download and complete in less than a minute but the next card with a similar number of images may take 10 minutes.
    Sometimes if I swap card readers, this can speed up download but then everything slowly grinds to a halt.
    I have closed down all other programmes, switched off Faces and Enable Gestures - switched off  Look up places.
    Situation driving me nuts as we shoot and view images with customers immediately after the session and download time effecting the amount of time we have to do this. Nothing seems logical as such a high spec machine should theoretically fly through this.

    Have you upgraded your OS ans all your Apps to the latest Versions and also rebooted the system?
    Turn on the Activity Viewer (Windows ->Show Activity) while importing, to see how Aperture is spending its time.
    Also launch some tools from the Application -> Utilities folder:
    launch the Console window, to check for diagnostics /error messages
    launch the Activity Viewer, to see how much RAM and CPU time Aperture and other processes are using.
    How much space is free on your system drive? And on the drive you are importing to?
    If you find in the Activity Viewer, that Aperture is slow, but de facto using not much CPU time, then the problem may be a corrupted Aperture Library or corrupted preference files.
    Try to import into a new, empty Aperture library and see if the problem persists.
    If you have no problems with a new Library, try the Aperture 3: Troubleshooting Basics:
    http://support.apple.com/kb/HT3805
    Start with repairing permissions and repairing the library.
    If none of the measures outlined there, or you need further help, please report back.
    Regards
    Léonie

  • Order Import program run very slow

    Hi
    I use Order Import program to create sale order from tables interface.But it run very slow,about two hours for one record.
    How can i do it run fast?
    Thanks

    Hi,
    before checking metalink for patches (there are some performance related patches for order entry)
    you should check if statistics OE related tables
    - OE_HEADERS_IFACE_ALL,
    - OE_ORDER_HEADERS_ALL
    - OE_LINES_IFACE_ALL
    - OE_ORDER_LINES_ALL
    are up to date (check column last_analyzed in dba_tables). If not, run gather schema statistics
    program to collect current table statistics for OE schema.
    If this does not help, have a look at metalink and query for "order import" "performance"...
    Regards

  • Updated to LR 5.4, images now importing extremely slow. After importing and rendering images are still very slow to go through and develop. No issues in LR 5.3 which was very fast.

    After importing and rendering images are still very slow to go through and develop. No issues in LR 5.3 which was very fast. Canon,Nikon, Leica and Fuji files all rendering slow. iMac 2.9 ghz 16gb RAM OSX 10.8.5. Catalogs and images are kept on external drives. Drives running fine. Really need a fix on this since I'm norally imported and processing at least 4,000 images per LR Catalog. Is there way to revert back to 5.3? I can't use 5.4 is its current state. Thanks.

    Download LR 5.3 and install it
    By the way, lots of people have this complaint about LR 5.4, and the advice I give is to use LR 5.3 unless you need one of the new camera models supported by LR 5.4

  • Importing and exporting from cc premiere pro is very slow

    importing and exporting from cc premiere pro is very slow, the system has i7 processor and memory is 64 gb and has internal raid storage. what could be the issue.

    If you would like to test the tuning and setup of your Premiere installation go to our Premiere Pro BenchMark PPBM, the Instructions page has the download file.  This single download can test Premiere 6, 7 or 8 with the identical benchmark.  We maybe able to tell you why you are having problems.

  • Very very slow import using LOG and Transfer

    I am using a CANON VIXAIA HF10.and the clips are importing very very slow. The max time on each clip is only 5 min or so. Is there a way to speed this up? Thanks.

    Log and transfer only supports AVCHD. It does not support AVCHD Lite, nor will it work with Motion JPEG media. You could use third party software to convert it, or iMovie and then export a QuickTime movie from there in the correct codec and format.

  • Can you help me, i try to import video from my gopro to final cut pro but when i play the video he become very slow and i lose qualiter

    can you help me, i try to import video from my gopro to final cut pro but when i play the video he become very slow and i lose qualiter ???

    Right click one of yur clips in the event browser and choose Transcode Media. Make a test project and playthat clip. See whether that solves your lag.
    Russ

Maybe you are looking for