Create table taking much time

Hi all,
I have to copy one table with more than 1 lakh records from one schema (deloitte) into another schema (dell). It is taking more time but not getting output as early as possible.
Is there any alternative way to make it more faster for copying the data.
Create table dell.newtable as
(select * from deloitte.oldtable)
Advance thanks

Costa wrote:
Hi all,
I have to copy one table with more than 1 lakh records from one schema (deloitte) into another schema (dell). It is taking more time but not getting output as early as possible.
Is there any alternative way to make it more faster for copying the data.
Create table dell.newtable as
(select * from deloitte.oldtable)the slow down is NOT CREATE TABLE itself.
The slowdown is populating the new table with old data.
Other sessions can not see uncommitted data.
So until the statement completes other sessions have no rows to access.

Similar Messages

  • Create table taking long time

    my select query(2m records) coming within a second but while creating a table (nologging) based on the select clause it is taking long time.
    can anybody give me the suggestion which part i will look to improve the performance..

    Plan
    SELECT STATEMENT ALL_ROWS Cost: 11 Bytes: 655 Cardinality: 1                                              
         19 FILTER                                         
              18 NESTED LOOPS Cost: 11 Bytes: 655 Cardinality: 1                                    
                   15 NESTED LOOPS Cost: 9 Bytes: 617 Cardinality: 1                               
                        12 NESTED LOOPS Cost: 8 Bytes: 481 Cardinality: 1                          
                             9 NESTED LOOPS Cost: 6 Bytes: 435 Cardinality: 1                     
                                  6 NESTED LOOPS Cost: 4 Bytes: 209 Cardinality: 1                
                                       3 TABLE ACCESS BY INDEX ROWID TABLE OYSTER_WEB3.TRANSACTION Cost: 2 Bytes: 155 Cardinality: 1           
                                            2 BITMAP CONVERSION TO ROWIDS      
                                                 1 BITMAP INDEX SINGLE VALUE INDEX (BITMAP) OYSTER_WEB3.IX_LINE_COMMODITY_ID
                                       5 TABLE ACCESS BY INDEX ROWID TABLE OYSTERPLUS_DATA.BRIO_SUPPLIERS Cost: 2 Bytes: 54 Cardinality: 1           
                                            4 INDEX UNIQUE SCAN INDEX (UNIQUE) OYSTERPLUS_DATA.PK_BRIO_SUPPLIERS Cost: 1 Cardinality: 1      
                                  8 TABLE ACCESS BY INDEX ROWID TABLE OYSTER3.FLAT_SITE_MV Cost: 2 Bytes: 226 Cardinality: 1                
                                       7 INDEX UNIQUE SCAN INDEX (UNIQUE) OYSTER3.PK_FLAT_SITE_MV Cost: 1 Cardinality: 1           
                             11 TABLE ACCESS BY INDEX ROWID TABLE OYSTER3.SITE_COMMODITY_CODING Cost: 2 Bytes: 46 Cardinality: 1                     
                                  10 INDEX UNIQUE SCAN INDEX (UNIQUE) OYSTER3.PK_SITE_COMMODITY_CODING Cost: 1 Cardinality: 1                
                        14 TABLE ACCESS BY INDEX ROWID TABLE OYSTERPLUS_DATA.BRIO_COMMODITIES Cost: 1 Bytes: 136 Cardinality: 1                          
                             13 INDEX UNIQUE SCAN INDEX (UNIQUE) OYSTERPLUS_DATA.PK_BRIO_COMMODITIES Cost: 0 Cardinality: 1                     
                   17 TABLE ACCESS BY INDEX ROWID TABLE OYSTER3.SUPPLIER_ALIAS Cost: 2 Bytes: 38 Cardinality: 1                               
                        16 INDEX UNIQUE SCAN INDEX (UNIQUE) OYSTER3.PK_SUPPLIER_ALIAS Cost: 1 Cardinality: 1

  • Create Table taking long time (more than 60 minutes) and still running

    Hi all,
    I was able to create and drop table until yesterday, but now the when I issue the below statement the server is stuck, its showing creating table in process.... and then it hangs till you cancel the operation. Query of table and running the procedures are ok, so I am confused why only create table is having issues.
    CREATE TABLE table1
    ( field1 VARCHAR2(20) );
    Any help at the earliest will be appreciated,
    Thanks

    Just check up your status of the log files
    select * from v$logfile
    If the status is inactive,
    Just do a
    alter system switch logfile
    The alert log should have the problem specified check for the file.
    also try to recreate another user and take a try in table creation:
    SQL> connect sys/******
    SQL> create user scott identified by tiger default tablespace users temporary tablespace temp quota 100m on users;
    SQL> grant connect, resource to scott;_
    SQL> connect scott/tiger
    SQL> create table xyz ( a number);
    SQL> create table abc ( d number);
    Could you successfully create the tables?

  • Database tableAUFM hitting is taking much time even secondary index created

    Hi Friends,
       There is report for Goods movement rel. to Service orders + Acc.indicator.
       We have  two testing Systems(EBQ for developer and PEQ from client side).
       EBQ system contains replica of PEQ every month.
       This report is not taking much time in EBQ.But it is taking much time in PEQ.For the selection criteria I have given,both systems  have same data(Getting same output).
    The report has the follwoing fields on the selection criteria:
    A_MJAHR     Material Doc. Year (Mandatory)
    S_BLDAT     Document Date(Optional)
    S_BUDAT     Posting Date(Optional)
    S_LGORT     Storage Location(Optional)
    S_MATNR     Material(Optional)
    S_MBLNR     Material Documen(Optional)t
    S_WERKS     Plant(Optional)
    Client not agrrying to make Material Documen as Mandatory.
    The main (first) table hit is on AUFM table .As there are non-key fileds as well in where condition,We have cretaed a secondary index as well for AUFM table on the following fields:
    BLDAT
    BUDAT
    MATNR
    WERKS
    LGORT 
    Even then also ,in PEQ sytem the report is taking very long time ,Some times not even getting the ALV output.
    What can be done to get teh report executed very fast.
    <removed by moderator>
    The part of report Soure code is as below:
    <long code part removed by moderator>
    Thanks and Regards,
    Rama chary.P
    Moderator message: please stay within the 2500 character limit to preserve formatting, only post relevant portions of the code, also please read the following sticky thread before posting.
    Please Read before Posting in the Performance and Tuning Forum
    locked by: Thomas Zloch on Sep 15, 2010 11:40 AM

    Hi Friends,
       There is report for Goods movement rel. to Service orders + Acc.indicator.
       We have  two testing Systems(EBQ for developer and PEQ from client side).
       EBQ system contains replica of PEQ every month.
       This report is not taking much time in EBQ.But it is taking much time in PEQ.For the selection criteria I have given,both systems  have same data(Getting same output).
    The report has the follwoing fields on the selection criteria:
    A_MJAHR     Material Doc. Year (Mandatory)
    S_BLDAT     Document Date(Optional)
    S_BUDAT     Posting Date(Optional)
    S_LGORT     Storage Location(Optional)
    S_MATNR     Material(Optional)
    S_MBLNR     Material Documen(Optional)t
    S_WERKS     Plant(Optional)
    Client not agrrying to make Material Documen as Mandatory.
    The main (first) table hit is on AUFM table .As there are non-key fileds as well in where condition,We have cretaed a secondary index as well for AUFM table on the following fields:
    BLDAT
    BUDAT
    MATNR
    WERKS
    LGORT 
    Even then also ,in PEQ sytem the report is taking very long time ,Some times not even getting the ALV output.
    What can be done to get teh report executed very fast.
    <removed by moderator>
    The part of report Soure code is as below:
    <long code part removed by moderator>
    Thanks and Regards,
    Rama chary.P
    Moderator message: please stay within the 2500 character limit to preserve formatting, only post relevant portions of the code, also please read the following sticky thread before posting.
    Please Read before Posting in the Performance and Tuning Forum
    locked by: Thomas Zloch on Sep 15, 2010 11:40 AM

  • ODS Activation is taking much time...

    Hi All,
    Some times ods activation is taking much time. Generally it takes 30 mins and some times it take 6 hours.
    If this activation is taking much, then if i check sm50 ...i can see that there is a piece of code is taking much time.
    SELECT
    COUNT(*) , "RECORDMODE"
    FROM
    "/BIC/B0000814000"
    WHERE
    "REQUEST" = :A0 AND "DATAPAKID" = :A1
    GROUP BY
    "RECORDMODE"#
    Could you please let me know what are the possiblites to solve this issue.
    thanks

    Hello,
    you have 2 options:
    1) as already mentioned, cleanup some old psa data or change log data from this psa table or
    2) create a addtional index for recordtype on this table via Tcode se11 -> indexes..
    Regards, Patrick Rieken.

  • Problem : SELECT from LTAP table takes much time (Sort in Database layer)

    Guys,
    Im having problem with this select statement. It takes much time just to get single record.
    The problem is with accessing the LTAP table and the ORDER BY DESCENDING statement.
    The objective of this select statement is to get the non blocked storage bin which is used by latest transfer order number.
    If the latest transfer order no storage bin is blocked, then it will loop and get the 2nd latest transfer order no's storage bin and
    checks whether it blocked or not. It will keep looping.
    The secondary index has been created but the it still taking much time (3 minutes for 10K records in LTAP)
    Secondary Indexes:
    a) LTAP_M ->MANDTLGNUM PQUIT MATNR
    b)LTAP_L  ->LGNUM PQUIT VLTYP VLPLA
    Below is the coding.
    ******************Start of DEVK9A14JW**************************
        SELECT ltaptanum ltapnlpla ltap~wdatu INTO (ltap-tanum, ltap-nlpla, ltap-wdatu)
              UP TO 1 ROWS
              FROM ltap INNER JOIN lagp                         "DEVK9A15OA
              ON lagplgnum =  ltaplgnum
              AND lagplgtyp =  ltapnltyp
              AND lagplgpla =  ltapnlpla
                WHERE lagp~skzue = ' '
                AND ltap~pquit = 'X'
                AND ltap~matnr = ls_9001_scrn-matnr
                AND ltap~lgort = ls_9001_scrn-to_lgort
                AND ltap~lgnum = ls_9001_scrn-lgnum
                AND ltap~nltyp = ls_9001_scrn-nltyp
          ORDER BY tanum DESCENDING.
         ENDSELECT.
        IF sy-subrc EQ 0.
          ls_9001_scrn-nlpla = ltap-nlpla.
          EXIT.
        ENDIF.
    ******************End of DEVK9A14JW**************************

    > Im having problem with this select statement. It takes much time just to get single record.
    This is not true. Together with the ORDER BY the UP TO 1 ROWS does not read 1 record but prepare all records, orders them and return one record, i.e. the largest in sort order.
    You must check what you need, either you need the largest record, then this can be your only possible solution.
    If you need only one recoird then the ORDER BY does not make sense.
    If you need the single largest record, then sometimes the aggregate function MAX can be an alternative.
    I did not look at the index support, this can always be a problem.
    Siegfried

  • Adding column is taking much time. How to avoid?

    ALTER TABLE CONTACT_DETAIL
    ADD (ISIMDSCONTACT_F NUMBER(1) DEFAULT 0 NOT NULL
    ,ISREACHCONTACT_F NUMBER(1) DEFAULT 0 NOT NULL
    Is there any way that to speed up the execution time of the query?
    It's more than 24 hrs completed after started running the above script.
    I do not know why it is taking much time.
    Size of the table is 30 MB.

    To add a column the row directory of every record must be rewritten.
    Obviously this will take time and produce redo.
    Whenever something is slow the first question you need to answer is
    'What is it waiting for?' You can do so by investigating by various v$ views.
    Also, after more than 200 'I can not be bothered to do any research on my own' questions, you should know you don't post here without posting a four digit version number and a platform,
    as volunteers aren't mind readers.
    If you want to continue to withheld information, please consider NOT posting here.
    Sybrand Bakker
    Senior Oracle DBA
    Experts: those who did read documentatiion and can be bothered to investigate their own problems.

  • Query taking much time Orace 9i

    Hi,
    **How can we tune the sql query in oracle 9i.**
    The select query taking more than 1 and 30 min to throw the result.
    Due to this,
    We have created materialsed view on the select query and also we submitted a job to get Materilazed view refreshed daily in dba_jobs.
    When we tried to retrive the data from Materilased view getting result very quickly.
    But the job which we has been assisgned in Dbajobs taking equal  time to complete, as the query use to take.
    We feel since the job taking much time in the test Database and it may cause load if we move the same scripts in Production Environment.
    Please suggest how to resolvethe issue and also how to tune the sql
    With Regards,
    Srinivas
    Edited by: Srinivas.. on Dec 17, 2009 6:29 AM

    Hi Srinivas;
    Please follow this search and see its helpful
    Regard
    Helios

  • Taking much time for loading data

    Dear All
    While loading data in <b>PRD system</b> (for master data and transaction data), it's taking much time. For ex., 2LIS_05_QOITM, I am loading delta data only in <b>PSA</b> from R/3. Some times (yesterday) it's taking 2 mins and some times (today) it's taking 5 hrs not yet completed (it's yellow). Actually yesterday, we went to SM58 in R/3 side and executed that LUW for some other data sources. Now also we can do but we don't want to like that. Because we are expecting permanent soulution. Could u please advice me. I am getting below message in status tab
    Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the source system.
    I am loading data through Process chain and user is <b>BWREMOTE (authorized user).</b>
    Please help me.
    Thanks a lot in advance
    Raja

    Dear Karthik
    No I couldn't resolve till now
    But Everything is fine.
    Now status is yellow only (209 from 209). Now what i want to do.
    Im getting below message <b>in status tab</b>
    Missing data packages for PSA Table
    Diagnosis
    Data packets are missing from PSA Table . BI processing does not return any errors. The data transport from the source system to BI was probably incorrect.
    Procedure
    Check the tRFC overview in the source system.
    You access this log using the wizard or following the menu path "Environment -> Transact. RFC -> Source System".
    Error handling:
    If the tRFC is incorrect, resolve the errors listed there.
    Check that the source system is connected properly to BI. In particular, check the remote user authorizations in BI.
    <b>In detail tab</b>, I am getting below message
    Info IDoc 2 : sent, not arrived ; IDoc ready for dispatch (ALE service)
    Thanks in advance
    Raja

  • Query taking much time.

    Hi All,
    I have one query which is taking much time in dev envi where data size is very small and planning to implement this query in production database where database size is huge. Plz let me know how I can optimize this query.
    select count(*) from
    select /*+ full(tls) full(tlo) parallel(tls,2) parallel(tls, 2) */
    tls.siebel_ba, tls.msisdn
    from
    TDB_LIBREP_SIEBEL tls, TDB_LIBREP_ONDB tlo
    where
    tls.siebel_ba = tlo.siebel_ba (+) and
    tls.msisdn = tlo.msisdn (+) and
    tlo.siebel_ba is null and
    tlo.msisdn is null
    union
    select /*+ full(tls) full(tlo) parallel(tls,2) parallel(tls, 2) */
    tlo.siebel_ba, tlo.msisdn
    from
    TDB_LIBREP_SIEBEL tls, TDB_LIBREP_ONDB tlo
    where
    tls.siebel_ba (+) = tlo.siebel_ba and
    tls.msisdn (+) = tlo.msisdn and
    tls.siebel_ba is null and
    tls.msisdn is null
    explain plan of above query is
    | Id | Operation | Name | Rows | Bytes | Cost | TQ |IN-OUT| PQ Distrib |
    | 0 | SELECT STATEMENT | | 1 | | 14 | | | |
    | 1 | SORT AGGREGATE | | 1 | | | | | |
    | 2 | SORT AGGREGATE | | 1 | | | 41,04 | P->S | QC (RAND) |
    | 3 | VIEW | | 164 | | 14 | 41,04 | PCWP | |
    | 4 | SORT UNIQUE | | 164 | 14104 | 14 | 41,04 | PCWP | |
    | 5 | UNION-ALL | | | | | 41,03 | P->P | HASH |
    |* 6 | FILTER | | | | | 41,03 | PCWC | |
    |* 7 | HASH JOIN OUTER | | | | | 41,03 | PCWP | |
    | 8 | TABLE ACCESS FULL| TDB_LIBREP_SIEBEL | 82 | 3526 | 1 | 41,03 | PCWP | |
    | 9 | TABLE ACCESS FULL| TDB_LIBREP_ONDB | 82 | 3526 | 2 | 41,00 | S->P | BROADCAST |
    |* 10 | FILTER | | | | | 41,03 | PCWC | |
    |* 11 | HASH JOIN OUTER | | | | | 41,03 | PCWP | |
    | 12 | TABLE ACCESS FULL| TDB_LIBREP_ONDB | 82 | 3526 | 2 | 41,01 | S->P | HASH |
    | 13 | TABLE ACCESS FULL| TDB_LIBREP_SIEBEL | 82 | 3526 | 1 | 41,02 | P->P | HASH |
    Predicate Information (identified by operation id):
    6 - filter("TLO"."SIEBEL_BA" IS NULL AND "TLO"."MSISDN" IS NULL)
    7 - access("TLS"."SIEBEL_BA"="TLO"."SIEBEL_BA"(+) AND "TLS"."MSISDN"="TLO"."MSISDN"(+))
    10 - filter("TLS"."SIEBEL_BA" IS NULL AND "TLS"."MSISDN" IS NULL)
    11 - access("TLS"."SIEBEL_BA"(+)="TLO"."SIEBEL_BA" AND "TLS"."MSISDN"(+)="TLO"."MSISDN")

    user3479748 wrote:
    Hi All,
    I have one query which is taking much time in dev envi where data size is very small and planning to implement this query in production database where database size is huge. Plz let me know how I can optimize this query.
    select count(*) from
    select /*+ full(tls) full(tlo) parallel(tls,2) parallel(tls, 2) */
    tls.siebel_ba, tls.msisdn
    from
    TDB_LIBREP_SIEBEL tls, TDB_LIBREP_ONDB tlo
    where
    tls.siebel_ba = tlo.siebel_ba (+) and
    tls.msisdn = tlo.msisdn (+) and
    tlo.siebel_ba is null and
    tlo.msisdn is null
    union
    select /*+ full(tls) full(tlo) parallel(tls,2) parallel(tls, 2) */
    tlo.siebel_ba, tlo.msisdn
    from
    TDB_LIBREP_SIEBEL tls, TDB_LIBREP_ONDB tlo
    where
    tls.siebel_ba (+) = tlo.siebel_ba and
    tls.msisdn (+) = tlo.msisdn and
    tls.siebel_ba is null and
    tls.msisdn is null
    ) ;explain plan of above query is
    | Id  | Operation                |  Name              | Rows  | Bytes | Cost  |  TQ    |IN-OUT| PQ Distrib |
    |   0 | SELECT STATEMENT         |                    |     1 |       |    14 |        |      |            |
    |   1 |  SORT AGGREGATE          |                    |     1 |       |       |        |      |            |
    |   2 |   SORT AGGREGATE         |                    |     1 |       |       | 41,04  | P->S | QC (RAND)  |
    |   3 |    VIEW                  |                    |   164 |       |    14 | 41,04  | PCWP |            |
    |   4 |     SORT UNIQUE          |                    |   164 | 14104 |    14 | 41,04  | PCWP |            |
    |   5 |      UNION-ALL           |                    |       |       |       | 41,03  | P->P | HASH       |
    |*  6 |       FILTER             |                    |       |       |       | 41,03  | PCWC |            |
    |*  7 |        HASH JOIN OUTER   |                    |       |       |       | 41,03  | PCWP |            |
    |   8 |         TABLE ACCESS FULL| TDB_LIBREP_SIEBEL  |    82 |  3526 |     1 | 41,03  | PCWP |            |
    |   9 |         TABLE ACCESS FULL| TDB_LIBREP_ONDB    |    82 |  3526 |     2 | 41,00  | S->P | BROADCAST  |
    |* 10 |       FILTER             |                    |       |       |       | 41,03  | PCWC |            |
    |* 11 |        HASH JOIN OUTER   |                    |       |       |       | 41,03  | PCWP |            |
    |  12 |         TABLE ACCESS FULL| TDB_LIBREP_ONDB    |    82 |  3526 |     2 | 41,01  | S->P | HASH       |
    |  13 |         TABLE ACCESS FULL| TDB_LIBREP_SIEBEL  |    82 |  3526 |     1 | 41,02  | P->P | HASH       |
    Predicate Information (identified by operation id):
    6 - filter("TLO"."SIEBEL_BA" IS NULL AND "TLO"."MSISDN" IS NULL)
    7 - access("TLS"."SIEBEL_BA"="TLO"."SIEBEL_BA"(+) AND "TLS"."MSISDN"="TLO"."MSISDN"(+))
    10 - filter("TLS"."SIEBEL_BA" IS NULL AND "TLS"."MSISDN" IS NULL)
    11 - access("TLS"."SIEBEL_BA"(+)="TLO"."SIEBEL_BA" AND "TLS"."MSISDN"(+)="TLO"."MSISDN")
    I dunno, it looks like you are getting all the things that are null with an outer join, so won't that decide to full scan anyways? Plus the union means it will do it twice and do a distinct to get rid of dups - see how it does a union all and then sort unique. Somehow I have the feeling there might be a more trick way to do what you want, so maybe you should state exactly what you want in English.

  • Taking much time when trying to Drill a characterstic in the BW Report.

    Hi All,
    When we are executing the BW report, it is taking nearly 1 to 2 mins then when we are trying to drill down a characterstic it is taking much time nearly 30 mins to 1 hour and througing an error message that,
    "An error has occared during loading. Plese look in the upper frame for further information."
    I have executed this query in RSRT and cheked the query properties,
    this quey is bringing the data directly form Aggregates but some chares are not avalable in the Agrregtes.
    So... after execution when we are trying to drill down the chars is taking much time for chars which are not avilable in the Aggregates. For other chars which are avilable in the Aggregates it is taking only 2 to 3 mins only.
    How to do the drill down for the chars which are not avilable in the Aggregates with out taking much time and error.
    Could you kindly give any solution for this.
    Thanks & Regards,
    Raju. E

    Hi,
    The only solution is to include all the char used in the report in the aggregates or this will the issue you will face.
    just create a proposal for aggregates before creating any new aggregates as it will give you the idea which one is most used.
    Also you should make sure that all the navigation characteristics are part of the aggregates.
    Thanks
    Ajeet

  • Discoverer report is taking much time to open

    Hi
    All the discoverer report are taking much time to open,even query in lov is taking 20 -25 min.s.We have restart the services but on result found.
    Please suggest what can be done ,my application is on 12.0.6.
    Regards

    This topic was discussed many times in the forum before, please see old threads for details and for the docs you need to refer to -- https://forums.oracle.com/forums/search.jspa?threadID=&q=Discoverer+AND+Slow&objID=c3&dateRange=all&userID=&numResults=15&rankBy=10001
    Thanks,
    Hussein

  • LOV is slow after selecting a value its taking much time to default

    Hi,
    I have a dependent LOV. Master LOV is executing fine and its populatin into the field fastly. But Child LOV is very slow after selecting a value its taking much time to default.
    Can any one please help me if there is any way to default the value fast after selecting a value?
    Thanks,
    Mahesh

    Hi Gyan,
    Same issues in TST and PROD instances.
    In my search criteria if i get 1 record, even after selecting that value its taking much time to default that value into the field.
    Please advice. Thanks for your quick resp.
    Thanks,
    Mahesh

  • HT1544 i have downloaded mac os 10.5 leopard update version successfully . but try to instal it is taking much time and the status bar is not going

    Hi < i have updated my mac 10.5 with 10.5.8 sucessfully , but while instaling it is taking much time and status bar is not showing any progress

    If I remember correctly one of the updates could hang after doing the update. Fixed by a restart but installing a combo update over the top of others rarely does any harm and it may be the best thing to do.

  • While saving the Ai file its taking much time

    I am using windows 7 and Ai CC, when i am working on files while saving the Ai file its taking much time. I got 4 GB RAM. now i got better system too with ram of 6GB. stillthe Ai is taking much time to save 300 MB file.

    Thank you Jdanek.
    1. I am saving the file in my local HDD only.
    2. Scratch disk is F Drive [100 GB free space], there no data is stored.
    3. Even i increased the Virtual memory too.
    4. Now I switched off the "Save with PDF compatability" too,  No luck!

Maybe you are looking for

  • I want to install Skype on my iPhone.

    I have got an iPhone. I want to install Skype on my iPhone. When installing it throught iTunes, it is showing that Skype is a free application but when I continue to download and install it, it asks for payment through credit cards!? So what is the m

  • How to create business charts in Forms 10g?

    In the old days, before and in 8i, we can still create business charts using Graphics Builder with the following graph types. How can we build business charts in forms 10g? Column: Shows discrete values using vertical columns Bar: Shows discrete valu

  • How to Export Aperture Library for Use by Windows PC

    Hello and thanks in advance for your help. I have recently returned to using a Mac after years of PC use. I'm using Parallels for some apps that are PC only, an astronomy program, TheSky6, for example. One of the first things I did was purchase Apert

  • Stored function with OUT parameter

    Hi there, anyone knows wether a stored function may have an OUT parameter other than a returning value and/or IN parameter. I think not but I've a doubt about it. Thanks Paolo Paolucci Oracle Consulting

  • No Aggregation Level Defined for Data Matrix ICHSM1

    Hey Folks, I see that there is no Aggregation Level defined for Data Matrix ICHSM1 ( the SMI cBook ). I now need to add new key figures to this Data Matrix. However, I am not being able to do so since there are no Aggregation Levels defined in the fi