Problem Copy big Table sqlserver 2012

Hi Friends,
Its 4th day I am trying to copy  table having 7 million record with text field  into same database (sqlserver 2012)  , but fail although I am using blade server with 256 Gb memory and multiprocessor. My query as followed
select * into newtable from  maintable where flag=0 
(Just while ago   I  canceled above statement (query) which was working since last 14 hours on server)
Please help mean how I can copy.
thank you.
regards,
asad
asad

Hi Friends,
Its 4th day I am trying to copy  table having 7 million record with text field  into same database (sqlserver 2012)  , but fail although I am using blade server with 256 Gb memory and multiprocessor. My query as followed
select * into newtable from  maintable where flag=0 
(Just while ago   I  canceled above statement (query) which was working since last 14 hours on server)
Please help mean how I can copy.
Is the data going to change while you are running select. If it is not going to change and its history data and its just there for movement you can use NOLOCK hint like
select * into newtable from maintable with (NOLOCK) where flag=0
Please make sure
NOLOCK query hint alos takes Schema stability lock and can be blocked if table from wher yo are selecting data is locked by alter table command. it also causes dirty reads and unexpected result so use at your own risk.
You should consider methods suggested by Saeid
Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
My Technet Articles

Similar Messages

  • In Pages, How to copy a table? How to make a table as a Header?

    Hi, How can I copy a table? I selected it , command c, then clicked where I wanted the copy, and clicked command v. Only the text from the table was copied. The table outline did not paste.
    I would like the table to appear under the Header of every page of the document. So I tried placing the table in the Header. Worked for the first page, but did not appear in the Header of the second page ... just a really big Header box.
    Thank you for your help.
    margaretfrommb

    Hi, margaretfromn
    I'm sorry this reply may not solve your problem. I have the same problem as you.
    What I have found that tables can be copied by using "option+drag" WITHIN the document. But, the same technique do not work across documents.
    In Pages '09, there is a function to convert text to table, but it has disappeared in Pages 5 as the other convenient functions.
    Cheers,

  • PDF export - problems with displaying tables

    Hi,
    I have nice and neat rtf template with some text and big table. It looks fine, but when I export it to PDF it starts the table from new page, not just below the text, which results in 75% blank page. How can I fix it?

    We are using the PDF Export 8.3.7 version, and are having a similar problem. We have a Word file (.doc) with an embedded graphic. Inside the graphic is a table (the original graphic was created some where else and had text and a table, then an image was made of that which was then embedded as a graphic in the Word document). The problem is that the table does not render corretly - the last row of data is not displayed. Oddly, if in the resulting PDF/A output you highlight the region where the data should be displayed, and copy that to Word/Notepad, the data is there, the issue seems to be in rendering the data to make it visible in the rendered PDF/A output. We see other problems with tables in other files, for example some tables are rotated ninety degrees.
    Will there be a fix to PDFExport anytime soon?
    Thanks,
    Kevin

  • GG for Sqlserver 2012

    Hi all,
    I want to replicate tables from Sqlserver 2012 to Oracle 11g. Is this version supported?
    Where can I find related docs for this activity?
    Thanks a lot,
    zxy

    Thanks Marie,
    Supposing I have a big table of Images/pictured LOB from Sqlserver( 1 million rows), and I want to migrate it to Oracle.
    I read in the docs that there is an "initial load" process in goldengate.  So does this mean GG can handle this bulk load of 1 millions lob images table?
    I just want to migrate the table. as another alternative/variation  just like doing expdp and impdp from oracle to oracle database, or extract to csv  and  sqlloading.
    (Since I can not do it with sqlserver-to-oracle)
    Thanks

  • How to copy a table with LONG and CLOB datatype over a dblink?

    Hi All,
    I need to copy a table from an external database into a local one. Note that this table has both LONG and CLOB datatypes included.
    I have taken 2 approaches to do this:
    1. Use the CREATE TABLE AS....
    SQL> create table XXXX_TEST as select * from XXXX_INDV_DOCS@ext_db;
    create table XXXX_TEST as select * from XXXX_INDV_DOCS@ext_db
    ERROR at line 1:
    ORA-00997: illegal use of LONG datatype
    2. After reading some threads I tried to use the COPY command:
    SQL> COPY FROM xxxx/pass@ext_db TO xxxx/pass@target_db REPLACE XXXX_INDV_DOCS USING SELECT * FROM XXXX_INDV_DOCS;
    Array fetch/bind size is 15. (arraysize is 15)
    Will commit when done. (copycommit is 0)
    Maximum long size is 80. (long is 80)
    CPY-0012: Datatype cannot be copied
    If my understanding is correct the 1st statement fails because there is a LONG datatype in XXXX_INDV_DOCS table and 2nd one fails because there is a CLOB datatype.
    Is there a way to copy the entire table (all columns including both LONG and CLOB) over a dblink?
    Would greatelly appriciate any workaround or ideas!
    Regards,
    Pawel.

    Hi Nicolas,
    There is a reason I am not using export/import:
    - I would like to have a one-script solution for this problem (meaning execute one script on one machine)
    - I am not able to make an SSH connection from the target DB to the local one (although the otherway it works fine) which means I cannot copy the dump file from target server to local one.
    - with export/import I need to have an SSH connection on the target DB in order to issue the exp command...
    Therefore, I am looking for a solution (or a workaround) which will work over a DBLINK.
    Regards,
    Pawel.

  • Copying a table from one databse to another

    Hi,
    I used the following code to copy a table from one database to another.
    set copycommit 1
    set arraysize 1000
    copy from username/passwd@tnsname -
    create <tablename> -
    using -
    select * from <tablename>
    But I get the following error:
    set copycommit 1
    ERROR at line 1:
    ORA-00922: missing or invalid option
    Could you please let me know how this can be done.
    Thanks,
    Narasimhan

    Thanks for your suggestions.I created a database link.I had no problems.
    Then when i issue the command
    copy from uname/password@db
    create <tablename>
    using
    select * from <tablename>
    <Here db is database link to source database identified by uname and password>.
    I still get the error
    copy from uname/password@db
    ERROR at line 1:
    ORA-00900: invalid SQL statement
    Does it mean that the command is wrong?Or I'm doing something different.
    Thanks!

  • I have a problem when update tables with a database link (APEX)

    Hi, I search answers for my problem and I not found a solution. My problem start when I open page with database link statement in APEX. I drop tables and execute this for update the tables in APEX.
    EXECUTE IMMEDIATE 'CREATE TABLE XXXXXXX
    AS
    SELECT A1,A2,A3
    FROM AAA@BBB';
    I use Mozilla Firefox and isn't always updated the tables because it shows dialog for downloading file "f". I rename the file to txt and I see the text "IBMPC/WIN_NT-8.1.0"...
    In the explorer always show a window only the text "IBMPC/WIN_NT-8.1.0"...
    What is the solution for my problem? I need update this tables.
    Thanks. John.

    It's easy. It doesn't work a database link, on apex, with a pl/sql code or store procedure.
    I saw that many people had the same problem but I don't see a solution.
    As I put up, I want to remove my tables on APEX and copy the table of my Oracle database with the update dates.
    The posts are from the same problem as my English is very bad, it got to help explain the problem.
    I have a page with a pl/sql code. This code in Oracle is correctly.
    For example, with this code I update the table PRUEBAS with the data of TABLEX.
    BEGIN
    BEGIN
    EXECUTE IMMEDIATE 'DROP TABLE "PRUEBAS"';
    EXCEPTION WHEN OTHERS THEN NULL;
    END;
    EXECUTE IMMEDIATE 'CREATE TABLE "PRUEBAS"
    AS
    SELECT AAA,
    BBB,
    CCC
    FROM TABLEX';
    END;
    But, in this code, I have the problem.
    BEGIN
    BEGIN
    EXECUTE IMMEDIATE 'DROP TABLE "PRUEBAS"';
    EXCEPTION WHEN OTHERS THEN NULL;
    END;
    EXECUTE IMMEDIATE 'CREATE TABLE "PRUEBAS"
    AS
    SELECT AAA,
    BBB,
    CCC
    FROM TABLEX@DB_LINK';
    END;
    I sometimes get an error. Sometimes it works fine, but half of the tries Apex returns "IBMPC/WIN_NT-8.1.0" in the browser and I am not using an IBM PC or Win NT...
    And i read that In the alert_xe.log file there are a error of core dump... ... :S or that the application return corrupt headers.... Is a problem of versions?
    I used a synonym and I have the same problem.      
    Even if I run the code on SQL COMMANDS (SQL WORKSHOP) it shows dialog for downloading file "f". I rename the file to txt and I see the text "IBMPC/WIN_NT-8.1.0"...
    The problem is in Mozilla Firefox and in IE.
    I have the OracleXE 10g (Application Express 3.0.1.00.08).
    And the question is.. ... ... ... ... What is the solution? :)
    Thanks.
    John.

  • Whitespace problem with embedded tables

    I have a datastructure with embedded groups. In my RTF template, I am using a table for each of these groups, and embedding them likewise. When I load my data, however, and view it, there is a random, undesirable amount of whitespace between where one table ends and where its parent table ends.
    To try to make it more clear:
    "TEXT in inner table"
    ----------------------------------------- (End of inner table)
    (Random Gap Here)
    "Text in outer table"
    ----------------------------------------- (End of outer table)
    The Size of the gap seems to depend on how much space is left on the page. If there is just enough data to fill a page, it will all fit nicely, exactly how I want it. If a page does not contain much data, then the gap gets really big, as if it was trying to cover the entire page.
    Anyone else run into this problem?
    Thanks,
    Matt

    Thanks very much!
    Before, I had a separate table for each group. When it was all said and done, I had about 5 levels of table embedding. YUCK!
    Now I made one big table (1x1) and only used tables where I couldn't get by with plain indenting. The most table embedding that is going on now is individual tables within the large 1x1 table. This method proves much more flexible and elegant.
    Thank you very much ashee1,
    Matt Soukup

  • HS ODBC GONE AWAY ON BIG TABLE QRY

    Hello,
    I have an HS ODBC connection set up pointing to a MySQL 5.0 database on Windows using mysql odbc 3.51.12. Oracle XE is on the same box and tnsames, sqlnet.ora, and HS ok is all set up.
    The problem is I have a huge table 100 mill rows, in MySQL, and when I run a query in Oracle SQL Developer it runs for about two minutes then I get errrors ORA-00942 lost connection, or gone away.
    I can run a query against a smaller table in the schema and it returns rows quickly. So I know the HS ODBC connection is working.
    I noticed the HS service running on Windows starts up and uses 1.5 gig of memory and the CPU time maxes to 95%, on the big table query, then the connection drops.
    Any advice on what to do here. There doesn't seem to be any config settings with HS service to limit or increase the rows, or increase the cache.
    MySQL does have some advanced ODBC driver options that I will try.
    Does anyone have any suggestions on how to handle this overloading problem??
    Thanks for the help,

    FYI, HS is Oracle Hetrogenous service to connect to non-oracle databases.
    I actually found a workaround. The table is so large the query crashes. So I broke table up with 5 MySql views, and now am able to query the views using select insert Oracle stored procedure into Oracle table.

  • Managing a big table

    Hi All,
    I have a big table in my database. When I say big, it is related to data stored in it (around 70 million recs) and also no of columns (425).
    I do not have any problems with it now, but going ahead I assume, it would be a bottleneck or very difficult to manage this table.
    I have a star schema for the application of which this is a master table.
    Apart from partitioning the table is there any other way of better handling such a table.
    Regards

    Hi,
    Usually the fact tables tend to be smaller in number of columns and larger in number of records while the dimension tables obey to the opposite larger number of columns, which is were the powerful of the dimension lays on, and very few (in some exceptions even millions of record) records. So the high number of columns make me thing that the fact table may be, only may be, I don't have enough information, improperly designed. If that is the case then you may want to revisit that design and most likely you will find some 'facts' in your fact table that can become attributes of any of the dimension tables linked to.
    Can you say why are you adding new columns to the fact table? A fact table is created for a specific business process and if done properly there shouldn't be such a requirement of adding new columns. A fact use to be limited in the number of metrics you can take from it. In fact, it is more common the oposite, a factless fact table.
    In any case, from the point of view of handling this large table with so many columns I would say that you have to focus on avoiding the increasing number of columns. There is nothing in the database itself, such as partitioning that could do this for you. So one option is to figure out which columns you want to get 'vertical partition' and split the table in at least two new tables. The set of columns will be those that are more frequently used or those that are more critical to you.Then you will have to link these two tables together and with the rest of dimensions. But, again if you are adding new columns then is just a matter of time that you will be running in the same situation in the future.
    I am sorry but cannot offer better advice than to revisit the design of your fact table. For doing that you may want to have a look at http://www.kimballgroup.com/html/designtips.html
    LW

  • Local client copy error - Table "BKPF" not edited due to excessive number

    Hi,
    I am performing a local client copy on our ECC 6.0 system. Client copy took more than 12 hours and its still in progress with 7 tables still to be copied. Logs in SCC3 -
    ===
    2 ETA163 Table "BKPF" not edited due to excessive number of errors
    2 ETA057 WARNING: Cancelled several times
    2 ETA163 Table "COEP" not edited due to excessive number of errors
    3 ETA311 Process "00002" started on server "sfrndevsap15"
    (time: "17:20:51")
    2 ETA297 Error: Table "ONROV" error in DDIC - Check table with SE14
    2 ETA057 WARNING: Cancelled several times
    2 ETA163 Table "FAGLFLEXA" not edited due to excessive number of errors
    3 ETA311 Process "00003" started on server "sfrndevsap15"
    (time: "17:20:52")
    4 ETA346 "FAGLFLEXC :"" 0 0 0 DEL.""
    0 0 17:20:52"
    2 ETA057 WARNING: Cancelled several times
    2 ETA163 Table "HRP1001" not edited due to excessive number of errors
    3 ETA311 Process "00004" started on server "sfrndevsap15"
    (time: "17:20:52")
    4 ETA346 "VSAFVU_CN :"" 0 86123 0 DEL.""
    0 29 17:21:21"
    2 ETA057 WARNING: Cancelled several times
    2 ETA163 Table "PCL4" not edited due to excessive number of errors
    ===
    Please help ASAP, as this is effecting our testing team.  I have tried to copy individual tables also using a transport request number and SCC1 to copy it to the target client.  This is also taking quite a long time.
    Thanks in advance,
    Abdul

    Hi Abdul,
    What does your system log say ?
    Check this SAP Note
    Note 579783 - CC ERROR: Loss of data - table not copied
    run the report RSCC_VERIFY to check if the table is consistent within the source client.
    SAP will only provide support for RSCC_VERIFY messages if there is a specific problem in an application in the target client that can be assigned immediately to the messages.
    RSCC_VERIFY is delivered with SAPKB46B49, SAPKB46C35 and SAPKB46D24. In older releases, you can create the report in the ABAP editor (transaction SE38) in accordance with the attached advance correction. Use the development class or the STRM package for this.
    Regards,
    Siddhesh

  • Very Big Table (36 Indexes, 1000000 Records)

    Hi
    I have a very big table (76 columns, 1000000 records), these 76 columns include 36 foreign key columns , each FK has an index on the table, and only one of these FK columns has a value at the same time while all other FK have NULL value. All these FK columns are of type NUMBER(20,0).
    I am facing performance problem which I want to resolve taking in consideration that this table is used with DML (Insert,Update,Delete) along with Query (Select) operations, all these operations and queries are done daily. I want to improve this table performance , and I am facing these scenarios:
    1- Replace all these 36 FK columns with 2 columns (ID, TABLE_NAME) (ID for master table ID value, and TABLE_NAME for master table name) and create only one index on these 2 columns.
    2- partition the table using its YEAR column, keep all FK columns but drop all indexes on these columns.
    3- partition the table using its YEAR column, and drop all FK columns, create (ID,TABLE_NAME) columns, and create index on (TABLE_NAME,YEAR) columns.
    Which way has more efficiency?
    Do I have to take "master-detail" relations in mind when building Forms on this table?
    Are there any other suggestions?
    I am using Oracle 8.1.7 database.
    Please Help.

    Hi everybody
    I would like to thank you for your cooperation and I will try to answer your questions, but please note that I am a developer in the first place and I am new to oracle database administration, so please forgive me if I did any mistakes.
    Q: Have you gathered statistics on the tables in your database?
    A: No I did not. And if I must do it, must I do it for all database tables or only for this big table?
    Q:Actually tracing the session with 10046 level 8 will give some clear idea on where your query is waiting.
    A: Actually I do not know what you mean by "10046 level 8".
    Q: what OS and what kind of server (hardware) are you using
    A: I am using Windows2000 Server operating system, my server has 2 Intel XEON 500MHz + 2.5GB RAM + 4 * 36GB Hard Disks(on RAID 5 controller).
    Q: how many concurrent user do you have an how many transactions per hour
    A: I have 40 concurrent users, and an average 100 transaction per hour, but the peak can goes to 1000 transaction per hour.
    Q: How fast should your queries be executed
    A: I want the queries be executed in about 10 to 15 seconds, or else every body here will complain. Please note that because of this table is highly used, there is a very good chance to 2 or more transaction to exist at the same time, one of them perform query, and the other perform DML operation. Some of these queries are used in reports, and it can be long query(ex. retrieve the summary of 50000 records).
    Q:please show use the explain plan of these queries
    A: If I understand your question, you ask me to show you the explain plan of those queries, well, first, I do not know how , an second, I think it is a big question because I can not collect all kind of queries that have been written on this table (some of them exist in server packages, and the others performed by Forms or Reports).

  • How to Slice big table in chunks

    I am trying to derive a piece of generic logic that would cut in chunks of definite size any big table. The goal is to perform update in chunks and avoid rollback too small issues. The full table scan on the update is unavoidable, since the update target every row of the table.
    The BIGTABLE has 63 millions rows. The purpose of the bellow SQL to give the ROWID every two million rows. So I am using the auto row numering field 'rownum' and perfrom a test to see I could. I expected the fist chunk to have 2 millons rows but in fact it is not the case:
    Here is the code +(NOTE I had many problems with quotes, so some ROWID appears without their enclosing quotes or they disappear from current output here)+:
    select rn, mod, frow, rownum from (
        select rowid rn ,  rownum frow, mod(rownum, 2000000) mod 
      from bigtable order by rn) where mod = 0
    SQL> /
    RN                        MOD       FROW     ROWNUM
    AAATCjAA0AAAKAVAAd          0    4000000          1
    AAATCjAA0AAAPUEAAv          0   10000000          2
    AAATCjAA0AAAbULAAx          0    6000000          3
    AAATCjAA0AAAsIeAAC          0   14000000          4
    AAATCjAA0AAAzhSAAp          0    8000000          5
    AAATCjAA0AABOtGAAa          0   26000000          6
    AAATCjAA0AABe24AAE          0   16000000          7
    AAATCjAA0AABjVgAAQ          0   30000000          8
    AAATCjAA0AABn4LAA3          0   32000000          9
    AAATCjAA0AAB3pdAAh          0   20000000         10
    AAATCjAA0AAB5dmAAT          0   22000000         11
    AAATCjAA0AACrFuAAW          0   36000000         12
    AAATCjAA6AAAXpOAAq          0    2000000         13
    AAATCjAA6AAA8CZAAO          0   18000000         14
    AAATCjAA6AABLAYAAj          0   12000000         15
    AAATCjAA6AABlwbAAg          0   52000000         16
    AAATCjAA6AACBEoAAM          0   38000000         17
    AAATCjAA6AACCYGAA1          0   24000000         18
    AAATCjAA6AACKfBABI          0   28000000         19
    AAATCjAA6AACe0cAAS          0   34000000         20
    AAATCjAA6AAFmytAAf          0   62000000         21
    AAATCjAA6AAFp+bAA6          0   60000000         22
    AAATCjAA6AAF6RAAAQ          0   44000000         23
    AAATCjAA6AAHJjDAAV          0   40000000         24
    AAATCjAA6AAIR+jAAL          0   42000000         25
    AAATCjAA6AAKomNAAE          0   48000000         26
    AAATCjAA6AALdcMAA3          0   46000000         27
    AAATCjAA9AAACuuAAl          0   50000000         28
    AAATCjAA9AABgD6AAD          0   54000000         29
    AAATCjAA9AADiA2AAC          0   56000000         30
    AAATCjAA9AAEQMPAAT          0   58000000         31
    31 rows selected.
    SQL> select count(*) from BIGTABLE where rowid < AAATCjAA0AAAKAVAAd ;
      COUNT(*)
        518712             <-- expected around 2 000 000
    SQL> select count(*) from BIGTABLE where rowid < AAATCjAA0AAAPUEAAv ;
      COUNT(*)
       1218270     <-- expected around 4 000 000
    SQL> select count(*) from BIGTABLE where rowid < AAATCjAA0AAAbULAAx ;
      COUNT(*)
       2685289    <-- expected around 6 000 000Amzingly, This code works perfectly for small tables but fails for big tables. Does anybody has an explanation and possibly a solution to this?
    Here is the full code of the SQL that is suppposed to generated all the predicates I need to add to the UPdate statements in order to cut them in piece :
    select line  from (
       with v as (select rn, mod, rownum frank from (
           select rowid rn ,  mod(rownum, 2000000) mod
               from BIGTABLE order by rn ) where mod = 0),
          v1 as (
                  select rn , frank, lag(rn) over (order by frank) lag_rn  from v ),
          v0 as (
                  select count(*) cpt from v)
        select 1, case
                    when frank = 1 then ' and rowid  <  ''' ||  rn  || ''''
                    when frank = cpt then ' and rowid >= ''' || lag_rn ||''' and rowid < ''' ||rn || ''''
                    else ' and rowid >= ''' || lag_rn ||''' and rowid <'''||rn||''''
                 end line
    from v1, v0
    union
    select 2, case
               when frank =  cpt then   ' and rowid >= ''' || rn  || ''''
              end line
        from v1, v0 order by 1)
    and rowid  <  AAATCjAA0AAAKAVAAd
    and rowid >= 'AAATCjAA0AAAKAVAAd' and rowid < 'AAATCjAA0AAAPUEAAv''
    and rowid >= 'AAATCjAA0AAAPUEAAv' and rowid < 'AAATCjAA0AAAbULAAx''
    and rowid >= 'AAATCjAA0AAAbULAAx' and rowid < 'AAATCjAA0AAAsIeAAC''
    and rowid >= 'AAATCjAA0AAAsIeAAC' and rowid < 'AAATCjAA0AAAzhSAAp''
    and rowid >= 'AAATCjAA0AAAzhSAAp' and rowid < 'AAATCjAA0AABOtGAAa''
    and rowid >= 'AAATCjAA0AAB3pdAAh' and rowid < 'AAATCjAA0AAB5dmAAT''
    and rowid >= 'AAATCjAA0AAB5dmAAT' and rowid < 'AAATCjAA0AACrFuAAW''
    and rowid >= 'AAATCjAA0AABOtGAAa' and rowid < 'AAATCjAA0AABe24AAE''
    and rowid >= 'AAATCjAA0AABe24AAE' and rowid < 'AAATCjAA0AABjVgAAQ''
    and rowid >= 'AAATCjAA0AABjVgAAQ' and rowid < 'AAATCjAA0AABn4LAA3''
    and rowid >= 'AAATCjAA0AABn4LAA3' and rowid < 'AAATCjAA0AAB3pdAAh''
    and rowid >= 'AAATCjAA0AACrFuAAW' and rowid < 'AAATCjAA6AAAXpOAAq''
    and rowid >= 'AAATCjAA6AAA8CZAAO' and rowid < 'AAATCjAA6AABLAYAAj''
    and rowid >= 'AAATCjAA6AAAXpOAAq' and rowid < 'AAATCjAA6AAA8CZAAO''
    and rowid >= 'AAATCjAA6AABLAYAAj' and rowid < 'AAATCjAA6AABlwbAAg''
    and rowid >= 'AAATCjAA6AABlwbAAg' and rowid < 'AAATCjAA6AACBEoAAM''
    and rowid >= 'AAATCjAA6AACBEoAAM' and rowid < 'AAATCjAA6AACCYGAA1''
    and rowid >= 'AAATCjAA6AACCYGAA1' and rowid < 'AAATCjAA6AACKfBABI''
    and rowid >= 'AAATCjAA6AACKfBABI' and rowid < 'AAATCjAA6AACe0cAAS''
    and rowid >= 'AAATCjAA6AACe0cAAS' and rowid < 'AAATCjAA6AAFmytAAf''
    and rowid >= 'AAATCjAA6AAF6RAAAQ' and rowid < 'AAATCjAA6AAHJjDAAV''
    and rowid >= 'AAATCjAA6AAFmytAAf' and rowid < 'AAATCjAA6AAFp+bAA6''
    and rowid >= 'AAATCjAA6AAFp+bAA6' and rowid < 'AAATCjAA6AAF6RAAAQ''
    and rowid >= 'AAATCjAA6AAHJjDAAV' and rowid < 'AAATCjAA6AAIR+jAAL''
    and rowid >= 'AAATCjAA6AAIR+jAAL' and rowid < 'AAATCjAA6AAKomNAAE''
    and rowid >= 'AAATCjAA6AAKomNAAE' and rowid < 'AAATCjAA6AALdcMAA3''
    and rowid >= 'AAATCjAA6AALdcMAA3' and rowid < 'AAATCjAA9AAACuuAAl''
    and rowid >= 'AAATCjAA9AAACuuAAl' and rowid < 'AAATCjAA9AABgD6AAD''
    and rowid >= 'AAATCjAA9AABgD6AAD' and rowid < 'AAATCjAA9AADiA2AAC''
    and rowid >= 'AAATCjAA9AADiA2AAC' and rowid < 'AAATCjAA9AAEQMPAAT''
    and rowid >= 'AAATCjAA9AAEQMPAAT''
    33 rows selected.
    SQL> select count(*) from BIGTABLE where  1=1 and rowid  <  AAATCjAA0AAAKAVAAd ;
      COUNT(*)
        518712
    SQL> select count(*) from BIGTABLE where  1=1 and rowid  >= 'AAATCjAA9AAEQMPAAT'' ;
      COUNT(*)
       1846369Nice but not accurate....

    Yes it works as intended now : +( still this annoying issue of quotes, so some rowid appear without enclosing quotes)+
    from (select rn, rownum frow, mod(rownum, 2000000) mod
            from  (select rowid rn from BIGTABLE  order by rn)
            order by rn
    where mod = 0
    SQL> /
    RN                        MOD       FROW     ROWNUM
    AAATCjAA0AAAVNlAAQ          0    2000000          1
    AAATCjAA0AAAlxyAAS          0    4000000          2
    AAATCjAA0AAA2CRAAQ          0    6000000          3
    AAATCjAA0AABFcoAAn          0    8000000          4
    AAATCjAA0AABVIDAAi          0   10000000          5
    AAATCjAA0AABoSEAAU          0   12000000          6
    AAATCjAA0AAB3YrAAf          0   14000000          7
    AAATCjAA0AACE+oAAS          0   16000000          8
    AAATCjAA0AACR6dAAR          0   18000000          9
    AAATCjAA0AACe8AAAa          0   20000000         10
    AAATCjAA0AACt3CAAS          0   22000000         11
    AAATCjAA6AAAPXrAAT          0   24000000         12
    AAATCjAA6AAAgO4AA5          0   26000000         13
    AAATCjAA6AAAwKfAAu          0   28000000         14
    AAATCjAA6AABAQBAAH          0   30000000         15
    AAATCjAA6AABREdAA9          0   32000000         16
    AAATCjAA6AABhFIAAT          0   34000000         17
    AAATCjAA6AABxyZAAj          0   36000000         18
    AAATCjAA6AACA5CAAm          0   38000000         19
    AAATCjAA6AACNJBAAN          0   40000000         20
    AAATCjAA6AACbLgAAV          0   42000000         21
    AAATCjAA6AACoukAAD          0   44000000         22
    AAATCjAA6AAFsS8AAO          0   46000000         23
    AAATCjAA6AAF36JAAa          0   48000000         24
    AAATCjAA6AAHJzoAAv          0   50000000         25
    AAATCjAA6AAKMCHAAv          0   52000000         26
    AAATCjAA6AAL2RbAAT          0   54000000         27
    AAATCjAA9AABLLbAAH          0   56000000         28
    AAATCjAA9AACmSyAAA          0   58000000         29
    AAATCjAA9AAEO/nAAe          0   60000000         30
    AAATCjAA9AAEbC7AAI          0   62000000         31
    31 rows selected.
    SQL> select count(*) cpt from BIGTABLE where rowid < AAATCjAA0AAAVNlAAQ ;
           CPT
       1999999
    SQL> select count(*) cpt from BIGTABLE where rowid >= 'AAATCjAA6AAAgO4AA5' and rowid < AAATCjAA6AAAwKfAAu ;
           CPT
       2000000

  • Problem insert JDBC Tables...

    hello all !
    I`m Sorry by open a new Thread about this problem ...but is very urgent !
    I have Two problems..
    1 - I need to insert in SQLServer 4 tables CRHE, CRFT, CRFV, TLINES !
    CHRE is only one data - Father Table ! its OK !
    The tables CRFT, CRFV exists many VML segments, but insert only one line in database ..only the first XML elements is inserted ...How a can to make to insert all xml segments for these tables !
    2 - I have problem is TDLINE Table --> i need to concatenate the items of TDLINE ...If you to analyze 3 first fields CRID, FTNO, FVNO are equal because of the key ok?
    The Last field TDLINE was broken in segments of the XML and I need to concatenate this fields to insert an one only line of the database !
    this is the XML come from R/3:
    thank you very much for help !
    Boni.

    Hi !
    Yes ..multiple items in RFC ...to insert all in DataBase !
    My problem is that insert only the first item ...hoe i do to insert all???
    I change the structure of XML Receiver Structure  !
    This is my complete XML Receiver Structure:
    statement_TABELA2           1 (ocur)
    CRFT                        O..unbounded
    action(Insert)              optional
    access                      0..unbounded
    CRID                        0..1
    FTNO                        0..1
    FTTYP                       0..1
    COSTR                       0..1
    VORNR                       0..1
    FTRNR                       0..1
    This is Rigth ??? ...you can Show me how i can make this !!
    Thank You...

  • Big tables from MySql to MSSql error Connection timeout

    HI,
    I have a big database around 46GB in Mysql format and I managed to convert all the database to MSsql except two tables, the biggest ones. When I try to migrate those 2 tables, one by one , after a while I get the error message "Connection timeout and
    was disabled"
    I encreased the timeout from SSMA option from 15 to 1440 and decreased the basc from 1000 to 500 and same thing, The tables have 52 mil rows and 110 milion rows with 1,5 GB and 6.5 GB.
    What can I do to migrate them
    Thank You

    Hi,
    According to your description, we need to verify that if you have installed the latest version of
    MySQL ODBC driver. If you have installed it, in order to make a shorter duration transaction and avoid the timeout, you can try to reduce the batch size to lower value such as 200 or 100 in
    SSMA option. Also, as Raju’s post, you can try to use incremental data migration in SSMA to migrate lager tables, then check if it is successful.
    In addition, you can use other methods to migrate big tables from MySQL to SQL Server. For example, you can copy the data directly from SQL Server using OpenQuery, and you could include WHERE clause to limit the rows. For more details, please review this
    blog :Migrate MySQL to Microsoft SQL Server. Or you can write queries for MySQL to export your data as csv, and then use the
    BULK INSERT features of SQL Server to import the csv data.
    Thanks
    Lydia Zhang

Maybe you are looking for

  • The MacBook pro, enough to edit AVCHD

    Ok, I am in the market to upgrade to a new Mac. The Mac I have now is a desktop, and it's not very powerful. I would get a desktop but I edit at friends houses a lot. So will the new 13" be enough with 2.7GHz dual-core Intel Core i7 processor with 4M

  • No video playback to external FS device 6.03

    Hello. I have read all the posts about not seeing video playback on my external device (TwinPact 100/FW). I can see the current frame on my external monitor so I know the connection is good. But the image does not scroll when I playback from the time

  • GRC 10.1 Access Request - Provisioning Logs Not Available

    Hello guys, I am currently running into an issue with the user provisioning logs, the Request Approval notification which is sent to the user are at the end of an approved access request are as below and the Provisioning Logs tab is throwing a timeou

  • How to represent java generics in UML

    Hello, Can anyone tell me how to represent java generics in UML? Say I have ClassA<ClassB>. How do I represent this in UML? Thanks in advance, Julien Martin.

  • Where to download BR*tools Studio 7.10?

    Hello experts, I am trying to download and install BR*Tool Studio 7.1 for our Oracle databases in our SAP environment but I cannot find the download anywhere. Please someone point me to the download link. Do I need to have license to run this softwar