Performance too Slow on SQL Azure box

Hi,
Performance is too slow on SQL Azure box:
Below query returns 500,000 rows in 18 Min. on SQL Azure box (connected via SSMS)
SELECT * FROM TABLE_1
Whereas, on local server it returns 500,000 rows in (30 sec.)
SQL Azure configuration:
Service Tier/Performance Level : Premium/P1
DTU       : 100
MAX DB Size : 500GB     
Max Worker Threads : 200          
Max Sessions     : 2400
Benchmark Transaction Rate      : 105 transactions per second
Predictability : Best
Thanks,

Hello,
Please refer to the following document too:
http://download.microsoft.com/download/D/2/0/D20E1C5F-72EA-4505-9F26-FEF9550EFD44/Performance%20Guidance%20for%20SQL%20Server%20in%20Windows%20Azure%20Virtual%20Machines.docx
Hope this helps.
Regards,
Alberto Morillo
SQLCoffee.com

Similar Messages

  • Performance is too slow on SQL Azure box

    Hi,
    Performance is too slow on SQL Azure box (Located in Europe)
    Below query returns 500,000 rows in 18 Min. on SQL Azure box (connected via SSMS, located in India)
    SELECT * FROM TABLE_1
    Whereas, on local server it returns 500,000 rows in (30 sec.)
    SQL Azure configuration:
    Service Tier/Performance Level : Premium/P1
    DTU       : 100
    MAX DB Size : 500GB     
    Max Worker Threads : 200          
    Max Sessions     : 2400
    Benchmark Transaction Rate      : 105 transactions per second
    Predictability : Best
    Any suggestion would be highly appreciated.
    Thanks,

    Hello,
    Can you please explain in a little more detail the scenario you testing? Are you comparing a SQL Database in Europe against a SQL Database in India? Or a SQL Database with a local, on-premise SQL Server installation?
    In case of the first scenario, the roundtrip latency for the connection to the datacenter might play a role. 
    If you are comparing to a local installation, please note that you might be running against completely different hardware specifications and without network delay, resulting in very different results.
    In both cases you can use the below blog post to assess the resource utilization of the SQL Database during the operation:
    http://azure.microsoft.com/blog/2014/09/11/azure-sql-database-introduces-new-near-real-time-performance-metrics/
    If the DB utilizes up to 100% you might have to consider to upgrade to a higher performance level to achieve the throughput you are looking for.
    Thanks,
    Jan 

  • BI Server performance too slow

    Hi experts,
    I am facing an issue, my OBIEE sits on a Unix Platform. And many time the server performance is too slow, when I check the processes, Java under Orabi user is using more than 50% of the CPU utilization. I need to find the root cause behind it, as it is the production system and is affecting many users.
    Thanks in advance.
    MT

    Thanks Prassu,
    Yes the Server is performing too much calculations, but they are required as need to cache reports at daily level also,
    I have an observation that the MAX_CACHE_ENTRY_SIZE [MAX_CACHE_ENTRY_SIZE = 1 MB;] parameter is set only for 1 MB, I can understand that if any query fetching more data than 1 MB will not be cached. But is there any chance that this will affect the server as well.
    Thanks
    MT

  • Help me! AWM "maintain cube ... " performance: too slow with a small cube

    Hi all,
    I have 3 dimensions: TIMES, PRODUCTS, STORES
    The hierarchy of TIMES, is in the order: year, quarter, month, week, day
    The hierarchy of Products is in the order: sector, family, sub-family, product
    The hierarchy of STORES is in the order: area, department, store
    each hierarchy is mapped with a data set and the size is
    730 lines (days) for the TIMES dimension
    282 lines (products) for the PRODUCTS dimension
    237 lines (stores) for the STORES dimension
    With these dimensions, I create a cube and 2 measures: QUANTITY, BENEFIT
    My measures used all 3 dimensions : TIMES, PRODUCTS, STORES.
    I mapped in the end my cube with a dataset contains: store, day, product, quantity, benefit
    This data set has 3921 lines.
    I chosed the rule to aggregate all dimensions is SUM.
    So when I do "maintain cube...", AWM takes much time to do it! (approximately 2h to maintain my cube)
    Some one knows why it's long?
    How can I do to accelerate it?
    Due I modify the parameters in AWM or in my cube?
    Thanks in advance!

    thanks all for your reply!
    I done maintain cube already in following your advice. I increased the rollback and tmp space and changed "NUMBER to DECIMAL" as type of data of cube.
    But I have other question. I hope you can help me!
    1. What's the effect of the option "Choose the regions of the cube to be presummarized and stored in the analytic workspace" in the table "Summarize To" when you create a cube? If I check this option for all hierarchy level, the next time, when I do the request,
    the time of calcul will be optimize so fastest?
    (I check this option in creating my cube but when I make a request for level of a particular hierarchy, it's so long to have the result. I use a software this can connect and do request from Oracle Olap DataBase.)
    2. I done maintain cube last week. Today I created a new cube and added more data set. But I cant maintain cube again. This is the message error. Can you explain to me why?
    ( I increase already the rollback and tmp space. To know how and what I done, view this thread please!
    Error when I do "Maintain Cubes ..." )
    --------------------------------------java.sql.SQLException: ORA-00933: SQL command not properly ended
    at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:111)
    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:330)
    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:287)
    at oracle.jdbc.driver.T4C8Oall.receive(T4C8Oall.java:742)
    at oracle.jdbc.driver.T4CStatement.doOall8(T4CStatement.java:206)
    at oracle.jdbc.driver.T4CStatement.executeForDescribe(T4CStatement.java:789)
    at oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:1030)
    at oracle.jdbc.driver.T4CStatement.executeMaybeDescribe(T4CStatement.java:829)
    at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1123)
    at oracle.jdbc.driver.OracleStatement.executeInternal(OracleStatement.java:1678)
    at oracle.jdbc.driver.OracleStatement.execute(OracleStatement.java:1644)
    at oracle.olap.awm.dataobject.dialog.olapcatalog.TableDialog.populateTableDataModel(TableDialog.java:164)
    at oracle.olap.awm.dataobject.dialog.olapcatalog.TableDialog.initialiseAndPopulate(TableDialog.java:68)
    at oracle.olap.awm.dataobject.dialog.olapcatalog.TableDialog.<init>(TableDialog.java:57)
    at oracle.olap.awm.wizard.awbuild.BuildWizardHelper.runBuild(BuildWizardHelper.java:252)
    at oracle.olap.awm.navigator.node.modeler.cube.ModelerMeasureNode.actionPerformed(ModelerMeasureNode.java:226)
    at javax.swing.AbstractButton.fireActionPerformed(Unknown Source)
    at javax.swing.AbstractButton$ForwardActionEvents.actionPerformed(Unknown Source)
    at javax.swing.DefaultButtonModel.fireActionPerformed(Unknown Source)
    at javax.swing.DefaultButtonModel.setPressed(Unknown Source)
    at javax.swing.AbstractButton.doClick(Unknown Source)
    at javax.swing.plaf.basic.BasicMenuItemUI.doClick(Unknown Source)
    at javax.swing.plaf.basic.BasicMenuItemUI$MouseInputHandler.mouseReleased(Unknown Source)
    at java.awt.Component.processMouseEvent(Unknown Source)
    at java.awt.Component.processEvent(Unknown Source)
    at java.awt.Container.processEvent(Unknown Source)
    at java.awt.Component.dispatchEventImpl(Unknown Source)
    at java.awt.Container.dispatchEventImpl(Unknown Source)
    at java.awt.Component.dispatchEvent(Unknown Source)
    at java.awt.LightweightDispatcher.retargetMouseEvent(Unknown Source)
    at java.awt.LightweightDispatcher.processMouseEvent(Unknown Source)
    at java.awt.LightweightDispatcher.dispatchEvent(Unknown Source)
    at java.awt.Container.dispatchEventImpl(Unknown Source)
    at java.awt.Window.dispatchEventImpl(Unknown Source)
    at java.awt.Component.dispatchEvent(Unknown Source)
    at java.awt.EventQueue.dispatchEvent(Unknown Source)
    at java.awt.EventDispatchThread.pumpOneEventForHierarchy(Unknown Source)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
    at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
    at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
    at java.awt.EventDispatchThread.run(Unknown Source)
    Thanks in advance!

  • Slow start up since wire got removed accidentally...imac perform too slow

    hi am sajith ...i am facing a slow in start up n using my mac from the time since, the wire got unplug accidentally ...it ws very fast before ..please help me out?

    Open Disk Utility (in the Utilities folder of the Application folder). Select your startup volume from the list on the left side & click the "Verify Disk" button. If it finds any problems follow the instructions to restart your computer from another volume (varies by the OS version you are using) & repair the drive.
    If & when any errors are fixed, run the Permissions repair.
    Note: If you fill in the model & OS version info in your user profile, that will help us provide more detailed advice.

  • Database is performing to slow

    Hi all,
    I have oracle 8i database with only 5GB data. My database is performing too slow. Cal any one tell me step by step which poction and where i have to look for increse the performance. Plz provide me some steps which i should follow.
    Thanks & Regards

    My database is performing too slowA very generic statement which is actually misleading.
    There could be a particular application/module which is performing slow, in turn slowing down the whole system. There could be some SQL that some user has fired, taking up all resources on database server (f.ex. cartesian join on seven tables as the developer thought that was a simple SQL before he could submit a more "complex" SQL).
    If you dont know where to start then you need to setup statspack in first place and get the snaps running at say 15 minutes interval. Then analyze the statspack reports to get a better picture of the database performance.
    If you know that after plugging in a particular module, your database is going down in terms of performance, then you can take it out, review the same, trace it, debug it and fix it.
    At the end I would go in for reviewing the instance parameters and tune them. If even after tuning the instance parameters, you are facing the performance problem, then I would review the hardware.
    That, in short, would be my plan of action.
    Message was edited by:
    Satish Kandi
    And I am slow in writing the same ;)

  • Select Data from aufm too slow

    Hi experts,
    I have a query in my report but its performance too slow.
    is there any proper way or any way to improve the performance of below query.
    SELECT  * FROM  aufm     into table i_aufm          " collecting input material document numbers
             WHERE mblnr GE '4900000000'
               AND mjahr GE '2008'
               AND zeile GT '0'
               AND matnr EQ matnr_101-matnr
               AND werks IN p_werks
               AND charg EQ matnr_101-charg
               AND bwart = '101'.

    Hi,
    1 .Dont use * from any table.
    2. Create  a internal table  that u want to retrieve the data from that table. (Structure).
    Eg:
    types : begin of str_ekko,
               ebeln type ekko-ebeln,
               bsart type ekko-bsart,
               aedat type ekko-aedat,
               endof str_ekko.
    data : itab type table of str_ekko,
            wa_itab type str_ekko.
    In the above u want to take care that while creating the structure u want to look the sequence of fields in the table. in the same sequence u want to create the structure.
    3 . The Select-options fields also wants be in the same field sequence of the table.
    eg : select ebeln bsart aedat from ekko into table itab where ebeln in so_ebeln and                                                                               
    bsart in so_bsart and
                                                                                    aedat in so_aedat.
    4.  If the above things are not working just look the indexes of that table. Just create a secondary index with ur requirement.

  • Why is performance so slow reading binary data from a SQL Azure DB with EF6.x

    I'm running a WPF client that hits a SQL Azure DB using EF 6.x. For the most part, everything seems to be working fine. The one exception is when I try to read a large binary column.
    I am storing files in the DB as a binary column.  When I test using the local DB, everything sings.  When I switch to the Azure DB, I get timeouts when I try to read the file contents.  I have no problem saving the binary data to the DB, just
    reading it.
    I don't know how to troubleshoot this.  I looked at the Query Performance page in the Azure portal, but it doesn't time stamp anything in there and you can't clear it, so I can't correlate what's running with the queries that show up there.
    I tried to start SQL profiler against the DB, but was denied because I'm not a member of the sysadmin fixed server role.
    If I query for the data directly, it comes back quickly.  So this seems to be an Azure via EF issue.
    Any help is appreciated.
    http://digitalcamel.blogspot.com/

    Hi Digital Camel,
    Since I don't know what your scenario is, I won't argue too much about not storing binaries in your SQL DB, but still: don't store binaries in your SQL DB :). The main reason is simple: first and foremost, in both the current and future pricing tier your
    levels are defines on the size of the DB. Basically, you pay way more by storing your binaries on your SQL layer rathern than storing them elsewhere, such as Azure Storage. Second, the protocol your binaries would be downloaded over the wire is prone to network
    connectivity issues: you could use HTTP(S) or FTP instead, if you'd use Azure Storage. Last but not least, when you download the binary from your DB, you keep a connection open which in the end is a connection other users might have used to query data instead.
    However, in regard to your question, how did you "query for the data directly"? Did you try to query the data using SSMS with the Client Statistics option on? This could tell you if the problem is network, server or client related.
    Hope this helps!
    Alex

  • Bumblebee performance is too slow

    Hi everyone,
    This is my second post in this forum and it has been only 2 days that I met Arch. Before, I was using Ubuntu for 3 years, but due to low performance in my PC, unfortunately, I decided to say good bye which was hard to say.
    Now, I am trying to have the same setup as my previous laptop and Bumblebee was one of them. I followed the instructions here: https://wiki.archlinux.org/index.php/Bumblebee.
    It seems that bumblebee is installed and working, however, the FPS is too slow:
    $ optirun glxspheres64 -info
    Polygons in scene: 62464
    Visual ID of window: 0x20
    Context is Direct
    OpenGL Renderer: GeForce GT 520MX/PCIe/SSE2
    0.023848 frames/sec - 0.021114 Mpixels/sec
    Without optirun I get:
    $ glxspheres64 -info
    Polygons in scene: 62464
    Visual ID of window: 0x20
    Context is Direct
    OpenGL Renderer: Mesa DRI Intel(R) Sandybridge Mobile
    0.033237 frames/sec - 0.029426 Mpixels/sec
    0.029968 frames/sec - 0.026533 Mpixels/sec
    This is impossible as I was getting very good results before.
    I am wondering if I did something wrong, or missed anything.
    Just for information, system specs:
    Intel i7 2670QM 2.2 GHZ
    4 GB RAM
    1 GB GeForce GT 520MX
    512 MB Intel Graphics
    I played 0ad with and without optirun and the performance was good in both of them, but not sure if it switches the video cards itself.
    I also have bbswitch installed.
    Any help would be appreciated. Thank you.
    Last edited by wakeup12 (2014-09-27 14:29:19)

    Hello,
    Can you please explain in a little more detail the scenario you testing? Are you comparing a SQL Database in Europe against a SQL Database in India? Or a SQL Database with a local, on-premise SQL Server installation?
    In case of the first scenario, the roundtrip latency for the connection to the datacenter might play a role. 
    If you are comparing to a local installation, please note that you might be running against completely different hardware specifications and without network delay, resulting in very different results.
    In both cases you can use the below blog post to assess the resource utilization of the SQL Database during the operation:
    http://azure.microsoft.com/blog/2014/09/11/azure-sql-database-introduces-new-near-real-time-performance-metrics/
    If the DB utilizes up to 100% you might have to consider to upgrade to a higher performance level to achieve the throughput you are looking for.
    Thanks,
    Jan 

  • SQL Azure very slow

    Hi Experts
    We have SQL Azure database that is running very slow, so i took a backup of the database and restored on my on premise sql server express edition. 
    I tried a update statement on a logs tables that have approximately about 1,062,367 records.
    UPDATE logs set FileId = NULL
    where LEN(FileId) < 36
    The above update statement updates about 910,593 records on localhost host in 20 sec and if i run the same updates statement on SQL Azure it took 40mins and 16sec. The update statement was issues using SQL Server management studio installed on my local machine.
    With reference to internet connection, we have fiber connection with download speed of 70.62mbps and upload speed of 84.09mbps 
    I am not sure whats going wrong with SQL Azure database i.e. could there be a specific fault on my database at Microsoft side etc.
    Any advice or suggestion will be highly appreciated.
    Kind Regards
    Bhavesh
    Bhavesh

    I have waited for 40 min for the update to complete and it has successfully and updated logs table so i can't re-run the update statement again because there will be no records to update.
    I have managed to find another example that took about 10 mins to complete on the SSMS but when ran the same query on the azure management portal it took only 6 secs which puzzled me. 
    Here is the query: 
    select   
     l.Value3 as ClientReference,
     l.[Key] as ActivityUser,
     l.Value2 as FormDisplayName,
     l.Value6 as Status,
     l.DateStamp as LastUpdated,
     l.[Action] as FileDescription,
     l.CompanyId as ClientID,
     l.FileId as FileID,
     f.id
    from Logs l
    LEFT OUTER JOIN files f on f.Id = l.FileId 
    where l.AccountId = 578
    and type = 1
    The above query returns about 3694 records only.
    Here is the details of the query performance 
    Azure Management Portal
    Duration(ms): 6487
    CPU(ms): 2443
    Logical Reads: 74729
    Physical Reads: 66147
    Logical Writes: 0
    SSMS
    Duration(ms): 615259
    CPU(ms): 3666
    Logical Reads: 74729
    Physical Reads: 71370
    Logical Writes: 0
    Any update will be appreciated.
    Bhavesh

  • Pl/sql block is too slow, should  procedure a better option

    Hi all,
    how to tune A PL/SQL block that traverse cursors and fetch millions of records then execute inserts in different tables,
    using execute immediate statement.
    It's too slow and takes 10 hours to populate 40 tables having millions of records,
    as i have to do some modifications in data so can not do it by CTAS,
    i.e. a single sql statement.
    Should i make a procedure, does it help .
    Please help or suggest As i am New to PL/Sql
    My code look like,
    declare
    cursor     cur_table1 is
         select field1,field2,field3,field4 from table1;
    begin
    for i in cur_table1
    loop
         execute immediate 'insert into table2 (field1,field2,field3,field4) '||
    'select :1,field2,field3,field4 '||
    ' from table1 where field3= :2'
    using i.field1||'_'||to_char(sysdate,'ddmmyyyy hh12:mi:ss',i.field1;
    commit;
    end if;
    end;
    Thanks and Regards,

    declare
    cursor cur_projects is
         select PROJECTID, PROJECTNAME, DESCRIPTION, DELETED, DELETINGDATE, ACTIVE, ADMINONLY, READONLY, SECURITYCLASS, PROJECTCONTACT, DEFAULTVERSION, DEFAULTSTARTPAGE, IMAGEPATH, MAXEXAMINEERRORS, LOCKTIMEOUT, MEMORYSAVINGLEVEL, PRELOADOBJECTS, PUBLICATIONSRCPROJNAME, CREATOR, CREATED, MODIFIER, MODIFIED from projects ;
    cursor cur_projectversion(p_projectid projects.projectid%TYPE) is
         select PROJECTID, PROJECTVERSIONID, PROJECTVERSIONNAME, DESCRIPTION, DELETED , DELETINGDATE, ACTIVE , ADMINONLY, READONLY, decode(EFFECTIVEDATE,null,trunc(sysdate),EFFECTIVEDATE) EFFECTIVEDATE, EXPIRATIONDATE, SECURITYCLASS, PROJECTCONTACT, DEFAULTVERSION, DEFAULTSTARTPAGE, IMAGEPATH, MAXEXAMINEERRORS, LOCKTIMEOUT, MEMORYSAVINGLEVEL, PRELOADOBJECTS, PUBLICATIONSRCPROJNAME, PUBLICATIONSRCPROJVERNAME, CREATOR, CREATED, MODIFIER, MODIFIED, PROFILELOADERCLASS /*, TRACKCHANGES */
         from projectversions where PROJECTID=p_projectid ;
    cursor cur_objects(p_projectid projects.projectid%TYPE,p_projectversionid projectversions.projectversionid%TYPE) is
         select PROJECTID , PROJECTVERSIONID, OBJECTID , OBJECTKEY , PARENTID, KIND , NAME , TITLE , OWNER , CREATED, MODIFIER , MODIFIED , READY_TO_PUBLISH, LAST_PUBLISHED_DATE , LAST_PUBLISHER , EFFECTIVE_PUBLISHING_DATE , PUBLISHER , PUBLISHING_DATE /*, to_lob(scripttext) */ from OBJECTS where PROJECTID=p_projectid and PROJECTVERSIONID=p_projectversionid /*order by objectid */;
    begin
    for i in cur_projects
    loop
    dbms_output.put_line('PROJECTID => '||i.projectid);
    dbms_output.put_line('_________________________________');
    execute immediate 'insert into &TARGET_USER\.projects(locktimeout, memorysavinglevel , preloadobjects, projectid, projectname, description, deleted, deletingdate, active, adminonly, readonly, securityclass, projectcontact, defaultversion, defaultstartpage, imagepath, maxexamineerrors ) values (:1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13,:14,:15,:16,:17) '
    using i.locktimeout, i.memorysavinglevel, i.preloadobjects,i.projectid ,i.projectname , i.description , i.deleted , i.deletingdate , i.active , i.adminonly , i.readonly, i.securityclass, i.projectcontact , i.defaultversion, i.defaultstartpage , i.imagepath, i.maxexamineerrors;
    for k in cur_projectversion(i.projectid)
         loop
    for l in cur_objects(k.projectid,k.projectversionid)
              loop
                   cnt:=cnt+1;
    select count(1) into object_exists from &TARGET_USER\.objects where objectid=l.objectid and projectversionid=1 and projectid=l.projectid;
              if object_exists = 0
              then
              if l.objectid = 1 ------Book Object , objectid = 1 and parentid = 0
              then
              execute immediate 'INSERT INTO &TARGET_USER\.objects(PROJECTID,PROJECTVERSIONID,OBJECTID, OBJECTKEY,PARENTID,NAME, KIND,LAST_PUBLISHED_DATE,LAST_PUBLISHER,REVISIONID,DISPLAYORDER,READONLY,DELETED) values( :1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13)'
                        using l.PROJECTID, 1, l.OBJECTID,l.OBJECTKEY, 0 , l.NAME,l.KIND, '' , '' , '', 0, 'N', 'N';
                   else
                        select count(1) into object_parentid_exists from objects where objectid=l.parentid and projectversionid=1 and projectid=l.projectid;
                        if object_parentid_exists = 0 ---Set Parentid as 1
                        then
                                  cnt_parentid_1:=cnt_parentid_1+1;
                                  execute immediate 'INSERT INTO &TARGET_USER\.objects(PROJECTID,PROJECTVERSIONID,OBJECTID, OBJECTKEY,PARENTID,NAME, KIND,LAST_PUBLISHED_DATE,LAST_PUBLISHER,REVISIONID,DISPLAYORDER,READONLY,DELETED) values( :1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13)'
                                  using l.PROJECTID, 1, l.OBJECTID,l.OBJECTKEY, 1 , l.NAME,l.KIND, '' , '' , '', 0, 'N', 'N';
                        else
                                  execute immediate 'INSERT INTO &TARGET_USER\.objects(PROJECTID,PROJECTVERSIONID,OBJECTID, OBJECTKEY, PARENTID, NAME, KIND,LAST_PUBLISHED_DATE,LAST_PUBLISHER,REVISIONID,DISPLAYORDER,READONLY,DELETED) values( :1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13)'
                                  using l.PROJECTID, 1, l.OBJECTID,l.OBJECTKEY,l.PARENTID,l.NAME,l.KIND, '' , '' , '', 0, 'N', 'N';
                        end if;
                   end if ;
         end if;
                   execute immediate 'INSERT INTO &TARGET_USER\.objectversions( PROJECTID, OBJECTID, PROJECTVERSIONID ,VERSIONNAME,OBJECTVERSIONID, REVISIONID,DESCRIPTION, TITLE , OWNER, CREATED, MODIFIER, MODIFIED, READY_TO_PUBLISH , LAST_PUBLISHED_DATE, LAST_PUBLISHER, EFFECTIVEDATE, SCRIPTTEXT, REVIEWSTATUS, READONLY, PUBLISHED, DELETED ) '||
                             'SELECT PROJECTID, OBJECTID, 1, owner||:1, PROJECTVERSIONID , '''', '''', TITLE, OWNER, CREATED, MODIFIER, MODIFIED, ''N'', '''' , '''', :2 , to_lob(SCRIPTTEXT), '''', ''N'', ''N'', '''' '||
                             'FROM OBJECTS '||
                             'WHERE PROJECTID= :3 and PROJECTVERSIONID= :4 and OBJECTID= :5'
                             using '_'||TO_CHAR(k.EFFECTIVEDATE,'DDMMYYHHMISS'),k.EFFECTIVEDATE,l.projectid,l.projectversionid,l.objectid;
         end loop;
         dbms_output.put_line(cnt||' OBJECTS, OBJECTVERIONS POPULATED');
         dbms_output.put_line(cnt_parentid_1||' DUMPED UNDER BOOK FOLDER ');
         cnt_parentid_1:=0;
         cnt:=0;
    ............

  • SQL Developer 2.1 working too slow

    Hi All,
    I am working on SQL Developer 2.1 after 3 years previously I used version 1.2. Comparatively so many changes came in 2.1 version but it is too slow and while debugging variable values not shows in Tooltip we need to depend on Smartdata tab. If any settings or patches available for following problem than please provide me.
    1.     Too slow
    2.     Variable values in Tooltip
    By
    Srinivas M. P.

    SQL Developer using 127232 KB, Free 1.32 GB
    Is there any setting is SQL Developer or you talking about my system memory.
    If you talking about my system memory than I am using 2 GB RAM in my system and system working well
    If I do anything in TOAD than TOAD is working fine but SQL Developer working too slow.
    Edited by: SrinivasMP on Feb 5, 2010 3:34 PM

  • My Dual 1.8 G5 too slow on color performances

    I send my FCP project to color but unfortunately my dual 1.8 with 3.5 ram, nvidia 6600 open gl won't perform at its best. I know for color application XT1900 is recommended but please the base recommended specs showed that i did have the baseline requirements.
    Any idea how to get my dual 1.8 go a little bit faster. adding more ram or what? Tried to put color FX into 20 secs clip takes 30 to 1 mins. Wow its way too slow though. Try to play the clip also takes more secs to mins to runs. I'm wondering why this happen and it should can go just a bit faster. Any help/advice much appreciated

    Oh ya i knew it. I owned this PMG5 since it first released and planning to buy the best and good one but not affordable. I love this machine but future technology is highly in demand. I just updated the graphic card from its original to the latest i have and pretty much can go to started the color app but way to far to execute a full project. Eager to buy the Mac Pro but have wait though. huh huh!
    Runs benchmark for nvidia GeFOrce 6600 and scored at 5326 - 400FPS. Not that i know, just did some test and open few applications simultoneously with color but got frustrated because its hang for a mins to runs on every apps.
    Thank you for your opinion and obviously my machine is going to obsolete in the near future. It can go for small graphic applications and final cut pro but FCP seems to slow either, render clips for 1 - 2 mins with transitions and fx would go for 30 mins to 40 mins. Harghhhh! i must get the new one!

  • SQL Azure - query with row_number() executes slow if columns with nvarchar of big size are included

    I am linking my question from Stack Overflow here. The link: http://stackoverflow.com/questions/27943913/sql-azure-query-with-row-number-executes-slow-if-columns-with-nvarchar-of-bi
    Appreciate your help!
    Gorgi

    Hi,
    Thanks for posting here.
    I suggest you to check this link and optimize your query on sql azure.
    http://www.sqlusa.com/articles/query-optimization/
    http://sqlblog.com/blogs/paul_white/archive/2011/02/23/Advanced-TSQL-Tuning-Why-Internals-Knowledge-Matters.aspx
    Also check this blog which had similar issue.
    https://social.msdn.microsoft.com/Forums/en-US/c1da08b4-265d-4ec8-a252-8d7090234e3e/simple-select-query-takes-long-time-to-execute-with-nvarchar-columns?forum=transactsql
    Girish Prajwal

  • I am using iphone 4s ios6.1.2 ...my iphone has become too slow once i have updated from ios 5.1.1..can anyone help me to make my iphone perform batter

    i am using iphone 4s ios6.1.2 ...my iphone has become too slow once i have updated from ios 5.1.1..can anyone help me to make my iphone perform batter

    Standard troubleshooting...
    Try a reset by pressing the home and sleep buttons until you see the Apple logo, ignoring the slider. Takes about 5-15 secs of button holding and you won't lose any data or settings.
    Remove all apps from Recently Used list...
    - From any Home Screen, double tap the home button to bring up the Recents List
    - Tap and hold any icon in this list until they wiggle
    - Press the red  to delete apps from this list.
    - Press the home button twice when done.
    If still a problem restore with your backup.
    If still a problem restore as new, i.e. without your backup. See how it runs with nothing synced to it.
    If still a problem, it's likely a hardware issue.

Maybe you are looking for